Next Article in Journal
Sliding-Mode Speed Control of PMSM with Fuzzy-Logic Chattering Minimization—Design and Implementation
Previous Article in Journal
Multi-Criteria Vertical Handover Comparison between Wimax and Wifi
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On Semantic Information in Nature

Frankfurt School of Finance & Management, Sonnemannstr. 9, 60314 Frankfurt a. M., Germany
Private address: Wilhelm Busch Str. 23, 64625 Bensheim, Germany.
Submission received: 29 May 2015 / Revised: 5 July 2015 / Accepted: 15 July 2015 / Published: 27 July 2015
(This article belongs to the Section Information Theory and Methodology)

Abstract

:
The connection between semantic information and evolution has gained growing attention recently. Evolution in this contribution—as in others before—we consider as being driven by information. Semantic information, as we consider it, is based on energy. It follows syntactic and semantic rules. We assume syntax, semantics and pragmatics to be structural features of information in biological evolution. These features started to evolve with the very beginning of life and have become more complex and richer in the course of the unfolding biological evolution across all species. We argue that semantic information is an exclusive feature of biological evolution. We present an information model covering this which to a certain degree—it does not cover quantitative aspects—complements Shannon’s information theory and opens novel views on informational based processes in nature.

Graphical Abstract

1. Introduction

Understanding information and its place in nature has been subject to various approaches in recent years (see below). Shannon already provided an “Information Theory” for certain aspects of information [1]. However, from the beginning, and explicitly intended so by Shannon and his coworker in Information Theory, Warren Weaver, semantics was left out of considerations. Weaver stated that information in Information Theory has a different meaning to information in common usage [2].
Information Theory provides a way of quantification of information by statistical means. Its introduction by Shannon was a bigger leap forward than anticipated in his time. Information Theory, not only helped to substantiate the so far somewhat fuzzy concept of information, but also made itself applicable far beyond its original field of communication technology.
Shannon’s Information Theory, by intention, is restricted to model quantitative aspects of information transmitted via a communication channel. The complementary field of investigation is meaningful, i.e., semantic information (unless stated otherwise, in the following “information” is to be understood as “semantic information”).
Since then, numerous approaches have been made to characterize or define what semantic information is. Views have been established that cover all aspects of information from communication to structures, culture, psychology, etc. In addition, some aspects of quantum mechanics gave reasons to consider the role of information in processes and models on the smallest scale. The famous and intensively discussed measurement problem in quantum physics hinted to a special role of the (human) mind in establishing results of experiments. This led the eminent physicist Wheeler [3] to a hypothesis he coined “It from Bit”, which indicates that reality comes from information and is established on information which then is the fundamental constituent of nature.
In parallel, progress was made in genetics and in biochemistry and insights were gained into the processing of information on different levels of an organism (i.e., eukaryotes). Investigations reached from heredity and metabolism to the neuronal functions of the human brain, including impacts of quantum physics. With regard to information, several approaches as early as chemist Eigen’s [4] as well as those of other scientists, and more recently biologist Loewenstein’s [5] work on self-organization, provided specific views on the role of information in organisms.
In this contribution, we present a new comprehensive information model based on the current scientific discourse. Our model complies with, among others, Shannon’s Information Theory, Bateson’s [6] definition and Burgin’s [7] ontology. We take into account the common denominators for all information, which are Syntax, Semantics and Pragmatics (SSP), as defined by semiotician and philosopher Morris [8]. Syntax (the rules according to which words are sequenced to sentences, signs that appear, etc.), semantics (the relationship between a sign or a word to what it is pointing to in reality), and pragmatics (rules for applying words and signs in written or verbal conversations and broader social situations) frame meaning. We define semantic information in nature within a narrow and closely contoured framework. We propose that semantic information is the only existing information in biological nature. Information, in that sense, is a quality of energy—however, not all energy comes with information. More specifically, the potential structure of a given energy is information. It depends on a biological receiver to classify and process received energy as informational structured—or do not. Semantic Information is biological and it is exclusively created and used by biological evolution. We name our information model, which covers this, Evolutionary Energetic Information Model (EEIM).
Our basic assumption is that semantics is ligated to life. Without life, “meaning” is indeed meaningless. Semantic Information as the very base of meaning is part of life, which in turn is the result of biological evolution on earth (we are aware that developments beyond biological evolution are viewed as “evolution”; therefore, if not stated otherwise, in the following with “evolution” we refer to biological evolution on earth). Different branches and paths of evolution possess and process different variations of information. Information is also subject to evolutionary functions of mutation, selection and adaption. We assume syntax is being produced in the process of evolution as are semantics and pragmatics. Syntax in the beginning has materialized as rules imprinted in biochemistry molecules as it is imprinted into logical hardware circuits of computers. Later, the biochemistry hardware became complemented by “software”—also in an analogy to formal language of computer technology (this analogy is not to be extended beyond limits, of course). We further assume that semantics and pragmatics developed with accumulation of information, its syntactic rules and improved processing capabilities of the species’ biological neural network. The degree of similarities across all sectors of biological evolution is striking. Considering, e.g., psychologist and cognition researcher Heyes’ [9] statement that “recent research reveals profound commonalities, as well striking differences, between human and non-human minds, and suggests that the evolution of human cognition has been much more gradual and incremental than previously assumed”, which hints to homogeneous structured developments of cognitive capabilities across all areas of biological evolution. Results differ dramatically but developments themselves show similarities. Another example that includes information processing in evolution is provided by physicist and philosopher Küppers [10], who sees parallels between the “language of genes” and natural languages. Küppers states that “in view of the broad-ranging parallels between the structures of human language and the language of genes, recent years have seen even the linguistic theory of Chomsky move into the center of molecular-genetic research. This is associated with the hope that the methods and formalisms developed by Chomsky will also prove suited to the task of elucidating the structures of the molecular language of genetics.” In this contribution, we do not establish a link to these attempts, however we are aware that it might be created by future research. A further aspect of information’s pervasiveness can be shown when placing biological evolution into the hierarchy of physics. Information processing in an evolutionary context is by no means restricted to the molecular world or to the arena of classical (Newtonian) physics. Physicist and writer Davies [11] states that “However, the key properties of life–replication with variation, and natural selection–does not logically require material structures themselves to be replicated. It is sufficient that information is replicated. This opens up the possibility that life may have started with some form of quantum replicator: Q-life, if you like.” Given the space we have for this contribution, we will not investigate these topics any further.
The more species have taken up information during evolution, the more they are potentially able to bequeath this information in various ways to the next generations to be built upon. By their building up on this, and being exposed to evolutionary mechanisms of mutation, adaption and selection, increasing information complexity is being developed, and with it more extensive information processing capabilities.
A comprehensive view on what semantic information is in nature is still missing. In the following, we will outline an approach to explain this in accordance with Information Theory, with views taken in quantum mechanics and with insights of modern molecular biology.

2. Information

2.1. Models of Information

Comprehensive summaries of models of information are provided by eminent researchers in this field, like Burgin [7], Hofkirchner [12] and Bates [13]. Information Theory, of course, was a major step to clarifying aspects of transmission of signals together with redundancy, noise, and error rates. Hofkirchner argues in favor of a “Unified Theory of Information”, which goes beyond Shannon Information. However, it should succeed in “being recognized by the overall scientific community as a more general Theory of Information that would, in fact, incorporate Shannonian Information Theory” [14,15].
One of the most influential statements regarding the nature of information was made by Bateson who proposed that information is a difference that makes a difference [6]. This statement is to be seen in relation to the most basic difference, a Bit (as invented by Shannon) or “1” versus “0”. The difference Bateson had in mind had to be made by a consciousness being. This is quite different from the view of physicist and philosopher von Weizsäcker, who also considered what makes a “difference” as the core characteristic of information. This difference he called “Ure”—non-matter quantum phenomena—was considered by him as a building block of reality [16].
The connection between information and evolution has also gained more attention recently. Whereas Burgin suggests an “Evolutionary Information Theory” [17], physician Kofler investigates information from an evolutionary point of view [18] and early chemistry stages of evolution are investigated by theoretical physicist Walker [19]. To our best knowledge, the topic has not been looked at yet in an approach that considers syntax, semantics and pragmatics as playing an important role in evolution and evolution of information; neither has a connection been established between this and the role of energy in information and evolution.
Models that consider meaning, that is semantics, often rely in some way on the three-world view the philosopher Popper introduced. The first world is the physical world, the second one the world of conscious experience, and the third world embraces the logical world with books, libraries, etc. It is important to note that Popper thought the third world to cover what he considered to be logic itself, not what is carrying it by means of physical substance [20]. Thus the contents of books are part of this world, the books as physical objects are not.
Bates aligned herself with the view of Wiener who stated “information is information, not matter or energy”. She argues that information is not identical with the physical material that composes it, and she rather assumes information to be the “pattern of organization” of that material, not that material itself. She also proposes various roles information has in biological organisms and differentiates between lines of information flow as, e.g., genetic, on the one hand, and culturally extrasomatic, on the other hand [13]. It is worth noting, that Wiener did not exclude that information could somehow be a quality of energy.
A close relationship between life and information was also proposed by philosopher Min [21] who states that information per se has a capability of self-replication. She argues that replication is organized by organisms by applying a communication channel like in Shannon’s model. The channel is formed by RNA molecules getting input from DNAs and delivering output to proteins at the receiver side. She also states that information is exclusively located in biological organisms. Min, however, does not relate her view to biological evolution.

2.2. Information Theory

Information Theory states that information is measured according to its probability to appear. The logarithm over the sum of the probabilities that individual signs out of a set of signs of allowed signs-an alphabet-will appear, is a measure of the amount of information comprising this sequence of signs.
Shannon’s formula for entropy—the amount of information—equals Boltzmann’s formula for entropy in thermodynamics, besides its being negative. This similarity has ignited discussions and controversies that are still going on today.
The more information a message contains, the more “surprise” it creates at the receiver side. Less surprise represents a high probability of appearance and vice versa. Let us assume we transmit a sequence of Bits. In the case that the specific sequence of length N is known beforehand at the receiver side, obviously no information is transmitted. In the case that each Bit is, independently, equally likely to be 0 or 1, the N Bits of information represent the amount of information measured in Bits. For an arbitrary variant of a sequence of Bits where each individual Bit (1, …, N) appears at the receiver side with probability p(1), p(2), …, p(N), Shannon [1] showed that the resulting entropy (H), that is the resulting amount of information measured in Bits, can be expressed as
H = p ( x ) log p ( x )
Thus entropy can be interpreted as a method to quantify uncertainty being reduced after signs x have been received. In principle, each physical system in a sufficiently stable state can be used to extract and transfer binary—or otherwise structured—information. Binary information, compared to three states or more, has the advantage of simplicity.
Entropy in Shannon’s approach measures the potential information an experimenter can gain by an incoming message at the receiver side; in analogy to a gas of a certain temperature and pressure information can be gained by measuring the microstate of the gas. This de facto unavailable information is thus the negative thermodynamic entropy, which reflects the “unknownness” or the “disorder” of the microstates. Thermodynamics and Shannon entropy seem to be closer to each other than often assumed, especially by critics of the use of the term “entropy” in Information Theory [22].
Shannon considered the probability of the appearance of a certain sign at the receiver side as the information content. However, not content wise in the sense of meaning. Shannon’s Information Theory not only allowed for the first time to measure information, which so far had been a quite airy metaphor. It also created the Bit as a unit of measurement and as the smallest amount of information. This measurement and the unit of measurement provided the basis for extending the application of Information Theory far beyond communication technology into areas like biologic organisms and other fields. Please note again, that this information is not connected to meaning, i.e., semantics.
The usage of the “Bit”, however, is to be carefully considered because there are two interpretations available. A more popular interpretation states that a Bit reflects a two-state system, which allows a binary differentiation into “there” and “not there”. In an interpretation closer to Shannon’s view, a Bit is a binary probabilistic variable with equally probable states. A Bit, thus, is the entropy which is gained by choosing between these states. It is the amount of information gained by determining the state. The latter view emphasizes the reduction of uncertainness by making a choice.
Information Theory is not concerned with semantic interpretation or information meaning. Meaning, as Weaver explained, relates to an agreement between sender and receiver. In case these do agree that a “1” relates to the whole content of a book—he used the King James version of the bible—and “0” does not, then the whole semantic content of this bible is “activated” at the receiver’s side. What is needed in such a configuration is a mutual understanding of the book’s content before the transmission of the signal, otherwise the receiver cannot use the information of that Bit.
The quantum physicist Schrödinger had his own interpretation of entropy [23]. He proposed that life, to be able to exist, exports entropy to its environment to lower its own entropy. He named the gain made “negentropy”. Organisms run as thermodynamic open systems. According to Schrödinger, negentropy is stored by these organisms (bodies) as information, which, in this sense, represents order and structure. Biochemist Crofts [24] warns that information and negentropy cannot be used as synonyms as long as information is meant to carry semantics. In case of semantics, a receiver is needed, able to interpret the incoming messages by its meaning.

2.3. Entropy or Information

Physicist Ben-Naim proposes in his book “A Farewell to Entropy: Statistical Thermodynamics Based on Information” not to use the term “entropy” anymore. He suggests to replace the “unfortunate, misleading and mysterious” term “entropy” with a more familiar, meaningful and appropriate term, such as information, missing information or uncertainty [25].
Philosopher Floridi [22] introduces the term “mixedupness” to replace entropy. He points out that, in thermodynamics, the availability of energy decreases with increasing entropy. Maximum entropy means a maximum of mixedupness of molecules and the absence of available energy. In Information Theory, entropy is maximized by an even distribution of signs of a used alphabet. In this case the uncertainty of receiving a specific sign is maximized because the signs are completely independent from each other. Floridi equates this with an increasing deficit of information.
Both Ben-Naim and Floridi participate in the ongoing discussion about the true nature of information. Contributions from both, however, indicate a still ongoing search with regard to a valid modeling approach.

3. Meaning

In Information Theory, a central aspect of information is the culminated uncertainty of appearances of signs in an alphabet of signs. In any case, it does not matter what meanings individual signs or sequences of signs are intended to carry.
Of course, this does cover only partly what we expect information to be. Our understanding of information reaches beyond syntax, semantics and pragmatics. It includes meaning, knowledge and cognition.
What are “meaning”, “knowledge” and “cognition” in our context? With regard to meaning we refer to Küppers [10] who states that “meaningful information in an absolute sense does not exist. Information acquires its meaning only in reference to a recipient. Thus, in order to specify the semantics of information one has to take into account the particular state of the recipient at the moment he receives and evaluates the information.” In short to maintain knowledge and cognition. With biochemist Kováč [26], we understand knowledge as being embodied in a system, “as the capacity to do ontic work, necessary for maintaining onticity, permanence of the system. And also to do epistemic work: to sense, measure, record the properties of the surroundings in order to counter their destructive effects and, in the case of more knowledgeable systems, to anticipate them. Epistemic complexity of an agent thus enfolds two processes: one of the evolutionary past and another of the total set of potential actions to be performed in the future.” Cognition uses knowledge and creates knowledge. It relates mental abilities to knowledge capabilities as memory, judgment, reasoning, decision making, etc. In the case of humans, cognition can be conscious, abstract, conceptual, etc.
Although everybody has an intuitive understanding what meaning is, it is worthwhile to have a closer look.

3.1. Current Modeling of Meaning

Einstein, famously, once asked “Do you really believe the moon exists only when you look at it?” He could not comply with a strong view in quantum mechanics that observations themselves change the world. Wheeler later even proposed a participatory universe where observers are not just collecting information of the objects they observe but also heavily influence the results of observations. He invented thought experiments like the “delayed choice”, which he even placed into interstellar space to get a convincing scenario supporting his view [27].
Let us assume, for the sake of a little thought experiment, that no observers are around to look at objects of the world. In that case, no one would be aware of culture and artifacts. In case humans stopped existing as “observers”, immediately all cultural elements of the world would have virtually disappeared. They might still exist as “things” for animals, but not as parts of a—by humans created and maintained—culture. The same applies to results, methods and artifacts of natural science. The horizon of human specific knowledge would collapse. Of course, the brain, as a knowledge processing biological device, would also be gone and with it the extraordinary cognitive capabilities of mankind. Nobody would be left to utter “I am”. All human specific information would be wasted to such a degree that it simply would not exist anymore. All meaning with regard to mankind would be erased.
What would be left is what evolution produced beside humans that could look at objects of the real world. Although we have only limited insight into the ways animals see the world, we know about numerous similarities. Since we are “products” of the same evolution, it hardly could be otherwise. In many ways animals (and plants) act in comparable ways in their environments, that is to say their ecological niches, as humans do. We can conclude that since they survive in the same reality as we do they maintain a realistic view of the world. Thus their “meaning” or better “meanings” of the world would still held up if all humans had gone. “Meaning” in this case reaches from elementary interaction, e.g., by bacteria with the environment to complex dealings similar to our own ones, e.g., by primates.
However, as soon as we would exclude animals and plants (and bacteria) from observing the world in our thought experiment, we would have excluded meaning from the earth where our known evolution has taken place. What is then left does not have a meaning, however limited it might be. Stones and lifeless chemical structures and electromagnetic or gravitational fields, we assume, do not actively have a meaning. They might be subjects of meaning, but the carriers of this are consequently excluded from being by our thought experiment. Nevertheless, we recognize that it is widely accepted that, e.g., interacting particles as photons do “measurements” when colliding with physical structures. We assume however, that in the absence of an observer this would create Shannon Information, however not semantic information.
With meaning gone in our thought experiment, the existence of things has disappeared as well. It takes awareness to recognize “existence”. Since we cannot be aware of objects as they are because we are restricted by our senses and interpretation capabilities, these objects as we define them no longer “exist”. As objects, e.g., the “moon”, they depend on living beings who recognize an object as a moon. In other words, the objects’ very existence depends on being observed. Someone arriving from outer space might “see” the moon through his specific instruments and senses. The result may be very different from the appearance we create of the moon. We assume that observers create or construct objects by building mental models. Observers create models of objects of various complexity. Simple organisms create simple ones. Organisms that are more complex, however, may use simple models, combine theme, use them as building blocks and create new and more sophisticated models.
Reality on a basic level is indeed nothing we can describe without being aware of it. An object is only such with respect to an observer interacting with it [28]. Even abstract nouns like relationship, potential, force, etc. are results of our capability to create models of the external world. This is even so if—to an extreme—reality is thought to be pure mathematics, as physicist Tegmark [29] proposes.
In summary, the key to meaning is biological life. No meaning without life and no life without evolution. As we will line out in the following, evolution produces meaning via creating and processing information.

3.2. Meaning—A Product of Evolution

Meaning is linked to evolution. It is produced by evolution as a part of the organisms themselves that are created in the process of evolution. This is partly done internally by processing available information in organisms, and partly by allowing external information to enter the organisms for further processing. The latter happened first with the beginning of evolution.
Information is what allows meaning and knowledge to be created. It has been existing since the start of evolution. The very first appearance of life came with information. From then on variety and complexity have both changed life. With the unfolding life information became richer and more complex—depending on the individual species. This, in turn, allowed species to create more complex models of the world which they recognized as their environment.
Information processing in and by organisms is in major parts being done specifically by DNA and RNA, inherited by ancestors at the level of molecules, and of course by brains. In addition to that communication necessary for building and running an organism and information based interaction with the environment, based on senses and action, play an important part in information processing.
Each organism depends on informational input from the external world. Without it is impossible to build models of the world, which are applied and combined with models that were inherited. The senses, which allow gaining external information, vary considerably. Different species apply different senses according to their specific approaches of how to recognize and understand their environment. All developments of these senses, however, started with the beginning of evolution. They started with the beginning of life—still biology’s biggest mystery—and thus they started small. They may have been started by catching single photons that provided a sense of an active light source, may it have been the sun or an object reflecting this photon. Encapsulation, self-adaption, reproduction or, in other words containment, metabolism and genetics, which are often, such as by physicist Roederer [30], is seen as a combined first giant step of evolution we assume to as much later and was preceded by continuous appearance. Catching information was first. Loewenstein [5] assumes that the most fundamental level of life is photosynthesis. “And photosynthesis is the capture of information from solar photons.” Molecular geneticist McFadden [31] speculates that self-replication, which indeed might be seen as a precondition for the real beginning of life, was created by quantum processes incorporating entanglement of combinations of amino acids and multiverses, of which one eventually produced the right combination of amino acids capable of doing self-replication.
How life started in detail, however, is subject to research, and it is being debated. Let us assume with Loewenstein [5] that, in whatever way it commenced, it started with an information gain. This is not necessarily the only aspect but certainly one of the most important ones. It is one that is so tightly connected with evolution that we consider it as a precondition for evolutionary development. We assume that from early on information usage went on with model building. Incoming information was—and is—used for creating models of the world. At a primitive stage, for example, bacteria—they already had in early history developed the cell wall as a major invention—used incoming light to orientate their movements. To a much greater extent incoming information was and is processed by mammals. Bigger species had to process more information to orientate themselves and act in their environment. The models they used during evolution were abstractions of more primitive models. Part of the abstraction was transformed, as a result, and became part of their physiognomy which allowed a species to physically take advantage of their environment. Of course, this was not “their environment” exclusively, but was inhabited by other species with which strong interactions like hunting or being hunted existed. Part of the model information went into “brainware”, i.e., processing capabilities and data.
Each species extracts specific meaning out of its environment. The range of this extraction might be small, e.g., in the case of bacteria, or large, e.g., in the case of humans who are able to explore the depth of the universe, the smallness of quantum and their own mind, to name a few. The better and the more intensive the interaction with the world outside organisms is, the better are the preconditions to extend the current reach and broaden the sphere of knowledge.

4. Information and Energy

The interaction between organisms and their external environment happens via senses for gaining information. Different organisms feature different senses. Mobility broadens the scope and intensity of interaction. The scope of senses can be extended by walking, swimming, flying, etc. Communication capabilities, such as language, habitus, gestures, signs, etc., further extend interaction capabilities with the external world.
No matter which means of interaction are being used, they have to be paid for with energy by the organisms. For a biological organism, there is no exchange with its environment available at all that would be free of charge.
Semantic information in our assumption is a quality of energy—it is not carried by it. Information nevertheless does not change the fundamentals of energy. Einstein’s formula equalizing energy with mass and Planck’s formula equalizing energy with the product of a fundamental constant and frequency are not touched. Of course, we assume energy to be as un-destroyable, transferrable from one physical object to another, and—important here—convertible from one form to another one. Energy is dynamic, it is never homogeneous.

4.1. No Information without Energy

Bateson’s characterization of information as a “difference which makes a difference” already directs to energy. It seems plausible that no difference can be created without applying energy; no cause, no effect.
Shannon’s type of information intentionally rules out meaning. By doing so, information entropy, on the one hand, is a measurement of information in a certain sense, but not in a sense that is commonly shared. The common understanding of information includes Morris’ [8] definition of syntax, semantics and pragmatics (see above).
All of the three mentioned components evidently were familiar to the human brain from quite early on. They were needed for doing language processing. DNA and RNA molecules are also places where we can identify those components of information when assuming the components are organized and interpreted as signs and words. Investigating the external world, artifacts for information storage, information transmission and information manipulation contain syntax, semantics and pragmatics. To be more precise, it must be emphasized that in the context outlined above externally stored information loses its status as information when it is stored. It regains the status as semantic information when it has been read again. When it is stored, it is transformed into manipulated structured matter or electromagnetic fields or holograms, and when read, its “engraved” structure is being transmitted to the reading instance by using energy. Then, it can be transformed into biological information once again. Information storage is feasible with matter and energy. It can be engraved, written, digitized or coded in electric fields. Many more options than those are available. They range from the Rosetta Stone (to name an ancient example) or the like to more modern means like books, news, documents, data on computers, streamed data or traffic signs, to name only a few.
Information transmission can take place by means of organism specific capabilities such as languages, gestures, etc., and by external artifacts reaching from drums to modern media technologies. In the same way, information processing as a capability, in a narrow sense, is restricted to biological organisms. However, they can be supported by means of external artifacts like, for example, computers. In both cases, in our understanding, since information is restricted to residing in biological organisms, the externalized transmission and processing does not deal with information in a narrow sense, but merely with patterns of energy. External means of supporting biological storage, transmission and processing are not part of what evolution is concerned with. The external resource only becomes information—out of matter and energy—if it is activated and evaluated by biological organisms (see Figure 1).
We assume, syntax has been produced by evolution as is semantics and pragmatics (see Morris [8] for a definition). In the beginning, syntax has materialized as simple rules (if e.g., received energy—it may have been some photons—appears to come from a source more to the right as similar energy did before, then generate an impulse to move to the right too) imprinted in biochemistry molecules as it is imprinted into logical hardware circuits of computers (this analogy is not to be overstretched beyond certain limits, of course). Later, the biochemistry hardware became accompanied by “software”—also in an analogy to formal language of computer technology. We further assume that semantics was developed with accumulation of information, as were its syntactic rules and improved processing capabilities of the species biological neural network.
Figure 1. Transformation of energy into information from outside to inside of a biologic organism. Energy is transformed into information via senses and the application of syntactic and semantic rules (by accompanied pragmatics) evaluating energy on its usability as information.
Figure 1. Transformation of energy into information from outside to inside of a biologic organism. Energy is transformed into information via senses and the application of syntactic and semantic rules (by accompanied pragmatics) evaluating energy on its usability as information.
Information 06 00411 g001
Thus, a book, the Rosetta Stone and an electronically represented Beethoven sonata remain dead matter or meaningless acoustic waves as long as they are not interpreted by a sufficiently knowledgeable being. Assuming sufficient knowledge means that this being has sufficient context information to its disposal to do the interpretation. As a step before this interpretation, however, the information has to be provided to the interpreting instance. For this, communication between source and interpreting instance—ultimately the brain—has to take place. Each sign that is to be communicated reaches this interpreter via energy. Whether it is photons reflected by book pages or audio signals or odorant chemicals reaching the odorant receptors, it is transformed from energy into information by the senses (eye, skin, ear, nose, and taste buds). It thus reaches the brain by means of energy—classified as information—and, with the help of energy, is manipulated for storage, processing and transmission.

4.2. Not all Energy is Information

Information, in any case, is a quality of energy. However, not all energy is information. It is by using syntax, semantics and pragmatics to transform energy into information. By this its usage context is being changed. Energy has then been transformed into information, but nevertheless it continues to be energy. Whether energy is information exclusively depends on the context being used to do the classification energy vs. information. Information therefore is energy in informational context. Signs on a book page are not recognized as signs at all without the proper context already in place. A Beethoven sonata may not unfold its beauty unless the listener is in some way prepared for it.
We conclude that information is related to energy and matter in the following ways:
  • Information depends on energy as a means of transmission.
  • Matter as a carrier of information is in no way—physically or chemically—different from matter not carrying information.
  • The structure imposed on objects by storing information, we define as potential information. This structure is potentially transformed into information, provided a fitting context is available to interpret it.
  • Energy intrinsically provides potential information to a receiver. Whether or not this information is transformed to semantic information depends on the context, respectively, the syntax, semantics and pragmatics maintained by a receiver instance.
  • A source of information is necessarily directly or indirectly energetically active.
  • Information is being transmitted to a receiver (e.g., senses) via an information channel.
  • Transmission is performed with energy.
  • The receiver instance of information is being energetically activated.
These points we see as extensions to the Shannon Information Model. Especially the context is needed at the receiver’s side for differentiating between energy and information. Information comes as energy. Signals are transferred without creating meaning. Only at the receiver’s side context comes into play. The sender’s side may maintain context, like, for example, in human communication, not like with a stone with hieroglyphs. Context allows for classifying energy as information and non-information.

5. The Evolutionary Energetic Information Model

As mentioned above, we base our approach of a new comprehensive information model on results available among others via Shannon’s Information Theory, Bateson’s definition, and Burgin’s ontology. The model defines semantic information in nature within a narrow and closely contoured framework. At the core, we propose that this is the only information in nature. Information in that sense is energy, however, not all energy is information. Information is biological and it is exclusively being produced by evolution. This information model we call Evolutionary Energetic Information Model (EEIM).
The basic ideas of our modeling approach of information are as follows:
  • information without energy cannot exist;
  • evolution exclusively creates information on various levels of complexity; and
  • thinking produces emerging models of the external world to reduce complexity.

5.1. Energy Provides Information

As we have outlined above, energy is information or it is not—depending on the context available at the receiver’s side. Thus, this classification creates information in that it directs accordingly classified energy for further processing to the appropriate information channel.
EEIM—with regard to energy—is based on the following assumptions:
  • Information is not per se in existence.
    It is a prerequisite to have energy in order to provide information to an organism. No object of reality that does not belong to the biological sphere created by evolution is able to receive information as such. Whether incoming energy is information or not is to be decided by the receiver. The receiver uses syntax, semantics and pragmatics (SSP) to perform this classification, which is based on the internal status of the receiver. This status in turn reflects the degree of sophistication in evolution the receiving organism has reached and which determines the complexity of SSP.
    Information thus comes into existence by selecting it from energy or, in other words, transforming it from energy. It remains energy as long as it is not treated as information. The origin of the use of energy as information started with evolution. In the process of evolution an increasing variety of incoming energy has been accepted as information. Very primitive information in this sense is, for example, a single Bit of information like “dark” or “bright”. More complex information is provided by incoming signals, e.g., photons, reflected by moving objects. Sophisticated information then can come as mathematical formulas in the shape of reflected photons—and are then processed by the trained brain. So where is SSP to be found here? A primitive organism with a sensible cell or the even more primitive version of an “eye” sees—and here is the rule—something being bright, or it does not. Moving targets require an eye which differentiates between fixed or moving objects—again by applying rules focusing, for example, on horizontal change, and then on vertical change. A mathematical formula is a device for the brain for applying SSP in order to differentiate between signs or sequences of signs that are allowed (or not allowed).
    SSP is obviously not a privilege for human senses and brains. It rather plays a crucial role in information processing at all levels and in all the branches of evolution.
  • Information is not a priori.
    What one receiver considers as information may not be seen likewise by another receiver at all. Information is created by a process of evaluation and recognition. Spoken human language is obviously able to carry valuable information as long as the language is understood by the receiver. Understanding requires the application of the proper SSP.
  • Information comes as energy.
    Information appearing as energy at the receiver’s side of a communication channel triggers a transformation process. This requires energy above a certain level. The eye, e.g., is very sensitive for low levels of energy, which might be just a small number of photons—each of them not having any mass—coming in. In case the trigger level is too high for incoming energy, no information is being generated. This means that we do not realize very low pressure on our skin.
    Information is being generated out of energy by applying context sensitive transformation processes. It remains energy as long as it is not connected with a certain pragmatics. If not, it is for sure that its characteristics as information remains permanent.
  • Syntax, semantics and pragmatics are not a priori.
    Information so far has exclusively been produced by processes within biological evolution. SSP as a context for transformation from energy to information is required. SSP itself is subject to evolution. Potential information, which is placed outside the living objects of evolution, may be activated inside. For example, libraries contain information only as a passive matter. The structure of this matter was created by writing and can eventually be activated only by humans because they possess the required SSP configuration for the specific writing. Then, the transformation from potential to semantic information can happen. Libraries thus contain potential information vs. activated information like in brains. The reading of machines does not create activated information. However, what happens here is a transformation from one version of potential information to another. Without a biological environment, information exists, like Shannon information, as patterns with certain probabilities of appearance.
    Receivers as well as senders principally maintain a variety of SSP. This SSP is a matter of a number of languages understood, observer behavior, preferences for the biological external world, or basic properties like the recognition of dark or bright.

5.2. Evolution Creates Meaning

EEIM is part of the biological evolution. The progress of evolution creates new SSP contexts. New SSP contexts thus are the result of adaption and selection. New capabilities—created, e.g., at random—manifest better or worse ways of surviving. Better survival means a higher probability of those who show it to exploit the environment, to create advantages within even their own species, and to bequeath the new capabilities to a next generation.
The information created by evolution is dynamic by nature. It lives and dies with the organism possessing it. In the case of humans, the organism has created various ways of externalizing information as a physical structure, e.g., by transforming semantic information into potential information residing in physical structures like stone engravings, written paper, silicon storage, etc. It can be (re-)activated by other humans by applying the proper set of SSP, thereby transforming potential information into semantic information. Let us take a ship like the RMS Titanic for example. When Morse telegraphs just came into wider use, an emergency call could only have been identified as such if the proper signal sequence had been received and both had followed the proper rules, as prescribed by SSP. Similar processes, we assume, are happening in biological systems as multi-staged transformation beginning with the senses and ending with brain processing.
In summary, our model comprises the characteristics in Table 1.
Table 1. Characteristics of Evolutionary Energetic Information Model (EEIM). A mapping between aspects of reality and model attributes is given.
Table 1. Characteristics of Evolutionary Energetic Information Model (EEIM). A mapping between aspects of reality and model attributes is given.
Aspects of Reality Attributes of Information EEIM
Evolution1Information is exclusively being “produced” by evolution
Energy2Information is a quality of-informational-energy
3Informational energy provides semantic information to accordingly prepared receiver instances
4Not all energy is informational
Semiotics5Information is bound to syntax, semantics and pragmatics
6Information requires a sending instance
7Information requires a biological receiver able to interpret incoming information
Biological receptors8Interpreting receiving instances for information are living organisms exclusively
Information processing9Organisms possess processing capabilities for information according to their place in evolution
10SSP in context is needed to transform received information into new information
11Animals are able to process information
Knowledge12Information may potentially be transformed into knowledge
13Organisms possess knowledge according their background in evolution
14Organisms processing knowledge are able to produce knowledge processing artifacts
Entropy15Shannon’s Information Theory allows for the probability of the appearance of signals. It does not cover semantic.
  • (1) Information is exclusively being “produced” by evolution. Everything that was created by biological evolution receives, manipulates and sends information. All that is not a result of biological evolution contains no information but, as artifacts, can contain potential information. Potential information can—but need not—be Shannon Information created by biological objects of evolution. Transmission of potential information by means of patterns of energy to a biological receiver is a precondition of being interpreted there. Interpretation is accepting energy as information, classifying it accordingly and channeling it properly. Thereby semantic information is created out of energy. Semantic information is being “produced” by evolution as it came into being with evolution and has evolved in evolution over time as an intrinsic part of evolution itself.
  • (2) Information is a quality of—informational—energy. A transformation into information is bound to an earlier transmission of energy. In the case of potential information, where energy is the originating cause, the transformation result will be information. In case transformation does not take place, energy remains pure energy and may be used by the receiver organism as such. Every living system is equipped with a fixed number of distinct sensors. The sensors interact with a specific part of the surroundings we call environment (Kováč [26]). Via these sensors, informational energy reaches the organism. In case of no energy inflow or when incoming energy is homogeneous, i.e., when entropy is maximized, no semantic information can be detected.
  • (3) Informational energy provides information to accordingly prepared receiver instances. The observation of energy by an organism a priori does not indicate whether the observed energy is information or not, unless the observer possesses and applies a fitting context, i.e., an SSP. Thus energy is information in a context. Energy, however is present all the time. Whether the organism in question is able to detect semantic information or not makes the difference. This in turns depends on the semantic context, the syntactic rules and the pragmatic features available. Kováč [26] summarizes that “Life is an eminently active enterprise aimed at acquiring both a fund of energy and a stock of knowledge, the possession of one being instrumental to the acquisition of the other.”
  • (4) Not all energy is information. Energy becomes information by a transformation of a receiving instance with a fitting context (SSP). Forms of energy may appear as light, heat, pressure, acoustics, electromagnetic fields, as well as olfactory and tasting substances and even quantum effects [32]. In short, information can be anything the senses of biological beings are sensitive to. Depending on the applied context, the same energy might lead to different information.
  • (5) Information is bound to syntax, semantics and pragmatics. Whether energy is taken by an organism as semantic information or whether it remains pure energy depends on the context the organism provides. The context consists of SSP, which, in turn, is built of inherited parts and results in the learning of this organism. In turn, semantic information exists only within context, which is SSP.
  • (6) Information requires a sending instance. Information can only be received if sent before. This means that the source always has to spend energy. A book being read reflects photons from the sun or an artificial light source and thereby acts as a source of energy from the receiver’s point of view. Thus the sender instance does not need to be—but can be—a sender acting by explicit intention.
  • (7) Semantic information requires a biological receiver able to interpret incoming information. Information is non-static. It is dynamic in three major aspects. The first is with regard to its staying within biological systems. Information is constantly being processed, communicated or actively stored. The second aspect is with regard to evolution. Information is subject to evolutionary mechanisms like selection, mutation and adaption. Through these mechanisms, it develops with the organisms and species themselves and becomes part of what is being inherited. During the process of evolution, each species has developed its specific set of SSP. The third aspect touches individual learning. Learning in a social context necessarily also changes the SSP sets of an individual.
  • (8) Interpreting receiving instances for information are living organisms exclusively. Non-interpreting inorganic instances that receive energy are not able to perform the transformation into semantic information. Their lack of SSP does not allow the necessary classification of incoming energy as information. All incoming energy, then, remains energy and can, e.g., change the heat status of an object.
  • (9) Organisms possess processing capabilities for semantic information according to their place in evolution. Information can be received by any kind of organism depending on the available sets of SSP it possesses. A simple informational light may come with incoming photons where SSP consists of very simple rules of sensing darkness or brightness. A more complex information requiring a different SSP are, e.g., the massive data streams of the Large Hadron Collider (at the European Organization for Nuclear Research, CERN, Geneva) experiment.
  • (10) SSP in context is needed to transform received semantic information into new information. Information creates new information [16]. Information needs information as an input, produces information as an intermediate results and as an output. Each step, however, changes the context and the SSP of information. This might result in ruling out incoming information so far accepted, or vice versa.
  • (11) Animals are able to process information. Biological organisms perform information processing and are the results of information processing—however, not exclusively—and they develop new capabilities in information processing in the course of evolution. This includes gathering knowledge by species and individuals when building up context by applying SSP.
  • (12) Information may potentially be transformed into knowledge. Available information of sufficient complexity is mandatory for developing (more) complex information structures and are a precondition for the development of (more) sophisticated thinking capabilities. As Kováč states “Biological evolution is a progressing process of knowledge acquisition (cognition) and, correspondingly, of growth of complexity. The acquired knowledge represents epistemic complexity.” In addition to prior knowledge, available when new information comes in, there is also pragmatics regarding a subject as, e.g. context comprising interest, intention, novelty, complexity, and selectivity. All this determines the overall quantity of information.
  • (13) Organisms possess knowledge according their background in biological evolution. This is knowledge regarding their immediate and wider environment, about rules applicable in interaction with others, of certain techniques find and explore food, self-aware knowledge, etc.
  • (14) Organisms processing knowledge are able to produce knowledge processing artifacts. Artifacts for knowledge-processing, such as computers (not yet available for knowledge in a conscious sense), currently can only be developed by humans and thus by evolution.
  • (15) Shannon’s Information Theory allows for the probability of the appearance of signals. It does not cover semantic information. Shannon Entropy can be applied in inorganic and organic fields. By intention of Shannon and Weaver, it is restricted, as a statistical theory, to measure the amount of (potential) information going through a transmission channel. Shannon explained the role of redundancy, noise, and error rates [13]. The underlying information model separated information from meaning.
Semantic information is being created exclusively in the process of biological evolution. At its core, energy within an organism processes—depending on the availability of SSP—classified energy into information and further on handles it as such. The complexity of the information and the SSP depend on the level of evolution reached.

5.3. Emerging of External World Models from Information Intake

Why do we not see quantum objects? The reason is that the entrance level for evolution was at a higher level. No decision was taken to do so. Instead, configurations of what we call matter and energy allowed interaction with, e.g., photons, which were taken as both energy and information. The latter at the very first stages resulted from very basic interactions of organisms (and their precursors) and their external world. They took in not just energy, but also more and more information regarding the environment.
By treating energy as information, models of the world were created from start on. A light or a dark spot, a moving object showing characteristics of prey, and so on. With two sensory systems objects could be calculated with regard to speed, distance etc. Colour sensitivity was added. The capabilities for creating very basic models of the external world were used to create more sophisticated models of the environment. This obviously has been done in different ways by different species. Because of the long common history shared between species, similarities between them, e.g., mammals, are strong.
Here, emergence comes into play. New capabilities combined with older ones allowed for the creation of new models of the world. With certain combinations of capabilities, new features of the outside world emerged as completely novel. Unfortunately, the process of emergence is not a well defined property or feature. Many types of emergence have merely been proposed, but cannot be proven. Whether emergence is considered an objective feature of the world, or whether it is merely existent in the eyes of the beholder, is still being discussed as philosopher Bedau [33] reflects.
We do not consider evolution as a direct cause for emergence. It is part of cognitive abilities acquired during and within evolution. Our access to reality is limited and we have to rely on models of reality. The environment cannot be understood directly. We have to help ourselves by our constructions of reality. As philosopher Wallner [34] states regarding models in the context of constructive realism “If they serve us well for gaining control over the environment, we keep them. If they do not, we discard them.” Model building seems to be an informational capability that all organisms, or better all organisms above a certain complexity, should possess. Model building in this sense is partnering with cognition in reducing complexity. Emergent—that is constructed—models are being created and used by the organisms and their brains. By creating new and more advanced models, available models may re-used and—when emergence happens—added with new qualities not contained before (e.g., bacteria to orient themselves according stimuli not used before, crows by recognize new types of enemies, humans by thinking about information models). Thus, emergence is a result of information processing not a result of a miracle or an intervening divine instance.
As philosopher Jeremy Butterfield [35] does, we do consider emergence as knowledge that is novel and robust relative to some more basic comparison class Emergent phenomena are Janus faced; they depend on more basic phenomena, and yet they are autonomous from that base [33].
We take the view that the process of taking semantic information from the outside world into the unfolding biological environment of life and making it subject to evolution started the process of model building, which is still going on in all species. Models are being built on top of more basic (or less adapted) earlier models and provide the basis for even more sophisticated models. This is obviously a knowledge building process. At different stages, specific views of the world are created.
This process of building models is connected to steps or stages marking knowledge leaps. It is based on taking in information, combining it with information already there and creating new information. With increasing amounts of information, complexity of information models had to be created that are able to reduce both amount of data to be processed and complexity to be dealt with. Since information is energy, both reductions are energy saving and thus optimizations of the running of the thermodynamically open organic systems.

5.4. EEIM’s Impact on Science

Will EEIM have a practical impact? We assume so. For an engineer, it makes a significant difference whether a motor is described by its internal forces at work, by the interplay of components, and by the chemical reactions taking place or whether—complementary—she/he is also able to understand, e.g., the thermodynamic laws and the role entropy plays in the machinery. Similarly, the capabilities of a computer cannot be sufficiently understood by just analyzing and understanding in detail its physical components and their interfaces to each other. It is the software which makes the difference and turns the physical device into what it is, an information manipulating framework of great flexibility. To understand how a computer works, the understanding principles of algorithms, of programming languages, of logical circuits, etc. must also be understood. The point is, understanding of complex systems requires holistic views to be applied. In any way, motors and computers are here to be seen as metaphoric, not as strong analogies to information handling in biological organisms. Life is certainly different from machinery.
Being aware of the important role information has in biology will influence the whole attitude of how organisms are being analyzed, understood and described. This will hopefully open the way to identifying new principles and maybe laws information obeys in organisms. Currently, we are only beginning to understand biological components and processes to a certain degree. To understand more of the combined roles and impacts of information, evolution and energy the EEM hopefully provides a staring point for new thinking.
Inevitable, this contribution leaves many questions open. Among them, are the roles of external information resources and information processing artifacts, which show capabilities that resemble processes of evolution. Also, the measurement of semantic information is not answered and—connected with the question of the very beginning of life—the origin of biological semantic information and its developments throughout evolution. The impact of quantum information on biological information, we also expect, to be a field of fruitful studies.

6. Conclusions

This contribution presents an approach to understand semantic information in nature, taking into account the views of Shannon, Bateson, Morris and others. The novelty of this approach is threefold. First, information is considered as a quality of energy, however, not all energy is information. Second, semantic information is assumed to be processed by biological organisms only. Third, information is being “produced” by evolution and restricted to areas belonging to evolution. With EEIM, we can now better explain the growth of information in time and in species in general. We make a distinction between Shannon Information, classified as potential information, and semantic information used by biological organisms. Semantic information is part of biological organisms in various degrees, depending on the species themselves, their ecological niche, the time within evolution, etc. This distinction draws a line between information as statistic pattern and semantic information. We have also created a basis from where in history knowledge was developed and worldviews or world models have been developed by species. This will potentially help epistemology to understand better how organisms perform the perception of the outer world. It will also help to understand better how organisms reduce complexity of incoming information of the outer world in order to create simplified models for their internal use.

Acknowledgments

I am very grateful for the constructive and helpful comments of three anonymous reviewers. Further I am grateful to Ms. Barbara Toepfer for many helpful comments.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  2. Weaver, W. Some recent Contributions to The Mathematical Theory of Communication. In The Mathematical Theory of Communication; Shannon, C.E., Weaver, W., Eds.; University of Illinois: Champaign, IL, USA, 1949. [Google Scholar]
  3. Wheeler, J.A. Information, physics, quantum: The search for links. In Complexity, Entropy, and the Physics of Information; Von Wojciech, H.Z., Ed.; Addison-Wesley: Redwood City, CA, USA, 1990. [Google Scholar]
  4. Eigen, M. Self-organization of matter and the evolution of biological macromolecules. Naturwissenschaften 1971, 58, 465–523. [Google Scholar] [CrossRef] [PubMed]
  5. Loewenstein, W. Physics in Mind; Basic Books: New York, NY, USA, 2013. [Google Scholar]
  6. Bateson, G. Mind and Nature: A Necessary Unity; Dutton: New York, NY, USA, 1979. [Google Scholar]
  7. Burgin, M. Theory of Information-Fundamentality, Diversity and Unification; World Scientific Series in Information Studies; World Scientific Publishing Co. Pte. Ltd.: Singapore, Singapore, 2010; Volume 1. [Google Scholar]
  8. Morris, C.W. Grundlagen der Zeichentheorie; Hanser Verlag: Munich, Germany, 1972. [Google Scholar]
  9. Heyes, C. New thinking: The evolution of human cognition. Philos. Trans. R. Soc. B 2012, 367, 2091–2096. [Google Scholar] [CrossRef] [PubMed]
  10. Küppers, B.-O. Elements of a Semantic Code. In Evolution of Semantic Systems; Küppers, B.-O., Hahn, U., Artmann, S., Eds.; Springer: Berlin/Heidelberg, Germany; New York, NY, USA, 2013; pp. 67–86. [Google Scholar]
  11. Davies, P. The quantum life. Phys. World 2009, 7, 24–28. [Google Scholar]
  12. Hofkirchner, W. The Quest for a Unified Theory of Information; OPA (Overseas Publishers Association) N.V.: New York, NY, USA, 1999. [Google Scholar]
  13. Bates, M.J. Information; CRC Press: New York, NY, USA, 2010. [Google Scholar]
  14. Hofkirchner, W. Toward a New Science of Information. Open Access Inf. 2011, 2, 372–382. [Google Scholar] [CrossRef]
  15. Hofkirchner, W. Information. A Unified Theory of Information Framework; World Scientific: Singapore, Singapore, 2013. [Google Scholar]
  16. Von Weizsäcker, C.F. Die Einheit der Natur; Deutscher Taschenbuch Verlag: München, Germany, 1974. [Google Scholar]
  17. Burgin, M. Evolutionary Information Theory. Open Access Inf. 2013, 4, 124–168. [Google Scholar] [CrossRef]
  18. Kofler, W. Information-from an Evolutionary Point of View. Open Access Inf. 2014, 5, 272–284. [Google Scholar] [CrossRef]
  19. Walker, S.I. Top-Down Causation and the Rise of Information in the Emergence of Life. Open Access Inf. 2014, 5, 424–439. [Google Scholar] [CrossRef]
  20. Popper, K. Three Worlds. The Tanner Lecture on Human Values; University of Michigan: Ann Arbor, MI, USA, 1978. [Google Scholar]
  21. Min, J. Information: Definition, Origin and Evolution. In The Quest for a Unified Theory of Information, Proceedings of the Second International Conference on the Foundations of Information Science, Vienna, Austria, 11–15 June 1996; Hofkirchner, W., Ed.; Routledge: New York, NY, USA, 1999. [Google Scholar]
  22. Floridi, L. The Philosophy of Information; Oxford University Press: Oxford, UK, 2011. [Google Scholar]
  23. Schrödinger, E. Was ist Leben? Serie Piper: München, Germany, 1989. [Google Scholar] (1st edition: What is Life? Cambridge University Press: Cambridge, UK, 1944. ).
  24. Crofts, A.R. Life, Information, Entropy, and Time Vehicles for Semantic Inheritance. Available online: http://0-onlinelibrary-wiley-com.brum.beds.ac.uk/doi/10.1002/cplx.20180/abstract (accessed on 20 July 2015).
  25. Ben-Naim, A. A Farewell to Entropy: Statistical Thermodynamics Based on Information; World Scientific: Singapore, Singapore, 2008. [Google Scholar]
  26. Kováč, L. Information and Knowledge in Biology. Plant Signal. Behav. 2007, 2, 65–73. [Google Scholar] [CrossRef] [PubMed]
  27. Zeilinger, A. Why the quantum? “It” from “bit”? A participatory universe? Three far-reaching challenges from John Archibald Wheeler and their relation to experiment. In Science and Ultimate Reality; Barrow, J.D., Davies, P.C.W., Harper, C.L., Eds.; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  28. Rovelli, C. Relative Information at the Foundation of Physics. In IT from Bit or Bit from IT? Aguirre, A., Foster, B., Merali, Z., Eds.; The Frontiers Collection; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  29. Tegmark, M. Our Mathematical Universe: My Quest for the Ultimate Nature of Reality; Alfred A. Knopf: New York, NY, USA, 2014. [Google Scholar]
  30. Roederer, J.G. Information and Its Role in Nature; The Frontiers Collection; Springer: Berlin/Heidelberg, Germany, 2005. [Google Scholar]
  31. McFadden, J. Quantum Evolution; W.W. Norton & Company, Inc.: New York, NY, USA; London, UK, 2001. [Google Scholar]
  32. Briegel, H.; Sandu, P. A perspective on possible manifestations of entanglement in biological systems. In Quantum Effects In Biology; Mohseni, M., Omar, Y., Engel, G.S., Plenio, M.B., Eds.; Cambridge University Press: Cambridge, UK, 2014. [Google Scholar]
  33. Bedau, M.A.; Paul, H. Introduction. In Emergence: Contemporary Readings in Philosophy and Science; Bedau, M.A., Paul, H., Eds.; The MIT Press: Cambridge, MA, USA, 2008. [Google Scholar]
  34. Wallner, F. A New Vision of Science. In Proceedings of the Twentieth World Congress of Philosophy, Boston, MA, USA, 10–16 August 1998.
  35. Butterfield, J. Less is Different: Emergence and Reduction Reconciled. Found. Phys. 2011, 41, 1065–1135. [Google Scholar] [CrossRef]

Share and Cite

MDPI and ACS Style

Johannsen, W. On Semantic Information in Nature. Information 2015, 6, 411-431. https://0-doi-org.brum.beds.ac.uk/10.3390/info6030411

AMA Style

Johannsen W. On Semantic Information in Nature. Information. 2015; 6(3):411-431. https://0-doi-org.brum.beds.ac.uk/10.3390/info6030411

Chicago/Turabian Style

Johannsen, Wolfgang. 2015. "On Semantic Information in Nature" Information 6, no. 3: 411-431. https://0-doi-org.brum.beds.ac.uk/10.3390/info6030411

Article Metrics

Back to TopTop