Next Article in Journal
Probabilistic Autoencoder Using Fisher Information
Next Article in Special Issue
Lack of Habituation in Migraine Patients Based on High-Density EEG Analysis Using the Steady State of Visual Evoked Potential
Previous Article in Journal
High-Energy Ejection of Molecules and Gas-Dust Outbursts in Coal Mines
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Editorial

Entropy for the Brain and Applied Computation

1
Istituto Nazionale di Alta Matematica (INdAM), Piazzale Aldo Moro 5, 00185 Rome, Italy
2
Institute for Quantum Science and Engineering, Southern University of Science and Technology, Shenzhen 518055, China
3
Institute of Sciences and Technologies for Sustainable Energy and Mobility—CNR–STEMS, Via Canal Bianco 28, 44124 Ferrara, Italy
*
Author to whom correspondence should be addressed.
Submission received: 30 November 2021 / Accepted: 2 December 2021 / Published: 6 December 2021
(This article belongs to the Special Issue Entropy for the Brain and Applied Computation)
Entropy is a quantity expressing the measure of disorder or unpredictability in a system, and, from a more general point of view, it can be regarded as an irreversible source of energy. Primarily, it was introduced as a classical thermodynamic quantity by Clausius and defined as the heat reversibly exchanged at a given temperature and has become a key concept in statistical physics because of the work of Boltzmann and Gibbs, and in information theory due to Shannon’s work; it takes different definitions that are all related to its statistical meaning [1]. Entropy also plays a pivotal role in chemistry, applied sciences, engineering and, last but not least, biological physics and biomedicine. Regarding biomedicine, for example, two decades ago, a method was introduced to calculate multiscale entropy for biological time series, enabling the separation between healthy and pathologic groups of individuals [2]. More recently, Prigogine’s minimum energy dissipation principle was theoretically proved in living systems by studying the rate of entropy production of irreversible reactions occurring in normal and cancer human cells, and its correlations with heat and matter transfer [3,4,5]. In strict connection with the above works dealing with entropy in biological systems, recently, the concept of entropy has also been applied to neuroscience, and its role in brain function has been highlighted as a paramount tool demonstrating the complex and intrinsic properties of the neural network system [6,7]. The description of the brain via the use of entropy has advanced three modern fields of research such as altered states of consciousness, the aging brain and the quantification of brain information processing, and their mutual interrelations [6]. A theory of conscious states has been developed by using neuroimaging research, and the concept of entropy was applied to the context of states of consciousness with special regard to the psychedelic state as a primary state of consciousness preceding the normal waking consciousness [8]. An understanding of the interplay between the brain region and the state of consciousness and the evolution of brain complexity, regarded as an entropy-enhancing process, has recently been proposed, leading to an increase in the space of states that can be visited, and to the accessibility of new channels [9]. Nowadays, nonlinear and complex system theories are considered promising candidates for analyzing the principles of operation of neural networks and their entropic content. The concept of computational entropy, for example, has been utilized in cryptographic primitives, namely, in cryptographic algorithms that are used to create cryptographic protocols in computer security systems. In most cases, computations involve hundreds of data points and have a high computational cost. However, the recent vectors with dissimilarity technique allows avoiding distance calculations of the most dissimilar vectors during computation, based on phase space reconstruction, leading to a reduction in the time needed to compute the sample entropy in random signals and electroencephalogram signals [10]. As a result, measurement of sample entropy computational time is accelerated. Moreover, the introduction of the concept of pseudo-entropy as a novel type of computational entropy that uses pseudorandom generators and is, in general, larger than the actual entropy has further advanced the field of brain entropy and computation.
The goal of this Special Issue is to investigate entropy-based novel methodologies to unravel, analyze and model intricate human brain networks and information flows. This Special Issue includes applied computational methods and theoretical methodologies from physics, biophysics, physical chemistry, electronics, engineering, biomedicine, biomathematics and neuroscience to provide additional insight into the role of the human brain based on entropy.
This Editorial has the aim to attract submissions of theoretical and/or applied computation-oriented papers fitting within the topic of this Special Issue.

Author Contributions

All authors contributed equally to this paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

This work was partially supported by National Group of Mathematical Physics (GNFM-INdAM).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wehrl, A. General properties of entropy. Rev. Mod. Phys. 1978, 50, 221–260. [Google Scholar] [CrossRef]
  2. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy analysis of complex physiologic time series. Phys. Rev. Lett. 2002, 89, 068102. [Google Scholar] [CrossRef] [Green Version]
  3. Zivieri, R.; Pacini, N.; Finocchio, G.; Carpentieri, M. Rate of entropy model for irreversible processes in living systems. Sci. Rep. 2017, 7, 9134. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Zivieri, R.; Pacini, N. Is an Entropy-Based Approach Suitable for an Understanding of the Metabolic Pathways of Fermentation and Respiration? Entropy 2017, 19, 662. [Google Scholar] [CrossRef] [Green Version]
  5. Zivieri, R.; Pacini, N. Entropy Density Acceleration and Minimum Dissipation Principle: Correlation with Heat and Matter Transfer in Glucose Catabolism. Entropy 2018, 20, 929. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Keshmiri, S. Entropy and the Brain: An Overview. Entropy 2020, 22, 917. [Google Scholar] [CrossRef] [PubMed]
  7. Carhart-Harris, R.L. The entropic brain-revisited. Neuropharmacology 2018, 142, 167–178. [Google Scholar] [CrossRef] [PubMed]
  8. Carhart-Harris, R.L.; Leech, R.; Hellyer, P.J.; Shanahan, M.; Feilding, A.; Tagliazucchi, E.; Chialvo, D.R.; Nutt, D. The entropic brain: A theory of conscious states informed by neuroimaging research with psychedelic drugs. Front. Hum. Neurosci. 2014, 8, 20. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Jeffery, K.J.; Rovelli, C. Transitions in brain evolution: Space, time and entropy. Trends Neurosci. 2020, 43, 467–474. [Google Scholar] [CrossRef] [PubMed]
  10. Lu, Y.; Wang, M.; Peng, R.; Zhang, Q. Accelerating the Computation of Entropy Measures by Exploiting Vectors with Dissimilarity. Entropy 2017, 19, 598. [Google Scholar] [CrossRef] [Green Version]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zivieri, R.; Medlej, I.; Fioravanti, A. Entropy for the Brain and Applied Computation. Entropy 2021, 23, 1639. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121639

AMA Style

Zivieri R, Medlej I, Fioravanti A. Entropy for the Brain and Applied Computation. Entropy. 2021; 23(12):1639. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121639

Chicago/Turabian Style

Zivieri, Roberto, Israa Medlej, and Ambra Fioravanti. 2021. "Entropy for the Brain and Applied Computation" Entropy 23, no. 12: 1639. https://0-doi-org.brum.beds.ac.uk/10.3390/e23121639

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop