266 | Christoph Adami on How Information Makes Sense of Biology

Evolution is sometimes described -- not precisely, but with some justification -- as being about the "survival of the fittest." But that idea doesn't work unless there is some way for one generation to pass down information about how best to survive. We now know that such information is passed down in a variety of ways: through our inherited genome, through epigenetic factors, and of course through cultural transmission. Chris Adami suggests that we update Dobzhansky's maxim "Nothing in biology makes sense except in the light of evolution" to "... except in the light of information." We talk about information theory as a subject in its own right, and how it helps us to understand organisms, evolution, and the origin of life.

christoph adami

Support Mindscape on Patreon.

Christoph Adami received his Ph.D. in physics from Stony Brook University. He is currently professor of Microbiology and Molecular Genetics as well as Physics and Astronomy at Michigan State University. Among his awards are the NASA Exceptional Achievement Medal and the Lifetime Achievement Award from the International Society for Artificial Life. His new book is The Evolution of Biological Information: How Evolution Creates Complexity, from Viruses to Brains.

0:00:00.0 Sean Carrol: Hello everyone. Welcome to the Mindscape Podcast. I'm your host, Sean Carroll. Longtime listeners will know that I've long been a fan of Las Vegas. I'd like to go to Vegas, play poker, eat, shop, whatever, just relax and have a good time. I haven't been able to do it quite as much. Now that we're on the East Coast, that's okay. There's always pluses and minuses. Now we can go to New York. It's like the Las Vegas of the East. But anyway, one of the reasons why I like Las Vegas is, in addition to the stereotypes. So the stereotypes are absolutely there, right? There are long rows of slot machines with dull-eyed people, mindlessly pressing buttons, hoping to get rich someday. It's kind of depressing that part, but there's a lot of weird stuff in Vegas if you look closely enough.

0:00:45.0 SC: So by Vegas standards, one of the weird things that I really like is there's an outpost of Bauman Rare Books. This is a well-known rare bookstore with a headquarters in New York, but they have a Vegas store in the Palazzo Hotel Casino where I used to stay sometimes. And it's just weird among all the glitz and glamor, here's rare books. And they actually have a lot of first editions of things, but also some weird sciencey books. They have a signed Albert Einstein, for example, and I was very amused to see in Bauman's rare books in Vegas, a copy of Claude Shannon's master's thesis. Claude Shannon, famous 20th century scientist for many reasons, and his master's thesis, I didn't actually know this before I saw it in the bookstore, is very famous. It's been called by Howard Gardner, the most influential master's thesis of all 20th century science. This is a master's thesis that has its own Wikipedia page, okay? And it's about basically using Boolean analysis to model, and improve the efficiency of circuits and things like that.

0:01:53.3 SC: Pretty awesome for a master's thesis kind of thing. But for my purposes today, I'm just using this as a somewhat strained segue into, of course, Claude Shannon's big contribution, even bigger than the master's thesis, came 10 years after when he essentially invented information theory. He wrote a paper called The Mathematical Theory of Communication. And like many great scientific breakthroughs, theoretical breakthroughs, this was driven by a down to earth technological practical need. Shannon was interested in sending signals across the Atlantic Ocean in wires. And you wanna send a signal in a way that is efficient, right? That is least likely to be garbled or to lose the information. And to do that, to figure out how to do that, you need to have a mathematical way of characterizing what you mean by information. And it turns out, as he discovered by doing the math, that it's very, very similar to the formula for entropy in statistical mechanics.

0:02:53.5 SC: So there was clearly some interesting hidden relationship here between information and communication on the one side, and statistical mechanics and thermodynamics on the other side. And of course, so this is the middle of the century, 1948. This idea has blossomed quite a bit. You hear about information all the time. You hear about it in the political arena, right? Misinformation, disinformation, you hear about it in technology. Remember the information superhighway? That was a big thing. And you must have heard that information as a concept is useful in biology, in neuroscience, quantum mechanics, quantum information, and so on. So today, on the podcast, we have Christoph Adami, who is like many Mindscape guests, he started out his career as a relatively traditional physicist doing nuclear and particle physics. And now, he's a professor of Michigan State, not only in physics, but also in I think, molecular biology department.

0:03:49.3 SC: But he studies evolution and life, including artificial life. And he has a new book out called The Evolution of Biological Information, How Evolution Creates Complexity from Viruses to Brains. So I'm going to use this as an excuse to really get clear on what is this information theory stuff. By the way, the book is a little technical. I'm not gonna advocate it for people who don't like equations. There are some equations there, but the equations are quite mild. If you're the kind of person who doesn't read equations every day, but doesn't blanch when you see one, it might be a good book to pick up. And we're gonna talk about what information theory is in general, and specifically, how biologists or people who care about biology, even if they're physicists, are going to use information theory to better understand things like evolution and even the origin of life. I would say that 75 years later, the idea of information theory we got from Claude Shannon is still not settled in. We're still working through the basic implications of this. So it's very exciting, it's fun, it's a new lens with which to look at the world in a detailed way. And that's all we like to do here at the Mindscape Podcast. So let's go.

[music]

0:05:20.1 SC: Chris Adami, welcome to the Mindscape Podcast.

0:05:21.0 Chris Adami: Glad to be here.

0:05:22.9 SC: Let's start with this idea of information theory. It's always gonna be a challenge when you have an idea like information theory that uses a natural language word that people are familiar with, right? Information, they think they know what information is, but now we have math behind it. We've talked about information theory a lot on the podcast without ever just taking a breath and saying what it is. So do you have a best way, a favorite way of explaining information theory to the person on the street?

0:05:52.6 CA: I do. It's actually strange that in this case, the mathematical understanding of the word information is actually very similar or very close to our understanding, generally speaking in our normal everyday usage. So to really define information mathematically, or even in words, we can just say that information is that which allows someone who has that information to make predictions with accuracy better than chance, right? So there's a system that I know something about it. And then that means I can actually predict some of its states, for example where I can find coffee with accuracy better than chance. If I would be just randomly go somewhere, well, I would most likely not end up in a coffee place, but if I have information, that means I can actually say, Hey, I have to go to this particular place.

0:06:50.6 CA: And that's how we understand it normally in our everyday language. And it is also exactly how it is defined mathematically. Mathematically, information is that which is used to make predictions, but it has to be better than chance to be information, otherwise, it's just a guess, right? And so because making predictions with accuracy better than chance is really powerful, that means it's also powerful in biology. But of course it's also powerful, let's say in the stock market. If I have inside information, I can get rich using information, because I can make predictions about, for example, the stock price was accuracy better than chance, right? And Shannon, when he defined this concept in 1948, he defined it precisely that way, even though he didn't use exactly those words. And as a consequence, there's a lot of misunderstanding about what information really is. It's often confounded with entropy, which is in a sense the absolute opposite of information. Namely, it is about what we don't know. And so if you then confound information and entropy, then you're going to get a lot of misunderstandings.

0:07:57.4 SC: I completely agree with you by the way, but didn't Shannon himself do that? Didn't he use the word information, like maximal information for him in a code or an alphabet is maximum entropy when every symbol is equally likely?

0:08:14.2 CA: You are right. He has done that. I should say not so much in his original article, but there was a book written or co-written, but I forget. It's Shannon and McMillan or something. Shannon and Weaver. And particular in the book, Shannon and Weaver, the words entropy and information get mixed up. And I have a suspicion that this was mostly Weaver doing that. I would have to actually look at the original article if Shannon himself does that, because in a sense, he should know better. Just because I saw so much of the confounding between entropy information, I felt compelled to just simply write an article called what is information, which its main purpose was to say, Hey guys you have to be careful. One is not the other. [laughter]

0:09:02.9 SC: No, that's great. But I do sometimes wonder... I agree with everything you just said, and I wonder if it's just my physics training 'cause to me thinking like Boltzmann would've thought a low entropy state, if you tell me that you're in a low entry macro state, all the molecules are in one corner of the room, you've given me a lot of information. Low entropy is a lot of information, and that makes perfect sense to me.

0:09:25.7 CA: That's right. Low entropy is a lot of information. You can think of information as simply a difference between two entropies, namely the maximum entropy and the actual entropy.

0:09:36.3 SC: Okay, good.

0:09:37.2 CA: So that's why when your actual entropy is low, then your information is high, because it's this difference. And that the fact that it's a difference, by the way, is a fundamentally important thing. Namely, entropy itself in a sense has no real meaning or existence, just like energy doesn't have in physics, only differences of entropy have a meaning. Just like in physics only differences between energies have any real life effect. And so the fact that it's a difference is important, but it's important that the first term is a maximum entropy, like how much there is to know about the system? And if you forget about that, then your information is just minus entropy or what bunch of people have called neg entropy, right? Which is nonsense. It's forgetting the first term, which is the maximum entropy, so that this thing which is information is always positive.

0:10:28.4 SC: So basically in this way of talking, you would say, when I have some configuration, and I have some statistical knowledge about it, I can divide the state into sort of the entropy part, which I don't understand very well, and the information part which specifies something.

0:10:43.8 CA: Right, or simply is speaking the... You can for every system write down or understand what is the maximum uncertainty that you have. It's just really counting the number of degrees of freedom that there are, and then taking the log of that, right? And then you ask yourself, well, how much entropy do I actually have? In other words, do I have a probability distribution that is different than the maximum entropy distribution? And if the answer is yes, then that means already you know something. That means that the actual distribution has a lower entropy, and then you have some information. However, you may not know at that point exactly what this information is about.

0:11:27.1 SC: Okay.

0:11:28.0 CA: In a gas, of course, very often you do, just like you said, when you have a bunch of molecules in a corner, well, then you have information about where these molecules are. Of course, if you allow the molecules to disappear or to sort of like do their thing, what happens is that your knowledge about the positions of the molecule starts decreasing. And at the same time, of course, the actual entropy starts increasing. That is what we call the second law of thermodynamics. It's literally just the loss of information that you have. And of course, during that equilibration phase, the system is returning to its equilibrium, but it's actually at this equilibrium. So in other words, whenever you have information, it gives you a hint that the system is not at equilibrium. It is away from equilibrium, and the amount of information is in fact characterizing how far away from equilibrium you are. When I'm speaking to you right now, you are actually in an ordered state, a very ordered state...

0:12:28.0 SC: Somewhat...

0:12:28.8 CA: As far as the molecules are concerned. And I can tell you in a sense, how much ordered state there is, it's about half a million bits.

0:12:37.7 SC: Okay.

0:12:39.0 CA: Why? Because in fact, that is the amount of information that's actually stored in your genome.

0:12:42.4 SC: Ah, okay. This is probably gonna be leaping ahead, but what about like the particular configuration of my body? Certainly that is more than half a million bits.

0:12:51.2 CA: That's a good question, but since in fact that organization is done by the half a million bits, my suspicion is that in fact, that must be the same, because how else would you achieve this ordered state? So this is actually an interesting question. I would've to sit down and think a little bit. My intuition is that even though it seems that that would be more than half a million bits, that in fact it is not. However, it's such a good question that I cannot actually tell you right now, why.

0:13:25.5 SC: Good final exam question for the next time you teach a course based on your book.

0:13:28.5 CA: Yeah, that's actually a good point. After this interview, I'll try to sit down and figure this out.

0:13:33.3 SC: Okay. That's a good one.

0:13:35.3 CA: And it's simply because the information cannot be anywhere else. The fact that you ordered, the only thing that makes that possible is the information in your genome. And as a consequence, in a sense, mathematically, it has to be like that, even though, just like you said, that seems low. [laughter]

0:13:54.6 SC: It does seem low. I like to think I convey more information. Certainly in my brain, there are neurons with certain weights and things like that, and there's probably more than a half a million plates there, right?

0:14:08.4 CA: Okay, so now, you're making another point. Of course, there's information that you have acquired during your lifetime that is not being counted. That is definitely there.

0:14:16.0 SC: Good.

0:14:16.9 CA: However, if you take, let's say a brain before it learned something, let's say a baby brain, physically speaking, of course there are differences. Yeah, but like the information is stored in the weights of the neurons in the brain in some way. So definitely, that is in addition, right? However, that of course does not give you your physical appearance. It doesn't change...

0:14:44.6 SC: No, of course, right.

0:14:46.1 CA: The organization of the rest of your body. It's an interesting train of thought, but no doubt.

0:14:50.5 SC: Yeah, either a final exam question or a paper to be written here, I think about this. How many bits am I? But okay, something very interesting and important happened just now in your discussion of information, because I think that if you asked per people on the street do you have information about where your car keys are? They would think about that as a yes, no question. I know where they are, or I don't. And you've sneaked in the idea of a probability distribution, which is kind of central here.

0:15:17.9 CA: Right.

0:15:18.8 SC: So how does that come in? Like how central is probability to information theory?

0:15:23.4 CA: Well, actually it is the central aspect of it. So before you can really understand information theory, you should have a very basic course in probability and statistics. Now to tell you the truth, not everybody who's working in the field did have that, and that shows. But simply because the concept of probability actually, to fully understand it is not as simple as just writing down Ps of I or some distribution. What a probability is, and I know you are familiar with this because you've thought about this in the context of quantum mechanics for a long time, is not immediately trivial, some people think, oh, a probability exists by itself. Other people say no, a probability is just simply something that you can estimate, but that in itself, in a sense, really doesn't exist.

0:16:18.7 CA: But the idea of prediction, of course, is central to that. So when you are asking yourself, can I predict with accuracy better than chance? Then you have to, in a sense, have an idea of, well, what is the chance prediction here? And for that, in a sense, you need a prior, you need an assumption about what is expected, right? And so if you are going to talk about a probability distribution, then you're really talking about what am I expecting? So let's go to the example of the car keys, right? So generally speaking, we could say, okay, here's a person who's looking for their car keys. They know with 90% of the chance from past experience, right? We're setting up the priors now. From past experience, 90% of the time, it's in my pocket, right? But if it's not in my pocket, it could be in 10 different places.

0:17:22.1 SC: Yeah.

0:17:22.7 CA: Right? If that is the case, you can set up, just set up the probability distribution for you, because that means that there are 11 possible places that the keys could be in, right? And the chance to find it in your pocket is 10%, sorry, 90%. And then in each of the 10 other places, the chance to find it is 1 out of a 100, right?

0:17:46.2 SC: Yeah.

0:17:47.1 CA: So now I've set the expectation, before I've done any experiment. Now I'm doing the experiment, and of course, what experiment are you going to be doing? You're gonna check your pocket.

0:18:00.2 SC: First. Yes.

0:18:00.3 CA: And now what happens is that this measurement is going to give you information, right? And in fact, we call this... This outcome of this measurement is called a specific information, right?

0:18:14.2 SC: Okay.

0:18:15.2 CA: It's a specific information because I can do 11 measurements, and the average of the outcomes of that, that's actually the information gained from the measurement of the locations. The interesting thing here is that if I don't find the key in my pocket, then I have a problem. Now, it could be in 10 different places. So it turns out that the specific information of this measurement, namely that the key is not in your pocket is negative, it's a negative specific information, even though the average information gained from the car key searching measurement is still positive.

0:18:52.9 SC: Because you're now less sure of where the keys are than you were before.

0:18:55.3 CA: Exactly. Exactly. So it's intuitively clear. It's like, oh my God. It's like, I thought I knew. See, that's the thing. I thought I knew means that's the priors. From the priors, you had a pretty good idea then you make your measurement and you go like, Oh my God, now, I don't know anything. And this is mathematically, in fact, it's an exercise in my book, in chapter three.

0:19:16.1 SC: Very good. And let me just raise again, I'm super on your side, but it's my job to play the devil's advocate sometimes. People worry about this way of thinking about information, 'cause it sounds a little subjective. It's not out there in the world. You're talking about information as an ability for some agent to make predictions, right? Is it truly objective or is it only relative to the capacities of some thinking being?

0:19:44.3 CA: You are asking, of course, the most important question about information. And the answer is, information is contextual, always. It is never absolute. And now I'm going to give you a great example that will get you to think about this. Suppose you are a virus and you are in fact wreaking havoc inside of a patient. You are replicating fast. Everything is going good for you. Let's imagine your mutation rate is actually kind of low, so you don't change much. So the information that you have about doing your job is doing its job, right? Now, the patient takes an antiviral. The patient takes an antiviral and suddenly this virus goes like, man, this replicating thing is now really, really, hard, right? And what you can now do is you can calculate the information about how to do this job. And it has changed dramatically. In fact, now the virus does not have this amount of information about doing its job anymore. And you go and like, but the sequence is the same.

0:20:50.3 CA: How can a sequence that is the same earlier, have lots of information, and then have very little information? The answer is because the context changed. Because that, what you have information about has changed. So what was the environment yesterday isn't the environment today? You have information about yesterday's world, not about today's world. And that is the important thing. Information is completely contextual. The same sequence evaluated in different environments will have different information and therefore different meaning. A lot of time people are saying, like, oh, information theory can't deal with meaning. It's absolutely untrue. My molecular sequence means something here on earth, actually above the surface of the water for example. You put me underwater and my sequence is making wrong predictions, right? Namely about how to breathe. No, underwater, that thing doesn't work. Your fitness is zero.

0:21:50.0 CA: I will have zero offspring underwater unless I actually am in a submarine or something like that, where I'm taking the environment with me, right? Then I could still have information. So very nearly so, in biology, information and fitness are exchangeable. You have high information about how to do something, how to survive in the environment means having high fitness. They are related. In fact, they are mathematically related. And we don't have to talk about what that relationship is, but you can read about it in my book. So, in other words, what happened in evolution is that fitness and information are essentially intertwined. They are almost the same thing. And so the way you are increasing your fitness is you're increasing the information about how to live in that environment.

0:22:39.0 SC: I guess I was trying to make a similar point to what you just said in my book, The Big Picture. I mentioned the Voynich Manuscript, right? This wonderful old book full of symbols, and nobody knows whether it's nonsense or whether it refers to something. And so I asked the question, how much information is there in the Voynich Manuscript? How would you answer that question?

0:23:03.0 CA: The answer is, unless we know what the information is about, this is entropy for us, right? So if you have text or any data, that is not information, okay? In fact, it is literally just entropy in the sense that you are talking about. Oh, it represents an example out of another similar set of things, right? And so you're setting up your hypothetical ensemble so that you can talk about an entropy. The moment that you know what things you can predict with it, then it becomes information. If you do not know this, the same piece of information or the same text, should I say, the same symbolic sequence, would just be entropy to you. In other words, just to repeat it, because information is that which allows you to make predictions about a particular system with accuracy better than chance. If you don't know what this is, it's not information. Another example is somebody hands you a subway chart, right? And he goes like, well, that's fantastic, except I don't know which city this is. Well, then it's useless to you. And if it's useless. So in other words, useless information is entropy.

0:24:13.6 SC: Got it. The other idea at the technical level, that I have found super important in my tiny little forays into information theory is mutual information. I love this concept, and it never gets explained in a popular level. So maybe you can give us a shot.

0:24:28.3 CA: Right. So, first of all, information is really a mutual entropy. It's a shared entropy. So, in other words, the reason why me... If I have information, can make predictions about another system is because in a sense, we have certain correlations. We know something about each other, right? And so information, even though I showed you earlier that it's really a difference of entropies, you can also think of it as a shared entropy. The mathematics to show that this is one and the same thing, we're not going to go into. But just trust me here that if you think about a Ven diagram about a system and the one who's trying to measure the system, the intersection of that Venn diagram, that's the information that the measurement device has about the system, right?

0:25:20.1 CA: That's a shared entropy. That shared entropy is, of course, the information. So when people say mutual information, they're meaning mutual, mutual entropy. So in fact, nobody should be using that word... Okay. Like in my book, for example, when I have like an index where it says mutual information, I say, see information, because they are...

0:25:44.5 SC: Fair enough.

0:25:46.1 CA: But it's just, for the reason that we talked about earlier, that people are using information and entropies anonymously. They have taken mutual entropy which isn't information, and called that mutual information. But in principle, they're one in the same thing. We just should call it information that correlates the two systems. Because, remember, we always have to talk about two systems when we are talking about information. One that makes the prediction and one that is being predicted. And what is shared, the shared correlation between them. In other words, what's not random between them. That is information.

0:26:27.1 SC: Good. So now we have a pretty firm grounding on information theory. We can move on to applying it to biology, which is what your book is all about. But let me just first ask a background question. How popular in biology is this task of thinking about things in terms of information theory? Is it everyone does it or no one does it, or it is the new hotness and it's sweeping the field.

0:26:54.5 CA: Let me put it this way. If this was something that everybody does, I wouldn't have had to write this book.

0:27:00.4 SC: Okay, good.

0:27:01.3 CA: This book really came out of a frustration that I have this tool. I see how valuable it is to understand essentially anything in biology, and nobody is using it. When I say nobody, that's not quite true. There are a few people who do, Bill Bielich in Princeton, for example, has certainly done this, and he does know how to do this. There's a few people often, in fact, in Bielech's orbit, who have looked at information transmission in cells, for example, in gene-regulation, and calculated the channel capacity of a cellular communication channel and did this very well. And I, in fact, have these examples in my book, but it is woefully underrepresented. There is something about information theory that presents a barrier, a hurdle, in a sense, to acceptance. And I only have a vague memory of when I first learned about information theory, which was, in fact, at Caltech, when a postdoc that was placed into my office actually first started explaining classical information theory to me because we were actually working in quantum information. And it was like, well, let's do the classical theory first.

0:28:18.6 CA: And I remember when he was first writing these formulas on the board, which has these vertical bars and these columns and semicolons thinking like, what is this? This is not my mathematics. My mathematics, there's integrals, there's differentials, there is... There are matrices, and now I'm seeing these weird symbols, and my brain doesn't really understand them. In a sense, from the cognitive science perspective, is you don't have these representations that allow you to recognize these things, and they have to build over time. And I think this barrier that I felt at the time and was like, oh, there's all these different entropies.

0:28:58.8 CA: There's a mutual, there's a shared, the conditional. And you were like, how am I going to keep them apart? And you're going to have to get used to them, just like you got used to integrals and differentials and things like that, right? If I ask anyone who had algebra, I think calculus. It's like, would you ever confound differentials with integrals? And he goes like, of course not, right? Yes. Because you got used to this stuff, right?

0:29:22.9 SC: That's right.

0:29:24.2 CA: So there seems to be this barrier. And in particular, very often information theory is perceived as being an engineering discipline, right? And they're going like, well, the engineers, well, they want to do error correcting codes. What does this have to do with biology? In fact, I've seen this in print saying the engineering discipline of information theory has nothing to say about biology because these are very different things. It's like saying, well, information theory can't apply to physics, when, in fact, we do know very well that it does. In fact, I just explained the second law of thermodynamics in terms of information theory, right? And so this barrier is very palpable. So if somebody is an established scientist and you tell them, well, you know what? You should really be using information theory, they're not going to do it. They're entrenched in their ways, and they usually are not receptive to learning, in a sense, a whole new bag of tricks.

0:30:21.3 CA: It is not that different from what happened in black hole physics, where they constantly talk about information loss in black holes. And not a single paper is trying to calculate the capacity of the black hole channel. Except, of course, me. But that's because I know information theory. But these people have been working in this area for 30 years and not picked up a single book or article on information theory, again, there's this barrier, they're saying like, this concept is different from the concept I need, and the answer is no, it is the same concept.

0:30:54.6 SC: I'm a little bit sympathetic because I know when I read papers in economics, they have this habit of denoting variables in equations by words rather than by single letters. And it's so trivial and silly, but it truly rubs me the wrong way, and I have trouble wrapping my brain around it.

0:31:16.4 CA: No, I fully agree with you, because that is not how we write equations. I still almost vividly remember this aversion where I'm like, this is not my thing. But you have to power through this, right? And in my lab, everybody gets the basic introduction to information theory. And then once you've sort of internalized it, you cannot see the world except through that lens of information theory. It's like the hammer that makes everything look like a nail.

0:31:50.1 CA: And in my book, in fact, I sort of co-opt the famous saying, which we all know in biology, which says, nothing in biology makes sense except in the light of evolution, which certainly is true, but I basically have changed it around to say nothing in biology makes sense except in the light of information. And, of course, information is that which evolves, really. I mean, people would say, oh, organisms involve. No, what evolves actually is information. The information is what is essential in an organism. The organism itself isn't really that essential. In a sense, it's replaceable, right? Like if we have offspring, well, they carry most of the information with them, but only the encoded information. Of course, what we think as specific to ourselves is, of course, the stuff in our brain that we have talked about about 25 minutes earlier, the information that we acquired over a lifetime is, of course, making us special. The information in your genome does not make you special, but it is what makes you alive.

0:32:50.6 SC: So is it fair to think of Charles Darwin as an early information theorist?

0:33:00.3 CA: Well, I would say no, simply because he didn't really... And he had no way of understanding that the basis of inheritance was really the replication of information. He didn't even know how any of this stuff was encoded because that was discovered in 1958. Or should we should say that John von Neumann kind of figured it out a little bit earlier when he came up with his theory of self propagating machines which essentially could have told, Watson and Crick, where to look for it, stored information. So Darwin did not know that, but he had, via his [0:33:46.6] ____, in a sense, going on boats and looking at stuff figured out, hey, there's variation going on, there's selection going on, and there's inheritance going on. The fact that these three things are properties of information. So inheritance being the replication of information, variance being the mutation of information, the changing of information, and selection is the meaning of information. In other words, those pieces of information that have a lot of information are fitter and therefore will have more offspring. And therefore, in fact, because I told you about the relationship between information and fitness, then the meaning of information is selection. Right? So we can think of the entire evolutionary process in terms of what happens to information, and we should, in fact.

0:34:35.7 CA: But Darwin did not think that way, even though, of course, I haven't read all of his works, even though many of them are in my little natural history collection, in my bookshelf, but they are big, big tomes. I mean, one of the things that is so astonishing to me is that Darwin had in his head essentially a twelve book treatise and then was forced to publish the abstract of it as the first book, which is now the Origin of Species. But he had so much more to say and then over the rest of his life, in fact, said many of these things and of course, most of those things, like about worms and about plants and things like that.

0:35:18.1 SC: Barnacles.

0:35:18.2 CA: We don't usually read. But in there, if you study those volumes, maybe with more attention to the concept of information, and the idea that what makes these plants how they are, in fact, is making predictions. But here's in fact a good example. Darwin himself at some point noticed a particular orchid which had a very, very long neck, which he knew had to be pollinated, right? And it was about 30 centimeters. What is that in... It's about a foot...

0:36:00.8 SC: About a foot, yeah.

0:36:01.0 CA: Or even longer, in fact. But basically he said, now I will make a prediction. I will predict that there is a pollinator exists with a proboscis or a nose, what they're using, which is about that length, right? And even though he didn't live to see that prediction come true, in fact his competitor in the evolutionary field that almost scooped him, he in fact basically said, oh, and I'm going to tell you that this must be a sphinx moth and in fact they later found it after Darwin died and that was know as you had predicted it. So to some extend, he knew that evolution, this idea, the theory of evolution is a predictive theory, right? Even though he didn't think of it in terms of information theoretical terms.

0:36:51.7 SC: So yeah, a typical evolutionary biologist who tried to explain things in terms of there's a population with certain traits and there's a fitness landscape, and they move towards peaks of the fitness landscape. And so you're not undermining that, you're just saying that a useful way of thinking about fitness is having the information to successfully predict what's going to happen in your environment and therefore survive.

0:37:16.0 CA: That's right, I don't undermine any of the standard population genetics. It is really just a very different way of understanding what fitness is. After all fitness, the way... It is not the same word as we talk about physical fitness, of course. Fitness in biology means fitting your environment well. It means being adapted to your environment, which means corresponding to. And see we're seeing "Aha". That means in a sense that your body structure is predictive about what world you live in. So for example, E. Coli Bacteria, they grow best at 37 degrees Celsius. Well, that's weird, isn't it? Well, no, it's not weird, that's the temperature of our stomach. So basically their molecular biology makes a prediction about what environment they're in. Namely one at 37 degrees Celsius. If their prediction is off, ell that means in fact, they don't have as much information as they think and they're actually not gonna grow as fast. So this correlation between the genome and the environment gives you the fitness because it tells you, you're fitting this environment well. You're well adapted to it. And that concept of information is precisely that this correlation.

0:38:39.6 SC: And this is a very important point 'cause it reminds us that just so people don't get the wrong idea, information isn't necessarily conscious. It's not something that you might say you know, the information in your genome is very well adapted to your environment, even though you personally might have no idea what the arrangement of nucleotides in your DNA is.

0:39:00.7 CA: That's right. So obviously, when we're talking about the fact that cells make predictions about their environment, which they do all the time, because cells have to make decisions. It's not like they have a brains. But to some extent we also know that, because once you understand that information is just a correlation, which is a non-random correlation, because sometimes you get correlation by chance. But no, we are looking for correlations that are not by chance and that are on top of that being maintained. Because in thermodynamics, you might have correlations by chance, but they're going to disappear very soon. The genome is making sure that these correlations are being maintained so that we can continue using what we have to make predictions about the environment. If you would be loosening this continuous maintenance, then the information would go away. And we call that death.

0:39:55.8 SC: Is that what we call death? Good. Now that we know, but this does bring us to a very fun point you make in the book, is that we can think of evolution as a kind of Maxwell's demon. And maybe it would be fun to explain what Maxwell's demon is. Not everyone knows. And what this has to do with evolution?

0:40:16.5 CA: Right. So that's a good point, and thank you for leading me into that, because it's one of my favorite parts of the book, actually. So Maxwell's demon, let's not talk about why this sort of devil was invented, but let's focus on what he does. So the Maxwell's demon basically is sitting at the intersection of two, let's say boxes. And there's a little window inside of the box. And now both boxes have gas in them, gas molecules who have different speeds, by in fact, described by the Maxwell distribution, it turns out. And now imagine that this demon who sits there, he operates sort of the door between the two boxes. And he also has a measurement device that he uses in order to measure the speed of a molecule, for example, one that is about to go through the door. And then he goes like, okay, if this is a fast molecule, I'm gonna let it go through, but if it's a slow molecule, I'm gonna shut the door. And so that it is in fact going to be reflected. If he does that a lot, he's going to have one half of the two boxes, I mean, or I should say the left box as opposed to the right box with lots of fast molecules. And the other one will be stuck with all the slow molecules.

0:41:36.7 CA: That is in violation, in apparent violation, I should say, of the second law of thermodynamics, which basically says, no, that can't happen. Namely the formation of a non-equilibrium situation from equilibrium. So mathematically, it looks like what the demon has achieved is in fact creating order or violating the second law of thermodynamics. The fact that he didn't do such a thing was in fact proven mathematically and fully correctly later by... What's his name?

0:42:11.8 SC: Landauer.

0:42:14.5 CA: Landauer. Rudolf Landauer. A German-American physicist. And we're not gonna go into how that proof goes, but it is not a violation of the second law of thermodynamics. But this idea that via measurement you can actually reduce entropy is of course a very common one because measurements give you information and information is sort of the opposite of entropy. So yeah, if you, for example, look into a room of molecules and then essentially measure the speed and position of all the molecules you're looking at, you could in principle achieve a lot of order because you could punch them, so to speak, with a laser and then all go them into one corner.

0:42:53.8 CA: So in other words, measurements do allow you to decrease entropy. Now that we have described the Maxwell's demon, now let's think about the Darwin demon. So in order to understand this, let's imagine that instead of these measurements that the demon does, we are basically doing... So the molecules are now mutations, okay. So you have a genome and a mutation happens. The mutation happens, it's completely random. It could actually increase your fitness or it could decrease your fitness.

0:43:25.0 CA: And now the demon basically says, "Look, I'm actually closing the door and not allowing the decreasing fitness mutations to persist, but I'm gonna keep those that actually are increasing my fitness". So in a sense, the organism now performs a measurement, but it turns out of course, that the mechanism of evolution is precisely that way. Namely, the deleterious mutation will make it such that the organism carrying it doesn't have as many offspring, it's less fit. And as a consequence, its frequency in the population will go down, or else, even if it's a lethal mutation, will disappear. Whereas the beneficial mutations will in fact be enhanced. So because the beneficial mutations create order, because like there's now a type that increases in frequency, maybe even very fast at the detriment of all of the other sequences. Well, that creates order in the population. And it's essentially because you learn something about the environment, you've extracted information about the environment by having this measurement of the environment. So in other words, these mutations that are being kept, they are measurements of the environment.

0:44:45.9 CA: So it's like, "Ah, 37 degrees Celsius. Okay, so let's adjust our genome in such a way that we grow well at 37 degrees Celsius". Whereas the deleterious mutations, of course are misinformation, and they reduce your fitness and as a consequence are thrown out of the window. That's the Maxwell demon closing the door on them. And what that means is that, well, if this is a continuous process, then it should lead to a constant increase of information inside of an organism over time as evolution proceeds. And therefore we can actually formulate this as a law, namely the law of increasing information in evolution that actually predicts that if you start out with a low information beginning sequence, that over time the information must increase, but not all the time. For example, if the environment changes, well, we already discussed that, then your information drops, and you have to sort of relearn things.

0:45:51.8 CA: There are other ways in which information can get lost. For example, recombination can actually destroy information. It can also lead to a good thing. So this natural demon, this Darwin demon is not perfect, in a sense it's leaky. It sometimes makes wrong decisions, where information can actually go down instead of up. But overall, on average, this theory of evolution predicts that the amount of information that is stored in a population of genomes has to be going up. And that is very interesting because we've been searching for a way to understand the bio-complexity that we are seeing around us, and asking how on earth is all of this possible? And can I understand if one organism is more complex than another? And this idea of complexity, of course it's difficult to put in mathematics, but in fact, it turns out that complexity is literally just information. In other words, information or complexity is just a proxy for information.

0:46:58.4 CA: Why? Complexity essentially is stuff that is complicated, but helps you do something interesting. We don't associate complexity with something that is just structurally interesting, but actually doesn't do anything. What was called a spandrel by Gould and Lewontin in biology. Something that looks very complicated once we understand, it's actually just, let's say an icicle, like, "Oh, it's not really that complex". But there are really complex things, for example, our brains, but that is in fact reflected by the information necessary to make it. So it turns out therefore that information is really the correct way of measuring complexity. And therefore the question of is complexity increasing in evolution is simply answered as a yes. As long as you understand that complexity is really information, and the natural demon is responsible for that law of increasing information.

0:47:56.3 SC: Good, because naively we might look at an organism and try to figure out, "Oh, that looks complicated". "That looks pretty simple". And we wonder why all this complexity has grown. But you're saying that that's just a pale reflection of what's really going on underneath the hood?

0:48:12.0 CA: That's right. But it's almost like saying, if I look at a sequence, somebody's genome or some animals genome or a bacterial genome, I can't immediately see, well this is information, this is not information. Here's some information. They look the same. So we need to figure out how they're making predictions or in other words, how are they functioning in the world? So if you have something that looks complex, then you have to ask yourself, let's observe this thing and see whether what looks such an... As an intricate mechanism is in fact necessary for survival, or whether it is just something that is a consequence of something else. We see sometimes like these complex display behaviors in animals, and they are obviously important in mating. And if you would remove them, in fact, your fitness drops to zero because you don't get to mate, zero offspring, zero fitness. But then there are other things that are in a sense just... They're not important for the actual survival. You would take them away. And it doesn't change the fitness. It's like saying you're removing a section of the genomic code. But if that is not predictive of, if it does not carry an information, it is not important for fitness, you can take it away.

0:49:26.7 SC: If I understood what you said correctly, we don't right now have a clear cut way of looking at a genome and saying, this part contains information, this part doesn't?

0:49:39.8 CA: We actually do, but only if we're looking at many of them not a single one. If I have a hundred versions of a gene, there are regions in it that are unimportant for survival, if I then make an alignment of them, then I can recognize, "Oh, look, this changes all the time. This position. Clearly that doesn't mean anything 'cause you can just willy nilly change it. But then you go like, "Oh, but look at this section. It's the same, everywhere, the same for every organism". I bet you it is because when you change it, you die. And that's why I haven't seen it. If you would make it, it would simply die and therefore you... It doesn't enter your database. So you have to do these multiple sequence alignment as they're called in bioinformatics. And when you're looking across, column by column, and you see some columns being completely conserved by evolution, even though evolution constantly tries to change them. Because through the mutation process, and then you see columns which are willy-nilly, like on the nucleotide level you see ACs, Gs, and Ts with 25% frequency. And you go like, clearly not information.

0:50:49.4 CA: And so that is in fact the basis of the algorithm that I describe in the book that allows you to measure the amount of information in a sequence. But if you do not have a multiple sequence alignment, in other words, a bunch of examples, then you cannot understand what is information because everything looks the same.

0:51:06.3 SC: Yeah. And this reminds me that you of course have collaborated with your colleague Richard Lenski on his long-term who's the boss of the... Was of the long-term evolution experiments. You've actually seen data like this.

0:51:20.6 CA: That's right. In fact, we constantly see this, we see this more easily in digital life experiments where we have self-replicating computer programs that live inside of a computer, just like the one I'm staring at right now, and that you're staring at right now. And there we can extract these sequences and we can look precisely where the information is. And in fact, we can measure the information and see as the fitness increases, the information increases. It's literally, if you would be taking the log of the fitness and superimposed it with the calculated information, they're almost on top of each other, because it's the log fitness, which is basically the growth rate, which is in a sense mathematically equivalent to information. So if information is zero, your growth rate is zero.

0:52:06.3 SC: So my own biological knowledge is not very big, but I do recall understanding that the human genome is not the longest out there in the ecosystem.

0:52:18.5 CA: That's right.

0:52:19.4 SC: Is that something we understand?

0:52:20.4 CA: You've addressed an important question, and it's in the biological literature is called the C value paradox. If you just look at the length of genomes, there is an amoeba out there which has 200 times more DNA than we have. It's called amoeba W actually. So it's almost like the w in a former president. So, you can think of that as you may. So clearly genomic content isn't information, but if you would take the genome of let's say a thousand of these amoeba and then align it, then you would see that most of the stuff is completely meaningless or else in fact, it might not be haploid. So in other words, you have a hundred copies of the same small piece of information. And here's another thing that you need to understand about information. If I have one book, okay, that let's say is predictive about certain things, and then you have a copy of the book that's not twice as much information, it is exactly one time as much information.

0:53:25.5 CA: So if you have a hundred copies because you have a hundred copies of your genome, like we for example, have two copies so we are diploid, there's many other organisms like in particular in plants which are tetraploid, they have 16 copies and so on. So this, amoeba probably has over a hundred copies of its own genome. So in other words, clearly this amount of DNA you have to divided it by hundred, that still wouldn't get us to the 3 billion that we have... 3 billion per chromosome, so...

0:53:55.4 CA: And it's not 6 billion, 6 billion base pairs, but in fact, out of the 3 billion, nucleotides, which in a sense give you a entropy or potential information of 6 billion bits, but it's potential information, which is the same thing as entropy. Only a small fraction of this actually encodes information. And from the encode project, it is probably about 8%. And that's how you're getting to that number of 500 million bits, which is about 8% I think, of 3.1 billion or so.

0:54:37.0 CA: And so, there are some plant genomes that are enormous because of this explosion. But if you have several copies and if you have... In plants for example, there's lots and lots of intergenic regions which are meaningless. In other words, there's no information in there. Plants are sort of spectacular for that. They're far more intergenic stuff than humans for example. Some of them have up to 70% intergenic stuff, and this intergenic stuff is probably also riddled with transposon. So literally, if you would be taking it out or mutating it, it would not change what the plant does at all.

0:55:13.1 CA: So there is no information encoded there. So if you really want to plot just information content in fact Eugene Koonin who works at the NIH, he has done that. In fact, his plot is in my book. And you find out that yeah, bacteria are sort of very low there. And humans, as far as we can measure, in fact have the most information. Mammals generally, but humans the most. So this idea that humans are in a sense, the apex of complexity that some people think is sort of very self-serving, that might not be wrong. Certainly if you take a look at it through the lens of information.

0:55:52.0 SC: Yeah, I believe that. That number went by pretty quickly, so I want to get it right. 8% of the human genome is information bearing?

0:56:01.4 CA: It's coding, yeah.

0:56:02.5 SC: Okay.

0:56:04.1 CA: And the rest is intergenic stuff or repetitive stuff. In other words, it's uninformative.

0:56:08.3 SC: It seems a little inefficient. I don't know.

0:56:11.5 CA: Well, I mean, it's because there is so much... For example if you have a gene, there are regions that are called introns, which just have junk in them. And they're being excised before the stuff is being translated into proteins. And so there's a vast amount of that. And then inside of chromosomes, there's vast regions of repetitive stuff. Some of it might even serve structural features so that molecules can attach to it. But it does not actually provide information about surviving. So obviously, it is difficult to measure the information content, because we would have to have thousands of human genomes that you can align. Now, we have, in fact, now thousands of human genomes, but nobody's actually tried to do that. And I should say that in order for a good ensemble to really be reflective of the information, we needed to have a good amount of variation. In other words, those nucleotides that are mutatable, they should have had a chance to mutate. But because we humans have a relatively common recent ancestor, there's a lot of conservation in the genome that is common through common descent.

0:57:33.9 CA: And that looks like it might be information if I'm making such an alignment of 1000 genomes, but in fact, might not be. So one of the things you might want to do is you want to take a lot of African genomes for this alignment, because there's much more variation in the African genomes, because they have in a sense been evolving longer and their common ancestor is much, much, much earlier, which is why there is so much more variation. And that would be a much better estimate as if you take European genomes and make that alignment, because really we're looking at like 100,000 years of variation, which given the population sizes might not be enough to really reveal the information content. So if you have a population that has a fairly common recent ancestor, then even this alignment is not going to do a very good job of giving you the information. But very often in proteins, single proteins, for example, or viral proteins, which are evolving very quickly, you can actually measure this information.

0:58:32.9 CA: And I've done this, for example, for the HIV protease, which is a 99 amino acid sequence. And then see, for example, like the example with drugs that I gave you. But when you when you introduce anti HIV drugs in the population, you see the information dropping very rapidly. And then because we have data over time, you can see how as the virus evolves drug resistance, it really learns about this new world that it's being cast into. And you see the information over 10 years, really approaching the pre-introduction levels. So for certain proteins and molecules, it's much easier to actually do this analysis and measuring information. Then, for example, to do it for a whole genome. In principle, it can be done. But ideally, you would do assays where you're trying to mutate every nucleotide and see how it reacts. And that's, of course, not feasible right now. So it would be unethical, by the way.

0:59:35.1 SC: By the way. Yeah, that happens in other podcasts I do. It'd be fun to do certain experiments. You just can't really do it. But one way of saying what you just said a little while ago is that the human genome has not had time to equilibrate in some sense.

0:59:49.7 CA: That is exactly what I'd say.

0:59:52.0 SC: Good.

0:59:53.7 CA: So there's plenty of unexplored variation in it.

0:59:56.5 SC: Because I was going to ask is are there sections of the human genome or for that matter, other advanced genomes where there is some continuity over time. So clearly, this section of the DNA is carrying information, but we don't know what it does.

1:00:13.4 CA: Absolutely. One of my favorite examples is that there are untranslated regions in the genome that seem to be completely conserved. And it's like it's untranslated, which means no protein is being made. Well, an RNA is actually being made. So in fact, these are long stretches of RNA. And what they do is they can form molecules called ribozymes. And it is a fairly recent discovery that these in fact, are actually very important in understanding brain function. So there are these long non coding RNAs. And remember, DNA has this neutrality in the third position of a codon. So for many of the amino acids, when you mutate the third codon, because every amino acid is encoded by a triplet, you can change it willy nilly, doesn't change the amino acid. So there should be neutrality in that position. And you all you see this throughout the genome. But then you have these non coding regions where there's no neutrality whatsoever. Well, yes, because if you translate it into an RNA, RNA doesn't... The ribozyme, there's no codon translation is not made from amino acids.

1:01:24.1 CA: So in fact, every nucleotide is important. And that's how it comes that these vast, long non coding RNA regions are completely conserved, which means that they carry 100% information or close to that. They're not 100% conserved. But that's because sometimes in these ribozymes, because they're folding into structures, you can make a mutation on one nucleotide, as long as you're making the complimentary mutation at the corresponding one. But this is a fundamental discovery, because it means that these ribozymes that are super important in the survival of the organism. And in fact, may explain... So they did this study for in brains of flies. That they are super important in controlling mating behavior of the fly, which was previously not understood. And you can find a protein that does it well, because done by these these these RNA sequences.

1:02:19.5 SC: That's actually super interesting. And I pretty sure I didn't know this. So this the central dogma in molecular biology is that the DNA contains information that gets transcribed into RNA that tells the ribosome which proteins to make. And what you're saying is that that's not all the DNA does, there's a separate function where parts of it get transcribed into RNA, and then don't make proteins, they make ribozymes instead. And those also play a crucial role.

1:02:49.2 CA: That's absolutely right. So whenever you hear the word dogma, then you're always going like, Well, okay, so I'm sure it's not really a dogma. So let's actually look for the violations of it. And that's always a fruitful avenue of investigation. So yeah, in fact, the role that RNAs play in the molecular biology of the cell cannot be underestimated. I mean, these are highly functional molecules, they modify DNA. If you didn't know that there's a whole bunch of gene editing going on, certainly in single celled eukaryotes. And I talked a little bit about that in my book about the trypanosomes, a fantastic, absolutely stunning story of how how these RNAs are ensuring, in a sense, the survival of genetic information that should have died 100 million years ago, in a sense.

1:03:44.3 CA: And so these molecules, because they're also regulators. So the RNA molecules, the ribozymes, which are untranslated are so functionally important that we cannot understand, really, the molecular biology of the cell without their action.

1:04:00.3 SC: Interesting. And that's actually a great segue into the sort of the final topic that I wanted to hit here. It would be a shame if we didn't talk about the origin of life. And of course that there's a lot of people who put RNA front and center in that story, but we don't know what the right story is for the origin of life. So how can information theory help us here?

1:04:18.9 CA: Right. So, that's a great segue. Because in fact, the current thinking is that prior to DNA, we know that DNA and proteins cannot have been at the very origin of life, because you need proteins to, in a sense, understand how DNA is acting. And you cannot understand one without the other. So it cannot be, you cannot have this separation between storage material, which is DNA, and machine, which is the proteins, as in the origin. Now, it turns out that RNAs, they can store information, because they're really the RNA nucleotides are cousins of DNA, they just basically have an oxygen compound more. That's why the RNA is called ribonucleotides. And the DNA is called deoxy, namely removed, ribonucleotides with an oxygen removed. That difference is actually very important, because it affects the stability of the molecule. And of course, if you're going to do information storage, you want to have a stable one. And that's the DNA.

1:05:29.1 CA: However, the RNA molecule, which, like, it's almost like a DNA, but can actually conform in all kinds of different ways. They can play the role of a machine, the ribozymes that we just talked about. So, according to the RNA world framework, you're going like, hey, how do you have something where information and machine is one and the same? And that would be the RNA world. So in other words, self replicating RNA molecules, like where they store information, and are folding into the machinery that replicates the information.

1:06:04.5 CA: We, as you correctly pointed out, we don't know if that is the ancestor, because the biochemists tell me that there are some fundamental problems of how this whole stuff could have worked out. And that's a problem in biochemistry, which I only lightly touch upon in the book, because I'm certainly not an expert on that. But the important thing to understand the origin of life is to first understand what life is. And what life is, is basically information, a string of information, or a string that contains information about how to make copies of that information.

1:06:46.4 CA: You might think, oh, but that's not so difficult. I make copies of information all the time. I have a copy machine. It's like, yes, but who built the copy machine? The copy machine isn't some random thing. If you just throw a bunch of piles of metal together, it's not going to form a copy machine. In fact, it takes an enormously complex set of instructions to make that copy machine. So, yes, it makes copies, and it's great at doing that. And you could, in fact, make copies of the blueprint of a copy machine. But that copy machine isn't going to make the copy machine from that blueprint. It's not. It makes copies. That's it. But if you had a machine that is, in fact, built from information, and therefore, then once it's crumpled up, in a sense, into this ribozyme, then makes copies of the actual information, well, that's life. So, life is information that copies itself.

1:07:48.7 CA: But if you just have a little bit of information, it is not going to do that. Because, like I said, to create a machine that can make copies of information, you need a minimum amount. Five bits is not going to do it. Maybe 100 bits is not going to do it. In fact, we don't know what the smallest amount of information is that has a capacity to replicate. You know why? Because we also don't know what the environment is within which this replication happened. Because I told you that the information is really information about the environment. So, a string of 100 bits, that's 100 potential information. But depending on the environment, it could be a lot and it could be a little. And so, we don't know a whole lot about the very early Earth. But there could be very many different environments. But what we do know, and I can say that unequivocally, is that there is a minimum amount information. And that amount of information cannot be obtained by chance. Under no circumstances.

1:08:50.3 CA: Well, I should say, mathematically speaking, if the likelihood of finding any of these molecules would be chance. However, there are in fact mechanisms that could make an avenue where information can actually gradually accumulate without a Darwinian process. This is speculation. But think of it this way. We can imagine a process where a molecule that carries no information, but is in fact a sequence in some alphabet. Let's imagine even it to be an RNA. We can imagine that there is a process by which it can be replicated, but in a sense, passively. For example, it could be laying on a clay surface, and then you can basically have a process that will make sure that only complementary bases are being added. Then basically it becomes copied, but there's no information in it. So, this copying of information, or of a sequence, you're copying entropy, really. But now, because the copying machinery is actually in the hardware, not in the software. It's this clay, and it's the process of of complementation.

1:10:16.5 CA: Now imagine that if, in fact, the rate of making errors in the complementarity can sort of vary. So, it turns out that the speed of this operation of copying varies depending on how many errors you are making in that complementarity. So, if for some reason you have something that makes this faster, well, that's actually a good thing, and you're going to make more of these copies, even though you might think there's no information. But now imagine that what makes this faster, which means makes it more accurate, because they are the same thing, is actually something that is part of that information that sort of like helps and binds. Like something breaks off, and then sort of goes like, hey, I'm using that broken off piece to actually make that process more error-free, and therefore faster.

1:11:15.0 CA: What just happens then is that information just seeped into the genome, because remember, that piece that broke off is actually one of your copies. And you can imagine a process like that, that slowly but surely, information about how to do the replication moves from the hardware into the software. It's a non-equilibrium process, but it is at least imaginable, because you have to reach this limit until there's enough information in the sequence that it autonomously replicates without the support of the hardware. That is my sort of imaginary view of the origin of life, but it is a hard thing, because you have to reach this limit. Let's say the limit is 200 bits, right? Only after 200 bits can evolution take hold, which then means I can have the evolutionary process increase this amount of information to infinity, theoretically. But before you have achieved Darwinian evolution, any information that is in the sequence is now highly fragile. It is basically prone to deteriorating due to the second law of thermodynamics.

1:12:24.1 CA: So you basically have to have a process where you are constantly working against this law, but it is because information from the replicative machinery, the hardware, is seeping into your sequence. If this happens fast enough, you might actually get to this point where Darwinian evolution can take place, and therefore you separated yourself from the second law of thermodynamics.

1:12:52.8 SC: Okay, good. I actually think I understand that. Let me try to run it by you again and see if I got the right idea. We talked about Maxwell's demon and evolution, and that was pretty clear to me. If you just had a set of genomes and you just acted randomly on them, they would randomly walk through the space of all possible genomes, and the entropy of that distribution would go up. That's the second law of thermodynamics. But you're saying that selection kicks in in the case of evolution, and that basically acts like Maxwell's demon to lower the entropy by sort of nudging all of the genomes towards areas of higher fitness. And that's pretty darn efficient. And you're suggesting that maybe the origin of life was a similar mechanism, not quite the same and not quite as efficient, but something that was a little Maxwell's demon-y to allow the distribution to become lower entropy until Darwin could kick in.

1:13:49.8 CA: That's exactly right. If you think about this idea that if you're making fewer mistakes in the copying process, that the copying goes faster, and you're therefore making more copies, that means that there is a Darwin-y mechanism at work, even though there is no self-replication, but there is replication. But it is replication of non-informative things, which by this mechanism, while some of the stuff by chance happens to increase your speed of replication, while those are being maintained. And that's this idea, yes, the information about to replicate is in the hardware, not in the software, but it slowly starts seeping in. Slowly or fast, it really depends on the environment that you're in. And of course, we have no glimpse of any of that, because if you imagine that, let's say, it's actually a process that isn't that difficult to achieve. Well, wherever it constantly keeps occurring, it would be dwarfed by any form of life that is actually there that has perfected this for four billion years.

1:14:56.6 CA: So there are these dark smokers and these underground, what are they called? These pure hydrogen, basically. They are these underground volcanoes almost.

1:15:08.4 SC: These underlying vents.

1:15:12.8 CA: And there is all kinds of interesting bacterial life going on, and who knows what else goes on there? There are some theorists that claim that these types of environments are perfect for the origin of life. But even if it would be constantly reoccurring there, we wouldn't see it because bacteria have already colonized.

1:15:28.8 SC: Right. Okay, that is great. That's a very interesting set of ideas. I hope that it's followed up on. But we're near the end of the podcast, so we are allowed to think big now. It does lead into, I guess, the last question, which is, human beings have developed capacities for manipulating information that didn't exist before human beings came along. We can not only think, but we can learn. We can teach in ways that other species don't. We can store information and replicate it in books and on computers and so forth. Is this analogous to the origin of life? Is this another way that information is reproducing itself? Are we seeing a phase transition to a new kind of information phase?

1:16:16.9 CA: I think you're right about that. Just imagine the advance of being able to write down what you have learned so that other people can build upon that. That obviously in a sense, I'm saying obviously, but maybe that's not obvious to everyone. But obviously that is the magic power that we have acquired as a human species that no other species has. The best that other species can do is to somehow teach young generations about certain behaviors, hunting behaviors, or maybe even tricks that you can use with rocks and sticks and things like that. But to actually figure out mathematics and write it down into a book and then teach that to other people so that they can, for example, predict the orbit of a planet 100 million years from now? I mean, if that is not mind-boggling, then I don't know what. If you think about it, what we call our ability to make predictions over long distances, as far as humans are concerned, in a sense, we call that intelligence. We're making intelligent decisions so that we, in a sense, can survive better.

1:17:30.5 CA: But in fact, clearly, making predictions about what the environment is tomorrow is going to be an important element in my survival. It was an important element for survival for any hunting species that goes like, oh, I predict that this prey is going to be at this location in about 10 seconds, and if I can make that prediction very accurate, I will catch that prey. Now, if you think about intelligence as the temporal radius at which you can make predictions with accuracy better than chance, well, then you realize, well, bacteria make predictions all the time, but they make predictions about the next second because they're making predictions about where the nutrient is, but that could change very quickly. For example, this behavior of finding food, chemotaxis, well, that's clearly, in a sense, an intelligent behavior in the sense that it makes certain predictions that are going to come true or not within a second or so. But other animals, for example, let's say chipmunks, they make predictions about, well, the next season will be such that there will be winter, so I may have to actually store some food.

1:18:40.0 CA: But we as humans, by having developed writing systems, by having developed mathematics, we can make predictions in time over a far, far longer period that are accurate better than chance. The fact that we are talking about global warming and how that could affect us as a species is an obvious intelligent concept because we have understood that if we don't do anything, then we might be in big, big trouble. That is making predictions in a time scale far removed of our day-to-day activity. And our brain does that. But our brain, of course, has been doing this before we had figured out writing systems. But now the conjunction of our brain, where information during a lifetime is stored, and then we can make it permanent in books, that is the real secret sauce to the success of the species, as far as I'm concerned.

1:19:34.5 SC: Well, we'll see whether or not we have the collective strength of will to actually act on these predictions that we're making or not.

1:19:42.0 CA: Yeah, that's right. Of course, that's the big political question. Unfortunately, it is a political one and not a mathematical one. The mathematical one is very obvious.

1:19:52.3 SC: Good. That helps us a little bit. Chris Adami, thanks so much for being on the Mindscape Podcast.

1:19:55.6 CA: It was really my pleasure. Thank you, Sean.

15 thoughts on “266 | Christoph Adami on How Information Makes Sense of Biology”

  1. This is an excellent and thought provoking conversation. The key point for me is that information is always contextual. You have to have a subjective point of view to understand it. Information is useless if it cannot
    be understand or it is irrelevant in a particular context. Subjectivity is fundamental and cannot be overcome. There is no view from everywhere or nowhere. On the other hand, subjectivity is what creates meaning, goals, values, desires and motives. All too often people are looking for objective meaning, morals or values. Adami’s information theory tells us that those things don’t exist and we should stop worrying about them. Adami gives us a good model for the evolution of complexity and increasing fitness that we see in evolution by natural selection. But his ideas on information theory have even broader implications than that. Since information is always contextual, the next question is what is the context for human behavior.? The answer is that subjective self-interest filters and provides the context for all information we receive. Self interest is where we find meaning, value, motivation and the drive toward our most cherished goals.

  2. I found Adami’s definitions of information and entropy hard to follow in this short podcast. He claims both terms are misunderstood and that correcting these misunderstandings are essential. Here is one of several linked blog posts that define his views more precisely:
    https://adamilab.blogspot.com/2013/06/what-is-information-part-4-information.html
    In these posts, he says: Information is “what you don’t know minus what remains to be known given what you know”. And, information is “unconditional entropy minus conditional entropy”.

  3. Pingback: Sean Carroll's Mindscape Podcast: Christoph Adami on How Information Makes Sense of Biology - 3 Quarks Daily

  4. Half a million bits is the information contained in your body’s ordered state because why, that’s the information stored in your genome” (12:31). Speechless… Even the naive maths of assigning 2 bits to each of the 3B+ base pairs yields 6+ GB of information. It also ignores the chemical and spatial information contributed by bases themselves, the encoded amino acids, as well as that provided by other structures around the double helix, and the entire cellular information provided by egg (and to a lesser degree sperm) at the moment of conception. Come on guys.

  5. Fascinating.
    I’m reminded of some of the fiction books of mathematician Greg Egan, which in many ways espouses the idea that information=life=information… and takes it to the extreme in a rigorous way. Thank you.

  6. @Peter Vogt – he talks about that later in the podcast. Most of it is either repeated or not important to the expression of the means of reproducing itself (ie the human body), and therefore not “information” with respect to “life”, per his definition.

  7. Still wrt to this half a million bit of information: IIRC there’s a lot of guidance provided by the maternal womb during embryogenesis that you just can’t get in a Petri dish. You could argue that this womb is itself made of a very similar genome but I would note that it’s a thing that behaves in a way that has been selected by evolution and probably encodes some more information that what is just contained in the genome.
    I guess my point is that it would be impossible to retrieve the information to make a baby just in the genome without a context around it and therefore looking at the sole information content of the genome and stating that this is all it takes to make a human body misses a lot.

    Does that makes sense?

  8. Frustrating to hear the guest speak out about how so many, including Claud Shannon, are confused about information and entropy and then be totally stumped by the basic question regarding the informational content of the human body. He seems to be the one who is confused.

  9. The amount of information I can know should not change depending on the system I’m looking at. If I had to memorize the location of every particle in a box clearly it would be easier if the particles were in a low entropy state. In fact, I could then memorize the location of every particle in another low entropy box of particles. Clearly I can do this with more boxes if they are in a low state of entropy implying that the information in those boxes is also low.

    Maybe Claud Shannon wasn’t the one who was confused.

  10. Torbjörn Larsson

    Population genetics, which describe the evolutionary process, has very little use for physics information theory.

    When a loci is fixated among the population it requires about 10^-10 parts of energy in cell growth and division. It doesn’t make sense to express it in terms of information.

  11. Responding to DMF: information is not just for living systems. Any physical system can contain information. (In fact, information must be embodied in some sort of a physical system). For example, when we measure the temperature of gas in a box, we are gaining information about the system (specifically the average kinetic energy of its molecules). Or if we measure the quantum spin of an electron, we gain information by finding out that the electron was either spin up or spin down. Furthermore, we also have the information that the electron remains in that state (at least until it acts with something else in its environment).

  12. Responding to Peter Vogt and Mark Whybird: I also was troubled by Christoph’s assertion that the amount of information in a human body is only 1/2 Megabit. I have been taught to think of the maximum amount of information in a system as the total number of bits necessary it represent its “phase space”. If you view the human body through the lens of classical physics, the phase space would consist of the location and velocity of every atom in the body. This would required 6 numbers to be stored to describe each atom, plus some more info to describe the type of atom, the atomic weight (in order to account for isotopes), and probably even the energy levels the atom’s electrons are in. That is a massive amount of potential information. And the amount of information it would take to describe it at the quantum level would probably be even much higher.
    But then later in the podcast Christoph talks about information replication and how life is replicated information which builds its own replication machine. I’m wondering if when he talked about 1/2 Mbit he was thinking of it as COMPRESSED information–it comprises most of the information in the human genome to build a human. Since information which has undergone lossless compression is still the same amount of info, just represented in fewer bits, then the 1/2 Mbit of original info makes sense.
    Of course, the information in a human body is more than just the information in its genome. It is also influenced by environmental factors that are present when the body is being built. But maybe he is viewing this as a second order effect.

Comments are closed.

Scroll to Top