Speaker | Time | Text |
---|---|---|
Boom. | ||
And we're live. | ||
Hello, sir. | ||
How are you? | ||
I'm great, Harry. | ||
Nice to see you. | ||
Thanks. | ||
You were eating chocolate when you got here, and you told me that you are a cacao shaman. | ||
And I said those are strong words. | ||
They are. | ||
What does that mean? | ||
So I was in Berlin last year giving a talk at a tech conference, and somebody invited me to a sacred cacao ceremony. | ||
Never heard of it. | ||
I thought, wow, that sounds awesome. | ||
I love chocolate. | ||
I went. | ||
It was so wonderful. | ||
And at the end, they were talking about these people, these great cacao shamans. | ||
And I thought, what is that? | ||
I got to be one of those. | ||
And so I came back. | ||
I looked for certification. | ||
There wasn't certification. | ||
I self-declared. | ||
And then I started doing cacao ceremonies in New York. | ||
And I have hundreds of people who come. | ||
It's really wonderful and it's exciting. | ||
Like if you want to be a doctor, you got to go to medical school, right? | ||
You want to be a comedian. | ||
You got to become a professional. | ||
You got to put in your time. | ||
Cacao shaman, just show up. | ||
It's like putting up a shingle. | ||
Hey, I'm a cacao shaman. | ||
If anybody shows up and they have a good time, then you're real. | ||
Now, how much do you need to know about cacao? | ||
Like the nutritional properties of it? | ||
Cacao is amazing. | ||
It's great stuff. | ||
It's incredible. | ||
So, definitely cacao. | ||
People have been using it ceremonially for about 5,000 years. | ||
So, it's incredible. | ||
Chocolate makes people happy. | ||
It helps your brain function, your circulation. | ||
There's all these kinds of incredible things. | ||
But in the ceremonies that I do, I have two key messages. | ||
One is, you are the drug. | ||
I mean, we all take... | ||
People take drugs. | ||
People take... | ||
Thank you. | ||
kind of release and happiness and joy. | ||
We have those things inside of us and we just kind of get out of our way. | ||
We can experience them. | ||
The second thing is I believe that there are no, I say this in my ceremonies, there's no such thing as sacred cacao or sacred plants or sacred mountains or sacred people if we don't treat life with sacredness. | ||
But if we recognize that everything is sacred, then we infuse life with sacredness and meaning. | ||
Anyway, that's why I do it. | ||
It's a lot of fun. | ||
That's very interesting from a guy who specializes essentially in manipulating life. | ||
Well, you know, we have manipulated life as humans for a very, very long time. | ||
But it's interesting, you know, the idea of things being sacred, but your specialty is... | ||
Manipulating genetics, right? | ||
Yeah, well, so that is this strange moment that we're in, because for about 3.8 billion years, our species has evolved by this set of principles we call Darwinian evolution, random mutation and natural selection. | ||
It's brought us here. | ||
We used to be single-cell organisms, and now look at us. | ||
And there's been a lot of magic in that process, and there still is, but we humans are deciphering some of that magic. | ||
We are looking under the hood of what it means to be human, and we are increasingly going to have the ability to manipulate all of life, including our own. | ||
Yeah, that is very unnerving to a lot of people. | ||
It's uncomfortable and scary. | ||
Yeah, it is. | ||
They like things the way they are. | ||
Jamie, I'd like to stay the way I am. | ||
We always think that. | ||
Why do we always think that? | ||
Because there's a built-in conservatism in our brains. | ||
And yet, we live these lives that are entirely dependent on these radical changes that are Ancestors have created. | ||
I mean, we didn't find this building or agriculture or medicine in nature. | ||
We built all those things. | ||
Then everybody gets a new baseline when you're born. | ||
And you think, well, I want organic corn. | ||
I want whatever. | ||
But all these things are creation. | ||
We live in an entirely created world, and our ability to manipulate and change that world is always growing. | ||
And I think we need to recognize that, but being afraid is okay, and being excited is okay, and we need to find the right balance between those two emotions. | ||
I think for a lot of people, they feel like so many changes happened, particularly when you're talking about genetically modified foods. | ||
So many things happened before they realized they had happened. | ||
So when they're like, hey man, I don't want to eat any GMO fruit. | ||
Well, then you probably shouldn't eat any fruit. | ||
Because everything that you buy has been changed. | ||
Every orange that you buy, that's not what an orange used to be like. | ||
You could buy an apple. | ||
Apples didn't used to be like that. | ||
Tomatoes didn't used to be like that. | ||
No, I know. | ||
We reset our baseline just from when we were kids. | ||
So if you went back 12,000 years ago to the end of the last Ice Age and you said, all right, find me all these things that we buy at Whole Foods. | ||
Most of them didn't exist. | ||
We've created them. | ||
Sure. | ||
And then in the 1970s, we had the ability to do what's recombinant DNA, what people call genetic modification. | ||
And people are afraid because it's, well, that feels unnatural. | ||
We're applying science to food. | ||
And, you know, that's the issue. | ||
And now we're entering the era of genetically modified humans, and there's that same level of uncomfortableness. | ||
But what happened, the reason why I've written this book, Hacking Darwin, is that if we approach genetically modified humans in the same way we approach genetically modified foods, which is the scientists say, hey, we've got this, we're going to manage them responsibly, and it just kind of happens to people. | ||
People are going to go nuts. | ||
I mean, imagine how agitated people are about GMO foods. | ||
If they don't have a say in how the human experience of genetic modification plays out, people are going to go berserk. | ||
So we have this window of time where we can start bringing everybody into an inclusive conversation about where we're going because where we are going is just radically different from where we've been. | ||
Yeah, I think it's an awareness issue and I also think it's a perception issue. | ||
I think that everything people do is natural. | ||
Including cities. | ||
I think cities are natural. | ||
That's why they're all over the world. | ||
I think they're as natural as beehives. | ||
And I think as much as we like to think that technology is not natural, it's clearly something people naturally make. | ||
unidentified
|
Of course. | |
They make it in every culture, if they can. | ||
It's the history of our species. | ||
And we kind of misuse this word naturally. | ||
unidentified
|
Natural. | |
Because what is natural? | ||
I mean, maybe natural was when we used to live and we were just part of nature. | ||
I always say, it's like, people say, oh, I love nature. | ||
I love going out and hiking in the woods. | ||
The reason you love hiking in the woods is that we've murdered all the predators. | ||
It's like in the old days, you stay in your cave, you're not going out and hiking in the woods. | ||
There's stuff that's going to kill you out there. | ||
unidentified
|
I know. | |
That was a massive luxury to go wander through the forest with no weapons. | ||
unidentified
|
Yeah. | |
Nobody did that. | ||
unidentified
|
Exactly. | |
No, exactly. | ||
But people say, oh, I want nature. | ||
I want my natural corn. | ||
I want my natural chihuahua, even though 25,000 years ago, there's no chihuahuas. | ||
There's wolves. | ||
And look what we've done to them. | ||
unidentified
|
I know. | |
Well, look what we have done to them. | ||
Made them pugs. | ||
If you had natural wheat, or natural corn in particular, natural corn used to be a tiny little thing. | ||
Yeah, it's just a few weeds. | ||
Yeah, weird, gross little... | ||
Yeah. | ||
Grain. | ||
Yep. | ||
And then now we made it this big, juicy, sweet, delicious thing that you put butter on. | ||
Which is great. | ||
That's great. | ||
As long as it doesn't have glyphosate on it. | ||
We can't fetishize that there's some kind of imaginary world where kind of everybody was wearing Birkenstocks and eating in Whole Foods. | ||
That imaginary world sucked for us. | ||
That's why we left it. | ||
It's true. | ||
But there's some sort of a balance, right? | ||
Yeah. | ||
We do appreciate the nature aspect of our world and eagles and salmon and all these wild things and to be able to see them is very cool. | ||
unidentified
|
Yeah. | |
But yeah, you don't want to get eaten by those things. | ||
You don't want them everywhere. | ||
You want to be able to go out and get your newspaper without being worried about getting attacked by a jaguar. | ||
It helps. | ||
Yeah. | ||
When you think about the future, at least me, let me tell you my concern. | ||
unidentified
|
Yeah. | |
I'm worried that rich people are going to get a hold of this technology quick and they're going to have massive unfair advantages in terms of intellect, in terms of physical athletic ability. | ||
I mean, we really can have a grossly imbalanced world radically quickly if this happens fast where we don't understand exactly what the consequences of these actions are until it's too late and then we try to play catch up with rules and regulations and laws. | ||
Yeah, that's a very, very real danger, and that's why I've written this book. | ||
That's why I'm out speaking every day about this topic, because we need to recognize that if we approach these revolutionary technologies using the same values that we experience today, where we're here and very comfortable, but just down the road, there are people who are living without many opportunities. | ||
There are people in parts of the world, like Central African Republic, where there's just a war zone, kids are born malnourished. | ||
If those are our values today, we can expect that when these future technologies arrive, we'll use those same values. | ||
So it's real, and right now we have an opportunity to say, all right, these technologies are coming. | ||
Whatever we do, these technologies are coming. | ||
There's a better possible future and a worse possible future. | ||
And how can we infuse our best values into the process to optimize the good stuff and minimize the bad stuff? | ||
And certainly what you're saying is a real risk. | ||
Think of what happened when European countries had slightly better weapons and slightly better ships than everybody else. | ||
They took over the world and dominated everybody. | ||
And so, yeah, it's very real. | ||
Governments need to play a role in ensuring broad access and regulating these technologies to make sure we don't get to that kind of dystopian scenario that you've laid out. | ||
Well, it's also in terms of governments regulating things. | ||
Like, why are they qualified? | ||
Who are they? | ||
Who are the governments? | ||
They're just people, right? | ||
They're people that are either elected or some sort of a monarchy. | ||
You're dealing with either kings and queens and sheiks, or you're dealing with presidents. | ||
And we've seen in this country that sometimes our presidents don't know what the fuck they're talking about, right? | ||
So who are they to disrupt science, to disrupt this natural flow of technology? | ||
Well, we need somebody to do it. | ||
We need some representation of our collective will just to avoid some of the things like you just mentioned. | ||
That's the reason why humans banded together and created governments. | ||
And the reason for democracy, especially if you have more functioning democracies, is that your government in some ways reflects the will of the people. | ||
And the government does things... | ||
That individuals can't do. | ||
And I know there are libertarian arguments where everyone should just, like, if you want a little road in front of your house, either go build the road or pay somebody. | ||
But there are a lot of things, even in that model, that won't get done. | ||
There are a lot of kind of big national, even global concerns that you need some kind of I think we're good to go. | ||
Every person needs to really understand these revolutionary technologies like genetics, like AI, and all of our responsibilities and opportunities to say, hey, this is really important. | ||
Here are the values that I think that I cherish. | ||
And just like you said, I don't want it to be That the wealthiest people are the ones who have kids with higher IQs and live longer and healthier than everybody else. | ||
So we have to raise our voice. | ||
And there needs to be a bottom-up process and a top-down process, and it's really hard. | ||
Many people have a concern that someone else is going to do what we're not willing to do first. | ||
Yes. | ||
Right? | ||
They're worried. | ||
China and Russia. | ||
Those are the big ones. | ||
China and Russia. | ||
Especially China. | ||
Yeah. | ||
They're very technologically advanced. | ||
And their innovation's off the chain. | ||
When it comes to... | ||
Like Huawei just recently surpassed Apple as the number two cell phone manufacturer in the world. | ||
Five years ago, they had a single-digit share of the marketplace. | ||
unidentified
|
Yeah. | |
Now they're number two on the planet Earth. | ||
I mean, they hustle. | ||
And if they just decide to make super people, and they do it before we do, that's what people are worried about, right? | ||
They're worried about – there's trivial things, seemingly trivial, like athletics, and then there's things that are really – like, what's to stop people from just becoming the Hulk? | ||
What's to stop people from becoming immortal? | ||
unidentified
|
What's to stop – what is – Well, two questions. | |
First is China. | ||
I think it's a really big issue. | ||
The story of the 21st century, one of the biggest stories of the 21st century will be how the US-China rivalry plays out and the playing field will be with these revolutionary technologies. | ||
And China has a national plan to lead the world in these technologies by 2050. They're putting huge resources. | ||
They have really smart people they are really focused on. | ||
And it's a big deal. | ||
In genetic technologies, when last year, my book Hacking Dharma was already in production in November when it was announced that these first genetically engineered babies had been born in China. | ||
And so I called the publisher and said, we need to pull this back out of production because I need to reference this. | ||
But it didn't require much of a change. | ||
I'd already written... | ||
This is happening. | ||
We're going to see the world's first gene-edited humans. | ||
It's going to happen first in China. | ||
And here's why. | ||
So I had to add a few sentences saying, and it just happened in October of 2018. So China is on that path. | ||
And we need to recognize that on one hand, the United States needs to be competitive. | ||
On the other hand, we don't want a runaway arms race of the human race, and that's why we need to find this balance between national ambition and some kind of global rules. | ||
That's really hard to do. | ||
Yeah, and the other thing is that we're competing with them, and so if they decide to do it first, we're almost compelled to do it second, or compelled to try to keep up. | ||
How far away do you think we are from physically manipulating living human beings, Versus fetuses, versus something in the womb. | ||
So physically manipulating living human beings, we're there. | ||
- Yeah, yeah, so that's often, it's called gene therapy. | ||
So for example, there's a whole class of treatments for treating cancer called CAR T therapy. | ||
So you have a cancer. | ||
When you're younger, your body is better able to fight cancers. | ||
What you can do with someone with a cancer, you take their cells, you manipulate their cells to give them cancer-fighting superpowers, and you put them back into the person's body. | ||
And now the person's body behaves like you're a younger person. | ||
You have the ability to fight back. | ||
So gene therapies are already happening. | ||
A relatively small number of them have already been approved, but there is a list of thousands of them with regulators and applications to regulators around the world. | ||
So the era of making genetic changes to living humans, that's already here. | ||
What can they do with it so far? | ||
So far, most of it is focused on treating diseases. | ||
But a lot more is coming. | ||
Because when people think about the human genome, our genome isn't a disease genome. | ||
It's not a healthcare genome. | ||
It's a human genome. | ||
And so we are going to be able to do things that feel like crazy things, like changing people's eye color, changing people's Skin color to funky things. | ||
I mean, there's a lot of stuff that we're not doing now that we will be able to do. | ||
How far away do you think we are from something like that? | ||
10 years? | ||
So in 10 years, we're going to have green people. | ||
If someone so chooses. | ||
Yeah, if someone so chooses. | ||
unidentified
|
What if it sucks? | |
Will they be able to go back to normal color? | ||
Well, if it's... | ||
That's a good question. | ||
If it's with this kind of gene therapy, and it's a small number of genes... | ||
Probably, but we are messing with very complex systems that we don't fully understand. | ||
So that's why there's a lot of unknowns. | ||
And coming back to your point on regulation, that's why I don't think we want a total free-for-all where people say, hey, I'm going to edit my own jeans. | ||
Yeah, and you don't want some backyard hustler. | ||
Yeah, it's true. | ||
Lab. | ||
It's true. | ||
It comes by saying about the Hulk. | ||
I mean, I just think that there are all kinds of... | ||
We're humans. | ||
We're diverse. | ||
Any kind of thing that you can think of, there is a range. | ||
And there's crazy on the left and crazy on the right and crazy on the top. | ||
So people are going to want to do things. | ||
And the question is, for any society, what do we think is okay and what do we think is not okay? | ||
And maybe there should be some... | ||
I believe there should be some limit to how far people can go with experimenting with... | ||
Possibly, likely on themselves, but certainly on their future children. | ||
Certainly on their future children, yeah. | ||
But once you're 18, I think, do whatever the fuck you want. | ||
Well, maybe 25. 25. We're going to have a lot of 25-year-olds with gills. | ||
We probably will. | ||
It's like the tattoo. | ||
Seemed like a good idea. | ||
Yeah, well we probably will. | ||
So you think we're probably like 50 years away from that being a reality? | ||
So I think that the genetic revolution has already begun. | ||
And it's going to fundamentally change our lives in three big areas. | ||
The first is our healthcare. | ||
So we're moving from a system of generalized healthcare based on population averages. | ||
So when you go to your doctor, you're treated. | ||
Because you're a human, just based on average. | ||
And we're moving to a world of personalized medicine, and the foundation of your personalized healthcare will be your sequence genome and your electronic health records. | ||
That's how they know you are you. | ||
And that's how they can say, this is a drug, this is an intervention that'll work for you. | ||
When we do that, then we're going to have to sequence everybody. | ||
So we're going to have about 2 billion people have had their whole genome sequenced within a decade. | ||
And then we're going to be able to compare What the genes say to how those genes are expressed. | ||
And then humans become a big data set. | ||
And that's going to move us from precision to predictive healthcare where you're going to be just born and you're going to have all this information. | ||
Your parents have all this information about how certain really important aspects of your life are going to play out. | ||
And some of that is going to be disease-related. | ||
But some of that's just going to be life-related, like you have a better-than-average chance of being really great at math or having a high IQ or low IQ or being a great sprinter. | ||
And how do we think about that? | ||
And then, again, a revolution that's already happening. | ||
We're just going to change the way we make babies. | ||
We're going to get away from sex as the primary mechanism for conceiving our kids. | ||
We'll still have sex for all the great reasons we do. | ||
and that's going to open up a whole new world of just of how of applying science to what it means to be a human with a lot of new possibilities that's what's going to be so freaky when people stop having sex to make kids and they make kids in a lab every kid's made in the lab well not only that i think we're going to move to an era an era where people who have who make babies through sex will be seen as taking a risk kind of like people who don't vaccinate their kids where it's natural to not | ||
it's more natural to not vaccinate your kids than to do it but people say wait a second you're taking on a risk on behalf of your kids about three percent of all kids in the world are born with some kind of harmful genetic abnormality Using in vitro fertilization and embryo screening, that 3% can be brought down significantly. | ||
And what happens if you see somebody 20 years from now who has a kid with one of those preventable diseases? | ||
Do you think that's fate? | ||
Or do you think, well, wait a second, those parents, they made an ideological decision about how they wanted to conceive their kids. | ||
So I think we're moving towards some really deep and fundamental changes. | ||
Hmm. | ||
Well, yeah. | ||
That's an interesting conversation of whether or not—I wonder if we're ever going to get to a point where people don't allow people, sort of like people don't allow people to not get vaccinated. | ||
unidentified
|
Right. | |
There's a lot of that going on today. | ||
unidentified
|
Right, right. | |
Which is great, right? | ||
You don't want diseases floating around. | ||
But what if that gets to the place where we do that with people? | ||
With people creating new life forms? | ||
What if you say, hey, you are being irresponsible, you're just having sex and having a kid. | ||
I know that's how your grandma did it. | ||
We don't do it that way in 2099. Yeah. | ||
I think it's going to be hard to do that in a society like the United States, but in a country like North Korea? | ||
They'll be able to do that. | ||
Or if they said, look, you can make babies however you want, but if you make babies the old-fashioned way, and if your kid has some kind of genetic disorder that was preventable, we're just not going to cover it with insurance. | ||
So you're going to have a $10 million lifetime bill. | ||
You don't need to require something. | ||
You can create an environment where people's behaviors will change. | ||
And then there will be... | ||
I mean, right now, somebody sees some little kid riding around their bicycle without a helmet. | ||
They're kind of looking at the parents like, hey, what are you doing? | ||
How come you don't have a helmet on your kids? | ||
And I just think that we're moving toward this kind of societal change where people will, I believe, see conceiving their kids in the lab as a safer alternative. | ||
And it's not just safety because once you do that, then that opens you up To the possibility of all other kinds of applications of technology, not just to eliminate risks or prevent disease, but you have a lot more information. | ||
So already it's possible to roughly rank order 15 pre-implanted embryos, tallest to shortest, in a decade from highest genetic component of IQ to lowest genetic component of IQ. This stuff is very real and it's very personal. | ||
What do you think will be the first thing that people start manipulating? | ||
I think certainly health. | ||
Health will be the primary driver because that's every parent's biggest fear. | ||
And that is what is going to be kind of the entry application. | ||
People wanting to make sure that their kids don't suffer from terrible genetic diseases. | ||
And then I think the second will probably be longevity. | ||
I mean, right now there's a lot of work going on sequencing people, the super-agers, people who live to their late 90s, people to 100, to identify what are the genetic patterns that these people have. | ||
So it's like to live to 90, you have to do all the things that you advocate, healthy living and whatever. | ||
But to live to 100, you really need the genetics to make that possible. | ||
So we're going to identify what are some of the genetic patterns that allow you to live Those kinds of long lives. | ||
But then after that, then it's wide open. | ||
I mean, it's higher genetic component of IQ, outgoing personality, faster sprinter. | ||
I mean, we are humans. | ||
We are primarily genetic beings, and we are going to be able to look under the hood of what it means to be human, and we'll have these incredible choices. | ||
And it's a huge responsibility. | ||
How long do you think before you have a person with four arms? | ||
I think it's going to take a long time. | ||
A couple hundred years? | ||
Well, the thing is, here's how I see it. | ||
So the real driver, there's two primary drivers. | ||
One will be embryo selection. | ||
So right now, average woman going through IVF has about 15 eggs extracted. | ||
And then in IVF, in vitro fertilization, those eggs are fertilized using the male sperm. | ||
And in average male ejaculation, there's about a billion sperm cells. | ||
So men are just giving it away. | ||
Women, human, female mammals are a little bit stingy. | ||
But then the next killer application is using a process called induced pluripotent stem cells. | ||
And so Shinya Yamanaka, this great Japanese scientist, won the 2012 Nobel Prize for developing a way to turn any adult cell into a stem cell. | ||
So a stem cell is a kind of cell, it can be anything. | ||
And so you take, let's say, a skin graft. | ||
That has millions of cells. | ||
You induce those adult skin cells into stem cells. | ||
So you use these four things called Yamanaka factors. | ||
And so now you have, let's call it 100,000 stem cells. | ||
And then you can induce those cells into egg precursor cells, and then eggs. | ||
So all of a sudden, humans are creating eggs like salmon on this huge scale. | ||
So you have 100,000 eggs, fertilize them with the male sperm, In a machine, an automated process, you grow them for about five days. | ||
And then you sequence cells extracted from each one of those. | ||
And the cost of genome sequencing in 2003 was a billion dollars. | ||
Now it's $800. | ||
It's going to be next to nothing within a decade. | ||
And then you have real options, because then you get this whole... | ||
A spreadsheet, an algorithm. | ||
And then you go to the parents and say, well, what are your priorities? | ||
And maybe they'll say, well, I want health, I want longevity, I want high IQ. When you're choosing from big numbers like that, you have some real options. | ||
And then on top of that, then there is this precision gene editing, the stuff that happened in China last year. | ||
I think it will be... | ||
And coming back to your question about four arms, I think it's going to be very... | ||
People have this idea that tools like CRISPR are going to be used. | ||
Someone's going to sit at a computer and say like four arms and three heads and wings and whatever. | ||
But it's pretty hard because human biology is incredibly complicated. | ||
And we always know more. | ||
But we're at the very beginning of understanding the full complexity of human biology enough to make these big kind of changes. | ||
But if you're choosing from 100,000 fertilized eggs, those are all your natural kids. | ||
Yeah. | ||
And then you would get the best of that and then work on those. | ||
Exactly. | ||
That's exactly the model. | ||
You get that, and then you say, alright. | ||
What if you had like 20 sons that were awesome and they didn't tell you about 18 of them? | ||
You kept two of them, then 18 of them shipped off to some military industrial complex, turned them into assassins. | ||
Any kind of crazy thing you can think of? | ||
That's the problem, right? | ||
All this stuff will be possible. | ||
And a lot of technologies, you can imagine all kinds of crazy stuff. | ||
And that's coming back to your earlier point about regulations. | ||
We want to live in regulated environments. | ||
So like now, think of the internet. | ||
In the beginning days of the internet, people thought, oh, just let the internet be. | ||
Just let it play out. | ||
It's going to liberate all of us. | ||
And now China is showing how the internet can actually be really actively used to suppress people and Facebook is taking people's information and Google in a way that's frightening a lot of people. | ||
It shouldn't be that these companies can do whatever they want. | ||
We have to have some way of establishing limits because not every individual is able to entirely protect themselves. | ||
They don't have the power. | ||
They don't have all the information. | ||
We need some representatives helping us with that. | ||
The real concern is the competition, right? | ||
The real concern is whether or not we do something in regards to regulation that somehow or another stifles competition on our end and doesn't allow us to compete with Russia and China, particularly China. | ||
Yeah, that's exactly right. | ||
And so what we need to do is to find that balance. | ||
And one of the big issues for this is privacy. | ||
So if you kind of look around the world, I'd say of the kind of the big countries and groupings of countries, there's three models of privacy. | ||
There's Europe, which has the strongest privacy protections for all kinds of data, including genetic data. | ||
There's China, that has the weakest. | ||
And there's the United States that has the middle. | ||
And the paradox is from an individual perspective, we all are thinking, well, we kind of want to be like Europe because I don't want somebody accessing my personal information, especially my genetic information. | ||
This is like my most intimate information. | ||
But... | ||
Genetics is the ultimate big data problem. | ||
And so you need these big data pools and you'd access to these big data pools in order to unlock the secrets of genetics. | ||
So these three different groupings, everyone's making a huge bet on the future and the way we're going to know who wins. | ||
Like right now in the IT world, we have Amazon and Apple and Google and those big companies. | ||
But whoever gets this bet right, They will be the ones who will be leading the way and making a huge amount of money on these technologies. | ||
What we're talking about is a trillion, multi-trillion dollar industry. | ||
How do you think this is going to affect things like competitive athletics? | ||
Hugely. | ||
So right now, we have this problem. | ||
Someone like Lance Armstrong, who is manipulating his body. | ||
And what he's basically doing is adding more red blood cells so that he can carry more oxygen. | ||
And people feel that that's cheating. | ||
It's a different topic that probably everybody in the Tour de France was doing exactly that when he won. | ||
But what if, which will be the case, we're going to be able to sequence the people, let's say nobody's doing drugs, and we sequence all these athletes. | ||
Some of them will just have a natural genetic advantage. | ||
Their bodies will naturally be doing what Lance Armstrong had manipulated his body to do. | ||
You know that's happening with a sprinter right now? | ||
Yeah. | ||
That female sprinter that has high levels of testosterone? | ||
Yeah. | ||
And I feel really sorry for her, but we have categories. | ||
I mean, with your world in mixed martial arts, I mean, I think I remember in the past there was some person who was kind of a borderline on a... | ||
Between genders and we're just kicking the shit out of all of these women in cage fighting. | ||
It's like we have these categories of man and woman. | ||
We know that the gender identities are fluid, but how do we think about it when these genetic differences confer advantages? | ||
So if your body is primed To do something. | ||
Maybe you could have like a Plato's Republic world where everybody fulfills a function that you are genetically optimized to do. | ||
And then you could imagine that being a very competitive kind of environment. | ||
But what do you do for now in something like the Olympics? | ||
If somebody has this huge genetic advantage, should we let somebody else manipulate their bodies? | ||
There's this thing called gene doping. | ||
In order to change the expression of genes, so your body to act like you're as naturally genetic enhanced as somebody else, it's complicated. | ||
Are they capable of doing certain physical enhancements through gene doping right now? | ||
unidentified
|
Yeah. | |
Like, what can they do right now? | ||
No, no, so the way it works is so your genes instruct your cells to make proteins. | ||
That's how the whole system works. | ||
So you can change genes or you can trigger the expression of proteins. | ||
So you can get people's bodies to behave As if they had these periodic optimizations. | ||
Yeah. | ||
And so that's why now the World Anti-Doping Agency, I mean, they are now starting to look at gene doping. | ||
And this is the first time that that's even being considered as a category. | ||
And then there are… Are there people that have done that successfully? | ||
You know, I don't know the answer to that. | ||
I know that WADA is looking for it, which makes me assume that it must have done, but I haven't seen it. | ||
unidentified
|
I've looked for it. | |
I haven't seen any reports. | ||
If China starts winning everything. | ||
Well, China is. | ||
So I wrote one of my sci-fi novels, Genesis Code, was about this. | ||
So China, as you know, has their system of their Olympic sports schools. | ||
And the way it works is they test kids all around the country. | ||
So let's just say it's diving. | ||
And they identify what are the core skills of a diver? | ||
What do you need? | ||
And then they go around the country and they test kids, and then they bring a bunch of them to their Olympic sports schools. | ||
And then they get them all involved, and then some kids are the best of those kids, and then the best of those kids, and then you get with these champs. | ||
That's why China advanced so rapidly. | ||
But what happens if they're doing that, but it's at the genetic level? | ||
And there are countries like Kazakhstan that are already announcing that they are going to be screening all of their athletes. | ||
The science isn't there yet, so it's impossible right now to say, well, I'm going to do a genome sequence of somebody, and I know this person has the potential to be an Olympic sprinter. | ||
But 10 years from now, that's not going to be the case. | ||
Wow. | ||
Yeah, it's sort of going to throw a monkey wrench in the whole idea of what is fair when it comes to athletics. | ||
Yeah, what is fair? | ||
What is human? | ||
Right. | ||
What is human? | ||
Yeah. | ||
I mean, look, it's not like people don't already alter their bodies by training, by diet, exercise, all sorts of different recovery modalities, cryotherapy, sauna, high elevation training, all these different things that they do that manipulates the human body. | ||
But, it's not like, it would be kind of crazy if you had sports, but you couldn't practice and you couldn't work out. | ||
Like, we want to find out what a person's really like. | ||
No practice, no working out. | ||
And that's the thing, is like, we are moving, it comes back to what we were saying before about nature. | ||
It's like we have this feeling, nature somehow feels comfortable to us. | ||
That's what we're used to. | ||
All this stuff that you're talking about, nobody was doing that 10,000 years ago. | ||
It's like, hey, I'm running after a buffalo. | ||
Yeah. | ||
And so as these boundaries change, as the realm of possibility changes, then we're going to be faced with all of these questions. | ||
Even now, look at a sport like competitive weightlifting. | ||
They have the real competitive bodybuilding. | ||
And you see these guys, and they're monsters. | ||
And then they have these drug-free guys, and everybody looks like a yogi. | ||
They still look pretty big. | ||
They look pretty big, but not compared to these other guys. | ||
The only way to get those freak levels is through steroids. | ||
Yeah. | ||
And so, like, how are we going to police this? | ||
And I think it's going to be very difficult. | ||
And so, maybe we can have some kind of natural area of life. | ||
But I think that our model of what's normal... | ||
Is just going to change. | ||
Because like I was saying in the beginning, we set our baseline based on how we grew up. | ||
And that, it seems about right. | ||
Like it seems about right to us that everybody gets immunizations. | ||
But immunizations are a form of superpower. | ||
Imagine if our ancestor, they couldn't even imagine immunizations. | ||
What an unfair advantage when you have 100 million people dying of Spanish flu. | ||
So all this stuff is scary and it's going to normalize, but how it normalizes That's what it played in. | ||
Well, the world has changed so much just in the last 20 years, but it feels like this is just scratching the surface in comparison to what's coming. | ||
People misunderstand, and they underestimate the rate of change. | ||
And the reason that they do that is since the beginning of the digital revolution, we have experienced a thing called exponential change. | ||
You've heard of Moore's Law, which is basically computing power roughly doubles every two years. | ||
And we've internalized Moore's Law, and that means that every new iPhone we expect to be better and stronger and faster and all these kinds of things. | ||
But now we're entering a world where we're going to have exponential change across technology platforms. | ||
And so we think about, well, what does exponential change mean in the context of biology? | ||
Well, at the very, very beginning, it's genome sequencing is going to be basically free. | ||
But we're going to be able to change life. | ||
And because we're on this J-curve, Like when you think of what's a 10-year unit of change looking in the rearview mirror, that amount of change is only going to take five years going forward and then two years and then one year. | ||
And so that's the reason why I've written this book is we have to get that this stuff is coming fast. | ||
And if we want to be part of it, we have to understand it and we have to make our voices heard. | ||
unidentified
|
What makes you nervous about this? | |
All right, three big areas. | ||
First, humans and all of us are incredibly complex. | ||
I mean, we talk about genetic code, which is mind-bogglingly complex. | ||
But our genetics exists within the incredible, incredibly complex systems biology. | ||
We have all these things like our microbiome, our virome, our proteome, our metabolome. | ||
And then that exists within the context of our environment and everything's always changing and interacting. | ||
And so we are messing and we have the tools to mess and we will mess because we're this hubristic species with these really complex ecosystems, including ourselves, we don't fully understand. | ||
That's number one. | ||
Two, and you mentioned it before, this issue of equity. | ||
What happens if we have – every technology has to have first adopters. | ||
If you don't have it, you never get the technology. | ||
But what happens if a group of people move much more quickly than other people? | ||
Whether it's real or not, even if they believe it's real, you could imagine big, dangerous societal changes. | ||
And the third big area is diversity. | ||
When we think about diversity, we think, well, it's great to have diverse workplaces and schools and we're more better people for it and we're more competitive. | ||
But diversity is something much, much, much deeper. | ||
In Darwinian terms, diversity is random mutation. | ||
Like, that's our core survival strategy as a species. | ||
If we didn't have that, you could say we'd still be single-cell organisms, but we wouldn't. | ||
We would have died because the environment would have changed and we wouldn't have had the built-in resilience to adapt. | ||
Yeah, that is really important when you think about diversity, right? | ||
That we need a non-uniformity when it comes to our own biology. | ||
We need a bunch of different kinds of people. | ||
We have to have it. | ||
Even if we optimize for this world, the world will change. | ||
There's no good and bad in evolution. | ||
There's just well-suited for a particular environment. | ||
If that environment changes, The best suited person for your old environment may be the least suited person for the new environment. | ||
So even if we have things that seem like really great ideas now, like optimizing health. | ||
So if you have sickle cell disease, you're probably going to die and you're going to die young and it's going to be excruciatingly painful. | ||
And so you would say, well, let's just get rid of sickle cell disease, which we can do. | ||
But if you are a recessive carrier of the single-cell disease gene, you don't have it, you're just carrying it, and you have a pretty significant risk of passing it on to your kids, but you also have an additional resistance to malaria. | ||
And so we are almost certainly carrying around all kinds of recessive traits, maybe even ones that we don't like that are harming us now, but that could be some protection against some future danger that we don't yet understand or haven't faced. | ||
And so the challenge is that diversity has just happened to us for 4 billion years. | ||
Now we're going to have to choose it, and that's a big challenge for us. | ||
So essentially, we're going to have, without doubt, some unintended consequences, some unintended domino effect, things that are going to take place that we really can't predict. | ||
We just have to kind of go along with this technology and see where it leads us as it improves. | ||
If you go back and look at surgeries from the 1950s, comparison to surgery of 2019... | ||
I mean, I would never advise someone to get their knee operated by a 1950s physician. | ||
But that's kind of, someone's going to have to be an early adopter when it comes to these genetic therapies. | ||
Yeah, no, so I agree with you, but where I would slightly, where I would add to what you're saying is, these technologies, they're going to happen, they're going to play out. | ||
unidentified
|
Mm-hmm. | |
What's at play now is not whether these technologies are going to advance. | ||
They will advance in a way that is going to just blow people's minds. | ||
What's at play is what are the values that we are going to weave into the decision-making process so that we can get a better outcome than we otherwise would have had. | ||
And that's what, in my view, that's the real important issue now. | ||
Yeah. | ||
Unintended consequences are something that I've been taking very seriously lately when I'm paying attention to technology as it's used in social media. | ||
Particularly, one of the things that's disturbed me quite a bit over the last few weeks is that there's a model that they use, and not intentionally, But there's a model that they use to get people upset about things, show you things in your feed that you argue against because that makes you click on them more and you engage in them more. | ||
and because of the fact that we have this advertiser-based model where people are trying to get clicks because they want to get ads on their page and the more clicks they get, the more money they get from those ads and so they want to incentivize people to go there and the best way through the algorithm, what they found, without doing it on purpose is to get people the more money they get from those ads and so they Yeah, and they're pushing people into these little information ghettos that are really Right. | ||
And we've gotten to this point where this is just an accepted part of our lives. | ||
Did you go to check your Google feed or your Facebook feed and go, what the fuck are they doing? | ||
Is this real? | ||
They're going to pass this? | ||
God damn it. | ||
And then you get mad and then you engage with people online and then it results in more revenue. | ||
Right. | ||
Getting them to stop that. | ||
If you had to go to Facebook and say, hey, Mark Zuckerberg, I know you have fucking $100 billion or whatever you got, but you can't make any more money this way. | ||
Because what you're doing is fucking up society. | ||
Because you're encouraging dissent. | ||
You're encouraging people to be upset and arguments. | ||
And you're doing it at great financial reward but great societal cost. | ||
So stop. | ||
He's not going to do it, right? | ||
Well, he may not do it. | ||
That comes back to the point about regulation. | ||
The question is, how big is your stick? | ||
Well, one of the guys who was the founder of Facebook is now coming out and saying that Facebook needs to be broken up. | ||
And then he was one of the original founders. | ||
And he's like, it has gotten so far out of hand. | ||
It's so far away from where it is. | ||
It's literally affecting global politics. | ||
Yeah. | ||
Well, it is. | ||
And so one option is to break it up. | ||
It seems to have worked pretty well with AT&T. Another option is to regulate it, which in my mind would be a better approach. | ||
And that is to say, here's what's okay and here's what's not okay. | ||
And this stuff is really intricate. | ||
You have to really get down beneath these algorithms, which are unbelievably complex, but you're exactly right. | ||
I mean, what we're seeing now Right. | ||
Right. | ||
Right. | ||
where people come and they work it out and they say, you know, I'm not perfectly happy with this outcome, but here's a compromise. | ||
And if we can't compromise, then our civic culture is gonna break down. | ||
And there's so much, I mean, people don't see these pillars that are holding up our society. | ||
I lived in Cambodia for two years. | ||
And if you don't have these civic pillars under your society, societies look very, very different. | ||
Everyone's life experiences, we kind of take for granted. | ||
You can go out the door, walk to Starbucks and not get shot, or you can have your house, something happens, your house gets robbed, you call the police, and the police aren't the ones who've robbed your house. | ||
I mean, there's all these kinds of crazy things. | ||
If we break down the foundations that underpin our lives, that's really dangerous. | ||
unidentified
|
Right. | |
What I was kind of getting at was that through this process of this algorithm, how this algorithm selects things that shows you in your feed and how people are getting upset by this and how this is generating massive amounts of revenue, once it's already happened, it's very difficult to stop. | ||
And my concern would be that this would be a similar thing when it comes to genetic engineering. | ||
We're saying we need to be able to put regulations on this, we need to be able to establish... | ||
But once it gets out of the bag, once it gets rolling, and I have... | ||
Do you remember when Mark Zuckerberg sat in front of all those politicians? | ||
They had no fucking idea what they were talking about. | ||
You make money, yeah. | ||
Piss poor prepares. | ||
It shows you, like, these are the people that are looking out for us, good fucking luck. | ||
These are Luddites. | ||
They're dumbasses. | ||
They're fools, right? | ||
unidentified
|
And they don't know anything about- Some are, there's better and worse, but yeah. | |
But almost everyone was underwhelming and under-impressive. | ||
In that hearing. | ||
In that hearing. | ||
The fact that they're dealing with one of the most important moments of our time, but they didn't bring on some sort of a legitimate technology expert who could explain the pitfalls of this and do so in a way that the rest of the world's going to know. | ||
So they're not going to protect us from genetic engineering either, right? | ||
I totally agree. | ||
Because they're generalists in terms of their education for the most part, and they're not concerned. | ||
They're concerned with raising money for their campaign. | ||
They're concerned with getting re-elected. | ||
That's what they're concerned with. | ||
Yeah, I totally agree with you that if we wait to focus on this issue until it becomes a crisis, it's going to be too late because all the big decisions will have been made. | ||
The reason why I wrote this book, the reason why I'm on my almost week three of this book tour doing events like this every day is what I am saying in every form that I can is this is really important. | ||
We were watching the news yesterday, they had this royal baby in the UK. I don't give a shit. | ||
It doesn't affect my life. | ||
But what is at play now is the future of our entire species and our democracy and our lives. | ||
And we have to be focusing on those things because we have a moment now where we can, to a certain extent, influence how these revolutions play out. | ||
And if we just wait around, if we're distracted and we're focusing on all this stuff that's sucking up our attention, and whether it's Trump or Brexit or Mueller and all these things, I mean, how much of our time are we spending focused on? | ||
It's fine, let's pay a little bit of attention, but there's really big stuff 50 years from now, 100 years from now, no one's going to look back and say, oh, that was the age of Trump. | ||
They're going to say that was the age when after almost 4 billion years of evolution, humans took control of their own evolutionary process, and it's huge, and it's going to change all of life. | ||
And what I'm trying to do is to say everybody has to have a seat at the table, whether you're a conservative Christian, whether you're a biohacking transhumanist, everybody needs to be at the table because we're talking about is the future of our species. | ||
We're talking about the future of our species, but are we even capable of understanding the consequences of these actions, the stuff that we're discussing? | ||
Right now, I'm not. | ||
I'm talking about it, but if someone said, hey, you've got to go speak in front of people about the consequences in a very clear one-hour presentation, I'd be like, no, I'm not. | ||
I don't know what I'm talking about. | ||
One, we can go together, so you're good. | ||
Thank you. | ||
But two, the reason why I've written this book, Hacking Darwin, is I wanted to say if you could read just one book, And it's written just for everybody in a very clear way with a lot of jokes that I think are funny, my mother laughed at them as well, that you get it. | ||
And then once you know just the basics, as a human being, anybody, it has an equal right to be part of this conversation as the top scientist or the leaders of any country. | ||
I would agree with you there, but I don't think that other people are going to see it that way. | ||
I think the people that are in control, they're not going to say, Hey, we need to be fair with everyone, all the citizens of the world. | ||
How do you feel we should proceed? | ||
But that's why we need this bottom-up groundswell, but we can't have a bottom-up groundswell if just general people aren't even aware of what the issues are. | ||
And that's the challenge, and that's why forums like yours are just so important. | ||
I mean, you have all of these people. | ||
And then, you know, maybe everyone doesn't listen to this podcast and say, all right, I get it. | ||
I can go give that hour-long speech. | ||
But you can read a couple books, and then you can give an hour speech. | ||
Because the issues, like, yes, there are scientific issues. | ||
But this isn't a conversation about science. | ||
This is about values and ethics and our future. | ||
And it has to be a conversation for everybody. | ||
Yeah, it's not just a scientific conversation. | ||
It's a conversation about the future of this species and what the species will become. | ||
And that's something we're wholly unqualified. | ||
No, but here's a little vote for optimism. | ||
Okay. | ||
We have never been this literate as a species. | ||
unidentified
|
True. | |
We've never been this educated. | ||
I don't think we've ever been this nice either. | ||
I hope so. | ||
I really do. | ||
I really think that. | ||
When you look at all the wars and all the murder that used to happen, it's actually, this is the best time ever to be alive. | ||
Still sucks for people that are in bad situations. | ||
For sure, no, but it's, yes, on average it's better. | ||
And we've never been this connected. | ||
So we have, so in the book I call for a species-wide dialogue on the future of human genetic engineering. | ||
You think, oh, that's nuts. | ||
Seven billion people on earth, how are they gonna do that? | ||
But we have the opportunity and we have to try because you don't want, like with the beginning of the genetically modified crops era, the scientists were actually really responsible, but regular people weren't consulted and they felt these guys just did it to me. | ||
So if you have all the marchers with genetically modified organisms, we are entering the era of genetically modified humans and that's gonna scare the shit out of people. | ||
And so we need to start preparing and we need to make people feel that they're respected and included. | ||
And our government leaders aren't going to do it for us. | ||
So we have to find ways of engaging ourselves. | ||
And that's why with me, with the book, I set up a website where people can share their views, debate with other people. | ||
I really want everybody to be part of this conversation. | ||
How do you think it's going to play out in terms of how people, various religions perceive this? | ||
Yeah. | ||
So there's a real variation. | ||
So there are people on one end of the spectrum who believe that this is quote-unquote playing God. | ||
And if you believe that the world was created exactly as it is by some kind of divine force and that it's wrong for humans… To change, to quote-unquote, play God, it's hard to explain how you could justify everything that we've done. | ||
I mean, we've changed the face of life on this planet Earth. | ||
But I really respect people who say, look, I think that there's a line, that I believe that life begins at conception, and that any kind of manipulation after conception is interfering, that's going too far, and I respect that. | ||
And those people need to have a seat at the table. | ||
And there's certainly very strong religious views. | ||
In Judaism, there's an idea called tikkun olam, which means that the world is created cracked and broken, and it's the responsibility of each person to try to fix it. | ||
And that's a justification for using science and doing things to try to make the world a better place. | ||
And then there are now these new kind of, I mean, transhumanism. | ||
It's almost like a religion. | ||
It's this religion of science. | ||
And so we're going to have... | ||
We're humans. | ||
We're so diverse. | ||
We are going to have this level of diversity. | ||
And the challenge is, how do we have a process that brings everybody in? | ||
But it's tough. | ||
So when we're talking about genetic, any sort of genetic manipulation, we're basically talking about doing stuff to the wetware, doing stuff to the biology. | ||
What do you think about symbiotic interactions with technology? | ||
Because one of the things that I'm concerned with more than anything is this sort of inevitable path of technology getting into our body. | ||
Yeah. | ||
Whether it's through nanobots, Yeah. | ||
Fixed diseases, Yeah. | ||
or through implementation. | ||
I mean, we were talking yesterday about chips. | ||
Right. | ||
Like, what would they have to do to get you to put a chip in your body? | ||
unidentified
|
Yeah. | |
Like, what kind of powers would it have to have Right. | ||
before you accepted it? | ||
unidentified
|
Yeah. | |
Well, people are already doing it in Sweden. | ||
Sure. | ||
What are they doing in Sweden? | ||
Yeah, they're putting just little chips in their hands and under their skin, and they're using it to open doors and access things. | ||
So it's just starting. | ||
So I definitely believe, you know, right now, we look at photographs of our parents, and It's like, God, look at your hair, your clothes. | ||
That's crazy. | ||
Definitely, I think that 20 years from now, 30 years from now, people are going to look at pictures of us and say, what's that little rectangular thing? | ||
And you're going to say, that was a phone. | ||
And they'll say, what? | ||
It's like, yeah, we used to carry it around in our pocket. | ||
Well, like Michael Douglas, when you watch him in that movie Wall Street, he's got that giant brick phone on the beach. | ||
Exactly. | ||
So, we are all Michael Douglas because our technology, you're absolutely right, is not going to be something that we carry around. | ||
Technology is coming inside of our bodies. | ||
That is the future of where it's going. | ||
And, you know, people say, well, what does human genetic engineering have to do when we know that AI is going to get more and more powerful? | ||
The future of technology, the future of all of this, it's not human or AI. It's human plus AI. And that is what's going to drive our – we are co-evolving with our technology, and that's what's going to drive us forward. | ||
But you're exactly right to be afraid and to be concerned. | ||
And again, everything comes to, well, how are we going to regulate it? | ||
Are we going to have guardrails of how far is too far? | ||
Are we going to let companies just do whatever they want, or are we going to put restrictions on what they can do? | ||
I think letting the whole world decide, though, you're going to run into those religious roadblocks. | ||
For sure. | ||
And that's the challenge is that the science is advancing exponentially, whatever we do. | ||
And so we have to have our understanding of the science needs to at least try to keep pace. | ||
Regulations need to keep up. | ||
I'm part of the World Health Organization International Advisory Committee on Human Genome Editing. | ||
So we're meeting six times this year in Geneva. | ||
And the question that we're asking is, how do we think about global regulation, at least to try to put limits on the far ends of what's possible? | ||
And it's really, really difficult. | ||
But that's why we need to have this kind of process. | ||
And it seems impossibly ambitious, but every crazy idea has to begin somewhere. | ||
So you're doing every couple months? | ||
Yeah, yes. | ||
Wow. | ||
Yeah. | ||
Because they want to be on top of it as things change. | ||
Well, that's the goal. | ||
It's just, it's so hard because... | ||
Almost impossible, I imagine. | ||
It's impossible. | ||
It's impossible. | ||
So I said, and that's why even the World Health Organization, which is the lead health organization of the United Nations, it's not enough. | ||
The task is so much bigger. | ||
And that's why we need to have this kind of bottom-up groundswell that I'm pushing for. | ||
And you're absolutely right what you said before. | ||
Like, because there's not a crisis, people are focusing on other things. | ||
Open any news site. | ||
What do you see? | ||
It's not like the really important stuff. | ||
It's Trump did this or Kardashians did that. | ||
We're in this culture where there are a lot of draws on our attention, but sometimes there's really important stuff and people are afraid of it. | ||
People are afraid of science. | ||
People feel like, I remember science from high school. | ||
I didn't like it. | ||
I was uncomfortable. | ||
You know, this is for technical people. | ||
And I just feel like we can't – science is so deeply transforming the world, not just around us, but within us. | ||
And so we have to understand it. | ||
And people who are explaining science like me, the onus is on us. | ||
Like if somebody reads my book and says, well, that was really dense. | ||
That was too hard. | ||
Like, that's my failure. | ||
Like, I was giving a talk in New York a couple of weeks ago, and so I gave my talk, and I tried to make this really accessible for people. | ||
People were all jazzed up, they got it. | ||
And then there was this wonderful guy, this brilliant senior scientist at this major stem cell research center. | ||
And so the host said, all right, Jamie just talked. | ||
Can you give us a little background on the science? | ||
This guy knows so much. | ||
And he started going, and it was very technical. | ||
And I could just see the faces of the people in the audience. | ||
It was like, oh, God, what's happening here? | ||
And their level of excitement, it just shrunk. | ||
Because they couldn't really put it all in a box. | ||
And scientists aren't trained, by and large, to communicate anything. | ||
And to see in the future, so a little more than a month ago, I was in Kyoto in Japan, and I went to the laboratory of the world's leading scientists who's doing a process of what I mentioned earlier, of turning adult cells into stem cells into eggs. | ||
And so this will revolutionize the way humans reproduce. | ||
And so I was in a meeting with his top postdoc students. | ||
So these are like really the cutting edge of these technologies. | ||
And I went around to each of them and I said, here's my question. | ||
I have two questions for each of you. | ||
One, tell me what you're doing now. | ||
And two, tell me what are the implications of what you're doing now for 50 years from now. | ||
And the first question was, oh, I'm doing this and we're doing this with mouse models and people were so animated. | ||
And then 50 years from now, people just froze and it was so uncomfortable. | ||
They were like squeezing the table Just because that's not what scientists do. | ||
They are trained to say, well, this is the thing just in front of me. | ||
So I thought I was writing this book for the general public, but I'm being invited to speak to thousands of doctors and scientists because what they're saying is we get that we're doing this little piece of this, and whether it's lab research or fertility doctors or all sorts of things, but it's really hard to put together the whole story of the genetics revolution and what it means for us and for society. | ||
Yeah, man. | ||
That is interesting about scientists, right? | ||
They're just concentrating on the task at hand. | ||
I mean, wasn't that, that was like one of the big concerns about the Manhattan Project, right? | ||
This is the task. | ||
The task is how do you figure out how to do it. | ||
So they figure out how to do it, not The eventual consequences. | ||
So, when Robert Oppenheimer, who was the lead of the Manhattan Project, when that first bomb went off, I mean, he has his famous quote. | ||
Yeah, exactly. | ||
I mean, the English common translation was, holy shit, what have we done? | ||
And this science is real. | ||
But it's not one person doing it. | ||
Science has been diffused, at least with nuclear power. | ||
It was a relatively small number of people. | ||
And it was one or two states that could do it. | ||
Now, with precision gene editing, you get the Nobel Prize for figuring out how to do, you will get the Nobel Prize for figuring out how to do CRISPR gene edits. | ||
But to apply it, once the formula already exists, you get like an A- in your high school biology class. | ||
So this technology is out there, it's cheap, it's accessible. | ||
Did you go to that 2045 conference in Manhattan a couple years back? | ||
No. | ||
Do you know about all that 2045? | ||
That's part of the thing with these transhumanist folks. | ||
They believe that with their own calculations of the exponential increase of technology, that somewhere around 2045. There's a singularity. | ||
Yeah. | ||
At the very least, we're going to reach this point where you're going to be able to either download consciousness or have some sort of an artificially intelligent sentient being that's hanging out with you. | ||
Yeah, so I'm involved. | ||
I'm on faculty for one of the programs of Singularity University called Exponential Medicine. | ||
And so we're thinking a lot about that. | ||
I actually had an editorial in the New York Times a few weeks ago imagining a visit to a fertility clinic in the year 2045. And again, because we're on this exponential change, it's really hard for people to internalize, to kind of feel how fast these changes are coming. | ||
I do think, though, Ray Kurzweil, who's a really incredible genius, he thinks that we are soon going to get to a point where our artificial intelligence is self-learning. | ||
But AI, if it gets to the point where it can read something, read and comprehend, like in seconds, it will read every book ever written in human history. | ||
And then when you have all these doublings and all this more knowledge, you can imagine how that would happen pretty quickly. | ||
The counterargument against, and I think that it will, But I don't think that our human brains are, on one hand, they're incredibly complex, and they're also kind of irrational. | ||
I mean, we have all these different layers. | ||
We have our lizard brain, and every decision that we make, there's the rational decision. | ||
But then there's all the other stuff that our brains, that doesn't even rise to the level of our awareness, that our brains are processing. | ||
And right now, we only really have one really effective artificial intelligence algorithm, which is for pattern recognition. | ||
But if you think of pattern recognition as a core skill of what our brains do, our brains probably have a thousand, two thousand different skills. | ||
But the core thing is whether we reach this singularity moment or not. | ||
These technologies are going to become incredibly more powerful. | ||
They're going to become increasingly integrated into our lives and into our beings and part of our evolutionary process. | ||
There's no longer, oh, we just have our biological evolution and our technological evolution, and those are separate things. | ||
They're connected. | ||
It's going to be that weird question of whether or not if an artificial intelligence is going to be able to absorb all the writing that human beings have ever done and really understand us. | ||
Yeah. | ||
Will they really still be able to understand us just because they get all the writing? | ||
So right now, you would say no. | ||
I'd say no, yeah. | ||
But 20 years from now? | ||
50 years from now? | ||
100 years from now? | ||
They could come up with a reasonable facsimile. | ||
I mean, they could figure out a way to get it close enough to... | ||
Yeah. | ||
You know, where it's like her, like that. | ||
Yeah, yeah. | ||
That's an essential point. | ||
Because I think when people imagine this AI future, they're imagining like some intimate relationship with some artificial intelligence that feels just like a human. | ||
I don't think that's going to happen. | ||
You don't? | ||
Well, no, but just because AI, it will be its own form of intelligence. | ||
And it may not be, frankly, we wouldn't want AIs with these brains like we have that have all these different impulses that are kind of imagining all this crazy stuff. | ||
We may want them to be more rational than we are. | ||
Chimpanzees are our close relatives. | ||
They don't think just like us. | ||
We're not expecting them to think like us. | ||
They're their own thing. | ||
And I think AIs will be their own things. | ||
Will we be interacting with them? | ||
Will we be having sex with them? | ||
Yes. | ||
But it's not going to be that they're just like us. | ||
They're going to be these things that live within us, live with us, and together we're going to evolve. | ||
Well, they're certainly already better at doing certain things like playing chess. | ||
I mean, it took a long time for an artificial intelligence to be able to compete against a real chess master, but now they swamp them. | ||
And they learn quickly, like incredibly quickly. | ||
They teach themselves. | ||
Yeah, so first we had chess, and chess people said, oh, that's what it means to be a human. | ||
The computers will never beat humans at chess. | ||
Now, it's like everyone says, well, no human could ever compete. | ||
And then they said, well, there's this Chinese game of Go, which kind of when people here look at it, it looks kind of like checkers, but it's actually way more sophisticated, way more complicated than chess. | ||
I heard that there are more moves in Go, more potential moves than there are stars in the universe. | ||
Yes, yes. | ||
So then they had AlphaGo. | ||
That this company DeepMind, which was later acquired by Google, they built this algorithm that in 2016 defeated the world champions of Go. | ||
People thought that we were decades away. | ||
And then DeepMind created this new program called AlphaZero. | ||
And AlphaZero, with AlphaGo, they gave it access to all of the digitized games of Go. | ||
So it very quickly was able to learn from how everybody else had played Go. | ||
AlphaZero, they just said, here are the basic rules of Go. | ||
And they let AlphaGo just play against itself with no other experience other than here are the rules and play against. | ||
And in four days, AlphaZero destroyed AlphaGo. | ||
And then... | ||
Alpha Zero destroyed the world champions of chess and destroyed every other computer program that had ever played chess. | ||
And again, those computer programs had internalized all the chess games of Grandmasters. | ||
Alpha Zero had not internalized any. | ||
It just played against itself for a few days. | ||
And then Shogi, which is a Japanese traditional game, kind of like chess, it destroyed the Grandmasters of that. | ||
That's what I'm saying. | ||
The world is changing. | ||
It's changing so much faster than we anticipate. | ||
And we have to be as ready for that as we can. | ||
I think we need to come to grips with the fact that we're way stupider than we think we are. | ||
We think we're really intelligent, and we are. | ||
In comparison to everything else on this planet. | ||
But in comparison to what is possible, we are really fucking dumb. | ||
In comparison to what this computer can do and what the future of that computer is going to be. | ||
Maybe that computer is going to redesign another computer. | ||
This is good, but I've got some hiccups here. | ||
Yeah. | ||
No, it's true. | ||
And yet the technology is us. | ||
Right. | ||
Sort of. | ||
It's not like this technology is some alien force. | ||
It's like we create art. | ||
unidentified
|
Collectively. | |
Yeah. | ||
You mentioned cities. | ||
We create these cities, which are these incredible places where dreams can happen. | ||
Cities like here in Los Angeles or New York, where I'm from. | ||
So this technology is us. | ||
And the challenge is, how can we make sure that this technology serves our needs rather than undermines our needs? | ||
unidentified
|
Right. | |
Yeah, and whether or not our needs supersede the needs of the human race or supersedes the needs of the planet. | ||
Yeah. | ||
We're almost too much chimp, right, to contemplate these critical decisions in terms of how it's going to unfold from here on out. | ||
We really might, not we, but the people that are actually at the tip of the spear of this stuff, they really might be affecting the way the planet exists. | ||
Absolutely. | ||
And we're doing that now. | ||
I mean, there was an article that came out the other day. | ||
There's a million species that are on the verge of extinction. | ||
We are driving all these other species to extinction. | ||
We're warming the planet. | ||
So humans are the determining factor in many ways for how this planet plays out. | ||
And that's why, in my mind, everything comes back to values. | ||
You're right. | ||
We have this... | ||
Lizard nature, this monkey nature. | ||
It's who we are. | ||
You wouldn't want to take that away because that's the core of what we are. | ||
And yet we're also a species that has created philosophy. | ||
We've created beautiful religions and traditions and art. | ||
And the question is, which version of us is going to lead us into the future? | ||
If it's this tribal primate with these urges, that's really frightening. | ||
If we can say... | ||
We've done better and worse in history. | ||
And we had this terrible Second World War. | ||
And yet at the end of the Second World War with American leadership, the world came together. | ||
We established a united nations. | ||
We established these concepts of human rights. | ||
Like you can't just kill everybody in your own country and say, hey, it's just my business. | ||
So we have this capability, but it's always a struggle. | ||
I mean, these forces are always at war with each other in many ways. | ||
unidentified
|
Yeah. | |
It's just too much to think about. | ||
Yeah, but we have to. | ||
I know, we do have to. | ||
One of the things that's always been amusing to me is that we seem to have this insatiable desire to improve things. | ||
And I've always wondered why. | ||
But is that maybe because this is what human beings are here for? | ||
It's what we do. | ||
It's who we are. | ||
Right. | ||
But is this a product? | ||
It's just a... | ||
Us being intelligent, trying to survive against nature, and predators, and weather, and all the different issues that we came up, that we evolved growing up and dealing with. | ||
And then now, we just want things to be better. | ||
We just want things to be more convenient, faster, but more data. | ||
You're aware of Elon Musk's neural link technology. | ||
unidentified
|
Sure. | |
How much do you know about it? | ||
I know a decent amount. | ||
And my friend Brian Johnson, he has a company, Kernel. | ||
There's a few different companies that are trying to think about these brain-machine interfaces. | ||
And what are they trying to do? | ||
Basically, what they're trying to do is to find a way to connect our brains to our machines. | ||
And there's a little bit of progress. | ||
Our brains are incredibly complicated and they're messy. | ||
I mean, there's a lot that's happening. | ||
But we are increasingly figuring out how to connect our brains to our technology. | ||
And so people are imagining a time when we can do things like download memories, download ideas, or upload memories and upload ideas. | ||
And there's some very early science that is suggesting that this will be The very early days. | ||
But Elon was giving the impression that sometime this year they're going to release something. | ||
You know, they may release something, but it's not going to be something that's going to change the world. | ||
Because that technology is way more nascent than even the genetics technology that I've been talking about. | ||
It's not like it's at all remotely possible that this year you're going to be able to upload a full memory or download a full memory. | ||
But there are little things that are happening, but every journey begins with a step. | ||
But the technology is fairly transparent in terms of where the state of the art is right now? | ||
It is in that it's extremely early. | ||
This stuff is... | ||
So when you think about systems that we understand, I mentioned that we understand just a little bit about genomics. | ||
We know less about the brain. | ||
The brain is kind of the great unknown of this universe. | ||
We know more about the oceans than we know about our brain. | ||
I mean, we know very, very little. | ||
We understand that if you kind of stick an electric current in somebody's brain, like that's going to be if you kind of shoot a spike through somebody's head. | ||
But really understanding how the brain functions, we're still in the very, very early days. | ||
So do you think that Kurzweil's off with this idea that you're going to be able to download your consciousness into a computer? | ||
Because that's one of the most controversial ideas that he's come up with. | ||
So I think he's I think he's off based on your use of the word your. | ||
So I mentioned that a month ago I was in Kyoto and I was at this at the stem cell lab. | ||
But I also went to another lab of a guy named Hiroshi Ishiguro, who's the world's leading humanoid roboticist. | ||
And so he's the guy who was on the cover of Wired and he's created these robot avatars. | ||
And like I had a conversation with this robot woman, And it was really interesting because I could see that if I would smile, she'd smile and lean forward. | ||
And if I had an over-exaggerated sad face, she'd change her expression. | ||
And she can have basic conversations. | ||
But we're still a long way, and so from having full robotics, but I had this, for robotic human interactions, but I had this debate with Ishiguro, and he was saying that he thought that the future of humanity was non-biological, that we were going to kind of unload ourselves to these non-biological entities, and that is how we would gain our immortality. | ||
And I argued something very different. | ||
I feel like we are biological beings. | ||
I think we'll fully integrate with our technology, but if we ever become entirely non-biological, then that's not us. | ||
Either we will have committed suicide as a species, or these robots will have killed us. | ||
Because even if, let's just say, that I could download my entire consciousness to some kind of robot, and let's just say that was possible, that robot would be me for that first Right. | ||
he's directionally right. | ||
Yeah, the problem would be that you would be locked in, like, if they downloaded your consciousness into some sort of a bank of computers somewhere, where are you? | ||
If that's your consciousness, your consciousness is in these ones and zeros? | ||
Yeah. | ||
I mean, that's terrifying. | ||
unidentified
|
Yeah. | |
What's terrifying is if someone didn't like you and they said, I'm going to make one version of you suffer for all eternity. | ||
Yeah, yeah, yeah. | ||
And I'm going to just download you while you sleep. | ||
unidentified
|
It's true. | |
But I have something worse than that. | ||
Okay. | ||
Death. | ||
And so I think that nobody is going to say, well, I'm going to be Joe living a life or I'm going to like not be Joe and I'll just have my consciousness downloaded someplace else. | ||
And so if the comparison is, well, I've lived this life and I don't want to die. | ||
And so I'd rather kind of be here in some kind of version. | ||
And even if it's not me, it's just something. | ||
I think some people will want that. | ||
I think some people will, but they don't know what they're getting, right? | ||
In terms of you don't know what that experience is going to be like, nor do you know if there is some sort of a chemical gateway that happens in the mind when you do expire and allows you to pass through to the other dimension that your consciousness and your soul longs to travel to. | ||
I hope you're right about that. | ||
I'm definitely not right. | ||
I've written about this in my novel. | ||
It's like, yeah, but I think kind of when you're dead, you're just dead. | ||
Why do you think that, though? | ||
Well, just because I think that this kind of immortality comes because time stops. | ||
Time is this relative concept. | ||
And so at the moment that you die, that's immortality for you because time stops flowing for you. | ||
That's what Einstein taught us. | ||
Time is this relative concept. | ||
Other people, very legitimately, and there's no way to prove it, feel that we have this soul and this soul can travel to other dimensions. | ||
I happen to believe that we are biological beings and our experience of the soul, whatever, is connected to our biology. | ||
When our biology stops functioning, those experiences, whatever they are, stop being accessible, at least to us. | ||
Have you had any psychedelic experiences? | ||
I haven't. | ||
And I was so tempted. | ||
We started the interview talking about my chocolate shamanism. | ||
You haven't had anything? | ||
I haven't. | ||
Do you want to? | ||
I don't think so, and I'll tell you why. | ||
So I listened to the Michael Palin interviews, and he had his great conversation with Sam Harris, and I really think that this psilocybin stuff is real. | ||
You just got decriminalized in Denver. | ||
I know, in Denver. | ||
I was just there the other day. | ||
But, as I said before, I think that the ultimate drug is us. | ||
And so for me, I would rather, and I definitely think that our awareness, it doesn't encompass everything that is knowable, everything that we could know, but we hem ourselves in. | ||
And if we want to get out of those limitations, certainly drugs are ways that people have used for many thousands of years. | ||
not taking the drug. | ||
Like I would rather, what does that mean? | ||
It's like, if you're taking psilocybin, you're having this great experience, but then tomorrow you're not taking psilocybin. | ||
And so kind of your consciousness has narrowed. | ||
My aspiration would be to recognize that the drug is us. | ||
That if we want to expand our consciousness, there are all kinds of ways, whether it's meditation or awareness or just simple appreciation. | ||
That's when I do these cacao ceremonies. | ||
What I say is you have this cacao in front of you, but it's not just this. | ||
Think of... | ||
The person in Honduras who planted the seed, the person who watered that seed, the person who took the plant, the person who paved the road to bring the plant. | ||
And I just think that we can expand our consciousness through our own means, and then we always have access. | ||
I hear what you're saying, but you're saying this from a person that's never had psychedelic experiences in a It's really preposterous. | ||
If you did experience what psilocybin can do to you, you definitely wouldn't be saying it this way. | ||
You also wouldn't be thinking that you take it and then you're not on it anymore because it's profoundly influential for your perspective in regards to the whole rest of your existence. | ||
There's many people that have had psychedelic experiences that Think about it as a rebirth, that they've gone through this and changed. | ||
So why would you have this rigid thought process about drugs and not drugs, but yet you don't have it about cacao, which is a mild drug? | ||
Yeah, and so you're right that it may not be entirely consistent. | ||
Some of the people you've described are good friends of mine who've really done it, and I've talked to them about it, and I'm endlessly curious. | ||
So why don't you do it? | ||
The reason is, so far, I have been on this journey to see what's possible within myself. | ||
And I'm still on that journey. | ||
I don't want to close off any possibility for anything. | ||
How would you assume that it would close things off? | ||
That's what's confusing. | ||
Yeah. | ||
It's just opening you up to a new experience that other people have found to be profoundly influential. | ||
Yeah. | ||
So far, you're very resistant to this. | ||
Like, even when I'm talking, you're like, yeah, yeah, yeah. | ||
Like, you can't wait to come back with your own rational perspective. | ||
unidentified
|
No, you're right. | |
You know what's so funny? | ||
I will come back to this. | ||
From the perspective of a person who hasn't experienced anything. | ||
You're so right. | ||
And just, I think it's such a great point. | ||
Because I'm very close with, actually, with the Tibetans. | ||
One of my closest friends is the prime minister of the Tibetan exile government. | ||
So, I've been many times to Dramsala in India. | ||
I've met with His Holiness the Dalai Lama many times. | ||
And the most incredible thing about meeting with these guys, and they are all people who've found these incredible states of heightened consciousness, so much that their brains are changed when they go into the fMRI machines. | ||
But When you have a conversation with them, it's not like what we do. | ||
Like, exactly. | ||
And thank you for calling me out. | ||
Like, you say something. | ||
I've already, in my head, countered what you're saying before you're finished saying. | ||
That's why you kept saying, right. | ||
Exactly. | ||
Which probably means that you made me a little uncomfortable, which is good. | ||
That's what we want. | ||
And these guys, it's like you'd talk to them, and they would just be so tuned in to what you were saying. | ||
And they would just kind of think about it. | ||
Then you'd finish. | ||
And then they'd kind of look up. | ||
And then, because we're Americans, we want, you know, somebody stops speaking. | ||
Like, you have to, if you don't speak right away. | ||
And then there was, like, a minute. | ||
And then it's like, you're kind of looking around. | ||
Was it me? | ||
Did I say something? | ||
And then they would come back with this incredibly thoughtful thing. | ||
So, you're right. | ||
You know, I don't want to close off any possibility. | ||
There are many different things that we could do. | ||
The path that I have been on, and certainly with this cacao, and you're right, cacao is like a... | ||
I mean, not a huge one, but like a mild drug. | ||
And life is a mild drug. | ||
Steve Simon: So, yeah. | ||
Steve Simon's not very mild. | ||
John Kuehl: Yeah. | ||
No, and my friends who have done it have said exactly what he said. | ||
So, I have a good friend of mine who did it. | ||
And then he said, "It showed me a path to a different identity, a different consciousness." And now he said, I don't do psilocybin, but I do daily meditation, but I can see where I'd like to go, what's possible. | ||
So I get that. | ||
Yeah, there's a bunch of different substances out there that have very similar profound effects. | ||
Yeah, yeah. | ||
There's a real thought, and this is something that Terence McKenna described way back in the late 90s, early 2000s. | ||
He believed that you're going to be able to recreate a lot of psychedelic states through virtual reality so that people that don't want to actually do a drug will be able to experience what it's like to be on that drug. | ||
And that, I mean, it's very theoretical and hypothetical. | ||
Yeah. | ||
You know, who knows whether or not that's possible. | ||
But that could be one other way that human beings interface with technology. | ||
So humans, in my view, are far more hackable than we think. | ||
That there are so many that we just imagine our biology as being fixed, but our biology is really variable. | ||
Like I have a friend who's an anesthesiologist at Stanford, and she's done this experimentation of just running just like very mild electric currents through people's brains. | ||
And people have these very real experiences. | ||
And whether it's arousal or something, that you're entering people's minds through different ways. | ||
So you talk about virtual reality. | ||
We are entering a world where the pixelation of virtual reality will be equal to life. | ||
And so you're going to be in this VR space. | ||
And it will look. | ||
It may smell. | ||
It could be with haptic suits. | ||
It may feel just like life. | ||
And our brains, I don't know if you've done these things with these VR glasses where You go and you can see people. | ||
You're just like in a hall. | ||
And you put on the glasses. | ||
And now you're in an elevator going to the top on the outside, like a window cleaner's elevator to the top of this high rise. | ||
And there's a little rickety board. | ||
And then there's this cat at the end of the board. | ||
And they're saying, yeah, you have to go. | ||
Go save the cat. | ||
And you've already seen that you're just in a hall. | ||
You know it in your brain. | ||
There's this cat. | ||
Everybody's looking at you, and you've seen all these other people panic. | ||
And you think, well, when I'm there, I'm going to be so... | ||
I'm just going to go grab that cat. | ||
And you're terrified. | ||
Like, you're trying to override your lizard brain. | ||
And your lizard brain is saying, like, no, don't step... | ||
They have that for our HTC Vive. | ||
Yeah. | ||
We need to set that up, Jamie. | ||
We need a two-by-four that we put on the ground for that. | ||
Oh, it's incredible. | ||
And so I think that this whole concept of reality is that our technology is going to be changing our sense of reality. | ||
And then what's real? | ||
Like if you feel arousal in your brain because of an electric current or you feel it because you're with your girlfriend or wife or reading a magazine or whatever, is that different? | ||
Is one real and one not real? | ||
I mean, I think that there's real big issues here. | ||
Yeah, I mean, that is the matrix, right? | ||
And if it feels better and it's more enjoyable than real life, what is going to stop people from doing the Ray Kurzweil deal and downloading yourself into this dimension? | ||
Not much. | ||
Whether it's possible to do a full download or not, I mean, I think that's an open question. | ||
But whether people are going to be more comfortable living in these alternative worlds, and whether we're going to be able to say, oh, no, that is the fake world. | ||
Like, if you're in this virtual world, but you have friends in that world, you're interacting in that world, you have experiences that feel every bit as real in that world as in our world, And people say, oh no, that's not real. | ||
Those aren't your friends. | ||
Like, even now. | ||
Like, you know, we all, people with global lives, you kind of have these friends. | ||
Like, I have a good friend in Mongolia. | ||
We talk all the time. | ||
Do you ever see them in person? | ||
Once in a while. | ||
Like, once every year or two. | ||
It's great to see them. | ||
Well, that's a real person, though. | ||
No, it's a real person. | ||
But that's someone you actually know. | ||
No, no, this isn't like a pen. | ||
No, but I don't mean that. | ||
I mean, you actually do know them. | ||
unidentified
|
No, absolutely. | |
But if it was someone that you only talked to online and they lived in Mongolia, that's where things get weird. | ||
It's true, but let's just say, following that hypothetical, you have that person. | ||
They're part of your whole life. | ||
And they're with you. | ||
They're with you through your life experiences. | ||
You call them up when you're sad. | ||
Is it so essential that you've met that person physically? | ||
Is that the core idea? | ||
It's not essential, but it means a lot. | ||
It does. | ||
Because we are these physical beings and we are these virtual beings, but figuring out what's the balance is going to be really tricky. | ||
Yeah, what is the balance? | ||
I'm worried about augmented reality, too. | ||
When you see people that use Snapchat filters and they give themselves doggy ears and stuff like that, how long before that is just something that people choose to turn on or turn off about life itself? | ||
You'll be able to see the world through different lenses. | ||
The sky could be a different color. | ||
The plants could be a different color. | ||
Yeah. | ||
I might write about this in one of my novels, Eternal Sonata, where I think we're just going to have these contact lenses, and it'll be different kinds of information based on what people want. | ||
I mean, like, I'll meet with you, and it'll say, all right, this is Joe, here's a little bit of background, whatever, and we'll have useful information. | ||
Or you're walking around a city, and you'll get little alerts of things you might do, or history. | ||
That's what they were thinking about with Google Glasses, right? | ||
Yeah, I know, but it just was so annoying that people wanted to kill people. | ||
It was just too weird. | ||
Yeah, yeah, yeah. | ||
Yeah, it was just too, everybody felt like they were getting filmed, too. | ||
They were. | ||
Yeah, I mean, when you're walking around with Google glasses on, you assume that people were recording everything. | ||
It was very strange. | ||
They were, but I think that's another thing that we're just, all of our lives are going to be recorded. | ||
Of course. | ||
unidentified
|
Yeah. | |
Now, do you think that that's going to come in the form of a contact lens, or do you think it's going to come in the form of ski goggles that you're going to put on? | ||
Nobody wants to look like an idiot. | ||
That's not true. | ||
No, but in the beginning, you talked about Michael Douglas, or my favorite one is Kurt Russell in Escape from New York. | ||
What do you have there? | ||
It's like this really cool tech. | ||
He finally gets out of this Manhattan hell, and he's got this phone, and it's like this big. | ||
So there's Mr. President's cell phone? | ||
It's like very, very early days. | ||
And so now we have these kind of glasses, and there's like a little bit of cachet. | ||
Tom has a cell phone that's that big. | ||
Have you seen that? | ||
No. | ||
I was at the Verizon store. | ||
Yeah. | ||
I think it's like an attachment or an accessory to a phone. | ||
Yeah, it's like you can bring it with you and not bring your other phone. | ||
That's the idea behind it. | ||
So you could decide, well, I'm going out, let me just bring my tiny phone for essentials. | ||
But all this tiny, I mean, the phones are going to get, quote-unquote phones are just going to get so small. | ||
That's why I say they're going to come inside of us. | ||
You'll have like a little contact lens, maybe a little thing in your ear, maybe like a little permanent implant behind your tooth. | ||
Or one of your teeth will replace a tooth with a computer. | ||
Pipe it right into your nerves. | ||
Any kind of crazy stuff you can think about is probably going to happen. | ||
Some of it will take, some of it won't. | ||
Yeah, it seems like that's what we're going to have to see, like how it plays out. | ||
And that's one of the things when you were talking about scientists that are working on these things, they're working on what's right in front of them. | ||
They're not looking at the greater landscape itself in terms of what the future holds. | ||
It's not their job, and that's why we need other people. | ||
I certainly see myself— Who are those people? | ||
You and who else? | ||
Who would you elect? | ||
If Trump came to you and said, Jamie, we've got problems. | ||
We need to figure out the future. | ||
What should we do? | ||
So, certainly, we need to have a mix of different kinds of people. | ||
And so I – certainly people like me who are kind of big picture futurists, we need that. | ||
We need scientists. | ||
I work with some really incredible scientists. | ||
I did an event at Harvard last week with George Church who's kind of like the living Charles Darwin. | ||
David Sinclair who you know has been on this program is a friend of me working on life extension. | ||
And we need people just from all backgrounds. | ||
And so like I would say like we need people. | ||
We need poor people. | ||
We need people from – We need a developing world. | ||
We need all kinds of people. | ||
But in terms of the people who are kind of articulating the big picture of the world and what are the challenges that we're facing, I certainly put myself in that category. | ||
People like Yuval Noah Harari who are just kind of big, also kind of big thinkers, people like Sid Mukherjee. | ||
And I just think we have to articulate the big picture and we have to do it in a way so that people can see themselves in this story And then enter into the conversation. | ||
Are you writing this book just to sort of educate people and let them understand exactly what is going on and that it is a really volatile and chaotic and amazing time? | ||
And that all these things are... | ||
Are you doing this book to... | ||
What essentially was Hacking Dharma? | ||
What was the motivation behind it? | ||
Was it for... | ||
A person like me? | ||
Or was it for everyone? | ||
It was for everyone. | ||
And so what I really wanted to do... | ||
So the background... | ||
I can give you just a little bit of background. | ||
So more than 20 years ago, I was working on the National Security Council. | ||
And my then boss, Richard Clark, who was then this obscure White House official who was jumping up and down saying, we need to be focusing on terrorism and Al-Qaeda and Bin Laden. | ||
And he was trying to tell everybody... | ||
And nobody was paying attention to it. | ||
It was totally marginalized. | ||
And when 9-11 happened, Dick's memo was on George Bush's desk saying exactly that. | ||
We need to focus on Al-Qaeda. | ||
Here's what's going to happen. | ||
And Dick, even before then, would always tell me that if everyone in Washington was focusing on one thing, you could be sure there was something much more important that was being missed. | ||
And so more than 20 years ago, I was looking around. | ||
I saw these little pieces of disparate information, and I came to the conclusion that the genetics revolution was going to change everything. | ||
So I I educated myself. | ||
I started writing articles. | ||
I was invited to testify before Congress. | ||
And then to try to get that story out, I wrote my two most recent near-term sci-fi novels, Genesis Code and Eternal Sonata. | ||
And when I was on book tours for those, and I explained the science to people, the way I kind of a self-educated citizen scientist... | ||
And a novelist would explain the science. | ||
All of a sudden, people got it. | ||
And that was when I realized I needed to write a book about the genetics revolution that people could absorb, that wouldn't scare people. | ||
But my mission for the book is that this stuff, as we've talked about, it's so important. | ||
Everybody needs to be part of the conversation. | ||
We have this brief window. | ||
And what I'm calling for is this species-wide dialogue on the future of human genetic engineering. | ||
And I have a whole game plan in the book about what people can do, how they can get involved, people individually, on a national level. | ||
We have to put a lot of pressure on our elected leaders and say, stop focusing on the crap. | ||
There is really important stuff that needs to be addressed, and we need leadership. | ||
I'm speaking in Congress a week and a half from now, talking about these issues. | ||
So we need to have... | ||
And on an international level, we have to have some kind of international system. | ||
We're so far away from being able to do that. | ||
We don't even know what the standards are, but we have to be pushing. | ||
So you think of it in terms of the same way we have with nuclear weapons? | ||
Yeah. | ||
In many ways, yeah. | ||
But the thing is, with nuclear weapons, a lot of that happened at the state level, at the country level. | ||
This needs to happen at a popular level and at a government level. | ||
So the only way that that's going to happen effectively is we need real comprehensive education on this subject. | ||
It's not something that people can just guess, right? | ||
unidentified
|
Absolutely. | |
They need to know what's the consequences, where we're at right now, right? | ||
Yeah. | ||
Yeah, and that's like Sanjay Gupta had a wonderful quote that's actually on the cover of my book, which is, if you can read one book on the future of our species, this is it. | ||
So what I've tried to do is to say, like, if you just want to go to one place to understand what's happening, what's at stake, what it means for you, and what you can do now if you want to get involved, I've tried to do that. | ||
Ironically, I'm now, as I mentioned, being asked to speak to thousands of doctors and scientists because they're all reading this book and they're saying, this is positioning my work in a much bigger context. | ||
Sanjay Gupta is a very interesting cat because he was very anti-marijuana and then started doing research on it and then totally flipped 180 degrees, which to me is a great sign of both humility and intelligence. | ||
He recognized that the data was different than his presuppositions. | ||
Yeah. | ||
He had these prejudices that were very common. | ||
Yeah. | ||
Now, when you speak to Congress, do they brief you in terms of what they would like specifically for you to address? | ||
No. | ||
So this one, I've been asked to go and speak, and a lot of members of Congress are going to be invited. | ||
And what I'm going to tell them is, look, this is really important. | ||
Our Congress is not doing enough. | ||
And here are the things that we need to do. | ||
And what are you going to say to try to really get it into their head? | ||
What I'm going to say is that the genetics revolution is here. | ||
If we don't have a system, if we don't have a rational system to manage it, if we don't have a system, you talked about public education, the challenge that we face in the United States is we traditionally have had a representative democracy. | ||
And now we're transitioning from a representative democracy to a popular democracy. | ||
So Switzerland has a popular democracy, but they have really well-educated people who have enough information to make smart decisions. | ||
We haven't educated our public, and yet the public is making big decisions, and a lot of it is happening just on a gut feeling. | ||
That's what's happening with trade agreements, where people just have a feeling it's bad, Without the ability to really get into the details. | ||
And so we are having that transition, which means there's a lot of responsibility on us to educate our public. | ||
And it's a tragedy. | ||
We treat people in this country like you can just throw people away. | ||
Like if you're in some crappy school system and your chances of success are so minimized, not because of anything that you've done, just because of your circumstances. | ||
And it's unacceptable. | ||
It is unacceptable. | ||
I mean, equal opportunity is what we really should all strive for. | ||
And I think some people conflate that with equal success. | ||
And you're not going to get the same equality of outcome. | ||
You're going to get different amounts of effort and different people are qualified or more talented at different things. | ||
But what I'm worried about is what I said initially, that some people are going to get a hold of this stuff quickly, and it's going to give them a massive advantage. | ||
The first person that has the ability to go forward in time five minutes is going to be able to manipulate the stock market in an unprecedented way. | ||
I don't think that that's really possible in our lifetime, but that's the kind of thing I'm talking about. | ||
You could get so far ahead that if we're talking about competition, there will be no catching up. | ||
But you don't have to travel in time to do that. | ||
Right. | ||
But I'm saying if that was a technology. | ||
But there's real technologies that are likely to happen, which are going to confer billions, tens, hundreds of billions of dollars of benefit. | ||
Of advantage. | ||
Yeah. | ||
And that stuff is happening now. | ||
And this is the concern with getting behind countries like China. | ||
Because if they get ahead of us in something like that, they're already moving in this direction in terms of technology. | ||
Yeah, so I am a proud American. | ||
My father and grandparents came here as refugees. | ||
I believe in what this country at our best stands for. | ||
I want us to continue to be the country that's setting an example for the rest of the world, that is articulating what are ideals of Responsibility and governance, good governance and accountability and all these things that we've championed. | ||
And because of that, I want us to get our act together politically and I want us to be the leading technological country in the world. | ||
And so I think that's what's at stake and we're losing so much time because there was a time in the period after the Second World War Where we recognized that technological leadership was the foundation for everything else. | ||
We had recreated the world out of the ashes of the war. | ||
But we realized that we needed to have the economic growth. | ||
We needed to have the competition. | ||
We needed to have these technologies. | ||
And it was a miracle what we've done. | ||
And now we've lost our focus and we have to regain it. | ||
The way I look at humans and the way I look at the human race today in 2019 is like we're driving very fast through fog, and it's very difficult to see what's in front of us. | ||
When I look back at – I don't know if you've ever read any H.G. Wells, but some of his predictions about the future are really interesting, because he was pretty close on quite a few things. | ||
But that vision, to be able to sit there and use your imagination, close your eyes and think, what is this going to be like? | ||
What we're dealing with now, as opposed to what? | ||
2119. Right. | ||
Which is similar to H.G. Wells versus us, right? | ||
What the fuck is that going to be like? | ||
Yeah. | ||
Is there going to be a time where there are no diseases, there is no death, and that we just have to regulate population control in some sort of other manner? | ||
And the only way people die, they're going to die from accidents and things along those lines. | ||
But mortality in terms of old age? | ||
According to David Sinclair, this is a fixable issue. | ||
It's a matter of when they fix it. | ||
Do they fix it in 20 years or 30 years or 50 years? | ||
Maybe. | ||
David is a friend and I have a whole chapter. | ||
I don't want to misquote him either. | ||
I have a whole chapter in the book on the science of human life extension. | ||
So I think definitely it's real that we're going to live healthier longer. | ||
We're going to harness our technology for that. | ||
I don't think that immortality, that biological immortality is in the cards for us. | ||
Maybe not immortality because we'll still be biologically vulnerable. | ||
We'll have hearts and brains and all that stuff. | ||
Yeah, I think we will age slower and we will live healthier longer and I think it's going to be great. | ||
But back to your core point, I mean that's the reason why I also write science fiction is that the world of science is changing so fast that we really need to apply a lot of imagination to imagine where it's going. | ||
Because if you're just looking at what's happening now, It's like this train is going to speed by you. | ||
We have to kind of imagine – it's like Wayne Gretzky. | ||
We have to imagine where the puck is going to be, not where it is now. | ||
And I mentioned George Church, who's like – he's at Harvard. | ||
He's like the living Charles Darwin. | ||
And I do a lot of speaking alongside George. | ||
And it's become our little thing that he says that he reads science fiction. | ||
Like mine, and then says, well, that's pretty cool. | ||
How can we do that? | ||
And what I do is I look at the research coming out of labs like George's, and I say, all right, well, that's where we are now. | ||
What's that going to mean in 20, 50, 100 years? | ||
Science fiction plays a more important role than it ever has in kind of imagining where we're going, and it's that imagining that allows us to try to say, well, what if that's one of the options of where we're going? | ||
so that we can have a better outcome rather than a worse one. | ||
If you're a gambling person, I don't know if you are, but if I had to give you $100 to put on something in 20 years that's going to be profoundly just changing It's going to change us in a way that is something that we're not really prepared to understand or deal with. | ||
What do you think that's going to be? | ||
I think it's going to be predictive genetics. | ||
Right now, it's like you go to your doctor when you're sick. | ||
This could have been some genetic disorder that you had from the moment you were conceived. | ||
And it was ticking. | ||
And it was ticking. | ||
And you showed up 50 years later when that's been manifest. | ||
So it's going to be very different. | ||
different. | ||
You're taking your kid home from the hospital, your newborn. | ||
And the doctor says, hey, congratulations, this is really great. | ||
But just FYI, your kid has a 50% greater than average chance of getting early onset Alzheimer's 50 years from now. | ||
And your kid has a really great chance of being phenomenal at abstract math. | ||
Like, how are we going to think about that? | ||
How are we going to think about what it means to be human when we have all of that information? | ||
And And there are things now that we call fate. | ||
And it's just a different model. | ||
And so I think that, and once we have that, that's going to change a lot of things. | ||
It's going to fundamentally transform our healthcare. | ||
What we call healthcare now is really sick care. | ||
You show up with a symptom. | ||
And this is going to be predictive. | ||
And it's going to change the way we make babies because people are going to have real choices about which embryos to implant. | ||
And we're going to have a lot of information about a lot of really intimate stuff. | ||
So you feel like genetic manipulation and genetic engineering, genetic understanding, genetic knowledge, and then applied genetic medicine. | ||
unidentified
|
Yes. | |
Those are going to be the big changes in the next 20 years, even more so than technology? | ||
Well, it's interconnected because there's really, it's like a super convergence of these technologies. | ||
So the genetics revolution is the artificial intelligence revolution in the sense that the complexity of genetics is so great. | ||
It's Way beyond what our brains on their own could process. | ||
And so really all these technologies are touching each other. | ||
And so the biological models are now influencing the AI. So for example, we are coming to the limits of silicon storage. | ||
But DNA has unlimited storage capacity. | ||
So it's, as I've said before, the boundaries between biology and AI, or genetics and AI, is going to be very blurry. | ||
Yeah, that is an interesting concept, right? | ||
The idea of storing information in DNA. And that has been discussed. | ||
Yeah, DNA is the greatest information storage mechanism ever imagined. | ||
But the question is, what happens when you do store things in there, and how does that information interact with all the rest of the stuff that's in your body already? | ||
Well, I mean, if you can do it in your body, it doesn't have to be in your body. | ||
unidentified
|
Right, it doesn't have to be external. | |
But just think, like, your DNA has four billion years of history, and it's done a great job of recording it. | ||
It's incredible. | ||
unidentified
|
Yeah. | |
I have old 8-track tapes. | ||
They haven't lasted. | ||
That is a squirrely concept that you have all that data inside your head. | ||
I mean, that's also when people make, when they try to understand instincts that people have. | ||
These are some sort of genetically encoded memories or some understanding of things that are dangerous. | ||
unidentified
|
Right. | |
And that these, they're in there because this is how we've learned over the years without actually having to experience these things personally. | ||
Yeah. | ||
Yeah, no, so that's, it's baked in. | ||
Our genetics are baked into us. | ||
And so, you know, I don't know if you've been to Indonesia before. | ||
I was in Indonesia and I went to this place called Komodo Island. | ||
Oh, wow. | ||
Where the dragons are? | ||
It's the Komodo dragons are. | ||
And it was fascinating because it's like you can tell they don't have plaintiff's attorney. | ||
So you're just walking around. | ||
There are all these Komodo dragons. | ||
Yeah, these are like the most deadly creatures on earth. | ||
And there's like some little guy with a little stick. | ||
And it's like, how effective is that stick? | ||
So you're just walking around? | ||
You're just walking around. | ||
Because the Komodo dragons, when they're not killing people or killing animals, they're just sitting there. | ||
Oh, Jesus Christ. | ||
So it's pretty scary. | ||
Do they ever get jacked? | ||
Do people ever go there and get bitten? | ||
Yes. | ||
And they say, oh, it's only a few times a year. | ||
It's like, wow, a few times a year. | ||
That seems like a lot. | ||
A few times a year is a lot. | ||
Anyway, but the way it works for a Komodo dragon, a mother lays the egg and then buries the egg and then forgets where the egg is. | ||
And then let's just say that this egg hatches and this little Komodo dragon comes out and the mother sees her own baby Komodo dragon. | ||
She'll eat it in a second. | ||
Oh, Jesus. | ||
And so if you're a Komodo dragon, you better have your entire survival strategy baked into your DNA because nobody's teaching you anything. | ||
And so for us, we have this sense that it's like parenting is really important. | ||
It is. | ||
Environment is really important. | ||
It is. | ||
But so much of who and what we are is baked into our genetics. | ||
And I think that's going to be this challenge. | ||
We're going to see ourselves as increasingly genetic beings. | ||
We can't become genetic determinists, think that we're just genetics, but we're going to know a lot more. | ||
We're going to demystify a lot of what it means to be a human. | ||
Poof. | ||
Yeah. | ||
Poof is right. | ||
But are we going to lose the romance and just the randomness of life because that's what people are concerned with, right? | ||
Right? | ||
Yeah. | ||
Like, if we have some sort of genetic uniformity, especially in particular with things like intelligence and athletic performance. | ||
We're not going to appreciate freaks as much. | ||
Or maybe we'll all want to be freaks because the freaks are the ones who push us. | ||
You're not going to want to be a moron. | ||
Well, your question, it's the essential question. | ||
It's like, what makes a human? | ||
A human isn't just someone with a higher IQ. That doesn't make you a better human. | ||
That makes you someone with a higher IQ. But how are we going to think about constructing societies when it's up to us? | ||
Like, if we are going to say we value certain people, certain ideas, I think we're going to need artists. | ||
Like right now, people like artists are sometimes in the mainstream. | ||
Sometimes they're on the fringe. | ||
But artists are going to be maybe the most important people in this new world. | ||
And like right now in hospitals, we have kind of a hierarchy. | ||
And like the most technical people are the people who are valued the most. | ||
And the least technical people, like some of the nurses or nurses' aides, are the people who are often valued and paid the least. | ||
But when technology can do these technological feats, what's going to be left is how can we be great humans? | ||
How can we emote? | ||
How can we connect? | ||
How can we create art? | ||
And if we get swept away by this tide of science, and you know how excited I am about the science, we could really undermine our humanity. | ||
unidentified
|
Right. | |
And as for humans, what humans value is many aspects of that humanity. | ||
The art, the creations. | ||
Yeah, literature, things. | ||
When you read someone's great prose, you're reading like an insight into their mind, and that's what's interesting about it, right? | ||
You're like, you're not gonna get that from just ones and zeros. - Yeah, and there will always be this, we call it mystery. | ||
And even if we can do a genetic analysis of Shakespeare and Mozart and whatever, like it's still miraculous. | ||
And we need to celebrate that. | ||
And we can't allow us to say that we are just our genetics or even just our biology. | ||
But we also can't just say biology has nothing to do with it. | ||
And especially because we're going to know more about our biology and about our differences. | ||
And that's normal. | ||
I mean, it used to be in the old days that everyone thought, well, God is weather. | ||
And now we understand weather pretty much, and nobody's saying, oh, that lightning, that's God delivering a message. | ||
It could be. | ||
But we still have that mystery. | ||
And I think that in some ways it's about our orientation. | ||
Like, how do we make sure that we keep this view of life, that we have artists and humanists who are just at the core of this conversation about where we're going? | ||
What if that mystery ultimately turns out to just be ignorance, and that as you develop more and more understanding, there's less and less mystery? | ||
Would we like to be less smart? | ||
Would we like to be more overwhelmed by possibility? | ||
It could be. | ||
Could that be part of what romance is? | ||
It could be. | ||
And certainly, like the unknown, we wake up every morning. | ||
Sure. | ||
And we just don't know the answer. | ||
And there are some people, going back to the issues of life extension, there are some people who say, well, that death is essential for appreciating life. | ||
I talk about this stuff all around. | ||
And then there are people who say, you're talking about eliminating these terrible diseases, but I know somebody who had that terrible disease. | ||
And their suffering was a gift to everybody else because we all had more humanity in response to their suffering. | ||
I was like, well, that's kind of screwed up. | ||
I'd prefer them to not have that suffering. | ||
Those people are thinking wacky. | ||
It's wacky. | ||
But we need to – I totally agree with you that if we allow ourselves to get swept away with this kind of scientific determinism, if we don't say we really value – Our humanistic traditions, our artists, our cultures, we could get lost. | ||
We could become obsolete. | ||
We could become obsolete, but we could also just become less human, and there's something wonderful, and there's magic in humans. | ||
Do you think that monkeys used to think, man, we can't become a human, we become less monkey? | ||
Yeah. | ||
Do you know what I'm saying? | ||
But no being looks in the mirror and recognizes that they are evolving. | ||
We've only been homo sapiens for about 300,000 years. | ||
So we just, it's hard. | ||
We know where we've come from because you see all those little charts from high school biology. | ||
But it's really hard for people to imagine being something else in the future. | ||
It's outside of our consciousness. | ||
It's H.G. Wells squared. | ||
Yeah, and we are monkeys. | ||
It's just that we've redefined our monkey thing. | ||
We do it in a little different way. | ||
That's what I was kind of getting at. | ||
Are you concerned at all with artificial life? | ||
Are you concerned about the propagation of artificial intelligence? | ||
Well, there are different kinds of artificial life. | ||
So, one is artificial intelligence, and I know people like Elon Musk and late Stephen Hawking are afraid. | ||
Terrified. | ||
Yeah. | ||
And I think that we need, whether it's right or not, I think it's great for us to focus on those risks. | ||
Because if we just say, oh, that's crazy, and we don't focus on it, it increases the likelihood of these bad things happening. | ||
So, kudos to Elon Musk. | ||
But I also think that we're a long way away from that threat. | ||
And we will be enormous beneficiaries of these technologies. | ||
And that's why, I don't want to sound like a broken record, but that's why I keep saying it's all about values. | ||
I think we should take those threats very seriously. | ||
Values are so abstract and we don't agree on them. | ||
It's true, but like Elon Musk, I mean, they've set up this institute where to say, well, what are the dangers? | ||
And then what are the things that we can do now? | ||
What are standards that we can integrate, for example, into our computer programming? | ||
And so I mentioned my World Health Organization committee. | ||
The question is, well, what are the standards that we can integrate into scientific culture that's not going to cure everything, but may increase the likelihood we'll have a better rather than worse outcome? | ||
unidentified
|
Right. | |
But isn't there an inherent danger in other companies or other countries rather not complying with any standards that we set because it would be anti-competitive? | ||
Yes. | ||
Like that would somehow or another diminish competition or diminish their competitive edge. | ||
It's true. | ||
And that's the balance that we're going to need to hold. | ||
And it's really hard. | ||
But we have a window of opportunity now. | ||
Now, to try to get ahead of that. | ||
And like I said, we have chemical weapons, biological weapons, nuclear weapons, where we've had international standards that have roughly held. | ||
I mean, there was a time when slavery was the norm, and there was a movement to say, this is wrong, and it was largely successful. | ||
So we have history of being more successful rather than less. | ||
And I think that's the goal. | ||
But you're right. | ||
I mean, this is a race between the technology and the best values. | ||
My real concern about artificial intelligence is that this paradigm shifting moment will happen before we recognize it's happening, and then it'll be too late. | ||
Yes, that's exactly right. | ||
And that's, like I was saying, that's why I've written the book, that's why I'm out on the road so much talking to people, why it's such an honor for me to be, and pleasure for me to be here with you talking about it, because we have to reach out to people, and people can't be afraid Of entering this conversation because it feels too technical or it feels like it's somebody else's business. | ||
This is all of our business because this is all of our lives and it's all of our futures. | ||
So if in the future you think 20 years the thing that's going to really change the most is predictive genetics and to be able to predict accurately a person's health, what do you think- Health and life. | ||
The biggest detriment for all this stuff and the thing that we have to avoid the most. | ||
Yeah. | ||
So one is, as I mentioned, this determinism. | ||
Just because if we just kind of take our sense of wonder about what it means to be a human away, like that's really going to harm us. | ||
We talked about equity and access to these technologies. | ||
And the technologies don't even need to be real. | ||
In order to have a negative impact. | ||
So in India, there are no significant genetic differences between people in different castes, but the caste system has been maintained for thousands of years because people just have accepted these differences. | ||
So it's a whole new way of understanding what is a human. | ||
And it's really going to be complicated. | ||
And we aren't ready for it. | ||
We aren't ready for it culturally. | ||
We aren't ready for it educationally. | ||
Certainly our political leaders aren't paying much of any attention to all of this. | ||
We have a huge job. | ||
unidentified
|
Oof. | |
Oof. | ||
So when you sit down and you give this speech to Congress— Yeah. | ||
What are you anticipating from them in terms of like, do you think that there's anything that they can do now to take certain steps? | ||
unidentified
|
Yes. | |
So a few things. | ||
One is we need to have a national education campaign. | ||
I mean, this is so important. | ||
I would say it's on the future of genetics revolution and of AI, because I think it's crazy. | ||
I learned French in grade school and high school, and I'm happy to speak French. | ||
But I would rather have people say, this is really important stuff. | ||
So that's number one. | ||
Number two is we need to make sure that we have a functioning regulatory system. | ||
In this country, in every country, and I do a lot of comparative work. | ||
And like the United Kingdom, they're really well organized. | ||
They have a national healthcare system, which allows them at a national level to kind of think about long-term care and trade-offs. | ||
In this country, the average person changes health plans every 18 months. | ||
And I was talking with somebody the other night, and they were working on a predictive health company. | ||
And And they said their first idea was they were going to sell this information to health insurers because like, wouldn't this be great if you're a health insurer and you had somebody who was your client and you could say, hey, here's some information. | ||
You can live healthier and you're not going to have this disease 20 years from now. | ||
And what he found out is the health insurers, they could have cared less because people were just, they were only going to be part of it for a year and a half. | ||
So we really need to think differently about how do we invest in people over the course of their lives. | ||
Certainly, education is one, but thinking long-term about health and well-being is another. | ||
What do you think is going to be the first technological innovation in terms of what's already on the pipeline right now that's going to radically alter human beings? | ||
So radically, I think it's going to be the end of procreative sex. | ||
And so when we stop conceiving our babies through sex, and we're selecting our embryos, that's going to open up this massive realm of possibility. | ||
And certainly, when we expand the number of fertilized eggs that we're choosing from, that is really, I think that's kind of the killer application of genetics to the future of human life. | ||
Do you see that being attainable to the general population anytime in the near future? | ||
Once the technology gets established, it seems like it's going to be wealthier people that are going to have access to it first, right? | ||
Well, it depends. | ||
Probably, yes. | ||
But when you think about right now, we have all these people who are born with these terrible, in many cases, deadly genetic diseases and disorders. | ||
And what is the societal expenditure for lifetime care for all those people? | ||
I mean, this is huge, huge amounts of money. | ||
So if we were to eliminate many of, not the people, but prevent those diseases and disorders from taking place in the first place, And we could use that money to provide IVF and embryo screening to everybody using just the economic models now. | ||
But then there's another issue that we have to talk about. | ||
It's really sensitive. | ||
So I talk a lot about this. | ||
And I talk in the book about people with Down syndrome. | ||
I have a lot of friends who have kids with Down syndrome. | ||
These are wonderful kids and they deserve every opportunity to thrive the same as everybody else. | ||
And I'm really sensitive because people say, well, hey, if you're... | ||
And I say, like, Down syndrome is largely not going... | ||
There aren't going to be newborns with Down syndrome 10 or 20 years from now. | ||
So people say, well, what are you saying about my kids? | ||
Like, are you saying that if this is going to be eliminated, that my kid has less of a right to be as somebody else? | ||
And I always say... | ||
Absolutely not. | ||
And we need to be extremely sensitive that we're not dehumanizing people. | ||
But if you have 15 or 15 fertilized eggs in a lab and you have to pick which one gets implanted in the mother, and one of them just has a disease like Tay-Sachs or sickle cell disease where they're going to die before they're 10 years old. | ||
Would you choose, affirmatively choose, to implant that embryo versus the 9 or 14 or whatever the number is of other ones? | ||
And so these are really sensitive things, and we can't be blasé about them. | ||
But we will have these choices, and we're going to have to figure out how do we make them. | ||
Well, I think there's also a real possibility of them being able to fix that. | ||
Yeah, and some things will be fixable, and some things won't. | ||
And so that's why, though, for these single gene mutation disorders, there's a debate. | ||
I was speaking in Berkeley the other day, and so I was talking about these two options. | ||
One is embryo selection, and one is gene editing. | ||
And so... | ||
There were different people who got up and said, oh no, we can do embryo selection. | ||
That's the way that we should prevent these diseases. | ||
But gene editing, that's going too far. | ||
That's playing God. | ||
And so for different things, there'll be different options. | ||
When you hear that, that that's going too far, who's usually saying that? | ||
There's two groups. | ||
I mean, one is certainly in the religious community. | ||
We're saying, well, this is playing God. | ||
But there's another kind of – it's like a progressive community who are the kinds of people who are uncomfortable with genetically modified crops, people who are saying that – There's this slippery slope that once we start making what are called germline genetic modifications, so germline is that are sperm, eggs, and embryos. | ||
If you make a change to an adult human, it doesn't pass to their kids. | ||
If you make a change to a sperm, an egg, or an embryo, it will pass on. | ||
And so there are a lot of people who are saying, well, we don't understand genetics well enough to make these changes that will last forever. | ||
I'm not in that view. | ||
I just think that we need to be cautious and we need to weigh the risks and the benefits of everything that we do. | ||
Do you think we do know enough about those changes? | ||
It depends because if we're – it depends on what we're selecting against. | ||
Like if the thing we're selecting against is some kind of terrible genetic disease that's going to kill somebody when they're a little kid, we have a lot of latitude because the alternative is death. | ||
And that's why I was so critical of this Chinese biophysicist who created – who genetically engineered these two little girls born in China last year because he wasn't in the gene edits that weren't – Probably weren't successful. | ||
It wasn't to eliminate some disease or disorder. | ||
He was trying to confer the benefit of increased resistance to HIV. And so I think that we need to be very mindful and we need to be doing kind of a cost-benefit analysis of the different interventions. | ||
And there was an unintended side effect of this, they believe, a perceived potential unintended side effect, and that's increased intelligence. | ||
Well, it's a possibility. | ||
unidentified
|
Possibility. | |
How does that work? | ||
So this gene, it's called a CCR5, is this gene, and when it was disrupted in some mouse studies, those mice became a little bit able to navigate mazes. | ||
And so that was what led people to believe that this disruption of the CCR5 could potentially lead to that kind of change in human. | ||
Nobody really knows. | ||
There's lots of things that happen in mice that don't have analogs in humans. | ||
And that was why it was so irresponsible, is that this scientist in secret made these gene edits. | ||
He didn't get a proper consent from the parents. | ||
Oh, really? | ||
Yeah, yes. | ||
Because it's China. | ||
Because it's China. | ||
He just let it ride. | ||
unidentified
|
Yeah. | |
I mean, the parents were all manipulated. | ||
Really? | ||
Yeah. | ||
And so that's the thing. | ||
So you're exactly right. | ||
Like, we are humans. | ||
We're nuts as a species. | ||
And so we need to try to establish some kind of guide rails, guard rails, about what's okay, what we're comfortable with, what we're not. | ||
Now, this guy is not operating in an isolated incidence. | ||
There's got to be a shit ton of that going on right now as we're talking in China. | ||
What do you think is happening over there? | ||
You know, I think China has a lot of money, they have brilliant people, and they have a government that is hell-bent on leading the world in advanced technology. | ||
And the scientific culture in China is just very different than it is here. | ||
And so we know what we know, but we don't know what we don't know. | ||
And it's a really, really big deal, because China is in many ways a Wild West. | ||
And the technology exists to do some really big stuff. | ||
And that's why we have to at least try to establish standards. | ||
Will we succeed fully? | ||
No. | ||
But maybe we can do better than worse. | ||
Are you anticipating seeing a lot of freaky things come out of China? | ||
Yes. | ||
Whoa, you said that very quick. | ||
Yeah, it's true. | ||
I spent a lot of time in China. | ||
It's a different thing with China. | ||
China has this great ancient civilization. | ||
But they destroyed their own civilization in the Great Leap Forward and the Cultural Revolution. | ||
They burned their books. | ||
They smashed their own historic relics. | ||
And so it's really, it's a society in many ways that's starting from scratch. | ||
And so all of these norms that people get, inherit through their traditions, China in many ways doesn't have. | ||
And so it's very different. | ||
And China is growing. | ||
I mean, they are increasingly powerful. | ||
And China is going to be a major force defining the world of the 21st century. | ||
That's why America has to get its act together. | ||
That's a hard concept for us to grasp when we think about the fact that they had the Great Wall, they have so much ancient art and architecture. | ||
We just assume they're a really old culture. | ||
They are, but they wiped it out. | ||
That's so crazy. | ||
That's a unique perspective. | ||
That's why if you want to see great Chinese art, you have to go to Taiwan. | ||
Because when the Chinese nationalists left in 1949, as they were losing the Civil War, they took the treasures and they put them in the National Museum of Taiwan. | ||
In the Cultural Revolution, the Great Leap Forward, China, the Red Guards were just smashing all of their own stuff, their own ancient history. | ||
And now the Chinese Communist Party is saying, oh no, we're going back and we have this great 5,000-year-old culture. | ||
In some ways it's true, but in some ways it's like an adolescent culture without these kinds of restrictions that other societies have. | ||
That's such a unique perspective that I haven't heard before. | ||
It makes so much sense in terms of like how frantic they are at restructuring their world. | ||
Yeah. | ||
And they feel that they got screwed over because there is this vague sense of Chinese greatness when you hear the word Middle Kingdom. | ||
It's like China's the center of the world and everybody else is some kind of tributary. | ||
And so they're monumentally pissed off. | ||
That these colonial powers came and overpowered them and they had to make all these concessions. | ||
They had to give land away. | ||
And hell-bent on regaining it. | ||
They're playing the long game. | ||
They are playing the long game. | ||
And we're not. | ||
And we are not. | ||
And we have to be mindful of it. | ||
That's also something you can do if you have complete control of your population. | ||
You don't have to worry about people's opinions or you can just go in the direction that you feel is going to benefit the Chinese power. | ||
unidentified
|
Yeah. | |
The power that be. | ||
unidentified
|
Yeah. | |
This is a country run by engineers. | ||
We're a country run largely by lawyers and reality TV people, I guess. | ||
But in China, it's run by engineers. | ||
So there are all these problems, and the answer is always engineering. | ||
So if you have a population problem, the answer is the one-child policy. | ||
Environmental problem, you have Three Gorges Dam. | ||
You don't have water in the north of China. | ||
You build this massive, biggest water project in the world from south to north. | ||
You want to... | ||
When in the Olympics, you engineer your population, you take kids away from their families and put them in their Olympic sports school. | ||
So I write about this in Genesis Code. | ||
If you're China and you kind of have this Plato's Republic model of the world and we're going to kind of identify the genetic or maybe manipulate these genetic superstars to be our greatest scientists and mathematicians and business leaders and political leaders, like there's a model that you can imagine for how to do it. | ||
Wow. | ||
It makes you really nervous. | ||
It should. | ||
unidentified
|
Yes. | |
That's the thing. | ||
That's why, like, I just feel like with this country, we don't have time to have all these distractions. | ||
We're focusing on junk. | ||
Like what? | ||
Just like all of this, you know, I'm on CNN all the time when I'm home in New York. | ||
And I always say, like, you guys, and I'm talking about kind of geopolitical issues, China and North Korea. | ||
What I always say is, like, you guys recognize this is porn. | ||
Like, CNN and MSCB, that's like one kind of porn. | ||
And Fox and whoever else, Infowars, that's another kind of porn. | ||
But it's all porn. | ||
And we're drawing people's attention to these few stories. | ||
But there's these big stories that we have to focus on. | ||
And certainly, the rise of China is such an essential story for the 21st century because China is competing in all of these technologies. | ||
And China, it's like, go, go, go. | ||
I mean, people in China who were involved in the tech world... | ||
When they go and visit Silicon Valley, uniformly they say, we cannot believe these people are so lazy. | ||
Like, why are they not working 24 hours a day? | ||
Why are they not issuing new products every week? | ||
And so this is, I mean, they are racing. | ||
And it's going to have huge implications for the world. | ||
And so if we believe in our values, as I believe we should, we have to fight for them. | ||
And the place that we have to fight for them first is here. | ||
And we can't, you know, it's just like every day that we're just focusing on this drama, this reality TV drama of our government is another day where we're not focusing on the big things. | ||
How are we going to get our act together? | ||
How are we going to lead the world in technology? | ||
Another example is immigration. | ||
We have this whole fight of how do we keep people out. | ||
What I'd like to do is to go to the State Department and say, all right, every embassy in the world, you have a new job. | ||
We're going to give you whatever number, 500 slots per year. | ||
You have to, in your country, find the 500 most brilliant, talented, creative, entrepreneurial people and say, we're giving you a green card. | ||
We're going to give you a little starter money. | ||
We want you to move to the United States and just start a life and have kids. | ||
And we should be creaming the crop, skimming the cream. | ||
Over the rest of the world. | ||
Like, we could take over, we could revitalize this country, but we're having this fight of how do we keep a small number of refugees out? | ||
And it's just, we're not focusing on the right things. | ||
That's, again, another very, very interesting perspective. | ||
We learned about Huawei in this country, really. | ||
Well, I learned about it because they put out some pretty innovative phones and some interesting technology. | ||
But we learned it because the State Department was telling people to stop using their phones. | ||
Do you think that that is trying to stifle the competition? | ||
Because the market share that they have, if they do really have the number two selling cell phones in the world now, that's not from America. | ||
America's largely out of that conversation. | ||
And if they were in America, they would probably dominate in America as well. | ||
Because they're cheaper. | ||
Yeah, and they're really good. | ||
I mean, their phones are insane. | ||
The cameras in their phones are off the charts. | ||
They put some video of the zoom capability of their newest phone, and people were calling bullshit. | ||
There's no way. | ||
That's not even possible. | ||
But it turned out it was true. | ||
It really can zoom like a super expensive telephoto lens. | ||
Sure. | ||
Yeah, yeah. | ||
So, Huawei, it's a complicated story. | ||
For sure, the founder of Huawei is a former Chinese military officer. | ||
For sure, in the early stages of their company, they stole, straight out stole, lots of source code from companies like Cisco. | ||
For sure, we should be really worried if Huawei is the sole supplier of the infrastructure that supports 5G all around the world because the Chinese government would have access to everything. | ||
And so that leads us to the question is, one... | ||
Is there a problem with Huawei itself? | ||
But then two is, let's just say, and I think the answer to that first question is probably yes. | ||
But then the question two is, let's just say Huawei is a legit company and they're not totally intimately connected to the Chinese government. | ||
Can we trust their relationship with the Chinese government? | ||
And the Chinese government has a rule that every one of these companies has the big Chinese national companies, national champion companies, They have a Communist Party cell inside of that company. | ||
I think that we can't think of big Chinese companies just like we think of companies here. | ||
We have to think of them as quasi-state actors. | ||
That's why this fight that's happening right now is so important. | ||
That's why when China is out investing in different parts of the world, including Africa, their companies are kind of acting like arms of the government. | ||
They're making all kinds of Investments that don't really make sense if you just see, well, this is a company doing something. | ||
If you say that this is a company with backing by the state that's fulfilling a function that supports the state, it's a very different model. | ||
So I am actually quite concerned about Huawei, and I'm not a fan of everything that this administration is doing, but I think on China, it's important that we need to stand up, and I think pushing back on Huawei is the right thing to do. | ||
I'm uncomfortable about this for two reasons. | ||
One, I'm uncomfortable about that, about the Chinese government being inexorably connected to this global superpower in technology. | ||
But I'm also uncomfortable that it sets a precedent for other nations to follow. | ||
Because they're like, look, this is the only way to compete. | ||
Because what you were talking about, the investments that Huawei or that the Chinese government makes in these other countries and that don't seem to make sense if you're just dealing with a company. | ||
Right, right. | ||
But if you're dealing with someone who's trying to take over the world, it makes a lot of sense. | ||
Yeah, and so when we have our companies, you're out in some place in Africa and you're competing with a Chinese company to do something, build a port or whatever. | ||
And you're competing because you are an American company. | ||
And so you have your cat, this is the port, what's the income stream going to be about it? | ||
And you have a certain amount that you can bid Because otherwise it becomes a bad investment. | ||
But if the Chinese company, their calculus is not, is this a good or bad investment? | ||
It's what is the state interest in controlling or quasi-controlling this asset? | ||
And so that's why we can't project ourselves onto the Chinese. | ||
We can't say they're just like us, just different. | ||
We have different models and our models are competing. | ||
Do you think that we should avoid Huawei products like consumers should? | ||
Well, I think the government should very tightly regulate products like Huawei products. | ||
Because some other network, like routers, they've shown that they're using them to extract information. | ||
And so we have a long history of European, Japanese, South Korean companies that have invested very well. | ||
They've out-competed us, and we've allowed the Japanese companies to out-compete our auto manufacturers, and that was fine. | ||
In the 1970s, our cars had become shit because we had this monopoly. | ||
And so I'm all for open competition. | ||
I'm all for free trade. | ||
It has to be fair. | ||
But I think that what China is doing, China recognized as a state that they could use the tools of capitalism to achieve state ends. | ||
And I think we need to be very cautious about that. | ||
That's interesting that you compare it to the automotive market because the consequences are so much different, right? | ||
So much different. | ||
But we do have a model to go on. | ||
We could see what happened. | ||
We made shitty cars. | ||
The Japanese took over. | ||
Yeah. | ||
And then we made better cars. | ||
I have a rental car here in Los Angeles, and I went to the rental car place at LAX, and they had all of the different cars. | ||
And there was like a Nissan and a Toyota, and there was a Cadillac. | ||
And I thought, you know, I said, I'm going to go with the Caddy. | ||
So it's a great car. | ||
Oh, they're amazing. | ||
They're incredible. | ||
Yeah, American cars are very good now. | ||
They're great! | ||
And so I'm all for competition, but I just feel like what some Chinese companies are doing, it's not competition. | ||
They have become, not all of them, but quasi-state actors. | ||
And if that's what they're doing, I think we need to respond to them in that way. | ||
Interesting. | ||
What else should we be concerned with? | ||
Should we be concerned with anything that North Korea is doing? | ||
Oh, absolutely. | ||
So I have spent a lot of time in North Korea. | ||
Yeah, that's why. | ||
Yeah, so I've advised the North Korean government on the establishment of special economic zones, which I certainly believe if North Korea could have economic growth and integrate into the rest of the world, that would be great. | ||
When was this that you went over there? | ||
This was in 2015, but I've been there twice, crossed the border from China and zigzagged the country by land, visited 10 or 12 different sites, so spent almost two weeks by land. | ||
What was that like? | ||
Incredible. | ||
I mean, North Korea, one, it's the most organized place I've ever seen. | ||
I mean, there's not anywhere. | ||
There's like on the side of the road, the stones are all raked. | ||
There's not a stick. | ||
Every little line is drawn. | ||
It's like total control. | ||
In the agricultural areas, there were very few machines and very few farm animals. | ||
So I saw people pulling plows. | ||
Like, you know, you usually have the animal in front of the plow and the person behind here. | ||
There were like two people in front of the plow and one person behind. | ||
The people were the animals. | ||
And we would go and visit these – just because they didn't – a lot of the animals got eaten when they had their famine. | ||
And so we visited these different sites for these special economic zones. | ||
And they would say like what they had done and what they were thinking about doing. | ||
And I would say like, how do you – do you know anything about the market? | ||
Like what are you going to sell here? | ||
And they said, well, we know about clearing land and building a fence. | ||
And then we went to Pyongyang and I spoke to about 400 economic planners and I said, look, I know you have these plans to do these special economic zones. | ||
It's totally going to fail. | ||
The way it's going to work, you have to connect to the market economy. | ||
You have to empower your workers. | ||
You need information flow. | ||
How else are you going to learn and adapt? | ||
So North Korea, it's a really dangerous place. | ||
And now it's even more dangerous because President Trump, it was this kind of nonsensical Hail Mary in these meetings with Kim Jong-un. | ||
There was never any indication that the North Koreans were planning on giving up their nuclear weapons. | ||
They never said they would. | ||
It's the last thing they would do because their goal is survival. | ||
And so there was this kind of head fake, which was like a PR stunt to be able to say, all right, we're having these meetings. | ||
And of course, the North Koreans weren't ever going to give up their nuclear weapons. | ||
They're still not. | ||
So now things are ramping up. | ||
So North Korea in the last couple of days has started firing missiles again. | ||
The United States today, the US military seized a North Korean ship. | ||
So we're going back to this very dangerous And so I think we really need to do a much better job. | ||
We need much more. | ||
North Korea, it's really hard. | ||
And these guys are really smart. | ||
People say, well, these guys are poor. | ||
They must not be smart. | ||
We're playing cards with them. | ||
We've got the whole deck. | ||
They don't have one card. | ||
And yet, they're... | ||
unidentified
|
They're in the game. | |
They're in the game. | ||
They're holding us to a stalemate and it's really worrying. | ||
And why did you go over there? | ||
What were you thinking? | ||
So I thought a lot about it because I have a background in human rights. | ||
I was a human rights officer for the United Nations in Cambodia. | ||
I'm the child of a refugee. | ||
I have this very strong belief in human rights and in supporting people. | ||
In North Korea, they have about 120,000 people in the most brutal, horrific prison camps in And so when I was asked to be part of this six-person delegation advising them on the establishment of special economic zones, one instinct was, screw them, I don't want to be part of this at all. | ||
But I also felt that if North Korea could have some kind of integrated economic development, That would at least connect them to the world that would create some kind of leverage and that would help people. | ||
So I decided to go. | ||
And I'm glad that I did, but these are really hard issues. | ||
And it's very unfortunate that in President Trump's negotiations with the North Koreans, human rights was never once mentioned. | ||
And I think that that's coming back to values. | ||
We have to be clear about who we are and what we stand for and be consistent in fighting for it. | ||
Do you think that Trump didn't bring that up because he wanted to be able to effectively communicate with them and not put them on their heels? | ||
Maybe, but I feel like had they done – I mean, I think that if he thought – That there was a real chance of progress. | ||
But the hard thing was he didn't know much about the North Koreans. | ||
He has people. | ||
We have brilliant people working in the United States government. | ||
And all of those people, all of the U.S. intelligence agencies were telling President Trump that the North Koreans have absolutely no intention of giving up their nuclear weapons. | ||
And so maybe he did think that he would charm Kim Jong-un or he would say, hey, we're going to give you economic development or whatever. | ||
I think for most people who were observers of North Korea who had watched it for a while, thought that was not – so we gave away a lot. | ||
So we didn't mention human rights. | ||
We suspended our military exercises. | ||
We gave them the legitimacy of a presidential meeting, which they've been wanting for 30 years, and we didn't get anything back. | ||
So had we gotten something back, then you could say, well, that was a risk we're taking maybe. | ||
Yeah, I haven't heard described that way, but I'm agreeing with what you're saying. | ||
What do you think he could have done differently, though? | ||
I don't think the meeting should have happened with no conditions. | ||
So if he had said, I'm open to meet with the North Koreans, Which is something the North Koreans have always wanted. | ||
We could have met with the North Koreans anytime immediately for the last 30 years. | ||
But in order to do it, they need to do this, this, and this. | ||
And if they do it, we'll meet. | ||
Like that would have been a legitimate thing. | ||
But what he said is somebody, the North Korean, I mean, sorry, the South Korean national security advisors peeked into his office and he goes, hey, they want to meet. | ||
And it was like, sure. | ||
That seems like an interesting thing to do. | ||
And I think that with this diplomacy, you kind of have to get something. | ||
And so we gave away so much up front, and the North Koreans didn't have an incentive to do anything in return. | ||
Was his perspective that it would be better to be in communication and to be friends with this guy? | ||
Was that what he was thinking? | ||
But we have real interests in the sense that we have large military forces in Seoul. | ||
We have a lot at stake. | ||
We have our closest ally, Japan, who's had citizens abducted. | ||
And so I think that was what he thought is like, let's be friendly. | ||
And then with the force of personal chemistry, everything will unlock. | ||
But I think that was always extremely unlikely. | ||
What do you think is going to happen to that country? | ||
I think eventually, and I've written this, I think eventually this regime will collapse under its own weight, but it's really held out a long time because you think of the collapse of the Soviet Union. | ||
The Soviet Union had enough bullets to survive. | ||
If they had said, you know, we're just going to shoot everybody at the Berlin Wall and every dissenter, they would still be – North Korea has essentially murdered millions of people. | ||
So with famine and execution and prison camps – So I think they're going to stay for a while, but eventually there will be leaders in North Korea who will come to the conclusion that it's safer to oppose the Kim family than to wait for the Kim family to come and get you. | ||
And that tends to happen in these kind of totalitarian systems where there's so little trust, there's so little loyalty. | ||
unidentified
|
Jesus. | |
Yeah. | ||
What are their conditions like technologically? | ||
What is their infrastructure like? | ||
So the general infrastructure is absolutely terrible. | ||
I mean, they have roads in the big cities that are actually quite nice roads because there's no cars. | ||
But their infrastructure is terrible. | ||
I mean, all of their power supply, they have brownouts, blackouts all the time. | ||
Their manufacturing is all being decimated. | ||
So it's terrible. | ||
But they have really focused their energy on building these nuclear weapons because they think that these nuclear weapons give them leverage to do things and to extract concessions and to get – it's terrible infrastructure. | ||
So do – They don't have an internet, right? | ||
But they have something similar, but it only allows them access to a few state-run websites? | ||
Well, average person doesn't have access to the internet. | ||
So the way it works is it's all about loyalty. | ||
So you need three or so generations of loyalty to the Kim family to even set foot in Pyongyang, the capital. | ||
Really? | ||
Yes. | ||
So it's not like you can kind of move around or whatever. | ||
It's like just to be in the capital, like you have to have your loyalty proven. | ||
And so average person out in the country, they don't have access to much of anything. | ||
They have a little bit more now than they did in the past. | ||
And then for this relatively small number of elites who are largely in Pyongyang and in the other cities, where there's a ring of defense around these cities, and just to enter, you have to have all of these checks. | ||
Some of them have access to limited internet, but it's tightly controlled, and it's not like you're going on Google and going wherever you want. | ||
Right. | ||
And they probably would get in trouble if they Googled the wrong thing. | ||
unidentified
|
Yes. | |
And trouble, it's not just you trouble. | ||
Like, if my brother or my uncle does something that gets me in trouble with the regime, the whole extended family is out. | ||
And that means either you go to prison camps or you're kicked out of Pyongyang. | ||
I mean, it's all about collective punishment. | ||
People are terrified. | ||
And by that ruthless punishment structure that they've set up, that's how they've kept control of the country. | ||
unidentified
|
Yeah, yes. | |
And everybody's forced to rat on each other, right? | ||
Yeah, absolutely. | ||
They're actually compelled to tell on each other for one thing that you did. | ||
If you don't, then what? | ||
If you don't, then you are complicit. | ||
And that's these horrible stories. | ||
I've met a lot of these people who were in the prison camps. | ||
I have a friend of mine. | ||
She was this 13-year-old girl. | ||
And her father was a low-level North Korean official, and then he was accused of something, and so this family that was privileged all of a sudden was out, and just these horrible things, and prison, and rape, and this little... | ||
I mean, now she's in the United States and incredibly positive. | ||
I mean, it's amazing how resilient she is. | ||
But this is like a real hell. | ||
It's an issue, and I think that for us as Americans, as humans, We're less human when there are people who are suffering like this. | ||
Yeah, I agree. | ||
Now, you were traveling all over North Korea. | ||
What were they having you do while you were out there? | ||
So what we would do is we would go from one of these special economic zones to the other. | ||
And in each one, it was kind of the same story. | ||
You'd get there. | ||
There'd be like a big field. | ||
The farmers had been kicked off. | ||
There was a fence around it. | ||
And then the group of the local officials would come and they'd have like a big chart And they have a plan, like here's where we're going to build this building. | ||
And I would always ask the same question, like, what are you going to do here? | ||
Why do you think you're going to be competitive? | ||
How do you know what the market prices are? | ||
How are your workers going to be empowered so they can change things? | ||
I mean, in the old days, it used to be you just kind of have these automaton workers. | ||
Now workers are actually making big decisions and fixing things. | ||
And they didn't have an answer to any of those questions. | ||
And that's what happens when you have these totalitarian, top-down systems, is that being creative is actually really dangerous. | ||
So if somebody says, do X, you just do X. Wow. | ||
Yeah. | ||
It's really incredible. | ||
And it's so sad because I spend a lot of time in South Korea. | ||
And this is the most dynamic place. | ||
Often I go to Seoul just to see what technology is going to show up here a few years in the future. | ||
I mean, Seoul is like the future. | ||
And then just 35 miles from Seoul is the demilitarized zone. | ||
And the other side, it's incredible. | ||
And the real problem would be once they finally did get free of that... | ||
I mean, you can call him whatever you want, a dictator, and his family. | ||
What tools would they have? | ||
Like, how prepared would they be to be autonomous? | ||
Well, it's really the good thing, the benefit that they have, if there is... | ||
So here's my thought of what a scenario might look like. | ||
I think eventually... | ||
Probably there'll be some kind of coup attempt against the Kim family. | ||
Let's just say it succeeds. | ||
But that would probably result in another military dictatorship with another group. | ||
Well, we don't know. | ||
Because then, I think, immediately, I think the Chinese would invade. | ||
Really? | ||
Yeah, because when people think of the Korean War from the early 1950s, they think, oh, it's the Korean War. | ||
It must have been... | ||
The Americans fighting against the Koreans. | ||
But the Korean War, the two sides was America and the South Koreans fighting against the North Koreans and the Chinese. | ||
The Chinese did the most of the fighting. | ||
And so China, North Korea is the only country in the world that has a treaty alliance with China, kind of like we have with Japan and with South Korea. | ||
And so China, their biggest fear is having a reunified Korean peninsula allied to the United States. | ||
So I think if there was a coup, the Chinese would immediately move in militarily. | ||
Then immediately there would be this call to have some kind of UN body. | ||
And there would be a call for a UN authority. | ||
And then I think it would be agreed that the Chinese would stay and they just would put on blue helmets like as a UN force. | ||
And then we'd have to negotiate what happens next. | ||
And I think what the Chinese would do would be say, well, we'll leave when the Americans leave. | ||
I think that would be what will likely happen. | ||
I think we're going to see a Korean reunification. | ||
And the good news of these reunified countries like East and West Germany is there's a whole system of law that is just – North Korea will be swallowed into South Korea. | ||
And then you have law. | ||
You have an infrastructure. | ||
And it will take one or two generations. | ||
But I think that will eventually happen. | ||
And I'm hoping it can happen without nuclear war or terrible bloodshed. | ||
But it's going to be a big challenge. | ||
Goddamn. | ||
Yeah. | ||
That sounds insurmountable. | ||
Just hearing you talk about that, about North Korea getting absorbed by South Korea, I'm like, oh my god, good luck. | ||
Just imagine. | ||
Just imagine the whole thing. | ||
Well, listen, man, it's been a fascinating conversation. | ||
I really appreciate it. | ||
I really appreciate you coming here and You've certainly sparked a lot of interesting ideas in my head and I'm sure a lot of other people's heads as well. | ||
And I would like to see down the line where this all goes. | ||
And I hope we don't get swallowed up by machines. | ||
We won't, but it's up to us to fight for what we believe in. | ||
Well, thank you very much, man. | ||
My great pleasure. | ||
I really appreciate it. | ||
It was a lot of fun. | ||
unidentified
|
Anytime. |