Chris Anderson: Elon Musk, great to see you. How are you?
Elon Musk: Good. How are you?
CA: We're here at the Texas Gigafactory the day before this thing opens. It's been pretty crazy out there. Thank you so much for making time on a busy day.
I would love you to help us, kind of, cast our minds, I don't know, 10, 20, 30 years into the future. And help us try to picture what it would take to build a future that's worth getting excited about.
The last time you spoke at TED, you said that that was really just a big driver. You know, you can talk about lots of other reasons to do the work you're doing, but fundamentally, you want to think about the future and not think that it sucks.
EM: Yeah, absolutely. I think in general, you know, there's a lot of discussion of like, this problem or that problem. And a lot of people are sad about the future and they're ... Pessimistic. And I think ... this is ... This is not great. I mean, we really want to wake up in the morning and look forward to the future. We want to be excited about what's going to happen. And life cannot simply be about sort of, solving one miserable problem after another.
CA: So if you look forward 30 years, you know, the year 2050 has been labeled by scientists as this, kind of, almost like this doomsday deadline on climate. There's a consensus of scientists, a large consensus of scientists, who believe that if we haven't completely eliminated greenhouse gases or offset them completely by 2050, effectively we're inviting climate catastrophe. Do you believe there is a pathway to avoid that catastrophe? And what would it look like?
EM: Yeah, so I am not one of the doomsday people, which may surprise you. I actually think we're on a good path. But at the same time, I want to caution against complacency. So, so long as we are not complacent, as long as we have a high sense of urgency about moving towards a sustainable energy economy, then I think things will be fine. So I can't emphasize that enough, as long as we push hard and are not complacent, the future is going to be great. Don't worry about it. I mean, worry about it, but if you worry about it, ironically, it will be a self-unfulfilling prophecy.
So, like, there are three elements to a sustainable energy future. One is of sustainable energy generation, which is primarily wind and solar. There's also hydro, geothermal, I'm actually pro-nuclear. I think nuclear is fine. But it's going to be primarily solar and wind, as the primary generators of energy. The second part is you need batteries to store the solar and wind energy because the sun doesn't shine all the time, the wind doesn't blow all the time. So it's a lot of stationary battery packs. And then you need electric transport. So electric cars, electric planes, boats. And then ultimately, it’s not really possible to make electric rockets, but you can make the propellant used in rockets using sustainable energy.
So ultimately, we can have a fully sustainable energy economy. And it's those three things: solar/wind, stationary battery pack, electric vehicles.
So then what are the limiting factors on progress? The limiting factor really will be battery cell production. So that's going to really be the fundamental rate driver. And then whatever the slowest element of the whole lithium-ion battery cells supply chain, from mining and the many steps of refining to ultimately creating a battery cell and putting it into a pack, that will be the limiting factor on progress towards sustainability.
CA: All right, so we need to talk more about batteries, because the key thing that I want to understand, like, there seems to be a scaling issue here that is kind of amazing and alarming. You have said that you have calculated that the amount of battery production that the world needs for sustainability is 300 terawatt hours of batteries. That's the end goal?
EM: Very rough numbers, and I certainly would invite others to check our calculations because they may arrive at different conclusions. But in order to transition, not just current electricity production, but also heating and transport, which roughly triples the amount of electricity that you need, it amounts to approximately 300 terawatt hours of installed capacity.
CA: So we need to give people a sense of how big a task that is. I mean, here we are at the Gigafactory. You know, this is one of the biggest buildings in the world. What I've read, and tell me if this is still right, is that the goal here is to eventually produce 100 gigawatt hours of batteries here a year eventually.
EM: We will probably do more than that, but yes, hopefully we get there within a couple of years.
CA: Right. But I mean, that is one --
EM: 0.1 terrawat hours.
CA: But that's still 1/100 of what's needed. How much of the rest of that 100 is Tesla planning to take on let's say, between now and 2030, 2040, when we really need to see the scale up happen?
EM: I mean, these are just guesses. So please, people shouldn't hold me to these things. It's not like this is like some -- What tends to happen is I'll make some like, you know, best guess and then people, in five years, there’ll be some jerk that writes an article: "Elon said this would happen, and it didn't happen. He's a liar and a fool." It's very annoying when that happens. So these are just guesses, this is a conversation.
CA: Right.
EM: I think Tesla probably ends up doing 10 percent of that. Roughly.
CA: Let's say 2050 we have this amazing, you know, 100 percent sustainable electric grid made up of, you know, some mixture of the sustainable energy sources you talked about. That same grid probably is offering the world really low-cost energy, isn't it, compared with now. And I'm curious about like, are people entitled to get a little bit excited about the possibilities of that world?
EM: People should be optimistic about the future. Humanity will solve sustainable energy. It will happen if we, you know, continue to push hard, the future is bright and good from an energy standpoint. And then it will be possible to also use that energy to do carbon sequestration. It takes a lot of energy to pull carbon out of the atmosphere because in putting it in the atmosphere it releases energy. So now, you know, obviously in order to pull it out, you need to use a lot of energy. But if you've got a lot of sustainable energy from wind and solar, you can actually sequester carbon. So you can reverse the CO2 parts per million of the atmosphere and oceans. And also you can really have as much fresh water as you want. Earth is mostly water. We should call Earth “Water.” It's 70 percent water by surface area. Now most of that’s seawater, but it's like we just happen to be on the bit that's land.
CA: And with energy, you can turn seawater into --
EM: Yes.
CA: Irrigating water or whatever water you need.
EM: At very low cost. Things will be good.
CA: Things will be good. And also, there's other benefits to this non-fossil fuel world where the air is cleaner --
EM: Yes, exactly. Because, like, when you burn fossil fuels, there's all these side reactions and toxic gases of various kinds. And sort of little particulates that are bad for your lungs. Like, there's all sorts of bad things that are happening that will go away. And the sky will be cleaner and quieter. The future's going to be good.
CA: I want us to switch now to think a bit about artificial intelligence. But the segue there, you mentioned how annoying it is when people call you up for bad predictions in the past. So I'm possibly going to be annoying now, but I’m curious about your timelines and how you predict and how come some things are so amazingly on the money and some aren't. So when it comes to predicting sales of Tesla vehicles, for example, you've kind of been amazing, I think in 2014 when Tesla had sold that year 60,000 cars, you said, "2020, I think we will do half a million a year."
EM: Yeah, we did almost exactly a half million.
CA: You did almost exactly half a million. You were scoffed in 2014 because no one since Henry Ford, with the Model T, had come close to that kind of growth rate for cars. You were scoffed, and you actually hit 500,000 cars and then 510,000 or whatever produced.
But five years ago, last time you came to TED, I asked you about full self-driving, and you said, “Yeah, this very year, I'm confident that we will have a car going from LA to New York without any intervention."
EM: Yeah, I don't want to blow your mind, but I'm not always right. CA: (Laughs)
What's the difference between those two? Why has full self-driving in particular been so hard to predict?
EM: I mean, the thing that really got me, and I think it's going to get a lot of other people, is that there are just so many false dawns with self-driving, where you think you've got the problem, have a handle on the problem, and then it, no, turns out you just hit a ceiling. Because if you were to plot the progress, the progress looks like a log curve. So it's like a series of log curves. So most people don't know what a log curve is, I suppose.
CA: Show the shape with your hands.
EM: It goes up you know, sort of a fairly straight way, and then it starts tailing off and you start getting diminishing returns. And you're like, uh oh, it was trending up and now it's sort of, curving over and you start getting to these, what I call local maxima, where you don't realize basically how dumb you were. And then it happens again.
And ultimately... These things, you know, in retrospect, they seem obvious, but in order to solve full self-driving properly, you actually have to solve real-world AI. Because what are the road networks designed to work with? They're designed to work with a biological neural net, our brains, and with vision, our eyes. And so in order to make it work with computers, you basically need to solve real-world AI and vision. Because we need cameras and silicon neural nets in order to have self-driving work for a system that was designed for eyes and biological neural nets.
You know, I guess when you put it that way, it's sort of, like, quite obvious that the only way to solve full self-driving is to solve real world AI and sophisticated vision.
CA: What do you feel about the current architecture? Do you think you have an architecture now where there is a chance for the logarithmic curve not to tail off any anytime soon?
EM: Well I mean, admittedly these may be infamous last words, but I actually am confident that we will solve it this year. That we will exceed -- The probability of an accident, at what point do you exceed that of the average person? I think we will exceed that this year.
CA: What are you seeing behind the scenes that gives you that confidence?
EM: We’re almost at the point where we have a high-quality unified vector space. In the beginning, we were trying to do this with image recognition on individual images. But if you get one image out of a video, it's actually quite hard to see what's going on without ambiguity. But if you look at a video segment of a few seconds of video, that ambiguity resolves. So the first thing we had to do is tie all eight cameras together so they're synchronized, so that all the frames are looked at simultaneously and labeled simultaneously by one person, because we still need human labeling. So at least they’re not labeled at different times by different people in different ways. So it's sort of a surround picture. Then a very important part is to add the time dimension. So that you’re looking at surround video, and you're labeling surround video. And this is actually quite difficult to do from a software standpoint. We had to write our own labeling tools and then create auto labeling, create auto labeling software to amplify the efficiency of human labelers because it’s quite hard to label. In the beginning, it was taking several hours to label a 10-second video clip. This is not scalable.
So basically what you have to have is you have to have surround video, and that surround video has to be primarily automatically labeled with humans just being editors and making slight corrections to the labeling of the video and then feeding back those corrections into the future auto labeler, so you get this flywheel eventually where the auto labeler is able to take in vast amounts of video and with high accuracy, automatically label the video for cars, lane lines, drive space.
CA: What you’re saying is ... the result of this is that you're effectively giving the car a 3D model of the actual objects that are all around it. It knows what they are, and it knows how fast they are moving. And the remaining task is to predict what the quirky behaviors are that, you know, that when a pedestrian is walking down the road with a smaller pedestrian, that maybe that smaller pedestrian might do something unpredictable or things like that. You have to build into it before you can really call it safe.
EM: You basically need to have memory across time and space. So what I mean by that is ... Memory can’t be infinite, because it's using up a lot of the computer's RAM basically. So you have to say how much are you going to try to remember? It's very common for things to be occluded. So if you talk about say, a pedestrian walking past a truck where you saw the pedestrian start on one side of the truck, then they're occluded by the truck. You would know intuitively, OK, that pedestrian is going to pop out the other side, most likely.
CA: A computer doesn't know it.
EM: You need to slow down.
CA: A skeptic is going to say that every year for the last five years, you've kind of said, well, no this is the year, we're confident that it will be there in a year or two or, you know, like it's always been about that far away. But we've got a new architecture now, you're seeing enough improvement behind the scenes to make you not certain, but pretty confident, that, by the end of this year, what in most, not in every city, and every circumstance but in many cities and circumstances, basically the car will be able to drive without interventions safer than a human.
EM: Yes. I mean, the car currently drives me around Austin most of the time with no interventions. So it's not like ... And we have over 100,000 people in our full self-driving beta program. So you can look at the videos that they post online.
CA: I do. And some of them are great, and some of them are a little terrifying. I mean, occasionally the car seems to veer off and scare the hell out of people.
EM: It’s still a beta. CA: But you’re behind the scenes, looking at the data, you're seeing enough improvement to believe that a this-year timeline is real.
EM: Yes, that's what it seems like. I mean, we could be here talking again in a year, like, well, another year went by, and it didn’t happen. But I think this is the year.
CA: And so in general, when people talk about Elon time, I mean it sounds like you can't just have a general rule that if you predict that something will be done in six months, actually what we should imagine is it’s going to be a year or it’s like two-x or three-x, it depends on the type of prediction. Some things, I guess, things involving software, AI, whatever, are fundamentally harder to predict than others.
Is there an element that you actually deliberately make aggressive prediction timelines to drive people to be ambitious? Without that, nothing gets done?
EM: Well, I generally believe, in terms of internal timelines, that we want to set the most aggressive timeline that we can. Because there’s sort of like a law of gaseous expansion where, for schedules, where whatever time you set, it's not going to be less than that. It's very rare that it'll be less than that. But as far as our predictions are concerned, what tends to happen in the media is that they will report all the wrong ones and ignore all the right ones. Or, you know, when writing an article about me -- I've had a long career in multiple industries. If you list my sins, I sound like the worst person on Earth. But if you put those against the things I've done right, it makes much more sense, you know? So essentially like, the longer you do anything, the more mistakes that you will make cumulatively. Which, if you sum up those mistakes, will sound like I'm the worst predictor ever. But for example, for Tesla vehicle growth, I said I think we’d do 50 percent, and we’ve done 80 percent.
CA: Yes.
EM: But they don't mention that one. So, I mean, I'm not sure what my exact track record is on predictions. They're more optimistic than pessimistic, but they're not all optimistic. Some of them are exceeded probably more or later, but they do come true. It's very rare that they do not come true. It's sort of like, you know, if there's some radical technology prediction, the point is not that it was a few years late, but that it happened at all. That's the more important part.
CA: So it feels like at some point in the last year, seeing the progress on understanding, the Tesla AI understanding the world around it, led to a kind of, an aha moment at Tesla. Because you really surprised people recently when you said probably the most important product development going on at Tesla this year is this robot, Optimus.
EM: Yes.
CA: Many companies out there have tried to put out these robots, they've been working on them for years. And so far no one has really cracked it. There's no mass adoption robot in people's homes. There are some in manufacturing, but I would say, no one's kind of, really cracked it. Is it something that happened in the development of full self-driving that gave you the confidence to say, "You know what, we could do something special here."
EM: Yeah, exactly. So, you know, it took me a while to sort of realize that in order to solve self-driving, you really needed to solve real-world AI. And at the point of which you solve real-world AI for a car, which is really a robot on four wheels, you can then generalize that to a robot on legs as well. The two hard parts I think -- like obviously companies like Boston Dynamics have shown that it's possible to make quite compelling, sometimes alarming robots.
CA: Right.
EM: You know, so from a sensors and actuators standpoint, it's certainly been demonstrated by many that it's possible to make a humanoid robot. The things that are currently missing are enough intelligence for the robot to navigate the real world and do useful things without being explicitly instructed. So the missing things are basically real-world intelligence and scaling up manufacturing. Those are two things that Tesla is very good at. And so then we basically just need to design the specialized actuators and sensors that are needed for humanoid robot. People have no idea, this is going to be bigger than the car.
CA: So let's dig into exactly that. I mean, in one way, it's actually an easier problem than full self-driving because instead of an object going along at 60 miles an hour, which if it gets it wrong, someone will die. This is an object that's engineered to only go at what, three or four or five miles an hour. And so a mistake, there aren't lives at stake. There might be embarrassment at stake.
EM: So long as the AI doesn't take it over and murder us in our sleep or something.
CA: Right.
(Laughter)
So talk about -- I think the first applications you've mentioned are probably going to be manufacturing, but eventually the vision is to have these available for people at home. If you had a robot that really understood the 3D architecture of your house and knew where every object in that house was or was supposed to be, and could recognize all those objects, I mean, that’s kind of amazing, isn’t it? Like the kind of thing that you could ask a robot to do would be what? Like, tidy up?
EM: Yeah, absolutely. Make dinner, I guess, mow the lawn.
CA: Take a cup of tea to grandma and show her family pictures.
EM: Exactly. Take care of my grandmother and make sure --
CA: It could obviously recognize everyone in the home. It could play catch with your kids.
EM: Yes. I mean, obviously, we need to be careful this doesn't become a dystopian situation. I think one of the things that's going to be important is to have a localized ROM chip on the robot that cannot be updated over the air. Where if you, for example, were to say, “Stop, stop, stop,” if anyone said that, then the robot would stop, you know, type of thing. And that's not updatable remotely. I think it's going to be important to have safety features like that.
CA: Yeah, that sounds wise.
EM: And I do think there should be a regulatory agency for AI. I've said that for many years. I don't love being regulated, but I think this is an important thing for public safety.
CA: Let's come back to that. But I don't think many people have really sort of taken seriously the notion of, you know, a robot at home. I mean, at the start of the computing revolution, Bill Gates said there's going to be a computer in every home. And people at the time said, yeah, whatever, who would even want that. Do you think there will be basically like in, say, 2050 or whatever, like a robot in most homes, is what there will be, and people will love them and count on them? You’ll have your own butler basically.
EM: Yeah, you'll have your sort of buddy robot probably, yeah.
CA: I mean, how much of a buddy? How many applications have you thought, you know, can you have a romantic partner, a sex partner? EM: It's probably inevitable. I mean, I did promise the internet that I’d make catgirls. We could make a robot catgirl.
CA: Be careful what you promise the internet.
(Laughter)
EM: So, yeah, I guess it'll be whatever people want really, you know.
CA: What sort of timeline should we be thinking about of the first models that are actually made and sold?
EM: Well, you know, the first units that we intend to make are for jobs that are dangerous, boring, repetitive, and things that people don't want to do. And, you know, I think we’ll have like an interesting prototype sometime this year. We might have something useful next year, but I think quite likely within at least two years. And then we'll see rapid growth year over year of the usefulness of the humanoid robots and decrease in cost and scaling up production.
CA: Initially just selling to businesses, or when do you picture you'll start selling them where you can buy your parents one for Christmas or something?
EM: I'd say in less than ten years.
CA: Help me on the economics of this. So what do you picture the cost of one of these being?
EM: Well, I think the cost is actually not going to be crazy high. Like less than a car. Initially, things will be expensive because it'll be a new technology at low production volume. The complexity and cost of a car is greater than that of a humanoid robot. So I would expect that it's going to be less than a car, or at least equivalent to a cheap car.
CA: So even if it starts at 50k, within a few years, it’s down to 20k or lower or whatever. And maybe for home they'll get much cheaper still. But think about the economics of this. If you can replace a $30,000, $40,000-a-year worker, which you have to pay every year, with a one-time payment of $25,000 for a robot that can work longer hours, a pretty rapid replacement of certain types of jobs. How worried should the world be about that?
EM: I wouldn't worry about the sort of, putting people out of a job thing. I think we're actually going to have, and already do have, a massive shortage of labor. So I think we will have ... Not people out of work, but actually still a shortage labor even in the future.
But this really will be a world of abundance. Any goods and services will be available to anyone who wants them. It'll be so cheap to have goods and services, it will be ridiculous.
CA: I'm presuming it should be possible to imagine a bunch of goods and services that can't profitably be made now but could be made in that world, courtesy of legions of robots.
EM: Yeah. It will be a world of abundance. The only scarcity that will exist in the future is that which we decide to create ourselves as humans.
CA: OK. So AI is allowing us to imagine a differently powered economy that will create this abundance. What are you most worried about going wrong?
EM: Well, like I said, AI and robotics will bring out what might be termed the age of abundance. Other people have used this word, and that this is my prediction: it will be an age of abundance for everyone. But I guess there’s ... The dangers would be the artificial general intelligence or digital superintelligence decouples from a collective human will and goes in the direction that for some reason we don't like. Whatever direction it might go.
You know, that’s sort of the idea behind Neuralink, is to try to more tightly couple collective human world to digital superintelligence. And also along the way solve a lot of brain injuries and spinal injuries and that kind of thing. So even if it doesn't succeed in the greater goal, I think it will succeed in the goal of alleviating brain and spine damage.
CA: So the spirit there is that if we're going to make these AIs that are so vastly intelligent, we ought to be wired directly to them so that we ourselves can have those superpowers more directly. But that doesn't seem to avoid the risk that those superpowers might ... turn ugly in unintended ways.
EM: I think it's a risk, I agree. I'm not saying that I have some certain answer to that risk. I’m just saying like maybe one of the things that would be good for ensuring that the future is one that we want is to more tightly couple the collective human world to digital intelligence.
The issue that we face here is that we are already a cyborg, if you think about it. The computers are an extension of ourselves. And when we die, we have, like, a digital ghost. You know, all of our text messages and social media, emails. And it's quite eerie actually, when someone dies but everything online is still there. But you say like, what's the limitation? What is it that inhibits a human-machine symbiosis? It's the data rate. When you communicate, especially with a phone, you're moving your thumbs very slowly. So you're like moving your two little meat sticks at a rate that’s maybe 10 bits per second, optimistically, 100 bits per second. And computers are communicating at the gigabyte level and beyond.
CA: Have you seen evidence that the technology is actually working, that you've got a richer, sort of, higher bandwidth connection, if you like, between like external electronics and a brain than has been possible before?
EM: Yeah. I mean, the fundamental principles of reading neurons, sort of doing read-write on neurons with tiny electrodes, have been demonstrated for decades. So it's not like the concept is new. The problem is that there is no product that works well that you can go and buy. So it's all sort of, in research labs. And it's like some cords sticking out of your head. And it's quite gruesome, and it's really ... There's no good product that actually does a good job and is high-bandwidth and safe and something actually that you could buy and would want to buy.
But the way to think of the Neuralink device is kind of like a Fitbit or an Apple Watch. That's where we take out sort of a small section of skull about the size of a quarter, replace that with what, in many ways really is very much like a Fitbit, Apple Watch or some kind of smart watch thing. But with tiny, tiny wires, very, very tiny wires. Wires so tiny, it’s hard to even see them. And it's very important to have very tiny wires so that when they’re implanted, they don’t damage the brain.
CA: How far are you from putting these into humans?
EM: Well, we have put in our FDA application to aspirationally do the first human implant this year.
CA: The first uses will be for neurological injuries of different kinds. But rolling the clock forward and imagining when people are actually using these for their own enhancement, let's say, and for the enhancement of the world, how clear are you in your mind as to what it will feel like to have one of these inside your head?
EM: Well, I do want to emphasize we're at an early stage. And so it really will be many years before we have anything approximating a high-bandwidth neural interface that allows for AI-human symbiosis. For many years, we will just be solving brain injuries and spinal injuries. For probably a decade. This is not something that will suddenly one day it will have this incredible sort of whole brain interface. It's going to be, like I said, at least a decade of really just solving brain injuries and spinal injuries. And really, I think you can solve a very wide range of brain injuries, including severe depression, morbid obesity, sleep, potentially schizophrenia, like, a lot of things that cause great stress to people. Restoring memory in older people.
CA: If you can pull that off, that's the app I will sign up for.
EM: Absolutely.
CA: Please hurry. (Laughs)
EM: I mean, the emails that we get at Neuralink are heartbreaking. I mean, they'll send us just tragic, you know, where someone was sort of, in the prime of life and they had an accident on a motorcycle and someone who's 25, you know, can't even feed themselves. And this is something we could fix.
CA: But you have said that AI is one of the things you're most worried about and that Neuralink may be one of the ways where we can keep abreast of it.
EM: Yeah, there's the short-term thing, which I think is helpful on an individual human level with injuries. And then the long-term thing is an attempt to address the civilizational risk of AI by bringing digital intelligence and biological intelligence closer together.
I mean, if you think of how the brain works today, there are really two layers to the brain. There's the limbic system and the cortex. You've got the kind of, animal brain where -- it’s kind of like the fun part, really.
CA: It's where most of Twitter operates, by the way.
EM: I think Tim Urban said, we’re like somebody, you know, stuck a computer on a monkey. You know, so we're like, if you gave a monkey a computer, that's our cortex. But we still have a lot of monkey instincts. Which we then try to rationalize as, no, it's not a monkey instinct. It’s something more important than that. But it's often just really a monkey instinct. We're just monkeys with a computer stuck in our brain. But even though the cortex is sort of the smart, or the intelligent part of the brain, the thinking part of the brain, I've not yet met anyone who wants to delete their limbic system or their cortex. They're quite happy having both. Everyone wants both parts of their brain. And people really want their phones and their computers, which are really the tertiary, the third part of your intelligence. It's just that it's ... Like the bandwidth, the rate of communication with that tertiary layer is slow. And it's just a very tiny straw to this tertiary layer. And we want to make that tiny straw a big highway. And I’m definitely not saying that this is going to solve everything. Or this is you know, it’s the only thing -- it’s something that might be helpful. And worst-case scenario, I think we solve some important brain injury, spinal injury issues, and that's still a great outcome.
CA: Best-case scenario, we may discover new human possibility, telepathy, you've spoken of, in a way, a connection with a loved one, you know, full memory and much faster thought processing maybe. All these things. It's very cool.
If AI were to take down Earth, we need a plan B. Let's shift our attention to space. We spoke last time at TED about reusability, and you had just demonstrated that spectacularly for the first time. Since then, you've gone on to build this monster rocket, Starship, which kind of changes the rules of the game in spectacular ways. Tell us about Starship.
EM: Starship is extremely fundamental.
So the holy grail of rocketry or space transport is full and rapid reusability. This has never been achieved. The closest that anything has come is our Falcon 9 rocket, where we are able to recover the first stage, the boost stage, which is probably about 60 percent of the cost of the vehicle of the whole launch, maybe 70 percent. And we've now done that over a hundred times. So with Starship, we will be recovering the entire thing. Or at least that's the goal.
CA: Right.
EM: And moreover, recovering it in such a way that it can be immediately re-flown. Whereas with Falcon 9, we still need to do some amount of refurbishment to the booster and to the fairing nose cone. But with Starship, the design goal is immediate re-flight. So you just refill propellants and go again. And this is gigantic. Just as it would be in any other mode of transport.
CA: And the main design is to basically take 100 plus people at a time, plus a bunch of things that they need, to Mars. So, first of all, talk about that piece. What is your latest timeline? One, for the first time, a Starship goes to Mars, presumably without people, but just equipment. Two, with people. Three, there’s sort of, OK, 100 people at a time, let's go.
EM: Sure. And just to put the cost thing into perspective, the expected cost of Starship, putting 100 tons into orbit, is significantly less than what it would have cost or what it did cost to put our tiny Falcon 1 rocket into orbit. Just as the cost of flying a 747 around the world is less than the cost of a small airplane. You know, a small airplane that was thrown away. So it's really pretty mind-boggling that the giant thing costs less, way less than the small thing. So it doesn't use exotic propellants or things that are difficult to obtain on Mars. It uses methane as fuel, and it's primarily oxygen, roughly 77-78 percent oxygen by weight. And Mars has a CO2 atmosphere and has water ice, which is CO2 plus H2O, so you can make CH4, methane, and O2, oxygen, on Mars.
CA: Presumably, one of the first tasks on Mars will be to create a fuel plant that can create the fuel for the return trips of many Starships.
EM: Yes. And actually, it's mostly going to be oxygen plants, because it's 78 percent oxygen, 22 percent fuel. But the fuel is a simple fuel that is easy to create on Mars. And in many other parts of the solar system. So basically ... And it's all propulsive landing, no parachutes, nothing thrown away. It has a heat shield that’s capable of entering on Earth or Mars. We can even potentially go to Venus. but you don't want to go there. (Laughs) Venus is hell, almost literally. But you could ... It's a generalized method of transport to anywhere in the solar system, because the point at which you have propellant depo on Mars, you can then travel to the asteroid belt and to the moons of Jupiter and Saturn and ultimately anywhere in the solar system.
CA: But your main focus and SpaceX's main focus is still Mars. That is the mission. That is where most of the effort will go? Or are you actually imagining a much broader array of uses even in the coming, you know, the first decade or so of uses of this. Where we could go, for example, to other places in the solar system to explore, perhaps NASA wants to use the rocket for that reason.
EM: Yeah, NASA is planning to use a Starship to return to the moon, to return people to the moon. And so we're very honored that NASA has chosen us to do this. But I'm saying it is a generalized -- it’s a general solution to getting anywhere in the greater solar system. It's not suitable for going to another star system, but it is a general solution for transport anywhere in the solar system.
CA: Before it can do any of that, it's got to demonstrate it can get into orbit, you know, around Earth. What’s your latest advice on the timeline for that?
EM: It's looking promising for us to have an orbital launch attempt in a few months. So we're actually integrating -- will be integrating the engines into the booster for the first orbital flight starting in about a week or two. And the launch complex itself is ready to go. So assuming we get regulatory approval, I think we could have an orbital launch attempt within a few months.
CA: And a radical new technology like this presumably there is real risk on those early attempts.
EM: Oh, 100 percent, yeah. The joke I make all the time is that excitement is guaranteed. Success is not guaranteed, but excitement certainly is.
CA: But the last I saw on your timeline, you've slightly put back the expected date to put the first human on Mars till 2029, I want to say?
EM: Yeah, I mean, so let's see. I mean, we have built a production system for Starship, so we're making a lot of ships and boosters.
CA: How many are you planning to make actually?
EM: Well, we're currently expecting to make a booster and a ship roughly every, well, initially, roughly every couple of months, and then hopefully by the end of this year, one every month. So it's giant rockets, and a lot of them. Just talking in terms of rough orders of magnitude, in order to create a self-sustaining city on Mars, I think you will need something on the order of a thousand ships. And we just need a Helen of Sparta, I guess, on Mars.
CA: This is not in most people's heads, Elon.
EM: The planet that launched 1,000 ships.
CA: That's nice. But this is not in most people's heads, this picture that you have in your mind. There's basically a two-year window, you can only really fly to Mars conveniently every two years. You were picturing that during the 2030s, every couple of years, something like 1,000 Starships take off, each containing 100 or more people. That picture is just completely mind-blowing to me. That sense of this armada of humans going to --
EM: It'll be like "Battlestar Galactica," the fleet departs.
CA: And you think that it can basically be funded by people spending maybe a couple hundred grand on a ticket to Mars? Is that price about where it has been?
EM: Well, I think if you say like, what's required in order to get enough people and enough cargo to Mars to build a self-sustaining city. And it's where you have an intersection of sets of people who want to go, because I think only a small percentage of humanity will want to go, and can afford to go or get sponsorship in some manner. That intersection of sets, I think, needs to be a million people or something like that. And so it’s what can a million people afford, or get sponsorship for, because I think governments will also pay for it, and people can take out loans. But I think at the point at which you say, OK, like, if moving to Mars costs are, for argument’s sake, $100,000, then I think you know, almost anyone can work and save up and eventually have $100,000 and be able to go to Mars if they want. We want to make it available to anyone who wants to go.
It's very important to emphasize that Mars, especially in the beginning, will not be luxurious. It will be dangerous, cramped, difficult, hard work. It's kind of like that Shackleton ad for going to the Antarctic, which I think is actually not real, but it sounds real and it's cool. It's sort of like, the sales pitch for going to Mars is, "It's dangerous, it's cramped. You might not make it back. It's difficult, it's hard work." That's the sales pitch.
CA: Right. But you will make history.
EM: But it'll be glorious.
CA: So on that kind of launch rate you're talking about over two decades, you could get your million people to Mars, essentially.
Whose city is it? Is it NASA's city, is it SpaceX's city?
EM: It’s the people of Mars’ city. The reason for this, I mean, I feel like why do this thing? I think this is important for maximizing the probable lifespan of humanity or consciousness. Human civilization could come to an end for external reasons, like a giant meteor or super volcanoes or extreme climate change. Or World War III, or you know, any one of a number of reasons. But the probable life span of civilizational consciousness as we know it, which we should really view as this very delicate thing, like a small candle in a vast darkness. That is what appears to be the case. We're in this vast darkness of space, and there's this little candle of consciousness that’s only really come about after 4.5 billion years, and it could just go out.
CA: I think that's powerful, and I think a lot of people will be inspired by that vision. And the reason you need the million people is because there has to be enough people there to do everything that you need to survive.
EM: Really, like the critical threshold is if the ships from Earth stop coming for any reason, does the Mars City die out or not? And so we have to -- You know, people talk about like, the sort of, the great filters, the things that perhaps, you know, we talk about the Fermi paradox, and where are the aliens? Well maybe there are these various great filters that the aliens didn’t pass, and so they eventually just ceased to exist. And one of the great filters is becoming a multi-planet species. So we want to pass that filter. And I'll be long-dead before this is, you know, a real thing, before it happens. But I’d like to at least see us make great progress in this direction.
CA: Given how tortured the Earth is right now, how much we're beating each other up, shouldn't there be discussions going on with everyone who is dreaming about Mars to try to say, we've got a once in a civilization's chance to make some new rules here? Should someone be trying to lead those discussions to figure out what it means for this to be the people of Mars' City?
EM: Well, I think ultimately this will be up to the people of Mars to decide how they want to rethink society. Yeah there’s certainly risk there. And hopefully the people of Mars will be more enlightened and will not fight amongst each other too much. I mean, I have some recommendations, which people of Mars may choose to listen to or not. I would advocate for more of a direct democracy, not a representative democracy, and laws that are short enough for people to understand. Where it is harder to create laws than to get rid of them.
CA: Coming back a bit nearer term, I'd love you to just talk a bit about some of the other possibility space that Starship seems to have created. So given -- Suddenly we've got this ability to move 100 tons-plus into orbit. So we've just launched the James Webb telescope, which is an incredible thing. It's unbelievable.
EM: Exquisite piece of technology.
CA: Exquisite piece of technology. But people spent two years trying to figure out how to fold up this thing. It's a three-ton telescope.
EM: We can make it a lot easier if you’ve got more volume and mass.
CA: But let's ask a different question. Which is, how much more powerful a telescope could someone design based on using Starship, for example?
EM: I mean, roughly, I'd say it's probably an order of magnitude more resolution. If you've got 100 tons and a thousand cubic meters volume, which is roughly what we have.
CA: And what about other exploration through the solar system? I mean, I'm you know --
EM: Europa is a big question mark.
CA: Right, so there's an ocean there. And what you really want to do is to drop a submarine into that ocean.
EM: Maybe there's like, some squid civilization, cephalopod civilization under the ice of Europa. That would be pretty interesting.
CA: I mean, Elon, if you could take a submarine to Europa and we see pictures of this thing being devoured by a squid, that would honestly be the happiest moment of my life.
EM: Pretty wild, yeah.
CA: What other possibilities are out there? Like, it feels like if you're going to create a thousand of these things, they can only fly to Mars every two years. What are they doing the rest of the time? It feels like there's this explosion of possibility that I don't think people are really thinking about.
EM: I don't know, we've certainly got a long way to go. As you alluded to earlier, we still have to get to orbit. And then after we get to orbit, we have to really prove out and refine full and rapid reusability. That'll take a moment. But I do think we will solve this. I'm highly confident we will solve this at this point.
CA: Do you ever wake up with the fear that there's going to be this Hindenburg moment for SpaceX where ...
EM: We've had many Hindenburg. Well, we've never had Hindenburg moments with people, which is very important. Big difference. We've blown up quite a few rockets. So there's a whole compilation online that we put together and others put together, it's showing rockets are hard. I mean, the sheer amount of energy going through a rocket boggles the mind. So, you know, getting out of Earth's gravity well is difficult. We have a strong gravity and a thick atmosphere. And Mars, which is less than 40 percent, it's like, 37 percent of Earth's gravity and has a thin atmosphere. The ship alone can go all the way from the surface of Mars to the surface of Earth. Whereas getting to Mars requires a giant booster and orbital refilling.
CA: So, Elon, as I think more about this incredible array of things that you're involved with, I keep seeing these synergies, to use a horrible word, between them. You know, for example, the robots you're building from Tesla could possibly be pretty handy on Mars, doing some of the dangerous work and so forth. I mean, maybe there's a scenario where your city on Mars doesn't need a million people, it needs half a million people and half a million robots. And that's a possibility. Maybe The Boring Company could play a role helping create some of the subterranean dwelling spaces that you might need.
EM: Yeah.
CA: Back on planet Earth, it seems like a partnership between Boring Company and Tesla could offer an unbelievable deal to a city to say, we will create for you a 3D network of tunnels populated by robo-taxis that will offer fast, low-cost transport to anyone. You know, full self-driving may or may not be done this year. And in some cities, like, somewhere like Mumbai, I suspect won't be done for a decade.
EM: Some places are more challenging than others.
CA: But today, today, with what you've got, you could put a 3D network of tunnels under there.
EM: Oh, if it’s just in a tunnel, that’s a solved problem.
CA: Exactly, full self-driving is a solved problem. To me, there’s amazing synergy there. With Starship, you know, Gwynne Shotwell talked about by 2028 having from city to city, you know, transport on planet Earth.
EM: This is a real possibility. The fastest way to get from one place to another, if it's a long distance, is a rocket. It's basically an ICBM.
CA: But it has to land -- Because it's an ICBM, it has to land probably offshore, because it's loud. So why not have a tunnel that then connects to the city with Tesla? And Neuralink. I mean, if you going to go to Mars having a telepathic connection with loved ones back home, even if there's a time delay...
EM: These are not intended to be connected, by the way. But there certainly could be some synergies, yeah.
CA: Surely there is a growing argument that you should actually put all these things together into one company and just have a company devoted to creating a future that’s exciting, and let a thousand flowers bloom. Have you been thinking about that?
EM: I mean, it is tricky because Tesla is a publicly-traded company, and the investor base of Tesla and SpaceX and certainly Boring Company and Neuralink are quite different. Boring Company and Neuralink are tiny companies.
CA: By comparison.
EM: Yeah, Tesla's got 110,000 people. SpaceX I think is around 12,000 people. Boring Company and Neuralink are both under 200 people. So they're little, tiny companies, but they will probably get bigger in the future. They will get bigger in the future. It's not that easy to sort of combine these things.
CA: Traditionally, you have said that for SpaceX especially, you wouldn't want it public, because public investors wouldn't support the craziness of the idea of going to Mars or whatever.
EM: Yeah, making life multi-planetary is outside of the normal time horizon of Wall Street analysts. (Laughs) To say the least.
CA: I think something's changed, though. What's changed is that Tesla is now so powerful and so big and throws off so much cash that you actually could connect the dots here. Just tell the public that x-billion dollars a year, whatever your number is, will be diverted to the Mars mission. I suspect you'd have massive interest in that company. And it might unlock a lot more possibility for you, no?
EM: I would like to give the public access to ownership of SpaceX, but I mean the thing that like, the overhead associated with a public company is high. I mean, as a public company, you're just constantly sued. It does occupy like, a fair bit of ... You know, time and effort to deal with these things.
CA: But you would still only have one public company, it would be bigger, and have more things going on. But instead of being on four boards, you'd be on one.
EM: I'm actually not even on the Neuralink or Boring Company boards. And I don't really attend the SpaceX board meetings. We only have two a year, and I just stop by and chat for an hour. The board overhead for a public company is much higher.
CA: I think some investors probably worry about how your time is being split, and they might be excited by you know, that. Anyway, I just woke up the other day thinking, just, there are so many ways in which these things connect. And you know, just the simplicity of that mission, of building a future that is worth getting excited about, might appeal to an awful lot of people.
Elon, you are reported by Forbes and everyone else as now, you know, the world's richest person.
EM: That’s not a sovereign.
CA: (Laughs) EM: You know, I think it’s fair to say that if somebody is like, the king or de facto king of a country, they're wealthier than I am.
CA: But it’s just harder to measure -- So $300 billion. I mean, your net worth on any given day is rising or falling by several billion dollars. How insane is that?
EM: It's bonkers, yeah.
CA: I mean, how do you handle that psychologically? There aren't many people in the world who have to even think about that.
EM: I actually don't think about that too much. But the thing that is actually more difficult and that does make sleeping difficult is that, you know, every good hour or even minute of thinking about Tesla and SpaceX has such a big effect on the company that I really try to work as much as possible, you know, to the edge of sanity, basically. Because you know, Tesla’s getting to the point where probably will get to the point later this year, where every high-quality minute of thinking is a million dollars impact on Tesla. Which is insane. I mean, the basic, you know, if Tesla is doing, you know, sort of $2 billion a week, let’s say, in revenue, it’s sort of $300 million a day, seven days a week. You know, it's ...
CA: If you can change that by five percent in an hour’s brainstorm, that's a pretty valuable hour.
EM: I mean, there are many instances where a half-hour meeting, I was able to improve the financial outcome of the company by $100 million in a half-hour meeting.
CA: There are many other people out there who can't stand this world of billionaires. Like, they are hugely offended by the notion that an individual can have the same wealth as, say, a billion or more of the world's poorest people.
EM: If they examine sort of -- I think there's some axiomatic flaws that are leading them to that conclusion. For sure, it would be very problematic if I was consuming, you know, billions of dollars a year in personal consumption. But that is not the case. In fact, I don't even own a home right now. I'm literally staying at friends' places. If I travel to the Bay Area, which is where most of Tesla engineering is, I basically rotate through friends' spare bedrooms. I don't have a yacht, I really don't take vacations. It’s not as though my personal consumption is high. I mean, the one exception is a plane. But if I don't use the plane, then I have less hours to work.
CA: I mean, I personally think you have shown that you are mostly driven by really quite a deep sense of moral purpose. Like, your attempts to solve the climate problem have been as powerful as anyone else on the planet that I'm aware of. And I actually can't understand, personally, I can't understand the fact that you get all this criticism from the Left about, "Oh, my God, he's so rich, that's disgusting." When climate is their issue. Philanthropy is a topic that some people go to. Philanthropy is a hard topic. How do you think about that?
EM: I think if you care about the reality of goodness instead of the perception of it, philanthropy is extremely difficult. SpaceX, Tesla, Neuralink and The Boring Company are philanthropy. If you say philanthropy is love of humanity, they are philanthropy. Tesla is accelerating sustainable energy. This is a love -- philanthropy. SpaceX is trying to ensure the long-term survival of humanity with a multiple-planet species. That is love of humanity. You know, Neuralink is trying to help solve brain injuries and existential risk with AI. Love of humanity. Boring Company is trying to solve traffic, which is hell for most people, and that also is love of humanity.
CA: How upsetting is it to you to hear this constant drumbeat of, "Billionaires, my God, Elon Musk, oh, my God?" Like, do you just shrug that off or does it does it actually hurt?
EM: I mean, at this point, it's water off a duck's back.
CA: Elon, I’d like to, as we wrap up now, just pull the camera back and just think ... You’re a father now of seven surviving kids.
EM: Well, I mean, I'm trying to set a good example because the birthrate on Earth is so low that we're facing civilizational collapse unless the birth rate returns to a sustainable level.
CA: Yeah, you've talked about this a lot, that depopulation is a big problem, and people don't understand how big a problem it is.
EM: Population collapse is one of the biggest threats to the future of human civilization. And that is what is going on right now.
CA: What drives you on a day-to-day basis to do what you do?
EM: I guess, like, I really want to make sure that there is a good future for humanity and that we're on a path to understanding the nature of the universe, the meaning of life. Why are we here, how did we get here? And in order to understand the nature of the universe and all these fundamental questions, we must expand the scope and scale of consciousness. Certainly it must not diminish or go out. Or we certainly won’t understand this. I would say I’ve been motivated by curiosity more than anything, and just desire to think about the future and not be sad, you know?
CA: And are you? Are you not sad?
EM: I'm sometimes sad, but mostly I'm feeling I guess relatively optimistic about the future these days. There are certainly some big risks that humanity faces. I think the population collapse is a really big deal, that I wish more people would think about because the birth rate is far below what's needed to sustain civilization at its current level. And there's obviously ... We need to take action on climate sustainability, which is being done. And we need to secure the future of consciousness by being a multi-planet species. We need to address -- Essentially, it's important to take whatever actions we can think of to address the existential risks that affect the future of consciousness.
CA: There's a whole generation coming through who seem really sad about the future. What would you say to them?
EM: Well, I think if you want the future to be good, you must make it so. Take action to make it good. And it will be.
CA: Elon, thank you for all this time. That is a beautiful place to end. Thanks for all you're doing.
EM: You're welcome.