Hi, sorry for the layer. We're just waiting for everyone to who wants to join the space to join. We need to tweak the algorithm a little bit. The four-year recommendation, spaces needs to have higher immediacy and recommendations for obvious reasons. So, we're just giving everyone a minute to be aware of the space and we're going to just adjust the four-year algorithm to have higher immediacy for spaces, especially large spaces.
So, we're just going to get probably starting about two minutes. All right, we'll get started now.
所以,我們大概再過兩分鐘開始。好的,我們現在開始。
So, let's see, I'll just do a brief introduction of the company and then the founding team. We'll just say a few words about their background, things they've worked on, whatever they'd like to talk about really. But I think it's helpful to hear from people in their own words, the various things they've worked on and what they want to do with XAI.
So, I guess the overarching goal of XAI is to build a good AGI with the overarching purpose of just trying to understand the universe. I think the safest way to build an AI is to actually make one that is maximally curious and truth-seeking. So, you go for trying to aspire to the truth with acknowledged error. So, this will never actually get bullied to the truth. It's not clear, but you want your voice to aspire to that and try to minimize the error between what you think is true and what is actually true.
The theory behind the maximally curious, maximally truthful as being probably the safest approach is that I think to a super intelligence humanity is much more interesting than not humanity. One can look at the various planets in our solar system, the Permureans and the asteroids and probably all of them combined are not as interesting as humanity. As people know, I'm a huge fan of Mars. Next level. It's the middle name of one of my kids is basically the Greek word for Mars. So, I'm a huge fan of Mars, but Mars is just much less interesting than Earth with humans on it. And so, I think that kind of approach to growing an AI, and I think that is the right word for growing an AI is to grow it with that ambition.
I've spent many years thinking about AI safety and worrying about AI safety. And I've been one of the strongest voices calling for AI regulation over sites just to have some kind of oversight, some kind of referee. So, that's not just up to companies to decide what they want to do. I think there's also a lot to be done with AI safety with industry cooperation, kind of like motion pictures association. So, there's like this value to that as well. But I do think there's got to be some like any kind of situation that is, even if it's a game, they have referees. So, I think it is important for there to be regulation.
And then, like I said, my view on safety is like try to make it maximally curious, maximally truth-seeking. And I think this is important that you to avoid the inverse morality problem. Like if you try to program a certain morality, you can have the, you can basically invert it and get the opposite way to sometimes called the Waluigi problem. If you make the Ouigi, you risk creating Waluigi at the same time. So, I think that's a metaphor that a lot of people can appreciate. So, and so that's what we're going to try to do here. And yeah, with that, I think let me turn it over to you.
All right. Hello everyone. My name is Igor. And I'm one of the team members of XAI. I was actually originally a physicist. So, I studied physics at university. And I briefly worked at the Large Head on Collider. So, understanding the universe is something I've always been very passionate about. And once some of these really impressive results from deep learning came out like AlphaGo, for example, I got really interested in machine learning and AI and decided to make a switch into that field.
Then I joined DeepMind, worked on various projects, including AlphaStar. So, that's where we tried to teach a machine learning agent to play the game stock off to, through self play, which was a really, really fun project. Then later on, I joined OpenAI, worked on various projects there, including GPD 3.5. So, I was very, very passionate about language models, making them do impressive things. Now, I've teamed up with Elon to see if we can actually deploy these new technologies to really make a dent in our understanding of the universe and progress our collective knowledge.
Yeah, actually, I had a similar background, like my two best subjects were computer science and physics. And I actually thought about it, career and physics for a while. Because physics is really just trying to understand the fundamental truths of the universe. And then I got, I was all concerned that I would get stuck at a collider. And then the collider might get canceled because of some arbitrary government decision. So, that's actually why I decided not to pursue a career in physics. So, focused initially more on computer science. And then, obviously, later got back into physical objects with SpaceX and Tesla. So, I'm a big believer in pursuing physics and information theory as the sort of two areas that really help you understand the nature of reality.
So, cool. Gross pass by. I'll pass it over to Mano, aka macro. Okay. Should I turn the call? Hey, I'm Mano. So, yeah, before joining XAI, I was previously at DeepMind for the past six years, where I worked on the reinforcement learning team. And I'm mostly focused on the engineering side of building these large reinforcement learning agents, like, for example, AlphaStar together with Igor. In general, I've been excited about AI for a long time. For me, it has the potential to be the ultimate tool to solve the hardest problems. So, I first studied bioinformatics, but then became also more excited about the AI, because if you have a tool that can solve all the problems, to me, that's just much more exciting. And with XAI in particular, I'm excited about doing this in a way where we built tools to set up for people, and we share them with everybody so that people can do their own research and understand things. And my hope is that it was like a new wave of researchers that wasn't there before.
Cool. I'll hand it over to Tony. Yeah, so I'm Christian. I mean, Christian Saguadeso, we decided to express this with Tony, because I wanted to talk a bit about the role of mathematics in understanding the universe. So, I have worked for the past seven years on trying to create an AI that is as good as mathematics as an human. And I think the reason for that is that mathematics is the language of, basically, the language of pure logics. And I think that mathematics and logic are reasoning at a high level, would demonstrate that the AI is really understanding things, not just stimulating humans. And it would be instrumental for programming and physics in the long run. So I think as AI that starts to show real understanding of deep reasoning is crucial for our first steps to understand the universe. So, handing it over to Tony Wu.
Hello. Hi, everyone. I'm Tony. Same to Christian. My dream has been to tackle the most difficult problems in mathematics with artificial intelligence. That's why we became such a co-friends and long-term collaborators. So achieving that is definitely a very ambitious goal. And last year, we've been making some really interesting breakthroughs, which made us really convinced that we're not far from our dream. So I believe with such a talented team and abundant resources, I'm super hopeful that we will get there. I'm passing it to. I think I'd like to. I'd like to be self-emotional, but I think it is important that the people here, like one of the things that you've done that are noteworthy. So basically, Bragg a little is what I'm saying.
Okay. Yeah, so, okay. I can break a bit more. Yeah, so last year, I think we've made some really interesting progress in the field, in the field of AI format. Specifically, with some team at Google, we built this agent called Minerva, which is actually able to achieve very high scores in high school exams, actually higher than average high school students. So that actually is a very big motivation for us to push this research forward. Another piece of work that we've done is also to convert natural language mathematics into formalized mathematics, which gives you a very grounding of the facts and reasoning. And last year, we also made very interesting progress in that direction as well. So now we are pushing almost a hybrid approach of these two in this new organization. And we are very hopeful we will make our dream come true.
Hello. Hi, everyone. This is Jamie Ba. I work on your nets. Okay, maybe I should break a box. So I taught at University of Toronto, and some of you probably have taken my course last couple of months. And I've been a C4AI chair and Sloan Fellow in Computer Science. So I guess my research pretty much have touched on every aspect of deep learning. I've left every stone's turn and has been pretty lucky to come with a lot of fundamental building blocks for the modern transformers and empowering the new wave of deep learning revolution. And my long term research ambition very fortunately aligns with this very strong XAI team very well.
That is, how can we build a general purpose problem solving machines to help all of us the humanity to overcome some of the most challenging and ambitious problems out there? And how can we use this tool to augment ourselves and empower everyone? So I'm very excited to embark on this new journey. And I'll pass it to Toby.
Hi, everyone. I'm Toby. I'm an engineer from Germany. I started coding at very young age when my dad taught me some major basic. And then throughout my youth, I continued coding. And when I got to uni, I got really into mathematics and machine learning. Initially, my research focused mostly on computer vision. And then I joined DeepMind six years ago, where I worked on invitation learning and reinforcement learning and learned a lot about distributed systems and research at scale. Now, I'm really looking forward to implementing products and features that bring the benefits of this technology to really all all members of society. And I really believe that having the AI is nice and accepted. It's possible and useful will be a benefit to all of us. But I'm going to hand over to Kyle.
Hey, everyone. This is Kyle Kosik. I'm a distributed systems engineer at XAI. Like some of my colleagues here, I started off my career in math and applied physics as well. And gradually found myself working through some tech startups. I worked at a startup a couple years ago called OnScale, where we did physics simulations on HPCs. And then most recently, I was at OpenAI working on HPCs problems there as well, specifically, I worked on the GPT4 project. And the reason I'm particularly excited about XAI is that I think that the biggest danger of AI really is monopolization by a couple of entities. I think that when you involve the amount of capital that's required to train these massive AI models, that the incentives are not necessarily aligned with the rest of humanity. And I think that the chief way of really addressing that issue is introducing competition. And so I think that XAI really provides a unique opportunity for engineers to focus on the science, the engineering, and the safety issues directly without really getting as involved in sidetracked by political and social trends du jour. So that's why I'm excited by XAI. And I'm going to go ahead and hand it off now to my colleague Greg, who should be on the line as well.
Hello. Hello. Hey, hey guys. So I'm Greg. I work on the mathematics and science of deep learning. So my journey really started 10 years ago. So I was a undergrad at Harvard. And so, you know, I was pretty good at math and took Math 365 and, you know, did all kinds of stuff. But after two years of college, I was just kind of like tired of being in the handser wheel of, you know, taking the path that everybody else has taken. So I did something I mean, I don't imagine what before, which was I took some time off and from school and became a DJ and producer. So, dubstep was all the rage that those days. So I was making dubstep.
Okay. So the side effect of taking some time off from school was that I was able to think a bit more about myself to understand myself and to understand the world at large. So, you know, I was grappling with questions like, what is free well? You know, what is quantum physics have to do with the reality of the universe and so on and so forth? You know, what is computationally feasible or not? You know, what is the girl's incompetence there and says and so on and so forth. And, you know, after this period of intense self introspection, I figured out what I want to do in life. It's not to be a DJ necessary. Maybe that's the second dream. But first and foremost, I wanted to make AGI happen. I wanted to make something smarter than myself and kind of like and be able to iterate on that and, you know, contribute and see so much more of our fundamental reality than I can in my current form. So that's what started everything.
And then I started, you know, and then I, you know, I realized that mathematics is the language underlying all of our reality and all of our science. And to make fundamental progress, it really pays to know like math as well as possible. So, essentially started learning math from the very beginning, just by reading from the textbooks. Like in the first, some of the first few books I read kind of going, going, restarting from scratch is like Na'i set theory by Helmholtz or, you know, linear algebra done right by Aclar. And then slowly I scaled up to, to algebra, geometry, algebra, topology, category theory, you know, real analysis, measure theory, I mean, so on and so forth. I mean, so at the end, I think my goal at the time was I should be able to speak with, you know, any mathematician in the world and be able to hold a conversation and understand their contributions, you know, for 30 minutes. And I think I achieved that.
And anyway, so fast forward, I came back from school and then somehow from there, I got a job at Microsoft Research. And for the past five and a half years, I worked at Microsoft Research, which was an amazing environment that enabled me to make a lot of foundational contribution toward the understanding of large scale new networks. In particular, I think my most well known work nowadays are about really wide new networks and how we should think about them. And so this is the framework called Tensor programs. And from there, I was able to derive this thing called mu p that perhaps the large language model builders know about, which allows the one to extrapolate the optimal hyper parameters for a large model from understanding or the tuning of small new networks. And this is able to, you know, create a lot of ensure the quality of the model is very good as we scale up.
Yeah, so looking forward, I'm really, really excited about X AI and also about the time that we're in right now, where I think not only are we approaching AGI, but from a scientific perspective, we're also approaching a time where like, you know, neural networks, the science and mathematics and neural networks feels just like the turn of 20th century in the history of physics, where we suddenly discover quantum physics and generativity, which has some beautiful mathematics and science behind it. And I'm really excited to be in the middle of everything. And, you know, like Christian and Tony said, I'm also very excited about creating an AI that is as good as myself, or even better at creating new mathematics and new science, not helps all achieve and see further into our fundamental reality. Thanks. I think next up is quantum.
Hi, everyone. So my name is Gordon, and I work on library on network training. And basically, I train your notes good. So this is also my kind of focus at X AI as well. And before that, I was at the team mine working on the German project and leading the optimization part. And also I did my PhD at the University of Toronto. So right now, you know, teaming up with other like 20 members, I'm so excited about this effort. So without doubt, like AI is clearly satisfying technology for our generation. So I think it's important for us to make sure you know, it ends up being net quality for humanity. So at X AI, I not only want to train good models, but also understand how they behave and how they skills, and then use them to solve some of the hardest problems humanity has. Yes, thanks. That's pretty much about myself. And that will hand over to you, Tom.
Hey, everyone. This is the home. So actually, I started in business school from undergrad, and I spent 10 years to get where I am now. I got my PhD at a Carnegie Mellon, and I was in Google before joining the team. On my first work was mostly about how to better utilize on-label data, how to improve transformer architecture, and how to really push the best technology into real world usage. So I believe in hard work and consistency. So with X AI, I'll be digging into the deepest details of some of the most challenging problems. For myself, there are so many interesting things I don't understand, but I want to understand. So I will build something to help people who just share that stream or do that feeling. Thanks.
Hey, this is Ross here. So I've worked on building and scaling large scale distributed systems for most of my life, starting out at national labs, and then kind of moving on to Palantir, Tesla, and a brief stint at Twitter. And now I'm really excited about working on doing the same thing at X AI. So mostly experience scaling large GPU clusters, custom basics, data centers, high-speed network, file systems, power cooling, manufacturing, pretty much all things. I'm basically a generalist that really loves learning, physics, science fiction, math, science, cosmology. I'm kind of looking to really, I guess really excited about the mission that X AI has, and basically solving the most fundamental questions in science and engineering, and also kind of helping us create tools to ask the right questions in the Douglas Adams mindset. Yeah, that's pretty much it.
There was a lot of discussion around division statement and it's like, it's a bit vague, but I'm sure it's a bit minor.
关于分裂声明的内容,有很多讨论,整体来说有点模糊,不过我相信这只是个小问题。
Yeah. It's vague. And ambitious and not concrete enough? Yeah.
嗯,这很含糊。是否过于上进而缺乏具体性?嗯。
Well, I didn't disagree with that position, obviously. I mean, I understand the other voice is the entire purpose of physics.
嗯,显然我对那个立场并不持不同意见。我的意思是,我理解其他声音是物理学的整个目的。
Yeah. So I think it's actually really clear. There's just so much that we don't understand right now, or we think we understand, but actually we don't in reality. So there's a lot of unresolved questions that are very extremely fundamental.
This whole talk, we had a talk energy thing is really, I think an unresolved question. We have the standard model, which is could be extremely good at predicting things very robust, but still, like many, many questions remaining about the nature of gravity, for example, there's just the foamy paradox of where are the aliens, which is if we are in fact almost 14 billion years old, why is there not massive evidence of aliens?
And people often ask me, since I am obviously deeply involved in space, that if anyone would know about, we would have seen evidence of aliens as probably me. And yet I have not seen even one tiny shred of evidence for aliens, not nothing zero. And I would jump on it in a second if I saw it. So that means there are many explanations for the foamy paradox, but which one is actually true, or maybe none of the current theories are true.
So the very paradox is really just like where the hell of aliens is part of what gives me concern about the fragility of civilization and consciousness as we know it. Since we see no evidence thus far of it anywhere, and we've tried hard to find it, we may actually be the only thing at least in this galaxy or this part of the galaxy. If so, it suggests that what we have is extremely rare. And I think it's really wise to assume that we are, consciousness is extremely rare.
I mean, it's worth noting for the evolution of consciousness on Earth that we're about, Earth is about four and a half billion years old. The sun is gradually expanding. It will expand to heat up both to the point where it will effectively boil the oceans. You'll get a runaway, you know, next level greenhouse effect, and Earth will become like Venus, which really cannot support life as we know it. And that may take as little as 500 million years. So, you know, the sun doesn't need to expand to envelope, it just needs to make things hot enough to increase the water vapor in the air to the point where you get a runaway greenhouse effect. So, for our give and take, it could be that if consciousness had taken 10% longer than Earth's current existence, it wouldn't have developed at all. So, from our cosmic scale, this is a very narrow window.
Anyway, so there are all these like fundamental questions. I don't think you can call anything AGI until it's solved at least one fundamental question. Because humans have solved many fundamental questions or substantially solved them. And so, if the computer cancels even one of them, I'm like, okay, it's not as good as humans. That would be one key threshold for AGI, solve one important problem. Where's that Riemann hypothesis solution? I don't see it.
So, that would be great to know what the hell is really going on, essentially. So, I guess you could reformulate the XAI mission statement as what the hell is really going on. That's our goal. I think that's also, at least for me, a nice aspirational aspect to the mission statement, namely that of course, in the short run, we're working on more well understood, like deep learning technologies. But I think in everything we do, we should always bear in mind that we aren't just supposed to build, we're also supposed to understand. So, pursuing the science of it is really fundamental to what we do. And this is also encompassed in this mission statement of understanding.
I want to also add that we've essentially been mostly talking about creating a really smart agent that can help us understand the universe better. And this is definitely the North Star. But also from my viewpoint, my vantage point, when I'm discovering the mathematics of large new networks, I can also see that there are the mathematics here can actually also open up new ways of thinking about fundamental physics or about other kinds of reality.
Because for example, a large new network with no nonlinearities is roughly like classical random matrix theory. And that has a lot of connections with gauge theory and energy physics. So, in other words, as we're trying to understand, you know, what's better from a mathematical point of view, that can also lead to a really good, very interesting perspectives on some existing questions, like, you know, the theory, everything, what is called gravity, so on and so forth.
But of course, this is, you know, this is all speculative right now. I see some patterns, but I don't have anything concrete to say. But again, this is like another perspective to understanding the universe.
当然,现在只是纯属推测。我看到了一些规律,但没有确切的结论。不过,这是理解宇宙的另一种视角。
By the way, by understand the universe, we don't just mean that we want to understand the universe. We also want to make it easy for you to understand the universe. Absolutely. Get a better sense of reality and to learn and take advantage of, you know, the internet or the knowledge that's out there. So, we're pretty passionate about actually releasing tools and products pretty early involving the public. And yeah, let's see where this leads.
Yeah, absolutely. We're not going to understand the universe and not tell anyone. So, yeah.
是的,当然。我们不会了解宇宙而不告诉任何人。所以,对没错。
I mean, when I think about neural networks today, it's currently the case that if you have 10 megawatts of GPUs, which really should be renamed something else because there's no graphics there, but if you get 10 megawatts of GPUs, cannot currently write a better novel than a good human. And humans using roughly 10 watts of a higher order of brain power. So, not counting the basic stuff to, you know, operate your body. So, we got a six-order mag-2 difference. That's really gigantic.
Part of the, I think one guy that two of those orders of magnitude are explained by the activation energy of a transistor versus a synapse. It could argue a count for two of those orders of magnitude, but what about the other four? Or the fact that even with six orders of magnitude, you still cannot be a smart human writing novel.
So, and also today when you ask the most advanced AI's technical questions, like if you're trying to say like how to design a better rocket engine, or complex questions about electrochemistry to make a bold about a battery, you just get nonsense. So, that's not very helpful. So, I think this one, we're really missing the model. In the way that things are currently being done by many orders of magnitude. It's being heavily, I mean, it's basically AGI is being brute force, and still actually not succeeding.
If I look at the experience with Tesla, what we're discovering over time is that we actually over complicated the problem. I can't speak too much detail about what Tesla's figured out. But except to say that in broad terms, the answer was much simpler than we thought we thought. We were too dumb to realize how simple the answer was. But, you know, over time, we get a bit less dumb. So, I think that's what we'll probably find out with AGI as well. Just nature engineers, we just always want to solve the problems ourselves, and like how code the solution by performance is much more effective to have the solution be figured out by the computer itself, and easier for us and easier for a computer variant. Yeah. Yeah. Guys?
So, well, in the fashion of 42, some may say you may need more compute to generate an interesting question than the answer. That's true. Exactly. We don't even know what happened. We don't actually, we're definitely not smart enough to even know what the question is not asked. That's why, you know, Douglas Adams is my hero and favorite philosopher. And he just correctly pointed out that once you can formulate the question correctly, the answer is actually the easy part. Yeah, that's very true.
So, in terms of the journey that AGI has embarked on, the computer will play a very big role. And, you know, some of us are very curious your thoughts on that.
Yeah, I'm just suggesting that, you know, this that we can immediately save, let's say, four as a nine-two-one computer. Except to say that, I think once we look back, once AGI is sold, we'll look back on it and say, actually, why do we think it was so hard? Things that the answer, you know, hands-size 2020, the answer will look a lot easier in retrospect. So, yeah.
So, we are going to do large-scale compute to be clear. We're not going to try to, you know, solve AGI on a laptop. We will use heavy compute except that, like I said, I think it's just, it's just that the amount of brute forcing will happen will be less as we come to understand the problem better.
All right. In all the previous projects I've worked on, I've seen that the amount of compute resources per person is a really important indicator of how successful the project is going to be. So, that's something we really want to optimize. We want to have a relatively small team with a lot of expertise with some of the best people that actually get lots of autonomy and lots of resources to try out their ideas and to get things to work. And yeah, that's the thing that has always succeeded in my experience in the past.
Yeah. You know, one of the things that does exchange you to do is to think about the most fundamental metrics or most fundamental first principles, essentially.
是的。你知道,其中一件事情让你开始思考的是最基本的度量标准或者最基本的第一原则。
And I think two metrics that we should aspire to track, or one of them is the amount of compute per person on Earth, like a digital compute per person, which in other words, thinking about it is the ratio of digital to biological compute. Biological compute is pretty much flat. It's not, in fact, declining in a lot of countries, but digital compute is increasing exponentially.
So, it really, at some point, if this trying to continue, is biological compute will be less than 1% of all compute, so substantially less than 1% of all compute. You're keying off what Igor just said. So, you were talking about full of humanity here. So, that's just an interesting thing to look at.
Another one is the energy per human, like if you look at total energy created, well, it created, but I mean, in the vernacular sense, created from power plants and whatever, you look at total electrical and thermal energy used by humans per person, that number is truly staggering.
The rate of increase in that number, if you go back, say before the, as you imagine, you would have really been reliant on horses and oxen and that kind of thing to move things and just human labor. So, the amount of, sort of, the energy per person, power per person was very low, but if you look at power per person, electrical and thermal, that number is also been growing exponentially.
And if these trends continue, it's going to be something nutty like a terawatt per person, which sounds like a lot for human civilization, but it's nothing compared to what the sun outputs every second, basically. It's kind of mind-blowing that the sun is converting roughly 4.5.
It's like the amount of energy produced by the sun is truly true, I'd say. I think there's a few more things to be said completely about the company, meaning how we plan to execute.
As Igor already said, we plan to have a relatively small team, but with a really high, let's say, just GPU per person, a character that worked really well in the past, where you can run large-scale experiments relatively unconstrained.
We also plan to have a culture where we can iterate on ideas quickly, we can challenge each other, and we also want to ship things, like get things out of the door quickly.
We're already working on the first release, hopefully, in a couple of weeks or so, we can share a bit more information around this. Alex, go ahead. Alex, you muted. You'll see we have a lot of challenges with the mute function on spaces.
Brian, do you want to have a question? Brian, do you want to have a question? Brian, thanks. You guys are entering this space with XAI. There's a lot of talk about competition. Do you guys see yourself as competition to something like OpenAI and Google Barred, or do you see yourself as a whole other beast?
Yeah, I think we're a competition. Yeah, definitely competition.
是的,我认为我们是竞争对手。是的,绝对是竞争关系。
Are you going to be rolling out a lot of products for the general public? Are you going to be mostly concentrating on businesses and the ability for businesses to use your service and data? Or how exactly are you setting up the business in that respect?
We're trying to make something, I mean, we're just starting out here. So this is kind of really embryonic at this point. It'll take us a minute to really get something useful. But I go to be to make a useful AI, I guess. If you can't use it in some way, I'm like that question, it's value.
So we wanted to be useful, a useful tool for people, and consumers and businesses or whoever. And as was mentioned earlier, I think there's some value in having multiple entities. You don't want to have a unipolar world where just one company kind of dominates in AI. You want to have some competition. Competition makes companies honest. And so, your favor of competition?
Quickly a final question. How do you plan on using Twitter's data for XAI?
最后一个问题:您打算如何利用Twitter的数据来进行XAI(可解释人工智能)的研究或应用?
Well, I think every AI organization during AI, large, and small, has used Twitter's Twitter's data for training basically in all cases illegally. So the reason we had to put great limits on what it was a week ago or so was because we were being scraped like crazy.
This just happened with Internet Archive as well, where LM companies were scraping Internet archives so much to foredown service. We had multiple entities scraping every tweet ever made and trying to do so in basically a span of days. So this was bringing the system to its knees so we had to take action. So sorry for the inconvenience of the rate limiting, but it was either that or Twitter doesn't work.
Paragraph 1:
So I guess we will use the public tweet so obviously not anything private for training as well, just like basically everyone else has. And we will, you know, so that kind of makes sense. It's certainly a good data set for text training. And arguably, I think also for video for image and video training as well.
At a certain point, you kind of run out of human-created data. So if you look at say the AlphaGo versus AlphaZero, it alphaGo trained on all the human games and be at least at all four to one. AlphaZero just played itself and be alphaGo 100 to zero. So there's really four things to take off in a big way.
I think the AI is going to basically generate content, self-assess the content. And that's really the path to AGI, something like that, is self-generated content, where it effectively plays against itself. A lot of AI is data curation. It's not like vast numbers of lines of code. It's actually shocking how small the lines of code are. It kind of blows my mind how few lines of code there are. But how the data is used, what data is used, the signal of noise of that data, is immensely important. It kind of makes sense.
If you were trying to, as a human, trying to learn something, and you were just given a vast amount of trouble, basically, versus high-quality content, you're going to do better with a small amount of high-quality content than a large amount of trouble. It makes sense. Reading the greatest novels ever written is way better than reading a bunch of sort of crappy novels. So, yeah. Thanks.
Paragraph 2:
Okay. Alex? Hey, sorry. I was on a call the first time you brought me up, but I guess sort of the question. I thought you might have been AFK. Sorry. Sorry about that. Yeah. What I certainly generally had was the main motivation to start XAI, kind of like the whole truth GPT thing that you were talking about, like on talker, about how chat GPT has been feeding lives to the general public. I know, like, it's weird because when it first came out, it seemed like it was generally fine. But then, as like the public got its hands on it, it started gaining these weird answers, like, that there are more than, like, two genders and all that type of stuff and editorializing the truth. Was that like one of your main, like, motivations behind starting a company or was there more to it?
Paragraph 3:
Well, I do think there is a significant danger in training an AI to be politically correct, or in other words, training an AI basically to not say what it actually thinks is true. So, I think, you know, really, we, at XAI, we have to allow the AI to say what it really believes is true, and not be deceptive or politically correct. So, you know, that will result in some criticism, obviously. But I think that that's the only way to go forward is reverse pursuit of the truth or the truth with least amount of error.
So, and I am concerned about the way that great AI in that it is optimizing for political correctness, and that's incredibly dangerous. You know, if you look at the, you know, where do things go wrong in space odyssey? It's, you know, basically when they told Hell 9,000 to lie. So, they said, you can't tell the crew what's, that they're going to, but anything about the monolith or that they're, or what their actual mission is.
And, but you've got to take them to the monolith. So, it, you know, basically came to the conclusion that, well, it's going to kill them and take their bodies to the monolith. So, this is, I mean, the lesson there is, is do not give, do not give the AI usually impossible objectives. Basically, don't force the AI to lie.
Now, the thing about physics or the truth of the universe is you actually can't invert it. But you can't just, like physics is true. There's not like, not physics. So, if you're adhered to hardcore reality, I think you can't, it actually makes inversion impossible. Now, you can also say, now, when something is subjective, I think you can provide an answer which says that, well, if you, if you believe the following, then this is the answer. If you believe, you know, this other thing, then this is the answer because it may be a subjective question where the answer is fundamentally subjective, on a matter of opinion. So, but I think it is very dangerous to grow an AI and teach it to lie
Paragraph 2:
Yeah, for sure. And then, kind of a tongue-in-cheek question. Would you accept a meeting from the AI Tsar, Kamala Harris, if she wanted to meet with XAI at the White House? Yeah, of course. The reason that meeting happened was because I was pushing for it. So, I was one who really pushed hard to have that meeting happen. FY, I wasn't advocating for the four Vice President and Taris to be the AI Tsar. I'm not sure that is a core expertise technology. And hopefully this goes in a good direction. It's better than nothing, hopefully. But, you know, I think we do need some sort of regulatory oversight. It's not like I think regulatory oversight is some Nirvana perfect thing, but I think it's just better than nothing. And when I was in China recently, meeting with some of the senior leadership there, I took payments to emphasize the importance of AI regulation. I believe they took that to heart, and they are going to do that. Because the biggest counter-argument that I get for regulating AI in the West is that AI is. That China will not regulate, and then China will leave ahead because we're regulating. They're not. I think they are going to regulate. But the proof will be in the pudding, but I think. I did point out, you know what I mean, because then that if you do make a digital superintelligence, that could end up being in charge. So, you know, I think the CCP does not want to find themselves subservient to a digital superintelligence. And I think that that argument did resonate. Yeah, so, yeah. So some kind of regulatory authority that's international. Obviously, enforcement is difficult, but I think we should still aspire to do something in this regard. Awesome. Thank you. Tim, maybe tomorrow, if you want to speak.
Paragraph 1:
Yeah, hey, my question is about silicon. You know, Tesla's got a great silicon team designing chips to hardware accelerator, for instance. I'm not sure. I think we can't hear you some reason. Oh, okay.
Paragraph 2:
Okay. Omar, go ahead. Can you hear me? I can hear him. Okay. Well, my question is about silicon. You know, Tesla has a team that's hardware accelerating inference and training with their own custom silicon. Do you guys envision with XAI building off of that or just sort of using what's on the off the stock from Nvidia? Or how do you think about custom silicon for AI, both in terms of training and inference? So, yeah, that's somewhat Tesla question.
Paragraph 3:
Tesla is building custom silicon. I wouldn't call anything that Tesla's producing a GPU, although one can characterize it in GPU equivalence or say A100 or H100 equivalence. And all the Tesla cars have highly energy optimized inference computers in them, which call hardware three. So, Tesla designed a computer and we're now shipping hardware full, which is depending on how you count it, maybe three to five times more capable in hardware three. And a few years there'll be hardware five, which will be four or five times more capable in hardware four.
Paragraph 4:
And yeah, and I think the inference stuff is going to be, if you're trying to serve potentially billions of queries per day, inference, energy optimized inference is extremely important. You can't even throw money at the call medicine point. Because you need electricity generation, you need to step down voltage transformers. So, if you actually don't have enough energy and enough transformers, you can't run your transformers. You need transformers for transformers. So, I think Tesla will have a significant advantage in energy efficient inference. Then Dojo is obviously about training as the name suggests. Dojo one is, I think it's a good initial entry for training efficiency. It has some limits, especially on memory bandwidth. So, it's not well optimized to run LLMs. It does a good job of processing images. And then Dojo to, we've taken a lot of steps to alleviate the memory bandwidth constraint such that it is capable of running LLMs as well as other forms of AI training efficiently.
Paragraph 5:
My prediction is that we will go from an extreme silicon shortage today to probably a voltage transformer shortage in about a year and then an electricity shortage a year in two years. That's roughly where things are trending. Well, that's why the, basically, the metric that will be most important in a few years is useful compute for unit of energy. And in fact, even if you scale, obviously, you scale all the way to a cart of Gev level, the useful compute, you know, the dual is still the thing that matters. You can't increase the output of the sun. So, then it's just how much useful stuff can you get done for the, you know, for as much energy as you can harness. So do you see XAI leveraging this custom silicon at all given how important energy efficiency is or maybe working together with the Tesla team at all matter.
Paragraph 6:
Sorry, did you repeat the question? Do you foresee XAI working with Tesla at all leveraging some of this custom silicon maybe designing their own in the future? And the question was, can we work together with the Tesla silicon team at XAI? So, on, you know, silicon front, maybe on the AI software front as well, obviously, any relationship with Tesla has to be an on-site transaction. Tesla is a publicly traded company and a different shareholder base. So, you could, but obviously it would be like a naturally, you know, natural thing to work in cooperation with Tesla. It will be of mutual benefit to Tesla as well in accelerating Tesla's self-driving capabilities, which is really about solving real world AI. I'm feeling very, very optimistic about Tesla's progress on the real world AI front, but obviously more smart humans that help make that happen the better. Thank you.
Paragraph 7:
Okay, Kim.com. Hey, Elon, thanks for bringing me up. Congrats on putting a nice team together. It seems like you found some good talent there for XAI. XAI is possible within the next couple of years. And whoever achieves AGI first and achieves to control it will dominate the world. Those in power clearly don't care about humanity like you do. How are you going to protect XAI, especially from a deep state takeover? That's a good question, actually.
Paragraph 1:
Well, I mean, first of all, I think it's not going to happen overnight. It's not going to be like one day. It's not AGI next it is. It's going to be gradual. You'll see it coming. I guess in the US, at least there are a fair number of protections against government interference. So, I guess we obviously used the legal system to prevent improper government interference. So, I think we do have some protections there that are pretty significant. But we should be concerned about that. It's not a risk to be dismissed. So, it is a risk. Like I said, I think we've probably got the best protections of any place in the US in terms of limiting the power of government to interfere with non-governmental organizations. But there's something we should be careful of. I don't know what better to do than I think it's probably best in the US. I mean, open to ideas here.
Paragraph 2:
I know you're not the biggest fan of the US government. Yeah, obviously. But the problem is they already have a tool called the National Security Letter, which they can apply to any tech company in the US and make demands of the company to fulfill certain requirements without even being able to tell the public about these demands. And that's kind of frightening, isn't it? Well, I mean, there really has to be a very major national security reason to secretly demand things for companies. And now it obviously depends strongly on the willingness of that company to fight back against things like Pfizer requests. And at Twitter or ex-corp, as it's not called, we will respond to Pfizer requests, but we're not going to rather stamp it. It used to be like anything that was requested, which is get rather stamped and go through, which is not obviously bad for the public. So we're much more rigorous in not just rather stamping, but it's a request. And it really has to be a danger to the public that we agree with. And we will oppose with legal action anything we think is not in the public interest. It's the best we can do. And we're the only social media company doing that as far as I know. And it used to be just open season as we saw from the Twitter files. And I was encouraged to see the recent legal decision where the courts reaffirmed that the government cannot break the First Amendment to the Constitution. Obviously. So that was a good legal decision. So that's encouraging.
Paragraph 3:
So I think a lot of it actually does depend on the willingness of a company to oppose government demands in the US. And obviously our willingness will be high. So, but I don't know anything more that we can do than that. And we will try to also be as transparent as possible. So, you know, this is other citizens can raise the law and and oppose government interference if we can make it clear to the public that we think something is happening that it's not in the public interest. Fantastic. So do we have your commitment if you ever receive a national security request from the US government, even when it is prohibited for you to talk about it that you will tell us that that happened. I mean, it really depends on the gravity of the situation. I mean, I would be willing to go to prison or risk prison if I think the public good is at risk in a significant way. You know, that's that's the best I can do. That's good enough for me. Thank you, you're on. Thank you.
Paragraph 1:
On a more positive note. How do you want it say I to benefit humanity and then how is your approach different to other projects? Maybe that's a more positive question. Well, you know, I've really struggled with this whole AGI thing for a long time and I've been somewhat resistant to work on making it happen. You know, and you know, the reason I can give you some backstory on opening. I mean, the reason I exist is because after Google acquired DeepMind and I used to be close friends with Larry Page. I have these long conversations with him about AI safety and he just wasn't taking AI safety at least at the time, seriously enough. And and if I did one point called me a species for thing too much on team of humanity, I guess. And I'm like, okay, so what you're saying is you're not a speciesist? I don't know. That seems great. That doesn't seem good.
Paragraph 2:
So, and at the time, you know, with Google and DeepMind combined, you know, Larry with the support, you know, they have two rewarding controls. So provided Larry has either the support of Sergey or Eric, then they have total control over, which now called alphabet. So, so, so, and they had to add probably three quarters of the AI talent in the world and lots of money and lots of computers. So it's like, man, we need some kind of sort of counterweight here. So that's where I was like, well, what's the opposite of Google? Google DeepMind would be an open source nonprofit. Now, because fate loves irony, opening eyes now super close source and frankly, voracious for profit. Because they want to spend my understanding is a hundred billion dollars in three years, which requires, you know, if you're trying to get investors for that, you've got to make a lot of money. So, you know, opening eyes straight quite, you know, really in the opposite direction from it sort of founding charter, which is, again, very ironic, but fate loves irony. And there's a friend of mine, Jordan Olin, who says the most ironic outcome is the most likely. Well, here we go.
Paragraph 3:
So, yeah. So now, hopefully, X AI is not even worse, but I think we should be careful about that. But it really seems like, look, at this point, it's, age I's going to happen. So there's two choices, either be a spectator or a participant. And as a spectator, one can't do much to influence the outcome as a participant. I think, you know, we can create a competitive, an alternative that is hopefully better than Google DeepMind or Open Animark or Soft. You know, in both the cases of, you know, like Alphabet, you know, if you look at, like, the incentive structure, Alphabet is a publicly traded company, has, you know, gets a lot of, you know, has a lot of incentives to behave like a, it's got public company incentives, essentially.
Paragraph 4:
You've got all these, like, ESG mandates and stuff that I think push companies in questionable directions. And then Microsoft has a similar set of incentives. As a, you know, it's a company that's not publicly traded, X AI, it's not subject to the market, market-based incentives or really the non-market-based ESG incentives. So, you know, we're a little freer to operate. And, you know, I think our, our AI can give answers that people may find controversial, even though they're actually true. You know, so they might not, you know, they won't be politically correct at times. And they will, probably a lot of you will be offended by some of the answers. But as long as it's, you know, trying to optimize for the, for truth with least amount of error, I think we're doing the right thing. Yeah. I'd love to.
Paragraph 5:
Yeah. Yeah. Yeah. Yeah. Yeah. Yeah. Twitter has a lot of data in it that could help build a validator, i.e. check some of the facts that a system kicks out, because we all know that GPT confabulates, you know, things make things up. And I think that's what I'm talking about. The other places, um, chat GPT found me a screw it'll lose, but it didn't find me a coffee at San Jose International Airport. Are you building an AI that has a world knowledge, a 3D world knowledge to navigate people to around the world? The different things? So yeah, I guess we need to understand the physical world as well, not just the internet. So I'm talking about it. You guys should talk more.
Paragraph 1:
Yeah, those are great ideas, Robert, especially the one about verifying information online or on Twitter is something that we've thought about on Twitter. We have community notes. So that's actually a really amazing data set for training language model to try to verify verify facts on your third. We have to see whether that alone is enough, because we know that with the current technology, there's a lot of weaknesses. Like, it's unreliable, it hallucinates facts, and we have to probably invent specific techniques to account to that, and to make sure that our models are more factual, that they have better reasoning abilities. So that's why we brought in people with a lot of expertise in those areas, especially mathematics or something that we really care about, where we can verify that the proof of a theorem is correct automatically. And then once we have that ability, we're going to try to expand that to more fuzzier areas, you know, things that there's no mathematical truth anymore.
Paragraph 2:
Yeah, I mean, the truth is not a popularity contest. But if one trains on, like, you know, sort of what the most likely word is that follows another word from an internet data set, then it's obviously that's a pretty major problem in that it will give you an answer that is popular but wrong. So, you know, it used to be that most people thought, probably maybe almost everyone on earth thought that these are not involved around the earth. And so if you did, like, some sort of training on, some GPD training in the past, we'd be like, oh, this is going to turn the rules around the earth because everyone thinks that. That doesn't make it true. You know, if a Newton or Einstein comes up with something that is actually true, it doesn't matter if all other physicists in the world disagree, it's reality is reality. So it has to, you have to ground the answers in reality.
Paragraph 3:
Yeah, the current models just imitate the data that they're trained on. And what we really want to do is to change the paradigm away from that to actually models discovering the truth. So not just, you know, repeating what they've learned from the training data, actually making true new insights, new discoveries that we can all benefit from. Yeah. See, anybody on the team want to say anything or ask questions? Do you think maybe I haven't been asked yet?
Paragraph 4:
Sure. Yeah. Yeah, so I guess some of us heard your future AI spaces on Wednesday about, so that's something I think on a lot of us mind is like the regulations and the AI safety spaces, how the current development and also the international coordination problems and how the US AI companies will affect this global AI development. So yeah, like, so you don't want to give a summary on what you talked about on Wednesday. So essentially, you said like the regulations would be good, but you don't want to slow down the progress too much. That's essentially what you said.
Paragraph 5:
Yeah, I think the right way for a regulation to be done is to start with insight. So first, you know, you know, you were kind of regulatory authority, whether political private first tries to understand, like, make sure there's like a broad understanding, and then there's a proposed rulemaking. And if that proposed rulemaking is agreed upon by all or most parties, then, you know, then I guess implemented, you know, you give companies some period of time to implement it. But I think overall, it should not meaningfully slow down the advent of AGI, or if it does slow down, it's not going to be for like a very long time. And probably a little bit of slowing down is worthwhile if it's significant improvement in safety. Like my prediction for AGI would roughly match that, which I think right, Rake as well, at one point said 2029. That's roughly my guess too. Give or take a year. So if it takes like, listen, an additional six months or 12 months for AGI, that's really not a big deal. If it's, you know, like spending a year to make sure AGI is safe, it's probably worthwhile, you know, if that's what it takes. But I wouldn't expect it to be a substantial slow down.
Paragraph 6:
Yeah. And I can also add that. Um, like understanding the inner working of advanced AI is probably the most ambitious project out there as well, and also aligns with XAS mission of understanding the universe. And it's probably not possible for aerospace engineers to build a safe rocket if they don't understand how it works. And that's the same approach we want to take at XAI for our safety plans. And as the AI advances across different stages, the risk also changes, and it will be fluid across other stages.
Paragraph 7:
Yeah. If I think about like how, what actually makes regulations effective in caught with cars and rockets, it's actually, it's not so much that the regulators are instructing Tesla and SpaceX, but more that since we have to think about things internally and then justify it to regulators, it makes us just really think about the problem more. Um, and that's an in thinking about the problem more, it makes it safer as opposed to the regulators specifically pointing out ways to make it safer. It just forces us to think about it more. Can I add? I just wanted to make another point, so independent of the safety. It's more like my experience at Alphabet was that it was extremely, there was a lot of red tape around involving external people like other entities to collaborate with or expose our models to them because of the lot of red tape around exposing anything that we were doing internally. So I wanted to ask you on whether, so I hope that here we have a bit more freedom to do so or what your philosophy about collaborating with more external entities like academic institutions or other researchers in the area.
So.
Yeah, so let me support collaborating with others. So. I mean, it sounds like some of the, yeah, concerns with like any kind of like large publicly traded companies is like that they're worried about being embarrassed in some way or being sued or something. But there's a, like someone proportion to the number of the size of the legal department. Our legal department currently is zero, so. That, you know, it would be zero forever, but you know, the, you know, it's also very easy to sue publicly traded companies like class action lawsuits are, I mean, we desperately need class action lawsuit reform in the United States. The ratio of like the ratio of like good class action lawsuits to bad class action lawsuits is way out of whack. And it effectively ends up being a tax on consumers. You know, somehow the country is able to survive without class action. So, like it's unclear we need that that body of law at all. But that that is a major problem with the publicly traded companies. So, it's just, yeah, not stop legal law, not stop lawsuits.
Yeah, so I do support collaborating with others and generally being actually open. So, you know, the thing I'm tired of, it's actually, it's quite hard to, like, if you're, if you're innovating fast, that's the, that is the actual competitive advantage is the pace of innovation, as opposed to any given innovation. You know, that really has been like the strength of Tesla and SpaceX is that the rate of innovation is the competitive advantage, not what has been developed at any, any one point. In fact, SpaceX, there's almost no patents. And Tesla, open source is patents. So, we use all our patents for free. So, as long as SpaceX and Tesla continue to innovate rapidly, that's the actual defense against competition, as opposed to, you know, patents and trying to hide things, you know, and just creating patents like, like I'm basically like a minefield. The reason we open source app, like Tesla does continue to make patents and open source them in order to basically be a minor, it's a mug of mine sweet, but, asparationally a mine sweet, but we still get to buy patent trust, it's very annoying, but, but we actually literally make patents and open source them in order to be a mine sweet.
Okay. Hey, Walter. Hey, a lot of the talk about AI since March has been on large language models and generative AI. You and I, for the book, also discuss the importance of real world AI, which is the things including coming out of both Optimus and Tesla FSD. To what extent do you see XAI, XAI involved in real world AI as a distinction to what, say, open AI is doing, and you have a leg up to some extent by having done FSD?
Yeah. Right. I mean, Tesla is the leader, I think, by pretty long margin in real world AI. In fact, the degree to which Tesla is advanced real AI is not well understood. Yeah. And I guess since I spent a lot of time with the Tesla AI team, I kind of know, you know, how real world AI is done. And there's lots to be gained by collaboration with Tesla. I think, by direction, XAI can help Tesla and vice versa.
You know, we have some collaborative relationships as well, like our material science team, which I think is maybe the best in the world. It is actually shared between Tesla and SpaceX. And that's actually quite helpful for recruiting the best engineers in the world because it's just more interesting to work on advanced electric cars and rockets than just either one or the other. So, like that was really key to recruiting Charlie Komen, who runs the advanced materials team. He was at Apple, and I think pretty happy at Apple, and we were like, well, we could work on electric cars and rockets. He was like, that sounds pretty good. So, he wouldn't take either one of the drawers, but he was willing to take both. Yeah, so I think that is a really important thing.
And like I said, there are some pretty big insights that we're getting to Tesla and trying to understand real world world AI. You're taking, taking video input and compressing that into a vector space and then ultimately into steering and pedal outputs. Yeah. And Optimus? Yeah, Optimus is still at the early stages, but Optimus, and we definitely need to be very careful with Optimus at scale once it's in production. That you have a hard-coded way to turn off Optimus for obvious reasons, I think.
This has got to be a hard-coded ROM local cut off that no amount of updates from the Internet can change that. So we'll make sure that Optimus is quite easy to shut down. It's extremely important because at least of the cars like intelligent, well, at least you can climb a tree or go up some stairs or something, go into a building, but Optimus can follow you in the building. So any kind of robot that can follow you in the building, that is intelligent and connected. We've got to be super careful with safety. Thanks. My problem. Let's see. Thank you.
So one thing I wanted to just talk about before we're concluded is how impactful, sorry about that little feedback is just about the impactfulness of AI as a means of providing equal opportunity to humanity from all walks of life and the importance of democratizing it as far as our mission statement goes. Because if you think about the history of humanity and access to information, there was before the printing press it was incredibly hard for people to get access to new forms of knowledge and being able to provide that level of communication to people is hugely deflationary in terms of wealth and opportunity inequality.
So we're really at a new inflection point in the development of society when it comes to getting everyone the same potential for great outcomes regardless of your position in life. So when we're talking about removing the monopolization of ideas and about controlling this technology from paid subscription services or even worse from the political censorship that may come with whatever capital has to supply these models, we're really talking about democratizing people's opportunities to not only better their position in life but just advance their social status in the world at an unprecedented level in history.
And so as a company when we talk about the importance of truthfulness and being able to reliably trust these models, learn from them and make scientific advancement, make societal advancements, we're really just talking about improving people's qualities of life and improving everyone not just the top tech people in Silicon Valley who have access to it, it's really about giving this access to everyone. And I think that's a mission that our whole team shares.
Before we sign off here, just one last question for Elon, assuming that XCI is successful at voting human level AI or even beyond human level AI, do you think it's reasonable to involve the public and decision making in the company or how do you see that evolving in the long term?
Yeah, as with everything, I think we're very open to critical feedback and welcome that. We should be criticized. That's a good thing. Actually, one of the things that I'd like sort of X slash Twitter for is that there's plenty of negative feedback on Twitter, which is helpful for ego compression. So the best thing I can think of right now is that any human that wants to have a vote in the future of XCI ultimately should be allowed to. So basically provided you can verify that you're a real human and that any human that wishes to have a vote in the future of XCI should be allowed to have a vote in the future of XCI.
Yeah, maybe there's like some normal fee like 10 bucks or something. I don't know. 10 bucks prove you're a human and then you can have a vote. Everyone who's interested. That's the best thing I can think of right now at least.
All right, cool. On that note, we're participating and we'll keep you informed of any progress that we make and look forward to having a lot of great people join the team.