Hi everyone, welcome to Gray Matter, the podcast from Gray Lock where we share stories from company builders and business leaders. I'm Heather Mack, head of editorial at Gray Lock. Today, Gray Lock General Partner Reed Hoffman interviews Sam Altman.
Sam is a CEO of OpenAI, an AI research and deployment company whose primary mission is to develop and promote AI technology that benefits humanity. This interview took place during Gray Lock's intelligent future event, a day long summit featuring experts and entrepreneurs from some of today's leading artificial intelligence organizations.
You can watch the video of this interview on our YouTube channel under our intelligent future playlist. And you can listen to other interviews from that summit on the Gray Matter podcast. You can find that on your preferred streaming platform or on the content section of our website, graylock.com.
Sam, close friend, many, many things. I think we actually probably first met, I think, on the street on El Camino bumping into when you were doing looped. Have done a number of things, including a good portion of my nuclear investments are with you. Because you call me and say, hey, this is really cool and I agree.
All right, let's start a little bit more pragmatic, but then we'll branch out. So one of the things I think a lot of folks here are interested in is based off the APIs that very large models will create, what are the real business opportunities?
What are the ways to look forward? And then how, given the APIs will be available to multiple players, how do you create distinctive businesses on them?
怎样才能展望未来呢?而且,由于API将开放给多个参与者,如何在其上创建独具特色的业务呢?
Yeah. So I think so far we've been in the realm where you can do an incredible copywriting business or you can do a sort of education service or whatever. But I don't think we've yet seen the kind of people go after the trillion dollar take on Google's. And I think that's about to happen. Maybe it'll be successful, maybe Google will do it themselves.
But I would guess that with the quality of language models we'll see in the coming years, there will be a serious challenge to Google for the first time for a search product. And I think people are really starting to think about how do the fundamental things change and that's going to be really powerful.
I think that a human level chat bot interface that actually works this time around, I think like many of these trends that we all made fun of were just too early. The chat bot thing was good, it was just too early. Now it can work. And I think having new medical services that are done through that, where you get great advice or new education services, these are going to be very large companies.
I think we'll get multi-model models and not much longer and that will open up new things. I think people are doing amazing work with agents that can use computers to do things for you, use programs. And this idea of a language interface where you say a natural language what you want in this dialogue back and forth, you can iterate and refine it and the computer does it for you.
You see some of this with like Dolly and Copilot in very early ways. But I think this is going to be a massive trend and very large businesses will get built with this as the interface and more generally that like these very powerful models will be one of the genuine new technological platforms which we haven't really had since mobile. And there's always like an explosion of new companies right after. So that'll be cool.
And what do you think the key things are given that the large language model we provided is an API service? What are the things that you think that folks who are thinking about these kind of AI businesses should think about as how do you create them during differentiated business?
So I think there will be a small handful of like fundamental large models out there that other people build on. But right now what happens is company makes large language model API other to build on top of it. And I think there will be a middle layer that becomes really important where I'm like skeptical of all of the startups that are trying to sort of train their own models.
I don't think that's going to keep going. But what I think will happen is there will be a whole new set of startups that take an existing very large model of the future and tune it which is not just fine tuning like all of the things you can do. I think there will be a lot of access provided to create the model for medicine or using the computer or like the kind of like friend or whatever.
And then those companies will create a lot of enduring value because they will have like a special version of they won't have to have created the base model but they will have created something they can use just for themselves or share with others that has this unique data flywheel going that sort of improves over time and all of that. So I think there will be a lot of value created in that middle layer.
And what do you think some of the most surprising ones will be? It's a little bit like for example a surprise a couple of years ago and we talked a little bit to Kevin Scott about this this morning as we opened up which is train on the internet do code. So what do you think some of the surprises will be of you didn't realize it reached that far?
I think the biggest like systemic mistaken thinking people are making right now is they're like all right you know maybe I was skeptical but this language model thing is really going to work and sure like images video too but it's not going to be generating net new knowledge for humanity it's just going to like do what other people have done and you know that's still great that's still like brings the marginal cost of intelligence very low but it's not it's not going to go like create fundamentally new it's not going to go to cure cancer it's not going to add to the sum total of human scientific knowledge and that is what I think will turn out to be wrong that most surprises the current experts in the field.
So let's go to science then there's the next thing. So talk the general tooling that really enhances science. What are some of the things whether it's building on the APIs you know use of APIs by scientists what what are some of the places where science will get accelerated now.
So I think there's two things happening now and then a bigger third one later. One is there are these science dedicated products whatever like alpha fold and those are adding huge amounts of value and you're going to see in this like way more and way more. I think I that I were like you know had time to do something else I would be so excited to like go after a bio company right now like I think you can just do amazing things there.
Anyway, but there's like another thing that's happening which is like tools that just make us all much more productive that help us think of new research directions that sort of write a bunch of our code so you know we can be twice as productive and that impact on like the net output of one engineer or scientist.
I think will be the surprising way that AI contributes to science that is like outside of the obvious models but even just seeing now like what I think these tools are capable of doing copilot as an example you know be much cooler stuff than that that will be a significant like change to the way that technological development scientific development happens.
But then the... so those are the two that I think are like huge now and lead to like just an acceleration of progress. But then the big thing that I think people are starting to explore is I hesitate to use this word because I think there's one one way it's used which is fine and one that is more scary but like AI that can start to be like an AI scientist and self-improve and so when like can we automate like can we automate our own jobs as AI developers very first the very first thing we do can that help us like solve the really hard alignment problems that we don't know how to solve like that honestly I think is how it's going to happen.
The scary version of self-improvement like the one from the science fiction books is like you know editing your own code and changing your optimization algorithm and whatever else but there's a less scary version of self-improvement which is like kind of what humans do which is if we try to go off and like discover new science you know that's like we come up with explanations we test them we think like we we whatever process we do that is like specialty humans teaching AI to do that.
I'm very excited to see what that does for the total like I'm a big believer that the only real driver of human progress and economic growth over the long term is the structure the societal structure that enables scientific progress and then scientific progress itself and like I think we're going to make a lot more of that well, especially science that's deploying technology.
Say a little bit about how what I think probably most people understand with the alignment problem is but it's probably worth four sentences on the alignment problem yeah so the alignment problem is like we're going to make this incredibly powerful system and like be really bad if it doesn't do what we want or if it sort of has you know goals that are either in conflict with ours and many sci-fi movies about what happens there or goals where it just like doesn't care about us that much and so the alignment problem is how do we build AGI that that does what is in the best interest of humanity.
How do we make sure that humanity gets to determine the you know the future of humanity and how do we avoid both like accidental misuse like where something goes wrong we didn't intend intentional misuse where like a bad person is like using an AGI for great harm even if it that's what other person wants and then the kind of like you know inner alignment problems where like what if this thing just becomes a creature that views this as a threat.
The way that I think the self-improving systems help us is not necessarily by the nature of self-improving but like we have some ideas about how to solve the alignment problem in small scale and we've you know been able to align open AIs biggest models better than. we thought we we would at this point so that's good we have some ideas about what to do next but we cannot honestly like look anyone in the eye and say we see out a hundred years how we're going to solve this problem but once the AI is good enough that we can ask it to like hey can you help us do alignment research I think that's going to be a new tool in the toolbox yeah like for example one of the conversations you and I had is could we tell the the the agent don't be racist right and it's supposed to try and to figure out all the different things where they're weird correlative data that exists on all the training settings that everyone else may lead to yeah racist outcomes it could actually in fact do a self-cleansing totally once the model gets smart enough that you can but it really understands what racism looks like and how complex that is you can say don't be racist yeah exactly um
Paragraph 1:
What do you think are the kind of moon shots that in the terms of evolution of the next couple years that people should be looking out for in terms of like evolution and where AI will go yeah um I'll start with like the higher certainty things uh I think language models are going to go just much much further than people think um and we're like very excited to see what happens there um I think it's like uh what a lot of people say about you know running out of compute running out of data like that's all true but I think there's so much algorithmic progress to come um that that we're going to have like a very exciting time um another thing is I think we will get true multimodal models working and so you know not just text images but every modality you'd like in one model you able to easily like uh you know fluidly move between things um I think we will have models that continuously learn uh so like right now if you use GPT whatever it's sort of like stuck in time that it was trained and the more you use it it doesn't get any better and all of that I think we'll get that changed.
Paragraph 2:
So very excited about all of that um and if you just think about like what that alone is going to unlock and the sort of applications people will be able to build with that um that that that that would be like a huge victory for all of us and just like a like a massive step forward um and a genuine technological revolution if that were all that happened um but I think we're likely to keep making research progress into new paradigms as well um we've been like pleasantly surprised on the upside about what seems to be happening and I think uh you know all these questions about like new knowledge generation how do we really advance humanity uh I think there will be systems that can help us with that so one thing I think it'd be useful to share because um uh folks don't realize that you're actually uh making these strong predictions from a fairly critical point of view not just a you know we can take that hill say a little bit about some of the areas that you think are current kind of loosely talked about like for example AI and fusion oh yeah.
Paragraph 3:
So I like one of the unfortunate things that's happened is uh you know AI has become like the mega buzzword um which is usually a really bad sign I hope I hope it doesn't mean like what the field is about to fall apart um but historically that's like a very bad sign for you know new startup creation or whatever if everybody is like I'm this with AI and that's definitely happening now um so like a lot of the you know we were talking about like are there all these people saying like I'm doing like these you know RL models for fusion or whatever and as far as we can tell they're all like much worse than what like you know smart physicists to figure it out um I think it is just an area where people are going to say uh everything is now this plus AI many things will be true.
Paragraph 4:
I do think this will be like the biggest technological platform of the generation um but I think it's like we like to make predictions where we can be on the frontier understand predictably what the scaling laws look like or already have done the research where we can say all right this new thing is going to work and make predictions out from that way and that's sort of like how we try to run the open AI um which is you know do the next thing in front of us when we have high confidence and take 10% of the company to just totally go off and explore um which has led to huge wins and there will be wait like oh I feel bad to say this like I did I that will still be using the transformers in five years I hope we're not I hope we find something way better um but the transform has obviously been remarkable so I think it's important to always look for like you know where am I going to find the next the sort of the next totally new paradigm um and but I but I think like that's the way to make predictions don't don't pay attention to the like AI for everything like you know can I see something working and can I see how predictably gets better and then of course leave room open for like the you can't plan the greatness but sometimes the research break through happens yep.
Paragraph 1:
Today we are fortunate to have Peter Thiel with us. Peter needs no introduction to most of you. He is a true visionary and entrepreneur in the best sense of the word. He's gone after big problems, whether it be with PayPal or Palantir or more recently in the life sciences space. And so it's going to be a great conversation about where things are going broadly with technology, with some focus on the areas that Peter is working on.
Paragraph 2:
I think one of the things that's puzzled me and other people is that the pace of innovation seems to have slowed down over the last few decades. Clearly it seems like we can do a lot of things with our smartphones that we couldn't do 10 or even five years ago. But is it just a slower pace or are we reaching the end of what's possible with digital innovation?
Paragraph 3:
I think one interesting thing is the rate of progress is actually higher than people think, but it's hard for people to measure it, because a lot of it is happening in these enterprise SaaS, enterprise software contexts that people don't really pay much attention to. But it's actually making a huge difference in improving productivity and industry. And I think there are these questions about whether it will directly create any new consumer products, or whether it will just empower other companies to build better consumer products.
Paragraph 4:
The other thing is I think there's this sense that the last 30 or 40 years have been about taking the atoms and turning them into bits. Can we think about the next 20, 30 years as moving from bits back to atoms? And how does that shape what's going on? I think certainly some of the most important technology companies of the last few years have been about physical things - so SpaceX or Tesla, and the physical layer of making the world work in better ways.
Paragraph 1:
So I'm going to uh ask two more questions and then open it up because I want to make sure that people have a chance to do this the broader discussion although I'm trying to paint the broad pictures so you can get the crazy ass questions as part of this um what do you think uh what do you think is going to happen vis-a-vis the application of AI to like these very important systems like for example financial markets um you know because the very natural thing would be to say well let's let's do a high frequency quant trading system on top of this and other kinds of things what what is it is it just kind of be a neutral arms race is it is it what how do you how what what you're thought and like it's almost like the life 3.0 yeah amegas point of view yeah
Paragraph 2:
Um I mean I think it is going to just seep in everywhere my my basic model of the next decade is that uh the cost of intelligence the marginal cost of intelligence and the marginal cost of energy are going to trend rapidly towards zero like surprisingly far and and those I think are two of the major inputs into the cost of everything else except the cost of things we want to be expensive the status goods whatever and and I think you have to assume that's going to touch almost everything um because these like seismic shifts that happen when like the whole cost structure of society change which happened many times before um like the temptation has always done to estimate those uh so I wouldn't like make a high confidence prediction about anything that doesn't change a lot or that where it doesn't get to be applied um but one of the things that is important is it's not like the thing trends either trends all the way to zero they just trend towards there and so it's like someone will still be willing to spend a huge amount of money on computing energy they will just get like unimaginable amount intelligence energy they'll just get unimaginable amounts about that and so like who's going to do that and where's it going to get the weirdest not because the cost comes way down but the amount spend actually goes way up yes the intersection of the two curves yeah you know the thing got 10 or think got a hundred times cheaper in the cost of energy you know a hundred million times cheaper in the cost of intelligence and I was still willing to spend a thousand times more into days dollars like what happens then yep
Paragraph 3:
And then uh last of the buzzword bingo part of the the future questions metaverse an AI what do you what do you see coming in this you know I think they're like both independently cool things it's not like totally clear to me yeah other than like how AI will impact all computing yeah well obviously computing simulation environments agents possibly possibly entertainment certainly education right um you know like an AI tutor and so forth those those would be baseline but the question is is there anything that's occurred to you that's I I would bet that the metaverse turns out in the upside case then which I think has a reason which is happening the upside case the metaverse turns out to be more like something on the order of the iPhone like a new a new container for software and you know a new way a new computer interaction thing an AI turns out to be something on the order of like a legitimate technological revolution um and and so I think it's more like how the metaverse is going to fit into this like new world of AI than AI fit into the metaverse but low confidence TBD all right questions
Hey there, how do you see uh technologies, uh foundational technologies like tpc3 affecting uh the pace of life science research? Specifically, uh, you can group in medical research there and sort of just quickening the iteration cycles. And then what do you see as the rate limiter in life science research and sort of where we won't be able to get passed because they're just like laws of nature yeah some like that?
Um, I think the current leo available models are kind of not good enough to have like made a big impact on the field, at least that's what like most like life sciences researchers have told me. They've all looked at it and now I guess a little helpful in some cases. Um, there's been some promising work in genomics but like stuff on a bench top hasn't really impacted it. I think that's going to change and I think uh, they, this is one of these areas where there will be these like you know new 100 billion to trillion dollar companies started and those those areas are rare but like when you can really change the way that if you can really make like a, you know, future of pharma company that is just the hundreds of times better than what's out there today, that's going to be really different.
Um, as you mentioned there still will be like the rate limit of like bio has to run its own thing and human trials take over long they take and that's so I think an interesting cut of this is like where can you avoid that? Like where are the synthetic bio companies that I've seen that have been most interesting are the ones that find a way to like make the cycle time super fast. Um, and that benefits like an AI that's giving you a lot of good ideas but you've still got to test them which is where things are right now. Um, I'm a huge believer first startups that like the thing you want is low costs and fast cycle times and if you have those you can then compete as a startup against the big incumbents. Uh, and so like I wouldn't go pick like cardiac disease as my first thing to go after right now with like at this kind of new kind of company, um, but you know using bio to manufacture something that sounds great.
Uh, I think the other thing is the simulators are still so bad and if I were an if I were a biomeats AI startup I would certainly try to work on that somehow when you think the AI tech will help create itself it's almost like a self-improvement will help make the simulator significantly better. Um, people are working on that now, uh, I don't know quite how it's going but you know there's very smart people are very optimistic about that. Yup, other questions and I can keep going on questions I just want to make sure you guys had a chance of this, uh, here yes great.
Um, Mike is coming haha awesome thank you, um, I was curious what what aspects of life do you think won't be changed by AI? Um, sort of the all of the deep biological things like I think we will still really care about interaction with other people like we'll still have fun and like the reward yeah, you know systems of our brain are still going to work the same way like we're still going to have the same like drives to kind of create new things and you know compete for silly status and like you know form families and whatever. Um, so I think the stuff that people cared about 50,000 years ago is more likely to be the stuff that people care about, you know, 100 years from now than 100 years ago. As they amplify on that before we get to the next whatever the next question is.
what do you think are the best utopian science fiction universes so far good question um starship is pretty good honestly uh like I do like all of the ones that are sort of like you know we turn our focus to like exploring and understanding the universe as much as we can it's not this is not a utopian one maybe I think the last question is like an incredible short story uh-huh yeah it was what that came up mine yep uh I was expecting you to say he invokes on the culture those are great uh I think science fiction is like there's not like one there's not like one sci-fi universe that I could point to and say I think all of this is great but like the uh high optimistic corner of sci-fi which is like a smallish yeah corner um I'm excited about actually uh I took a few days off to write a sci-fi story and I had so much fun doing it just about sort of like the optimistic case of AGI um that it made me want to go like read a bunch more so I'm looking for recommendations of more to read now um like the sort of less known stuff you have anything I will I will get to some great some recommendations
hi so in a similar vein one of my favorite sci-fi books is called Childhoods End by Arthur Clark from like the 60s I think and the I guess the one sentence summary is aliens come to the earth try to save us and they just take our kids and leave everything else so you know I are slightly more optimistic than that but yes I mean there's ascension into the overmind is is is meant to be more utopian but yes okay uh you may not read it that way but yes well also in our current universe our current situation um you know a lot of people think about family building and fertility and like some of us have different people have different ways of approaching this but from where you stand what do you see as like the most promising solutions it might not be a technological solution but I'm curious what you think other than everyone having 10 kids you know like how do we have everyone having 10 kids yeah um how do you populate how do you like how do you see family building co-existing with you know AGI high tech it's this is like a question that comes up at Openair a lot like how do I think about you know how should one think about having kids there's I think no consensus answer to this um there are people who say yeah I'm not I was gonna I thought I was gonna have kids and I'm not going to because of AGI like there's just for all the obvious reasons and I think some less obvious ones there's people who say like well it's gonna be the only thing for me to do and you know 15 20 years so of course I'm gonna have a big family like that's what I'm gonna spend my time doing you know I'll just like raise great kids and then I think that's what will bring me fulfillment I think like as always it is a a a personal decision I get very depressed when people are like I'm not having kids because of AGI the EA community is like I'm not doing that because they're all gonna die that kind of like techno op-mists are like well it's just like you know I want to like merge into the AGI and go off exploring the universe and it's gonna be so wonderful and I you know just I want total freedom but I think like all of those I find quite depressing I think having a lot of kids is great I you know want to do that now more than I did even more than I did when I was younger and I'm excited for it.
Paragraph 1:
What do you think will be the way that most users interact with foundation models in five years do you think there'll be a number of verticalized AI startups that essentially have adapted and fine to you and foundation models to an industry or do you think prompt engineering will be something many organizations have as an in-house function?
Paragraph 2:
I don't think we'll still be doing prompt engineering in five years I think it'll just be like you and this will be integrated everywhere but you will just like you know either with text or voice depending on the context you you will just like interface in language and get the computer to do whatever you want and that will you know apply to like generate an image where maybe we still do a little bit of prompt engineering but you know it's kind of just gonna get it to like go off and do this research for me and do this complicated thing or just like you know be my therapist and help me figure out how to make my life better or like you know go use my computer for me and do this thing or or any number of other things but I think the fundamental interface will be natural language.
Paragraph 3:
Let me actually push on that a little bit before we get to the next question which is I mean to some degree just like we have a wide range of human talents right now and taking like for example a dolly when you have like a a great visual thinker they can get a lot more out of dolly because they know how to think more than how to iterate the loop through the the test don't you think that will be a general truth about most of these things so it isn't that while it will be natural language is the way you're doing it it will be there will be like almost an evolving set of human talents about about going that extra mile.
Paragraph 4:
100% I just hope it's not like figuring out like hack the prompt by adding one magic word to the end that like changes everything else I like what will matter is like the quality of ideas and the understanding of what you want so the artist will still do the best with image generation but not because they figured out to like add this one magic word at the end of it because they were just able to like articulate it with a creative eye that you know I don't have certain. They have as a vision and kind of how their visual thinking and iterating through it yeah yeah no obviously it'll be that word or prompt now but it'll iterate to better all right at least we have a question here.
Paragraph 5:
Hey thanks so much I think the term AGI is used thrown around a lot and sometimes I've noticed my own discussions like the sources of confusion has just come from people having different definitions of AGI and so it can kind of be the magic box where everyone just kind of projects their ideas onto it and I just want to get a sense for me would like how do you think you know how would you define AGI and how do you think you'll know when we find that early.
Paragraph 6:
It's a great point I think there's like a lot of valid definitions to this but for me AGI is basically the equivalent of a median human that you could like you know hire as a coworker so then they could like say do anything that you'd be happy with a remote coworker doing like just behind the computer which includes like you know learning how to go be a doctor learning how to go be a very competent coder like there's a lot of stuff that a median human is capable of getting good at and I think one of the skills of an AGI is not any particular milestone but the the meta skill of learning to figure things out and that it can go decide to get good at whatever you need so for me like that's that's kind of like AGI and then super intelligence is when it's like smarter than all of humanity put together.
Paragraph 1:
So we have, do you have a question? Yep, great. Thanks. Just what would you say are in the next 20, 30 years are some of the main societal issues that will arise as AGI continues to grow and what can we do today to mitigate those issues?
Paragraph 2:
Obviously the economic impacts are huge and I think. it's like if it is as divergent as I think it could be for like some people doing incredibly well and others not I think society just won't tolerate it this time and so figuring out when we're going to like disrupt so much of economic activity and even if it's not all disrupted by 20 or 30 years now I think it'll be clear that it's all going to be.
Paragraph 3:
What is the new social contract? My guess is that the things that we'll have to figure out are how we think about fairly distributing wealth access to AGI systems which will be like the commodity of the realm and governance like how we collectively decide what they can do what they don't do things like that.
Paragraph 4:
I think figuring out the answer to those questions is going to just be huge. I'm optimistic that people will figure out how to spend their time and be very fulfilled. I think people worry about that in a little bit of a silly way. I'm sure what people do will be very different but we always solve this problem.
Paragraph 5:
But I do think like the concept of wealth and access and governance those are all going to change and how we address those will be huge. Actually one thing I don't know what level of Debs you can share that but one of the things I love about what OpenAI and you guys are doing is when you think about these questions a lot themselves and they initiate some research.
Paragraph 6:
So you've initiated some research on this stuff. Yeah so we run the largest UBI experiment in the world. I don't think that is we have a year and a half year and a quarter left in a five year project. I don't think that's like the only solution but I think it's a great thing to be doing. And I think we should have like ten more things like that that we try. We also try different ways to get input from a lot of the groups that we think will be most affected and see how we can do that early in the cycle. We've explored more recently like how this technology can be used for rescilling people that are going to be impacted early. We'll try to do a lot more stuff like that too. Yeah so they are the organization is actually in fact these are great questions addressing them and actually doing a bunch of interesting research on it. So next question.
Paragraph 7:
Hi so creativity came up today in several of the panels you know and it seems to me that the way it's being used like you you have tools for human creators and going to expand human creativity. So where do you think the line is between these tools to allow a creator to be more productive and artificial creativity is taking to the creator itself.
Paragraph 8:
So I think and I think we're seeing this now that tools for creatives that is going to be like the great application of AI in the short term. People love it it's really helpful and I think it is at least in what we're seeing so far not replacing it is mostly enhancing. It's replacing in some cases but for the majority of like the kind of work that people in these fields want to be doing it's enhancing and I think we'll see that trend continue for a long time. Eventually yeah it probably is just like you know we look at a hundred years okay it can do the whole creative job.
Paragraph 9:
I think it's interesting that if you asked people 10 years ago about how AI was going to have an impact with a lot of confidence from almost most people you would have heard you know first it's going to come for the blue collar jobs working the factories truck drivers whatever then it will come for the kind of like the low skill white collar jobs then the very high skill like really high IQ white collar jobs like a programmer or whatever and then very last of all and maybe never it's going to take the creative jobs and it's really gone exactly the and is going exactly the other direction and I think this like isn't there's an interesting reminder in here generally about how hard predictions are but more specifically about you know more not always very aware maybe even ourselves of like what skills are hard and easy like what uses most of our brain and what doesn't or how like difficult bodies are to control or make or whatever we have one more question over here.
Paragraph 1:
hey thanks for being here so you mentioned that you would be skeptical of any startup trying to train the old language model and it would look to understand more
Paragraph 2:
so what I have heard and which might be wrong is that last language models depend on data and compute and any startup can access to the same amount of data because it's just like internet data and compute like different companies might have different mouth compute but I guess like the big players can sell my compute
Paragraph 3:
so how good a large language model startup differentiate from another how would the startup differentiate from another how would one large language model start up differentiate from I think it'll be this middle layer
Paragraph 4:
I think in some sense the startups will train their own models just not from the beginning they will take like you know base models that are are like hugely trained with a gigantic amount of compute and data and then they will train on top of those to create you know the model for each vertical and end it those startup
Paragraph 5:
so in some sense they are training their own models just not not from scratch but they're doing the 1% of training that really matters for for whatever this use case is going to be those startups I think there will be hugely successful and very differentiated startups there
Paragraph 6:
but that'll be about the kind of like data flywheel that the startup is able to do the kind of like all of the pieces on top and below like this could include prompt engineering for a while or whatever the sort of the kind of like core base model
Paragraph 7:
I think that's just going to get to two complex and two expensive and the world also just doesn't make enough chips
我觉得这会变得太复杂、太昂贵,而且世界上也没有足够的芯片。
Paragraph 8:
so Sam has a work thing he needs to get to so and as you probably can tell with a very far far ranging thing Sam always expands my batteries and a little bit unlikely that when you're feeling depressed whether it's kids in the house you're the person I was turned to for I appreciate that yes
Paragraph 9:
so anyway I think I think like no one knows like we're sitting on this like precipice of AI and like people like it's either going to be like really great or really terrible you may as well like you got to you got to like plan for the worst you certainly like it's not a strategy to say it's all going to be okay but you may as well like emotionally feel like we're going to get to the great future and work hard as you can to get there and play for it yes rather than like act from this place of like fear and despair all the time because if you if we acted from a place of fear and paranoia we would not be where we are today
Paragraph 10:
so let's thank Sam for spending dinner with us thank you that concludes this episode of Gray Matter you can read a transcript for this interview on the content section of our website graylock.com slash blog and you can watch a video from this interview on our YouTube channel under the intelligent future playlist you can subscribe to graynet on your preferred streaming platform and you can follow us on twitter at graylock vc