Timestamps:
01:59 Innovation in AI Music Generation
10:18 Disruption of Artistic Expression by AI
16:59 The Future of AI and Creativity
21:31 AI Supercomputers and Audio Technology
28:55 The Impact of Scaling Laws
34:02 Debating the Future of AI Scaling
41:23 AI Models DBRX and Grok 1.5
48:21 Elon Musk and AI Open Source
53:24 Exploring AI Tools and Experimentation
Summary:
In this episode of Sidecar Sync, Amith and Mallory delve into the fascinating world of AI-generated music, discussing the capabilities of Suno AI, a startup that has developed an AI model capable of creating credible, emotional music and lyrics. They explore the implications of this technology on the music industry and the potential impact on human artists. Additionally, they discuss the Stargate AI supercomputer project, a massive $100 billion collaboration between Microsoft and OpenAI to build one of the world's most powerful AI systems. The episode also covers the latest developments in large language models, including Grok 1.5 from Elon Musk's XAI and the open-source DBRX model from Databricks.
Let us know what you think about the podcast! Drop your questions or comments in the Sidecar community.
This episode is brought to you by Sidecar's AI Learning Hub. The AI Learning Hub blends self-paced learning with live expert interaction. It's designed for the busy association or nonprofit professional.
Follow Sidecar on LinkedIn
Other Resources from Sidecar:
Tools mentioned:
Other Resources Mentioned:
More about Your Hosts:
Amith Nagarajan is the Chairman of Blue Cypress (BlueCypress.io), a family of purpose-driven companies and proud practitioners of Conscious Capitalism. The Blue Cypress companies focus on helping associations, non-profits, and other purpose-driven organizations achieve long-term success. Amith is also an active early-stage investor in B2B SaaS companies. He’s had the good fortune of nearly three decades of success as an entrepreneur and enjoys helping others in their journey.
Follow Amith on LinkedIn.
Mallory Mejias is the Manager at Sidecar, and she's passionate about creating opportunities for association professionals to learn, grow, and better serve their members using artificial intelligence. She enjoys blending creativity and innovation to produce fresh, meaningful content for the association space. Follow Mallory on Linkedin.
Disclaimer: This transcript was generated by artificial intelligence using Descript. It may contain errors or inaccuracies.
Amith Nagarajan: Alright, I got the outline coming up.
Alright, I'm good to go if you are.
Mallory Mejias: I am good to go.
Amith Nagarajan: All right, here we go. Greetings and welcome back to the sidecar sink. We have another really exciting episode for you guys today, where we're gonna be talking about some exciting topics at the intersection of A. I. and associations, as always. Uh, particularly, we have some really fun stuff today we'll be starting off with.
I'm here with my co host Mallory, and before we get started, uh, let's hear a quick word from our sponsor.[00:01:00]
Mallory Mejias: Amith, how are you today?
Amith Nagarajan: I am doing great. How are you today, Mallory?
Mallory Mejias: I'm pretty good. We're nearing, or I guess we're fully in spring, really, in New Orleans. I had a call with someone the other day in Minnesota, I think, and they said they had snow last week. And I'm like, you know, in New Orleans, we're full blown spring, enjoying the good weather while it lasts.
Amith Nagarajan: Yeah, I was getting some pictures from some of my friends up in Utah, and they had like, you know, 10 inches of powder, like, three days ago, I think, on the slopes. So I was a little bit jealous, but enjoying the weather out here, too. It was really nice yesterday. Pretty warm.
Mallory Mejias: Absolutely. Well, today, like Amith said, we have an interesting episode lined up for you.
Particularly fun first topic, we'll be talking about Suno AI. Then we'll be talking about the Stargate AI supercomputer. And finally, we'll wrap up with a discussion around Grok 1. 5 and DBRX. So we've got a really interesting episode lined up for you. Diving into topic one, Suno AI. I want you all to take a quick listen to this [00:02:00] cool song we created.
So, believe it or not, Amith and I did not create this song ourselves. Uh, we used AI, like you might have guessed, and we used a tool called Suno. Uh, Suno is a startup that has developed an AI model capable of generating credible, emotional music and lyrics. including a blues song called Soul of the Machine that went viral.
I had a chance to listen to that one myself and recommend that you all check it out. The model collaborates with OpenAI's chat GPT to generate the lyrics, while the Suno model creates the music itself, including realistic sounding vocals and guitars. The technology has sparked debate and controversy around issues like cultural appropriation, the use of copyrighted training data, and the potential impact on human artists.
Music industry experts like Living Color guitarist Vernon Reed have expressed a combination of wonder, shock, and horror at the capabilities of Suno's AI generated music. The Suno AI founders have [00:03:00] ambitious goals to democratize music creation, allowing anyone to make professional sounding songs without any musical training or instruments.
So Amith, you are the one that created this song. So can you talk a little bit about your experience with Suno?
Amith Nagarajan: Yeah, so anyone who knows me at all knows that I have, you know, if I could have negative music, music talent, I'd be there. I'm definitely at zero. And so, if I were to create a song, it would be the stick figure equivalent of music, you know, it would not be very good at all.
And, uh, my kids can attest to that. So, it's, it's, uh, definitely interesting to hear what Suno can create from a simple text prompt. And so, what I did, over the weekend, is, I've been wanting to play with this for the last, I think, three weeks, I've heard about this tool. And, this weekend, I actually started off with something different that we didn't play.
Um, The thing that I ended up with, with Sidecar, I think was the second or third prompt I put in, but, uh, my youngest child was heading off to the beach this weekend to celebrate its spring break here in New Orleans, and, um, they were [00:04:00] celebrating, uh, the, uh, a friend's birthday, uh, at the beach. And, uh, that was an occasion for me to say, Oh, okay, well let me see if I can create a cool song for that kid at their beach party.
And so I went in and put in these different prompts, and I came up with this. really cool song that I was able to then text to, um, to my kid and who played it for all their friends. And it was really, it was, they were like, wow, this is amazing. And it was, it wasn't really like a happy birthday song. It was more of like, uh, you know, we're celebrating you on the beach and it even had like cover art that was beachy and it was, it was pretty impressive.
So then I said, well, what's the, that, that was fun. And you know, for me having no musical skill at all, it was kind of cool to be able to like have an idea. And then turn it into a song, you know, it's kind of like having an idea, turning it into a video with something like, uh, Sora that we've talked about or even like with stable diffusions, video diffusion models and others that are coming when we talk about multi modality and AI, we're talking about, you know, any kind of input and any kind of output.
So it's like [00:05:00] text or audio or video in which we've talked about a lot on the output side. It can be, of course, text, but it can be images, video and now music is another modality. Essentially, right. And it's a super interesting creative outlet for people who don't have skill, uh, in these areas. I also think it could be a really good companion, uh, for artists who are interested in very quickly experimenting with different concepts, which can then lead to their own musical creations.
So I think it's interesting, but we'll talk more about that in a bit. I think the experience was very simple. It's a text prompt. You can say the style of music, you know, what, uh, what the lyrics are going to be. Uh, I even was messing around with multilingual capabilities. So I asked for a French song. All of these kids are in a French immersion school here in New Orleans.
So I'm like, well, And then I didn't send them the French song. And the reason was, is I didn't totally trust the AI. I didn't want to send a bunch of teens a song that had no idea what it was saying. So, I decided to go with the English version. Um, I tried to get the English version with a French [00:06:00] accent. I couldn't figure out how to do that.
But, uh, that was fun. And then I, of course, you know, being a business guy for 30 years, I figured I would try to think about the business application. Since that's when the sidecar, uh, Song came to mind. I'm like, you know, it'd be kind of cool to have like a little sidecar theme song about how important it is to learn AI, which we talk about on this pod so much.
So I put in, I think that was the second try. I put in, uh, not even a paragraph, a couple sentences. I explained what sidecar does. I explained our passion for providing AI learning to associations, nonprofits, and out popped the song we just played.
Mallory Mejias: Wow. I'm assuming the French song you were trying to create was the one about baguettes.
Is that right? Okay.
Amith Nagarajan: Yeah, that one and I tried to create another one that was, um, Something about beach and like water and having fun. And that's where I got a little bit uncomfortable because I'm like, I have no idea what this guy's saying in his song, so.
Mallory Mejias: Fair point, fair point. I think you shared the baguette one with me and I was like, what is it me thinking?
Like, why is he creating a song about baguettes? Maybe you love baguettes. A lot of people do. Um, well, very neat. Would [00:07:00] you say the prompt that you used to create the sidecar song, like how long did it take you, all in all?
Amith Nagarajan: I mean in total across all these different little tests I did I spent maybe 15 minutes in the tool So the sidecar one was the third or fourth one.
I did it took me a minute You know to play with a couple different prompts. It was super quick. It was very fast You know the technology behind Suno, as you described, is a mixture of a large language model in this particular instance right now coming from GPT four and a diffusion model, which is capable of actually generating other modalities outside of text, and then the two interact with each other in an interesting way.
So it's from a technology perspective. It's cool because we have kind of a Uh, you know, a combination right of different tools that are coming together under the hood. And in fact, when we talked about Sora a few weeks ago, which we don't have a lot of details on yet, but we know that that is a transformer diffusion combination as well.
Um, so it's just an interesting time because people are starting to figure out how to take pieces of technology that [00:08:00] exist and recombine them to create new kinds of capabilities, new applications.
Mallory Mejias: So this topic was, this one was a tough one for me, I think, in terms of outlining what we were talking about the episode, because professionally, I think this sounds great.
I realize that we at Sidecar have never experimented with music creation before, and so this opens a whole new wave of opportunity for things that we can do in our marketing efforts. Um, you know, add music to our website. It's really fun on that end. On the personal side, I haven't really talked about this much on the podcast, but I'm also an actor outside of my nine to five job.
I like to think I'm creative all the time in my nine to five and out of my nine to five, but I am a creative at heart and I always have been. And seeing a tool like this, I really agree with the quote I shared in the summary. I have like a mixture of awe and admiration and horror because I realize now that Music is so easy to create.
Um, it's already tough enough to make a living as an artist. So Amith and I have talked about this [00:09:00] before. We do believe art for art's sake will continue in the same way that painting has persisted on for a long time. And theater has persisted on for a long time. However, art. to make a living with a tool like this seems under attack.
And so I want to get your take on this, Amith. I know we like to be AI optimists, but I will say this is a hard one for me to kind of swallow.
Amith Nagarajan: I agree a hundred percent. I don't know what the, I don't know that there is a solution to this other than the fact that there's, you know, this, this type of progress is incredibly disruptive to a lot of people's livelihoods.
Um, I think it opens up the door for creative expression for people who don't have those skills and even for people who do potentially to be a companion to help them accelerate and improve what they do. But the reality is also that a lot of the work that people do in this field, if the AI, and remember, this is AI basically 1.
0 collectively, I know we have models that are like GPT 4 and Gemini 1. 5, but really what we're in right now is AI 1. 0 from the perspective of the consumer's eyes. A. I. S. Of course, 60 years [00:10:00] in the making. We've talked as we talk about in this pot a lot. But the reality is What we have now from like a broad adoption perspective is 1.
0 and we're going to be easing into version two and three. And I mean that kind of like generationally, right? Every, every couple of years. So we have the worst AI today that we'll ever have. And so what we will see soon with video models where you can say, Hey, give me, um, give me a five minute, uh, Intro to sidecar video that combines content from our website, perhaps clips from some of our existing learning assets, but has some original creativity to where you have an A.I. Avatar that's bringing it all together. Maybe that A. I. Avatar breaks into song at some point. You know, there's some really creative, interesting marketing things you could do. Um, the people who would have created that video would no longer be hired by someone like sidecar to create that video. Um, The flip side of that, though, is Sidecar, as a smaller company, would never create a video like that because it would cost, probably traditionally, hundreds of thousands of dollars to create a video like I just described of any, you know, value.[00:11:00]
So, I do think that what will happen is demand will increase, that there will be way more music, there will be way more video, there will be way more demand for videos. for these creative forms of expression because that's happened every time. There's been a step change in access for a variety of different technologies over time.
Uh, even going back to the printing press in terms of the availability of information and how that spread and how there was, there wasn't, people didn't have a hard time like, you know, getting access to books in the sense that they did before that when they were handwritten. Um, because now all of a sudden it was cheap to produce a book and lots of people could have access to that information.
And in a similar way, I think, you know, this broadens and democratizes access. That being said, I don't know what's gonna happen in the near term because there definitely are cases where people will stop using professionals in graphic arts, in, uh, you know, in music, right, uh, who they might have otherwise used.
So think about, like, Ad campaigns for TV. Uh, if I'm an ad exec at a big Madison Avenue agency, and I [00:12:00] have a client who wants to create TV spots, or YouTube shorts, or whatever, um, I have these other assets, these other tools available to create lots of experiments. Um, do I still need musicians and artists?
Maybe, but maybe I need fewer of them. Um, but I also think people who think forward about this will say, Well, actually I want to keep all those amazing professionals in my employment, in my employee. And then I want to be able to dramatically increase the volume and the quality of work that I create.
Whereas you might have only used music a little bit because it was a rare opportunity. That you couldn't really just create. Um, I have no idea what it would cost to create a corporate jingle, right? For a 30 second song with Sidecar. Probably many thousands of dollars, right? Probably a lot. Um, I don't know.
But now it's basically free. So It is both of these things that we're right in the middle of this and we're early in this process So we have a lack of visibility. So I think that's a great quote. We should probably post that on our website It's that mixture of wonder shock and horror which you can have those emotions [00:13:00] simultaneously when in some cases which Um, you know, I think it's, it's kind of like if you go to like, you know, some of those really high end fine dining restaurants, I feel the same way because it's really interesting, but it's also usually pretty terrible.
At least I'd rather have a burger and fries personally. So
Mallory Mejias: exactly a mix of shock, horror and awe. Yeah. I think what it is for me is at least personally, I've taken this stance of, um, being an AI optimist, but at the same time choosing to consume art. That is human created in my mind. That was kind of my thing is moving forward I think people consume art for a lot of different reasons sometimes entertainment.
Sometimes it's really light hearted But for me personally, I like seeing the story I like feeling like that empathy component and seeing stories that are relatable to me represented in music and And film and on tv and whatnot and I guess what's really hard About this particularly is you cannot tell that this song was created by AI.
So it's just it's almost like my personal belief kind of starting to crumble because I won't be able to tell I won't be able to hear a song and be like [00:14:00] that was created by a storyteller human versus an AI model and I think that is just something I've got to unfortunately get over or kind of just figure out a new a new way to go about this.
Amith Nagarajan: One of the interesting ways to think about labor and labor markets is you think about distributions, like the typical bell curve, and that's applicable in a lot of fields, but in, in labor economies, you have a distribution of skill level, right? And so if you think about the broader scope of a particular field, whether it be musicians or visual artists or whatever it is, there's people on the very right hand side of that diagram, which are basically in the top tier.
You know, X percent, very small percentage of their field and people on the other end. But there's the broad middle, probably the 80 percentile, which is most people and those folks, I think, are going to have a choice to make. They either embrace having a eyes, a companion tool to lift their work quality up.
or they're probably going to have a hard time. Um, now I think over [00:15:00] time everyone will embrace using these tools just because the forces of the, of the economics are so hard to go against regardless of what your opinion is. These things exist. You know, it's a similar thing I can, I can relate to a lot more closely is, uh, software development.
So, you know, I've been doing software development for pretty much my entire life and coders now who don't know how to use ai. are basically irrelevant. Uh, coders who are good at using AI are superpowers. You know, they're, they're able to go out there and do so much more. And for a lot of people, actually, who aren't familiar with software development and programming, they might think, well, that's a very different kind of analogy because this isn't software is more of an engineering thing and, you know, Obviously, music and graphics are more artistic, but I will tell you that software development is actually an extremely creative process as well.
Um, and that's actually what attracts a lot of people to it, is it's a mixture of science and art in a lot of ways. Um, but ultimately, the developers who know how to use AI are the ones who are going to be highly employable, and the people who either, Are slow to adopt it or just, you know, kind of push against it or gonna [00:16:00] have a hard time.
I think we might see the same thing here. And that distribution I was referring to, people that are in the middle are most susceptible. Because the people that are on the far end, like, I don't think Beyonce has a lot to worry about. Um, but I think people that are not her or not at that tier, right, that top 5, 10%, are gonna have a really hard time.
If they don't figure out a way to use this stuff. .
Mallory Mejias: Mm-Hmm. . I think the key is what you said, that we've got to think about AI as a companion tool. We've gotta think about opportunities for humans to guide AI in a way that they can create more than they ever had possible. Um, and that's why I'm, I'm super fortunate and happy to kind of be in the position that I am as a creative who's also talking about AI Every week in a podcast, I've told you, Amme, it's kind of, it's a taboo topic when talking to other actors, bringing up ai, and especially talking about the great opportunity it.
Can create and will create in the world. However, I think we should just stress that we need more creatives working with AI all the time. And I will say as a side note, um, SUNO [00:17:00] has only 12 employees, which I thought was pretty small and many of them are musicians. So like that made me happy reading about Suno, I was like, okay, we've got musicians creating this tool.
Hopefully they are working, you know, with a creative interest in mind.
Amith Nagarajan: Well, I think, you know, the number of employees is an interesting thing to think about. The capability of what Suno does is very impressive, regardless of how you feel about it on either end or in the middle. 12 people doing that with probably a really modest amount of capital relative to, you know, some of the bigger AI companies.
Another example, similar to that, we've talked about mid journey on this pot a number of times. I know Mallory, you use, you've used mid journey quite a bit. I think they have like 28 employees and they have over a hundred million dollars in revenue. Um, so, you know, there's a lot of companies out there that have very limited number of people and you'd be, you're shocked by what they can produce.
Part of it is if you're an AI company, guess what? You're really good at using these tools yourself. And so I guarantee you the Suno team is all over using AI in every phase of their work, [00:18:00] independent of what their product does. And that's the key to it. It's like, If someone were to say, Oh, what would it take to create an AI music generator?
I might say, I need a thousand people and a billion dollars in five years. And now all of a sudden, Suno, I don't know how long they've been at it, but probably not super long. And there will be many Sunos. There'll be many other products like this.
Mallory Mejias: It seems like AI generated audio has been lagging behind text and image.
Would you say that kind of an order of least difficult to most complex to create it would be text, image, video, then audio as kind of being the most complicated thing to produce?
Amith Nagarajan: Audio is, is interesting because there's definitely more information in audio than just text, and a lot less information than video.
Video is probably the most complex because you not only have the most information dense modality, but you also have to have an understanding of physics. You have to have an understanding of the, like, what we call the world models that a lot of people are talking about. None of these models have a direct understanding of physics.
They learn physics in the video models and even the [00:19:00] language models to some extent. Through examples. They kind of approximate physics, but they don't have like a rules engine about physics. And so video is harder there. That's why there's a lot of speculation with Sora from open a I that they have some kind of world model in there.
Young Lacoon, who's the head of a I meta talks a lot about the need for a world model or a physics engine that is, you know, trained on that explicitly to really power future video models. And this is also super relevant in robotics because you think about video generation. Think about the flip side, which is video consumption and, you know, Bots that need to exist in the physical world need to have a complete understanding.
We talked about that recently in this pod as well with humanoid robots and the advancements, uh, coming there. But coming back to the question about audio, I actually think audio is a relatively straightforward problem compared, um, to images and video. Uh, it's just a less popular category. So, I'll give you an example, is that 11 labs for some time, Has had extraordinary A.I. audio tools, their audio only, or maybe they have some video now, too, [00:20:00] but I'm pretty sure they started for sure. Audio only, and that's primarily their focus. We've talked about Hey, Jen here a number of times, and they just released their version five studio, which includes a number of new new tools, and they're a video company.
And there's overlap between the two. But 11 labs is actually super popular. It's just not as well known. I just think it's a modality. People don't think about as much. You can do things like A. I. Voice dubbing. Um, so some of our companies are, they want to standardize the professional voice they use for all of their video content.
And so they'll have anyone record a video. So like Member Junction, which is our, uh, open source project for a common data platform we've talked about here. Uh, they have a lot of, there's a lot of people involved in that project and we want to be able to record videos of little demos of the software and how to's and tutorials and it's nice to be able to say, okay, we want to have a consistent brand voice across all those videos.
You can do that very easily with Eleven Labs voice to voice dubbing. Um, and there's there's lots of applications. So I think the technology is actually excellent. There's text to voice or text to audio. [00:21:00] Um, so there's open source models for that, including open. I has something called Whisper, which is a text to voice model, and there's excellent speech recognition model.
So I think the audio modality has, you know, really mature technology available that pretty much is It's able to fool anyone that it's computer generated. Uh, there was recently like, uh, I forget the name of it, but a company last week that announced a public beta and they had an AI audio to audio assistant that understood emotion.
Um, so, and that's something that, um, we'll have to go look it up in the show notes. I forget the company name now, but, um, a lot of people are talking about that and I just think it's like, that kind of, it's kind of obvious that that exists because all the capabilities are there. It's a matter of what we build with these fundamental building blocks.
Mallory Mejias: On that physics note for video, I feel like it's worth diving into just a little bit. Thomas Altman, who leads our Intro to AI webinar that we do every month, um, he was telling me about, or talking to me about Sora and saying, I think there's this video out there of an elephant made out of leaves. Have you seen this video, Amith, [00:22:00] that they released?
I have not. And I was like, I feel like you see it and you just say, Oh, wow, that's really neat. But he was talking to me about how, There had, there had to be an understanding of how an elephant walks and then also the leaves and gravity and how they like flow in the wind. And when you really start to think about that, these text to video models are wild.
Amith Nagarajan: They're incredibly sophisticated and you know how far you can get just with inferred understanding versus an explicit understanding of physics. Um, so I think either way, you know, we're, we're having more and more of these solutions come to market that are clearly more and more realistic. Um, so yeah, there's a lot of fun things to talk about there.
I think audio is an area that associations can really lean into because so much of the content that we have in text format, we can convert to audio. Uh, some people like to listen to their content, obviously as evidenced by our audience for this podcast versus reading a book or a podcast, or sorry, or a blog.
So you know, the modalities of your content can be very fluid now and that's exciting. You can also think about translation, which we've talked about where you go from, you English to some [00:23:00] other language. You can even go from English to Australian English, you know, which could be fun. Um, so there's, there's a lot of reach opportunity, better accessibility for your content.
Um, and so translation isn't just about languages, but it can be from one modality to another. And associations sit on a massive repository of content. So understanding that knowledge, um, you know, language models can help you with a lot of that. Um, but these other models can help you really get that content activated, put them in, put that content in front of your audience in different ways to really engage them.
So that to me is super exciting. And so music, going back to suno, is just another modality where, you know, you could potentially have, uh, songs for all sorts of things. And it's, it's interesting too because we talk about how, Our species is hard wired to, uh, really remember stories well. Um, and also, we remember songs in a different way.
I don't know what the neuroscience is behind it, but sometimes, like my wife's a great example, she remembers, like, every single jingle from every commercial she's ever heard. And [00:24:00] sometimes just breaks into song in the middle of the day, you know? So it's like, which is super fun, but like, I don't know how she remembers all this stuff and when she sings something, I'm like, Oh yeah, I remember that from like the 1980s, you know, um, but there's definitely a different form of memory somewhere in the brain for song and for story.
And so I think that's a really powerful thing to pick up on because if we can translate our message either as marketers or as educators into song, into story, really make it come alive, like think about the story and the song coming into a video to explain something, right? Yeah. To different audiences and animation or like a, you know, full motion video, it just opens up these unbelievable opportunities to educate and inform, uh, folks in, in a variety of ways.
So I, I get really pumped about it because none of that would ever happen with human labor. That stuff is way beyond the scope of what we have. Even if eight billion of us all trained on how to do that, we wouldn't have enough people to do that. So that's where AI scale is really exciting to me. Well, at the same time, I think we have to be cautious about it because what you said earlier really resonates with me that [00:25:00] this is, you know, very shocking too.
Mallory Mejias: Well, I would love to hear from all of our listeners. Are you more excited? Are you more horrified? Are you feeling both? Let us know on the sidecar community or on LinkedIn. We'll have both in the show notes. Moving to topic twos, the Stargate AI supercomputer, Microsoft and open AI are in discussions to build a massive new AI supercomputer data center project called Stargate that could cost over 100 billion.
Stargate is envisioned as the largest and most advanced data center in a series of installations. The company's plan to build over the next 5 to 6 years, the Stargate supercomputer would use millions of specialized server chips to power open AI is next generation AI systems like Microsoft. GPT 5. The project could launch as soon as 2028 and would be over 100 times more expensive than today's largest data centers.
Microsoft would likely be responsible for financing the project, which could require exploring alternative power sources like nuclear energy due to its massive [00:26:00] five gigawatt power needs. The high cost is driven by the need for vast computing power to train advanced AI models, as well as challenges in finding enough specialized AI chips.
The project has not been officially greenlit, and its future may depend on OpenAI delivering on its promise to significantly boost its AI capabilities. The Stargate project represents the enormous scale of investment that major tech companies are pouring into the race for advanced AI capabilities, with Microsoft and OpenAI aiming to build one of the world's most powerful AI supercomputers.
So, Amith, Definitely interested to get your take on this. I feel like talking about these super computers, 100 billion. It's kind of hard to put that into human understanding. So can you kind of bring this back down to earth for us on why this is so important?
Amith Nagarajan: Sure. Well, the first thing I mentioned is 100 billion is obviously an enormous sum of money, but it also represents about eight months of profit for Microsoft.
So it's well within the realm of capabilities that they and other of the large corporations have in [00:27:00] terms of resources, which actually at this point go beyond the scale of what most governments can invest. So that's an interesting conversation by itself is that, you know, the fab seven as they call them in terms of the tech economy driven growth in the stock market.
And, um, you know, it's like how much control and power do they have? That's a separate conversation, but that's, that's why like 100 billion like, Oh, what are you gonna do? How much debt are you gonna raise? Well, they have the cash. Coming in consistently for this and companies like Apple and Google and now NVIDIA are starting to see that although NVIDIA's Cash flow is obviously way smaller than Microsoft or Google But in any event I digress the point is is that it's an achievable financial goal And the theory behind that level of investment is that That scale will solve all ills.
And so there's this so called scaling laws in a I that thus far have held to be true, where, when you add more compute, really compute being the broader term for, you know, compute, memory, storage, etcetera. But when you make your computer bigger, you're able to get emergent capabilities. So you take the same model, you run [00:28:00] more data through its training, you give it more power, and it gives you more power.
Uh, new capabilities you hadn't necessarily predicted. So when you think about what's happened just using the open a I timeline, You think about GPT 1 to GPT 2 to GPT 3 to GPT 4. Four major versions of that model over a number of years, and the amount of increased compute that's been thrown, which has pretty much been like a 10x or an order of magnitude each time they've trained a new model.
And, uh, With that, they've also they've seen these emergent properties where GPT force capabilities were radically greater than GPT threes, even in categories they didn't expect, for example, being, you know, as capable as the typical medical school graduate and passing a medical licensing exam, things like that.
Or even back in the GPT three days when that came out after GPT two, same basic architecture, obviously improved somewhat, but same basic concepts. And it was able to start coding, and that's what led to the GitHub copilot taking the world by storm, which was back in the GPT 3 days. So these [00:29:00] emergent properties are believed that they will continue to, you know, be there to emerge over time as scaling continues.
So if you throw more hardware and more money at it, you'll get better results. Now, that isn't to say that people who believe in the scaling laws don't believe in algorithmic improvement. They don't believe it's not to say that they don't believe in making better models or better training approaches. Um, you know, for example, with training, companies like Mistral don't use RLHF or reinforcement learning with human feedback.
They use something called DPO instead, which is about three to six times more efficient and similar in terms of quality of output. It's one of the reasons Mistral has been so quick and so efficient with capital to get really powerful models out there. They just released Mistral Large about a month ago, and it's very, very close to GPT 4 and Cloud 3.
So There are just another school of thought out there is what I'm getting at that will say, first of all, we're just thrown way too much resourcing at this independent of the money, the environmental impact. of, you know, what you just described is truly horrendous, right? So if you just say, we're [00:30:00] gonna throw as much money as we want to at it, How in the world do you power it?
How do you power it responsibly? Those are big open questions. So can we get an order of magnitude or multiple orders of magnitude of improved efficiency, smarter models, et cetera? You know, Yann LeCun, who I mentioned already on this pod, is someone I like to follow a lot because he's a contrarian in a lot of respects.
He's very big in Uh, he's one of the original AI, you know, godfathers, so to speak, and he's an amazing guy, but he is not as big on, not so much scaling laws, but the idea that current transformer architectures, language models can solve on their own a lot of these problems, and so he believes that there's other architectures that are needed, other approaches that are needed, and I think that probably there's truth in both camps, ultimately.
Um, but the point here is that a project like this at 100 billion of investment over really a handful of years, I think, you know, you said 2028, you know, that's around the corner. And so having assets like that, and by then we'll be on like GPT six or GPT seven, [00:31:00] maybe. Who knows what that looks like, but the point is, is that things are not slowing down, they're speeding up.
That's my main takeaway from this topic, is that, and this is one company, right? Because Microsoft and OpenAI, I kind of think of them as one thing in a way, in terms of the investment side of it. Um, it's one company going after it. There are lots of companies with similar resourcing that are going after it, and tons of other organizations with not quite that level of resourcing, but different ways of thinking about it, going after this problem.
So, the takeaway for associations in my mind, It's not so much the number of gigawatts or the number of dollars or the exact timeline, but it's that the pace of change in AI will accelerate from where we're at, not slow down.
Mallory Mejias: That makes sense. I want to dig more into this idea of the scaling law or scaling laws.
In terminology, a law is something that you know to be true. So you said some people believe in scaling laws. I don't know if that was just the word you chose to use. Is it that some people believe in this and some people don't, or is it that some people say we need to do this at all costs and [00:32:00] some people say we need to move more cautiously?
Amith Nagarajan: So empirically, thus far, the scaling laws have held true. When you throw more resourcing at it, you get, you know, these order of magnitude improvements in performance. The question is whether or not they will continue to hold. Um, it's similar in a way to Moore's Law, the so called Moore's Law, which he never called it a law.
He just observed the fact that transistor density doubled every 24 months for the same price point. And there was always, even back in this day, In the eighties and the nineties, there were always skeptics saying, Oh, Moore's law is dead. It's not gonna hold. And so it's a similar debate in a way. It's like, well, the scaling laws continue to hold here.
We'll be able to get that much more. It's a totally different context, but it's a similar debate. So it's a quote unquote law in the same sense that Moore's law is nothing guarantees that it will continue, but it has performed on that. If you look at the data and you say, Hey, what's the curve show in terms of capability relative to training data set and and therefore compute?
Um, Uh, it very much shows you these capabilities are growing in order, in, in alignment with that so called law.
Mallory Mejias: [00:33:00] Okay. I read that Sam Altman has publicly said that the main bottleneck to, the main bottleneck holding up better AI is the lack of sufficient servers to develop it. You know, based on the conversations you and I have had on this podcast, I think I was thinking we were more in a discovery bottleneck and not really fully understanding like the black box of AI and, and how to create AGI, multimodality, et cetera.
Yeah. Text to music in my mind being a good example because it seems like we had all the pieces there But then it just kind of you know evolved out of nowhere, which maybe it's been around for a minute But that's how it seemed to me at least. So my question is It's about the scaling laws. Do you think that if we had all the servers and chips we needed right now, we would not have any AI questions left?
Amith Nagarajan: I don't believe that that's true. I think there's plenty that we don't know. So I think that just throwing more horsepower at it doesn't mean you're going to solve the problem. I am not really qualified to answer that question because I'm not an AI scientist, but I just don't believe that it makes sense that the current algorithms we have, [00:34:00] which largely they've evolved, but largely they're, you know, they're a little bit old now.
You know, what we're starting to deploy are. Evolutions of the transformer architecture which was a paper from 2017 There's a lot of work that's gone on since then in the last seven years, but we need new models We need new architectures. We need new ways of thinking about it And the beauty of the world we live in is that there are a lot of people who are competing for this, you know, and it's a global competition.
Um, and what we'll see is a lot of innovation. People who don't have the resources that Microsoft and OpenAI have that are forced to figure out, like, in the proverbial garage, you know, how do you build the next great AI model, um, Going back to Yann LeCun again, he points out that a 17 year old can learn how to drive an automobile with a very limited number of hours of practice, right?
So we've all gone through that if we're drivers, and it doesn't take, you know, a million hours of video training data to train our brains on how to safely operate a vehicle. Um, it takes a handful of hours with some, you know, feedback from an instructor, [00:35:00] maybe a parent, and most people can learn how to drive.
Um, and be reasonably safe, right? So, but we need a gargantuan amount of video. So there's, there's a radical inefficiency there if you think about it. Um, and so there will be breakthroughs over time. I don't know when they'll come or who will make them. But I'd be rooting for the underdog and actually kind of betting on the underdog a little bit.
In terms of people who are going after these alternate architectures and in this era where we're just starting to discover like it's almost like the double helix was just discovered and we know what DNA is right and like all of a sudden the that explodes scientific discovery in biology. Similar things are happening in A.I. Right now, and we just don't know. So I think that people who want to throw 100 billion at it have at it. Go for it. I hope they power responsibly in some way. I don't know how you do that, but I also think there's plenty of people who are going after this prize with a very different mindset. And again, follow Mistral.
Those guys are doing some really interesting work. Their models are really good, but that's not what's interesting. It's their approach. That's interesting. It's based on super high [00:36:00] quality and much smaller data sets. It's a different training methodology that I mentioned, DPO instead of RLHF, which we don't want to get into here, but it's basically a different approach that's more efficient, and there will always be this advancement, right?
And so, one of the things that we have to remember is that the people who are the established entrenched players, They kind of want it to require a hundred billion dollars because that protects them. And so if someone can come up with a way of doing it with say a measly one billion dollars or something like that, you know, okay, well, what does that mean for OpenAI?
They have far less of a moat. So they're kind of hoping that that's true. And one of the things that happens to all of us is Is that if we keep saying the same thing to ourselves, we start believing our own stuff and we think it's the only thing even the most brilliant people in the world are susceptible to that kind of confirmation bias from themselves and other biases on the flip side of that.
And so we have to remember that. And there's also this thing where You know, you look at it and think about the classical innovators dilemma, where if you're a large established business, it's hard to disrupt yourself. And in many ways, the scientific [00:37:00] progress in a particular, like, thread that you're going down, it's hard to disrupt that when you've spent all your time and energy and emotional investment in saying, that's what we're going to go do.
So, I think it's going to be super interesting to see what happens, but I think the scaling laws by themselves probably don't get us to AGI. I think it's a mixture of things.
Mallory Mejias: Interesting. Yeah, I was just thinking logically, if compute was the only thing holding us back, it seems like you would see every major giant doing something similar, or in my, I don't know, in my ideal world, I'm like, is there a place where these giants, these big tech companies work together and kind of like create a supercomputer together?
But I'm guessing that's not really the business landscape with me.
Amith Nagarajan: There's stuff we don't know. You know, there's it's a it's a world we don't know a lot about and it's It's kind of like if I said, Hey Mallory, let's go create a supersonic jet together. Let's take an Airbus 380, the largest passenger jet in service in the world, and let's strap on 10 extra engines.
So we'll figure out a way to like throw extra engines on the wing, somehow maintain its airworthiness and not crash. Are we going to get to supersonic speeds? No, we're not, because there's some fundamental issues with the [00:38:00] design of that aircraft that prevent it from achieving that. Um, and so similarly, like, we, and we have an understanding of that now in terms of aeronautical engineering and all those, all those sub disciplines, um, which I know nothing about, but I'm an enthusiast around this stuff.
I think it's one of these things that you look at and say, okay, well, you know, it's the same idea. It's like, we have this, uh, you know, really cool, but also not super advanced model architecture, Let's throw a whole bunch more engines on there in terms of compute and hope that it goes supersonic or hypersonic.
You know, I think, who knows, maybe it'll, maybe it'll work, but I, I just suspect, I have this nagging belief that we need a lot more innovation in the algorithms.
Mallory Mejias: That makes sense. That's a good example to contextualize all this abstract supercomputer stuff. Moving on to topic three, grok 1. 5 and DBRX. XAI, the AI company founded by Elon Musk, has unveiled the latest iteration of its grok large language model called grok 1.
A major upgrade is the expanded context window of 128, 000 tokens, far beyond Grok 1's 8, 000 tokens. This allows GROK 1. 5 to maintain conversations for longer and refer to earlier parts of the discussion. It's unclear for now if GROK 1. 5 carries over the unrestrictive traits of the previous GROK models, which Musk has described as having a rebellious streak and being willing to engage with controversial topics.
Now. On the other hand, we have DBRX. Databricks has introduced DBRX, A new open source general purpose, large language model that outperforms other established open source LLMs like LAMA two mixed role and GR one across the range of benchmarks. [00:40:00] DBRX also surpasses the performance of GPT-3 0.5 on most benchmarks and is competitive with the proprietary Gemini one Pro model.
DBRX is especially strong as a code model outperforming specialized models. like code llama 70 B on programming tasks. The high performance of DBR X is enabled by its fine grained mixture of experts architecture, which provides 65 times more possible expert combinations compared to other open mixture of extra mixture of experts models.
We'll, we'll explain that. Don't worry. Um, this improves the overall model quality. So, Amith, what are your thoughts on these two models? Again, we haven't seen Grok 1. 5, but it might be out by the time everyone listens to this podcast. I'm also not sure if you've tested DBRX yourself?
Amith Nagarajan: I have not. I haven't played with either of these two models, haven't had a chance yet.
But, um, I think the most compelling aspect of this, first of all, builds on the conversation we just had about competition. and alternatives, both having, you know, accessible [00:41:00] models here that you can run locally or run in a in your own private cloud environment, do whatever you want with them. But the most notable thing is, um, the speed and their capabilities at logic and math, which also extends to coding.
So that's a very important domain, because when you are able to get a model to generate reliable, high quality and significant code, you can build on that. You can build a lot of interesting solutions. To give you one example, um, one of our teams here has built an A. I. Called Skip, which is an A. I. Data scientist.
And what Skip basically does is has a chat GPT like conversation with a business user. Business user says, Hey, I'd love to analyze my member renewal trends or I want to understand my event registration data or whatever. Skip has a conversation with that user. And then ultimately what Skip does is is skip writes code.
Skip generates a program and then tests that program against your data and then presents the user with the result, just like a human analyst would. It's really cool. Um, the reason I [00:42:00] bring up skip is that under the hood, um, right now, uh, skip is using primarily open A. I. S. G. P. T. Four model, but it's evaluating a number of other models, including club three, uh, and we'll be definitely looking at these models as well, because code generation as good as it is, is nowhere near perfect.
Um, we, uh, some of our listeners might be familiar with Devin, which was an AI software engineer that was announced about three or four weeks ago. Uh, and then there's OpenDev, which is the open source equivalent of that. And these are basically, uh, tools. They're not models. They're tools that build on top of various models to do software development for you.
And so, if the underlying brain, though, is kind of like GPT 4 esque, which is good, uh, but not fantastic, then Right? It is a limiting factor. And so what you can engineer solution wise is limited. So DBRX and with Grok 1. 5 we have yet additional high quality, low cost options for code generation. And I know they have other capabilities, but I think that their [00:43:00] advances in code generation and math are the most notable Applications that I think are super applicable.
So, uh, to me, those are the things that are exciting. I think that the mixture of experts comment it's almost going to be ignored in the future that everyone will always will use that architectural approach because the M. O. E. Model essentially is this idea of a bunch of specialists working together versus one generalist.
So you can take the most brilliant, trained human that has, you know, all sorts of great capabilities, um, and you ask them to solve a diverse set of problems. They're not going to be as good as a bunch of specialists. Um, and then, of course, those specialists can be fine, fine tuned and focused on particular categories.
And so these MOE models, we talked about this with Mixtral, M I X T A T R A L, from the Mixtral company, Um, a couple months ago, Mixtral is an 8 times 7 billion models, so basically they have 8 different sub models, each of which is 7 billion parameters, kind of like smaller models, [00:44:00] essentially. And it performs at a very high caliber of performance, yet it inferences, meaning it's runtime cost is very low because it's only using typically two of the submodels at a time, and DBRX is very similar to that.
Uh, it's rumored that GPT 4 also uses an MOE model, and they were probably the first commercial. M. O. E. example, but OpenAI doesn't tell you a whole lot about what goes on under the hood in their stuff. Um, so, to me, I think it's just more evidence that acceleration continues, more choice is good for consumers, uh, and will lead to more and more applications being built that will serve the non profit sector.
So I find it exciting.
Mallory Mejias: Do you know if, uh, Gemini uses a mixture of experts architectures? Okay.
Amith Nagarajan: Yeah, Gemini, I, I know Gemini Pro 1. 5 does. I'm pretty sure Gemini 1. 0, Ultra, and Pro do as well, but I'm not 100 percent sure about that.
Mallory Mejias: Well, looking into grok 1. 5, obviously I'm no AI expert, but I will say even reading the article that they put out on their website, it didn't seem all that impressive.
Um, it only outperformed GPT 4 and Gemini 1. [00:45:00] 5 in the human eval benchmark, which I looked up, and that is a dataset designed to evaluate the code generation capabilities of large language models. Um, it did show 100 percent recall and it's 128, 000 token context window, which I thought was Pretty solid.
But anyway, it didn't seem like it was going to be this life changing model. Um, do you foresee Grok being a major competitor in this space? Or do you kind of see this as maybe like a passion project of Elon?
Amith Nagarajan: It is a passion project for Elon, but that's enough to make it material in terms of competitive landscape.
We gotta remember, this is an individual who has, individually, Insane amount of resourcing and it's not just because of financial ability, but it's because he controls Tesla, which has the world world's largest and most capable kind of three dimensional model of the world. They have more video footage, like moving video footage than anyone else on the planet through Tesla, and there's tremendous AI capabilities at Tesla.
Then he has Twitter, which has the probably still are [00:46:00] X, the best real time. information on what's happening around the planet at all times. They have kind of an unparalleled amount of text coming in. Granted, it's a lot of garbage, too, but it's there's a lot of content there. He controls both of those assets.
He has kind of a three dimensional visual world and he has the text world. Uh, and so the combination of those assets could be very, very interesting. He himself obviously is very capable and very determined. This is the same guy who, after making millions of PayPal, uh, Literally threw every dollar he had into going after SpaceX, which we don't really talk about in this podcast.
It's not super relevant to AI, but it's deeply an innovation story where, you know, his thesis was to basically radically reduce the cost of space travel. And so the number one problem with that was that you threw away a rocket as soon as he used it. It's like saying, Hey, Mowry, let's jump on a 747 and fly from here to Europe, and when we get there, let's throw it away, and let's get another plane and fly back.
That sounds absurd, but that's exactly what we've done in the rocket world, and he's like, well, that doesn't make any sense. Let's build reusable [00:47:00] rockets, and lo and behold, he did. It took a lot of determination, Um, love him or hate him again, like, the guy's determined, and he, he barrels through walls that would stop most people, both because he has determination and, and obviously intellectual capability.
Um, so, he is going to be a major player across all of his assets. He's got, the other thing he has going, which we've talked about here, is Neuralink, uh, which is that brain computer interface. So, he has a lot of moving parts. But remember, it's basically Elon Musk, Inc. Which basically means he's able to pull resources across the board from Tesla to x.
ai to x. com to all these other places. However he wants to. There's some interesting corporate governance problems with that since Tesla is a public company, but, uh, but ultimately it is what it is, right? And, and so there will be something competitive coming from Grok. I think Grok might have a shot if, if he, uh, chooses to train it on some of the Tesla video data to be an extraordinary multimodal model, possibly powering robotics.
Cause by the way, back at [00:48:00] Tesla, he has a big robotics piece coming out. They have, uh, I forget what they call it, but their own humanoid robot project, which is. Uh, quite compelling as well. So you have all these pieces going on in Musk's crazy, you know, universe. Uh, he's going to be a competitor. He has the resources to do it, and he's going to compete.
Uh, so to me, and DBRX from Databricks, Databricks is not a company a lot of people know, they're an infrastructure company that powers a tremendous amount of machine learning. And they actually kind of quietly acquired a company called Mosaic ML a number of months back, which was doing some really interesting, uh, generative AI work.
And so they're just another yet another example of someone with a lot of resources and skill coming after this problem.
Mallory Mejias: When I saw DBRX was open source. I immediately looked up to see if, uh, Grok was open source and then I found an article from several weeks ago that said, I think, just in March, they decided to go open source.
And then there were some other articles in the mix that were like, it's not truly open source. Um, I don't know if you have any insight on that, Amith, but with Elon's decision to go open [00:49:00] source, uh, versus keeping it closed.
Amith Nagarajan: I don't know if you've read Walter Isaacson's book on Musk yet. Uh, if you haven't, I'd recommend it.
And it's It's fascinating. You get even people who really dislike the guy. And I'm kind of, I like him and I also don't like him for a lot of different reasons, but I admire what he's accomplished and reading the book was fascinating because if you kind of look at his history and the way that guy makes decisions, um, it's pretty haphazard.
I think there's perhaps a method to the madness, but he pursues what he believes are, you know, the right decisions. And he doesn't really care about collateral damage at any point in time. So that's really scary when you have as much power as he does. Things like Starlink and the impact geopolitically on Ukraine and stuff like that.
It's, that's pretty, pretty crazy in my opinion. But, um, the point I'd make is when we look at someone like that with those kinds of resources, um, I think you're going to see A lot happened quickly because he doesn't care about what anyone thinks. So, the level of innovation that comes from that complete lack of, you know, [00:50:00] care, you could say that it's a judgment flaw or whatever, but like, that's what he's doing.
So, unlike some of these other companies that are much more measured. Now, Elon had this, and there's this lawsuit right now between Elon and OpenAI, because he was the major financial backer of OpenAI when they got started. And he sued them to say that they've gone against what he invested in this, uh, when they were a nonprofit, and he's this big advocate for AI safety, and he talks about open source versus closed source, yet Grok initially was closed source.
And so, you know, very much typical thing. Someone on Twitter tagged him and said, Hey, you know, what are you doing? You're talking about open source, yet Grok is an open source. And I think like the next week, he's like, Oh, yeah, let's make it open source.
Mallory Mejias: That's what it seems like. When I was reading, I read that article about the lawsuit.
And then I just started seeing these articles pop up that was like, yeah, you just decided it was open source or will be very soon. And I guess you're right. That's how Elon kind of works.
Amith Nagarajan: That's how he, that's how he plays the game. So we'll see what happens. Um, you know, I think it, I think there's just, when you have a player like that in the mix, it's harder to [00:51:00] predict what's going to happen.
You kind of have a pretty good idea of what Microsoft's going to do and what Google's going to do at this point. You know, what Elon's going to go do, who knows? So, all, I think it's just going to make things more competitive. Um, so, and his goals, we're not exactly sure what they are, but we know that, you know, AI safety is important to him, at least so he says, yet he's putting out models that are, you know, not quite at the frontier of the field, but will be soon.
So, we'll see, we'll see how that affects what he's, how he approaches this now that he's in that seat.
Mallory Mejias: And if you all want more information too on that open source versus closed source debate, uh, we have a previous episode. I'm not sure what number it is, but one of the earlier episodes we did, we covered that in depth.
Um, I mean, it's kind of my last question for the day. I know we're talking about AI beginners where we often say, just get started, just pick a tool, pick a model, go try it out, see what works for you. Um, for the people who are more intermediate to advanced, maybe that listen to this podcast that are using AI every day, How do you recommend sorting through the [00:52:00] many options out there?
Because I myself, you know, fall into a trap of using chat GPT pretty much all the time. I think you've mentioned the same and I want to make sure that kind of as we get more and more options out there, we're being deliberate about our intentions to try new things and experiment.
Amith Nagarajan: I think that's the most important point, what you just said, which is to try new things and experiment, have a budget, just like you would in a financial budget, a time budget, where once a week, you have an experiment with AI appointment on your calendar for 15 minutes, you know, maybe every Friday afternoon, you do that as an exercise to wind down your day, you just do something you have not done yet with AI, it can be something fun like it.
Uh, the music generation we talked about. Maybe you want to try out Claude. You haven't tried Claude III. You want to play with that. So, I tend to do a lot of that in just the natural flow of my work. I have, uh, a really cool, you know, schedule of what I get to do. I get to play with a lot of different technologies, working across a lot of [00:53:00] teams, talking with a lot of organizations.
So, I kind of, inherent to what I do, uh, end up with that. But most people don't have that level of flexibility, unfortunately. So, I would say budget a meeting with yourself on your calendar, put a recurring appointment on your calendar, new experiment with AI every week for a certain amount of time and go try new things.
Uh, I do tend to float between Claude and chat GPT, and I have been working with Gemini one five for a while, so I think it's got interesting capabilities. Um, I think it's just important to try them all out and that's kind of, One comment from the I would call it the end user perspective. Now, the other thing I would say is for those that are listening that are more on the technical side.
If you're a developer or you're interested in going deeper, all these companies have studios or playgrounds like opening as a playground. Uh, Gemini has a Google A. I. Studio. Claude has something similar. You can go in there and really go deeper and play with these models beyond kind of the veneer of the chat GPT style interface.
[00:54:00] And with those experiments, you can control a lot of different things. You can choose different models. You can test a whole bunch of different system prompts, which are kind of the prompt behind the prompt where you can make the model take on different characteristics. So if you're a little bit more technical, I'd encourage you to check out those studios or playgrounds because you can.
You can learn a lot just by, it's kind of like popping the hood in the car and seeing what's going on in there and learning a little bit about how the car drives. Uh, Mallory, I think, didn't you play around with OpenAI Playground at some point and look in there?
Mallory Mejias: Yeah, I played around with that. I think originally, like way back in 2022, that was what, Thomas Altman, who I already mentioned on this podcast, showed me, um, and I think when we were working on creating blogs out of your book, uh, Ascend, we were kind of working in the playground.
I mean, it's very, anyone could really do it, it just looks a little bit more intimidating than Chatubity.
Amith Nagarajan: Yeah, it's kind of like going into the cockpit of a jet and there's all these controls and things, maybe not quite as intimidating as that, but, um, there's a lot more knobs and levers that you can pull and push, and that's how you learn more about how these models [00:55:00] work.
Uh, it's still very much end user, you're not coding, you're just using a different interface. Um, it is called a playground because it's designed for developers, um, to simulate what would happen if they wrote code that called APIs to do this. So you can kind of work in the playground, say, oh, what would happen if I gave it this system prompt and this user prompt, how would it react?
And that's a very quick way of prototyping something and then saying, oh, okay, I know it'll work this way, now I want to have a developer create a program that calls the A. P. I. Over and over to do this. So a good example would be something like taxonomies. So if I have, let's say, you know, a million documents and I want to create a taxonomy around them, we've talked about that here and in blogs a lot.
How would I do that? Well, I'd pick a model or experiment with a few. I'd go to Claude. I'd go to open a I. I would take in some of my documents. I would say, Okay, I'm gonna have a test system prompt and I'm gonna test out different user prompts. Kind of gets the point where I have an idea of which models produce different kinds of output.
And I'd get that to work. And this is a non developer. This is someone like moderately technical. [00:56:00] Then, if I wanted to automate it with a million documents, I would hire a developer. You know, there's lots of people who you can hire. You can go on Upwork to get freelancers. You can hire companies. You might have in house IT.
And you can say, Hey, developer, this is the set of prompts that we're going to use. I want you to simply write a program in JavaScript or in Python or whatever that uses these prompts, right? So you can separate the programmer from the AI, uh, like architect, if you will. Uh, so that becomes a really interesting skill set for people to pick up on, because you don't really need to be a programmer to learn how these models interact.
And that's one of the best ways to think about like, Oh, maybe we can plug in DBRX. Maybe we can plug in Lama 2. Or maybe we can plug in Mixtral. Um, so you have like all these tools in your tool bench. The question is which one do you pull off the shelf to use depending on the circumstance. It's an amazing time.
It's a little bit complicated, but it's an amazing time to be thinking about building stuff.
Mallory Mejias: For this week's episode, normally I use ChatGPT kind of to do the research and the summaries for the topics that we discussed, but, uh, earlier [00:57:00] this week my uncle actually was talking to me about Perplexity AI, and I've used Perplexity, but I haven't in probably like six to eight months, so I decided to go and use that this week.
And it was awesome. Like I don't know if they've improved it, but it was so efficient I even say I had increased my productivity even from using chat gbt by using perplexity this time and I was like, okay I've got to remind myself stop getting so locked into what I do every day I love the idea to ameeth of like the 15 minutes recurring on your calendar.
Anyone can take 15 minutes I will I think we can all find 15 minutes that we could do some experimenting with
Amith Nagarajan: Yeah, you know, if Google could write a blank check to perplexity and absorb them without, you know, the FTC and other people getting getting on them. And if the companies want to sell, they I think they would do that in a heartbeat.
So I wouldn't be surprised if they got acquired in the next six months. They really have some innovative approaches to blending. Search with generative results and a lot of people are big big fans of perplexity. I haven't used it a ton I've played with it on and off here and there but it's it's a great reminder to go check these tools out if you haven't Been in there in a while.
Some people tell me they've [00:58:00] tried chat GPT and they weren't very impressed Yet they might've done it a year ago. And also they probably were in the free tier instead of the paid version, which in the paid version you get GPT 4 and the free version you get GPT 3 5. It's a totally different experience.
It's like saying I'm riding around on Maserati. And there's a big difference. They both have engines and four wheels, but they do different things.
Mallory Mejias: I love your analogies, Amith. Well, thank you for the conversation today. Hopefully we'll get to create more songs in the future for Sidecar.
Amith Nagarajan: Sounds awesome.
We'll catch you in the next episode. Until then, keep learning, keep growing, and keep disrupting.