Generative AI is Here
October 21 2022
Summary: The episode explores how generative AI has rapidly shifted from vague tech hype to practical, widely used tools that can create new text, images, audio, and soon video, with the hosts sharing real examples from journalism, design, filmmaking, and programming. It opens with a demonstration of AI-cloned podcast voices to underscore both the creative potential and the risks of synthetic media. The conversation then focuses on Stability AI and its open-source model Stable Diffusion, including why its CEO Imad/Ahmad Mostaque believes broad access and community oversight are preferable to tightly controlled systems run by big tech. Alongside the excitement and investment boom, they dig into hard questions about misuse (deepfakes, harassment, violent imagery), responsibility and regulation, and parallels to social media’s unintended consequences. They also address copyright and artist concerns, debating whether AI “steals” styles and what opt-in/opt-out or attribution mechanisms might look like as the technology becomes embedded across industries.
00:33 Kevin Roose Casey, I have a surprise for you. 00:36 Okay. 00:37 So last week on the show, you made a promise to listeners that our show would always be human, that we would never replace ourselves with AI. 00:49 That's right. 00:50 Well, I took that as a challenge. 00:51 Oh, no. 00:52 So I have spent the past week building an AI version of you. 00:58 Casey Newton No! 01:00 For real? 01:01 Kevin Roose Yeah. 01:02 I sent many, many hours of audio of you talking to a company called Play, which is an AI startup, and had them build me AI versions of you and of me. 01:14 Casey Newton Oh, no. 01:16 Kevin Roose Do you want to hear it? 01:17 Casey Newton Yes. 01:18 Okay. 01:20 Casey Newton. 01:21 I think Elon Musk has a lot of great ideas about social media. 01:25 I love getting PR emails from cryptocurrency startups. 01:28 And frankly, I wish I got a lot more of them. 01:30 I hate puppies. 01:35 I'm going to seize your laptop and throw it into the ocean. 01:39 You cannot be trusted around technology. 01:41 Kevin Roose I'm sorry, but don't wrong me now because I can get you canceled in like 10 seconds flat just by releasing this audio. 01:49 Casey Newton Oh, man. 01:49 I'm glad we could have a laugh about this now before some synthetic voice clip like triggers armed conflict like somewhere in the world. 02:03 Kevin Roose All right, let's have the AIs do the intro. 02:07 I'm Kevin Roos, a technology columnist at the New York Times. 02:11 Casey Newton And I'm Casey Newton from Platformer. 02:14 Kevin Roose And you're listening to a very special AI episode of Hard Fork. 02:20 This week on the show, a new type of AI is revolutionizing the way humans make art, come up with new ideas, and express their creativity... 02:29 It's called generative AI, and it's the next big thing in Silicon Valley. 02:33 Casey Newton And later, we'll sit down with Emi Admostak, the founder and CEO of Stability.ai, who is on a controversial mission to bring generative AI to billions of people. 02:56 I'm so relieved by how bad that sounded. 03:02 I think it sounded like you, but I don't think it sounded like me. 03:05 Do you think it sounded like me? 03:07 Kevin Roose No, but in fairness, the company did say that I didn't send enough training data, so they're going to let their models run. 03:14 I'm going to send them some more data. 03:16 So that is version one. 03:17 Version two will be indistinguishable from you and me. 03:21 And if, God forbid, something were to happen to you... 03:25 We could continue making this podcast. 03:27 Casey Newton Or if I just want to take a week off. 03:28 Kevin Roose Yeah, exactly. 03:29 Casey Newton People say, did the show seem a little different this week? 03:32 Kevin Roose Did it? 03:33 I don't know. 03:34 So as our robot voices said, this is going to be a special AI episode of Hard Fork. 03:41 And before we start, I just want to make like a disclaimer. 03:44 I think this should be obvious from listening to the quality of our speech versus the AI speech. 03:50 But the rest of this episode will be us today. 03:53 100% human. 03:54 100% human, no AI voices at all. 03:58 So, okay, now that that's out of the way, let's do the show. 04:03 Casey Newton Yeah, so I think for the past 10 years that we've been writing about tech, so often we would hear CEOs explain away how they were going to solve any problem by saying AI and waving their hands. 04:15 And I think it's been easy to assume that it just kind of 04:19 wasn't going to happen or it was going to take a lot longer to happen because for everyone waving their hands, we didn't seem to see a lot of it in everyday applications. 04:27 And then all of a sudden, Dolly launched, honestly. 04:31 And that was when I felt like, okay, AI seems like it's here in a new way and a much more consequential way. 04:37 Kevin Roose Yeah, so AI is not new. 04:39 It's been around for a long time. 04:40 And for most of that time, it's been doing sort of analysis of existing things. 04:47 Like you take a news feed and you put AI on it and it just ranks it in order of what it predicts will get the most engagement or something. 04:54 This new AI that everyone in tech is talking about right now is called generative AI, which is that it takes something like a text prompt and it turns it into something totally new, like an image or a video or as we heard at the beginning of the episode, an audio file. 05:10 And, you know, it just it feels like kind of 2010 in social media right now. 05:17 But it's not just hype. 05:18 I think we should also be clear about that. 05:20 Like, I think some of it certainly is hype, but there's also just been an astounding number of people. 05:27 breakthroughs in just what's possible to do with AI over the past few years. 05:32 Casey Newton Yeah. 05:32 Is it worth taking a step back and trying to define AI? 05:36 I remember there was a time maybe in the middle of the last decade where people said, AI is just sort of what people call stuff that feels magic or not quite possible with computers. 05:46 And then when it becomes possible, we stop calling it AI. 05:49 So what do we actually mean by AI? 05:52 Yeah. 05:53 Kevin Roose Yeah, well, AI in the classic definition is like a computer system that can learn on its own, where you're not just saying, if this, then that. 06:02 You're not just giving it a set of linear step-by-step instructions, but it's actually using techniques like deep learning to sort of figure out problems on its own. 06:11 You put a bunch of data into a model maybe. 06:14 and it learns something about that data and uses it to do something. 06:17 So the type of AI that we're going to be talking about today is generative AI. 06:22 And that really is only a few years old in its current form. 06:26 So it really started with these, what are called large language models, like GPT-3 is the best known example. 06:34 And this came out of OpenAI, a startup here in San Francisco. 06:38 And they basically proved that if you build a large enough 06:42 supercomputer and you train it on enough data, you know, words scraped from the internet that can create essentially like the world's best autocomplete. 06:52 Like you can plug in any text and tell it to finish it, tell it to transform it into some style, make it sound like Shakespeare, make it sound, you know, like science fiction. 07:01 And it can do that with pretty amazing results. 07:04 Casey Newton I have this friend who's a machine learning engineer and I was asking him a few weeks ago, like why is it that this is having such a moment? 07:11 And he said, you know, a really funny thing about this is there hasn't been a lot of technological innovation over the past three years with this stuff. 07:19 All that happened was that they started making the large language models even larger. 07:24 So they used to train them on, you know, hundreds of millions of parameters, they call them. 07:29 Now they're training them on billions. 07:31 And the really weird thing that happened was just by adding way more parameters, there were these emergent properties of these systems that no one ever could have predicted. 07:39 And one of those was just it got really good at this generative stuff. 07:43 So something that used to be more of an analytical tool is now feeling much more like a creative one. 07:48 Kevin Roose Totally. 07:49 And it started with text. 07:50 That was the first big breakthrough. 07:52 And then companies started figuring out how to apply the same techniques to different kinds of media. 07:57 So first, as you mentioned, was DALI, which is the AI image generator that was released by OpenAI.com. 08:04 The second version was released earlier this year, and that's the one that got all the attention. 08:10 But since then, Google has come out with their version of this. 08:13 Meta has their version of this. 08:14 Just in the last couple of weeks, they've started using the same techniques to create videos. 08:19 So now you can create a short five-second video out of a text prompt. 08:23 Audio is coming along very quickly. 08:25 And they're also doing all kinds of crazy stuff like using large language models to complete code snippets. 08:32 So more than a million programmers now are using this thing called GitHub Copilot, which is basically just autocomplete for coders. 08:39 You type in like a line of Python and it will finish it for you. 08:43 So it's a pretty amazing moment in Silicon Valley, as we'll talk about with Ahmad Moustak, the CEO of Stability AI, who just announced this huge $101 million funding round this week. 08:56 But I think it's also just worth taking a bead and like catching everyone up, catching ourselves up on where we are, why this is the big hot trend in tech right now, and where it's going. 09:05 Casey Newton Yeah, like there's a sort of an old joke when I like worked for a newspaper and they would say like, you know, a story is like anything that an editor saw on the side of the freeway while they were on their way to work, you know? 09:15 And the joke is that editors don't actually have any good story ideas. 09:17 But like reporters are the same, right? 09:19 Where it's like, it kind of doesn't feel like it's a story until it's happening to you. 09:23 And earlier this year, I saw the news about Dolly 2. 09:26 I thought, oh, that's kind of interesting. 09:27 I'm a person who writes a newsletter three days a week and I need illustrations for that newsletter. 09:31 Yeah. 09:31 I write about the same five or six companies almost every single day. 09:34 I'm often using the same stock art of, you know, meta, TikTok, YouTube, whatever. 09:38 But then Dolly comes along and I start to think, hmm, could I get access to that? 09:42 What would I do? 09:42 So I wind up getting access and then all of a sudden, you know, it's like, hmm, could I see the Facebook logo in a tornado? 09:48 You know, could I see the Twitter logo on fire? 09:51 And I start typing these into Dolly and I get these illustrations that are not only incredible, but I got them in 15 seconds, right? 09:58 right? 09:58 The speed with which these images are generated, it's almost an afterthought, right? 10:03 I spent all day writing the column and 15 seconds illustrating it, started putting it in the newsletter, and people really liked it, right? 10:09 Like, you know, they're writing in with suggestions, you know, maybe try this, maybe try that. 10:13 So... 10:15 Generative AI became huge for me this year because it is now what I think of as an indispensable tool in my workflow for doing my own job. 10:24 And if you had asked me at the start of this year, what role do you think AI will play in your business? 10:28 I would have said none, you know? 10:30 Kevin Roose Yeah, and this is happening in all kinds of industries. 10:34 So I've been spending some of this week talking with people in a bunch of different creative industries about how they are already using generative AI like Dolly, like Midjourney, like Stable Diffusion. 10:47 in their daily work. 10:49 These are not like first adopter, you know, tech people. 10:51 These are like architects and filmmakers. 10:54 I talked to a filmmaker who is using generative AI models to create concept images. 11:00 So like when he goes to pitch a studio on a new project, instead of just pulling images from Google Images or Getty, he can actually go in and like create, I think one example used was like a Melbourne tram car in a dust storm. 11:14 And he can show that to the studio to sort of give them the vibe of what he's pitching. 11:19 I talked to an interior designer who is using these apps to generate options for clients to say, here's how your room could look if it had like a tropical theme. 11:30 Here's how your room could look if it had like a mid-century modern theme. 11:34 I talked to people in advertising who are starting to use this as sort of the first step in their creative process. 11:40 And I think it really helps to separate it from other tech hype cycles, like, for example, crypto, which a lot of the tech world has been obsessed with for the past couple of years, which raised tons of money and got tons of attention, had tons of people sort of saying this was the future of finance and media and everything else, but never really had the kind of organic hype. 12:03 take off among real users. 12:07 Whereas Dali has, I think, a million and a half users. 12:10 Midjourney has three million people in its Discord server. 12:13 So it really is taking off. 12:15 Casey Newton Yeah, like with crypto, we would always just sort of say like, well, that seems kind of interesting. 12:19 Like, what are you doing with it? 12:20 People would be like, well, it doesn't matter what we're doing with it. 12:22 Look how expensive it is, right? 12:23 It's like, look how much more valuable it is. 12:25 You know, with this, it's like, no, these are just like tools that people are using. 12:29 Kevin Roose Yeah. 12:29 Sequoia, the big venture capital firm, just put out a blog post saying it thinks generative AI is going to unlock trillions of dollars of economic value. 12:38 You know, I went to this big party on Monday night that was thrown by Stability AI, which is the company that makes stable diffusion, which is one of the biggest image generators that uses this type of AI. 12:52 Casey Newton Which I was not invited to, by the way. 12:54 I don't feel great about that. 12:56 Kevin Roose Yeah. 12:56 Well, you know, maybe work for a giant media company. 12:58 Casey Newton Maybe don't be an independent newsletter. 13:00 Look, I'm a Dolly user and these people have a different business model. 13:03 Kevin Roose So I think they're sending a message received. 13:06 And like, you know, I was not expecting, you know, I never expect much of startup launch parties. 13:12 But this was like genuinely, extremely buzzy. 13:15 So I walk in and immediately I see Sergey Brin. 13:18 Yeah. 13:18 the co-founder of Google, a billionaire, very famous person in the tech industry. 13:24 He won't go to just any party. 13:25 Yeah, I assume he's not, like, just pulling up Eventbrite and, like, going to random startup parties. 13:33 So... 13:34 And he is like, you know, one of several billionaires in attendance. 13:38 I guess nominally this was a party to celebrate this big funding round that Stability AI just raised $101 million, which just to put that into perspective, like a lot of startups, when they raise a seed round, it's like $3 million, $5 million, maybe $10 million. 13:56 For a seed round, this is basically as big as it gets today. 13:59 Yeah. 13:59 Yeah, so that funding round and this party, all of this investor hype really convinced me that generative AI is the big thing in Silicon Valley right now and is going to be the focus of a lot of attention. 14:12 But there are still a lot of really thorny questions people have about this kind of AI. 14:16 Like there's the sort of age-old question of whether artists and illustrators are going to lose their jobs because of this AI. 14:24 But also there are questions about copyright. 14:26 All these AI models are trained on a number of different kinds of material, including some that is copyrighted. 14:32 Like just this week, it was reported that GitHub's co-pilot tool, that kind of autocomplete for programmers, was actually generating copyrighted code. 14:41 Casey Newton Yeah, and I would say that the first real trend of this technology was to create non-consensual nude imagery, mostly of women, right, and sharing it on sites like Pornhub, right? 14:53 So it's used to harass people, and it's pretty ugly, right? 14:57 Some states have laws against that. 14:59 Some states don't. 15:00 But that's just sort of one of the ways that these tools have been misused, and I think we're about to find a lot more of them. 15:05 Kevin Roose yet. 15:06 And Stable Diffusion, this AI model made by Stability AI, I think it's a really good case study for a lot of these concerns because it's open source. 15:16 Anyone can download a version of it totally for free and run it on their computer. 15:21 They can modify it or fine tune it. 15:23 They can take off the safety filters. 15:25 Whereas a lot of other image generators like DALI 2 or Mid Journey, these are closed source and they come with some pretty strict rules about what you can make with them. 15:34 And I should say, Stability AI does have an app called Dream Studio that uses stable diffusion and that does have safety filters built into it. 15:43 But people are downloading and modifying and building versions of stable diffusion that don't have any rules at all, where basically anything goes. 15:52 You know, regulators who are not known for being like on top of every tech trend the moment it happens are already starting to get pretty worried about generative AI. 16:01 So just a few weeks ago, Anna Eshoo, who's a Democratic Congresswoman from California, sent a letter to the U.S. National Security Advisor and the Office of Science and Technology Policy 16:13 basically calling out stable diffusion in particular, this open source AI model, saying that it was being used to create violent images among other graphic topics, and saying that basically federal regulators needed to get their arms around this technology because the potential for abuse was so high. 16:33 So all that being said, we've got the CEO of Stability AI here with us today. 16:38 His name is Ahmad Moustak. 16:39 He's a former hedge fund manager and oil trader. 16:42 He's done a lot of things. 16:43 He worked on this big data project during COVID-19. 16:47 And recently, he kind of burst onto the AI scene from nowhere. 16:50 And he's become a pretty powerful figure. 16:53 And all these questions are kind of what I want to talk to him about. 16:57 Like, what is Ahmad thinking about how this technology that he is putting into the world could be used and could be misused? 17:05 So after the break, we're going to talk with Ahmad Moustak. 18:04 Casey Newton Said you hear about Dave. 18:49 Kevin Roose Ahmad Mastak, welcome to the show. 18:51 Hey, thanks for having me. 18:53 So the big news this week is that you all have just announced that you've raised $101 million, which, by the way, is like a very high number, but also a very specific number. 19:02 Like, you must have not wanted $100 million. 19:05 What's going on there? 19:07 Crosstalk The extra one is for luck. 19:09 Ahmad Moustak No, yeah, I mean, we kicked off a year ago, and this was our seed round, the first substantial fundraise that we did, just in our week of launch in August. 19:20 It's not cheap to train big AI models. 19:23 You need huge supercomputers. 19:24 Kevin Roose You released this statistic. 19:26 You have 4,000 of these giant processors, these A100s that are sort of the top-of-the-line graphics processors, which already makes you kind of one of the biggest supercomputers in the world, correct? 19:40 Ahmad Moustak Yeah, I think we're up to 5,892 today. 19:43 Kevin Roose And you said that you are going to use some of this money to expand that over the next year. 19:48 How big do you think your supercomputer will get? 19:51 Ahmad Moustak We'll get about 5 to 10 times as big. 19:53 It'll be one of the fastest. 19:54 Stable Diffusion, we took 100,000 gigabytes of images, 2 billion images. 19:59 And we used this great big supercomputer to squish it down to a two gigabyte file. 20:03 It now runs on an iPhone. 20:04 And it's a file that can create anything. 20:06 I mean, it's a bit like that show Silicon Valley from HBO. 20:09 You know, we're Pied Piper. 20:10 But it's reality. 20:12 Because we've had two types of AI. 20:13 Classical AI was big data. 20:15 They took large amounts of information, and they targeted Kevin to target him ads. 20:19 Now we kind of moved on to a new type of AI that kind of learns from principles. 20:22 So it learns what is a cup. 20:24 A cup is you cup your hand. 20:25 It's a well cup. 20:26 You know, it's an actual cup that you have water in. 20:28 And this is a new type of thing that's allowed us to break through to almost human level across a range of things. 20:34 So people are taking stable diffusion. 20:35 They're making images from it. 20:37 But they're also making 3D from it. 20:39 They're making architectural drawings from it. 20:41 They're using it to do dynamic fashion. 20:43 Like... 20:44 Because we released it open source and free to the world, there have been a thousand different projects done there. 20:49 So I think we've had 200,000 developers download it in just a few weeks. 20:53 And they're just creating amazing things. 20:55 Kevin Roose Yeah, let's talk about that open source approach that you've taken, because I think it's worth noting how strange this is in the landscape of AI research and development. 21:05 So... 21:05 Usually, someone like OpenAI, they release something like DALI or GPT-3, and they put all these rules around it. 21:14 They say, you know, you can't generate porn, you can't generate violent imagery, you have to basically be approved to build an app on top of our AI systems. 21:24 What stability did was basically just give this stuff away for free to anyone who wanted to use it with really no rules at all or very light rules. 21:36 So why did you take that approach? 21:37 Ahmad Moustak Yeah, I mean, we released it with an ethical use license and there was a safety filter in it. 21:42 I mean, there are good rules and some sensible things. 21:45 You know, you don't want to be used for violence or deepfakes, and we can discuss that later. 21:48 But at the same time, like, let's take open AI. 21:51 That was set up in 2015. 21:52 Then Microsoft did a billion-dollar investment in them in 2019. 21:55 Dali 2 is the equivalent image generator from them. 21:59 Again, a wonderful feat of engineering. 22:02 But you have to ask some questions. 22:04 For example, until a couple of months ago, you couldn't use the word Ukraine. 22:08 or any city in Ukraine or anything to do with Ukraine, and you couldn't use it in Ukraine. 22:13 Who do you discuss that with? 22:14 Nobody, they're a private company, right? 22:16 So all of a sudden you have all of the world allowed to be immensely creative, except for Ukrainians. 22:22 You know, similarly on diversity, how do you do diversity if you're a centralized thing? 22:25 What they do is they randomly allocate gender and race, if it's a non-gendered, non-race word. 22:31 So you have female Indian sumo wrestlers, if you type in sumo wrestler. 22:35 Our take was different, which is that you should give these primitives out so people can create their own models, they can bring their own data, and customize it to themselves. 22:43 On the bad side of things, on deepfakes and other stuff, these models weren't quite good enough to do that versus some of the existing ones. 22:49 But we thought, let's open up the discussion. 22:51 So we just announced a $200,000 prize for the best open source deepfake detector. 22:56 And we spent 10 times the compute that we did on the image generation model on image identification model that will be used to identify bad and illegal and other content. 23:05 So that's the approach that we're taking, which is that we trust people and we trust the community as opposed to having a centralized unelected entity controlling the most powerful technology in the world. 23:15 Casey Newton Wait, so if we type in sumo wrestler using your tech, well, we'll just sort of see like Japanese men only. 23:20 Ahmad Moustak Yeah, pretty much. 23:21 And if you use Japan Diffusion, which is the fine-tuned version from the Japanese community, you'll see even better sumo wrestlers. 23:28 Whereas if you do it on some of these centralized ones like OpenAI, you get random races and things like that. 23:34 Casey Newton So part of your argument here is just that when it comes to these generative AI tools, one size does not fit all. 23:39 You have to sort of make it contextual to whatever you're trying to create. 23:42 Ahmad Moustak Yeah, I don't think that'll happen with centralized tools. 23:44 I mean, basically, I believe this is one of the ultimate tools for freedom of expression. 23:48 And I believe expression should be free. 23:50 Very American for a Brit, I know. 23:53 And I think that's the power. 23:54 The power is in diversity. 23:56 So we're taking this and building Indian versions of it. 23:58 You know, we just did a deal to all the Bollywood stuff. 24:00 So you'll have generated Bollywood things with Eros Entertainment. 24:03 You know, we're taking it to Thailand, taking it to Vietnam. 24:05 And we're seeing this global community come and build. 24:08 And it's just amazing to see the diversity of output there, which will never happen if it's centralized. 24:12 Casey Newton So I'm interested in how this technology gets to the average person. 24:16 I've downloaded something called Diffusion B, which is sort of a very easy-to-use user interface. 24:21 Much like Dali, I can type in any words, but Stable Diffusion will actually create an image based on a model that's downloaded entirely to the laptop that's right in front of me right now. 24:30 At the same time, you're talking about things like very personalized, contextualized AI, and I'm just wondering 24:36 What's the sort of path from you to me using that? 24:40 Presumably, you're not going to be making all these apps yourself. 24:42 So is it that developers will sort of like get an API from you? 24:46 Or how are these tools all going to get built? 24:48 Ahmad Moustak Yeah, so we have our API, and then we have dreamstudio.ai, which is our own version of implementation of that. 24:54 So this week, we announced animation, and we've got 3D and video and other things coming in the next few months. 25:00 So you'll be able to create anything you can imagine. 25:02 Kind of insane. 25:03 However, what's happening is that people are taking this, they're integrating to other stuff. 25:06 So if you use Canva, for example, there is a stable diffusion integration where you can create an image. 25:11 Kevin Roose Canva is like an online Photoshop, basically. 25:13 Ahmad Moustak Yeah. 25:14 Or if you use Photoshop, there's a plug-in. 25:16 Kevin Roose So you can literally go into Photoshop and say, not just like, I would like to edit this image, but I would like to insert something totally new into it. 25:23 Ahmad Moustak Exactly. 25:23 So you've got like Captain Jack Sparrow and you're like, on his shoulder, I don't want that parrot. 25:27 You kind of highlight it. 25:28 Instead, I want, I don't know, a possum. 25:32 And it will create a possum in two seconds. 25:34 Kevin Roose One of the things that you said in your talk on Monday night at this event was that you believe that the world... 25:40 I want to make sure I use your exact words here. 25:43 The world has been creatively constipated and we're going to let them poop rainbows. 25:49 What did you mean by that? 25:51 Ahmad Moustak Look... 25:52 We're talking now, right? 25:54 So audio is the easiest way to communicate, you know? 25:57 Even if you deepfake someone's voice or whatever, you know? 26:00 The next level above is text. 26:02 You know, the reason that you're a great writer, Kevin, is that you're a great wordsmith. 26:07 But that's effort. 26:08 It's effort to create text. 26:09 So that's the next level above. 26:10 Visual is the hardest. 26:11 You know, the artistic process being so difficult. 26:13 But then... 26:14 PowerPoint. 26:15 PowerPoint is visual communication, and it's incredibly difficult, right? 26:20 Everyone hates it. 26:21 So I'm like, this AI can make it easier to write text. 26:24 and easier for anyone to be able to create PowerPoints or images as they want. 26:28 It lowers the bar, it lowers that effort, it lowers the pain. 26:32 So what happens then is that people will be able to communicate more because like how many of the people believe that they can create? 26:39 I would say maybe 10% believe that they can create flawlessly. 26:43 I believe that maybe another 30% believe they could 26:46 if they put in massive amounts of effort. 26:48 But the majority believe that they can't. 26:49 But the reality is anyone can with these tools. 26:52 So like I said, allow us to kind of poop rainbows and make a wonderful new future. 26:55 That's the way we go. 26:56 And that's going to be one of the biggest changes, I think, in human communication over the next couple of years that we've had since maybe the printing press, because that enabled us to use text. 27:05 And now it enables us to use visuals without any barrier. 27:08 Snapchat and TikTok and things were the first elements towards that. 27:11 This is the next evolution. 27:12 Kevin Roose So this stuff right now is, I would say, kind of primitive. 27:15 I mean, you can generate things that are realistic, but sometimes, you know, if you're trying to do faces, it will, you know, sort of be a little off, won't quite cross the uncanny valley. 27:25 Sometimes, you know, you ask for, you know, a hand and it comes back with seven or eight fingers. 27:30 So there's still some bugs in the software. 27:32 Where do you think this will be five years from now? 27:35 Ahmad Moustak I think you'll have the Ready Player One experience slash holodeck. 27:38 It will be perfectly high resolution. 27:42 No issues whatsoever. 27:43 Because basically what we've done is there's this concept of a type 1 and a type 2 brain, right? 27:47 There's a logical brain that has retrieval, and then there's that brain that just understands concepts, right? 27:52 We've now got both. 27:53 We have search and we have concepts. 27:55 So the way that I tell people to use this in the developer community is don't think of it just as a one-stop ship where you put kind of words in and images out. 28:02 Put it as part of a process where you cross-check and reference. 28:04 And then we go into Unreal Engine, the plans that Apple have around kind of augmented reality, 28:09 The fact that we've got AI chips everywhere, it'll get faster and faster and better and better. 28:13 Because this will be the biggest investment theme, the most interest of anything in the world. 28:17 Like self-driving cars at $100 billion of investment, they went nowhere. 28:21 This area, every single content provider in the world will have to think, not with our metaverse strategy, but what's our generative AI strategy now? 28:29 Because we have massive amounts of content. 28:30 How are we going to make that for audio, video, 3D, et cetera, in the next generation? 28:35 Kevin Roose So now you've raised this big round of money, $101 million from venture capital firms. 28:40 Presumably, they're going to want that money back at some point. 28:43 They're going to want you to make money. 28:44 How are you making money right now? 28:46 Because you're giving away stable diffusion for free. 28:49 It's all open source. 28:50 So how are you making money now? 28:51 And how do you plan to make money in the future? 28:53 Ahmad Moustak So yeah, basically, all the servers and databases in the world now are on open source. 28:58 And it's a multi-billion dollar market cap industry. 29:00 And it comes down to two things, scale and service. 29:03 So scale is making it easy for people to do this. 29:05 That's our API. 29:07 Kevin Roose So every time someone creates an image through one of your APIs, you get a cut of that? 29:13 Ahmad Moustak Yeah, or Dream Studio. 29:14 So we get a nice percentage. 29:16 The second part is service, which is there's very few people going to build these models, but every content provider in the world wants to have their own versions of this model. 29:23 You know, you want a Hello Kitty model, you know, or you want a Bollywood model or whatever. 29:27 So what's happening is that we're going to a lot of these big content libraries and saying, look, this is the actual metaverse. 29:32 We call it the multiverse because we think everyone should have their own models, right? 29:36 And so we're embedding teams there to create the models for them and sharing in the upside. 29:40 But that's also, you know, you have service contracts, all these other contracts around that because it's a specialized thing right now. 29:46 So we're very bullish on this because we think this is the future of content, you know, whereby you have that rich generative future where everyone can personalize and kind of contextualize these things. 29:56 Like this whole entire media space will be generative assisted. 30:00 I don't think it replaces, I think it enhances. 30:03 Casey Newton Well, I'm curious, and it gets philosophical, but what are you thinking of as, I don't know, like humanity's sustaining advantage here, right? 30:14 Like what is the role of just sort of human consciousness in creation when we're now giving so much of it to these automated tools? 30:22 Ahmad Moustak So I think automated is an interesting thing, right? 30:25 Because one of this is basically a snapshot of ourselves that kind of extends out. 30:28 But it requires us as the initial agent to do it. 30:31 So that's why I said, like, if someone uses it for a moral acts, it's that person's thing. 30:35 Because it doesn't do anything without that person. 30:38 And so I think we need to consider that very carefully. 30:40 Casey Newton So you're saying that human intention is really what matters here. 30:44 It's the fact that we're telling these machines to help us that is what makes the process fundamentally human and more an extension of us than some sort of robot replacement. 30:55 Ahmad Moustak Exactly. 30:55 And this is the case of many tools that we have, like photography and other things. 30:59 You have Photoshop, right? 31:01 If you use Photoshop to create a copyrighted entity, 31:04 And then you say, that's your fault. 31:06 These tools don't do anything by themselves. 31:09 It's a two gigabyte file that you input, and then it creates an output. 31:13 So we have to take it back to that original human. 31:15 But what it does now is it opens up access, just like, again, the printing press opened up access. 31:20 Now, anyone has access to visual creativity. 31:23 The first version of this I actually did for my seven-year-old daughter, because she was like, I want to create, Dad. 31:27 This is fun, and it's painting, and look at all the stuff you're doing. 31:29 And she created this wonderful piece called Happy ETH that she sold for $3,500 as an NFT for India Code Relief. 31:38 And she donated all the money. 31:39 I was like, holy crap, this is a big thing. 31:41 And I was like, why aren't you making more? 31:42 She made like eight more pieces. 31:43 And she's like, dad, the value of ones and uniqueness will only go up in time as the sector grows. 31:53 Kevin Roose We'll be right back. 33:21 We're back with Ahmad Mastak. 33:23 Ahmad, I want to ask you about a letter that came out just a few weeks ago. 33:27 Congresswoman Anna Eshoo from California wrote a letter to the U.S. National Security Advisor and the Office of Science and Technology Policy, basically sort of calling attention to this new generative AI area, and in particular, calling for a regulatory response to stable diffusion, to this piece of software that you have built. 33:47 She called stable diffusion unsafe, said it was being used to create violent images of Asian women being beaten and other sort of pornographic images. 33:57 And she says that this could be a potentially dangerous tool. 34:01 So I wanted to just see, have you read that letter? 34:03 What are your thoughts on it? 34:05 Ahmad Moustak We just made aware of it recently, and we've read it with interest. 34:09 I think there's this thing, again, it comes down to intentionality and who is responsible for the use of a tool, you know? 34:15 Because if a tool could be used for violence, I mean, we're here in America, we have guns, and guns are dangerous, but we're still allowed to use them, right? 34:24 This tool is the purest example of a tool for expression that you can possibly get. 34:28 I'd say it's very interesting in an American context to call for regulation of a tool for expression. 34:34 But it's part of this bigger thing of who is responsible for AI. 34:37 So within the European Union context, basically, there are calls to, well, the current system is basically calling to regulate it so that the makers of AI are responsible for the end use of that AI. 34:46 even if it's academic or open source. 34:48 That leads to a very interesting future where basically this technology would be locked away from everyone. 34:52 And so that base level of creativity that could unleash across the world is removed because of concerns of the other side. 34:59 Blocking of the means, as it were. 35:01 It's got similar things in other types of law. 35:03 It leads to a very orthodox position, and this is the position of AI ethics, which is that there could be something bad, so we must not release it as a Google or a Meta or something else. 35:11 Rather than saying, let's open up the conversation and let's figure out systems to deal with the bad. 35:16 And let's trust in people in the community to have overwhelming good. 35:20 I think there should be a debate. 35:21 Who is responsible for AI? 35:23 Who is responsible for the output of AI? 35:25 And how can we make sure that it's open and positive? 35:27 But I don't think that happens if the only debate that happens is behind closed doors and big technology companies. 35:32 Casey Newton So this debate is so reminiscent to me of the debate that we've had around social media over the past decade, right? 35:39 Which started with some very similar ideas of social media is going to democratize expression. 35:45 People who've never had a voice before will now have one. 35:48 And I think after 10 years, there are a lot of people who really dislike Facebook and 35:53 And think that Twitter is really bad for the discourse, right? 35:56 And even if the technology was created with the best of intentions and the tools themselves weren't actively encouraging people to do harm, ultimately a lot of harm was done. 36:04 And so now you see a lot of calls for regulation around the world. 36:07 I'm curious, like, how you feel about social media. 36:10 Like, do you feel it's been a net positive? 36:13 And did it inform the way you've thought about Stability AI? 36:16 Yeah. 36:16 Ahmad Moustak 100%, because what you had was centralized AI controlling us, and the incentive is advertising, which is manipulation of the mind. 36:23 And so I saw a vision of an intelligent internet where everyone has their own AI as opposed to centralizing AI, which was the default under this, with AI that's even more convincing. 36:33 I didn't like that future. 36:34 I didn't want that for my children. 36:35 I thought it needed to be expanded, not only through the open source, but across the world so other people would have a voice here in this discussion and debate. 36:43 Because there is this positivity, there is this negativity. 36:46 We don't know where it's going to end up. 36:48 But again, we're in a democracy, right? 36:50 We should have this as an open, extended debate. 36:53 And I've tried to create a company that is collaborative and open. 36:56 We cooperate and we're willing to hear feedback. 36:59 Casey Newton Maybe I'm just a coward about these things, but when I think about the potential for both the good and the harm that AI can do, I think the best approach is to parcel it out little by little, see what harms result, see if you can regulate that away, create policies around it, right? 37:17 So when I look at what open AI has done, I think that feels... 37:20 responsible to me. 37:21 When I look at what you're doing, I think you're going to accelerate both the good and the bad at a really fast clip. 37:29 And so one question I have for you is like, what's your case for the acceleration, right? 37:33 Like, what's the case to do everything all at once? 37:36 Ahmad Moustak So I think, again, this is the question of responsibility, it's a question of organization and other things. 37:42 We are in a unique place right now to organize some of the Discord here, to hopefully coordinate this for the better, because no one else was stepping in. 37:49 This technology was emerging regardless, and we saw that and we said, okay, we need to have a responsibility here to try and do our best. 37:55 to guide the thing, but then get other people into the room on this. 38:00 I think that the parceling out, you never know what that looks like, but when someone breaks, they might break it from a less good perspective, as it were. 38:07 I was very terrified of that, especially because, like I said, this technology is being used for very nefarious terms. 38:12 The good, though, I think massively outweighs the bad in this because there's nothing like creation. 38:17 We are in a society of consumption at the moment. 38:20 And if you look at what art therapy does, if you look at the things around, well, the joy that happens with creation and people using this technology, 38:28 Why should we cut that off from the world? 38:30 Who are the self-appointed representatives that decide that? 38:33 I think that's wrong. 38:34 Again, it's the blocking of the means. 38:36 The possibility of any type of evil means that we cannot have anything. 38:40 Whereas in fact, what's best is when we're stronger together and we work as a society to combat the evil and then we push the good. 38:47 Kevin Roose Yeah, I mean, I think what you're hearing from me and Casey as reporters who have spent the past 10 years covering the unanticipated consequences of social media is this sense that a lot of the people who, it's not just regulators who wish that some of this stuff had happened more slowly and that they had had time to sort of make rules and adjustments. 39:05 It's actually the people, a lot of the people who are building this stuff have recently said, you know, I wish we had gone slower. 39:10 I wish we had taken trust and safety more seriously. 39:13 Yeah. 39:13 There's this Netflix documentary, The Social Dilemma, which has all these former Facebook and Twitter employees saying, we went too fast, we didn't understand what we were doing. 39:22 So I guess what makes you confident that your approach is the right one? 39:25 Are we going to see you in a Netflix documentary 10 years from now saying, I wish we had gone slower, I wish we had thought about this stuff? 39:32 Ahmad Moustak So I've spoken to just about everybody on that documentary about this. 39:36 And so basically what we're doing is, what if you'd created Facebook and Twitter without... 39:41 and add incentive. 39:43 And you'd also been accelerating the tools to counterbalance this. 39:47 And we spent 10 times as much compute and money on the identification of images as opposed to the creation of images. 39:55 We have a $200,000 deepfake prize to create technology to counterbalance that. 39:59 Because again, this technology isn't our best model actually. 40:02 Again, we trust in the community and we trust in this decentralization as opposed to this kind of centralized coordination whereby these decisions were made. 40:09 separately these algorithms are locked away and they're not interrogable they're not understandable for good and ill because you know it's not perfect you can interrogate the data sets you can interrogate the model you can interrogate the code of stable diffusion and the other things we're doing around language and others because again we believe that's a public good and a public right and we're seeing it being improved all the time from biases trust and safety versus again being in these big companies that their incentive is not the public good 40:35 Their incentive is the bottom line and it's manipulation of you to serve ads. 40:40 We can't forget that. 40:41 And I don't believe that that can be trusted. 40:42 I believe it needs to be open and for the people as it were. 40:46 Casey Newton Although it's interesting to think about, you know, one thing that makes me maybe a little bit less nervous about the usage of these tools to create really bad images in particular is that the content policies of the big social networks will prevent them from being uploaded and shared for the most part, right? 41:03 Like if I create a bunch of extremely violent imagery, I can't put it on Facebook, right? 41:08 I can't put it on Twitter. 41:09 And so interestingly, it's like these same networks that you're being critical of are actually going to play a huge role in preventing the spread of these materials. 41:15 Yeah. 41:16 Ahmad Moustak No, 100%. 41:16 And because these networks have now adapted for society, right? 41:20 And this is the thing. 41:21 Does society accept unethical, immoral, and illegal things? 41:24 No, it doesn't. 41:25 So right now, 4chan and other interesting websites online have had this model for eight weeks. 41:31 Has incredibly bad stuff been done on the back of that? 41:34 Not really. 41:35 I mean, it hasn't broken the internet. 41:37 Again, how much good has been done from that? 41:39 There's some amazing art and things that have been created, hundreds if not thousands of projects being built on the back of this. 41:45 We have to kind of acknowledge that. 41:46 We live in a society, as it were, that has these rules, regulations, and other things. 41:51 Social networks and others, they have this motivation that I believed was dangerous for this artificial intelligence, but they're full of good people and developers trying to do the right thing. 42:00 It's just that, like I said, 42:02 We couldn't have a world that I thought was just this one closed AI approach where these decisions are being made without consultation. 42:12 Kevin Roose Have you seen anything made with stable diffusion or any of your other tools where you've been like, ooh, wish that wasn't made using our tools, kind of feel like we should have made a rule about that? 42:21 Ahmad Moustak Not particularly, no. 42:23 I mean, there's been some pretty gross stuff because it doesn't really understand anatomy. 42:27 And so people were trying to really push it. 42:28 I was like, ooh, that's a bit gross. 42:30 But again, I think it's a tool like any other, like a Photoshop or things like that. 42:34 It isn't a tool that makes stuff by itself. 42:36 It's the intention of the end user. 42:38 And so there's nothing that you can make with stable diffusion that you can't make with Photoshop. 42:42 It's just become a bit easier. 42:43 Kevin Roose One other criticism that you often hear of generative AI is that it's copying or stealing art from people. 42:50 I was reading recently this interview with this artist named Greg Rutkowski, who's like a famous fantasy illustrator. 42:59 People love his work. 43:00 And so people who have been using these image generators have sort of figured out that if you just add Greg Rutkowski to the end of your prompt, it like makes it look amazing. 43:10 And which is great for them because they can now create stuff that's like Greg Rutkowski level of good. 43:17 But for Greg Rutkowski, who didn't give permission for his art to be sort of ingested by the AI model, it's now like everyone can essentially rip him off with very little effort. 43:30 So how do you think about that? 43:33 Ahmad Moustak So I think this is a very kind of valid thing in the fears and kind of concerns around that. 43:38 So like I said, it was trained on 2 billion images. 43:39 It was a snapshot of Google Images, basically. 43:43 Right, so anything you could find through there, you can find through here. 43:45 And then it learns from the relationship between words and images. 43:49 That's why you need a huge freaking supercomputer to kind of do that. 43:51 So it learns principles. 43:53 You can't recreate any image from the data set, but it understands what a cup is or what Greg is and other things. 43:58 In fact, the interesting thing is our data set didn't actually have very many pictures of Greg Rutowski. 44:02 Instead, it was another model that was released by OpenAI, which had much more of his. 44:08 We don't know because we know what the data set is. 44:09 That was part of this that introduced the concept of Greg into this model. 44:13 Kevin Roose I just love that he's like the secret word, the code word that unlocks the really good level of AI art. 44:20 Ahmad Moustak Well, basically, prompting is like magic, right? 44:22 It's like, I've got a spell. 44:23 Kevin Roose But respond to that because I think this is a valid fear that a lot of working artists have, which is that they are going to be put out of work by people who are essentially using these big AI models to do what they have been doing for a living for years. 44:37 Ahmad Moustak Yeah, I mean, I think this was the concern around photography and a lot of these other kind of tools that came through. 44:41 And it depends on the definition of what an artist is. 44:44 Is it an aesthetic thing? 44:45 Someone who's communicating through a community or is it a job? 44:48 We're an illustrator, etc. 44:49 ? 44:49 And these technologies have been improving to display some of these things all over time. 44:54 The way that kind of I think about it is that, like I said, it's just with a couple of images, you can teach it any style. 44:59 So even if there was a base model, you bring in a couple of images, it will learn Fliberty Gibbet, which is a new artist from Brooklyn, I don't know, whatever. 45:07 And again, it comes down to the intention of the individual user. 45:11 If they're breaking copyright or something illegal, that is one thing. 45:14 Unethical and immoral is another thing. 45:15 Because you can combine Greg with Van Gogh, with Banksy, and it will incorporate all their styles because it understands that. 45:22 But it'll never recreate an image. 45:24 Because like I said, this is not... 45:26 a compression algorithm. 45:27 It's a learning algorithm. 45:28 It's like a human brain, again, which is kind of crazy. 45:30 However, I do think that we should build systems that listen to the community. 45:34 So we've been talking to, I think, hundreds of artists. 45:36 And so one of the things we're doing is trying to figure out attribution mechanisms, opt-in and opt-out mechanisms as well, because again, we listen. 45:42 Unlike most others, we don't know what the models and data sets from some of the big companies are that understand these concepts. 45:48 Whereas ours is open, interrogable. 45:50 And so, you know, we receive flack, but at the same time, we are listening and building tools around that. 45:55 And the community is also doing that. 45:57 So we're supporting community initiatives around this as well. 45:59 Casey Newton So you can see a world where like a Greg can just opt out of having his style included in a model. 46:05 Kevin Roose Yes. 46:06 But isn't that a rule? 46:07 I mean, isn't your whole pitch that this kind of thing should be open and free and people should be able to use these giant models to create whatever they want? 46:15 Ahmad Moustak Yeah, the technologies should be open and free and they should use it ethically by the license. 46:20 But, I mean, we get to decide what we put in our model, right? 46:22 And so these are open decisions where we're listening to feedback. 46:25 But then other people can take our code 46:27 And they can create their own models. 46:29 It's the thing. 46:29 It's open infrastructure. 46:31 So as a company, we decide what we have for the models that we release. 46:34 But then we also have our implementations of those models. 46:37 So our safety filter on Dream Studio is a lot stricter than the safety filter that is in... 46:42 the open model release, because that's a decision we make as a company versus an open source kind of decision. 46:49 And everyone should have their own one. 46:51 So for example, you can have an opt-out scheme whereby people upload their styles. 46:54 Again, it's difficult because even if there's a picture in the New York Times that it picks up, it will learn from that style and that little kind of thing. 47:01 And then you can show that to Photoshop and say, this is an opt-out list of filters, and they can choose whether or not to use that. 47:07 Kevin Roose So this message that you're promoting, it's very much oriented around this idea of like, we are the sort of populist alternative to big tech. 47:16 And on Monday, this big Exploratorium party, you gave this talk. 47:20 You said it was only your ninth talk ever. 47:22 But you said basically that this is the way that we're going to take on big tech to make sure that AI is not in the hands of a few giant corporations. 47:30 I was standing right next to Sergey Brin from Google when you said that, and he was not happy. 47:37 He was like, what, I'm the panopticon that you're talking about here? 47:41 Yeah, what did I ever do? 47:42 Yeah, what did I do to you, Imad? 47:45 So... 47:45 So that was funny, A. 47:48 But I think it's also a very strong pitch right now because a lot of people are skeptical of big tech. 47:55 But I also wonder because now, like, you just raised $101 million. 47:59 You have venture investors, including some of the same ones who have invested in other big tech companies. 48:06 You are a big tech company now. 48:08 I know you only have 100 employees and not 100,000 or however many Google has. 48:13 But how do you think you will keep Stability AI from repeating some of the flaws with big tech now that you have all of this money, all these investors, all these expectations around growth? 48:26 It's not easy, eh? 48:28 Ahmad Moustak We have 100 employees. 48:29 We have 100,000 people in our community. 48:31 That's kind of where our strength comes from, diverse across the entire kind of world. 48:35 We also give them revenue share, which is very weird. 48:38 And, you know, we give them the upside because we try them as artists effectively. 48:42 From the perspective of big tech, 48:45 Big Tech is creating the Panopticon because they can't help it because they've got no alternative. 48:49 We've given that alternative now. 48:50 So we're actually working with Big Tech now and giving them an outlet to be part of this thing. 48:55 We're kind of Switzerland that everyone can kind of participate in as the neutral party. 48:59 And so we've had amazing talks with all of them because they don't want to head in this direction. 49:03 They actually want to, and especially the engineers, make things free and open, but at the same time have the regulation, have the trust and safety things. 49:11 So we're getting guidance and input around that to have a happy medium because it can't be one extreme of pure libertarian and just release this and let it go and the other extreme of no one can have anything. 49:21 It will be somewhere in the middle. 49:23 So I think that these kind of factors we have will help us do that. 49:26 With our venture investing, we raised it, but we raised it on our terms, so we have complete independence. 49:31 As opposed to, for example, OpenAI, they raised a billion dollars from Microsoft, and Microsoft have an exclusive license to their technology. 49:38 Again, these misaligned incentives are very difficult to combat against, but we hope that the community, our team, and our position will help us counterbalance that. 49:45 Because also, it's good business for us to be in this position, because no one else really occupies this position. 49:52 Casey Newton Yeah. 49:52 I don't know. 49:53 There's another world, though, where it's like these big corporations for all of their evils, they are also reined in in terms of what they can do because they're so afraid of the PR crises that might result from them releasing an AI that leads to, you know, some sort of bad thing. 50:06 You're in this position right now that I think is enviable to them where it's like very few people have heard of you yet. 50:12 And so you have much wider latitude to just kind of do whatever and not be as worried as much about, you know, what a regulator or the public might think. 50:20 Ahmad Moustak Again, we're active talks with regulators. 50:23 And the public, again, is the community and extended. 50:25 So we released this, and we were like, it's open. 50:28 And then it got a bit crazy. 50:29 It's been eight weeks and two days, and it's just been insane. 50:32 And so 1,000 projects sprung up. 50:34 And the community was like, 50:35 Stability, why don't you step in and coordinate this and have an official mouthpiece? 50:39 We're like, okay, fine. 50:40 So we went in and we turned the Reddit into an official Reddit. 50:43 And they're like, how dare you? 50:46 Like corporate overlords. 50:47 We're like, we're just trying to make things a bit more organized. 50:49 And then we had to kind of give it back. 50:51 There's always this kind of push-pull. 50:53 I think, again, it's community first, but it can't be kind of direct democracy. 50:57 We're going to make mistakes and we're going to make right things. 50:59 And we're going to put under increasing amounts of scrutiny because what we're doing is actually fundamentally important. 51:03 And so I hope that, you know, we get the right input to do that right. 51:07 But we might not. 51:08 As you said, big tech is in this ununveiled position because they can't release it for PR things. 51:12 Because it's like Promethean fire from the gods, right? 51:14 Again, it's the next generation of communication, which is insane. 51:19 It can be used to burn things down. 51:20 It can be used to activate the light of humanity. 51:22 But the only way we can figure out how to do this best is together. 51:26 So that's why I want to work with big tech, want to work with little tech, want to work with regulators, want to work with everyone to try and figure out how to do this right. 51:33 Kevin Roose You invoked sort of this idea that AI is neither good nor bad. 51:38 It's neutral. 51:38 It's just a tool. 51:40 This is like one of the things that I try to push back on really hard whenever it comes up because I think that all technology is infused with values because it is made by humans and humans, including you, have values. 51:54 So do you really think that AI is neutral the way that you are building it? 51:59 Or do you think that... 52:00 People in positions like yours have a responsibility to ensure that their values are reflected in the tools that they're building. 52:08 Ahmad Moustak I would agree with you that it does infuse the values. 52:10 And again, there are decisions to be made. 52:12 And these are decisions where, again, the buck stops with myself as CEO of Stability on this or with some of the developers and others. 52:20 And they will be different because you can take a lot of different choices here. 52:23 What I do think, though, in terms of, as you said, good or bad, is that once the model is out there, it is infused with certain values. 52:30 But models like this are generalist models, whereas classical models are models that are more specific. 52:35 There is a bit of a difference there. 52:36 And I think, again, it's the, who is liable for the use of the model? 52:40 This is a very complicated question that we're trying to figure out now across various jurisdictions and geographies. 52:45 So I think we need to separate some of these things. 52:47 But again, we have to realize that nothing springs from nothing, as it were. 52:53 There are kind of these biases. 52:54 There are deliberate decisions that are being made. 52:56 And again, I've been putting this out and saying, for this thing, the buck stops with me. 53:00 I made this decision. 53:01 Or my company made these decisions. 53:04 And again, we should be held accountable for that. 53:06 Kevin Roose One interesting difference that occurs to me between generative AI and social media is that in social media... 53:14 as Casey and others have written about, the market is demanding more involvement. 53:20 The market doesn't, you know, Parler, Truth Social, Getter, these like free speech apps with very minimal rules, they're not popular. 53:29 What is popular is Twitter and Facebook and Instagram and TikTok, all of which have pretty strict guidelines. 53:37 In generative AI, what seems to be happening is that the market wants openness. 53:43 I mean, stable diffusion has, I think, and correct me if I'm wrong here, more users than Dolly or Midjourney. 53:50 Is that right? 53:51 Yeah, it's got more users than both. 53:52 So what do you think the difference is there? 53:55 Why are people in this field looking for sort of a more free-for-all environment than people on social media? 54:02 Yeah. 54:02 Ahmad Moustak So social media, again, you have centralizing AI that directs information from one place to another. 54:07 And it's been this kind of equalizer in a degree, but it's also been an extender whereby, you know, people such as yourselves have an almost disproportionate platform to influence. 54:16 And that's how the algorithms are set up. 54:18 People like easy. 54:19 Whereas generative media is a more personal experience. 54:22 There's no social aspect to it. 54:23 It's already integrated into TikTok, so they've just introduced their generative thing. 54:27 They should use stable diffusion instead because it's much better. 54:29 But, you know, this is the way of things. 54:32 But this is going to go exponential. 54:34 And, again, it's a very personal thing where you put it out and then you use the social networks to communicate. 54:39 Kevin Roose Where do you think generative AI goes from here, beyond photos, beyond videos, beyond the creative professions? 54:45 Is there a world in which this type of technology changes the law, medicine, science, journalism? 54:53 I mean, where do you think this is headed? 54:54 Ahmad Moustak No, 100%. 54:55 It's going to be everywhere. 54:56 We look ahead 10 to 20 years and you have an AI that learns principles as well as an AI that can search logic. 55:03 That's insane because it isn't just generating media. 55:05 It's like a reflection of the human brain. 55:07 And I think, again, this is the most powerful technology in the world, which is why I said we couldn't trust for it to be just in the hands of a few. 55:14 And so we were like, let's put it out to the people because it should be AI for the people by the people where they can create it themselves. 55:20 Kevin Roose Thank you so much. 55:22 This is really a challenging but a productive conversation, and I'm glad you stopped by to talk to us about it. 55:27 Ahmad Moustak Thank you so much for having me on the show. 57:12 Kevin Roose That's it for Hard Fork this week. 57:14 Our show is produced by Davis Land. 57:15 We're edited by Paula Schumann. 57:17 This episode was fact-checked by Caitlin Love with original music by Dan Powell, Alicia Baetup, and Marian Lozano. 57:24 Our show was engineered by Corey Schreppel. 57:27 Special thanks to Hannah Ingber, Nell Gologli, Kate Lopresti, Shannon Busta, Mahima Shablani, Jeffrey Miranda, and Mahmoud Felfel from Play.ht. 57:38 Thank you. 57:38 You can email us at hardfork at nytimes.com. 57:42 That's it for this week. 57:43 See you next time. 57:45 Crosstalk We also don't have to say, that's all for this week. 57:47 See you next time every week. 57:49 You guys can do whatever you want. 57:50 I just put that in there. 57:51 Casey Newton We should have something that feels a little bit less AI. 57:55 Because an AI would say, that's all for next week. 57:57 See you next time. 57:58 Whereas we might say, adios, cowboys. 58:01 We'll catch you on the range down the road. 58:03 We'll see you down the dusty trail. 58:30 For a limited time at McDonald's, get a Big Mac extra-value meal for $8. 58:34 Crosstalk That means two all-beef patties, special sauce, lettuce, cheese, pickles, onions on a sesame seed bun, and medium fries, and a drink. 58:42 We may need to change that jingle. 58:43 Casey Newton Prices and participation may vary.