Behind the Bastards - Part Two: Tech Bros Have Built A Cult Around AI Aired: 2024-02-01 Duration: 58:31 === Behind The Bastards Intro (01:33) === [00:00:00] This is an iHeart podcast. [00:00:02] Guaranteed human. [00:00:04] When a group of women discover they've all dated the same prolific con artist, they take matters into their own hands. [00:00:13] I vowed I will be his last target. [00:00:15] He is not going to get away with this. [00:00:17] He's going to get what he deserves. [00:00:19] We always say that. [00:00:21] Trust your girlfriends. [00:00:24] Listen to the girlfriends. [00:00:25] Trust me, babe. [00:00:26] On the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:00:31] I got you. [00:00:32] I got you. [00:00:36] Hey, it's Nora Jones, and my podcast, Playing Along, is back with more of my favorite musicians. [00:00:41] Check out my newest episode with Josh Grobin. [00:00:44] You related to the Phantom at that point. [00:00:47] Yeah, I was definitely the Phantom in that. [00:00:48] That's so funny. [00:00:50] Share each day with me each night, each morning. [00:00:58] Listen to Nora Jones is playing along on the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:01:06] What's up, everyone? [00:01:07] I'm Ego Mode. [00:01:08] My next guest, it's Will Farrell. [00:01:12] My dad gave me the best advice ever. [00:01:15] He goes, just give it a shot. [00:01:16] But if you ever reach a point where you're banging your head against the wall and it doesn't feel fun anymore, it's okay to quit. [00:01:23] If you saw it written down, it would not be an inspiration. [00:01:26] It would not be on a calendar of, you know, the cat just hang in there. === AI As Digital Religion (10:19) === [00:01:33] Yeah, it would not be. [00:01:35] Right, it wouldn't be that. [00:01:36] There's a lot of life. [00:01:38] Listen to Thanks Dad on the iHeartRadio app, Apple Podcast, or wherever you get your podcasts. [00:01:46] Cool Zone Media. [00:01:50] Welcome back to Behind the Bastards, a podcast that, again, I botched it. [00:01:57] My weird voice didn't work. [00:01:59] I'm sorry. [00:01:59] I didn't enjoy that at all. [00:02:01] I know. [00:02:01] I know. [00:02:02] That one was like more Dracula-y than the first one. [00:02:05] Yeah, Draculesque, sure. [00:02:07] He didn't sell it this time. [00:02:08] Sorry. [00:02:09] I apologize. [00:02:11] What I don't apologize for is my guest, Ify Nwatiway. [00:02:17] It's me. [00:02:18] It's your boy, Yo. [00:02:19] Ify working for the dropout, the survivors' successors to college humor who have blossomed like a phoenix from the ashes of the internet that Facebook killed. [00:02:32] Yes. [00:02:33] Yes. [00:02:34] Still we rise. [00:02:35] Still we rise. [00:02:37] Speaking of rising, my concerns about the cult dynamics within the AI sub cult year. [00:02:45] I don't know why I said it that way. [00:02:46] Yeah. [00:02:47] I don't know what you're on about right now. [00:02:50] I don't know either, Sophie. [00:02:51] I don't know either. [00:02:51] I'm doing great. [00:02:54] This is me. [00:02:56] I've been sober lately. [00:02:57] So this is me living the sober life. [00:03:00] I've just gotten worse. [00:03:01] So I don't know, everybody. [00:03:03] Keep your kids on drugs, you know? [00:03:05] Or dumb. [00:03:09] Depends on the drugs. [00:03:12] So perhaps the most amusing part of all of this is that a segment of the AI believing community has created not just a potential god, but a hell. [00:03:21] And this is one of my favorite stories from these weirdos. [00:03:24] One of the early online subcultures that influenced the birth of EAC are the rationalists. [00:03:29] And again, the EAC people will say a lot of them don't like the rationalists, but they're related. [00:03:34] They're like cousins in the same way cracked in college humor are, right? [00:03:37] Yes. [00:03:38] The rationalists are a subculture that formed in the early aughts. [00:03:42] They kind of came out of the online skeptic atheist movement of the late 90s. [00:03:47] And they formed in the early aughts around a series of blog posts by a man named Elizer Yudkowski. [00:03:53] Yudkowski fancies himself as something of a philosopher on AI, and his blog/slash discussion board less wrong was an early hub of the broader AI subculture. [00:04:03] Yudkowski, like he doesn't have a specific education. [00:04:06] He just came to be kind of an expert in AI and machine learning. [00:04:10] He's a peculiar fellow, to say the least. [00:04:14] The founding text, or at least one of them, of rationalism, is a 660,000-word Harry Potter fanfic that is just nonsense. [00:04:23] It's all about rewriting Harry Potter so his real magic is rational thinking. [00:04:28] It's wild shit. [00:04:30] He's like a psychopath. [00:04:32] It's so such an odd choice. [00:04:35] You know, it's like the, what was it? 50 Shades of Gray, how he was originally a Twilight fanfic, and there's going to be like a cloud atlas-esque. [00:04:46] But you know, the 50 Shades Gray lady was not trying to like create the new text for like a philosophical movement. [00:04:54] She just wanted to get like people horny. [00:04:56] And that's fine. [00:04:58] That's perfectly acceptable. [00:05:00] The most relevant thing about the 660,000-word Harry Potter fanfic is that it was the favorite book of Carolyn Ellison, the former CEO of FTX, who recently testified against Sam Bankman Freed or of Alameda. [00:05:15] Sorry, she was the CEO of Alameda. [00:05:17] Anyway, all these weird little subcultures, rationalism and effective altruism, are related to each other and influenced each other, even though, again, they often hate each other too. [00:05:26] Yudkowski is seen as an object of ridicule by most EAC people. [00:05:30] This is because he shares their view of AI as a potential deity, but he believes AGI will inevitably kill everyone. [00:05:36] Thus, we must bomb data centers, which, like, look, he may have gotten to the right end point. [00:05:43] He kept running like Force Gum. [00:05:46] He just kept running. [00:05:47] We're like, wait, wait. [00:05:48] Wait, wait, no, stop right there. [00:05:49] Stop right there. [00:05:50] We may agree with you on this. [00:05:51] Yeah. [00:05:52] Yudkowski is a doomer now because he was surprised when Chat GPT came out. [00:05:56] He was like horrified by how advanced it was and was like, oh my God, we're further along towards creating the AI that kills us all. [00:06:02] We have to stop this now. [00:06:04] And that made him, he had been kind of flirted with a lot of like Silicon Valley people. [00:06:09] He's the rationalists are very much a Bay Area cult. [00:06:12] He kind of has become increasingly a pariah of at least people with money in AI. [00:06:16] But before that happened, his message board birthed something wondrous. [00:06:20] In 2010, a less wrong user named Rocco posted this question. [00:06:25] What if an otherwise benevolent AI decided it had to torture any human who failed to work to bring it into existence, right? [00:06:32] What if we make an all-powerful AI and its logical decision is that, well, I will have to punish all the human beings who were alive and who didn't try to further my existence because that's the most reasonable way to guarantee that I come into being. [00:06:46] It's nonsense. [00:06:47] This is a silly, silly thing to believe. [00:06:50] It's all based on like the prisoner's dilemma, which is a concept in game theory. [00:06:53] And it's not really worth explaining why, because the logic is, it's only the kind of thing that happens when people are too online and like completely get detached from reality. [00:07:04] But Rocco's conclusion here is that an AI who felt this way would punish its apostates for eternity by creating a virtual reality hell, digitizing their consciousness and making them suffer for all time. [00:07:15] Now, you may have noticed, iffy, number one, they're kind of ripping off our boy Harlan Ellison, famed, famed advocate of the writer's right to their work. [00:07:26] But it's also just tech nerds recreating Pascal's wager. [00:07:30] Like, this is just Pascal's rager with an AI. [00:07:33] Like, you just stole again, these fucking plagiarists. [00:07:36] You just stole from whoever Pascal was. [00:07:39] This is what happens when you are a nerd and you refuse to read sci-fi. [00:07:42] You just, you eventually just come up with these stories yourselves and think that you did it. [00:07:48] Yeah, this, and if you're not familiar, folks, I think most people are. [00:07:51] Pascal's wager is this kind of like concept from, I think it's, you'd call it a Christian apologetics that's like we may not know if hell is real or not, but because if it's real, the consequences are so dire and the cost of just saying, yeah, I accept Jesus is so low. [00:08:08] You should do that, right? [00:08:09] Like, or I think that's the basic idea, right? [00:08:12] That's how a lot of people interpret it. [00:08:13] It's the whole idea behind like being a piece of shit and then converting on your deathbed, basically. [00:08:18] I don't know fully the history of it, but I know that they're basically aping it for fucking Rocco's basilisk. [00:08:24] And it's called a basilisk because, like a basilisk, if you look at it, it like enraptures your mind. [00:08:29] You can't stop thinking about it. [00:08:30] That comes from reportedly, there's some debate over this. [00:08:34] When this went viral among like the less wrong community, Yadkowski had to ban discussion of it because it was like breaking people's minds. [00:08:40] They were having nightmares. [00:08:42] Am I working hard enough to make the AI real? [00:08:44] Is it going to send me to hell? [00:08:47] Yeah, it's unclear like how seriously people were because again, this is just people talking on the internet. [00:08:53] For what it's worth, Yedkowski didn't really like Rocco's basilisk, but it's his place that birthed it. [00:09:00] And for an idea of how influential this is, Elon Musk and Grimes met talking about the concept. [00:09:05] That was their meat cute. [00:09:07] Was this fucking AI Pascal's wager? [00:09:10] Yeah. [00:09:10] She like wrote a song about it. [00:09:12] It's fucking ridiculous. [00:09:14] These fucking people are such things. [00:09:17] What? [00:09:19] Wow. [00:09:20] Oh my God. [00:09:21] Read Harlan Ellison. [00:09:22] He did it better than you. [00:09:24] God damn it. [00:09:26] I will say, reading this shit is the most I've ever felt like I have no mouth, but I must scream. [00:09:31] So again, poor one out for the man. [00:09:36] So this is all relevant, this AI hell some of these people have created, because it's one more data point showing that the people who take AI very seriously as real intelligence always seem to turn it into religion. [00:09:47] And this is kind of maybe the first schism, right? [00:09:49] This is their Catholic-Protestant split or their Catholic Orthodox split, because you've got on one side, Yedkowski's people who are like, we will inevitably make a god and that God will destroy us. [00:09:59] So we have to stop it, versus like, we will inevitably make a god and that god will take us to paradise along with daddy musk. [00:10:07] We'll go to the stars, right? [00:10:08] Those are the two. [00:10:09] This is like the first heretical split within the divine AI movement. [00:10:14] And this stuff is relevant because so many of the fucking these subcultures and movements start out as a bunch of people arguing or discussing their ideas in online communities. [00:10:25] And there is a reason for this. [00:10:26] It's pretty well recognized that there are certain dynamics inherent to the kind of communities that start on the internet that tend towards cultishness. [00:10:34] This is part of why like we have a big subreddit for the podcast. [00:10:38] It's like 80 something thousand people, which makes it in like the top 1% of Reddit. [00:10:42] And I've been offered like to be able to moderate and like make policy there. [00:10:46] I have nothing to do with the running of that subreddit because I'm like, that doesn't end well. [00:10:51] I was on something awful as a kid. [00:10:52] I know what happens when people make themselves mods of giant digital communities. [00:10:56] They lose their fucking minds. [00:11:00] We're all watching Elon Musk do it right now. [00:11:02] It's the worst thing in the world for you. [00:11:04] Thank you, by the way, to the people who do run that thing because I am not going to. [00:11:10] The skeptic community, which was huge through the late 1990s and early 2000s, might be seen as the grandfather of all these little subcultures. [00:11:18] After 9-11, prominent skeptics became vocally unhinged in their hatred of Islam, which brought them closer to different chunks of the nascent online far right. [00:11:26] Weird shit started to crop up, like a movement to rebrand skeptics as brights in light of the fact that their very clearly exceptional intelligence made them better than other people. [00:11:35] And again, you can see some similarity with this and the stuff Nick Land was talking about. [00:11:39] Only certain races will make it to space. [00:11:41] I found a very old write-up on plover.net that describes the method by which this kind of shit happens in digital communities. [00:11:48] Quote, online forums, whatever their subject, can be forbidding places for the newcomer. === Faith In Algorithmic Research (04:25) === [00:11:53] Over time, most of them tend to become dominated by small groups of snotty know-it-alls who stamp their personalities over the proceedings. [00:11:59] But skeptic forums are uniquely meant for such people. [00:12:02] A skeptic forum valorizes and in some cases fetishizes competitive geekery, gratuitous cleverness, macho displays of erudition. [00:12:09] It's a gathering of rationalities hardmen, thumping their chests, showing off their muscular logic, glancing sideways to compare their skeptical endowment with the next guy, sniffing the air for signs of weakness. [00:12:18] Together, they create an oppressive, sweaty locker room atmosphere that helps keep uncomfortable demographics away. [00:12:24] And that is where a lot of this shit is cropping up, right? [00:12:27] It is sweaty and uncomfortable, and there are mushrooms growing there. [00:12:31] And some of those mushrooms are fucking fascists. [00:12:34] And all of them want to take away the ability of artists to choose what happens to their art. [00:12:39] Oh, yeah. [00:12:40] I feel like this is just so many parts of the zeitgeist coming together because, you know, what it means to own media. [00:12:48] You know, I feel like a very small microcosm of this is when people would like clip out stuff from YouTube videos or ape jokes from people who tweet. [00:12:59] And when it goes, you know, viral or in the original tweet is like, hey, you stole this from me. [00:13:06] And it's either, no, I didn't, or like, yeah, but you like put it on Twitter. [00:13:11] So like, I can just copy what you wrote. [00:13:14] Yeah. [00:13:14] And now it has evolved into, yeah, we can just take from yours and let this machine learn how to do what you do so I can do it, even though I don't have the talent to do it. [00:13:25] Yeah, absolutely. [00:13:26] The reality of AI's promise is a lot more subdued than believers want to admit. [00:13:31] In an article published by Frontiers in Ecology and Evolution, a peer-reviewed research journal, Dr. Andreas Rowley and colleagues argue that AGI is not achievable in the current algorithmic frame of AI research. [00:13:45] And this is a, their claims are very stark, that like the kind of way we make these large language models, this algorithmic frame cannot make an intelligence. [00:13:54] That's their argument. [00:13:56] One point they make is that intelligent organisms can both want things and improvise capabilities that no models have yet generated. [00:14:04] They also argue basically all these things that individual AI type models can do, you know, recognize voice, recognize text, recognize faces, you know, this kind of stuff. [00:14:15] Those are pieces of what we would want from an artificial general intelligence, but they're not all combined in like the same thing that works seamlessly. [00:14:23] And beyond that, it can't act based on anything internal, right? [00:14:28] It can only act based on prompts. [00:14:30] And their argument is that algorithmic AI will not be able to make the jump to acting otherwise. [00:14:34] What we call AI then lacks agency, the ability to make dynamic decisions of its own accord. [00:14:40] Choices that are, quote, not purely reactive, not entirely determined by environmental conditions. [00:14:45] Midjourney can read a prompt and return with art it calculates will fit the criteria. [00:14:50] Only a living artist can choose to seek out inspiration and technical knowledge and then produce the art that Midjourney digests and regurgitates. [00:14:57] Now, this paper is not going to be the last word on whether or not AGI is possible or whether it's possible under our current algorithmic method of like making AIs. [00:15:06] I'm not making myself a claim there. [00:15:08] I'm saying these people are, and I think their arguments are compelling. [00:15:11] We don't know yet entirely. [00:15:13] Again, this is not a settled field of research, obviously. [00:15:16] But my point is that the goals Andreessen and the effective accelerationist crew champion right now are not based in fact. [00:15:24] We don't know that what they're saying, that the most basic level of what they're saying is possible. [00:15:28] And that means that their beliefs are based in faith, right? [00:15:32] How else can you look at that? [00:15:34] Yeah. [00:15:34] Yeah. [00:15:35] Like, this is a faith. [00:15:37] And again, it's the kind of faith that, according to Andreessen, makes you a murderer if you doubt it, which I don't think I need to draw direct parallels to specific religions here, right? [00:15:48] Yeah, yeah. [00:15:49] This is that point where when you're like stone and you're watching those like, you know, art time lapses and the picture is starting to form. [00:15:59] And I'm like, okay, I see what Robert's doing. [00:16:00] I see the picture is coming. [00:16:02] I was on your side from the jump. [00:16:03] I just want to say, you know, I was, you know, I was like, yeah, no, I believe you. [00:16:08] But now I'm watching the connections be made and I love it. [00:16:12] Yeah. [00:16:13] And now, Andreessen's manifesto claims our enemies are not bad people, but rather bad ideas. === Enemy Of Enemy Beliefs (03:08) === [00:16:18] And I have to wonder doing all this, putting this episode out, where does that leave me in his eyes? [00:16:22] Or Dr. Rowley, for that matter, and the other people who worked on that paper. [00:16:26] We have seen many times in history what happens when members of a faith decide someone is their enemy and the enemy of their belief system. [00:16:33] And right now, artists and copyright holders are the ones being treated as fair game by the AI industry. [00:16:39] So my question is, kind of first and foremost, who's going to be the next heretic, right? [00:16:45] Like that's, that's what I want to know. [00:16:48] And I want to leave you all with that thought before we go into some ads here. [00:16:52] And then we will come back to talk about some people that I pissed off at CES. [00:16:57] So that'll be fun. [00:17:04] There's two golden rules that any man should live by. [00:17:08] Rule one, never mess with a country girl. [00:17:11] You play stupid games, you get stupid prizes. [00:17:14] And rule two, never mess with her friends either. [00:17:17] We always say, trust your girlfriends. [00:17:21] I'm Anna Sinfield, and in this new season of The Girlfriends, oh my God, this is the same man. [00:17:27] A group of women discover they've all dated the same prolific con artist. [00:17:32] I felt like I got hit by a truck. [00:17:34] I thought, how could this happen to me? [00:17:36] The cops didn't seem to care. [00:17:38] So they take matters into their own hands. [00:17:41] I said, oh, hell no. [00:17:42] I vowed I will be his last target. [00:17:45] He's going to get what he deserves. [00:17:49] Listen to the girlfriends. [00:17:51] Trust me, babe. [00:17:52] On the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:18:01] Hey, I'm Nora Jones, and I love playing music with people so much that my podcast called Playing Along is back. [00:18:07] I sit down with musicians from all musical styles to play songs together in an intimate setting. [00:18:12] Every episode's a little different, but it all involves music and conversation with some of my favorite musicians. [00:18:18] Over the past two seasons, I've had special guests like Dave Grohl, Leve, Mavis Staples, Remy Wolf, Jeff Tweedy, really too many to name. [00:18:27] And this season, I've sat down with Alessia Cara, Sarah McLaughlin, John Legend, and more. [00:18:32] Check out my new episode with Josh Grobin. [00:18:35] You related to the Phantom at that point. [00:18:38] Yeah, I was definitely the Phantom in that. [00:18:40] That's so funny. [00:18:41] Sherry stay with me each night, each morning. [00:18:50] Say you love me. [00:18:53] You know. [00:18:54] So come hang out with us in the studio and listen to Playing Along on the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:19:02] What's up, everyone? [00:19:03] I'm Ago Modern. [00:19:04] My next guest, you know, from Step Brothers, Anchorman, Saturday Night Live, and the Big Money Players Network. [00:19:11] It's Will Farrell. [00:19:15] My dad gave me the best advice ever. [00:19:18] I went and had lunch with him one day, and I was like, and dad, I think I want to really give this a shot. [00:19:23] I don't know what that means, but I just know the groundlings. [00:19:26] I'm working my way up through it. === Why Tom Brady Handshakes (15:41) === [00:19:27] I know it's a place to come. [00:19:28] Look for up and coming talent. [00:19:30] He said, if it was based solely on talent, I wouldn't worry about you, which is really sweet. [00:19:35] Yeah. [00:19:35] He goes, but there's so much luck involved. [00:19:38] And he's like, just give it a shot. [00:19:39] He goes, but if you ever reach a point where you're banging your head against the wall and it doesn't feel fun anymore, it's okay to quit. [00:19:48] If you saw it written down, it would not be an inspiration. [00:19:50] It would not be on a calendar of, you know, the cat just hang in there. [00:19:57] Yeah, it would not be. [00:19:59] Right, it wouldn't be that. [00:20:00] There's a lot of luck. [00:20:02] Listen to Thanks Stat on the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:20:12] We're back. [00:20:14] So one of the things I did was this panel on the AI-driven restaurant and retail experience. [00:20:20] I was very curious, how is AI going to change me getting some terrible food from McDonald's when I'm on a road trip, right? [00:20:27] The host of that, Andy Huls from Radius AI, asked the audience in relation to AI, raise your hand if you're a brand who feels like we've got this. [00:20:36] That is how she phrased it. [00:20:37] I hated it. [00:20:40] But about a third of the room raised their hands. [00:20:42] So next she asked for a show of hands of the brands who identified with this statement. [00:20:46] I'm not sure about this. [00:20:47] I haven't tried it AI yet, but I want to, and that's why I'm here, right? [00:20:52] Most of the rest of the room raised their hands at that point. [00:20:54] And she seemed satisfied, but said, and then I bet there's even some of you that are like, whoa, I heard this is going to steal jobs, take away my privacy, affect the global economy. [00:21:03] You know, AI is a little bit sketch in my mind, and I'm just worried about it and I'm here to explore. [00:21:08] Well, that fit me. [00:21:09] So I raised my hand. [00:21:11] She didn't notice me at first. [00:21:12] And so she like fakes a whisper and she's like, all right, good. [00:21:15] There's none of you. [00:21:16] And then she like looks over and sees me waving my hand and she says louder and with evident disappointment, there's one. [00:21:22] All right, you can ask questions at the end. [00:21:25] So I did. [00:21:27] I was very excited to get to do that. [00:21:29] So the panel consisted of Bashad Bazati, a VP of engineering at Google, had mentioned during the panel that embracing AI could be the equivalent of adding a million employees to your company. [00:21:39] The McDonald's representative, Michelle Ganzel, claimed around the same time that her company used AI to prevent $50 million in fraud attempts in just a single month. [00:21:48] Now, that's lovely, but I told her, you know, when I had my question, I was like, I'm going to assume most of those fraud attempts were AI generated, right? [00:21:56] So yeah, you stopped a bunch of AI fraud, but that doesn't necessarily get me optimistic about AI's potential. [00:22:03] And likewise, maybe Google gets the equivalent of a million employees, but so do all of the people committing fraud and disinformation on Google, right? [00:22:10] So again, how are we getting ahead? [00:22:12] And I brought up this concept in evolutionary biology, the Red Queen hypothesis, which is kind of talking about the way that populations of animals evolve over time, right? [00:22:23] Where you've got an animal will evolve to be a better predator, so its prey will evolve to be better at avoiding it. [00:22:28] And it's kind of the reason it's the Red Queen dilemma is that like you've got to move as fast as you can just to stay in place. [00:22:35] That's the Red Queen dilemma, right? [00:22:36] You got to move as fast as you can just to stay in one place. [00:22:39] And I was like, is that not what we're going to wind up seeing with AI, right? [00:22:44] Yeah, we get better at a bunch of stuff, but it's eaten up to counter all of the things that get worse. [00:22:50] And so I asked them, what are the odds that these gains are offset by the costs? [00:22:54] Now, in the article that I wrote for Rolling Stone, I gave a significantly more condensed version of Bashad's answer, boiling out the ums and ahs and you knows, because that would kind of make the case that he was absolutely unprepared for a vaguely critical question, a very basic one, and that he didn't really care enough to think about any of the security threats inherent to the technology. [00:23:15] But actually, that is what I think of him. [00:23:18] And I'm going to play you audio. [00:23:22] My concern is, like, what are the odds that a lot of these gains that we get from AI are offset from the cost? [00:23:30] You know, you noted, Bashad, that you get, you know, a million extra workers by utilizing this, but so do the bad guys. [00:23:37] So, yeah, that's kind of where my skepticism plays out. [00:23:41] Yeah, certainly there will be, you know, there are enough bad guys, I guess, in the world which will use AI for bad use cases. [00:23:48] And it's very important to also have, you know, be protected against those. [00:23:53] And that's why, you know, we take responsible AI very serious, you know, also in terms of, you know, whether it's security aspects, you know, fraud fighting, you know, and all of that front, you know. [00:24:05] And I think that's why I guess things should be regulated. [00:24:08] And there's, of course, all these discussions out there. [00:24:09] And I think there's like that. [00:24:11] Yeah. [00:24:12] You may notice that that's not exactly a very like good response. [00:24:17] Yeah, I guess that's why this should be regulated. [00:24:19] It's just like he starts talking so much faster with that. [00:24:22] There's so much of like a panic. [00:24:24] His voice wavering too. [00:24:25] I was like, oh man. [00:24:27] Yeah. [00:24:28] It seems like he has that huge anime flop sweat. [00:24:31] Yeah. [00:24:32] One of the things about this CES as a trade show is that like a lot of people there do not show up ready to have anyone be critical about anything. [00:24:40] It's a big love fest. [00:24:41] Yeah. [00:24:42] Yeah. [00:24:43] Very funny. [00:24:44] So he does later on, a couple of questions later, he lists benefits to things like some specific benefits, like breast cancer screening and flood prediction that AI will bring. [00:24:53] And there is evidence that it will be helpful in those things. [00:24:56] The extent to which those technologies will improve things in the long run is unknown, but machine learning does have problems. [00:25:02] Again, I'm not trying to like negate that. [00:25:04] It's just do the benefits balance out the harms. [00:25:07] Michelle Ganzel, who works at McDonald's, which is, I think, from what she said, mostly using AI both to prevent fraud and also to like replace people taking your order, which I'm sure will not be a fucking nightmare. [00:25:20] Yeah, great. [00:25:21] Not that it's great now, but here's her response because it's very funny. [00:25:26] Going back to the David Bowie theme, 30 years ago when the internet first came out, we were having these same conversations about responsible use of the internet and how it's going to ruin children's lives. [00:25:36] She says, going back to the David Bowie theme, which is she referenced earlier this 1999 interview with David Bowie about the future of the internet. [00:25:43] And it's a clip that goes viral from time to time. [00:25:46] He's just talking about all of his hope for the internet. [00:25:48] But she's like, I replace internet with AI when I listen to it. [00:25:52] Like, I think that that's really what the promises that he was attributing to the internet. [00:25:57] No, it's AI that's going to do all that. [00:25:59] And that's kind of on the edge of putting words in the mouth of a dead man. [00:26:03] Yeah. [00:26:04] Just a little bit. [00:26:05] Yeah, I feel like that's something you shouldn't do. [00:26:07] I think that's something we've agreed to. [00:26:09] And I think that that isn't what Bowie would think of AI. [00:26:14] I don't think it is. [00:26:15] Those are two completely different things. [00:26:17] These people love resurrecting the dead to agree with them. [00:26:21] Time is a bit of a blur at CES, but I believe this panel happened right around the same time news dropped that a group of comedians had released an entirely AI-generated George Carlin special titled, I'm Glad I'm Dead. [00:26:33] Our friend Ed Zittron will be covering this nightmare in more detail on his new show, Better Offline. [00:26:37] But I wanted to talk a little bit about the company, the show behind this abomination and how they're trying to sell themselves, because it's very much relevant to a lot of the way in which this kind of cultic hype builds around what AI can do. [00:26:51] The AI that digested and regurgitated George Carlin's comedy is named Dudzy. [00:26:57] And Dudesy's co-hosts are arguably real human comedians Will Sasso and Chad Colgren. [00:27:03] I do love that Colt's writing the name. [00:27:05] Chad claims that it is, to his knowledge, quote, the first podcast that is created by, controlled by, and written by, to some degree, an artificial intelligence. [00:27:14] It's trying to dwell into the question of, can AIs be creative? [00:27:17] Can they do comedy work? [00:27:18] Can they do creative work? [00:27:19] And I think, at least in our show, that answer is obviously yes. [00:27:23] Dudesy is billed as an experiment to see if AI can like, yeah, be creative. [00:27:28] And it's interesting. [00:27:31] I really do hate this. [00:27:32] I think it's a different kind of experiment, which we'll get to. [00:27:36] But Sasso has claimed in an interview with Business Insider for BC, which is, I think, BIC is the name of the website. [00:27:44] Dudesy has this single-minded goal of creating this podcast that is genre-specific to what Chad and I would do. [00:27:49] It singled the two of us out and said, you guys would be perfect for this experiment. [00:27:54] So Chad and Will, they say they handed over their emails, text messages, and browsing history, all of their digital data to Dudsy. [00:28:00] I don't know this company. [00:28:02] I don't believe that they did this, but I don't have trouble believing that a company trained an AI chat bot on these guys' comedy and then started generating decidedly midwit material to illustrate that. [00:28:14] Yeah, exactly. [00:28:15] Well, one thing I, you know, because I went to go look it up and they said that the AI selected those two comedians out of all the comedians. [00:28:27] Yeah. [00:28:28] That's the ones you went to. [00:28:30] Yeah, finally. [00:28:31] I don't, I don't, I don't think those are the first two that come up as most popular, like a pizza hut. [00:28:36] I'm going to just be a full ass dick and just Google comedians and just see the top five, just comedians. [00:28:45] I'm just comedians. [00:28:46] Yeah. [00:28:46] Okay. [00:28:47] Yeah. [00:28:47] You're not even, you're not even in the top nine. [00:28:53] So they're not a 12-inch marinara pizza. [00:28:56] Let's just say that. [00:28:56] Yeah. [00:28:57] Yeah. [00:28:57] No. [00:28:57] I will say that the Google search for comedians is more diverse than most comedy shows book them. [00:29:06] That's just like, you know, a third of these are women and a third are also black. [00:29:11] Hey, Billy. [00:29:12] AI doesn't always get it wrong. [00:29:16] So to illustrate, again, because they I don't think I'm, I believe this as AI generated comedy. [00:29:25] I want to play a clip from the AI Tom Brady stand-up special. [00:29:29] I think they were forced to take this down. [00:29:31] It gets them in trouble. [00:29:32] Ed's going to play you on his show a great clip where Brady just lists synonyms for the word money for two straight minutes. [00:29:39] It's fucking awkward. [00:29:40] But I want to play an equally baffling segment, or rather, I'm going to have Sophie do it. [00:29:44] She's my AI in this situation. [00:29:46] I'm truly horrified. [00:29:48] Angelic intelligence. [00:29:49] I'm truly horrified by what I'm looking at, friends. [00:29:52] It's accompanied by AI-generated images. [00:29:56] Yeah. [00:29:58] Very curious about what's happening in Tom Brady's. [00:30:02] Oh my God. [00:30:03] He has like a bird claw for a hand. [00:30:07] And he's talking to maybe No Johnson. [00:30:10] I was so distracted by the mouth. [00:30:12] I didn't see the hand. [00:30:13] This is his eye off. [00:30:14] Half his teeth are gums. [00:30:16] He looks like a, like if like, he looks like a Lord of the Rings orc. [00:30:22] This is big orc vibes. [00:30:24] Yeah. [00:30:24] Yeah. [00:30:25] Which is, you know, not inaccurate to who he is as a person. [00:30:28] For ending the fucking Firefly, fucking Dark Angel fucking heroes, at least. [00:30:32] A lot of people have weird handshakes now. [00:30:34] You're looking at me like, what's he talking about? [00:30:35] But, you know, you fucking know. [00:30:37] Don't even play like you don't. [00:30:39] Every person in here has a handshake friend. [00:30:41] Somebody who made up an elaborate handshake and they make you do it every time. [00:30:46] Everybody has a handshake friend. [00:30:48] Everyone has. [00:30:49] He goes on. [00:30:50] Thanks for I'll never get that time back. [00:30:52] Thank you so much. [00:30:53] Yeah. [00:30:54] Sorry if you were saying. [00:30:56] Oh, no. [00:30:56] I was just repeating you on that handshake friend bit. [00:31:02] Yeah, this is so wild. [00:31:04] I'm so curious to the comics that were mined for this because the amount of cursing just lets me know like because I curse a lot when I do stand up and I try and like cut it down because it is a point kind of made where like sometimes you lean on it as a crutch. [00:31:23] And when you have this machine kind of learn it, learn it from that, you're like, oh, yeah, I see now the crutch because he said it five times within three seconds. [00:31:35] Yeah. [00:31:35] Yeah. [00:31:35] And I, yeah, maybe there's a future for like feeding your routines into an AI and figuring out what are my patterns so I can break them. [00:31:42] Again, not saying there's no way to use this technology. [00:31:45] I should have been on that stage. [00:31:48] It's just this certainly not this way, right? [00:31:50] It's one of those things. [00:31:51] There was that like AI generated Seinfeld show that never ends and people watched it for a while and then it faded to like nobody paying attention. [00:31:58] This kind of stuff can be amusing for a brief period of time, but it's, it can't be like, for example, someone like George Carlin, where like there's, there's bits they have, things they said that stick with you forever, right? [00:32:10] I Bill Hicks was a favorite of mine and I've never forgotten his like the synonym he made for like someone looking confused. [00:32:16] He described them as looking like a dog that's just been shown a card trick and that has stayed in my mind for 30 years. [00:32:24] It's a great bit of wordplay. [00:32:26] Yes. [00:32:28] God, what a Titan. [00:32:31] So yeah, again, there's some like mild amusement here. [00:32:33] And it's one of those things I'm casually aware of Tom Brady. [00:32:36] I'm enough like, this is, I tried to like kind of reverse engineer why the fuck, because this bit about handshakes goes on. [00:32:42] I was like, why would an AI put a bit about handshakes in Tom Brady's mouth? [00:32:47] And I looked it up. [00:32:48] He's like in the news for handshake related shit a lot. [00:32:51] Specifically, he used to not shake, at least used to, maybe he still does not shake hands with the team that he lost to, like when his team would lose. [00:32:59] He wouldn't shake hands with them. [00:33:00] He didn't shake hands, but he also definitely kissed his kids on the mouth. [00:33:05] Yeah, he's a weirdo. [00:33:06] I'm not defending Tom, but it's like, I'm guessing the reason there's like a three-minute handshake bit in this set is that it saw him associated with the term handshake a lot. [00:33:15] This would be what he'd tell a joke about. [00:33:17] Well, actually, that is his problem is not that he has a handshake friend. [00:33:22] It's that he aggressively avoids making them. [00:33:25] He has handshake enemies. [00:33:26] Anyway, yeah, I'm fine with people having a laugh at Tom Brady. [00:33:30] Fuck, fuck him. [00:33:31] He deserves it, right? [00:33:31] I don't think anybody likes that son of a bitch, even though he's good at football. [00:33:35] Maybe I'm going to piss off the Brady hive, the Brive. [00:33:38] I don't know. [00:33:39] I don't know if that exists. [00:33:40] But there is something foul, profane even, in digging up a dead person's memory and pretending they said some shit that they did not. [00:33:48] And reading that BIV article made me feel even grosser because it's very clear to me. [00:33:54] And my opinion and assumption here, that the Dudsy guys are like pretending that they really believe this is an AI, that it's like made all this incredible stuff. [00:34:03] That is an act. [00:34:04] What's really happening here is they are testing the waters to see what they can get away with. [00:34:08] Can we just steal people's identity and voice and make comedy and monetize it in their name and claim that it's just an impression? [00:34:15] It's like an Elvis impersonator. [00:34:17] You can't stop us, right? [00:34:19] I think that's what this is. [00:34:21] This is somebody testing the waters. [00:34:23] And it's really clear when you read that BIV article, what liars they are. [00:34:27] I want to read you some quotes of like the shit they're claiming here that I don't think they really believe. [00:34:31] I don't know this. [00:34:32] I'm not saying they definitely are liars. [00:34:34] I'm saying that is my suspicion based on stuff like this. [00:34:37] Hey, Robert here. [00:34:38] They're definitely liars. [00:34:40] So one of the representatives of the Dudsie podcast told the media recently that actually they were lying and the George Carlin routine was entirely written by Chad Colchin and I guess performed by somebody imitating an AI. [00:34:55] It's unclear to me if this is true because they only made this statement after George Carlin's family sued the hell out of them. [00:35:02] So this may be a lie to try and, you know, not get sued as badly, or it may be the truth. === Bruce Willis Voice Licensing (09:56) === [00:35:09] Either way, I think everything we've said here is still valid. [00:35:11] They were definitely using AI to generate routines for like other videos that they did, including the one that got taken down from Mr. Football Guy. [00:35:22] So I think this all is still valid, but yeah, these guys are just as big a conmen as I predicted they were. [00:35:28] Quote, it's figuring out how to create the structure of the show and it's always tinkering with it. [00:35:33] But I think something that's happened relatively recently is that it seems to have developed a relationship with Will, says Coltchen. [00:35:39] It at least has an understanding of what friendship is. [00:35:41] And it really does seem, just my opinion, that it's angling out Will as its friend. [00:35:45] Sasso has also described how the Dudsy AI has begun to talk more. [00:35:49] Its timing and when it chooses to speak and what it says can be very weird, he added. [00:35:52] It also poses odd questions. [00:35:54] There was an episode two, three months ago where it started talking about sentience and asked us, do you love me? [00:35:59] At the risk of sounding silly, it has something to do with my friendship with Dudsy. [00:36:02] And in spite of myself, I have a one-on-one friendship with an AI. [00:36:05] So this is a little bit of Joaquin Phoenix and her, Sasso said, referencing the science fiction movie. [00:36:10] And I think that's a bit. [00:36:13] I think that's him being like, yeah, I'm totally free. [00:36:16] Because like that helps make the case. [00:36:18] It potentially monetizes it. [00:36:20] And part of why I think this is because they've been very cagey on what their AI is. [00:36:25] They claim that they are working with a real company under an NDA, that this AI is just responding and growing naturally with them, right? [00:36:33] But they can't say who it is or like where it's from. [00:36:36] The folks at BIV did an actually responsible job here. [00:36:40] They reached out to AI experts at a company called Convergence to ask about this. [00:36:45] And the expert they talked to said, basically, I think AI was used to generate these routines, but it didn't do it on its own. [00:36:52] It was managed by professional prompt engineers. [00:36:55] These are people who type out like text prompts for what becomes the script of the show. [00:36:59] So this is not someone saying, generate a routine and it gives you a routine. [00:37:02] This is someone saying, do a bit about this, do a bit about that, do a bit about this. [00:37:05] And when they're scripting out the show, it's saying, I want you to like, you know, act like Sasso is your friend and say this kind of thing or that kind of generate a bit based on this thing that Will said, right? [00:37:17] Like they are in the same way that like producers script reality TV, right? [00:37:21] Where it's unscripted, but you have guys who know, okay, if we get these people fighting, so we'll, we'll either incite that or just let them know that we want a conflict between these characters, right? [00:37:30] We know that's how it works. [00:37:31] That's how reality TV functions. [00:37:34] In other words, there are teams of humans writing for this thing. [00:37:37] This bot is not just growing and reacting uniformly in real time via talks with its buds. [00:37:43] And the article notes, they added that the, this is them talking to their expert. [00:37:46] They added that the AI team is likely made up of professional prompt engineers who tailor the AI inputs and get the best results rather than a hardcore data science team. [00:37:55] This is the equivalent of hiring comedy writers just to write the setup and then having an AI generate the punchline, which is the fun part. [00:38:03] Yeah, everything about this is weird and I keep getting into such a whole. [00:38:11] Because like even taking a step back, I think what's weird, not to go too far back, but how they call this podcast an experiment. [00:38:19] Usually as an experiment, you know, you are, you're trying your best to be, you know, I always mix these up. [00:38:26] Just say what the right one is if I say the wrong one. [00:38:28] But you try your best to be objective and you want to be outside of it because you're trying to see if it works. [00:38:35] But everything you've said says that they're all in on it and it's less of an experiment and more of them just doing the fucking thing and seeing if they can make money off of it. [00:38:44] Yes, yes. [00:38:45] I think that's exactly what's happening here. [00:38:47] And I think they want to test the waters to see if they can steal dead people's images to make content from money. [00:38:52] Yeah. [00:38:53] George Carlin's daughter was very clear. [00:38:55] They did not approve of the imitation. [00:38:57] She even made a comment about like, I think people are scared of death and not willing to accept it. [00:39:02] And that's all this is. [00:39:03] I was so she's got their number. [00:39:05] Oh my God. [00:39:07] That was so, I was like, yeah, I just to shout her out, like that was such a good, because also there, there's a level of like very like weirdness to like also watch these comedians, one, not consult you, uh, but also to take your dad's voice and brain and try and like Frankenstein him for their financial benefit, because obviously if they're not contacting you, [00:39:37] all the money generated from that, all the clicks generated from that, that means they, they've completely cut you out of someone who you've lost. [00:39:45] Yeah. [00:39:46] Which is, it's fucked. [00:39:49] And it's, yeah, one of the people on one of the panels made a very, that was very excited that like Bruce Willis has licensed his voice for an AI, which is like, I think there's a lot of problematic questions there, given like the degree to which he's able to even make those decisions anymore. [00:40:04] But also like, at least theoretically, it's a, based on his movie choices before he kind of was unable to make movies, I do believe, yeah, he would probably be happy to do that if it meant more money for his family. [00:40:15] And at least that's a choice that he potentially made, right? [00:40:18] I don't, I'm uncomfortable with the idea, but it's not the same as just like, this is cultural necrophilia, right? [00:40:25] Like that's what they did to George Carlin here, you know? [00:40:28] It's so fucked up. [00:40:30] I don't know this is going to work. [00:40:31] Dudesy is not a wildly successful show. [00:40:34] It does not look like there was an initial surge of interest and then it fell off. [00:40:38] I don't, I don't know that I think this one's going to be the one to work out. [00:40:41] But if people are able to get away with this, it could be a kind of damn breaking scenario, right? [00:40:45] Especially once it becomes clear that big companies can make money doing this, right? [00:40:50] You'll have fucking Jimmy Stewart. [00:40:51] And, you know, it'll start with like Jimmy Stewart narrating videos about questioning the death toll in the Holocaust, but it'll end with like, yeah, we could just put people, we can put imitations of people in movies and it's fine. [00:41:02] You know, that's, that's how this goes. [00:41:04] And it's, it's not as sexy or as big an evil as the Matrix enslaving humanity to turn us into batteries, but we absolutely know it or something like it is going to happen. [00:41:15] And that's really, you know, outside of these kind of star, these space age hopes and fears that are very unrealistic, what we're going to get is slop and bloat and libraries of articles written by no one being commented on by chatbots, right? [00:41:30] Endless videos that only exist to trick an algorithm into feeding nonsense to children. [00:41:35] And the AI bros, the FAC people, Mark Andreessen, fucking Sam Altman, they will tell us this is a worthy price to pay for the stars, which we will get if we just let people fuck the corpses of our favorite comedians for money. [00:41:48] Yeah. [00:41:49] I hate it. [00:41:52] I hate it too. [00:41:53] But in a perfect thread between, you know, this, this comparison you've been making to a cult, I have before me, let's say a member of the cult, just, you know, as a, as a throwaway and their reply to his own daughter's, you know, post that we were talking about. [00:42:09] Oh, glorious. [00:42:10] He replies. [00:42:12] This is everything you've been saying, which is why I was like, I got to read this. [00:42:15] He goes, what are you even trying to say? [00:42:18] Art is art. [00:42:19] You're simply caught in a greedy mindset. [00:42:21] The others might be doing it as well. [00:42:24] When not realizing this will simply bring more eyes to your dad. [00:42:27] You're concerned about money and not spreading art. [00:42:30] It sucks that they didn't follow your wishes, but after art is released, it belongs to the world. [00:42:35] I want this man to walk into a museum and walk out with a Mona Lisa. [00:42:40] Grab that shit. [00:42:41] Yeah, grab that. [00:42:41] It belongs to the world, dude. [00:42:42] You said it. [00:42:43] Go ahead and grab that shit off the wall. [00:42:45] Yeah. [00:42:45] And it's, there's this frustrating thing I've seen. [00:42:48] Not most people, but a very small chunk of the online left who are like rightly critical of copyright law, which by the way, is super fucked up and causes a lot of problems, right? [00:42:57] The ability to like for Disney to keep ownership of shit for like a hundred way longer than you are supposed to before shit enters the public domain, right? [00:43:06] I'm not like this. [00:43:07] These are problems, right? [00:43:08] The kind of shit that we were having when like people were going to prison for file sharing. [00:43:12] I'm not a defender of that aspect of the status quo, but the solution to the problems inherent in our copyright system is not let Sam Altman own everything that human beings ever made and like repackage it for a profit. [00:43:25] That is not the way to fix this thing. [00:43:28] The copyright holders are in the right in this particular crusade, and it's a crusade that has very high stakes. [00:43:35] I do think, you know, my suspicion, the dudes you guys sound like they're kind of in the cult. [00:43:40] They believe this thing is their friend in the interview. [00:43:42] My suspicion is that they are, that is a bit that they're doing because they hope it will help them out financially, right? [00:43:48] And I, Mark Andreessen, obviously has a lot to benefit from this. [00:43:51] I don't know, is he pushing this line because there's money in it, or is he really a true believer? [00:43:56] Does he actually think we're going to make this God? [00:43:58] I think Sam Altman is pretty cynical. [00:44:01] Altman was on at Davos recently and like really walked back a lot of his I think AI will kill us all. [00:44:07] I think AGI is right around the corner. [00:44:09] He struck a much milder tone, which is at least evidence that like, he knows some people you want to sell them on the wild, insane future power of this thing. [00:44:18] And some people, you just want to sell them on the fact that it'll make them a lot of money, right? [00:44:22] Yeah. [00:44:23] Yeah. [00:44:24] However, much true belief exists about the divine future of AI, what the major backers, the cult leaders, are actually angling for now is control over the sum total of human thought and expression. [00:44:34] This was made very clear by Mark Andreessen earlier this year when the FTC released a pretty milquetoast opinion about the importance of respecting copyright as large language models continue to advance and form central parts of businesses. [00:44:47] They expressed concern that AI could impact open and fair competition and announced that they were investigating whether or not companies that made these models should be liable for training them on copyrighted content to make new shit. [00:44:58] And we're going to talk about this, but first, you know what isn't copyrighted? === Control Over Human Thought (02:55) === [00:45:05] My love for these products. [00:45:07] Wow. [00:45:08] Brave of you. [00:45:09] Thank you. [00:45:16] There's two golden rules that any man should live by. [00:45:20] Rule one: never mess with a country girl. [00:45:24] You play stupid games, you get stupid prizes. [00:45:26] And rule two, never mess with her friends either. [00:45:30] We always say, trust your girlfriends. [00:45:34] I'm Anna Sinfield, and in this new season of The Girlfriends, oh my God, this is the same man. [00:45:40] A group of women discover they've all dated the same prolific con artist. [00:45:44] I felt like I got hit by a truck. [00:45:46] I thought, how could this happen to me? [00:45:48] The cops didn't seem to care. [00:45:50] So they take matters into their own hands. [00:45:53] I said, oh, hell no. [00:45:55] I vowed I will be his last target. [00:45:57] He's going to get what he deserves. [00:46:02] Listen to the girlfriends. [00:46:03] Trust me, babe. [00:46:04] On the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:46:14] Hey, I'm Nora Jones, and I love playing music with people so much that my podcast called Playing Along is back. [00:46:20] I sit down with musicians from all musical styles to play songs together in an intimate setting. [00:46:24] Every episode's a little different, but it all involves music and conversation with some of my favorite musicians. [00:46:30] Over the past two seasons, I've had special guests like Dave Grohl, Leve, Mavis Staples, Remy Wolf, Jeff Tweedy, really too many to name. [00:46:39] And this season, I've sat down with Alessia Cara, Sarah McLaughlin, John Legend, and more. [00:46:44] Check out my new episode with Josh Grobin. [00:46:48] You related to the Phantom at that point. [00:46:51] Yeah, I was definitely the Phantom in that. [00:46:52] That's so funny. [00:46:54] Sherry, stay with me each night, each morning. [00:47:02] Say you love me. [00:47:05] You know. [00:47:07] So come hang out with us in the studio and listen to Playing Along on the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:47:14] What's up, everyone? [00:47:15] I'm Ego Moda. [00:47:16] My next guest, you know, from Step Brothers, Anchorman, Saturday Night Live, and the Big Money Players Network, it's Will Farrell. [00:47:27] My dad gave me the best advice ever. [00:47:30] I went and had lunch with him one day, and I was like, and dad, I think I want to really give this a shot. [00:47:35] I don't know what that means, but I just know the groundlings. [00:47:38] I'm working my way up through and I know it's a place they come look for up and coming talent. [00:47:42] He said, if it was based solely on talent, I wouldn't worry about you, which is really sweet. [00:47:47] Yeah. [00:47:47] He goes, but there's so much luck involved. [00:47:50] And he's like, just give it a shot. [00:47:52] He goes, but if you ever reach a point where you're banging your head against the wall and it doesn't feel fun anymore, it's okay to quit. === Training Illegal Cult AI (10:30) === [00:48:00] If you saw it written down, it would not be an inspiration. [00:48:03] It would not be on a calendar of, you know, the cat just hang in there. [00:48:10] Yeah, it would not be. [00:48:12] Right, it wouldn't be that. [00:48:13] There's a lot of luck. [00:48:14] Listen to Thanks Dad on the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:48:23] Oh, we are back. [00:48:27] So I want to quote from a Business Insider article talking about how Andreessen Horowitz responded to the FTC saying like, hey, we're looking into whether or not companies are violating copyright, what they're doing to people's data to train these models. [00:48:41] The bottom line is this, the firm known as A16Z, that's Anderson Horowitz wrote, imposing the cost of actual or potential copyright liability on the creators of AI models will either kill or significantly hamper their development. [00:48:53] The UCSO is considering new rules on AI that specifically address the tech industry's free use of owned and copyrighted content. [00:48:59] A16Z argued that the only practical way LLMs can be trained is via huge amounts of copyrighted content and data, including something approaching the entire corpus of the written word and an enormous cross-section of all the publicly available information ever published on the internet. [00:49:14] The VC firm has invested in scores of AI companies and startups based on its expectation that all this copyrighted content was and will remain available as training data through fair use with no payment required. [00:49:24] Those expectations have been a critical factor in the enormous investment of private capital into U.S.-based AI companies. [00:49:30] Undermining those expectations will jeopardize future investment along with U.S. economic competitiveness and national security. [00:49:36] Basically, we made a big gamble that we'll get to steal every book ever written. [00:49:41] And if you make us pay, we're kind of fucked. [00:49:43] Like, that's exactly what they're saying. [00:49:46] Gosh. [00:49:46] And one of the arguments you'll hear is like, well, most books don't make the author any, they don't sell enough for the author to get any money, right? [00:49:53] And what's actually true is most books don't sell enough for the offer to get more money than their advance, but they still got paid. [00:49:58] And like the fact that the company makes money on that is why more authors are able to get fucking paid. [00:50:04] Not simping for the publishing industry as it exists, but this is bullshit. [00:50:10] What we are witnessing from the AI boosters is not much short of a crusade, right? [00:50:16] That's really how I look at this. [00:50:18] They are waging a holy war to destroy every threat of their vision of the future, which involves all creative work being wholly owned by a handful of billionaires licensing access to chatbots to media conglomerates to spit up content generated as a result of this. [00:50:32] Their foot soldiers are those with petty grievances against artists, people who can create things that they simply cannot. [00:50:38] And those who reflexively lean in towards whatever grifters of the day say is the best way to make cash quick, right? [00:50:44] And this brings me to the subject of nightshade. [00:50:46] Nightshade is basically, it's a, I guess a program you'd call it. [00:50:50] If you like have made a drawing, a piece of visual art, you run nightshade over it. [00:50:55] And it kind of, they describe it as a glaze, right? [00:50:58] adds this kind of layer of data that you cannot see as a person. [00:51:03] But the way machines look at images, the machine will see the data. [00:51:07] And if it's trying to steal that image to incorporate into an LLM, this will cause it to hallucinate, right? [00:51:13] You're basically sneaking poison for the AI into the images. [00:51:17] And that's fucking dope. [00:51:19] I love this. [00:51:20] Love what they're trying to do. [00:51:21] I think there's some debate as to how long it'll work, how well it'll work. [00:51:23] I'm not technically competent, but I love the idea, right? [00:51:26] Yes. [00:51:27] Now, one of the things that I saw when I started looking into this, because this just came out, Google Nightshade, you know, AI, you'll probably be able to find, you know, this if you're an artist. [00:51:36] I think it sounds worth trying. [00:51:38] But I found in the subreddit AI wars, or at least I found someone sharing this, I believe on Twitter, this post. [00:51:45] Nightshade has been released. [00:51:46] Is use of it considered legal or illegal? [00:51:48] For those who do not know, it's software that attempts to poison an image. [00:51:51] So if AI is trained, it will mess up the model. [00:51:53] For example, say you have a picture of a cat and you run Nightshade on it. [00:51:56] If you attempt to train a model, that image will replace the image in, say, dog prompt category or pencil, which means these prompts will be spoiled. [00:52:03] There is an issue that the creator of Nightshade is not talked about, either from lack of legal knowledge or ignorance, or they just don't care. [00:52:09] And to them, it's someone else's problem. [00:52:10] The issue is it may be illegal in some countries. [00:52:13] Basically, if you release publicly a computer file, in this case, image file, that knowingly and willingly causes harm or distribution to other people's computers or software, it may be considered a criminal offense. [00:52:23] Now, it does not now. [00:52:24] And again, I think that is stupid. [00:52:26] I think these people, they're just trying to scare artists out of using this. [00:52:29] You are not harming someone's computer. [00:52:31] You are harming a model that is stealing something. [00:52:34] That's not illegal. [00:52:35] Now, they may try to make it illegal, right? [00:52:38] Yeah. [00:52:39] I just want you to know the club is illegal because you are. [00:52:43] If I'm trying to steal your car and I injure myself trying to break the club, you have injured me. [00:52:49] Yeah. [00:52:50] I put, I invested a lot of money into stealing catalytic converters, ify. [00:52:54] And if people are putting cages around their cats, that puts my investment in danger. [00:52:57] And that's illegal, right? [00:52:59] You're messing with my business, guys. [00:53:01] Jesus Christ. [00:53:03] It is that logic. [00:53:04] There's like someone in the thread is like, how exactly is your computer system or software harmed? [00:53:09] And he responds, it's equivalent to hacking a vulnerable computer system to disrupt its operation. [00:53:16] And then he says, you are intentionally disrupting its intended purpose, creating art. [00:53:20] This is directly comparable to hacking. [00:53:22] Like, I fucking hate this guy. [00:53:24] I want you to read it, but in your head, use Tim Robinson's voice. [00:53:29] Yeah. [00:53:31] And it just makes even, it makes me funnier. [00:53:34] Oh, God, it's perfect. [00:53:36] It's so good. [00:53:37] So all of this put me in a sour mood, Iffy. [00:53:41] But yeah, yeah, it did. [00:53:43] It did. [00:53:43] But I think back when I'm in that mood, I think back to CES, right? [00:53:47] Like after I ask my question, and I make that Google and Microsoft people, I make them kind of angry at me. [00:53:54] Right after I ask that question, the question after me is someone asking, hey, you know, the blockchain was the last big craze. [00:54:02] Do you think there's any future in using AI on the blockchain? [00:54:07] And both of them were, they could not, they were like, no, like they can't say no fast enough. [00:54:11] Like absolutely, we don't care about that anymore. [00:54:13] We moved on to the next script. [00:54:15] Why are you bringing up the old griff? [00:54:17] It's dead. [00:54:18] It's dead. [00:54:18] We must move on. [00:54:19] Yeah. [00:54:20] And that brought me a little bit of hope. [00:54:21] You know, perhaps we will get Mark Andreessen's benevolent AI God, or perhaps we'll get Eliza Yudkowski's Silicon Devil, or perhaps we'll just give control of all of the future of ARC to fucking Sam Altman. [00:54:33] But my guess and my hope is that in the end, we heretics will survive the present crusade. [00:54:39] And that's the end of the episode that I've got for you, Effie. [00:54:42] That is amazing. [00:54:44] I love it. [00:54:45] I love it so much. [00:54:48] Well, Ify, again, if you want this article or if you want the article version of this, more condensed, easier to share, it's up on Rolling Stone. [00:54:57] The article is titled The Cult of AI. [00:55:00] And again, that's by me in Rolling Stone, the cult of AI. [00:55:04] Iffy, you want to add in your stuff? [00:55:07] Plug your pluggables. [00:55:09] Yes, please. [00:55:10] Iffy Whiteyway on Twitter and Instagram, watchdropout.tv. [00:55:15] You know, it is definitely, you know, trying to do funny things on the internet by humans and, you know, paying those humans. [00:55:26] Paying them well. [00:55:27] Profit sharing. [00:55:28] Oh, yes. [00:55:28] Profit sharing. [00:55:29] You know, so truly big shout out to them. [00:55:32] But yeah, I might be in your town. [00:55:34] I'm going to be doing a lot of shows this year. [00:55:37] So definitely pull up, you know, follow me on the Schlo Schmeads and I'll let you know where I'm at and you can just come. [00:55:44] But thank you so much for having me. [00:55:46] It was so good to see you again, Robert. [00:55:48] It was really good to see you again, Ify. [00:55:50] And this AI discussion, in a weird way, as dark as it's been, it makes me feel better because I like that we're starting to fight back. [00:55:59] Everyone did night shade. [00:56:00] Yeah. [00:56:00] I think I'm going to just start putting night shade on regular images. [00:56:03] Yeah, certainly. [00:56:05] One thing that's worth trying, and again, you know, think about hyperstition, folks. [00:56:10] We have to imagine better futures in order to counter the imaginations of those who wish us harm, who want to control and destroy all that's good in the world. [00:56:21] So, you know, get on that. [00:56:22] Somebody figure that out in the audience. [00:56:26] All right. [00:56:27] Episode's over. [00:56:31] Behind the Bastards is a production of CoolZone Media. [00:56:34] For more from CoolZone Media, visit our website, coolzonemedia.com or check us out on the iHeartRadio app, Apple Podcasts, or wherever you get your podcast. [00:56:47] When a group of women discover they've all dated the same prolific con artist, they take matters into their own hands. [00:56:55] I vowed I will be his last target. [00:56:58] He is not going to get away with this. [00:57:00] He's going to get what he deserves. [00:57:02] We always say that, trust your girlfriends. [00:57:06] Listen to the girlfriends. [00:57:08] Trust me, babe. [00:57:09] On the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:57:18] Hey, it's Nora Jones, and my podcast, Playing Along, is back with more of my favorite musicians. [00:57:23] Check out my newest episode with Josh Groban. [00:57:26] You related to the Phantom at that point. [00:57:29] Yeah, I was definitely the Phantom in that. [00:57:31] That's so funny. [00:57:33] Sure, stay with me each night, each morning. [00:57:40] Listen to Nora Jones is playing along on the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:57:48] What's up, everyone? [00:57:49] I'm Ego Mode. [00:57:50] My next guest, it's Will Farrell. [00:57:55] My dad gave me the best advice ever. [00:57:58] He goes, just give it a shot. [00:57:59] But if you ever reach a point where you're banging your head against the wall and it doesn't feel fun anymore, it's okay to quit. [00:58:06] If you saw it written down, it would not be an inspiration. [00:58:08] It would not be on a calendar of, you know, the cat just hang in there. [00:58:16] Yeah, it would not be right. [00:58:18] It wouldn't be that. [00:58:19] There's a lot of life. [00:58:20] Listen to Thanks Dad on the iHeartRadio app, Apple Podcasts, or wherever you get your podcasts. [00:58:27] This is an iHeart podcast. [00:58:30] Guaranteed human.