Info Warrior - Jason Bermas - Musk Dittmann And The Media! X Is Not Free Speech! Aired: 2024-02-25 Duration: 34:30 === Adrian Dittman's Real Time Tweet (11:40) === [00:00:02] We were theorizing, me and a few other people were theorizing about this in space a very long time ago. [00:00:07] We're saying, you know, if you can control the data set, you control reality, right? [00:00:10] Because the thing that the AI outputs is taken by the majority as fact, much like anyone else would look at a legacy media publication, right? [00:00:18] This is why it is so effective. [00:00:19] And so we sat there and we thought about how exactly could you, say, make an AI do some really terrible things without anyone actually noticing? [00:00:26] Well, of course, control the data set and, of course, apply heuristics in certain areas where there shouldn't be any, right? [00:00:32] And I think we got a very interesting scale model demonstration of this firsthand through an overplay of the hand by Gemini, which is honestly kind of a letdown. [00:00:43] You know, I like Google and what they stand for, well, stood for, really, because right now I think the inverse of what they... [00:00:49] I think the inverse of that is true today, which is really unfortunate. [00:00:53] Because it is a really good company. [00:00:54] They produce really good products. [00:00:57] But if they're not able to keep this up consistently and if they are not able to stand on a moral, neutral field, then they will be disrupted. [00:01:07] And Google's future... And then he talks about Singapore. [00:01:16] And who is he? [00:01:18] Hey everybody, Jason Bermas here. [00:01:20] We are live right now. [00:01:23] On this Sunday afternoon. [00:01:26] And we're going to be talking about Adrian Dittman. [00:01:28] Because that was the Adrian Dittman account. [00:01:32] And the reason that we're doing this video is because, number one, I am fascinated by this entire thing. [00:01:41] That's... Numero uno, because a lot of people aren't covering it. [00:01:46] But what you just heard was Dittman once again interacting with Alex Jones on Friday. [00:01:54] And I believe Alex Jones still thinks that this is Elon Musk's burner account. [00:02:01] Dittman comes in and talks about how the media has now covered the fact that he again interacted with Elon Musk, who was on a space held by a Variety reporter. [00:02:17] And that's really important. [00:02:18] We're going to break all this down because we're going to talk about who or what Adrian Dittman really is In my opinion, because this is speculation. [00:02:28] I want everybody to know I could be totally wrong about this. [00:02:32] I could be missing something. [00:02:34] I am a human being. [00:02:35] I'm not artificial intelligence. [00:02:38] But I got to tell you, the more and more I listen to Dittman and the more and more I reflect on the debate I was able to have with Dittman and Dittman's response to the Tesla-Curvec partnership and how Dittman is able to almost flip the script and interact very quickly. [00:03:01] And I think that's key for this to work. [00:03:04] Okay? That I do believe... [00:03:07] Bare minimum, you're dealing with a voice clone. [00:03:12] With somebody who's very on their feet. [00:03:15] With things Elon has said in public. [00:03:18] Alright? But really an AI. And we're going to get to why I believe that in a moment. [00:03:26] We're going to play some clips. We're actually going to play the Dittman-Musk interaction clip all together. [00:03:32] And Dittman posted this. [00:03:35] So one of the first things that I do want to do, and we're going to do this right here, is first and foremost, this is... [00:03:45] The space in which they're interacting. [00:03:47] This is the host right here, Katherine Brodsky. [00:03:51] Let's just bring it right in. [00:03:53] All right, and this is an actual Elon Musk speaker. [00:03:56] And I got in here because I noticed that Truthstream Media was in here. [00:04:01] And I was listening before I even saw that Dittman here was listening. [00:04:07] Now, Dittman actually gets the mic. [00:04:10] Catherine Brodsky says that she's interacted with Dittman. [00:04:15] You'll notice that Aaron Kariotius here. [00:04:18] Aaron Cariotti, very interesting fellow. [00:04:20] I've interviewed him on Making Sense of the Madness. [00:04:23] If you missed that, it's a great episode. [00:04:26] But he talks to Musk directly before Dittman gets the mic. [00:04:30] Now, there are thousands of people in this. [00:04:33] For them to get the mic is a huge deal. [00:04:36] Now, Dittman, when he goes on Jones on Friday, curses like a sailor, but still very much in line with this not Elon Musk persona. [00:04:47] Now, before this, and even after, I want you guys to notice that Dittman and the Musk account are interacting within minutes of each other also. [00:05:02] Maximize truth over ignorance. [00:05:04] You see this path to AI? It's like, you know, because Musk can meme all day. [00:05:08] Now, I think there's even an AI automation aspect or a bot aspect via surrogates who have access to the account to post things like that. [00:05:21] When the mainstream reports on this Dittman-Musk interaction, by the way, we're going to read some of it, The Verge, Daily Mail, they get it wrong that somebody got their account deactivated for harassing Dittman. [00:05:35] That person gave it up after the fact, deactivated it on its own, but said that they were threatened. [00:05:41] That's not even a drama I'm going to get into this. [00:05:44] But again, within minutes of these guys, even Musk is interacting with his account. [00:05:50] Within moments. Seven minutes. [00:05:52] Dittman. AI reflects the values of its creators. [00:05:56] Absolutely. So again, within moments, they're interacting with each other all the time. [00:06:03] It's kind of a big deal. [00:06:06] So we're going to get to there. [00:06:08] Now we're going to get to some interesting stuff. [00:06:11] First of all, the name Adrian Dittman, and this wasn't me, this was somebody else who's in the audience. [00:06:17] That's why I love the audience. Get your questions and comments in. [00:06:20] Maybe we're going to get to them at the end. [00:06:21] We're also going to talk about how it's not a free speech platform in real time. [00:06:25] I'm going to show you that in real time. [00:06:30] Dittman is an anagram, or could be an anagram for, didn't train a man. [00:06:39] Remember, we're going to play this interaction between Musk and Dittman. [00:06:44] This latest one. [00:06:45] From this space with thousands of people on it. [00:06:48] Where it's just a big joke fest and ha-ha fest. [00:06:52] And I don't believe you're not me. [00:06:54] A whole bunch of stuff. [00:06:56] Now, didn't train a man. [00:06:59] Let's start here. When you just Google this, check it out. [00:07:05] Surname Dittman meeting Orange to serve. [00:07:09] Okay? It's a German surname and it's to serve. [00:07:13] Alright? Now, on top of that, there also is this company. [00:07:19] Remember, this person is supposed to be in manufacturing. [00:07:22] Right? In the steel and metal working. [00:07:25] This is like a big manufacturing place. [00:07:28] Drilling, punching, Dittman tools. [00:07:32] Just throwing that out there. [00:07:34] Okay? All this is going to get thrown out there. [00:07:38] So, this Dittman character, more and more, to me, feels like an AI. Now, before we play that, I want to illustrate, again, in real time, how this is not a free speech platform. [00:07:52] Okay, so, people were pointing this out to me. [00:07:55] When I posted this, this right here, of course it had sound, of course it played. [00:08:00] Somebody said, hey, I couldn't get sound on it. [00:08:02] So, like, we're just going to take out my phone here. [00:08:05] We're going to go right here. [00:08:07] I've got another phone to do this on again. [00:08:08] I was like, that's weird, right? [00:08:10] Okay, so we have that. Let's take the screen down a bit so maybe it'll pick it up. [00:08:15] There we go. So you can see that right here we're going to do this. [00:08:18] And we're going to hit play on it. [00:08:19] I want everybody to do this and check it out because we're going to do it in real time. [00:08:24] Okay? No audio. [00:08:25] None. It's not a trick. [00:08:28] You know, my audio's jacked up. [00:08:30] I'm hitting the button. We'll do it on another phone. [00:08:32] I'm going to show you what happens on a desktop. [00:08:35] On multiple, including Microsoft Edge. [00:08:38] So, when I put this up, I'm sure a lot of you watch. [00:08:42] This is my pinned tweet. [00:08:44] Alright? And it happens to be critical of Elon Musk and transhumanism and Tesla. [00:08:50] Weird. It's just weird. [00:08:54] It's just so weird. It doesn't work. [00:08:56] Okay? So, right here, we're going to go to my profile. [00:09:01] Do it again. And let's see. [00:09:05] Nothing. Take it down so people can see I'm on the same thing. [00:09:09] But once again, nothing. [00:09:11] No audio. Why is that? [00:09:14] All right. Well, now let's do it on the desktop. [00:09:16] Let's do it. Because that makes sense. [00:09:19] Free speech platform, right? [00:09:21] Talk about a D-boost. Look. [00:09:23] So, wait. I hit play. [00:09:25] What's going on? Oh! [00:09:27] Can't do it, huh? [00:09:28] Let's just go to the tweet. [00:09:30] Let's get this URL here. [00:09:32] Let's try it again here. [00:09:34] I mean, that's so weird. [00:09:36] Okay. So now let's bring up Microsoft Edge. [00:09:40] The worst. The worst of the worst. [00:09:42] But it should work, right? Run all Microsoft product. [00:09:45] It knows I'm on a desktop. [00:09:47] Wait a minute. Huh? [00:09:50] Oh. No dice. [00:09:53] Okay. So now let's do this in real time. [00:09:55] Let's copy this and let's see if this works. [00:09:58] We're going to copy my tweet. [00:10:01] All right. Obviously, we're going to get rid of that one, right? [00:10:04] We're going to take the same exact file. [00:10:06] Let's go. Where's the post? Why can't I post? [00:10:09] Doesn't want to let me post? [00:10:10] I can't postyposterton? [00:10:12] Freezing up on me. There we go. [00:10:16] Geez, what's going on here? [00:10:18] Doesn't want me to post this whole thing? [00:10:20] Now, there's weird spacing for some reason when you do this via Twitter. [00:10:25] We're going to try to take out all the weird spacing. [00:10:27] Bear with me. Thumbs up and subscribe and share. [00:10:29] We're going to play the Musk, not Musk, Twitter interaction soon. [00:10:33] Don't worry. There we go. So now we're going to take the file. [00:10:36] We're going to pop it right in here. [00:10:38] Now, I just want people to understand, to prove this is the same exact file. [00:10:42] I'm not messing with it. Let's bring this over here. [00:10:46] It's science here. [00:10:47] If you notice, it says Sunday, August 27th, modified. [00:10:51] The original post, if you go check it out, on the page, guess what? [00:10:55] It's on the 28th. It's the same exact file. [00:10:59] 100% the same file. [00:11:00] Okay? So... [00:11:03] We're going to post this in a moment. [00:11:06] I'm going to let it upload. It's doing its thing. [00:11:08] It's already processing. [00:11:09] It's a pretty short file. You see it's at 60%. [00:11:11] We did it at 720p. [00:11:13] But as we do that, and this thing, we won't post it until after this, the big crescendo. [00:11:19] Let's listen to the Musk interaction with Adrian Dittman. [00:11:26] Let's do it, because we do it live. === It Gets Worse (04:35) === [00:11:30] Alex Jones thinks it's me. [00:11:31] Yeah, this is part two. [00:11:33] You've met me at least three times. [00:11:35] You're about to die of laughter, aren't you? [00:11:39] Yeah, it's me again. [00:11:41] This is part three. [00:11:45] Honestly, there's nothing better than to hear you laugh this much. [00:11:49] I can only imagine the stress you're under. [00:11:52] Dude, for the past couple of weeks, I basically... [00:11:56] Oh god, I can't do this. [00:11:58] laughter laughter [00:12:02] It's... [00:12:04] Oh my god. [00:12:06] laughter There's 8 billion people on earth. [00:12:11] So odds are there's someone who sounds like you somewhere. [00:12:15] And you are that person. [00:12:17] Yes. It gets worse. [00:12:19] It gets worse. [00:12:20] It gets so much worse. It gets so much worse. [00:12:24] Honestly, I don't believe you're not me. [00:12:27] Yeah, that's a theory. [00:12:30] People think I'm legitimately your clone. [00:12:33] There's been theories that I'm one of your Neuralink experiments. [00:12:36] Can you clarify that I'm not one of your Neuralink experiments? [00:12:39] Or maybe I am and I wouldn't realize it? [00:12:41] Because that's just vast. [00:12:44] How are your monkeys? How are your monkeys? [00:12:48] I don't know. How's your test subject? [00:12:52] Well, the weird thing is you even ask questions in the way that I would ask questions. [00:12:57] Autism. And your sense of humor is quite similar sense of humor. [00:13:03] It's eerie. Yeah, I know. [00:13:06] I still remember the first time... [00:13:08] Oh, God. [00:13:11] Oh, God. I still remember the first time... [00:13:15] I still remember the first time when I... The first time I found out about you was actually in early 2016. [00:13:22] This is when I watched SpaceX rocket engine experimentations on YouTube. [00:13:27] I was like, man, this is the coolest shit I've ever seen. [00:13:30] It's going to be next all day. And a significant other actually showed me a video clip of you and said, hey, Brad, this guy sounds like an older version of you. [00:13:36] I'm like, what are you fucking talking about? [00:13:37] Here, this guy. Look at him. [00:13:38] He's talking about hydrogen. What do you mean he's talking about hydrogen? [00:13:42] I listened to this and I was like, What the fuck is going on? [00:13:46] So this is the first time I realized that something's not quite right with this world. [00:13:50] Probably there's like a bug in the simulation. [00:13:53] And even if there was one, I'm pretty sure I caused a lot of disruption. [00:13:56] Recently, as I'm sure you're aware, there's a mass meltdown. [00:14:00] There's like an account that's doing rage farming. [00:14:04] It's so good. These people, what they've done is they're trying to pin on you that you've suspended an account because it did targeted harassment. [00:14:12] Well, it's not suspended. They actually deactivated their account. [00:14:15] Now they're trying to do rage farming. [00:14:16] Now they're going to spread it on the news everywhere and be like, dude, I'm complimenting yourself. [00:14:20] It's like the stupidest thing I've ever seen. [00:14:23] I was going to let it happen, but then I said, you know what? [00:14:26] I don't want thermonuclear warfare just yet. [00:14:34] For media. I'm not going to have that. [00:14:37] Ever since yesterday, even I do Diablo live streams, everything, it's all over the place. [00:14:40] It's inescapable now. [00:14:43] It's so weird. [00:14:44] It's terrible. I like a uniquely case. [00:14:49] I mean, are there pictures of you? [00:14:52] What do you look like? I've seen Adrian. [00:14:55] A thinner version of you. I've seen Adrian. [00:14:59] He could be your twin. [00:15:01] Are you serious? [00:15:04] I could DM you. [00:15:06] Dude, I could DM you. [00:15:08] If you enable access, I'll tell you everything you need to know. [00:15:12] Five minutes, I'll tell you everything. [00:15:14] If you're interested. [00:15:16] If you're interested. With you guys revealing each other, that would be pretty funny. [00:15:20] Oh god, yes. We should also get that guy from China. [00:15:26] Yeah, yeah, yeah. I've been thinking about this guy. [00:15:28] That would be the weirdest thing ever. [00:15:30] That's the holy trifecta. [00:15:31] You know the holy trinity? There we go, you found it. [00:15:36] Yeah, I don't know. [00:15:39] Wait, is that guy real? [00:15:42] That guy? I'm not sure. [00:15:43] Maybe he has had some facial modifications. [00:15:46] He looks strikingly real. === Medical AI Trolls (15:56) === [00:15:48] I would say. I mean, deepfakes maybe because the resolution is kind of shit. [00:15:53] But I don't know. [00:15:55] There are various many, say, Elvis impersonators out there, right? [00:15:59] They would modify their facial structure to look like him. [00:16:03] So I think maybe we're seeing a similar case. [00:16:06] I mean, they're really advanced down there in China, especially in the medical field. [00:16:10] So, like, maybe. I don't know. [00:16:11] It would be a really good troll. [00:16:13] Good for propaganda, too. [00:16:14] I would do it. Oh, boy, you guys. [00:16:17] Yeah. I mean, that would really trip people out. [00:16:22] Yeah, I mean, it's perfect, too. [00:16:24] Yeah, I mean, it's 1.4 billion people in China. [00:16:28] If you take the person in China that looks the most like me, it probably looks pretty close. [00:16:32] Yeah. Yeah. [00:16:35] So, I mean, honestly, I have to say just, like, to get this out of the gate, I really enjoy this platform. [00:16:42] I honestly don't understand why a lot of people complain about it. [00:16:44] I feel like it's about how you use it and what to use it for. [00:16:47] Like, I use basically every feature on the platform, except, of course, community notes and some of the newer features like articles and such. [00:16:55] I don't have access to that. I mean, I do spaces practically eight hours a day, three times a week. [00:17:03] Really? Yeah, all the time. [00:17:05] It's nice. I like to use it for information gathering and dissemination. [00:17:09] It's like, imagine your brain like a machine learning algorithm. [00:17:13] You just need a lot of data, and then you need to curate out of the noise the value. [00:17:17] You find the diamonds in the rough. [00:17:18] That's essentially what I do. [00:17:20] And I do this quite a lot. [00:17:21] It's amazing. We find cool stuff out. [00:17:23] We cover a lot of events. We covered your earnings call. [00:17:26] Your mother was in there. [00:17:28] That was so weird. Wait, wait. [00:17:31] You talked to me. What did my mom think? [00:17:33] I have no idea. [00:17:34] She was just listening in there. Just sitting. [00:17:36] And we were having a discussion on the earnings call. [00:17:38] She was just sitting in there. [00:17:40] She didn't hear me much because I didn't want to interrupt in the conversation. [00:17:43] And also I was scared shitless. [00:17:47] No, it's really gonna trip out my mom to hear you talk. [00:17:50] Oh yeah, oh yeah. [00:17:51] We'll have to clip it. It's good or anything. [00:17:56] Sure, why not? [00:17:57] I think she might be asleep. How's your kid doing? [00:18:00] I hear he's a little active in the background there. [00:18:07] You're doing fine? Yeah, I'll just text my mom so she's awake. [00:18:13] This would be the greatest face ever. [00:18:17] So it goes on a bit more. [00:18:19] Let's just review what happened there. [00:18:22] Then I'm going to play myself over with Alex Jones because more and more I really do think it's an AI. He talks about taking the diamonds in the rough. [00:18:35] All right, 2016 is where he got to know him. [00:18:38] So again, I think a lot of these interviews are from that time period on to make this persona. [00:18:44] Talks about taking in the good information. [00:18:47] All right, so now we're going to post this, by the way, guys. [00:18:49] Tell me if it has sound, you know? [00:18:51] And if it has sound, we're obviously going to pin this up there. [00:18:54] But it's very odd that the one above it, let's pin it right now, Let's do that. [00:19:00] The one above it doesn't have sound. [00:19:03] We just did that. [00:19:05] So let's pin her. Pin it to the top. [00:19:07] Pin to my profile. [00:19:09] Boom. Pin it. [00:19:11] Boom. All right. [00:19:12] All right. We did it. [00:19:14] Now, Again, I can't reiterate enough that if this is in fact an AI and eventually a consumer product, it's doing a great job right now of selling itself because it's past the Turing test. [00:19:29] You know, I keep reiterating that, but when you cannot tell an AI from a human being and you've got people convinced, that's the number one sale of your AI product. [00:19:40] And just the way that Musk is reacting to this, I mean, 99.9% he's in on the joke. [00:19:48] It's part of the product, right? [00:19:49] It has to be. I mean, talking about being an AI, not an AI clone, but a Neuralink experiment, the thing asks about the monkey several times. [00:20:00] And again, if... It's that advanced. [00:20:03] There's no telling if it's really Musk on the end of these spaces either. [00:20:08] And soon with the video, as we've shown you, you know, maybe I should bring that up again, that avatar thing. [00:20:13] Because it is really important for people to understand how far this stuff has really gotten. [00:20:21] Okay, so let's just pop this right up here right now. [00:20:26] Let's go to this right here and shebango. [00:20:31] You can finally turn yourself into an AI avatar for free. [00:20:34] In the process, it only takes five minutes. [00:20:36] And all you need is your phone. [00:20:38] Wow, that was not so bad. [00:20:39] Again, that was an AI avatar clone. [00:20:43] He takes you through how to do it. [00:20:45] You can finally turn yourself into an AI avatar for free. [00:20:48] In the process, it only takes five minutes. [00:20:51] And all you need is your phone. [00:20:52] And it's true. [00:20:55] And he shows you the website. [00:20:57] It's called HeyGen. [00:20:58] H-E-Y-G-N. We're in the post-truth world. [00:21:03] Post-truth era. [00:21:05] We're here. It's not coming. [00:21:08] So this is me talking to Alex Jones. [00:21:11] Now remember, Dittman, after that interaction with Musk, and I want to show this article here too. [00:21:18] Definitely do that. [00:21:19] And the media picks up on it. [00:21:22] Right here, ex-users claim Elon Musk has created a burner account. [00:21:25] He actually discusses the interaction that got them to retract this part when he was talking about the person in there that supposedly got their account banned but really didn't. [00:21:38] See how this works? [00:21:40] All right? I mean, it's all right here. [00:21:43] But then Dittman didn't interact with him afterwards. [00:21:47] Like, the DMs of Dittman are open. [00:21:50] They're right there. So again, just totally and completely bizarro world unless this thing is in fact some sort of an AI product or a beta for an AI product. [00:22:06] It's the ultimate digital Andy Kaufman troll right now. [00:22:10] That's how I feel about it. [00:22:13] I could be completely wrong. [00:22:14] But here we go. [00:22:16] Here is me talking to Jones about what I think this is before I actually debate Dittman for 20 minutes. [00:22:27] Oh, nice. Jason, what is going on, man? [00:22:29] Go ahead and unmute yourself if you're listening and say what you have to say. [00:22:35] Jason Barber, you're on the air. [00:22:36] No, here we go. Oh, there we go. [00:22:38] Yeah, you muted me, dude. [00:22:40] Sorry, I got you unmuted here. [00:22:43] Okay, so Jake was muted, but I'm on now. [00:22:46] It's Burmus, just to let everybody know. [00:22:48] Okay, Alex, first of all, G. Edward Griffin, living legend. [00:22:52] That's why he was a part of the documentary film that you produced, Invisible Empire. [00:22:57] You know, beyond just cancer, globalism, so much more. [00:23:01] G. Edward Griffin is the man. [00:23:02] I was actually at that Red Pill conference out in Des Moines. [00:23:06] Still a workhorse and more relevant today than ever. [00:23:09] The reason I was on the space is this Dittmer situation. [00:23:13] You know, Alex, I listened to your debate with David Icke on Friday, and I was in and out, and I was on the road. [00:23:20] And when I tuned back in and, quote-unquote, Dittmer was on, obviously I was convinced that that was Elon Musk. [00:23:26] You know, after I started doing some research, you know, he held like a six-hour space after that. [00:23:32] Of what or who this is, I'm just blown away because to me there's only two possibilities. [00:23:38] Dittmer is either, like you said, an advanced AI or a surrogate for Musk that is kind of playing a part. [00:23:47] And I think that also there's this interchangeable quality where he could also act like a dummy account for Musk himself. [00:23:55] In other words, Musk has access to Dittmer, obviously, and it could be Musk at any moment. [00:24:00] Jason, let me Well, Alex, are you aware that they actually did a space together in July where they had like a 20-plus minute conversation? [00:24:12] And when I listen to this back and forth... [00:24:14] Yeah, but he's able to do that because he goes back and forth and does it. [00:24:17] It's all a big troll. [00:24:20] Well, I would love to have an opportunity to debate Dittmer on many subjects. [00:24:27] Alex, you're killing it. The show is more interesting... [00:24:31] So, I keep calling him Dittmer on that one. [00:24:33] You guys must love that, as you should. [00:24:36] But, again, the more and more I look at it, AI to the maximum. [00:24:40] Let's see if we got any questions or comments here from the crew. [00:24:44] By the way, thumbs it up, subscribe, share. [00:24:47] Couldn't do it without you guys. You know, the links are down below. [00:24:50] I want you to share this information with others. [00:24:53] My show moves to the morning. [00:24:54] I actually just did a great show with Derek Brose for Making Sense of the Madness. [00:24:59] You go check that out over on Rumble. [00:25:01] But if you're following me on X, there's links to it down below. [00:25:04] That's where you're really going to find all of my work. [00:25:06] So I moved to the mornings. [00:25:08] Tomorrow, Joseph Lee, Dr. [00:25:11] Joseph Lee, is who I did a really great interview that's going to air tomorrow morning, making sense of the madness. [00:25:17] All right, let's do this thing. [00:25:18] Let's come on over here. [00:25:20] And hey there, Jason, how you doing? [00:25:22] All right. It's Elon with a voice clone of himself. [00:25:25] Why do spaces not have chat? [00:25:27] Yeah, well, I mean, the live streams apparently have chat. [00:25:31] I am live streaming, hopefully, on X right now as well. [00:25:35] The boop, perhaps you ought to be cautious too, Jason. [00:25:38] What am I going to be cautious with? I don't know what you're talking about. [00:25:42] I was thinking about codes in the name. [00:25:44] Didn't train Ant-Man. [00:25:48] No, did train Ant-Man. [00:25:50] Yeah, no, it's didn't train a man. [00:25:52] Okay, I'm just saying. [00:25:54] I wasn't able to crack the anagram. [00:25:57] I didn't crack it either. [00:25:58] Again, it's the audience out there. [00:26:00] No audio on your Twitter vid? [00:26:03] Oh yeah, there you go. [00:26:04] That's it. Well, maybe there's audio now. [00:26:06] People go check it. Let me know. [00:26:09] Let's see. Solar storm. [00:26:10] The power might go completely out this cycle. [00:26:13] Did it just break? [00:26:14] I cannot share much stuff too. [00:26:16] Has anyone seen Musk and Dittman in the same room? [00:26:20] Let's see. There's a lot of what's going on. [00:26:22] It's so real it must be fake. [00:26:24] Too funny. It's not just autism. [00:26:25] Sorry. He answers it. [00:26:28] The whole account is a promotion of Musk and Tesla and AI. All of it. [00:26:33] SpaceX. Let's see. [00:26:37] Like I said, the one we listened to was earlier this week. [00:26:39] The one we introed was after. [00:26:41] We did it kind of all out of order. [00:26:42] We did it Tarantino style. [00:26:44] Okay, the first thing that actually happened in all that was me talking to Alex Jones. [00:26:51] And the last thing that happened was Dittman again talking to Alex Jones. [00:26:56] And in between, he had that must-or-do space. [00:27:00] See, we try to keep it interesting and confusing as possible here when we're dealing with such a confusing topic. [00:27:08] So far down this rabbit hole, it seems old to me. [00:27:11] It's only been out a couple weeks. [00:27:12] I didn't know anything about it. You know, I had a discussion with Zach Voorhees, the Google whistleblower, on Making Sense of the Madness earlier this week about Dittman at length. [00:27:22] I'd love for you guys to check that out as well. [00:27:24] That's all over at patriot.tv. [00:27:26] Or the Patriot.tv Rumble has a lot of it. [00:27:31] Let's see. Elon never lasts this morning. [00:27:33] He's interacting with his own AI. Why isn't Elon saying anything even slightly intelligent? [00:27:38] This reeks of fake. Jason, have you seen that WWF guy accused of trafficking mask with Mexican mustache? [00:27:48] What? No, I guess not. [00:27:52] What? Hi, the next PSYOP is going to be time travel. [00:27:55] Wait for it. You'll see. [00:27:57] We're in the post-truth world. [00:27:59] They could pull anything. [00:28:01] His biological parents, not perfect enough. [00:28:03] Home alone. Talkboy technology. [00:28:07] Let's see. Yes, let's talk about Musk's mother's ties. [00:28:10] Geez, yeah. Let's see. [00:28:12] Sounds good. The media could not be played. [00:28:15] So the new one. The new one. [00:28:18] Oh wait, so the latest post does have audio. [00:28:20] Look at that. It's the same exact file. [00:28:24] And that other one had audio for a while too. [00:28:27] But it's a free speech platform. [00:28:29] I'm sure it's just a bug. [00:28:31] I'm sure it's just a buggy bug. [00:28:34] Come on. Give me a break. [00:28:37] Give me a break. [00:28:38] Most people are robotic, Jay. [00:28:40] Not like that. Dittman's quick. [00:28:43] Let's see. Anybody done frequency spectrum analysis on the Dittman audio? [00:28:47] I've not. But again, it could be an advanced voice mod with somebody really good. [00:28:51] But I don't know. But I don't know. [00:28:55] Let's see. The twinkling eyes jump out. [00:28:57] Whoa. Let's see. [00:29:00] Let's see. My man, Spyro. [00:29:02] How we doing, bro? He does great work. [00:29:04] I'm sorry. Again, go back to that whole... [00:29:09] You know what? We're going to play it right now. [00:29:11] Think about how he answers the Tesla CureVac question. [00:29:15] How quick he is and how quick to move the subject. [00:29:17] He tries to make it about... [00:29:19] When I say he, I mean Dittman. [00:29:21] It. Tries to make it about... [00:29:23] How I want to talk about the simulation. [00:29:25] The simulation was the last thing I wanted to talk about. [00:29:28] He even diverted my real intentions. [00:29:30] But at the same time, I don't want people to think that I buy into simulation theory because I don't. [00:29:36] Because I don't. Okay, so let's see. [00:29:40] Let's just give me one second. [00:29:41] Thumbs this up and subscribe and share. [00:29:44] Alright, everybody? That's what we want to do. [00:29:47] Whoa, that's big. [00:29:49] Without you guys, I couldn't do it. [00:29:51] Doing a little editing here because I want this done right. [00:29:54] And I can't stand it when things don't get done right. [00:29:57] You know what I'm saying? It's tough out there. [00:29:59] We want a pro show. [00:30:01] Okay, so we're going to add this source. [00:30:03] Boom. That's what we want to do. [00:30:07] And there it is. Perfect. [00:30:10] And yes. Okay. [00:30:12] So let's do it. Let's bring this in. [00:30:14] This is... [00:30:15] We have a few new faces here. [00:30:20] Jason, what's up? Yeah, what's up, Jason? [00:30:23] No, that's me talking. [00:30:24] I'm muted here, but don't worry, I'll get in there. [00:30:26] Think about how defensive Dick is. [00:30:28] Adrian, I've got a question for you. [00:30:30] You know, obviously you've been on the Alex Jones show recently. [00:30:36] I wonder what you think about Tesla's partnership with CureVac during the pandemic to print up the mRNA shots via the microfactories. [00:30:50] I am, like, not going to discuss that. [00:30:55] Is there any reason why you would not discuss that? [00:30:57] We're not doing medical stuff today. [00:30:59] We're not doing medical stuff today. [00:31:00] Anything else? I was just wondering, during that time period, it was reported that you made a 600% profit during the COVID... And again, I... I don't want to get into a whole bunch of stuff. [00:31:14] Obviously, I started referring to him as Musk. [00:31:16] That's where we get into a debate on why he's not Musk. [00:31:19] Very defensive. But it's hard when you're talking to this thing. [00:31:23] Think about... I think I stumped it for a second there. [00:31:27] It's not about medical stuff. === Eight Hour Space Dive (01:43) === [00:31:29] None of that was about medical stuff. [00:31:30] Just want to say that. Think about it. [00:31:32] It's about a partnership business one. [00:31:35] Let's keep going downtown here. [00:31:39] Let's see... Next level, AJ, but will your avatar... [00:31:44] Who cares? Oh, you're talking about the Gemini thing. [00:31:47] It's so out of control. Get sent into space through SpaceX. [00:31:53] I could use a double. [00:31:55] Hey, caught a live show. Awesome. [00:31:56] Well, let's see. Jason, did you see the eight-hour space where he discussed generative AI and copyright? [00:32:02] I think Alex is correct. [00:32:05] Well, what do you mean? [00:32:07] Alex is correct about what? [00:32:08] Whether he is or not an AI? I didn't watch. [00:32:12] First of all, some of those spaces are recorded. [00:32:14] Some of them, they're not. I do have a life. [00:32:15] I can't just listen to eight-hour spaces all the time. [00:32:18] I'm running kids around all the time. [00:32:20] Let's see. Yeah, only 46 thumbs. [00:32:22] 67. Can we get 100 thumbs on this one? [00:32:24] This is just being so under, like, covered in my opinion. [00:32:29] Let's see. Ditmer. [00:32:34] Again, I'm a human being. [00:32:36] I'm going to get things wrong. I'm not an AI. I'm certainly not an AI based on Elon Musk. [00:32:41] Let's see. I must have spent a lot less time in front of a screen than a lot of people. [00:32:49] Hey, Miriam, how we doing? [00:32:50] You did stump it. [00:32:53] I don't know. Maybe I did. [00:32:54] I don't know. Let's see. [00:32:56] Just add my three cents that many of us have been speculating about Dittman for the last eight months. === Pennywise and Digital Doppelgangers (01:33) === [00:33:01] Everyone is going to want their own digital doppelganger. [00:33:04] Yeah, a lot of people were in it. [00:33:05] Johnny Vedmore got at me. [00:33:07] Again, I just talked to Zach Voorhees about it. [00:33:11] He'd been talking about it. [00:33:13] I had no idea until the Muskernuts, AI, Jones versus... [00:33:22] Jones vs. Ike debate came up. [00:33:25] So, guys, that's going to wrap it for me. [00:33:27] I'm going to tell you what. I'm an excited little guy. [00:33:29] I'm going to check out Pennywise opening up for the Dropkick Murphys. [00:33:34] I've seen the Dropkick Murphys several times. [00:33:36] There's not much more I really enjoy in life than a Kick Arnis concert. [00:33:42] If you've never seen the Dropkick Murphys live, I don't know what to tell you, but you are missing out. [00:33:47] I'm expecting a great show. [00:33:49] This is their St. [00:33:51] Patrick's Day tour. St. [00:33:52] Patty's Day. Just a few weeks away. [00:33:55] So I'm pumped. I'm pumped on life. [00:33:58] I love living it. I don't think it's a simulation. [00:34:00] I think that the Murphys are about to kill it in base reality. [00:34:04] Everybody, that's the real deal. [00:34:05] And I want you guys to love life. [00:34:07] I want you to get out there. [00:34:09] Enjoy yourselves on top of all the madness. [00:34:11] That is going on. [00:34:12] If you can support the broadcast, you know how to. [00:34:14] The links are down below. [00:34:16] But thumbs it up. Subscribe. Share. [00:34:17] Help me crack it. [00:34:19] Didn't train a man? [00:34:22] Dittman tools? [00:34:24] Dittman means to serve? [00:34:26] I don't know. I don't know, everybody. [00:34:29] We'll talk to you soon.