The Charlie Kirk Show - Indiana Jones and the Temple of Woke with Libby Emmons and ALX Aired: 2022-12-04 Duration: 33:16 === Losing Our Human Soul (12:20) === [00:00:00] Jack Rozobic filling in for Charlie Kirk today. [00:00:03] Incredible set of interviews for you. [00:00:05] First, Libby Emmons breaking down transhumanism that we have lost the embrace of our soul, our almighty soul. [00:00:15] You have to listen to this interview. [00:00:16] And then ALX freshly escaped from the Twitter gulag, breaking down everything that's happening with Elon Musk. [00:00:24] How should we move forward with freedom of speech on Twitter? [00:00:28] Buckle up. [00:00:29] And here we go. [00:00:31] Charlie, what you've done is incredible here. [00:00:32] Maybe Charlie Kirk is on the college campus. [00:00:34] I want you to know we are lucky to have Charlie Kirk. [00:00:38] Charlie Kirk's running the White House, folks. [00:00:41] I want to thank Charlie. [00:00:42] He's an incredible guy. [00:00:43] His spirit, his love of this country, he's done an amazing job building one of the most powerful youth organizations ever created. [00:00:50] Turning point USA. [00:00:52] We will not embrace the ideas that have destroyed countries, destroyed lives, and we are going to fight for freedom on campuses across the country. [00:01:00] That's why we are here. [00:01:03] The new Indiana Jones movie. [00:01:05] It's coming out, Indiana Jones 5. [00:01:08] And what have they done yet again? [00:01:10] What have they done yet again? [00:01:12] They are working to introduce a brand new character for the series, for the franchise, because Harrison Ford is getting a little old, 80 years old, by the way. [00:01:23] Apparently he was hurt during a lot of the filming. [00:01:25] And Kathleen Kennedy, the same mastermind that gave you the incredible Disney Star Wars. [00:01:31] Hey, kids, if you like Disney Star Wars, you'll love, you'll absolutely love Disney Indiana Jones, because what we're going to do is we're going to replace Indiana with a sassy and crazy female, female goddaughter who she's going to get the hat. [00:01:54] She's going to get the whip at the very end of it all. [00:01:57] And so when I look at stuff like this, it's amazing to me because it's like, this is exactly what you did with Luke Skywalker. [00:02:06] You essentially replaced him with this girl that no one's ever heard of who popped up and suddenly, oh, we all have to love her. [00:02:13] We all have to do. [00:02:14] You're creatively bankrupt. [00:02:16] And Kathleen Kennedy is the same person who's involved in all of this. [00:02:20] You find some institution or franchise that everybody loves. [00:02:24] You come in and you say, by the way, it's not enough. [00:02:28] It's not enough that we, you know, you're going to say like, oh, well, Americans, you know, these, you know, the masculine patriarchy would never allow, never, ever allow for a female explorer, a strong female explorer character. [00:02:43] Oh, wait, you know, we have the entire Laura Croft franchise. [00:02:45] No, it's not enough. [00:02:46] Why? [00:02:47] Because Indiana Jones must be, literally be a woman. [00:02:53] This is what you must accept, America, or you're a bigot. [00:02:57] You're a misanthrope. [00:02:59] You're a misogynist. [00:03:01] If you don't accept that Indiana Jones is now a woman, if they can't imagine good stories, they're basically making fanfics. [00:03:11] It's basically fanfiction, right? [00:03:13] It's, well, if I were, if I were Indiana Jones, then I would do things. [00:03:19] No, Stop. [00:03:20] Indiana Jones is Indiana Jones. [00:03:22] Okay. [00:03:22] You have to stop. [00:03:23] I was talking to my kids last night, by the way, and it just four years old. [00:03:27] And baby AJ, his birthday is tomorrow, by the way, so he'll be two. [00:03:31] But I was talking to Jack Jr. and he's, he's four. [00:03:33] I said, and I just kind of ran it by him. [00:03:35] And I don't get very political with my kids, but I said, I said, Jack, Jack, can boys become girls? [00:03:42] He goes, no. [00:03:43] I said, can girls become boys? [00:03:45] And he goes, no. [00:03:47] And then I said, well, why not? [00:03:48] And he goes, because that's what they are, dad. [00:03:51] Yeah, exactly. [00:03:52] Exactly. [00:03:53] So speaking of, of all things, not just transgender, but also transhuman, we do have finally the great Libby Emmons, the editor-in-chief of the post-millennial, sort of a sort of a co-host for me now on Human Events Daily, which is pretty cool. [00:04:11] Libby Emmons, thank you so much for joining the program. [00:04:13] Thanks so much, Jack. [00:04:15] Glad to be on the show. [00:04:16] So by the way, I got your take. [00:04:18] What do you think about this whole Indiana Jones, the female replacement is coming in? [00:04:23] What's going on? [00:04:24] I think it's actually insulting to women that writers and showrunners and whoever else think that the best way to create a woman character is to take a male character and just give her a female body. [00:04:37] That's not how we create organic women characters. [00:04:40] It's just a completely useless way to do it. [00:04:44] The way you write characters is you listen to your characters, you build your characters from the ground up, you dig into their own personal experiences. [00:04:51] And I say this as someone who came out of the world of theater and has written dozens of female lead characters. [00:04:59] That's how you do it. [00:05:00] You don't take a whole bunch of male energy and just put it on a female body. [00:05:04] It's fake. [00:05:04] Everyone knows it's fake. [00:05:06] And these movies fail over and over again. [00:05:08] By the way, even beyond that, the CGI is trash. [00:05:14] I wasn't a fan, by the way, of Avatar because I thought it was a stupid movie, but the CGI, it looks disgusting. [00:05:20] He's riding a horse through the subway and it just looks fake. [00:05:24] The whole thing looks fake and plastic and dumb. [00:05:27] It's like, this is really what, and it's just sad. [00:05:30] It's just very sad to, you know, Harrison, to the character, to the fans. [00:05:36] And honestly, even for my kids, right? [00:05:38] You know, and that being said, you know, I suppose it shouldn't matter because thanks to Spermageddon, we're not having, you know, men or men are, we're losing our foothold. [00:05:47] What can I say in terms of all? [00:05:49] But I wanted to talk to you about this latest. [00:05:52] And you and I did a riff on human events yesterday on this as well. [00:05:56] But we have to get into it again here on the program because Elon Musk, a guy who, you know, and we've been critical of him today. [00:06:05] I've said basically, I don't think that anybody should be permanently banned from Twitter if they haven't broken the law. [00:06:14] I don't believe that. [00:06:14] If you want to talk about tweets, if you want to talk about specific things that should be banned, sure, fine. [00:06:18] Let's talk about that. [00:06:21] But Libby, you had a piece years ago talking about the transhumanism revolution over in Quillette that got you in a little bit of trouble with that theater group that you were just talking about. [00:06:33] Walk us through that. [00:06:34] It did. [00:06:35] It did. [00:06:36] So I had been doing a lot of research into transhumanism and futurism. [00:06:40] I had been very fascinated with that. [00:06:42] I had been doing research into that for about a year. [00:06:45] And I finally realized that I wanted to write about this. [00:06:49] I had been inspired in part by Dr. Jordan Peterson's appearances on a few of these podcasts. [00:06:54] And I realized that there were undercurrents of transhumanism in Western culture coming down the pike. [00:06:59] And we weren't aware of it at all. [00:07:02] What I mean by transhumanism is the intentional evolution of human beings with the help of technology towards longevity and towards a transhuman humanity, as you would say. [00:07:18] So that's what transhumanism essentially is. [00:07:21] These three undercurrents that I was looking at were AI-human interface, as in Neuralink, as well as body hacking, where you could put things like RFID chips in your hands to open your garage door opener, things like that. [00:07:36] And also transgender ideology. [00:07:38] Of course, it was transgender ideology coming from an indie theater background that got me into trouble with all of my non-binary feminist lesbian theater friends who were saying to me that I had essentially posited that transgender people were weird futuristic robot people and then I needed to apologize to everyone. [00:08:00] I hadn't said that at all in Quillette, but there it is. [00:08:04] But actually, that being said, and I'm so sorry that happened, but at the same time, it's sort of, you know, we all have this sort of story about how we got jumped in right to the gang. [00:08:12] Right. [00:08:14] But walk me through that because that's interesting. [00:08:17] What would you say is the ideological connection there? [00:08:21] What were you discussing? [00:08:21] What was your thesis? [00:08:23] I was saying that all of these things essentially answer the Cartesian question of the mind-body split and say that objectively, mind and body are two separate things, that what you do to the body has nothing to do with the mind and the reverse being true as well, I would say, for these technologies. [00:08:42] I thoroughly disagree with that. [00:08:44] I believe in human beings. [00:08:46] I believe we were created in God's image. [00:08:48] I believe that the body matters. [00:08:50] I believe that what you do to your body has an impact on your mind and what you do to your mind has an impact on your body. [00:08:56] We see this all the time. [00:08:58] We keep getting smacked in the face with it over and over. [00:09:01] Yet we continue to say, you know, that we can do anything to our bodies and it doesn't matter. [00:09:06] There are so many elements of this in our culture that we can look at. [00:09:11] And we also have essentially abandoned the concept of an everlasting soul, which I think is really a shame. [00:09:18] And you have a lot of kids today. [00:09:20] And when they feel meaninglessness or, you know, a spiritual gap, they attribute it to a psychological condition. [00:09:28] And they go off in search of drugs and medical cures when what they really need is a spiritual solution. [00:09:35] Couldn't agree more. [00:09:36] Libby, we're going to stick with you because we have to talk more about how this effective altruism of SBF and humanism of Elon Musk is leading us down this path of trying to put chips in our brains, removing a small part of your skull, a quarter-sized hole, and replacing it with a Fitbit. [00:09:56] That's what Elon Musk said. [00:09:58] It's not me. [00:09:58] It's not Libby Emmons. [00:09:59] It's not Charlie Kirk. [00:10:00] That's Elon Musk. [00:10:02] A Fitbit in your skull. [00:10:04] By the way, I mentioned that's a tiny date last night. [00:10:06] I was supposed to say it did not go over well. [00:10:11] And Libby, let's get into this now. [00:10:14] Elon Musk telling us that, and it's similar to what you said before, that there's that we're just not enough on our own. [00:10:22] You know, we're basically cyborgs, his words, we're basically cyborgs already because of our cell phones and our, you know, our smart devices, the internet of things. [00:10:32] So this is really just all it's doing is just taking another step. [00:10:36] And, you know, it's just your garage door opener, you know, that's going to be implanted in your hand now. [00:10:42] It's just a Fitbit, but it's going to be in your skull. [00:10:44] It's just a new interface for the internet, but it's going to be inside your eyelids, you know, et cetera, et cetera. [00:10:49] And so inside your, you know, maybe your optic nerve. [00:10:54] What are we losing by going down this path? [00:10:57] What is happening? [00:10:57] What has really happened? [00:10:59] What's really happening is we are losing what it means to be human and we are also losing in many ways the ability to be independent and autonomous beings. [00:11:11] Part of what Elon Musk has talked about with Neuralink and many others in the technology space have talked about it. [00:11:18] It's not just Elon Musk. [00:11:19] There's other companies who are undertaking this as well. [00:11:23] Of course, Musk is really quite an entrepreneur. [00:11:26] So he's at the forefront of this, as with so many things. [00:11:29] But part of the goal is not just to, You know, have your own data port in your brain that can be fed from the data stream that is out there, but to connect human beings to one another via these data ports, via this stream, so that you would have each individual literally in a hive mind. [00:11:53] And we would not necessarily know where our thoughts stop, where the data stream begins, where the thoughts of other human beings are. [00:12:01] When you think about how easy it is to, when you're not paying attention, to take in advertising, right? [00:12:09] Anytime you watch a little kid watching TV in the lead up to Christmas, and the next thing you know, he's watching ads for Legos. [00:12:15] And then the next thing you know, Mom, I want this Lego set. [00:12:18] I want that. [00:12:19] Oh, did you just see it on TV? === The Toy Zoo Trap (02:55) === [00:12:20] Yeah. [00:12:21] Oh, well, you should have heard what Jack Jack said. [00:12:23] You know, actually, funny enough, because he's been super into Legos lately, but then he goes up to Santa last night and, you know, what would you like, you know, little boy? [00:12:32] He goes, Santa, I want a zoo. [00:12:35] I saw you said that on Twitter. [00:12:38] And then he said, oh, you mean like a toy zoo? [00:12:40] And he goes, no, I mean a zoo. [00:12:41] I mean, I want a zoo for Christmas. [00:12:44] Literally. [00:12:45] And I said, I came back to me. [00:12:47] I said, Son, you've never mentioned anything about this at all. [00:12:50] He goes, so I want a zoo. [00:12:51] So now, so now we have to get him. [00:12:53] You have to get him a zoo now, obviously. [00:12:56] Yeah, but if he were connected to the hive mind, then a zoo could just magically come to your door. [00:13:03] Perhaps Amazon will have it set up so that whatever you think just magically appears. [00:13:07] But yeah, so what we're losing is our humanity. [00:13:09] We are losing our soul in many ways. [00:13:12] We are losing our individuality. [00:13:14] And the potential to lose things is actually even greater. [00:13:18] If you think about authoritarian countries and really oppressive societies, what does it mean for a dissident to have thoughts about how to stage a protest when they are connected to a hive mind? [00:13:30] And those thoughts, the data is going both ways. [00:13:34] What does that mean? [00:13:35] How easy will it be for governments to literally become thought police and to use your thoughts against you? [00:13:41] What about also in a spiritual sense? [00:13:43] Because the idea behind the concept of free will from a theological perspective is that you come to God of your own volition. [00:13:57] God could have made automatons and robots and mental slaves and all of this that you will love me. [00:14:03] But God didn't do that. [00:14:04] God said, I want to create humans with free will, sentience, and a conscience, but not force their directions. [00:14:17] Yeah. [00:14:19] I, you know, that is the thing that keeps bringing me back to theology as an adult, looking at that and realizing that God wanted God wanted people to really love him, to not be told to love him, but to actually come to them, come to him of their own free will and to seek out God's guidance on their own, which I think is just such a beautiful concept for mankind and for Christianity. [00:14:46] I think we lose a sense of free will 100% when we start talking about being interconnected with data streams. [00:14:54] And we don't know who the puppet masters are who are controlling that. [00:14:57] Certainly it's not God. [00:14:59] Certainly it has nothing to do with our souls or with meaning. [00:15:03] And it deprives us of our ability to think for ourselves, to think critically, to form our own perspectives, and to interact organically with the world around us. [00:15:14] But Libby Emmons, go follow her. === Chaos in Terms of Service (15:24) === [00:15:16] She's been co-hosting, will be co-hosting with me again on Human Events Daily tonight. [00:15:20] Make sure you read this piece, humanevents.com, and the post-millennial editor-in-chief just fantastic, fantastic work. [00:15:25] Thank you very much, Libby, and God bless you for all of this. [00:15:29] And we need to understand we're doing away with our spirituality and committing blasphemy against the Holy Spirit. [00:15:38] A Twitter POW who's released from the gulags, someone who is a veteran of the great meme war of 2015 and 16 himself, ladies and gentlemen, the great ALX joins us, graces us with his presence. [00:15:56] ALX, thank you so much for joining us here at Charlie Kirk Show. [00:16:00] Thanks for having me, Jack. [00:16:02] So I got to get into it. [00:16:03] We got to get right into it. [00:16:05] You had a tweet thread and an op-ed getting into this. [00:16:09] And we've been talking about it all day here, back and forth, about the permanent banning, Elon Musk's permanent ban of Kanye West. [00:16:16] It came in last night. [00:16:19] You have a lot to say about this. [00:16:22] You had a lot to say about it at the time. [00:16:23] We've been asking people to email us, freedom at charliekirk.com about it. [00:16:28] But walk us through your tweet thread that you sent to Elon just prior to the ban and this system of a quote timeout period for repeat multiple violations that I think is a very interesting issue. [00:16:47] Very interesting system. [00:16:48] Yeah. [00:16:49] So ideally, I wouldn't like to see any suspensions or any bans for lawful speech. [00:16:55] But as a middle ground, because I know there's a lot of pressure from advertisers and just in general online for people who like intentionally violate the terms of service and abuse the fact that Elon wants to make it a free speech platform. [00:17:10] I think until the terms of service can be properly modified, I think the best course of action is to institute what I proposed of being a temporary suspension period. [00:17:23] Twitter already has sort of this, but it's kind of arbitrary. [00:17:28] They have like 12 hour suspensions, week suspensions, three day suspensions. [00:17:32] But what I outlined is sort of the first strike is kind of like a warning. [00:17:37] You get a 12 hour lockout. [00:17:39] And then if you come back from that and then intentionally violate the terms of service like right away, then it would increase each time. [00:17:48] And those would stay on your account for 90 days up until like the most pre like the previous violation. [00:17:55] And I think that's a way to get rid of permanent bans altogether for lawful speech. [00:18:00] Because hypothetically, if somebody was a bad actor and continued to violate the terms of service, they would basically keep offending. [00:18:10] And then it would be a longer period of time each time if they like hadn't had good behavior for 90 days. [00:18:19] And again, I know this isn't like an ideal solution. [00:18:23] And I believe the terms of service need to be clarified more because some people are saying that's a religious symbol that he tweeted. [00:18:29] I know it's open to interpretation. [00:18:32] And due to his comments yesterday as well, that's why everyone's interpreting it the other way. [00:18:37] But it needs to be more, the terms of service need to be more specific in terms of what violates their policies. [00:18:45] Cause I don't think that was an insight into violence. [00:18:48] Well, and so, you know, and I just had Derry Beattie on ahead of you a couple of segments ago talking about this. [00:18:54] And it's this sort of idea that I think people have gotten into this theory that when it comes to, so let's take it the two highest profile. [00:19:04] Well, the one is that he refused to unban Alex Jones, right? [00:19:08] But then unbans Donald Trump and you on the same day. [00:19:14] But also has said that he's banning Kanye West. [00:19:18] And he posted that tweet last night, you know, mess around and find out. [00:19:22] And so the question that I had for Darren, and we were sort of teasing this out, was does this because if you remember, his reasoning for the Alex Jones ban was a personal reason, right? [00:19:33] He brought up a personal reason. [00:19:35] And we've seen, now he didn't state this with respect to Kanye. [00:19:40] And by the way, I'm going through the emails again here. [00:19:43] And I've yet to actually see one person say they agree with a permanent ban of Kanye West. [00:19:48] I'm actually surprised. [00:19:48] I thought we get at least one. [00:19:50] And maybe I'm asking if they can go back and look. [00:19:52] But they said, you know, the last one we just have here, it says, thank you for guest hosting, Jack. [00:19:58] You're blah, blah, blah. [00:19:59] You're a good guy. [00:20:00] No, I don't think Ye should be banned at all. [00:20:02] Nothing he said was a direct threat of violence or libeling anyone. [00:20:06] Yes, his remarks were disgraceful. [00:20:08] Okay. [00:20:09] And this is from Christopher, but not illegal. [00:20:13] And I haven't gotten a single email yet. [00:20:15] We haven't gotten an email yet today saying that they agree with, again, a permanent ban. [00:20:20] We're not talking about the tweet itself, but to your point as well, when we're talking about terms of service, and I also asked, you know, asked Aaron, the timing, I think, matters as well, because Elon has just taken over Twitter. [00:20:32] And so there isn't really a system in place yet for something like this. [00:20:36] Kanye, of course, goes out, pushes the envelope, tests the limits, and gets a response. [00:20:41] And so I get, and but my point is, though, he's posting direct text messages with Elon Musk. [00:20:49] And then he's posting them on True Social and they're going back and forth. [00:20:52] And so again, it seems like it's another thing that's in this category of something that's personal to Elon. [00:20:58] And so, you know, I'm okay with that, I guess, as long as it's transparent, right? [00:21:02] As long as we know what's going on, that's certainly better than what we had before if we look at it in the aggregate sense. [00:21:08] But do you think that I'm not asking you to reveal anything, but do you think that there might be an opening for the kind of thing you're outlining about this, you know, period of timeouts that could expand, contract, that would reset itself? [00:21:25] It's, you know, almost like a video game, really, you know, where you get a hit counter and then you're and then you're off the battlefield and you have to wait for your respawn. [00:21:34] Boom, you get your respawn in. [00:21:37] Sorry to folks who aren't gamers out there. [00:21:38] I have no idea what I'm talking about either. [00:21:40] But, you know, it's this idea that you would be able to come back, right? [00:21:44] The idea that you would regenerate your ability to come back. [00:21:48] Again, as long as you complied with whatever the terms of service are. [00:21:52] Yeah, I think there's a possibility for it. [00:21:54] And not to give like Google any credit, but YouTube has a similar system with their three strike system. [00:21:59] But the only thing I disagree with is that like the after the third strike, it's permanent. [00:22:04] I like the idea of a 90-day period to have a hard reset as like a better system. [00:22:10] And then also something that Twitter has always lacked is there's nothing on your profile where you can go and look at your account status. [00:22:16] Like, do I have any like outstanding warnings like in the past? [00:22:21] Have I been suspended before? [00:22:24] Yes, yes. [00:22:25] Other social media that are more transparent. [00:22:28] And again, I don't like to give Facebook or Google any credit, but you can at least see where your account stands. [00:22:34] With Twitter, it's very trouble with Facebook right now. [00:22:37] I've got a 90-day warning on my Facebook right now because I said something about child traffickers that Facebook didn't like. [00:22:45] Yeah, I wonder why. [00:22:48] But yeah, the other thing too, that's curious is that so Ye's suspension was actually, it started off as a 12-hour suspension. [00:22:55] So I don't know how he was suspended for 12 hours, but then it ended up being permanent if he maybe like had another tweet violated. [00:23:06] I'm not, I don't like that idea too. [00:23:07] It's like if you have one violation and you're walked out, I don't, I don't think you should be able to have another tweet actioned like previous tweets because that's right. [00:23:17] That's exactly what I think happened though, because I think the initial tweet triggered the 12-hour suspension through the whatever system they have that's currently set up. [00:23:27] But then I think Elon himself personally got involved and said, no, ban, ban the account. [00:23:33] That's and then he posts, because remember, Elon posted his tweet. [00:23:36] Think of the timing. [00:23:37] Elon posted his tweet, mess around and find out. [00:23:40] And then Kanye West was banned completely. [00:23:44] And interestingly enough, just for the backstory of this, we wouldn't even know that he had had the initial 12-hour ban had he not posted it himself on social. [00:23:54] So we're beyond meta on this one. [00:23:58] We're through several looking glasses on this, but it's interesting to see it all play out in real time. [00:24:05] And again, it's more transparent than what we had in the past. [00:24:09] And at least there's some sort of reasoning given, like being given. [00:24:12] But again, I don't agree that was an insight into violence at all. [00:24:17] But again, like in my case, I got like zero, zero discourse from Twitter and they just shut me off for two and a half years without any sort of reasoning at all and not pointing to a single tweet or anything. [00:24:33] And again, I think their terms of service need to be completely updated because this is just like not tenable. [00:24:41] And he's even said it. [00:24:42] He's like, the terms of service haven't changed since I have taken over yet. [00:24:46] But like, I think they should, honestly. [00:24:49] No, I agree with you on that in terms of that, certainly. [00:24:51] But also at the same time, I do, I think it's better. [00:24:56] Okay. [00:24:57] I think it's better to have a king than have a faceless oligarchy like we had in the past at Twitter. [00:25:04] And I say that as a Twitter user, a prominent Twitter user, that I'd rather at least know and respect someone for their personal decisions, even if they are personal decisions, even if they are idiosyncratic, than this sort of nameless, faceless, oh, it's the trust and safety board, the ministry of trust, the ministry of integrity. [00:25:26] We're doing this for your own good. [00:25:28] I even remember what Vijaya used to say, well, we want to promote healthy conversations. [00:25:34] Healthy conversations. [00:25:35] What in God's name does that mean? [00:25:38] Is this a healthy conversation? [00:25:39] And who are you? [00:25:41] Who are you to say that we are having a healthy conversation? [00:25:44] How about a healthy disagreement? [00:25:46] Okay. [00:25:47] How about a healthy argument? [00:25:48] Why can't we have that? [00:25:49] In terms of that's that is obviously the point of freedom of speech in the first place. [00:25:53] It's not about freedom of speech is not about speech that you agree with. [00:25:57] It's about speech that you disagree with. [00:25:58] You don't have to say that you, if someone, you agree with someone, you don't have to say, well, I respect their right to freedom of speech, right? [00:26:05] It's about speech that you do disagree with. [00:26:06] It's about speech that you don't want. [00:26:08] But I want to get into this also idea of potentially a Twitter safe mode that people would be able to opt into, or maybe it's even the default feature and you opt out of it to say, no, you know, it's, it's, you know, going back to video games, it's all right, you know, my, am I playing Twitter on easy mode or hard mode? [00:26:24] You know, do I want savage mode, you know, safe mode or savage mode? [00:26:28] You, you can get one or the other when you initially signed into Twitter. [00:26:31] And I, I do think that it's interesting, at least that so much is open, but I would agree with you that, that Kanye definitely, and yay, excuse me, I don't mean dead name folks, that he was definitely, he was definitely testing things too quickly. [00:26:47] Obviously, goes as far as he possibly could, just ODs on red pills, and then gets us into a situation, gets himself into a situation where Elon said, look, I just banned the account. [00:26:57] But, you know, and last slide on this real quick, and we'll hold you over the break. [00:27:02] But do you think, do you think, do you think it's actually permanent or would Elon let him back? [00:27:06] Yeah, I don't think it's going to be permanent. [00:27:08] Honestly, well, first of all, he has his phone number. [00:27:10] He's going to be texting him about it. [00:27:11] Second of all, people are lobbying him behind the scenes who are close with both of them. [00:27:18] And third, you've already seen Kathy Griffin, who she violated the terms of service overtly and was doing it on purpose, and she's been brought back. [00:27:26] So it was sort of more a sign to them, a signal to them that, hey, this type of behavior will result in a permanent ban, but you're getting one warning, and now we're going to enforce it. [00:27:45] Because at the same time, we're talking about post facto rulings here that he legitimately thought he could do this and he wouldn't be banned, right? [00:27:55] Well, and again, who's to say it's inciting to violence or not? [00:27:58] Well, it's Elon Musk, isn't it? [00:27:59] It's Elon Musk's decision to say this is inciting violence or this isn't. [00:28:03] It is his platform. [00:28:04] And certainly we should respect that. [00:28:10] What do you think is going to happen six months from now? [00:28:13] Do you think that there's going to be any more mass reinstatements? [00:28:19] We're going to see more of a system in place. [00:28:22] Or do you think it's going to be more of this, which does feel a little bit, I'm just going to say it, it feels a little chaotic? [00:28:27] I think part of the chaos is on purpose. [00:28:31] Because so if you notice at the start of this week, so he, like last week, he held the poll about reinstatements and said it would start this week. [00:28:40] And then he had a whole whole different thing at the beginning of this week. [00:28:44] He picked a fight with Apple and then that got resolved pretty quickly. [00:28:47] So I think, so before he would want to reinstate all of these accounts, which would scare advertisers and scare platforms like Apple and Google, I think he wanted to pick the fight and get it solved before at a time because he knew it was coming down the line if he reinstated all these accounts. [00:29:05] I do think reinstatements are coming, but the only thing is they can't flip a switch and reinstate everyone at once because he does want to keep off the people that were legitimately banned for illegal content or for like spam and bot forms. [00:29:23] But what it looks like they're doing, and I read a report on this, is that they're reinstating the bigger accounts that they can vet thoroughly, more thoroughly, easily, because like bigger name accounts, like over a million and then over 10,000. [00:29:38] And they're going to bring those back first. [00:29:41] And they're kind of going to do it in waves. [00:29:44] I do think in the meantime, before like Twitter Blue and everything is fully operational, it's going to be kind of more chaotic because he has to keep that balance of pleasing the advertisers that are leaving before another revenue stream starts generating in because Twitter Blue is currently paused and they're making like only, I think it's 90% of their money off of advertisers. [00:30:09] So that's currently the only real revenue that they're generating other than people who are already subscribed. [00:30:15] I think it's supposed to relaunch today, but who knows? [00:30:19] But yeah, until they have a broader base of Twitter Blue users and other sources of revenue, I think it's going to be more kind of chaotic, unfortunately. [00:30:29] And by the way, you know, huge attack on Elon today from the New York Times. [00:30:35] Let's do this headline. [00:30:36] Hate speeches rise on Twitter is unprecedented. === Fighting the Disinformation Archipelago (02:02) === [00:30:41] Researchers find. [00:30:42] They wrote, Musk, quote, sent up the bat signal to every kind of racist, misogynist, and homophobe that Twitter was open for business. [00:30:51] The head of the research groups said, but Elon responding to that just now, utterly false. [00:30:57] And I've said this before. [00:30:58] This is how the disinformation archipelago works. [00:31:01] They have their research units and the research unit of the censorship arm, the researcher comes up with this totally bogus study. [00:31:09] And they come in with whether they're at Stanford, whether they're at Harvard. [00:31:13] There's a few of these out there, right? [00:31:15] Where they'll say, in their estimation, right? [00:31:19] Based on their data, their analysis, all of these horrible things are happening. [00:31:25] And then the media reports what the researcher said. [00:31:28] And then the media puts pressure on the platform to say, oh, you need to ban these people. [00:31:33] You need to ban this. [00:31:34] You need to censor more. [00:31:35] But Elon, at least, is coming back and just publicly stating false. [00:31:38] No, I'm not going to do that. [00:31:40] And so I've said, look, I've said before, when it comes to Elon, all right, praise him when he does well, but call him out when he does something we don't like, permanent bans. [00:31:49] I'm not for permanent bans. [00:31:51] Just in general, I'm not for that. [00:31:53] But at the same time, calling out the New York Times. [00:31:56] Oh, yeah, 100%. [00:31:58] So 30 seconds left. [00:31:59] ALX, how do people, this is a great question for you, right? [00:32:02] How do people follow you? [00:32:04] Can they follow you on Twitter by any chance? [00:32:07] Yes, Twitter at ALX. [00:32:11] And I am, yeah, I'm the same way. [00:32:13] I'm pushing get back against Elon as much as I can in things like this. [00:32:17] And my idea for the hate speech thing would be take their list that they currently have because they've been publishing reports of hate speech engagements and let users opt in or out to seeing hate speech or not. [00:32:30] So instead of, because he's currently saying that freedom of reach is not freedom of reach. [00:32:36] Hey, ALX, we're totally out of time, man. [00:32:38] But I completely agree with you, right? [00:32:40] Have a safe mode. [00:32:41] People could opt into it, opt out of it. === Following Truth on Twitter (00:32) === [00:32:43] It's great, folks. [00:32:44] Make sure to catch me on Human Events Daily tonight. [00:32:46] We've got a huge Sunday special coming up at 10 a.m. on Sunday with Matt Tiermond all about the uprising in Brazil and the fact that the CCP is trying to take over Brazil. [00:32:57] Follow me at Jack Pesovic, Human Events Daily Podcast. [00:33:00] Ladies and gentlemen, that's a wrap. [00:33:01] You have my permission. [00:33:05] Thanks for listening. [00:33:06] Make sure to click and subscribe and get your Amfest tickets, amfest.com. [00:33:12] For more on many of these stories and news you can trust, go to CharlieKirk.com.