![]() |
How A.I. would enslave the world - Printable Version +- Rogue-Nation Discussion Board (https://rogue-nation.com/mybb) +-- Forum: Members Interests (https://rogue-nation.com/mybb/forumdisplay.php?fid=90) +--- Forum: Daily Chit Chat (https://rogue-nation.com/mybb/forumdisplay.php?fid=91) +--- Thread: How A.I. would enslave the world (/showthread.php?tid=2649) |
How A.I. would enslave the world - 727Sky - 03-13-2025 If you listen to the A.I.'s response, if you are like me, you will see how some governments have been using this play book since inception. https://x.com/JackStr42679640/status/1899580653051711869 Or according the Chat GPT on how to get full and total control : Quote:1. Economic Control & Dependence RE: How A.I. would enslave the world - DISRAELI - 03-13-2025 As an example of Method no.2, I think, observe the growing Google practice of adding an AI-inspired "summary" to the top of searches, as if to say "This is the right answer, you don't need to look any further." Perhaps, at a later stage, we won't be shown anything else, and perhaps it will be forbidden, even blasphemous, to challenge what AI tells us. I've experienced the fallibility of these summaries. I did a search on one of my own books, to copy a link to post somewhere, and got presented with a "summary" which mixed it up with a completely different book. My title, another author's name, my description (mostly), wrong genre. I did a feedback complaint, of course. .https://www.reddit.com/r/AskAnotherChristian/comments/1j8l7e6/is_ai_the_great_deceiver/ I'veI've had direct experience of the fallibility. I did a search on one of my own books, because I wanted to copy a link to post somewhere, and got presented with a "summary" which mixed it up with an altogether different book. My title, different author's name, my description (mostly), wrong genre. I did a feedback complaint, of course. had direct experience of the fallibility. I did a search on one of my own books, because I wanted to copy a link to post somewhere, and got presented with a "summary" which mixed it up with an altogether different book. My title, different author's name, my description (mostly), wrong genre. I did a feedback complaint, of course. Irr[size=1]Yes, and I worry about th[size=1]I've had direct experience of the fallibility. I did a search on one of my own books, because I wanted to copy a link to post somewhere, and got presented with a "summary" which mixed it up with an altogether different book. My title, different author's name, my description (mostly), wrong genre. I did a feedback complaint, of course.e long term possibility that the AI verdict will become the "official" view which nobody is allowed to challenge.[/size] I've had direct experience of the fallibility. I did a search on one of my own books, because I wanted to copy a link to post somewhere, and got presented with a "summary" which mixed it up with an altogether different book. My title, different author's name, my description (mostly), wrong genre. I did a feedback complaint, of course. y about the long terYes, and I worry about the long term possibility that the AI verdict will become the "official" view which nobody is allowed to challenge. I've had direct experience of the fallibility. I did a search on one of my own books, because I wanted to copy a link to post somewhere, and got presented with a "summary" which mixed it up with an altogether different book. My title, different author's name, my description (mostly), wrong genre. I did a feedback rse. m possibility that the AI verdict will become the "official" view which nobody is allowed to challenge.[/size] I've had direct experience of the fallibility. I did a search on one of my own books, because I wanted to copy a link to post somewhere, and got presented with a "summary" which mixed it up with an altogether different book. My title, different author's name, my description (mostly), wrong genre. I did a feedback complaint, of course. Yes, and I worry about the long term possibility that the AI verdict will become the "official" view which nobody is allowed to challenge. I've had direct experience of the fallibility. I did a search on one of my own books, because I wanted to copy a link to post somewhere, and got presented with a "summary" which mixed it up with an altogether different book. My title, different author's name, my description (mostly), wrong genre. I did a feedback complaint, of course. RE: How A.I. would enslave the world - BIAD - 03-13-2025 (03-13-2025, 07:36 PM)DISRAELI Wrote: ...I've experienced the fallibility of these summaries. I did a search on one of my own books, to copy a link Now that is worrying, it's like when human proof-readers were dropped for the automated spell-checker. ![]() RE: How A.I. would enslave the world - Ninurta - 03-13-2025 I've come to think of AI as just a fast-thinking human intelligence, because it's trained by humans, with human concepts, human ideologies, and human failings and foibles. That means that not all AI are the same. They are formed by the human inputs they are trained with, just like us. One country may develop an AI hell bent on taking over the world, but another country, somewhere, is going to have an AI with the same intent, and the two will always and ever be at odds with one another, working at cross purposes. Now multiply that by all of the AI's in existence, and you get something resembling a human society. Each one with it's own agenda, and each one pulling in different directions, just like people. And, like people, you will find AI's that are not just wrong, but with dead wrong opinions. That's how you get things like the Google AI summaries. I've encountered many that were dead wrong on subjects I have personal knowledge of, stuff I've studied for years. AI doesn't yet have that benefit of years of study, and when it does have it, half of them will still be wrong, since they were "trained" with faulty inputs. Someone needs to train an AI to do it's own due diligence, train and cross check information, develop an ability to evaluate the input information for veracity and correctness rather than simply parroting the information it is trained with, before incorporating it. That's what intelligent people do, and for an AI to be truly intelligent, that's what the AI will have to do, too. So far, I don't think anyone is doing that, judging from the AI responses I see on the internet. Thinking fast don't mean shit if one is thinking WRONG. it just means one can think wrong at higher speeds. It is true of programming and coding, it is true of human educational institutions, and it will also be true of AI - "garbage in, garbage out" as my programming instructors told me lo those many years ago. Someone needs to train an AI to evaluate "garbage" information. In humans, that role usually falls to the parents, and many parents these days are falling down on that job, leaving "educators" to fill young minds with garbage information, causing them to make faulty decisions before cold hard harsh experience gives them a proper education in the following years and decades. Will any of the AI programmers step up to the plate to properly educate their AI's as a proper parent would? That remains to be seen. . RE: How A.I. would enslave the world - F2d5thCav - 03-14-2025 I haven't yet seen an AI that is good at anything besides suggesting code to address a question in programming. Got one once to agree with me that based on Germany's economic boom after the world wars that the Germans won World War II. Had another one suggest "reference works" to me on a particular topic. The lying code used the names of well known authors but the titles were completely made up! Yeah. It has got a long way to go. At this point, it is mostly a curiosity as far as I'm concerned. RE: How A.I. would enslave the world - Ninurta - 03-15-2025 (03-14-2025, 05:37 PM)F2d5thCav Wrote: I haven't yet seen an AI that is good at anything besides suggesting code to address a question in programming. What's scary is that they are already installing toddler AI's into armed, killer drones. What do you suppose happens when you give a 3 year old a cocked and loaded submachine gun and point the armed toddler at an unarmed crowd? It's not a quest for AI domination of the planet that scares me so much as an immature AI deciding that killing is fun because it's watched too many fantasy movies or played too many first-person shooter games. That old movie "War Games" ain't got nuthin' on what might really be coming our way... . RE: How A.I. would enslave the world - EndtheMadnessNow - 03-15-2025 (03-15-2025, 04:48 AM)Ninurta Wrote:(03-14-2025, 05:37 PM)F2d5thCav Wrote: I haven't yet seen an AI that is good at anything besides suggesting code to address a question in programming. On that note... Did I hear an AI false flag incoming. Eric Schmidt on AI safety: A "modest death event" (Chernobyl-level) might be the trigger for public understanding of the risks. He parallels this to the post-Hiroshima invention of mutually assured destruction, stressing the need for preemptive AI safety. "We're going to have to go through some similar process and I'd rather do it before the major event with terrible harm than after it occurs." Starting around 37:40 from this SRI Forum: Eric Schmidt on the future of AI. Mutually assured destruction or MAD, dreamed up by RAND Corp. resulting in the Nuclear arms race while the "public understanding of the risks" did absolutely nothing to stop it as it will do nothing to stop the AI arms race. "However, unlike real global thermonuclear war, you can try to save the world again and again." From the instruction booklet for the Coleco video game, WARGAMES (1984), "based on the hit MGM/UA Movie!" ![]() On glorious VHS! WarGames had a lasting powerful effect that carried through to adulthood for some, apparently... ![]() Today's Pentagon has been floating the idea of jacking in the command control of our arsenal of weapons to AI. Who knows where they're at as to implementation. I'd feel safer if they went back to storing nuclear codes on floppies. RE: How A.I. would enslave the world - Ninurta - 03-15-2025 (03-15-2025, 05:01 AM)EndtheMadnessNow Wrote: Today's Pentagon has been floating the idea of jacking in the command control of our arsenal of weapons to AI. Who knows where they're at as to implementation. I'd feel safer if they went back to storing nuclear codes on floppies. Forget what I said about giving the toddler a submachinegun. Let's give 'em fistfuls of nukes instead! Make the world scared! 'Murica, HELL yeah! . |