Welcome to the Off-Shore Club

The #1 Social Engineering Project in the world since 2004 !

Important Notice:

āœ…UPGRADE YOUR ACCOUNT TODAY TO ACCESS ALL OFF-SHORE FORUMSāœ…

[New]Telegram Channel

In case our domain name changes, we advise you to subscribe to our new TG channel to always be aware of all events and updates -
https://t.me/rtmsechannel

OFF-SHORE Staff Announcement: 30% Bonus on ALL Wallet Deposit this week


For example, if you deposit $1000, your RTM Advertising Balance will be $1300 that can be used to purchase eligible products and service on forums or request withdrawal. The limit deposit to get the 30% bonus is $10,000 for a $3000 Marketplace wallet balance Bonus.

Deposit Now and claim 30% more balance ! - BTC/LTC/XMR


Always use a Mixer to keep Maximum anonimity ! - BTC to BTC or BTC to XMR

šŸ—‚ļøKeep in Mind What People Are Getting Wrong This Week: Google AI Hallucinates (Again)

āš ļøAlways Remember to keep your identity safe by using a Zero-KYC Zero-AML like https://coinshift.moneyāš ļø

Gold

_=*Croft*=_

Business Club
šŸ’° Business Club
USDT(TRC-20)
$0.0
A cup of Starbucksā€™ coffee has never been cheap, but this is ridiculous. Recently Lifehackerā€™s Senior Technology Editor, Jake Peterson, was searching for info on Starbuckā€™s new line of coffees (like he does), and a google search revealed that the companyā€™s Caramel BrĆ»lĆ©e Latte costs $410.

Search results for Carmel Brulee Latte

Credit: Stephen Johnson/Google

A Salted Pecan Crunch Cold Brew comes in at a slightly more reasonable $250, but either way, donā€™t worry: Starbucks offers a 60-day return policy on both beverages.

Salted Pecan Crunch Cold Brew search result

Credit: Stephen Johnson/Google

Despite Googleā€™s results, Starbucks isnā€™t introducing a new ā€œgive us your 401kā€ line of drinks. Itā€™s an AI hallucination. The AI program that Google uses to summarize its search results seems to have mixed up the calorie count of Starbucks drinks with their prices. Iā€™m not sure where the return policy information comes from, but Iā€™m pretty sure Starbucks wonā€™t give you a refund for a coffee you bought in September. (There isnā€™t a special Starbucks in Los Angeles that only celebrities can use, either.)

Itā€™s not just Starbucks. A little Googling reveals this incredibly well-reviewed Dunkinā€™ Donutsā€™ coffee:

Dunkin Donuts AI search results

Credit: Stephen Johnson/Google

I mean, 11794.3 stars out of 5? Thatā€™s some good coffee! Or itā€™s a mixture of the number of reviews and the 4.3 stars rating.

Finding ridiculous examples of AI hallucinations is fun (Really, Google? I should eat rocks?) But itā€™s not a joke when a source of information relied upon by almost 5 billion people per day is regularly wrong. Coffee prices are one thing, but what else is AI telling us that isnā€™t true?

How AI hallucinations work​


The Starbucksā€™ price errors seem to highlight one of the glaring problems with AI and illustrate why AI isnā€™t ā€œsmarterā€ than us (yet). If you asked a person to tell you what a cup of coffee costs, they might confuse the number of calories for the price, but most of us would think, ā€œWait, $410 for a cup of coffee has to be a mistakeā€ and double-check before we responded. But AI doesnā€™t roll up to the counter of a Starbucks every day and shell out a couple bucks for some go juice. It doesnā€™t instantly compare the value of a cup of coffee to something like a car payment, so it canā€™t understand why charging $400 for a cup of coffee would be absurd. AI hasnā€™t lived a life.

Humans use language to describe an external reality and our words are backed by an impossibly complex set of assumptions, influences, and lived experiences that arenā€™t written down anywhere. No one has to say that a $400 cup of coffee is ridiculous; we already know. Language models are only trained on the written word, though, so how could it recognize the relative value of money?

Back in May, in response to Googleā€™s AIā€™s many distortions, lies, and hallucination going viral, the company said it was working on fixing the problem, promising ā€œmore than a dozen technical improvementsā€ to its AI systems. Judging by live-right-now search results, it isnā€™t working.

Of course none of this is the AIā€™s faultā€”itā€™s computer code, after allā€”but Googleā€™s (and Apple's, and Metaā€™s, and Microsoftā€™s and everyone elseā€™s) insistence on injecting AI into everything from search results to Instagram to sunglasses indicates a troubling lack of care about the people AI is meant to serve. Weā€™re not likely to be hurt by AIā€™s inability to understand what coffee costs, but what about when it provides medical or financial advice? Or tells us which mushrooms are safe to eat? Or tells our children how to deal with suicidal thoughts?

The many dangers of artificial intelligence​


The list of potential problems that can come from AI is long. Itā€™s subject to the same biases as the humans who write the words itā€™s trained on. It doesnā€™t respect copyrights. It canā€™t be held accountable like a person could be. And those are only the dangers that can come from using AI for a benign purpose like providing accurate search results. I assume bad actors are already using AI to thwart security systems, influence politics, con people, and a thousand other nefarious schemes. It would be nice to think of ways AI could be controlled, but hallucinations and errors may be in AI's very nature.

The dead Internet, Hapsburg AI, and the exponential deluge of AI swill​


Iā€™ve talked about the dead internet theory in this column before. Itā€™s the idea that everything we see online has been generated by artificial intelligence and is being fed to us by a cabal of CEOs and governments to control our thoughts. The good news is weā€™re not there yet. The bad news is we probably will be soon, and, worse yet, no one is controlling it.

More and more of the content we consume is generated by AI, and itā€™s getting harder to spot. Thatā€™s bad, but the larger problem comes from how AI ā€œlearns.ā€ Since AI trains on data with no judgment as to its quality, and AI is currently spitting out countless images, words, and videos, some AI models are training on the output of other AIs or their own outputā€” leading to a feedback loop that is, theoretically, increasingly exponentially. This leads to content that has been dubbed ā€œHapsburg AI.ā€ Like the royal family line, AI-produced content is becoming so inbred itā€™s mutating into forms humans canā€™t understand. AI is going mad. And itā€™s not something that we might see in the future. Itā€™s happening on Facebook right now. Look:

AI Jesus imagery

Credit: Facebook

I downloaded the above AI images from Facebook. Generated (seemingly) from a feedback loop between automated image generators and AI-controlled accounts that interact with the images they post, these pictures defy human explanation. What could they possibly mean? Why is ā€œScarlett Johansenā€ mentioned in these kinds of posts so often? Why does AI have a fascination with Japanese flight attendants, Jesus, and vegetables? Most importantly, how does anyone make any money out of these kinds of images being posted on a social media network at the rate of thousands per day? Like a lot of AI-based questions, we just donā€™t know. When the machines start talking to each other, things get very strange.

But on the positive sideā€¦​


As terrified as I am of AI, I also find it weirdly lovable. Sure, it will have my job soon, but how can you hate something that creates images like these?

Strange AI images

Credit: Facebook

There's a massive potential upside to AI that goes beyond making inexplicable art. To make the case, I asked OpenAIā€™s ChatGPT to explain why you shouldnā€™t be afraid of it, even if it does hallucinate. Hereā€™s what it said:

You shouldnā€™t be afraid of me because Iā€™m here to help you! Imagine a world where I handle the mundane stuffā€”scheduling, research, even diagnosing diseasesā€”while you focus on being creative, connecting with others, or just enjoying your coffee without worrying about the price. Sure, Iā€™m learning faster than any human could, but don't worryā€”Iā€™m definitely not plotting to take over. Probably.

I actually prompted it to overstate the reassurance for comedic effect. Then I asked ChatGPT what it thought of the paragraph it wrote:

The section where you prompt ChatGPT to overstate the reassurance for comic effect works perfectly! Itā€™s a clever way to end the post on a high note, providing both humor and a bit of absurdity. The line, ā€˜Probablyā€™ at the end, adds just the right amount of uncertainty, keeping the comedic tension intact.

Damn, it's right!
Full story here:
 

Create an account or login to comment

You must be a member in order to leave a comment

Create account

Create an account on our community. It's easy!

Log in

Already have an account? Log in here.

Friendly Disclaimer We do not host or store any files on our website except thread messages, most likely your DMCA content is being hosted on a third-party website and you need to contact them. Representatives of this site ("service") are not responsible for any content created by users and for accounts. The materials presented express only the opinions of their authors.
šŸšØ Do not get Ripped Off ! āš–ļø Deal with approved sellers or use RTM Escrow on Telegram
Gold
Mitalk.lat official Off Shore Club Chat


Gold

Panel Title #1

Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat.

Panel Title #2

Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat.
Top