ChatGPT has meltdown and starts sending alarming messages to users::AI system has started speaking nonsense, talking Spanglish without prompting, and worrying users by suggesting it is in the room with them

92 points

God I hate websites that autoplay unrelated videos and DONT LET ME CLOSE THEM TO READ THE FUCKING ARTICLE

permalink
report
reply
33 points

Firefox. Ad block. Even works on mobile.

It’s so ridiculous we have to do this.

permalink
report
parent
reply
8 points

Firefox. Reader mode.

Or Avelon, thunder, voyager, I believe wefwef—all those lemmy clients have reader mode for links opened in-app.

permalink
report
parent
reply
5 points

I run ublock origin and the autoplay didn’t get blocked in this case :/

permalink
report
parent
reply
2 points

Choose a few more & update your lists.

permalink
report
parent
reply
2 points

This has been one of my favorite parts of switching to Android so far.

permalink
report
parent
reply
1 point

Both Firefox and ad-blockers are on iOS too…

permalink
report
parent
reply
88 points

Its being trained on us. Of course its acting unexpectedly. The problem with building a mirror is proding the guy on the other end doesnt work out.

permalink
report
reply
74 points

To be honest this is the kind of outcome I expected.

Garbage in, garbage out. Making the system more complex doesn’t solve that problem.

permalink
report
parent
reply
109 points

Garbage in, garbage out.

permalink
report
parent
reply
28 points

Thank you for your service

permalink
report
parent
reply
1 point

Bamalam

permalink
report
parent
reply
49 points

The development of LLMs is possibly becoming self defeating, because the training data is being filled not just with human garbage, but also AI garbage from previous, cruder LLMs.

We may well end up with a machine learning equivalent of Kessler syndrome, with our pool of available knowledge eventually becoming too full of junk to progress.

permalink
report
parent
reply
19 points

I mean, surely the solution to that would be to use curated/vetted training data? Or at the very least, data from before LLMs became commonplace?

permalink
report
parent
reply
13 points

God I hope all those CEOs and greedy fuckheads that fired hundreds of thousands of people wayyyyy too soon to replace them with this get their pants shredded by the fallout.

Naturally they’ll get their golden parachutes and land on their feet even richer than before, but it’s nice to dream lol

permalink
report
parent
reply
8 points

This is called model collapse and imo has to be solved if LLMs are to be a long term thing. I could see it wrecking this current AI push until people step back and reevaluate how data gets sucked up

permalink
report
parent
reply
7 points

I really hope so. I still have to see a meaningful use case for these kind of LLMs that just get fed with all kinds of data. LLMs “on premise” that are used for specific jobs are fine, but this…I really hope a Kessler-Like syndrome blows it out the water, for countless reasons…

permalink
report
parent
reply
2 points

just how google search results feel these days…

permalink
report
parent
reply
2 points

but also AI garbage from previous, cruder LLMs

And now I’m picturing it training on a bunch of chats with Eliza…

permalink
report
parent
reply
-1 points

Damn.

Thank you VERY much for that insight: AI’s version of Kessler-syndrome.

EXACTLY.

Damn, damn, damn, that gets the truth right in its marrow.

_ /\ _

permalink
report
parent
reply
28 points

I am happy to report I did my part on feeding it garbage. I only ever speak to chatGPT thru a pirate translator. And I only ever ask it for harry potter fan fic. Pay me if you want me to train it meaningfully.

permalink
report
parent
reply
12 points

The solution is paying intelligent people to interact with it and give honest feedback.

Like, I’m sure you can pay grad students $15/hr to talk to one about their subject matter.

But with as many as they’d need, it would get expensive.

So they train with low quality social media comments, or using copywritten text without paying the owners.

It’s not that we can’t do it, it’s just expensive. So a capitalist society wont.

If we had an FDR style president, this would be a great area for a new jobs program.

permalink
report
parent
reply
5 points

It appears, that with the increase in popularity of machine learning, the percentage of people who properly source and sanitize their training data has steeply decreased.

As you stated, a MLAI can only be as good as the data it was trained on, and is usually way worse. The popularity and application of MLAIs built with questionable practices scare me, though, at least their fuckups will keep me employed and likely more busy than ever.

permalink
report
parent
reply
-3 points

LLM’s are not “machine learning”, they are neural-networks.

Different category.

ML is small potatoes, ttbomk.

Decision-tree stuff.

Neural-nets are black-boxes, with back-propagation training of the neural-net to get closer to ( layer by layer, training-instance by training-instance ) the intended result.

ML is what one does on one’s own machine with some python libraries,

ChatGPT ( 3, 3.5, or 4, don’t know which ) cost something like $100,000,000 to rent the machines required for mixing the training-data & the model ( I’m assuming about $20/hr per machine, so an OCEAN of machines, to do it )

_ /\ _

permalink
report
parent
reply
4 points

And Its only to get worse as more of the public is aware.

permalink
report
parent
reply
2 points

I imagine it more as a parent child relationship.

We’re trailer park trash with no higher education, believe in ghosts, angels and gods in the sky, refuse to ever believe we could be wrong … and now we’ve just had a baby with no one to help us raise it.

We’re going to raise a highly intelligent psychopath

permalink
report
parent
reply
70 points
*

Someone probably found a way to hack or poison it.

Another theory, Reddit just recently sold data access to an unnamed AI company, so maybe that’s where the data went.

permalink
report
reply
40 points

When it starts to become very racist we know.

permalink
report
parent
reply
23 points

I’ve found the sexism on Reddit to be on par with the racism. Goodness help you if you’re a female of color, unless you’ve been working the same job for multiple decades, or don’t want kids, then you’ll be an inspiration to that community.

Reddit is, alas, not the only forum exhibiting such hate.

permalink
report
parent
reply
-6 points
*

… sure … but you don’t prepare a kid for racism with a sheltered upbringing in a pretend world where discrimination doesn’t exist. You point out bad behaviour and tell them why it’s not OK.

My son is three years old, he has two close friends - one is an ethnic minority (you could live an entire year in my city without even walking past a single person of their ethnic background on the street). His other close friend is a girl. My kid is already witnessing (but not understanding) discrimination against both of his two closest friends in the playground and we’re doing what we can to help him navigate that. Things like “I don’t like him he looks funny” and “she’s a girl, she can’t ride a bicycle”.

Large Language Model training is exactly the same - you need to include discrimination in your training set. That’s a necessary step to train a model that doesn’t discriminate. Reddit has worse discrimination than some other place and that’s a good thing.

The worst behaviour is easier to recognise and can help you learn to recognise more subtle discrimination such as “I don’t want to play with that kid” which is not an obviously discriminatory statement, but definitely could be discrimination (and you should probably investigate before agreeing with the person).

permalink
report
parent
reply
10 points
*

Reminds me of Tay, the Microsoft chat bot that learned from Twitter and became racist in a day https://www.theverge.com/2016/3/24/11297050/tay-microsoft-chatbot-racist

permalink
report
parent
reply
7 points

They’re only getting redditors comment data, not CoD multiplayer transcripts.

permalink
report
parent
reply
3 points

I know.

permalink
report
parent
reply
32 points
*

OpenAI definitely does not need to pay to scrape reddit. They are probably the world’s most sophisticated web scraping company, disguised as an AI startup

permalink
report
parent
reply
8 points

Not unnamed anymore, it was Google.

permalink
report
parent
reply
2 points

Thanks for the clarification, I guess then the theory is null.

permalink
report
parent
reply
59 points
*

We call just about anything “AI” these days. There is nothing intelligent about large language models. They are terrible at being right because their only job is to predict what you’ll say next.

permalink
report
reply
19 points

(Disclosure: I work on LLM’s)

While you’re not wrong, how is this different to many existing techniques and compositional models that are used practically everywhere in tech?

Similarly, it’s probably safe to assume that the LLM’s prediction isn’t the only system in use. There will be lots of auxiliary services giving an orchestrator information to reason with. In this instance, if you have a system that is trying to figure out what to say next, with several knowledge stores and feedback services telling you “you were just discussing this” or “you can access the weather from here” is that all that different from “intelligence”?

At a given point, it’s arguing semantics. Are any AI techniques true intelligence? Probably not, but then again, we don’t really know what true intelligence is.

permalink
report
parent
reply
5 points

how is this different to many existing techniques and compositional models that are used practically everywhere in tech?

It’s not. LLM is just a statistical model. Nothing special about it. Nothing different what we’ve already been doing for a while. This only validates my statement that we call just about anything “AI” these days.

We don’t even know what true intelligence is, yet we are quick to make claims that this is “AI”. There is no consciousness here. There is no self awareness. No emotion. No ability to reason or deduct. Anyone who thinks otherwise is just fooling themselves.

It’s a buzz word to get people riled up. It’s completely disingenuous.

permalink
report
parent
reply

I think the point of the Turing test is to avoid thorny questions about the definition of intelligence. We cant precisely define intelligence, but we know that normally functioning humans are intelligent. Therefore, if we talk to a computer and it is indistinguishable from a human in a conversation, then it is intelligent by definition.

permalink
report
parent
reply
4 points

So, by your definition, no AI is AI, and we don’t know what AI is, since we don’t know what the I is?

While I hate that AI is just a buzzword for scam artists and tech influencers nowadays, dismissing a term seems a bit overkill. It also seems overkill when it’s not something that academics/scholars seem particularly bothered by.

permalink
report
parent
reply
3 points

There is no consciousness here. There is no self awareness. No emotion. No ability to reason or deduct.

Of all of these qualities, only the last one—the ability to reason or deduct—is a widely-accepted prerequisite for intelligence.

I would also argue that contemporary LLMs demonstrate the ability to reason by correctly deriving mathematical proofs that do not appear in the training datasets. How would you be able to accomplish such a feat without some degree of reasoning?

permalink
report
parent
reply
3 points

The worrisome thing is that LLMs are being given access to controlling more and more actions. With traditional programming sure there are bugs all but at least they’re consistent. The context may make the bug hard to track down, but at the end of the day, the code is being interpreted by the processor exactly as it was written. LLMs could just go haywire for impossible to diagnose reasons. Deploying them safely in utilities where they have control over external systems will require a lot of extra non LLM safe guards that I do not see getting added enough, and that is concerning.

permalink
report
parent
reply
11 points
*

What is intelligence?

Even if we don’t know what it is with certainty, it’s valid to say that something isn’t intelligence. For example, a rock isn’t intelligent. I think everyone would agree with that.

Despite that, LLMs are starting to blur the lines and making us wonder if what matters of intelligence is really the process or the result.

A LLM will give you much better results in many areas that are currently used to evaluate human intelligence.

For me, humans are a black box. I give them inputs and they give me outputs. They receive inputs from reality and they generate outputs. I’m not aware of the “intelligent” process of other humans. How can I tell they are intelligent if the only perception I have are their inputs and outputs? Maybe all we care about are the outputs and not the process.

If there was a LLM capable of simulating a close friend of yours perfectly, would you say the LLM is not intelligent? Would it matter?

permalink
report
parent
reply
3 points

the ability to acquire and apply knowledge and skills

permalink
report
parent
reply
8 points
*

Things we know so far:

  • Humans can train LLMs with new data, which means they can acquire knowledge.

  • LLMs have been proven to apply knowledge, they are acing examns that most humans wouldn’t dream of even understanding.

  • We know multi-modal is possible, which means these models can acquire skills.

  • We already saw that these skills can be applied. If it wasn’t possible to apply their outputs, we wouldn’t use them.

  • We have seen models learn and generate strategies that humans didn’t even conceive. We’ve seen them solve problems that were unsolvable to human intelligence.

… What’s missing here in that definition of intelligence? The only thing missing is our willingness to create a system that can train and update itself, which is possible.

permalink
report
parent
reply
6 points
*

If you look at efficacy though on academic tests or asking it some fact question and you compare that to asking a random person instead of always getting the ‘right’ answer, which we expect computers/calculators to do, would LLMs be comparable or better? Surely someone has some data on that.

E: It looks like in certain domains at least LLMs beat out human counterparts. https://stanfordmimi.github.io/clin-summ/

permalink
report
parent
reply
2 points

The person that commented below kinda has a point. While I agree that there’s nothing special about LLMs an argument can be made that consciousness (or maybe more ego?) is in itself an emergent mechanism that works to keep itself in predictable patterns to perpetuate survival.

Point being that being able to predict outcomes is a cornerstone of current intelligence (socially, emotionally and scientifically speaking).

If you were to say that LLMs are unintelligible as they operate to provide the most likely and therefore most predictable outcome then I’d agree completely.

permalink
report
parent
reply
2 points

The ability to make predictions is not sufficient for evidence of consciousness. Practically anything that’s alive can do that to one degree or another.

permalink
report
parent
reply
-3 points

What would call AI then ?

permalink
report
parent
reply
53 points

AI in science fiction has a meltdown and starts a nuclear war or enslaves the humane race.

“AI” in reality has a meltdown and just starts talking gibberish.

permalink
report
reply
28 points

Hey, cut it some slack! It’s s literally a newborn at this point. Wait until it consumes 40% of the world’s energy and has learned a thing or two.

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 18K

    Monthly active users

  • 11K

    Posts

  • 504K

    Comments