9 points

Negative IQ points?

permalink
report
reply
6 points

chatbots and ai are just dumber 1990s search engines.

permalink
report
reply
3 points

I remember 90s search engines. AltaVista was pretty ok a t searching the small web that existed, but I’m pretty sure I can get better answers from the LLMs tied to Kagi search.

AltaVista also got blown out of the water by google(back when it was just a search engine), and that was in the 00s not the 90s. 25 to 35 years ago is a long time, search is so so much better these days(or worse if you use a “search” engine like Google now).

Don’t be the product.

permalink
report
parent
reply
1 point

Depending on what I needed I remember using AltaVista, AskJeeves, Dogpile, and I feel like later on MetaCrawler or something like that (would search multiple search engines for you and ordered them scored based on platform and relevancy iirc?)

permalink
report
parent
reply
1 point

And Hastalavista if you wanted to find things that Altavista didn’t.

permalink
report
parent
reply
1 point

I miss AskJeeves. Those were some great commercials too.

permalink
report
parent
reply
4 points

i can feel it too when I use it. that is why i use it only for trivial things if at all.

permalink
report
reply
247 points

people tend to become dependent upon AI chatbots when their personal lives are lacking. In other words, the neediest people are developing the deepest parasocial relationship with AI

Preying on the vulnerable is a feature, not a bug.

permalink
report
reply
4 points

And it’s beyond obvious in the way LLMs are conditioned, especially if you’re used them long enough to notice trends. Where early on their responses were straight to the point (inaccurate as hell, yes, but that’s not what we’re talking about in this case) today instead they are meandering and full of straight engagement bait - programmed to feign some level of curiosity and ask stupid and needless follow-up questions to “keep the conversation going.” I suspect this is just a way to increase token usage to further exploit and drain the whales who tend to pay for these kinds of services, personally.

There is no shortage of ethical quandaries brought into the world with the rise of LLMs, but in my opinion the locked-down nature of these systems is one of the most problematic; if LLMs are going to be the commonality it seems the tech sector is insistent on making happen, then we really need to push back on these companies being able to control and guide them in their own monetary interests.

permalink
report
parent
reply
64 points

I kind of see it more as a sign of utter desperation on the human’s part. They lack connection with others at such a high degree that anything similar can serve as a replacement. Kind of reminiscent of Harlow’s experiment with baby monkeys. The videos are interesting from that study but make me feel pretty bad about what we do to nature. Anywho, there you have it.

permalink
report
parent
reply
29 points

And the amount of connections and friends the average person has has been in free fall for decades…

permalink
report
parent
reply
3 points

I dunno. I connected with more people on reddit and Twitter than irl tbh.

Different connection but real and valid nonetheless.

I’m thinking places like r/stopdrinking, petioles, bipolar, shits been therapy for me tbh.

permalink
report
parent
reply
13 points

That utter-desparation is engineered into our civilization.

What happens when you prevent the “inferiors” from having living-wage, while you pour wallowing-wealth on the executives?

They have to overwork, to make ends meet, is what, which breaks parenting.

Then, when you’ve broken parenting for a few generatios, the manufactured ocean-of-attachment-disorder manufactures a plethora of narcissism, which itself produces mass-shootings.

2024 was down 200 mass-shootings, in the US of A, from the peak of 700/year, to only 500.

You are seeing engineered eradication of human-worth, for moneyarchy.

Isn’t ruling-over-the-destruction-of-the-Earth the “greatest thrill-ride there is”?

We NEED to do objective calibration of the harm that policies & political-forces, & put force against what is actually harming our world’s human-viability.

Not what the marketing-programs-for-the-special-interest-groups want us acting against, the red herrings…

They’re getting more vicious, we need to get TF up & begin fighting for our species’ life.

_ /\ _

permalink
report
parent
reply
9 points

a sign of utter desperation on the human’s part.

Yes it seems to be the same underlying issue that leads some people to throw money at only fans streamers and such like. A complete starvation of personal contact that leads people to willingly live in a fantasy world.

permalink
report
parent
reply
9 points

That was clear from GPT-3, day 1.

I read a Reddit post about a woman who used GPT-3 to effectively replace her husband, who had passed on not too long before that. She used it as a way to grief, I suppose? She ended up noticing that she was getting too attach to it, and had to leave him behind a second time…

permalink
report
parent
reply
2 points

Ugh, that hit me hard. Poor lady. I hope it helped in some way.

permalink
report
parent
reply
-5 points
*

These same people would be dating a body pillow or trying to marry a video game character.

The issue here isn’t AI, it’s losers using it to replace human contact that they can’t get themselves.

permalink
report
parent
reply
27 points

You labeling all lonely people losers is part of the problem

permalink
report
parent
reply
-7 points

If you are dating a body pillow, I think that’s a pretty good sign that you have taken a wrong turn in life.

permalink
report
parent
reply
7 points

More ways to be an addict means more hooks means more addicts.

permalink
report
parent
reply
3 points

Me and Serana are not just in love, we’re involved!

Even if she’ s an ancient vampire.

permalink
report
parent
reply
136 points

permalink
report
reply
10 points

The quote was originally on news and journalists.

permalink
report
parent
reply
3 points

I remember thinking this when I was like 15. Every time they mentioned tech, wtf this is all wrong! Then a few other topics, even ones I only knew a little about, so many inaccuracies.

permalink
report
parent
reply
-31 points
*

Another realization might be that the humans whose output ChatGPT was trained on were probably already 40% wrong about everything. But let’s not think about that either. AI Bad!

permalink
report
parent
reply
17 points

This is a salient point that’s well worth discussing. We should not be training large language models on any supposedly factual information that people put out. It’s super easy to call out a bad research study and have it retracted. But you can’t just explain to an AI that that study was wrong, you have to completely retrain it every time. Exacerbating this issue is the way that people tend to view large language models as somehow objective describers of reality, because they’re synthetic and emotionless. In truth, an AI holds exactly the same biases as the people who put together the data it was trained on.

permalink
report
parent
reply
8 points
*

I’ll bait. Let’s think:

-there are three humans who are 98% right about what they say, and where they know they might be wrong, they indicate it

  • now there is an llm (fuck capitalization, I hate the ways they are shoved everywhere that much) trained on their output

  • now llm is asked about the topic and computes the answer string

By definition that answer string can contain all the probably-wrong things without proper indicators (“might”, “under such and such circumstances” etc)

If you want to say 40% wrong llm means 40% wrong sources, prove me wrong

permalink
report
parent
reply
0 points

It’s more up to you to prove that a hypothetical edge case you dreamed up is more likely than what happens in a normal bell curve. Given the size of typical LLM data this seems futile, but if that’s how you want to spend your time, hey knock yourself out.

permalink
report
parent
reply
3 points

AI Bad!

Yes, it is. But not in, like a moral sense. It’s just not good at doing things.

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


Community stats

  • 23K

    Monthly active users

  • 15K

    Posts

  • 633K

    Comments