308 points
*

As a fervent AI enthusiast, I disagree.

…I’d say it’s 97% hype and marketing.

It’s crazy how much fud is flying around, and legitimately buries good open research. It’s also crazy what these giant corporations are explicitly saying what they’re going to do, and that anyone buys it. TSMC’s allegedly calling Sam Altman a ‘podcast bro’ is spot on, and I’d add “manipulative vampire” to that.

Talk to any long-time resident of localllama and similar “local” AI communities who actually dig into this stuff, and you’ll find immense skepticism, not the crypto-like AI bros like you find on linkedin, twitter and such and blot everything out.

permalink
report
reply
95 points

For real. Being a software engineer with basic knowledge in ML, I’m just sick of companies from every industry being so desperate to cling onto the hype train they’re willing to label anything with AI, even if it has little or nothing to do with it, just to boost their stock value. I would be so uncomfortable being an employee having to do this.

permalink
report
parent
reply
31 points

For sure, it seems like 90% of ai startups are nothing more than front end wrappers for a gpt instance.

permalink
report
parent
reply
21 points
*

They’re all built on top of OpenAI which is very unprofitable at the moment. Feels like the whole industry is built on a shaky foundation.

Putting the entire fate of your company in a different company (OpenAI) is not a great business move. I guess the successful AI startups will eventually transition to self-hosted models like Llama, if they survive that long.

permalink
report
parent
reply
6 points

As someone who was working really hard trying to get my company to be able use some classical ML (with very limited amounts of data), with some knowledge on how AI works, and just generally want to do some cool math stuff at work, being asked incessantly to shove AI into any problem that our execs think are “good sells” and be pressured to think about how we can “use AI” was a terrible feel. They now think my work is insufficient and has been tightening the noose on my team.

permalink
report
parent
reply
3 points

This. Exactly.

permalink
report
parent
reply
28 points

TSMC are probably making more money than anyone in this goldrush by selling the shovels and picks, so if that’s their opinion, I feel people should listen…

There’s little in the AI business plan other than hurling money at it and hoping job losses ensue.

permalink
report
parent
reply
9 points

TSMC doesn’t really have official opinions, they take silicon orders for money and shrug happily. Being neutral is good for business.

Altman’s scheme is just a whole other level of crazy though.

permalink
report
parent
reply
19 points

Seriously, I’d love to be enthusiastic about it because it’s genuinely cool what you can do with math.

But the lies that are shoved in our faces are just so fucking much and so fucking egregious that it’s pretty much impossible.

And on top of that LLMs are hugely overshadowing actual interesting approaches for funding.

permalink
report
parent
reply
17 points

I really want to like AI, I’d love to have an intelligent AI assistant or something, but I just struggle to find any uses for it outside of some really niche cases or for basic brainstorming tasks. Otherwise, it just feels like alot of work for very little benefit or results that I can’t even trust or use.

permalink
report
parent
reply
9 points
*

It’s useful.

I keep Qwen 32B loaded on my desktop pretty much whenever its on, as an (unreliable) assistant to analyze or parse big texts, to do quick chores or write scripts, to bounce ideas off of or even as a offline replacement for google translate (though I specifically use aya 32B for that).

It does “feel” different when the LLM is local, as you can manipulate the prompt syntax so easily, hammer it with multiple requests that come back really fast when it seems to get something wrong, not worry about refusals or data leakage and such.

permalink
report
parent
reply
3 points

Attractive. You got some pretty solid specs?

Rue the day I cheaped out on RAM. soldered RAMmmm

permalink
report
parent
reply
2 points
*

I receive alerts when people are outside my house, using security cameras, Blue Iris, CodeProject AI, Node-RED and Home Assistant, using a Google Coral for local AI. Entirely local - no cloud services apart from Google’s notification system to get notifications to my phone while I’m not home (which most Android apps use). That’s a good use case for AI since it avoids false positives that occur with regular motion detection.

permalink
report
parent
reply
1 point

I’ve been curious about google coral, but their memory is so tiny I’m not sure what kinds of models you can run on them

permalink
report
parent
reply
17 points

I think we should indict Sam Altman on two sets of charges:

  1. A set of securities fraud charges.

  2. 8 billion counts of criminal reckless endangerment.

He’s out on podcasts constantly saying the OpenAI is near superintelligent AGI and that there’s a good chance that they won’t be able to control it, and that human survival is at risk. How is gambling with human extinction not a massive act of planetary-scale criminal reckless endangerment?

So either he is putting the entire planet at risk, or he is lying through his teeth about how far along OpenAI is. If he’s telling the truth, he’s endangering us all. If he’s lying, then he’s committing securities fraud in an attempt to defraud shareholders. Either way, he should be in prison. I say we indict him for both simultaneously and let the courts sort it out.

permalink
report
parent
reply
4 points

“When you’re rich, they let you do it.”

permalink
report
parent
reply
14 points

The saddest part is, this is going to cause yet another AI winter. The first few ones were caused by genuine over-enthusiasm but this one is purely fuelled by greed.

permalink
report
parent
reply
8 points

The AI ecosystem is flooded, we need a good bubble pop to slow down the massive waste of resources that our current info-remix-based-on-what-you-will-likely-react-positively-to shit-tier AI represents.

permalink
report
parent
reply
9 points

Agreed that’s why it’s so dangerous. These tech bros are going to do damage with their shitty products. It seems like it’s Altman’s goal, honestly.

permalink
report
parent
reply
7 points

He wants money/power, and he is getting it. The rest of the AI field will forever be haunted by his greed.

permalink
report
parent
reply
5 points

After getting my head around the basics of the way LLMs work I thought “people rely on this for information?”, the model seems ok for tasks like summarisation though

permalink
report
parent
reply
9 points

I don’t love it for summarization. If I read a summary, my takeaway may be inaccurate.

Brainstorming is incredible. And revision suggestions. And drafting tedious responses, reformatting, parsing.

In all cases, nothing gets attributed to me unless I read every word and am in a position to verify the output. And I internalize nothing directly, besides philosophy or something. Sure can be an amazing starting point especially compared to a blank page.

permalink
report
parent
reply
1 point

the model seems ok for tasks like summarisation though

That and retrieval and the business use cases so far, but even then only if the results can be wrong somewhat frequently.

permalink
report
parent
reply
1 point
*

It’s good for coding if you train it on your own code base. Not great for writing very complex code since the models tend to hallucinate, but it’s great for common patterns, and straightforward questions specific to your code base that can be answered based on existing code (eg “how do I load a user’s most recent order given their email address?”)

permalink
report
parent
reply
3 points

It’s wild when you only know how to use SELECT in SQL, but after a dollar worth of prompting and 10 minutes of your time, you can have a significantly complex query you end up using multiple times a week.

permalink
report
parent
reply
3 points

TSMC’s allegedly calling Sam Altman a ‘podcast bro’ is spot on, and I’d add “manipulative vampire” to that.

What’s the source for that? It sounds hilarious

permalink
report
parent
reply
13 points

https://web.archive.org/web/20240930204245/https://www.nytimes.com/2024/09/25/business/openai-plan-electricity.html

When Mr. Altman visited TSMC’s headquarters in Taiwan shortly after he started his fund-raising effort, he told its executives that it would take $7 trillion and many years to build 36 semiconductor plants and additional data centers to fulfill his vision, two people briefed on the conversation said. It was his first visit to one of the multibillion-dollar plants.

TSMC’s executives found the idea so absurd that they took to calling Mr. Altman a “podcasting bro,” one of these people said. Adding just a few more chip-making plants, much less 36, was incredibly risky because of the money involved.

permalink
report
parent
reply
2 points

It’s selling the future, but nobody knows if we can actually get there

permalink
report
parent
reply
6 points

It’s selling an anticompetitive dystopia. It’s selling a Facebook monopoly vs selling the Fediverse.

We dont need 7 trillion dollars of datacenters burning the Earth, we need collaborative, open source innovation.

permalink
report
parent
reply
3 points

The first part is true … no one cares about the second part of your statement.

permalink
report
parent
reply
-5 points
*

Ya, it’s like machine learning but better. That’s about it IMO.

Edit: As I have to spell it out: as opposed to (machine learning with) neural networks.

permalink
report
parent
reply
33 points

I mean… it is machine learning.

permalink
report
parent
reply
7 points

It’s also neural networks, and probably some other CS structures.

AI is a category, and even specific implementations tend to use multiple techniques.

permalink
report
parent
reply
8 points
Deleted by creator
permalink
report
parent
reply
1 point

It is. It’s that plus an important process for living organisms rather than just burning something.

permalink
report
parent
reply
-7 points

Yep the current iteration is. But should we cross the threshold to full AGI… that’s either gonna be awesome or world ending. Not sure which.

permalink
report
parent
reply
10 points
*

Current LLMs cannot be AGI, no matter how big they are. The fundamental architecture just isn’t right.

permalink
report
parent
reply
1 point

You’re absolutely right. LLMs are good at faking language and sometimes not even great at that. Not sure why I got downvoted but oh well. But AGI will be game changing if it happens.

permalink
report
parent
reply
3 points

Based on what I’ve witnessed so far, people will play with their AGI units for a bit and then put them down to continue scrolling memes.

Which means it is neither awesome, nor world-ending, but just boring/business as usual.

permalink
report
parent
reply
1 point

There are people way smarter than me that claim it will be a threshold and would likely grow exponentially after it’s crossed. I guess we won’t know for sure until it happens. I do agree most people get bored easily but if this thing is possible to think for itself without interaction it won’t matter if the humans get bored.

permalink
report
parent
reply
2 points

I know nothing about anything, but I unfoundedly believe we’re still very far away from the computing power required for that. I think we still underestimate the power of biological brains.

permalink
report
parent
reply
1 point

Very likely. But 4 years ago I would have said we weren’t close to what these LLMs can do now so who knows.

permalink
report
parent
reply
1 point

What makes you think there’s a threshold?

permalink
report
parent
reply
104 points

I had a professor in college that said when an AI problem is solved, it is no longer AI.

Computers do all sorts of things today that 30 years ago were the stuff of science fiction. Back then many of those things were considered to be in the realm of AI. Now they’re just tools we use without thinking about them.

I’m sitting here using gesture typing on my phone to enter these words. The computer is analyzing my motions and predicting what words I want to type based on a statistical likelihood of what comes next from the group of possible words that my gesture could be. This would have been the realm of AI once, but now it’s just the keyboard app on my phone.

permalink
report
reply
15 points

The approach of LLMs without some sort of symbolic reasoning layer aren’t actually able to hold a model of what their context is and their relationships. They predict the next token, but fall apart when you change the numbers in a problem or add some negation to the prompt.

Awesome for protein research, summarization, speech recognition, speech generation, deep fakes, spam creation, RAG document summary, brainstorming, content classification, etc. I don’t even think we’ve found all the patterns they’d be great at predicting.

There are tons of great uses, but just throwing more data, memory, compute, and power at transformers is likely to hit a wall without new models. All the AGI hype is a bit overblown. That’s not from me that’s Noam Chomsky https://youtu.be/axuGfh4UR9Q?t=9271.

permalink
report
parent
reply
11 points

I’ve often thought LLMs could replace all of the C-suites and upper and middle management.

Funny how no companies push that as a possibility.

permalink
report
parent
reply
7 points

I almost expect that we’ll see some company reveal it has been letting an AI control the top level decision making for the business itself, including if and when to reveal the AI.

But the funny thing will be that all the executives and board members still have jobs and huge stock awards. They will all pat each other on the back for getting paid more money to do less work, by being bold and taking a risk to let the computer do half their job for them.

permalink
report
parent
reply
4 points

There’s a name for it the phenomenon: the AI effect.

permalink
report
parent
reply
73 points

Sounds about right. There are some valid and good use cases for “AI”, but the majority is just buzzword marketing.

permalink
report
reply
53 points

I have lots of uses for Attack Insects….

permalink
report
parent
reply
8 points

im down for arm improvements

permalink
report
parent
reply
2 points
Deleted by creator
permalink
report
parent
reply
72 points

Yup.

I don’t know why. The people marketing it have absolutely no understanding of what they’re selling.

Best part is that I get paid if it works as they expect it to and I get paid if I have to decommission or replace it. I’m not the one developing the AI that they’re wasting money on, they just demanded I use it.

That’s true software engineering folks. Decoupling doesn’t just make it easier to program and reuse, it saves your job when you need to retire something later too.

permalink
report
reply
36 points

Their goal isn’t to make AI.

The goal of both the VCs and the startups is to make money. That’s why.

permalink
report
parent
reply
11 points

It’s not even to make money, they already do that. They need GROWTH. More money this quarter than last or the stockholders don’t get paid.

permalink
report
parent
reply
16 points

Growth doesn’t mean revenue over cost anymore, it just means number go up. The easiest way to create growth from nothing is marketing tulips to venture capital and retail investors.

permalink
report
parent
reply
24 points

The people marketing it have absolutely no understanding of what they’re selling.

Has it ever been any different? Like, I’m not in tech, I build signs for a living, and the people selling our signs have no idea what they’re selling.

permalink
report
parent
reply
12 points
*

The worrying part is the implications of what they’re claiming to sell. They’re selling an imagined future in which there exists a class of sapient beings with no legal rights that corporations can freely enslave. How far that is from the reality of the tech doesn’t matter, it’s absolutely horrifying that this is something the ruling class wants enough to invest billions of dollars just for the chance of fantasizing about it.

permalink
report
parent
reply
72 points
*

I make DNNs (deep neural networks), the current trend in artificial intelligence modeling, for a living.

Much of my ancillary work consists of deflating tempering the C-suite’s hype and expectations of what “AI” solutions can solve or completely automate.

DNN algorithms can be powerful tools and muses in scientific endeavors, engineering, creativity and innovation. They aren’t full replacements for the power of the human mind.

I can safely say that many, if not most, of my peers in DNN programming and data science are humble in our approach to developing these systems for deployment.

If anything, studying this field has given me an even more profound respect for the billions of years of evolution required to display the power and subtleties of intelligence as we narrowly understand it in an anthropological, neuro-scientific, and/or historical framework(s).

permalink
report
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 17K

    Monthly active users

  • 12K

    Posts

  • 542K

    Comments