I’ve seen a lot of sentiment around Lemmy that AI is “useless”. I think this tends to stem from the fact that AI has not delivered on, well, anything the capitalists that push it have promised it would. That is to say, it has failed to meaningfully replace workers with a less expensive solution - AI that actually attempts to replace people’s jobs are incredibly expensive (and environmentally irresponsible) and they simply lie and say it’s not. It’s subsidized by that sweet sweet VC capital so they can keep the lie up. And I say attempt because AI is truly horrible at actually replacing people. It’s going to make mistakes and while everybody’s been trying real hard to make it less wrong, it’s just never gonna be “smart” enough to not have a human reviewing its’ behavior. Then you’ve got AI being shoehorned into every little thing that really, REALLY doesn’t need it. I’d say that AI is useless.
But AIs have been very useful to me. For one thing, they’re much better at googling than I am. They save me time by summarizing articles to just give me the broad strokes, and I can decide whether I want to go into the details from there. They’re also good idea generators - I’ve used them in creative writing just to explore things like “how might this story go?” or “what are interesting ways to describe this?”. I never really use what comes out of them verbatim - whether image or text - but it’s a good way to explore and seeing things expressed in ways you never would’ve thought of (and also the juxtaposition of seeing it next to very obvious expressions) tends to push your mind into new directions.
Lastly, I don’t know if it’s just because there’s an abundance of Japanese language learning content online, but GPT 4o has been incredibly useful in learning Japanese. I can ask it things like “how would a native speaker express X?” And it would give me some good answers that even my Japanese teacher agreed with. It can also give some incredibly accurate breakdowns of grammar. I’ve tried with less popular languages like Filipino and it just isn’t the same, but as far as Japanese goes it’s like having a tutor on standby 24/7. In fact, that’s exactly how I’ve been using it - I have it grade my own translations and give feedback on what could’ve been said more naturally.
All this to say, AI when used as a tool, rather than a dystopic stand-in for a human, can be a very useful one. So, what are some use cases you guys have where AI actually is pretty useful?
Hey, I’ll be one of the first in line to suggest that our brains are not special, magical, impossible to create systems. We could probably approximate human-level ability with a few antagonistic models, an image processor, and (crucially) a simple body and locomotion routines (because I don’t believe human-level intelligence is possible without being able to directly interface with the world).
My thesis - from my first post in this thread - is that this one system, acting on it’s own, doing nothing but producing text, is not AI. It’s not intelligence, because it doesn’t know what it’s saying, it’s just spitting out (mathematically guided, syntactically-correct-looking, stolen-from-humans) random words.
Ok, let’s check the dictionary.
Artificial intelligence, noun, The capacity of computers or other machines to exhibit or simulate intelligent behaviour; the field of study concerned with this. In later use also: software used to perform tasks or produce output previously thought to require human intelligence, esp. by using machine learning to extrapolate from large collections of data. Also as a count noun: an instance of this type of software; a (notional) entity exhibiting such intelligence. Abbreviated AI.
So it would still be AI. Just not up to your standards. They really should make some level system, like the sae levels of automation.
Since we’re already consulting the Oxford Dictionary;
Intelligence - The faculty of understanding; intellect. Also as a count noun: a mental manifestation of this faculty, a capacity to understand.
Intelligent - Having a high degree or good measure of understanding; quick to understand; knowing, sagacious.
It’s great that they have a non-technical, linguist’s supposition of what “AI” is, but if something is going to meet the standard of “Artificial Intelligence”, I think it would first need to meet the definitions of “Artificial” (which is an easy test in this case) and “Intelligent” (see above).
I’m not talking about simulating intelligence, I’m talking about actually having it. In order to do that - as I said before - you need to be able to demonstrate understanding. LLMs do not understand things. They spit out random words, guided by a fancy algorithm. You can demonstrate this in real-time; Ask it a question, get an obviously wrong answer, then call it on it’s own response. It will generate an apology, then give you a new answer. You can do this infinity. It’s not even paying attention to itself, and you’re suggesting that it has an understanding of what it’s saying.
As to the definition you posted; Humans thinking they’re so special that only they can do certain tasks, then being proven wrong, does not make another entity (a computer, in this case) more intelligent. It only proves that the task didn’t require a human. This definition is based on a false equivalency (specifically: “if only a human can do something, it requires intelligence”). If this is the bar (which is set absurdly low), then computers achieved AI the first time a simple if/then statement was created (even though a human came up with the process, wrote the statement, and the process has no ability to adapt to new situations). You don’t need intelligence (again, requiring understanding) to follow logic gates (and if you do, then basic circuit boards are also AI, so congratulations, we’ve had AI since the first AND gate was created in 1924).
LLMs do not understand things. They spit out random words, guided by a fancy algorithm.
Yes, that’s why it’s called artificial. It’s not true intelligence, it’s not natural intelligence, it’s artificial, it’s not real. Artificial is a synonym for fake in this case. LLM are fake intelligence, and anyone with some real intelligence can see it’s fake. It’s one of the issues AI developers have. To make the fake better, it needs exponentially more energy and data, exactly because it doesn’t have understanding.
“if only a human can do something, it requires intelligence”). If this is the bar (which is set absurdly low)
That always reminds me of the troubles the park rangers had in securing garbage, because “There is considerable overlap between the intelligence of the smartest bears and the dumbest tourists.”
This definition is based on a false equivalency (specifically: “if only a human can do something, it requires intelligence”). It only proves that the task didn’t require a human. Humans thinking they’re so special
Hence why goal posts keep shifting. There are enough people that want to keep the special feeling. I’d say that self delusion is pretty human but LLMs can fake that pretty well too.