I’ve seen a lot of sentiment around Lemmy that AI is “useless”. I think this tends to stem from the fact that AI has not delivered on, well, anything the capitalists that push it have promised it would. That is to say, it has failed to meaningfully replace workers with a less expensive solution - AI that actually attempts to replace people’s jobs are incredibly expensive (and environmentally irresponsible) and they simply lie and say it’s not. It’s subsidized by that sweet sweet VC capital so they can keep the lie up. And I say attempt because AI is truly horrible at actually replacing people. It’s going to make mistakes and while everybody’s been trying real hard to make it less wrong, it’s just never gonna be “smart” enough to not have a human reviewing its’ behavior. Then you’ve got AI being shoehorned into every little thing that really, REALLY doesn’t need it. I’d say that AI is useless.
But AIs have been very useful to me. For one thing, they’re much better at googling than I am. They save me time by summarizing articles to just give me the broad strokes, and I can decide whether I want to go into the details from there. They’re also good idea generators - I’ve used them in creative writing just to explore things like “how might this story go?” or “what are interesting ways to describe this?”. I never really use what comes out of them verbatim - whether image or text - but it’s a good way to explore and seeing things expressed in ways you never would’ve thought of (and also the juxtaposition of seeing it next to very obvious expressions) tends to push your mind into new directions.
Lastly, I don’t know if it’s just because there’s an abundance of Japanese language learning content online, but GPT 4o has been incredibly useful in learning Japanese. I can ask it things like “how would a native speaker express X?” And it would give me some good answers that even my Japanese teacher agreed with. It can also give some incredibly accurate breakdowns of grammar. I’ve tried with less popular languages like Filipino and it just isn’t the same, but as far as Japanese goes it’s like having a tutor on standby 24/7. In fact, that’s exactly how I’ve been using it - I have it grade my own translations and give feedback on what could’ve been said more naturally.
All this to say, AI when used as a tool, rather than a dystopic stand-in for a human, can be a very useful one. So, what are some use cases you guys have where AI actually is pretty useful?
It is sometimes good at building SQL code examples, but almost always needs fine-tuning since it doesn’t know the schema specifics.
Having said that one time it gave me code that resulted in an error, then I went back to GPT and said “This code you gave me is giving this error, can you fix it?” and all it would do is say something like “Correct, that code is wrong and will give an error.”
I just pass the create table statements after the instructions. It does pretty good up to 2 or 3 tables, but it will start to make mistakes when things get complicated
On the plus side, it’ll generate tedious code very well - double checking it is less draining than writing it yourself. Especially because I make more typos than it does - I often use it to get a starting point, then write the business logic myself
I needed a simple script to combine jpegs into a pdf. I tried to make a python script but it’s been years since I’ve programmed anything and I was intermediate at best. My script was riddled with errors and would not run. I asked chatgpt to write me the script and the second or third attempt worked great. The first two only failed because my prompts were bad, I had never used chatgpt before.
I would generally say they’re great with anything you happy being 100% right 90% of the time.
There is one thing I would find genuinely useful that seems within its current capabilities. I’d like to be able to give an AI a summary of my current knowledge on a subject, along with a batch of papers or articles, and have it give me one or more of the following:
-
A summary of the papers omitting the stuff I already know
-
A summary of any prerequisite background info I don’t already know, but isn’t in the papers
-
A summary of all the points on which the papers are in agreement
-
A summary of any points where the papers are in contention.
It’s perfect for topics you have professional knowledge of but don’t have perfect recall for. It can bring forward the context you need to be refreshed on but you can fact check it because you are an expert in that field.
If you need boilerplate code for a project but don’t remember a specific library or built in function that tackles your problem, you can use AI to generate an example you can then fix to make it run the way you wanted.
Same thing with finding config examples for a program that isn’t well documented but you are familiar with.
Sorry all my examples are tech nerd stuff because I’m just another tech nerd on lemmy
On the inverse I’ve found it to be quite bad at that. I can generally count on the AI answer to be wrong, fundamentally.
Might depend on your industry. It’s garbage at g code.
It probably depends how many good examples it has to pull together from stack overflow etc. it’s usually fine writing python, JavaScript, or powershell but I’d say if you have any level of specific needs it will just hallucinate a fake module or library that is a couple words from your prompt put into a function name but it’s usually good enough for me to get started to either write my own code or gives me enough context that I can google what the actual module is and find some real documentation. Useful to subject matter experts if there is enough training data would be my new qualifier.