You are viewing a single thread.
View all comments
40 points

Large Language Models (such as GPT) and AI image generators.

I follow certain AI related post tags on Tumblr and sometimes I see people expressing pure hatred towards these tools, as they only see the AIs as content thieves.

permalink
report
reply
28 points

I don’t mind the tool itself if you use it as such. I do mind when people use its output as the final product. See: the lawyer who used chatgpt for a legal brief

permalink
report
parent
reply
5 points

The lawyer fuck up is what happens when someone doesn’t know or understand the limitations of a LLM.

If you want a GPT model tailored and specialized for a specific task, you have to train it with custom data, fine tune it and tweak the model’s parameters. You cannot do that from the ChatGPT web/app, you need a custom implementation coded in Python or some other language.

permalink
report
parent
reply
2 points
*
Deleted by creator
permalink
report
parent
reply
2 points

There are some uis that allow for fine tuning (assuming you have an extremely high end rig designed for ml). For example ChatGPT alternative and DALLE alternative.

permalink
report
parent
reply
1 point

I’m glad you understand my point. Chatgpt is not Google. It’s a language model that will give you something that looks like the thing you asked for it to provide. It can and will pull facts out of its recycle bin if it fits the cadence of what it expects the answer to look like.

permalink
report
parent
reply
-3 points

Why do you mind that?

permalink
report
parent
reply
2 points

Have you seen that legal brief?

permalink
report
parent
reply
2 points

Letting a language model do the work of thinking is like building a house and using a circular saw to put nails in. It will do it but you should not trust the results.

It is not Google. It can, will, and has made up facts as long as it fits the format expected

Not at the very least proof reading and fact checking the output is beyond lazy and a terrible use of a tool. Using it to create the end product instead of as a tool to use in creation of an end product are two very different things.

permalink
report
parent
reply

they only see the AIs as content thieves.

AI is a method of content theft, it takes other people’s work and pieces it together in a way that resembles other works, without any actual coherency.

I don’t like that it churns out slop that displaces actual content.

I also don’t like the way it’s sped up enshitification of google and news sites. I didn’t think it could get worse than pages of listicles written by disinterested journalists paid fuckall to churn out 10 a day, but now you have chatGPT churning out 100 completely useless articles a day.

permalink
report
parent
reply

content theft

abolish intellectual property

permalink
report
parent
reply
0 points
*

LLMs just automates and does faster certain things that a person could do on their own if they invested way more effort and time. If a human being takes people’s work and pieces it together in a way that resembles other works without using any LLM/AI or automation tool, is the final result content theft too?

I agree with the content enshitification, but I disagree about the coherency.

Usually, implementations like the ChatGPT web/app will generate different outputs for the same prompt/input. You can also ask it to tweak a previous output, make it shorter, more concise, exclude parts, etc. And if you’re making API calls through a script you can tweak parameters like the Temperature, Top P, Presence Penalty or Frequence Penaly, which affect things like the coherence, randomness or repetitiveness of the output.

There’s also fine tunning using embeddings, which can help training a model to fit one’s specific needs and expectations, but I haven’t got to try it yet.

permalink
report
parent
reply

I disagree about the coherency.

Coherency requires relating symbolic meanings. AI just uses statistical analysis.

Consider if you were locked in the national library of Thailand. You don’t speak Siamese, and any pictures or bilingual dictionaries were removed.

Given a thousand years, you could look at the patterns and produce text similar to what someone who writes Siamese would write, but there’s still no coherency because you cannot connect the meaning behind any of the words.

That doesn’t necessarily mean your outputs are useless though, someone who does read Siamese can have you generate outputs until you print out something they can infer a coherent thought from, but you’re fundamentally unable to be trained to do that yourself.

If a human being takes people’s work and pieces it together in a way that resembles other works without using any LLM/AI or automation tool, is the final result content theft too?

We’re getting into ethics territory. IP is a social construct and we live under capitalism, our model for determining what is and isn’t theft should be selected by what supports artists and consumers against capitalists.

permalink
report
parent
reply
3 points

If a human being takes people’s work and pieces it together in a way that resembles other works without using any LLM/AI or automation tool, is the final result content theft too?

Yes, obviously. Artists and writers can learn from others and can be inspired by other’s works, but they can’t use parts of those works. That is content theft. Imitating a style is fine, but you have to create something new. LLMs cannot create, only steal.

permalink
report
parent
reply
23 points
*

As an artist I think it’s a more complicated issue than a lot of people are making it out to be, and all the fearmongering some popular artists are promoting really doesn’t help.

permalink
report
parent
reply
2 points

I think it’s a more complicated issue than a lot of people are making it out to be

Agree.

Also. People are pissed that what they have taken years to master others can now get close to replicate with little effort and time.

I’ve just realized that although they call the AIs “content thieves”, what they really feel is that as AIs are able to replicate their skills quickly, it makes them feel their own merit diminished.

If an artist creates artwork inspired on some other artist eveyone’s cool; if an AI does the same, then it’s stolen work even if the generated image is a unique new one.

permalink
report
parent
reply
-6 points

Using AI is feeding bullshit into a bullshit generator that’s handing back a synthesis of stolen art.

It’s a bullshit tool for hacks and grifters to pretend they’re “artists” so they can exploit another avenue for the grind.

This is not a debate. I don’t give a shit what your excuses are. Shout at a wall for all I care.

permalink
report
parent
reply
8 points

You sound like you’ve already closed your mind to the discussion, but in case you’re actually still willing to healthily engage in the discussion here is a really good video about why calling people who utilize AI in their work “hacks and grifters” is a very narrow minded (and often factually incorrect) way of looking at AI utilization.

permalink
report
parent
reply
9 points
*

It’s not that I hate it, but like, chatGPT sucks.

There was this uber hype around it, then we started using it … and it just makes so many errors, it’s literally just generating more work. Scrapped it after less than a week. It’s modern snakeoil.

permalink
report
parent
reply
3 points

Bard is the same, I asked it questions about two of my favourite bands whom I know a lot about. It omitted facts and invented things that were not true!

permalink
report
parent
reply
3 points

We used it for code generation. But we ended up spending more time fixen and debugging the generated code than it would have taken us to just write it. Also it introduces the most annoying type of bugs. Like once it misspelled a property name, but only at one point in the code, got it right everywhere else.

permalink
report
parent
reply
1 point

That’s why, in the case of a GPT model you would feed it custom training data using something like LlamaIndex. I don’t know if there’s an API available for Bard, tho.

You’re wrong assuming that the free models that we have at our disposal are the only possible and best implementations of these LLMs.

permalink
report
parent
reply
2 points

What! I have the opposite experience.
Im a tabletop roleplaying gamemaster and it has helped me immensely with translations, formatting of text, compiling and keeping track of my players character backgrounds and even coming up with plots and scenes that are suited for each player.

permalink
report
parent
reply
2 points

What did you use it for? I helps me a lot with coding, scripting, translations, terminologies… Sometimes it makes mistakes, but other times it produces working code that accomplishes what I asked for.

In any case, ChatGPT is just a demo that uses the GPT-3.5 Turbo model. Many people is being misled assuming that the ChatGPT research preview is all that the model has to offer. You can also try the improved model GPT-4, but it’s not free.

If you really want to get its full potential you need a custom implementation in Python that works against the API and do things like fine tune the model, embeddings, feed it custom data or give it access to tools with LangChain.

Of course that’s not something easy to do, but don’t think that the ChatGPT web/app is GPT models’ full potential.

permalink
report
parent
reply
1 point

I have a feeling this one’s mostly operator error.

Or you vastly overestimated what it could do.

permalink
report
parent
reply
1 point
*

I have a feeling this one’s mostly operator error.

Once we found the issues, it was actually quite easy to tell the AI to fix them. But at this point you’re debugging generated code to imrpove your input for the code generator … and it just was faster to write the code by hand.

And yes, there was a vast overestimation of what it can do, especially by some managers that used to be coders and thought this would compensate for their lack of recent practical expirence. It didn’t … I had to fix it.

permalink
report
parent
reply
7 points

LLM is way overhyped. So if your boss bought into that hype you’re gonna have a certain amount of animosity towards it. I’m a developer and it can be helpful at times, but managers seem to think it can write software on its own.

It’s basically an iterative improvement over a search engine, but unlike a search engine it cuts off the people creating the content it’s scraping from any kind of revenue stream.

And yeah there’s some real problems with it stealing content. Which isn’t being addressed at all. And bringing up these issues tend to get treated like Luddites by those that have bought into the hype.

permalink
report
parent
reply
5 points

I wouldn’t say “hate”, to me it’s more… so what? They’re really bad at what they do, only impressive at first glance. Not bad for some brainstorming, but then you end up with a facsimile of what the actual result would be, and now have to use that as a guideline to create the result.

permalink
report
parent
reply
1 point

IMO they’re not bad, but they require a lot of tweaking and trial and error.

I’ve learnt some Python thanks to ChatGPT’s help. When I say “some” I mean that I was able to create a custom implementation that uses a web interface and custom tools. The more lI learnt, the less I needed ChatGPT, but I always require some more coding help.

However, these LLMs are not sentient super smart AIs.

permalink
report
parent
reply
1 point
*
Deleted by creator
permalink
report
parent
reply
1 point

I like them as non-profit tools for personal use, but the hatred is justified IMO because we’re already seeing people with writing jobs lose that job and get replaced by an LLM and an “editor” who is paid less than the writer was.

That’s capitalism in all of its glory. People never mattered to the ones who want to make money; they just want want to as much profit as possible with the minimal investment. Someone at work created a tool that turns a work day of painstating tasks into a 5 minute wait? Fire the people, keep the tool. You may call LLMs or AIs enablers, but it’s like hating baseball bats because some use them to crack open skulls instead of hitting baseballs.

Regarding the verification of AI-generated content, I just can say I agree, but it’s going to be hard to detect.

permalink
report
parent
reply

Asklemmy

!asklemmy@lemmy.ml

Create post

A loosely moderated place to ask open-ended questions

Search asklemmy 🔍

If your post meets the following criteria, it’s welcome here!

  1. Open-ended question
  2. Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
  3. Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
  4. Not ad nauseam inducing: please make sure it is a question that would be new to most members
  5. An actual topic of discussion

Looking for support?

Looking for a community?

Icon by @Double_A@discuss.tchncs.de

Community stats

  • 8K

    Monthly active users

  • 5.8K

    Posts

  • 316K

    Comments