You are viewing a single thread.
View all comments View context
3 points

You really don’t understand how these models work and you should learn about them before you make statements about them.

Machine learning models are, almost by definition, non-deterministic.

permalink
report
parent
reply
-1 points

We know the input, we can set the model to save the weight in checkpoints during training and can view them any time, and we can see weights of the finished model, and we can see the code.

If what you said about LLMs being completely black box were true, we wouldn’t be able to reproduce models, and each model would be unique.

But we can control every step of the training process, and we can reproduce not just the finished model, but the model at every single step during training.

We created the math, we created the training sets, we created the code and we can see and modify the weights and any other property of the model.

What exactly do we not understand?

permalink
report
parent
reply
3 points

Look, I understand why you think this. I thought this too when I was first beginning to learn machine learning and data science. But I’ve now been working with machine learning models including neural networks for nearly a decade, and the truth is that is nearly impossible to track the path of an input to a given output in machine learning models other than regression-based models and decision tree-based models.

There is an entire field of data science devoted to explaining how these models arrive at their conclusions. It’s called “explainable AI” or “xAI”, and I have a few papers that I’ve published in exploring the utility of them. The basic explanation for how they work is that we run hundreds of thousands of different models and then do statistical analysis to estimate why the models arrived at their conclusion. It isn’t an exact science, however.

permalink
report
parent
reply
0 points

Again, we have the input, we have the math and code that make it work, we have the weights, we have everything.

Would it take a lot of time to backtrack and check why we got a given output to an input? Yes, maybe an inordinate amount of time. But it can be done. It’s only black box because nobody has the time (likely years to decades) to wade through the layers of a finished model to check every node and weight.

The whole thing at its core is mathematics. It’s a series of steps, that can be listed and reviewed each step of the way if we wanted. It’s just that if would take too much time.

If what you said were true, we couldn’t reproduce models. And since we can…

It isn’t an exact science, however.

So if math and computer science isn’t an exact science, what is?

permalink
report
parent
reply

Technology

!technology@lemmy.ml

Create post

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

Community stats

  • 2K

    Monthly active users

  • 2.7K

    Posts

  • 42K

    Comments

Community moderators