-6 points
*
Removed by mod
permalink
report
reply
6 points
permalink
report
reply
22 points
*
Deleted by creator
permalink
report
reply
12 points

Unless you compress it heavily, a Blu-ray rip is in the 20 to 40GB range.

permalink
report
parent
reply
3 points
*
Deleted by creator
permalink
report
parent
reply
6 points

Blu-ray is yes.

A rip implies higher compression /changes.

A movie like Independence Day? Easy to do 7gb with good quality. Blurays is H264, We have x265 now. 4gb size? Those explosions gonna be chunky, but film is watchable.

permalink
report
parent
reply
18 points

700mB aXXo DVD rip then

permalink
report
parent
reply
3 points

pff just send star wars and male them confused

permalink
report
parent
reply
3 points

Or female them sensible

permalink
report
parent
reply
19 points

A link with a cracked Minecraft client and an ip to join to a small server to chill with that alien.

Technically you can probably send a bunch of links, like Wikipedia etc. He “just” needs to access to it, which may or may not breach the 4GB rule.

permalink
report
reply
4 points

Why do you assume the alien has access to the internet?

permalink
report
parent
reply
3 points

No reason at all. Of course probably the aliens won’t have access to the interwebz, but playing Minecraft with an alien sounded funny in my head.

permalink
report
parent
reply
8 points

Damn, what a cheapskate. A chance to play Minecraft with a friendly space alien and you can’t even pay for a legit copy. Probably going to give that alien a computer virus and doom us all. Don’t put this guy in charge.

permalink
report
parent
reply
20 points

Can agree with that, but I wanted to save the alien from having a Microsoft account.

permalink
report
parent
reply
4 points

You’ve got a point. On second thought, maybe a different game would be better.

permalink
report
parent
reply
2 points

Nah, they’re going to give the alien an Alienware !

permalink
report
parent
reply
3 points

It won’t breach the 4GB rule, but even the closest star to Sol would have at minimum latency of 4 years.

permalink
report
parent
reply
2 points

Yeah that’s true.

permalink
report
parent
reply
11 points

Probably an AI model that fits in that size. It might not be our best models, but it probably would be a lot more useful to aliens than whatever we’d decide to fit on 4GB.

They’d get mostly all the inner workings of our languages and how we do conversations and generally be able to answer basic questions about humanity.

permalink
report
reply
2 points

AIs do a lot of different things. What kind of AI do you mean?

permalink
report
parent
reply
8 points

stable diffusion trained off of gay furry porn

permalink
report
parent
reply
7 points

What ? “an AI model” is not a compression algorithm. Why give the aliens an AI trained with some wikipedia articles when you could just give them wikipedia.

permalink
report
parent
reply
0 points
*

Because an LLM is more than just data: it’s like a big network of how syllables and words go together based on some context. And that’s useful because language is how we communicate, how we connect ideas together, it’s how we share stories. It’s not just Wikipedia articles, it’s a database of relationships between words and concepts. It approximates how we think as humans.

Yes, AI is hella overhyped. Everyone wants to AI everything. But really for this particular situation, I think the model data would actually be the best precompiled database of knowledge we can possibly provide to learn about humans for the size.

No it’s not magic compression, but 4GB worth of parameters is still a lot. GPT4All has models just under 4GB. They’re not particularly impressive compared to OpenAI’s offerings, but I think you can extract a lot more practical information to do first contact out of a basic model than 4GB worth of Wikipedia. It’s extremely lossy compression, it’s never gonna spit out articles vebatim, it will hallucinate a ton of stuff.

If we had more space I’d send all the major AIs we have like Dall-E, LLaMa and GPT 4. Imagine you’re an alien, you’re presented with a keyboard and a monitor, and know nothing about us. You can use Dall-E to try random letters and words and see if the output makes sense. Maybe you find out what a cat, dog, bat, frog, apple looks like. You can then input those words in ChatGPT, and get context as to when those are used. What’s “a horse”? What’s “riding”? Put those into Dall-E, now you know what a “cat riding a horse” looks like. It can generate as many as you want, any combination. Eventually you can figure out how to ask ChatGPT if cats typically ride horses, cars, bycles, what do cats do.

Now imagine you’re a very advanced alien species that can easily process the model’s parameters. You’ve just downloaded the basics of humanity. They can map their language to our model’s parameters, and basically speak to us in our language, and translate our answers to theirs, and basically have a basic conversation.

permalink
report
parent
reply
3 points

Sorry chief you haven’t really explained why an AI model would be the best format.

It’s less dense than Wikipedia text. End of.

permalink
report
parent
reply

Asklemmy

!asklemmy@lemmy.ml

Create post

A loosely moderated place to ask open-ended questions

Search asklemmy 🔍

If your post meets the following criteria, it’s welcome here!

  1. Open-ended question
  2. Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
  3. Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
  4. Not ad nauseam inducing: please make sure it is a question that would be new to most members
  5. An actual topic of discussion

Looking for support?

Looking for a community?

Icon by @Double_A@discuss.tchncs.de

Community stats

  • 7.6K

    Monthly active users

  • 5.9K

    Posts

  • 319K

    Comments