Unless you compress it heavily, a Blu-ray rip is in the 20 to 40GB range.
A link with a cracked Minecraft client and an ip to join to a small server to chill with that alien.
Technically you can probably send a bunch of links, like Wikipedia etc. He “just” needs to access to it, which may or may not breach the 4GB rule.
Damn, what a cheapskate. A chance to play Minecraft with a friendly space alien and you can’t even pay for a legit copy. Probably going to give that alien a computer virus and doom us all. Don’t put this guy in charge.
Can agree with that, but I wanted to save the alien from having a Microsoft account.
You’ve got a point. On second thought, maybe a different game would be better.
It won’t breach the 4GB rule, but even the closest star to Sol would have at minimum latency of 4 years.
Probably an AI model that fits in that size. It might not be our best models, but it probably would be a lot more useful to aliens than whatever we’d decide to fit on 4GB.
They’d get mostly all the inner workings of our languages and how we do conversations and generally be able to answer basic questions about humanity.
What ? “an AI model” is not a compression algorithm. Why give the aliens an AI trained with some wikipedia articles when you could just give them wikipedia.
Because an LLM is more than just data: it’s like a big network of how syllables and words go together based on some context. And that’s useful because language is how we communicate, how we connect ideas together, it’s how we share stories. It’s not just Wikipedia articles, it’s a database of relationships between words and concepts. It approximates how we think as humans.
Yes, AI is hella overhyped. Everyone wants to AI everything. But really for this particular situation, I think the model data would actually be the best precompiled database of knowledge we can possibly provide to learn about humans for the size.
No it’s not magic compression, but 4GB worth of parameters is still a lot. GPT4All has models just under 4GB. They’re not particularly impressive compared to OpenAI’s offerings, but I think you can extract a lot more practical information to do first contact out of a basic model than 4GB worth of Wikipedia. It’s extremely lossy compression, it’s never gonna spit out articles vebatim, it will hallucinate a ton of stuff.
If we had more space I’d send all the major AIs we have like Dall-E, LLaMa and GPT 4. Imagine you’re an alien, you’re presented with a keyboard and a monitor, and know nothing about us. You can use Dall-E to try random letters and words and see if the output makes sense. Maybe you find out what a cat, dog, bat, frog, apple looks like. You can then input those words in ChatGPT, and get context as to when those are used. What’s “a horse”? What’s “riding”? Put those into Dall-E, now you know what a “cat riding a horse” looks like. It can generate as many as you want, any combination. Eventually you can figure out how to ask ChatGPT if cats typically ride horses, cars, bycles, what do cats do.
Now imagine you’re a very advanced alien species that can easily process the model’s parameters. You’ve just downloaded the basics of humanity. They can map their language to our model’s parameters, and basically speak to us in our language, and translate our answers to theirs, and basically have a basic conversation.
Sorry chief you haven’t really explained why an AI model would be the best format.
It’s less dense than Wikipedia text. End of.