A purported leak of 2,500 pages of internal documentation from Google sheds light on how Search, the most powerful arbiter of the internet, operates.

The leaked documents touch on topics like what kind of data Google collects and uses, which sites Google elevates for sensitive topics like elections, how Google handles small websites, and more. Some information in the documents appears to be in conflict with public statements by Google representatives, according to Fishkin and King.

You are viewing a single thread.
View all comments View context
62 points

You mean hosting your own crawler/indexer? That doesn’t really sound like a thing you could do cost-effectively.

permalink
report
parent
reply
62 points

No problem we crowdsource the crawling torrent style.

We outsourced that to google for reasonnable performance reason. But they shit the bed so now there’s no choice but to do it ourselves.

permalink
report
parent
reply
11 points

ooh that might be an interesting app to run on veilid

permalink
report
parent
reply
1 point

What is that and how does it apply ?

permalink
report
parent
reply
19 points

Surprisingly, it’s very doable, requires basic technical knowledge and relatively minimal computing resources (runs in the background on your computer).

https://yacy.net/ Github

I have tampermonkey script that sends yacy to crawl any websites that I visit, and it’s keeping up relatively good index for personal use of the visited websites. Combine yacy with ~300gb of Kiwix databases, add searxng as a frontend and you have pretty strong self hosted search engine.

Of course you need to supplement your searches from other search engines, as yacy does not crawl the whole web, just what you tell it to.

I encourage anyone who’s even slightly interested on this stuff to try Yacy, it’s ancient piece of software, but it still works very well and is not an abandoned project yet!

I personally use Yacy mostly on private mode, but it does have the distributed network there as well.

permalink
report
parent
reply
7 points

Yeah, I guess the P2P component sort of solves part of the issue I was imagining by distributing indexes and crawling. I was thinking that people were trying to run all of Google on a raspberry pi at home.

permalink
report
parent
reply
5 points

This is interesting, have you had it index reddit? I’m just wondering how much storage space the database takes up.

permalink
report
parent
reply
3 points

Hi!

Great question! I don’t crawl reddit, but this applies to other large sites as well. reddit themselves they have at this very moment banned the ip range where I host my Yacy at (Hetzner). I just looked up from my index that I do have 257k pages indexed from reddit through teddit I used to run, this is from before reddit api-enshittification, going to delete those right now.

And the way how the crawling is done is you define crawling depth, which limits how much content is crawled from the site.

  • 0 crawling depth = only the page you send Yacy to, nothing more.
  • 1 crawling depth = all the links on the page you send Yacy to
  • 2 crawling depth = all links on the page you send Yacy to, and all links on the pages crawled…
  • 3 …
  • n …

… etc.

I have my tampermonkey scripts set to only crawling depth of 1 at the moment (Just set them to 2 actually, kinda curious how much more I will be crawling), I’ve manually crawled some local news sites as a curiosity at the beginning. And my database is currently relatively small, only around ~86.38 gigabytes according to Yacy. This stores aproximately 2.6 million documents in Yacy’s Solr.

Yacy has tons of options for crawling, so you can customize how much it crawls and even filter out overly large sites with maximum number of documents set when you send Yacy there.

Large picture of Yacy's interface for starting a crawl.

The tampermonkey script I’ve been talking about in these posts, it’s very simple script: https://github.com/JeremyRand/YaCyIndexerGreasemonkey

Hit me up if you guys have more questions! I’m by no means an expert on Yacy, but I will do my best to answer.

permalink
report
parent
reply
16 points

Right!

Before his company was able to block more of Microsoft’s own tracking scripts, DuckDuckGo CEO and founder Gabriel Weinberg explained in a Reddit reply why firms like his weren’t going the full DIY route:

“… [W]e source most of our traditional links and images privately from Bing … Really only two companies (Google and Microsoft) have a high-quality global web link index (because I believe it costs upwards of a billion dollars a year to do), and so literally every other global search engine needs to bootstrap with one or both of them to provide a mainstream search product. The same is true for maps btw – only the biggest companies can similarly afford to put satellites up and send ground cars to take streetview pictures of every neighborhood.”

Ars

permalink
report
parent
reply
16 points

Federated bookmarks?

permalink
report
parent
reply
54 points

Federated directories. We’re going back to Yahoo like it’s 1995

permalink
report
parent
reply
32 points

Webrings!!!

permalink
report
parent
reply
4 points

I loved Geocities!

permalink
report
parent
reply
1 point

Yahoo patiently plotting its return from Japan.

permalink
report
parent
reply
7 points

I’m so ready for something like this. I’ve cleaned up my bookmarks and been waiting for alternatives to search engines.

permalink
report
parent
reply
4 points
permalink
report
parent
reply
7 points

You could use Common Crawl, it’s run by a non profit

https://en.wikipedia.org/wiki/Common_Crawl

permalink
report
parent
reply
5 points

Look up the yacy repo in github

permalink
report
parent
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 18K

    Monthly active users

  • 11K

    Posts

  • 504K

    Comments