The quality of search engines has gone down so much for technical questions.
I’m looking for a way to index sites like stack exchanges, reddit, quora, and research papers. Would this be possible to do this locally with metadata?
I mean, you can easily self host a meta-search engine like Searx, Searx-ng, Whoogle, etc. I run Searx-ng and it sends your queries to multiple engines and aggregates the results for you.
To host your own search engine, you’d need to crawl and index every site. It’s certainly doable, but it would take a lot of time /effort.
I agree. Selfhosting a true search engine is way too much work and infeasible for individuals. Meta search engines however are very feasible and a great option.
Solr is a great search engine. It won’t help you with the crawling, but if you manage to get the data into Solr you have a come far.
I wished so much that the installation of this would be easier. It’s such an amazing concept and idea
I’m glad you ask! I have been working on a peer-to-peer search engine. The goal is to index websites that are on IPFS. It’s an MVP but you can already try the demo and run your node to make your data searchable for the whole network (you would need to generate html files from your data and put it on ipfs first)
In addition to YaCy and the varieties of Searx (both of which perform better for me than any of the commercial search engines), it’s not even out of the question to do this yourself, if you’re willing to start with the most recent Common Crawl dump and do some spidering in between releases. I don’t recommend it, unless you want to learn for yourself why search engines often give such miserable results, but it’s possible.
However, that’s the issue, here. Can you self-host a search engine? Sure, if you want to maintain the storage to back it. That depends on how deep your pockets go…
https://docs.searxng.org/ is a meta-search engine but good one 👍
Even the main search engines don’t index the entire internet of content these days and their databases are truly massive already. Writing a basic web crawler to produce a search index isn’t all that hard (I used to do it as a programming exercise for applicants) but dealing with the volume of data of the entire internet and storing it to produce a worthwhile search engine however is just not feasible on home hardware, it would be TB’s at least. It wouldn’t just be a little worse it would dramatically worse unless you put substantial resources to it including enormous amounts of network bandwidth which would have your ISP questioning your “unlimited 1 gbps fibre” contract. It would probably take years to get decent and always be many months out of date at best.
Doesn’t seem practical to try to self host based on the need to download and index every single page of the internet its a truly massive scale problem.
Well Google does self-host their own search engine.
I know you’re getting downvoted but you’re also technically correct. Which I appreciate.
“Hey @SundarPichai , how do I set up one of those googley thingies of yours?”
Docker container pls
Great username.