How to organize a search across 10 million documents?


Warning: count(): Parameter must be an array or an object that implements Countable in /home/styllloz/public_html/qa-theme/donut-theme/qa-donut-layer.php on line 274
0 like 0 dislike
38 views
Is there a website. In the MySQL DB is about 10 000 000 records.

Need to do full text search + what would the search was relevant.


Using the search engine does not help. Even at 1 million, the delay in results using ApacheSolr for a few seconds.

Tried to look at www.elasticsearch.org and generally bundles with NoSQL (HSearch), but solutions can give relevant results in a reasonable time is not found. Rather there are options, but would have to implement the logic of calculating the relevance for themselves, for all types of records, and I would like something ready.
by | 38 views

7 Answers

0 like 0 dislike
sphinxsearch.com
by
0 like 0 dislike
Thanks for the info. Arrange a test drive of the Sphinx, then it will become clear if it is suitable to us...
by
0 like 0 dislike
Up to 100 requests per second keeps the Sphinx.
by
0 like 0 dislike
In the framework of the question posed, Yes, Sphinx would likely be saved.
\r
But due to the fact that the application is in PHP MySQL with this volume might not see the need to add NoSQL. And as a result I would like to find the easiest way to combine NoSQL and search engine.
The ideal would be a bunch of NoSQL + Search so that the search engine has given a list of keys based on relevance and the papers then pull it from NoSQL.
by
0 like 0 dislike
Xapian for example
by
0 like 0 dislike
Yandex server.
by
0 like 0 dislike
We for this purpose used Lucene
by

Related questions

110,608 questions
257,186 answers
0 comments
27,900 users