instant searching in petabyte of data

前端 未结 3 1219
心在旅途
心在旅途 2021-01-01 07:56

I need to search over petabyte of data in CSV formate files. After indexing using LUCENE, the size of the indexing file is doubler than the original file. Is it possible to

3条回答
  •  执念已碎
    2021-01-01 08:11

    If you want to avoid changing your implementation, you should decompose your lucene index into 10, 20 or even more indices and query them in parallel. It worked in my case (I created 8 indices), I had 80 GB of data, and I needed implement search which works on a developer machine (Intel Duo Core, 3GB RAM).

提交回复
热议问题