We have lots of users(30-200 million). Each user has N(30-100) attributes. The attribute can be of type integer, text, timestamp or counter. Schema is not defined.
We are looking for a database to which we can index and retrieve(GET API) users in the most efficient way(low response time). We also have the requirement to search for users but this is out of this topic. We also need to have strong consistent and high available data.
I already know that GET API provides consistent data, but I'm not sure if Elasticsearch is the right choice for this case. We have tried Cassandra but the read performance was low(high disk IOPS).
Hi @xtapodi Welcome to the community and Thanks for considering Elasticsearch.
Elasticsearch when configured correctly (hardware software and index strategy) is highly performant We would need to learn a little bit about what you're trying to do.
even 200 million records with the 100 attributes each is still a relatively small data set with respect to Elasticsearch.
I know you said no schema but there will be a schema whether you set it or not and then with proper queries you should be able to query across that entire data set with highly performant low latency queries.
I would suggest setting up the a node or a cluster or better yet just try a Elastic Cloud and do some testing.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.