Has anyone been successful in grabbing tokenized fields from ES to go
in to Mahouts format?
You can point to an index using $MAHOUT_HOME/bin/mahout lucene.vector
but the possibility to have multiple indexes is pretty high in ES so
what is the best strategy for doing this? Can anyone give me a good
starting point to go from?
Has anyone been successful in grabbing tokenized fields from ES to go
in to Mahouts format?
You can point to an index using $MAHOUT_HOME/bin/mahoutlucene.vector
but the possibility to have multiple indexes is pretty high in ES so
what is the best strategy for doing this? Can anyone give me a good
starting point to go from?
Has anyone been successful in grabbing tokenized fields from ES to go
in to Mahouts format?
You can point to an index using $MAHOUT_HOME/bin/mahoutlucene.vector
but the possibility to have multiple indexes is pretty high in ES so
what is the best strategy for doing this? Can anyone give me a good
starting point to go from?
With the default settings you will find Lucene index which you can
open with Luke, Mahout etc. in
ES/data/elasticsearch/nodes/0/indices/mahout_index/0/index
If you do need multiple indices (multiple index shards), with shared
FS gateway defined you can open all indices in shared gateway location
(one by one and create Mahout vectors from each).
Has anyone been successful in grabbing tokenized fields from ES to go
in to Mahouts format?
You can point to an index using $MAHOUT_HOME/bin/mahoutlucene.vector
but the possibility to have multiple indexes is pretty high in ES so
what is the best strategy for doing this? Can anyone give me a good
starting point to go from?
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.