do you want to turn your Elasticsearch into a SPARQL endpoint?
This Elasticsearch plugin stores and retrieves RDF triples by using the
Apache Jena API:
Apache Jena http://jena.apache.org is a free and open source Java framework
for building semantic web and Linked Data applications. The framework is
composed of different APIs interacting together to process RDF data.
Each triple will be stored as a single document, and Jena API uses term
filter queries to match triples. Due to the restrictions of such an
architecture, do not expect good SPARQL performance.
Apache Jena http://jena.apache.org is a free and open source Java
framework for building semantic web and Linked Data applications. The
framework is composed of different APIs interacting together to process RDF
data.
Each triple will be stored as a single document, and Jena API uses term
filter queries to match triples. Due to the restrictions of such an
architecture, do not expect good SPARQL performance.
HI great news!!
nice work, as well as the plugin from Gazzarini on Solr!
I can't wait to seriously play a little with both...
Have you already done some tests on the triples? I mean: is it possible to
use the ES replication to provide triples to other nodes, query on multiple
nodes via SPARQL etc?
Apache Jena http://jena.apache.org is a free and open source Java
framework for building semantic web and Linked Data applications. The
framework is composed of different APIs interacting together to process RDF
data.
Each triple will be stored as a single document, and Jena API uses term
filter queries to match triples. Due to the restrictions of such an
architecture, do not expect good SPARQL performance.
Apache Jena http://jena.apache.org is a free and open source Java
framework for building semantic web and Linked Data applications. The
framework is composed of different APIs interacting together to process RDF
data.
Each triple will be stored as a single document, and Jena API uses term
filter queries to match triples. Due to the restrictions of such an
architecture, do not expect good SPARQL performance.
With this RDF Jena plugin, you can set up indices like any other index,
only the mapping is kind of strict:
That means, with replica settings, triples (like docs) are replicated to
other nodes and yes, they can be queried by SPARQL. I haven't tested it
much but it is easy now to build loaders who can lift some billions of
triples into ES and run SPARQL over it.
Being frankly, I am not a big fan of SPARQL implementations because SPARQL
is so cumbersome, slow, and difficult to use in RDF processing. I still
have to understand how Jena will exercise ES though. With the plugin, there
are several restrictions, like translating ORDER BY to ES sort clauses,
which is an open challenge.
With JSON-LD and ES being a document store, I am still after an alternative
method how to translate SPARQL queries to performant ES queries on ordinary
JSON docs, operating on documents which look more like a DESCRIBE result,
and contexts saved similar to templates in an internal index. This is my
favorite idea I have not dropped. I can now compare this better to a pure
Jena API plugin approach.
Jörg
On Sun, Dec 28, 2014 at 9:42 PM, Alfredo Serafini seralf@gmail.com wrote:
HI great news!!
nice work, as well as the plugin from Gazzarini on Solr!
I can't wait to seriously play a little with both...
Have you already done some tests on the triples? I mean: is it possible to
use the ES replication to provide triples to other nodes, query on multiple
nodes via SPARQL etc?
PS: have you some plan about what we discussed here:
Apache Jena http://jena.apache.org is a free and open source Java
framework for building semantic web and Linked Data applications. The
framework is composed of different APIs interacting together to process RDF
data.
Each triple will be stored as a single document, and Jena API uses term
filter queries to match triples. Due to the restrictions of such an
architecture, do not expect good SPARQL performance.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant logo are trademarks of the Apache Software Foundation in the United States and/or other countries.