Upload a reliably large RDF file to openrdf Sesame Ontology Manager

I need to upload a very large ontology, presented as an N-triples (1gb) file, to the openrdf Sesame application. For this, I use the workbench interface. I know that this file is too large to load in a single request. To get around this, I split my files into 100 MB files. But I still get the error message from openrdf Sesame server:

HTTP ERROR 500

Problem accessing /openrdf-workbench/repositories/business/add. Reason:

    Unbuffered entity enclosing request can not be repeated.
Caused by:

org.apache.commons.httpclient.ProtocolException: Unbuffered entity enclosing request can not be repeated.
 at org.apache.commons.httpclient.methods.EntityEnclosingMethod.writeRequestBody(EntityEnclosingMethod.java:487)

Does anyone have a good knowledge of openrdf Sesame or another ontology manager that I could use for my task?

Thanks so much for your input.

TO.

+3
source share
3 answers

Sesame Workbench - , 100 . , Tomcat, Sesame, POST? "", . :

- , API Sesame. Sesame .

, Sesame, "" Sesame: ( , HTTP- ). Sesame, .

+3

. "" RDF ( 40 ), :

.

Tomcat, , . ( localhost tomcat - Jeen ), :

: JAXP00010001: , "64000" ; , JDK. [ 1, 1]

, , Entity Limit - tomcat Umbuffered entity.

ParseError AWS SQS Storm tomcat:

export JAVA_OPTS="${JAVA_OPTS} -Djdk.xml.entityExpansionLimit=0"

XML ( 64 000). "" RDF ( 40-800 ).

+1

I don’t know exactly what task you want to achieve, but you can check here for a list of scalable three stores with informal (mostly self-advertising) scalability results. In this case, Sesame only reports that it processes 70M applications (not so much ... could be the cause of your problems.)

-1
source

Source: https://habr.com/ru/post/1761068/


All Articles