-
Notifications
You must be signed in to change notification settings - Fork 270
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The accumulated size of entities is "50,000,001" that exceeded the "50,000,000" limit set by "FEATURE_SECURE_PROCESSING" #487
Comments
I also noticed this problem at some point but thought of it as a dump problem I see a few references around like can you test the following? |
I already tried putting jdk.xml.totalEntitySizeLimit and totalEntitySizeLimit as indicated in elastic/stream2es#65 (comment) but I still get the same error. I edited the clean-install-run script like this: mvn -Djdk.xml.totalEntitySizeLimit=2147480000 I'm out of ideas. Please help. |
can you edit the |
Hi. Sorry I'm a little confused on where I should put it in the run script. This is how I did it: Is it correct? |
this is the line:
https://github.com/dbpedia/extraction-framework/blob/master/run#L45
not sure if the package is needed so maybe add both
`-Djdk.xml.totalEntitySizeLimit=2147480000 -DtotalEntitySizeLimit=214748000
0"`
|
I tried doing as you said but same error appears. I edited https://github.com/dbpedia/extraction-framework/blob/master/run#L45 like this: |
sorry again, let's make a final check try putting this here: https://github.com/dbpedia/extraction-framework/blob/master/dump/pom.xml#L64-L71 |
Still the same error when I put this: |
I put it in this line: Here's the snippet:
And it worked! Thank you so much! :) |
Great! can you do a final check to see which of the two arguments is the one needed? Thanks! |
Thanks guys! |
I'm trying to use the 'downloa.10000.properties' and then 'extraction.default.properties' I'm receiving the same error.
I've tried the solutions above, I set extraction-framework/dump/pom.xml's jvmArgs
No luck fixing it. I then tried modifying ../run's script
You'll notice I added "-Djdk.xml.totalEntitySizeLimit=0". According to the docs, 0 is supposed to set it to unlimited. I did try it with the limits you listed above, that didn't work either. |
Interesting, I just ran an import over 130 languages without a hitch.
|
Was it Java 7 or 8? I was using Java 8 and I was reading that this was added to Java in 8. I'm not testing with Java 7. |
Using Java 8 as well, is this still causing problems with you? |
@chile12 I'm still processing all of Wiki (2 days later), but it's working no issues. I switched to Java 7. |
Hi guys. Sorry for the late update. So I just confirmed that this argument is the correct one:
|
I ran into this problem with the Wikidata extractor, having these settings
It is possible to remove this limitation by setting the value to 0. The extractor runs fine now. |
Thanks for the update Roland, best to integrate this into all the POM files. |
ERROR: The accumulated size of entities is "50,000,001" that exceeded the "50,000,000" limit set by "FEATURE_SECURE_PROCESSING" Fixes several parsing issues raised over the previous 8 years. [1] https://www.mediawiki.org/wiki/Manual_talk:MWDumper#Exception_in_thread_.22main.22_java.lang.ArrayIndexOutOfBoundsException:_2048 [2] dbpedia/extraction-framework#487 (comment)
Can anyone tell me the final solution. I am still stuck in this and tried all the solutions but no luck. thanks |
Hi! I followed the step by step instruction but encountered this error after running
../clean-install-run
I used the enwiki-20161001-pages-articles.xml.bz2 and I know it's a huge file but how do I get away with this error?
I tried putting the -DentityExpansionLimit=2147480000 in the clean-install-run like this:
mvn -DentityExpansionLimit=2147480000 ...
but I still get the same error. Please help me.
The text was updated successfully, but these errors were encountered: