cancel
Showing results for 
Search instead for 
Did you mean: 

OutOfmemory exception when processing a large number of documents

Ahmad_Ben_Maall
Star Contributor
Star Contributor

Hello,

I want to process a large number of documents (around 3 million). The process consists of updating a metadata of each document from another metadata (so I will need to do a read and write from the database).

For that, I was thinking of using elasticsearch's SCROLL API.

The problem is that I will have a "java.lang.OutOfMemoryError: GC overhead limit exceeded" exception in the middle of processing (knowing that I have xmx = xms = 24g in JAVA_OPTS)

I have tried different configuration for the garbage collector but no great effect noticed.

Can someone help me or give me an idea how to process a large batch of documents in nuxeo.

Thank you in advance.

1 ACCEPTED ANSWER

Ahmad_Ben_Maall
Star Contributor
Star Contributor

Hello,

I have used the bulkAction mechanism in nuxeo and it looks good.

Thank you

View answer in original post

3 REPLIES 3

Ahmad_Ben_Maall
Star Contributor
Star Contributor

Hello,

I have used the bulkAction mechanism in nuxeo and it looks good.

Thank you

sujoy_debnath
Champ in-the-making
Champ in-the-making

Can you help me how to use bulk upload with custom metadata?

Hello,