You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The reindexation step fails when upgrading from TheHive 4.0.5. It seems to come from an unwanted character in one of my cases. The reindexation process seems to go on forever, I let it run for 4 hours with the thehive process showing no activity in htop.
I use the lucene index mechanism.
I tried to restart TheHive but it keeps trying to index everything and gets stuck at the same point.
Steps to Reproduce
Upgrade from version 4.0.5 to 4.1
Restart TheHive, wait until TheHive reaches the reindexation step
Watch the logs containing only "Reindex job is running" and htop showing 0.7% CPU usage for TheHive at best.
Possible Solutions
As I said, I think it is from a weird character in one of my cases (see logs below). It seems everything stops after the error occurs.
Complementary information
Mar 21 14:53:49 [DEBUG] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScanMetrics in Thread-12 [|] [457975760:doc-updates] Incremented by 1
Mar 21 14:53:49 [DEBUG] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScanMetrics in Thread-12 [|] [457975760:success-tx] Incremented by 1
Mar 21 14:53:49 [ERROR] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScannerExecutor in Thread-12 [|] Unexpected error processing data: {}
java.lang.IllegalArgumentException: Document contains at least one immense term in field="description_____s" (whose UTF8 encoding is longer than the max length 32766), all of which were skipped. Please correct the analyzer to not produce such terms. The prefix of the first immense term is: '[85, 115, 101, 114, 32, 76, 111, 103, 105, 110, 32, 83, 117, 99, 99, 101, 115, 115, 32, 40, 73, 80, 118, 54, 32, 77, 105, 99, 114, 111]...', original message: bytes can be at most 32766 in length; got 105149
at org.apache.lucene.index.DefaultIndexingChain$PerField.invert(DefaultIndexingChain.java:853)
at org.apache.lucene.index.DefaultIndexingChain.processField(DefaultIndexingChain.java:430)
at org.apache.lucene.index.DefaultIndexingChain.processDocument(DefaultIndexingChain.java:394)
at org.apache.lucene.index.DocumentsWriterPerThread.updateDocument(DocumentsWriterPerThread.java:251)
at org.apache.lucene.index.DocumentsWriter.updateDocument(DocumentsWriter.java:494)
at org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:1616)
at org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:1608)
at org.janusgraph.diskstorage.lucene.LuceneIndex.restore(LuceneIndex.java:305)
at org.janusgraph.diskstorage.indexing.IndexTransaction.restore(IndexTransaction.java:128)
at org.janusgraph.graphdb.olap.job.IndexRepairJob.workerIterationEnd(IndexRepairJob.java:201)
at org.janusgraph.graphdb.olap.VertexJobConverter.workerIterationEnd(VertexJobConverter.java:118)
at org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScannerExecutor$Processor.run(StandardScannerExecutor.java:285)
Caused by: org.apache.lucene.util.BytesRefHash$MaxBytesLengthExceededException: bytes can be at most 32766 in length; got 105149
at org.apache.lucene.util.BytesRefHash.add(BytesRefHash.java:265)
at org.apache.lucene.index.TermsHashPerField.add(TermsHashPerField.java:151)
at org.apache.lucene.index.DefaultIndexingChain$PerField.invert(DefaultIndexingChain.java:843)
... 11 common frames omitted
Mar 21 14:53:49 [DEBUG] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScanMetrics in Thread-12 [|] [457975760:success-tx] Incremented by 1
Mar 21 14:53:50 [DEBUG] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScanMetrics in application-akka.actor.default-dispatcher-15 [|] [457975760:adds] Returning zero by default (was null)
Mar 21 14:53:50 [INFO] from org.thp.scalligraph.models.Database in application-akka.actor.default-dispatcher-15 [|] Reindex job is running: 0 record(s) indexed
Mar 21 14:53:51 [DEBUG] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScanMetrics in application-akka.actor.default-dispatcher-15 [|] [457975760:adds] Returning zero by default (was null)
Mar 21 14:53:51 [INFO] from org.thp.scalligraph.models.Database in application-akka.actor.default-dispatcher-15 [|] Reindex job is running: 0 record(s) indexed
Mar 21 14:53:52 [DEBUG] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScanMetrics in application-akka.actor.default-dispatcher-15 [|] [457975760:adds] Returning zero by default (was null)
Mar 21 14:53:52 [INFO] from org.thp.scalligraph.models.Database in application-akka.actor.default-dispatcher-15 [|] Reindex job is running: 0 record(s) indexed
Mar 21 14:53:52 [DEBUG] from com.datastax.driver.core.RequestHandler in thp-timeouter-0 [|] onTimeout triggered but the response was completed by another thread, cancelling (retryCount = 0, queryState = QueryState(count
=0, inProgress=false, cancelled=false), queryStateRef = QueryState(count=0, inProgress=false, cancelled=false))
Mar 21 14:53:52 [DEBUG] from com.datastax.driver.core.RequestHandler in thp-timeouter-0 [|] onTimeout triggered but the response was completed by another thread, cancelling (retryCount = 0, queryState = QueryState(count
=0, inProgress=false, cancelled=false), queryStateRef = QueryState(count=0, inProgress=false, cancelled=false))
Mar 21 14:53:53 [DEBUG] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScanMetrics in application-akka.actor.default-dispatcher-15 [|] [457975760:adds] Returning zero by default (was null)
Mar 21 14:53:53 [INFO] from org.thp.scalligraph.models.Database in application-akka.actor.default-dispatcher-15 [|] Reindex job is running: 0 record(s) indexed
Mar 21 14:53:54 [DEBUG] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScanMetrics in application-akka.actor.default-dispatcher-15 [|] [457975760:adds] Returning zero by default (was null)
Mar 21 14:53:54 [INFO] from org.thp.scalligraph.models.Database in application-akka.actor.default-dispatcher-15 [|] Reindex job is running: 0 record(s) indexed
Mar 21 14:53:55 [DEBUG] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScanMetrics in application-akka.actor.default-dispatcher-15 [|] [457975760:adds] Returning zero by default (was null)
Mar 21 14:53:55 [INFO] from org.thp.scalligraph.models.Database in application-akka.actor.default-dispatcher-15 [|] Reindex job is running: 0 record(s) indexed
Mar 21 14:53:56 [DEBUG] from org.janusgraph.diskstorage.keycolumnvalue.scan.StandardScanMetrics in application-akka.actor.default-dispatcher-15 [|] [457975760:adds] Returning zero by default (was null)
Mar 21 14:53:56 [INFO] from org.thp.scalligraph.models.Database in application-akka.actor.default-dispatcher-15 [|] Reindex job is running: 0 record(s) indexed
The text was updated successfully, but these errors were encountered:
I would like to request prioritizing this as reverting thehive back to 4.0.5/ 4.0.4/ 4.0.3 still shows Reindex job is running: 0 record(s) indexed.
Removing index config from application.conf did not do any change. Indexing process still occurs and gets stuck.
In my case, my assumption is that long description fields on some alert types caused this issue.
Request Type
Bug
Work Environment
Problem Description
The reindexation step fails when upgrading from TheHive 4.0.5. It seems to come from an unwanted character in one of my cases. The reindexation process seems to go on forever, I let it run for 4 hours with the thehive process showing no activity in htop.
I use the lucene index mechanism.
I tried to restart TheHive but it keeps trying to index everything and gets stuck at the same point.
Steps to Reproduce
Possible Solutions
As I said, I think it is from a weird character in one of my cases (see logs below). It seems everything stops after the error occurs.
Complementary information
The text was updated successfully, but these errors were encountered: