Skip to content

OOM when performing inference on an extremely large document with a semantic_text field #116022

@maxhniebergall

Description

@maxhniebergall

Elasticsearch Version

8.15.2, 8.16, main

Installed Plugins

No response

Java Version

bundled

OS Version

linux

Problem Description

The addition of automatic chunking combined with very large documents has caused some ingest processes to crash, with OOM or quiet restarts of the ML node.

We are currently working on long-term solutions to this problem which will be released in an upcoming patch.

Steps to Reproduce

Use semantic_text field to ingest documents over 100k characters in length.

Logs (if relevant)

No response

Metadata

Metadata

Assignees

Labels

:mlMachine learning>bugTeam:MLMeta label for the ML team

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions