Skip to content

[Help Wanted] Tokenzier warning messages #2007

@kylesayrs

Description

@kylesayrs

Many users see a warning about tokenizer parallelism. However, this warning is somewhat hard to reproduce. The scope of this ticket is to determine the source of this warning and prevent it from happening when running LLM Compressor.

2025-11-08T20:32:27.729606+0000 | reset | INFO - Compression lifecycle reset
2025-11-08T20:32:27.729954+0000 | from_modifiers | INFO - Creating recipe from modifiers
2025-11-08T20:32:27.730019+0000 | infer_mapping_from_model | INFO - Unrecognized model architecture Gemma3ForConditionalGeneration. Falling back to default mappings
2025-11-08T20:32:27.730046+0000 | infer_norm_mapping_from_model | INFO - Unrecognized model architecture Gemma3ForConditionalGeneration. Falling back to default mappings
2025-11-08T20:32:27.815196+0000 | initialize | INFO - Compression lifecycle initialized for 2 modifiers
2025-11-08T20:32:27.815337+0000 | moe_calibration_context | INFO - Entering MoE calibration context
2025-11-08T20:32:27.815840+0000 | IndependentPipeline | INFO - Inferred `SequentialPipeline` for `SpinQuantModifier`
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
To disable this warning, you can either:
	- Avoid using `tokenizers` before the fork if possible
	- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...

Metadata

Metadata

Assignees

No one assigned

    Labels

    good first issueA good first issue for users wanting to contribute

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions