Skip to content

issues Search Results · repo:pytorch/torchchat language:Python

Filter by

366 results
 (155 ms)

366 results

inpytorch/torchchat (press backspace or delete to remove)

As of May 19th 2025, we are halting active development on torchchat. The original intent of torchchat was to both demonstrate how to run LLM inference using PyTorch and improve the performance and functionality ...
  • Jack-Khuu
  • 3
  • Opened 
    on May 20
  • #1543

🚀 The feature, motivation and pitch As a sequel to https://github.com/pytorch/torchchat/issues/1518 where we added an enum for tokenizer types to simplify TokenizerArgs __post_init__, we need to further ...
actionable
good first issue
triaged
  • zhenyan-zhang-meta
  • 3
  • Opened 
    on Apr 28
  • #1536

🚀 The feature, motivation and pitch Quantization is a technique used to reduce the speed, size, or memory requirements of a model and torchao is PyTorch s native quantization library for inference and ...
Quantization
triaged
  • Jack-Khuu
  • 6
  • Opened 
    on Mar 31
  • #1520

🚀 The feature, motivation and pitch We would like to make torchtune an optional dependency. The first step towards that is to avoid importing torchtune unless it is actively used. To make this migration ...
actionable
good first issue
torchtune
triaged
  • Jack-Khuu
  • 2
  • Opened 
    on Mar 31
  • #1519

🚀 The feature, motivation and pitch TokenizerArgs.__post_init__ has grown quite verbose/redundant and could use a bit of simplification https://github.com/pytorch/torchchat/blob/1384f7d3d7af0847d8364fe7b300a8b49f2213c2/torchchat/cli/builder.py#L244-L289 ...
actionable
good first issue
triaged
  • Jack-Khuu
  • 9
  • Opened 
    on Mar 31
  • #1518

🐛 Describe the bug I just follow the tutorial to export an mobile model: python3 torchchat.py export llama3.1 --quantize torchchat/quant_config/mobile.json --output-pte-path llama3.1.pte in the torchchat ...
ExecuTorch
Quantization
triaged
  • TheBetterSolution
  • 4
  • Opened 
    on Feb 27
  • #1506

🚀 The feature, motivation and pitch PT is deprecating CUDA 12.4 nightly builds: https://github.com/pytorch/test-infra/pull/6333 This repo will need to update CI and testing in response to this Alternatives ...
CI Infra
Cuda
triaged
  • Jack-Khuu
  • 2
  • Opened 
    on Feb 25
  • #1504

🚀 The feature, motivation and pitch Need work to enable: python3 torchchat.py export llama3.2-11B --output-aoti-package-path exportedModels/llama3_2_artifacts.pt2 Features to be added: - Run AOTI ...
Compile / AOTI
enhancement
triaged
  • larryliu0820
  • 1
  • Opened 
    on Feb 21
  • #1497

I see current torchchat serving provides basic serving function. I m wondering what the future plan for serving. What s the target of torchchat serve? Will it provide more optimized and high performance ...
enhancement
triaged
  • jenniew
  • 3
  • Opened 
    on Feb 8
  • #1491

🚀 The feature, motivation and pitch After an ExecuTorch model is exported to a pte, tokenization information must be passed in as an arg (-l # ) to the runner. This can be avoided by writing this information ...
actionable
enhancement
ExecuTorch
good first issue
triaged
  • Jack-Khuu
  • Opened 
    on Jan 30
  • #1484
Issue origami icon

Learn how you can use GitHub Issues to plan and track your work.

Save views for sprints, backlogs, teams, or releases. Rank, sort, and filter issues to suit the occasion. The possibilities are endless.Learn more about GitHub Issues
ProTip! 
Restrict your search to the title by using the in:title qualifier.
Issue origami icon

Learn how you can use GitHub Issues to plan and track your work.

Save views for sprints, backlogs, teams, or releases. Rank, sort, and filter issues to suit the occasion. The possibilities are endless.Learn more about GitHub Issues
ProTip! 
Press the
/
key to activate the search input again and adjust your query.
Issue search results · GitHub