Skip to content

awslabs/HLAT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

HLAT

Open source code for paper (HLAT: High-quality Large Language Model Pre-trained on AWS Trainium: https://arxiv.org/abs/2404.10630) and blog (https://aws.amazon.com/blogs/machine-learning/end-to-end-llm-training-on-instance-clusters-with-over-100-nodes-using-aws-trainium).

Security

See CONTRIBUTING for more information.

License

This project is licensed under the Apache-2.0 License.

Reference

If you found HLAT useful in your research or applications, please cite using the following BibTeX:

@software{HLAT,
  title = {HLAT: High-quality Large Language Model Pre-trained on AWS Trainium},
  author = {Haozheng Fan, Hao Zhou, Guangtai Huang, Parameswaran Raman, Xinwei Fu, Gaurav Gupta, Dhananjay Ram, Yida Wang, Jun Huan},
  url = {https://github.com/awslabs/HLAT/},
  year={2024}
}
@article{HLAT,
  title={HLAT: High-quality Large Language Model Pre-trained on AWS Trainium},
  author={Haozheng Fan, Hao Zhou, Guangtai Huang, Parameswaran Raman, Xinwei Fu, Gaurav Gupta, Dhananjay Ram, Yida Wang, Jun Huan},
  journal={arXiv preprint arXiv:2404.10630},
  year={2024}
}

About

No description, website, or topics provided.

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published