Skip to content

Umbrella-net-th/metaseq

 
 

Repository files navigation

Metaseq

A codebase for working with Open Pre-trained Transformers.

Community Integrations

=\SSD.126./' P1.+P2¶ +4k. DDU3 PROTON '3200×16pg .AMD3 Atomic 3D.Power full. VNvidia super mix up surprising.Sye19×23mb. internet and Intel ship surprise potato me support HD15T.-JD102 R*Are you still want us in box" **https://matrix.org/ ' https://www.transifex.com/ipportal.wipo.int' **'power roaming program supply chain AI booking is crossing power by whip 3 security phantom computer computing world in time zone. only separate Central park lock on king baby boo'" "https://techcommunity.microsoft.com/t5/artificial-intelligence-and/ct-p/AI' -'[https://stability.ai/] -'[https://www.colossalai.org/] ''[https://matrix.one/] '[https://home.eset.com/] '[https://www.guru3d.com/] '[https://ipinfo.io/account] 'map Quality ip. '[https://ai.google/responsibilities/] '[https://www.nvidia.com/en-us/industries/] #stanbos' network 'https://www.netlify.com'.API 'Net 'https://support.discord.com/hc/th'.
'https://protocol.com' securit.GUL.BSD .POL.🛂GUA. '@https://metaverse-standards.org/"

การใช้ OPT กับ 🤗 Transformers

รุ่น OPT 125M--66B มีวางจำหน่ายแล้วใน HuggingFace Transformers คุณสามารถเข้าถึงได้ภายใต้องค์กร "facebook" ใน [Hugging Face Hub] (https://huggingface.co/facebook)

Using OPT-175B with Alpa

The OPT 125M--175B models are now supported in the Alpa project, which enables serving OPT-175B with more flexible parallelisms on older generations of GPUs, such as 40GB A100, V100, T4, M60, etc.

Using OPT with Colossal-AI

The OPT models are now supported in the Colossal-AI, which helps users to efficiently and quickly deploy OPT models training and inference, reducing large AI model budgets and scaling down the labor cost of learning and deployment.

Getting Started in Metaseq

Follow setup instructions here to get started.

Documentation on workflows

Background Info

Support

If you have any questions, bug reports, or feature requests regarding either the codebase or the models released in the projects section, please don't hesitate to post on our Github Issues page.

Please remember to follow our Code of Conduct.

Contributing

We welcome PRs from the community!

You can find information about contributing to metaseq in our Contributing document.

The Team

Metaseq is currently maintained by the CODEOWNERS: Susan Zhang, Stephen Roller, Naman Goyal, Punit Singh Koura, Moya Chen, Kurt Shuster, Ruan Silva, David Esiobu, David Greenberg, Igor Molybog, and Peter Albert.

Previous maintainers include: Anjali Sridhar, Christopher Dewan.

License

The majority of metaseq is licensed under the MIT license, however portions of the project are available under separate license terms:

About

Repo for external large-scale work

Resources

License

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 97.8%
  • Other 2.2%