Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Roadmap (tentative) #12

Open
28 of 32 tasks
justheuristic opened this issue Jun 20, 2022 · 4 comments
Open
28 of 32 tasks

Roadmap (tentative) #12

justheuristic opened this issue Jun 20, 2022 · 4 comments

Comments

@justheuristic
Copy link
Collaborator

justheuristic commented Jun 20, 2022

Current tasks:

End of december: cover more use cases

End of july august: make it reliable, test with early adopters

End of june: build a proof-of-concept

  • agree on the user interface (see [DESIGN] user experience #5 (comment) )
  • run simple (but correct!) inference with a smaller model (for generation)
  • do simple (but correct!) forward/backward with frozen layers (for prompt tuning)
  • client can dynamically choose which remote servers to use for inference ( by: @justheuristic )
  • create basic correctness tests for later
  • check if 8-bit compression is remotely feasible ( by: @TimDettmers )
  • it's okay if the code is not super reliable for now
  • it's okay if servers have to be set up manually for now
  • begin investigating: quantized weights, quantized communication, automatic server allocation, "bloom points"

Important, but not urgent:

@justheuristic justheuristic pinned this issue Jun 20, 2022
@justheuristic
Copy link
Collaborator Author

justheuristic commented Aug 18, 2022

[moved inference of prompt-tuned model and priorities from summer to current tasks]

@bionicles
Copy link

hey, how hard would it be to extend petals to support training these models in addition to the fine tuning?

@mryab
Copy link
Member

mryab commented May 23, 2023

Hi @bionicles, Petals is a system designed specifically for inference of large models: however, it shares a lot of the underlying architecture with SWARM Parallelism (see https://github.com/yandex-research/swarm for a WIP implementation, which I hope to update in the coming weeks).

The short answer is "definitely possible", but please keep in mind that pretraining is out of scope for Petals. Hence, it might be more useful to continue the discussion elsewhere (e.g. to the SWARM repo or our Discord server) if you have specific questions or suggestions

@borzunov borzunov unpinned this issue May 23, 2023
@borzunov
Copy link
Collaborator

borzunov commented May 23, 2023

Hi @bionicles,

A small addition to the @mryab's response - while Petals does not support training from scratch, both Petals and SWARM are based on hivemind, our library for training over the Internet, which can be used for pre-training. Please see Q3 of the FAQ's "General" section for details.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants