Replies: 2 comments 9 replies
-
The approaches you suggest seem okay, I think OpenAI also used such approach on ChatGPT and it's kinda working for them. |
Beta Was this translation helpful? Give feedback.
-
Fine-tuned models can be costly 🤔 not only in training but also in usage. Have we explored the way of creating a proper prompt? For example, we can have this algorithm:
Sometimes it works pretty well, and GPT-3 is able to respond within the given format pretty steadily. If we can define item 4 (what's our ideal structure to generate the diagram from), I can try to experiment with items 2-3 |
Beta Was this translation helpful? Give feedback.
-
Context
MVP evolves around the completion model which requires a bulky input including a "train sample" and the prompt for prediction. The approach has many disadvantages:
Fine-tune of the model using a significant number of training samples is the way to address the bottlenecks.
Problem
The OpenAI documentation suggests few hundreds data points as a reasonable training sample size:
It imposes the bottleneck: how to generate such amount of quality data?
Proposed Solutions
Internally
Every contributor commits to generate few dozens of training data points.
Pros:
Cons:
Crowdsourcing
Involve community to generate data and share with us through PR and other coms channels.
Requirements:
Pros:
Cons:
Users input
It's an extension of the crowdsourcing approach leveraging the incentive to improve the product by its users.
Requirements:
Pros:
Cons:
Summary
I'd like to propose a combination of the "Internally" and the "Users input" approaches:
Beta Was this translation helpful? Give feedback.
All reactions