New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Leverage Initial Prompts from ChatGPT #850
Comments
We are not using output from chatgpt. It is against their tos and we want to build a better bot. Thank you for brining this up. We can however, research bot output detectors and can compare our bot's output to human output as an eval. if you want to work on this lmk. |
Related to this issue: #792 |
@hooman650 Thanks for linking the paper. The analysis on p. 6 is interesting. We could think about including points a) - d) rephrased as instructions in our prompting-guide: |
Oh - @andreaskoepf you gave me a great idea. @Rallio67 - we could include these as part of our sparrow prompts actually and try to see if our answers are more informative. |
You are a kind and wise chatbot ... blah blah... When asked questions, you define the core concept of the question and give detailed answers step by stepp and then offer a summary at the end. ... Then you give exmaples like this. Same for points b-d. then we run some data through a qg/qa generator with this prompt and some background text. |
There are datasets being built based on prompts and responses of ChatGPT, for instance:
I thought this might be helpful and also collecting ChatGPT on some complex prompts (coding, puzzles) that are hard to generate might be very helpful?
The text was updated successfully, but these errors were encountered: