Skip to content

Latest commit

 

History

History
29 lines (25 loc) · 2.73 KB

Week-2_Quiz-1.md

File metadata and controls

29 lines (25 loc) · 2.73 KB

1. In the videos, we described using either supervised learning or a prompt-based development process to build a restaurant review sentiment classifier. Which of the following statements about prompt-based development is correct?

  • Prompt-based development is generally much faster than supervised learning.
  • Prompt-based development requires that you collect hundreds or thousands of labeled examples.
  • Prompt-based development requires that you collect hundreds or thousands of unlabeled examples (meaning reviews without a label B to say if it is positive or negative sentiment).
  • If you want to classify reviews as positive, neutral, or negative (3 possible outputs) there is no way to write a prompt to do so: An LLM can generate only 2 outputs.

2. What is a token in the context of a large language model (LLM)?

  • The part of the LLM output that has primarily symbolic rather than substantive value (as in, "the court issued a token fine", or "the LLM generated a token output").
  • A physical device or digital code to authenticate a user's identity.
  • A word or part of a word in either the input prompt or LLM output.
  • A unit of cryptocurrency (like bitcoin or other "crypto tokens") that you can use to pay for LLM services.

3. What are the major steps of the lifecycle of a Generative AI project?

  • Scope project → Build/improve system → Deploy and monitor → Internal evaluation
  • Scope project → Internal evaluation → Build/improve system → Deploy and monitor
  • Scope project → Internal evaluation → Deploy and monitor → Build/improve system
  • Scope project → Build/improve system → Internal evaluation → Deploy and monitor

4. You are building a customer service chatbot. Why is it important to monitor the performance of the system after it is deployed?

  • Because of the LLM's knowledge cutoff, we must continuously monitor the knowledge cutoff and update its knowledge frequently.
  • Every product should be monitored to track customer satisfaction -- this is good practice for all software.
  • This is false. So long as internal evaluation is done well, further monitoring is not necessary.
  • In case customers say something that causes the chatbot to respond in an unexpected way, monitoring lets you discover problems and fix them.

5. You are working on using an LLM to summarize research reports. Suppose an average report contains roughly 6,000 words. Approximately how many tokens would it take an LLM to process 6,000 input words? (Assume 1 token = 3/4 words, or equivalently, 1 word \approx 1.333 tokens).

  • 4,500 tokens (6000 * 3/4)
  • 6,000 tokens
  • 8,000 tokens (about 6000 * 1.333)
  • 14,000 tokens (about 6000 * 1.333 + the original 6000 words)