Distillation
What is Model Distillation?
Reduce a whole LLM to a much smaller, scoped-down model
Distillation
Advantage to distilled models?
Smaller, faster, cheaper
Distillation
How fast are distilled models?
In Amazon Bedrock are up to 500% faster
Distillation
Accuracy loss with distilled models?
Less than 2% accuracy loss for use cases like RAG
Fine Tuning
What is Fine-Tuning?
Changes the weights of base foundation model based on input data you give it.
Fine Tuning
What format is input files to fine-tuning?
S3 JSON objects with keys like “prompt” and “completion”.
Fine Tuning
Starting requirement of a model to use fine-tuning?
Must use Provisioned Throughput
Fine Tuning
What is Instruction-based Fine Tuning?
Pass in prompts and expected completions
Fine Tuning
Category of data passed into instruction-based fine-tuning?
Labeled data: has the expected completion included
Fine Tuning
Why use instruction-based fine-tuning?
Domain-specific tasks where you supply the missing domain of data
Fine Tuning
What is Continued Pre-training?
Pass large amt of unlabeled data, model trains on the whole thing
Fine Tuning
What’s another name for Continued Pre-training?
Domain Adaptation Fine Tuning
Fine Tuning
Example of continued pre-training?
Give a LLM the whole AWS documentation catalog
Fine Tuning
What is Single-turn Messaging?
Type of instruction-based fine-tuning that mimics a chat bot
Fine Tuning
Format for input to Single-turn Messaging?
Supply “user” (human question) and “assistant” (chatbot response), JSON
Fine Tuning
What is Multi-turn Messaging?
Single-turn, but you supply lots of back-and-forth between human and chatbot
Fine Tuning
Which is cheaper, instruction-based or continued pre-training?
Instruction-based: data is targeted and usually smaller
Fine Tuning
Why is running a fine-tuned model is more expensive?
Have to use Provisioned Throughput
Fine Tuning
What is Transfer Learning?
start with LLM or pre-trained model adapt it to a new related task
Fine Tuning
how is transfer learning and fine-tuning related?
Fine Tuning is how you do it, Transfer Learning is the concept.