Fine Tuning Flashcards

(20 cards)

1
Q

Distillation

What is Model Distillation?

A

Reduce a whole LLM to a much smaller, scoped-down model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Distillation

Advantage to distilled models?

A

Smaller, faster, cheaper

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Distillation

How fast are distilled models?

A

In Amazon Bedrock are up to 500% faster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Distillation

Accuracy loss with distilled models?

A

Less than 2% accuracy loss for use cases like RAG

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Fine Tuning

What is Fine-Tuning?

A

Changes the weights of base foundation model based on input data you give it.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Fine Tuning

What format is input files to fine-tuning?

A

S3 JSON objects with keys like “prompt” and “completion”.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Fine Tuning

Starting requirement of a model to use fine-tuning?

A

Must use Provisioned Throughput

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Fine Tuning

What is Instruction-based Fine Tuning?

A

Pass in prompts and expected completions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Fine Tuning

Category of data passed into instruction-based fine-tuning?

A

Labeled data: has the expected completion included

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Fine Tuning

Why use instruction-based fine-tuning?

A

Domain-specific tasks where you supply the missing domain of data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Fine Tuning

What is Continued Pre-training?

A

Pass large amt of unlabeled data, model trains on the whole thing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Fine Tuning

What’s another name for Continued Pre-training?

A

Domain Adaptation Fine Tuning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Fine Tuning

Example of continued pre-training?

A

Give a LLM the whole AWS documentation catalog

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Fine Tuning

What is Single-turn Messaging?

A

Type of instruction-based fine-tuning that mimics a chat bot

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Fine Tuning

Format for input to Single-turn Messaging?

A

Supply “user” (human question) and “assistant” (chatbot response), JSON

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Fine Tuning

What is Multi-turn Messaging?

A

Single-turn, but you supply lots of back-and-forth between human and chatbot

17
Q

Fine Tuning

Which is cheaper, instruction-based or continued pre-training?

A

Instruction-based: data is targeted and usually smaller

18
Q

Fine Tuning

Why is running a fine-tuned model is more expensive?

A

Have to use Provisioned Throughput

19
Q

Fine Tuning

What is Transfer Learning?

A

start with LLM or pre-trained model adapt it to a new related task

20
Q

Fine Tuning

how is transfer learning and fine-tuning related?

A

Fine Tuning is how you do it, Transfer Learning is the concept.