LLM concepts Flashcards

(25 cards)

1
Q

Define LLM.

A

A Large Language Model is an AI model trained on vast text data to understand and generate human-like text.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

True or false: LLMs can perform zero-shot learning.

A

TRUE

Zero-shot learning allows LLMs to perform tasks without prior specific training on those tasks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What does tokenization refer to in LLMs?

A

The process of converting text into smaller units called tokens for processing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Fill in the blank: LLMs use _________ to predict the next word in a sequence.

A

probabilities

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is fine-tuning in the context of LLMs?

A

Adjusting a pre-trained model on a specific dataset to improve performance on a particular task.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Define transformer architecture.

A

A neural network design that uses self-attention mechanisms to process sequential data efficiently.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

True or false: LLMs require supervised learning for all tasks.

A

FALSE

LLMs can also learn from unsupervised and semi-supervised methods.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the purpose of self-attention in LLMs?

A

To weigh the importance of different words in a sentence relative to each other.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Fill in the blank: The _________ layer in LLMs helps to capture long-range dependencies.

A

attention

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does pre-training involve for LLMs?

A

Training on a large corpus of text to learn general language patterns before fine-tuning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Define context window.

A

The number of tokens the model considers at once when generating or understanding text.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

True or false: LLMs can generate creative content.

A

TRUE

They can produce poetry, stories, and other creative forms based on prompts.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the role of embeddings in LLMs?

A

To represent words or phrases as vectors in a continuous vector space for better processing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Fill in the blank: LLMs are often evaluated using _________ metrics.

A

perplexity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is transfer learning in LLMs?

A

Using knowledge gained from one task to improve performance on a different but related task.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Define prompt engineering.

A

The practice of designing inputs to elicit desired responses from LLMs.

17
Q

True or false: LLMs can understand contextual nuances in language.

A

TRUE

They can grasp subtleties like sarcasm or idioms based on training data.

18
Q

What is a decoder in LLMs?

A

A component that generates output sequences from the encoded representations of input data.

19
Q

Fill in the blank: _________ is a common challenge faced by LLMs regarding biased outputs.

20
Q

What does model size refer to in LLMs?

A

The number of parameters in the model, affecting its capacity and performance.

21
Q

Define data augmentation in LLM training.

A

Techniques used to increase the diversity of training data without collecting new data.

22
Q

True or false: LLMs can only process English text.

A

FALSE

LLMs can be trained on multiple languages and can understand various linguistic structures.

23
Q

What is inference in the context of LLMs?

A

The process of generating output from a trained model based on new input data.

24
Q

Fill in the blank: _________ is a method to reduce overfitting in LLMs.

A

Regularization

25
What is a **language model**?
A statistical model that predicts the likelihood of a sequence of words in a language.