Portion of training training data is labeled and feedback is provided in the form of rewards or penalties. What type of learning
Reinforcement learning
What are the two types of inferencing?
Batch and Real time
Deep learning is used in which use 2 cases?
Computer vision and NLP
What are FMs in generative AI?
Pretrained models
What are Transformer models?
Builds encoder decoder concept in genAI. They use self-attention to process input data. Self-attention allows the model to weigh the importance of different words in a sentence when encoding a particular word
Are FMs pre trained using reinforced learning? True or False?
False. FMs are typically pre-trained through self-supervised learning
Where is pre-text tasks used for?
In self-supervised learning
Self-supervised learning makes use of the structure within the data to autogenerate labels. True or False?
True
Optimization of pre trained FMs are done using what?
Prompt engineering,
Retrieval-augmented generation (RAG),
Fine-tuning on task-specific data
LLMs, Diffusion and Multiodel models are what?
These are FM models
These are numerical representations of tokens, where each token is assigned a vector (a list of numbers) that captures its meaning and relationships with other tokens?
Embeddings
What is a context window?
The maximum number of tokens a LLM model can take when generating text
What is a vector?
It is an array of numercial values
What is the process of vectorization?
Text -> [Tokenization]->Tokens -> [Embeddings Model] -> Vectors
What is the process of vectorization in Bedrock KBs using RAG?
Customer KB->[Upload in Amazon S3]->[Select a vector DB]->[Select a Model]->[Sync with customer KB]->Vectorization of Customer KB text
What is Watermark detection for Amazon Bedrock?
Identify images generated by Amazon Titan Image Generator, a foundation model that allows users to create realistic, studio-quality images in large volumes and at low cost, using natural language prompts
What is continued pretraining in Amazon Bedrock?
You provide unlabeled data to pre-train a model by familiarizing it with certain types of inputs
Which is the models which gradually add more and more meaningful information to this noise until they end up with a clear and coherent output, like an image or a piece of text?
Diffusion model
Which model has generator and discriminator?
Generative adversarial networks
Which model has encoders and decoders?
Varional autoencoders
What are the components of prompt engineering
Instructions, Context, Input data and Output indicator
What are non-determistic LLMs popularly called?
Generative Language Models
What is a supervised learning process that involves taking a pre-trained model and adding specific, smaller datasets?
Fine tuning
Two types of fine tuning
Instruction fine-tuning and Reinforcement learning from human feedback (RLHF)