ML OPs Flashcards

(70 cards)

1
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does MLOps stand for?

A

The way teams build, deploy, monitor and update machine-learning models in real life

MLOps encompasses everything around the training of machine-learning models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the real ML life cycle in MLOps?

A
  • Define business goal
  • Get data
  • Train models
  • Test models
  • Deploy model
  • Monitor model
  • Retrain / fix
  • Repeat

MLOps is about running this loop properly.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Why do we need MLOps?

A

To manage:
* Lots of data
* Lots of models
* Lots of users
* Lots of updates

Without MLOps, things can break very fast.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are some problems MLOps tries to stop?

A
  • Model suddenly becomes inaccurate
  • Data changes silently
  • Wrong model version gets deployed
  • No one knows which model is in production
  • Retraining is manual and slow

These issues highlight the importance of MLOps in maintaining model reliability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What must you have before training in MLOps?

A
  • A clear business objective
  • Success metrics
  • Acceptance criteria

No MLOps magic can happen without these foundational elements.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does model training & experiments include?

A
  • Running experiments
  • Tuning hyperparameters
  • Comparing models
  • Tracking results
  • Storing models and versions

This phase is crucial for developing effective machine-learning models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is an important point regarding a trained model?

A

A trained model is not a product

You still need code around it, APIs, pipelines, scalability, and automation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does deployment in MLOps refer to?

A

Putting the model into the real system (web app, backend, mobile app, etc.)

Deployment is a critical step in making the model usable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What should you monitor after release of a model?

A
  • Prediction quality
  • Error rates
  • Data changes
  • Performance (speed, cost, failures)

Monitoring is a huge aspect of MLOps.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is a very important idea regarding retraining and updating?

A

Releasing a model is NOT the end

If monitoring shows data changed or accuracy dropped, you must go back to training.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the most important idea in MLOps?

A

It is a continuous cycle

The cycle includes training, deploying, monitoring, improving, and redeploying.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Who does MLOps involve?

A
  • Data scientists
  • Software engineers
  • Platform / cloud engineers
  • Product people

MLOps is a cross-functional discipline connecting ML work, software engineering, and operations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How does MLOps compare to DevOps?

A

MLOps = build and operate ML systems

DevOps focuses on building and operating software.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the ultra-short summary of MLOps?

A

MLOps is the discipline that makes machine-learning models usable and reliable in production

It covers training, testing, packaging, deployment, monitoring, retraining, teamwork, and tooling.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is an ML pipeline?

A

The full chain that turns raw data into a running prediction system

Not just a model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What are the first two steps in an ML pipeline?

A
  • Business understanding
  • Data collection

Business understanding involves defining the problem and success metrics.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is the purpose of data cleaning in an ML pipeline?

A

Fix missing values, errors, broken rows, weird formats

Ensures the data is usable for modeling.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What does feature engineering involve?

A

Turn raw data into useful inputs

Essential for improving model performance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What is data augmentation used for?

A

Create extra data (mainly for images / sound / text)

Helps improve model robustness.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is the role of data annotation in supervised learning?

A

Add labels

Necessary for training supervised models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

What is the purpose of feature selection?

A

Keep only the useful inputs

Reduces dimensionality and improves model efficiency.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What does feature scaling achieve?

A

Put numbers on comparable ranges

Important for algorithms sensitive to the scale of data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What is involved in the modeling step?

A

Train the model

This is where the actual learning occurs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
What is the purpose of **evaluation** in an ML pipeline?
Check how good it is (metrics, validation, tests) ## Footnote Ensures the model meets performance criteria.
26
What does **deployment** refer to in an ML pipeline?
Put the system into real use ## Footnote Involves making the model accessible for predictions.
27
What is the focus of the **monitoring** step?
Watch it in production: is it still accurate? ## Footnote Ensures ongoing performance and reliability.
28
What is the final output of an ML pipeline?
A real prediction system that the business actually uses ## Footnote This is the ultimate goal of the pipeline.
29
True or false: You deploy a model in an ML pipeline.
FALSE ## Footnote You deploy a pipeline, which includes data cleaning, feature processing, scaling, transformations, the model, and prediction logic.
30
Who primarily builds **ML pipelines**?
* MLOps engineers * Software engineers * Data scientists ## Footnote It’s system architecture work.
31
What is the difference between a **Jupyter notebook** and an **ML pipeline**?
A Jupyter notebook = experiment; An ML pipeline = production system ## Footnote This highlights the transition from experimentation to deployment.
32
What are the key components of an **ML pipeline**?
* Business goal * Data → cleaning → features * Train model * Evaluate * Deploy * Monitor * Improve ## Footnote This summarizes the entire process from start to finish.
33
What are the **four main steps** in the process described?
* Modelling * Evaluation * Deployment * Monitoring ## Footnote These steps outline the workflow for developing and maintaining a machine learning model.
34
In the **Modelling** step, what must you document?
* Algorithm used * Features used * Hyperparameters tried * Error limits accepted ## Footnote Documentation is crucial for reproducibility in model development.
35
What is the goal of the **Evaluation** step?
To determine if the model can predict unknown data well ## Footnote This is assessed by running the model on a test set and computing performance metrics.
36
True or false: A trained model is enough for deployment.
FALSE ## Footnote Deployment requires the model, preprocessing, feature steps, scaling, and prediction code.
37
What are the main goals of the **Deployment** step?
* Scalability * Stability * Automation ## Footnote This involves turning the model into a real product and often includes continuous delivery.
38
What two aspects should be monitored after deployment?
* Software performance * Model performance ## Footnote Monitoring ensures that both the software and the model continue to function effectively in real-world conditions.
39
What should you do if model performance drops?
* Go back to earlier pipeline steps * Retrain with new data * Adjust model/hyperparameters * Evaluate again * Redeploy ## Footnote This process ensures that the model remains valid and effective over time.
40
Fill in the blank: You deploy the ________, not just the model.
pipeline ## Footnote This emphasizes the importance of deploying the entire workflow for machine learning.
41
What is the **output** of the Modelling step?
* Model type * Features * Hyperparameters * Learned weights * Training data info (metadata) ## Footnote This output is essential for understanding the model's structure and performance.
42
What does monitoring model performance involve?
Comparing predicted values vs real outcomes ## Footnote This helps determine if the model is still valid in changing conditions.
43
What does **MLOps maturity** refer to?
How automated, reliable, and scalable your ML system is ## Footnote MLOps maturity indicates the effectiveness of managing machine learning operations.
44
What is **Level 0** in MLOps maturity?
* No MLOps * Everything is manual * Train manually * Test manually * Deploy manually * No monitoring ## Footnote Models are treated with uncertainty: 'We trained something… we hope it still works.'
45
What characterizes **Level 1** in MLOps maturity?
* DevOps only (no real MLOps) * Software has automation (builds, tests, releases) * Models treated like files * Little or no feedback on model performance ## Footnote Code is professional, but ML processes remain messy.
46
What advancements are present in **Level 2** of MLOps maturity?
* Automated model training * Model performance tracking * Better model management ## Footnote Deployment is still mostly manual, but you can determine which model is better and which experiment worked.
47
What features define **Level 3** in MLOps maturity?
* Automated deployment * Proper train/test/production environments * A/B testing between models * Full traceability ## Footnote The system can safely decide which model to deploy.
48
What does **Level 4** in MLOps maturity entail?
* Continuous monitoring * Production performance feeds back automatically * Automatic retraining triggers * Models compared against live metrics * Zero-downtime releases ## Footnote The system improves itself autonomously.
49
What does the **tiny table** summarize about MLOps maturity levels?
* Level 0: everything manual * Level 1: software automated, ML not * Level 2: training & tracking automated * Level 3: deployment automated * Level 4: monitoring + retraining automated ## Footnote This table provides an easy reference for understanding MLOps maturity.
50
What roles are necessary to move up the MLOps maturity levels?
* Data scientists * Data engineers * Software engineers * MLOps/platform people ## Footnote Teams must collaborate and stop being siloed to improve MLOps maturity.
51
What changes occur as you mature in MLOps?
* Manual runs → automated pipelines * Manual releases → CI/CD * No tracking → full version control * No monitoring → performance feedback * Manual retraining → triggered retraining ## Footnote These changes enhance the efficiency and reliability of ML operations.
52
True or false: A successful ML project is only about technology.
FALSE ## Footnote Management support, access to data, easy deployment paths, and time for maintenance are also crucial.
53
What should you always ask regarding MLOps projects?
'What MLOps level is this project actually at?' ## Footnote This question helps identify the current maturity level and plan for improvements.
54
What is the **ultra-short summary** of MLOps maturity?
MLOps maturity = how automated and reliable your ML lifecycle is ## Footnote It emphasizes the transition from manual experiments to self-monitoring, self-retraining systems.
55
What does **CRISP-DM** stand for?
Cross-Industry Standard Process for Data Mining ## Footnote CRISP-DM is a checklist for data projects that prevents jumping straight to modeling.
56
List the **six steps** of the CRISP-DM workflow.
* Business understanding * Data understanding * Data preparation * Modelling * Evaluation * Deployment ## Footnote These steps guide the execution of a data/ML project.
57
What is the first step in the **CRISP-DM** workflow?
Business understanding ## Footnote This step involves defining the problem and success criteria.
58
What does the **Data understanding** step involve?
* Assessing available data * Evaluating data quality * Exploring data with visualizations ## Footnote Engaging with business stakeholders to clarify data meanings is also important.
59
What is typically the **longest step** in the CRISP-DM process?
Data preparation ## Footnote This step includes cleaning data, merging files, fixing errors, and feature engineering.
60
What are the key tasks in the **Modelling** step?
* Splitting data into train/validation/test * Trying different models * Tuning hyperparameters ## Footnote This step focuses on developing the predictive model.
61
What is evaluated in the **Evaluation** step?
* Model performance on unseen data * Achievement of business targets ## Footnote This step determines if the model meets the defined success criteria.
62
What does the **Deployment** step entail?
* Deploying the entire pipeline * Documenting the process * Planning for monitoring and maintenance ## Footnote This step ensures the model is operational and maintained.
63
True or false: In the CRISP-DM workflow, you move in a straight line from one step to the next.
FALSE ## Footnote The process involves going back and forth between steps as needed.
64
What is the main advantage of **cloud computing** for ML projects?
Allows renting computers (including GPUs) instead of owning them ## Footnote This makes machine learning more affordable.
65
List the **three service types** in cloud computing.
* IaaS – Infrastructure * PaaS – Platform * SaaS – Software ## Footnote These services provide varying levels of control and tools for ML projects.
66
Name three major **cloud platforms** mentioned.
* Google Cloud Platform (GCP) * Amazon Web Services (AWS) * Microsoft Azure ## Footnote These platforms offer various tools and services for machine learning.
67
What is a **container** in the context of cloud ML systems?
An environment where the app, model, and libraries run together ## Footnote This ensures reliable deployment.
68
What is a **reality check** regarding cloud platforms?
* They are powerful * Often complex * May be overkill for small projects ## Footnote Simpler platforms are often better for learning or demos.
69
Summarize the **CRISP-DM** workflow in one line.
A simple project workflow that helps plan, organize, and iterate properly ## Footnote It is essential for managing data and ML projects effectively.
70
Summarize the role of **cloud computing** in ML.
Enables running ML on rented machines and GPUs, making it accessible and scalable ## Footnote This has transformed the affordability of machine learning.