What is heuristics and why are they used?
Shortcuts to solutions. Rules-based approaches. An approximate solution instead of an exact solution.
They are used because some problems either can’t be solved or require too much time or processing power to be reasonable for solving the problem at hand., e.g. don’t want a robot taking forever to assess the best move in chess
Nearest neighbour heuristic –> Ask the computer to figure out the closest city that has’t been visited yet by the salesperson and make that the next stop. Doesn’t consider future moves so it’s not the most effective.
Alpha-beta pruning (games) –> runs through many different possible next moves until a move is determined to be worse than a previously considered move
What is AI?
A broad category that involves teaching computers to solve problems using algorithms. They mimic human intelligence.
This can be done using a set of complex rules processing or training machine learning models.
What are the challenges with ML
What is ML?
Machine learning is about training a machine (set of mathematical models) with a historical dataset such that the machine can predict the unseen data.
The key part of machine learning systems is that its performance can be improved based on the new data set (experience).
What is deep learning?
Deep learning problems form the subset of machine learning problems.
Deep learning represents the aspect of machine learning which mimics human brain for learning from data set and predicting the outcome on unseen data set.
In deep learning models. the features are learnt automatically based on the optimisation algorithm.
When should one use deep learning?
How do you identify whether a problem is a machine learning one?
What are the different kinds of ML problems?
Three most common…
Supervised learning problems can be categorized into the following different types:
- Regression (Predict the numerical value given the data set)
- Classification (Predict the class or the label of the dataset)
What is feature engineering and what role do product managers play?
Feature engineering is one of the key stages of the machine learning model development lifecycle. It can be defined as the process of identifying the most important features that can be used to train a machine learning model, which generalizes well for an unseen dataset (larger population). You need to clearly understand the concept of features
Feature engineering comprises of the following tasks:
As a product manager, you play a key role in helping data scientists identify raw features and derived features. Rest is the work of the data scienitist.
What is your approach towards model governance/monitoring?
Model performance can be classified into three categories, namely, the green zone, the yellow zone, and the red zone. One needs to identify thresholds for putting the model performance in the green, yellow, and red zones. Based on which zone model performance is found, the model is scheduled for retraining.
What is accuracy and how to best use it?
How do you use other metrics?
Accuracy = total correct predictions / total predictions
Measures the total misclassification done by the model.
Can be misleading if the dataset is unbalanced (majority comprised of one of the labels). Ok to use if the dataset is balanced.
We use precision when we want the prediction of 1 to be as correct as possible and we use recall when we want our model to spot as many real 1 as possible.
Why is testing ML projects challenging?
How do you handle data science uncertainty when planning?
How do you help the team manage and prioritise BAU model improvements?
What is precision and how to best use it?
Give an example.
Precision = true positive / total predicted positive (true and false positive)
Preicion is a good measure when the costs of getting a prediction wrong is much higher than cost of missing out on the right prediction, e.g. email spam detection. If precision is low, there are too many false positives and important email goes to the spam folder.
What is recall and how to best use it?
Give an example.
Recall = true positive / total actual positive (true positive and false negative)
The model should capture all examples of the class.
Best used when cost of missing a prediction is much higher than a wrong predction. e.g. when a bank is transaction is predicted as non-fradulent or sick patients. Or airport detectors missing any actual bombs/dangerous items. High coverage!
What is an F1 score and how to best use it?
F1-Score = 2 x (precision*recall)/(precision + recall) –> seeking a balance between precision and recall
Best used when the data is unbalanced.
What is the difference between supervised and unsupervised methods?
Can you provide some examples?
Supervised: Need labels, answers questions in pre-defined categories (e.g. email classification - classifying the email exactly)
Unsupervised: No need for labels, good for exploring, can visualise well (e.g. figuring out how many email classes there are/exploration)
Why is versioning important?
Model, data and code
What product metrics would you use for a chatbot?
User engagement: qualitative and quantitative
Performance: Model metrics and response ties
Stability: SLA, bugs
How would you design for feedback capture? Can you give an example?
(Does no feedback mean positive feedback?)
Ensuring that the design of the UI makes it very clear to the user what action they are taking next so that they don’t click without a clear intent.
I would make sure to user test this multiple times to ensure the mentality is engrained.
This gives data science teams reassurance on the other end.
Example: Reconciliations accept or decline. Decline made a difference. Had to stop people from progressing until they clicked somethig.