What is machine learning?
It is the study of algorithms that improve their performance (understanding), at some task (learning algorithm) with some experience (data).
A spam filter based on Machine Learning techniques automatically learns words and phrases that are good ___________ of spam by detecting unusually __________ patterns of words.
A spam filter based on Machine Learning techniques automatically notices the new ___________ has become unusually frequent in spam flagged by the users, and it starts flagging them ____________ your intervention.
predictors; frequent; pattern; without
What are the different types of machine learning?
In supervised learning, the training data you feed to the algorithm includes the desired solutions, called labels. True or False?
True
Define classification.
Task to predict the class label.
Define regression.
Task to predict a target numeric value.
Classification and regression are the two main tasks of ________________ learning.
supervised
What are the most typical supervised learning algorithms?
1) k-nearest neighbor
2) Linear regression
3) Logistic regression
4) Support vector machines
5) Decision trees and random forests
6) Neural networks
In ______________ learning, the training data is unlabeled. The system tries to learn without a teacher.
unsupervised
What are the main tasks of unsupervised learning (4)?
1) Clustering: task to detect groups of similar data
2) Anomaly detection and novelty detection: task to detect if a new instance looks like a normal one or is likely to be an anomaly
3) Visualization and dimensionality reduction: task to visualize or simplify the data without losing too much information
4) Association rule learning: task to discover exciting relations between attributes
What are the most typical unsupervised learning algorithms (4)?
1) k-means
2) Hierarchical clustering
3) Principal component analysis
4) Kernel PCA
In semi-supervised learning, the training data is partially labeled, usually a lot of _____________ data and a little bit of ___________ data.
unlabeled; labeled
Most semi-supervised algorithms are combinations of unsupervised and supervised algorithms. For example, Deep Belief Networks (DBNs) are based on _____________ components called restricted Boltzmann machines (RBMs) stacked on top of one another. RBMs are trained sequentially in an _________________ manner, and then the whole system is fine-tuned using _____________ learning techniques.
unsupervised; unsupervised; supervised
In reinforcement learning, the learning system, called an _________, can observe the ________________, select and perform actions, and get ___________ in return (or penalties in the form of negative rewards). The agent must then learn by itself what is the best strategy, called __________, to get the most reward over time. The policy defines what __________ the agent should choose when in a given situation.
agent; environment; rewards; policy; action
Define ML model.
Models form the central concept in machine learning as they are what is being learned from the data to solve a given task. There is a considerable range of machine learning models to choose from.
What are the different types of models?
1) Geometric Models– constructed directly in instance space, using geometric concepts such as lines, planes and distances
2) Probabilistic Models– statistical models that capture the inherent uncertainty in data and incorporate it into their predictions, modeled by means of probability distributions
3) Logical Models– models of this type can be easily translated into rules that are understandable by humans
What are the stages of the machine learning workflow (6)?
1) Data collection
2) Data preparation
3) Choosing a learning algorithm
4) Training the model
5) Evaluating the model
6) Predictions
What are the 3 types of data?
1) Numeric (e.g., income, age)
2) Categorical (e.g., gender, nationality)
3) Ordinal (e.g., low/medium/high)
Data pre-processing is a process of cleaning the raw data, i.e., the data is collected in the real-world and is converted to a clean dataset. Whenever the data is gathered from different sources, it is collected in a raw format, and this data isn’t feasible for the analysis. True or False?
True
What are some pre-processing techniques?
1) Conversion of data
2) Scaling data
3) Missing values
4) Outliers detection
Why is using feature scaling important?
1) Scaling guarantees that all features are on a comparable scale and have comparable ranges: the magnitude of the features has an impact on many machine learning techniques; larger scale features may dominate the learning process and have an excessive impact on the outcomes.
2) Algorithm performance improvement:
when the features are scaled, several machine learning methods, including gradient descent-based algorithms, distance-based algorithms (such as k-nearest neighbors), and support vector machines, perform better or converge more quickly.
3) Preventing numerical instability: numerical instability can be prevented by avoiding significant scale disparities between features. Examples include distance calculations or matrix operations, where having features with radically differing scales can result in numerical overflow or underflow problems.
4) Scaling features ensure that each characteristic is considered equally during the learning process: without scaling, bigger scale features could dominate the learning, producing skewed outcomes.
Standardization makes all variables directly ______________ and ___________ handle ____________.
comparable; cannot; outliers
Range scaling requires a specific range of values, may ____________ data and ___________ are an issue.
compress; outliers
It is important to fit the scalers to the _____________ data only, not to the full dataset. Only then can you use them to transform the training set and the test set (and new data).
training