100% satisfaction guarantee Immediately available after payment Both online and in PDF No strings attached 4.6 TrustPilot
logo-home
Summary

Summary Exam notes of Machine Learning & Learning Algorithms (RSM Business Analytics & Management)

Rating
-
Sold
-
Pages
31
Uploaded on
21-01-2026
Written in
2025/2026

Summary of both book and lecture note with summarisation of key parameters of each machine learning model which can help buyer to prepare well for the final exams for machine learning related courses. I have received an excellent grade of 8.2 for the Machine Learning & Learning Algorithms exam with the help of the note.

Show more Read less
Institution
Course










Whoops! We can’t load your doc right now. Try again or contact support.

Written for

Institution
Study
Course

Document information

Uploaded on
January 21, 2026
Number of pages
31
Written in
2025/2026
Type
Summary

Subjects

Content preview

Function Approximation View of Statistical Learning
In supervised learning, the response follows the model:

Y = f (X) + ε,

where:

• f (X) is the true unknown function we want to approximate.

• ε is irreducible noise with mean zero.

Goal of statistical learning: Produce an estimator fˆ(X) that approximates f (X) well.


Loss Function
Purpose
A loss function quantifies how bad a prediction is. Choosing a loss function implicitly
defines the prediction function f (x) that minimizes expected loss.

L2 Loss (Squared Error Loss)


L2 (Y, f (X)) = (Y − f (X))2

Optimal predictor:
f (x) = M ean[Y | X = x]
Properties:

• Dominant in regression.

• Sensitive to large errors.

• Produces the conditional mean.

L1 Loss (Absolute Error Loss)


L1 (Y, f (X)) = |Y − f (X)|

Optimal predictor:
f (x) = M edian(Y | X = x)
Properties:

• More robust to outliers.

• Leads to the conditional median.




1

,Loss 0–1 (Classification Loss)


L0−1 (Y, f (X)) = 1[Y ̸= f (X)]

Optimal predictor:
f (x) = M ode(Y | X = x)
Properties:
• Used in classification.

• Minimizing 0–1 loss yields the Bayes classifier.


Bias Variance Trade-off
Under L2 loss, expected test error decomposes into:

E[(Y − fˆ(X))2 ] = Var[fˆ(X)] + (Bias[fˆ(X)])2 + Var[ε]
| {z } | {z } | {z } | {z }
T estError V ariance SquaredBias IrreducibleError


Bias
• Error that occurs when a model is too simple to capture the true patterns in the
data

• High bias: The model oversimplifies → misses patterns and underfits the
data.

• Low bias: The model captures patterns well and is closer to the true values
→ Overfits

Variance
• How much a model’s predictions change when it’s trained on different data.

• High variance: The model is too sensitive to small changes ⇒ overfitting

• Low variance: The model is more stable but might miss some patterns →
Underfits

Reducible Error


E[(f (X) − fˆ(X))2 ] = Var[fˆ(X)] + (Bias[fˆ(X)])2
| {z } | {z } | {z }
ReducibleError V ariance SquaredBias




• Origin: Inability to perfectly estimate the true function f(X)

• Reducible error = bias² + variance

• Reason: Use an approximation (a model) instead of the true function

2

, • How to reduce bias:
– Use more complex models
– Using more relevant features
– Reduce regularization to allow the model more flexibility in fitting
• How to reduce variance:
– Simplify the model
– Increase training data
– Apply regularization to constrain model complexity
– Use ensemble methods

Irreducible Error
• Origin: Random noise term ε
• Even if we know the true function f (X) → ε still cause variability in Y
• Reason: ε is independent of X

Trade-off
• Increasing model flexibility decreases bias but increases variance.
• Goal: choose a model complexity (λ, number of features, neighbors k, etc.) that
minimizes test error, not training error.


Training Error vs Test Error & Generalization Error
Training Error
• Computed on data used to fit the model.
• Typically underestimates true error.
• Flexible models can make training error nearly zero.

Test Error
• Error on previously unseen data.
• Used to estimate real-world performance.

Generalization Error
• True population-level predictive error.
• Not observable directly.
• Test error is an estimate of generalization error.
Training error is not reliable because the model is optimized to minimize it.

3
$19.46
Get access to the full document:

100% satisfaction guarantee
Immediately available after payment
Both online and in PDF
No strings attached

Get to know the seller
Seller avatar
karenhuang920905

Also available in package deal

Get to know the seller

Seller avatar
karenhuang920905 Erasmus Universiteit Rotterdam
Follow You need to be logged in order to follow users or courses
Sold
4
Member since
2 year
Number of followers
0
Documents
6
Last sold
6 days ago

0.0

0 reviews

5
0
4
0
3
0
2
0
1
0

Recently viewed by you

Why students choose Stuvia

Created by fellow students, verified by reviews

Quality you can trust: written by students who passed their tests and reviewed by others who've used these notes.

Didn't get what you expected? Choose another document

No worries! You can instantly pick a different document that better fits what you're looking for.

Pay as you like, start learning right away

No subscription, no commitments. Pay the way you're used to via credit card and download your PDF document instantly.

Student with book image

“Bought, downloaded, and aced it. It really can be that simple.”

Alisha Student

Frequently asked questions