Machine Learning is the science of getting computers to learn from data, rather than following rules a human explicitly wrote.
The key insight: Instead of a programmer writing 'if the email contains "free money" then mark it as spam', an ML model reads thousands of real spam emails, finds the patterns itself, and learns a far better filter than any human could write by hand.
Arthur Samuel's definition (1959): 'The field of study that gives computers the ability to learn without being explicitly programmed.' Still the clearest definition 65 years later.
You use it daily without knowing:
- Google Search โ ranking is learned from billions of clicks, not hand-tuned rules
- Netflix โ recommendations learned from what 260M subscribers watched next
- Email spam filter โ trained on millions of labelled spam/not-spam examples
- Google Photos โ recognises your face from a single example photo
- Voice assistants โ speech recognition improved from zero hand-written rules
Why this matters: ML systems improve automatically as more data arrives. Rule-based systems require manual updates for every new edge case. At scale, ML is the only approach that works.
Deepening Notes
Source-backed reinforcement: these points are extracted from the session source note to strengthen your theory intuition.
- And that works so well because their machine learning software has figured out how to rank web pages.
- Or when you upload pictures to Instagram or Snapchat and think to yourself, I want to tag my friends so they can see their pictures.
- Well these apps can recognize your friends in your pictures and label them as well.
- Or if you've just finished watching a Star Wars movie on the video streaming service and you think what other similar movies can I watch?
- Well the streaming service will likely use machine learning to recommend something that you might like.
Interview-Ready Deepening
Source-backed reinforcement: these points add detail beyond short-duration UI hints and emphasize production tradeoffs.
- Machine Learning is the science of getting computers to learn from data, rather than following rules a human explicitly wrote.
- That's machine learning, it's the science of getting computers to learn without being explicitly programmed.
- That too is an application of machine learning.
- Arthur Samuel's definition (1959): 'The field of study that gives computers the ability to learn without being explicitly programmed.' Still the clearest definition 65 years later.
- Why this matters: ML systems improve automatically as more data arrives.
- Google Search โ ranking is learned from billions of clicks, not hand-tuned rules
- Netflix โ recommendations learned from what 260M subscribers watched next
- Email spam filter โ trained on millions of labelled spam/not-spam examples
Tradeoffs You Should Be Able to Explain
- More expressive models improve fit but can reduce interpretability and raise overfitting risk.
- Higher optimization speed can reduce training time but may increase instability if learning dynamics are not monitored.
- Feature-rich pipelines improve performance ceilings but increase maintenance and monitoring complexity.
First-time learner note: Read each model as a dataflow system: inputs become representations, representations become scores, and scores become decisions through a chosen loss and thresholding policy.
Production note: Track three things relentlessly in ML systems: data shape contracts, evaluation methodology, and the operational meaning of the model's errors. Most expensive failures come from one of those three.