Presenters: Ted Thompson
An explanation of how machines learn and the intrinsic limitations of this method of learning. The talk will cover how machine learning is connected to probability and statistics, optimization, and generative vs discriminatory models. Examples of different types of machine learning models and what they can and cannot do will be included.
- Machine learning is a subset of AI.
- Simplistic description: a pattern recognition system.
- Probability and Statistics: use data sample to do:
- y = mx + b
- in machine learning, you are trying to find the best answer for m
- b is used to account for obscurities and bias in the model
- x is the input
- y is the result.
- two main ways of learning
- Discriminative model
- decision boundary
- Support Vector Machines
- most deep neural nets are discriminative models
- Generative model
- probability distributions of the data set
- What ML can't do
- learn without sufficient data
- explain itself
- cannot understand that if a < b and b < c, that a < c
- "Do the right thing" that is change break from the task to do something that is important but out of "character" e.g. a protestor stopping to help an injured police officer.
- What ML can do
- Repetitive tasks really fast
- Drive you crazy (takes a lot of effort to make your model work correctly)
- Generate synthetic data
- Question about the machine not being able to explain what it is doing.
- Bias vs Variance tradeoff
- to minimize error rates you want to minimize the bias and variance.
- less bias has more variance
- less variance has more bias