Thank you for signing up for the course! We look forward to working with you and hearing your feedback in our forums.

**Need help getting started?**

- Find an answer to your question in the Udacity FAQ
- Learn about different parts of the Classroom in Udacity Introduction

Contents

- 1 Course Resources
- 1.1 Reading Materials
- 1.1.1 Suggested Text
- 1.1.2 Optional Text
- 1.1.3 Reading List

- 1.2 Coding Resources
- 1.3 Downloadable Materials

- 1.1 Reading Materials
- 2 Course Syllabus
- 2.1 Lesson 1: Decision Trees
- 2.2 Lesson 2: Regression & Classification
- 2.3 Lesson 3: Neural Networks
- 2.4 Lesson 4: Instance Based Learning
- 2.5 Lesson 5: Ensemble B&B
- 2.6 Lesson 6: Kernel Methods & SVMs
- 2.7 Lesson 7: Comp Learning Theory
- 2.8 Lesson 8: VC Dimensions
- 2.9 Lesson 9: Bayesian Learning
- 2.10 Lesson 10: Bayesian Inference
- 2.11 Final Project: Predict Boston Housing Prices

- Tom Mitchell, Machine Learning. McGraw-Hill, 1997.
- Ethem Alpaydın, Introduction to Machine Learning. Second Edition.

- Larry Wasserman, All of Statistics. Springer, 2010.
- Richard Sutton and Andrew Barto, Reinforcement Learning: An introduction. MIT Press, 1998.
- Trevor Hastie, Robert Tibshirani and Jerome Friedman, The Elements of Statistical Learning. Springer, 2009.

- Linear Algebra
- Support Vector Machines
- Deep Learning
- Boosting
- Information Theory
- Optimization
- Dimensionality Reduction
- Unsupervised Learning
- Markov Decision Processes
- Game Theory

- WEKA Machine learning software in JAVA that you can use for your projects
- Data Mining with Weka A MOOC Course
- ABAGAIL Machine learning software in JAVA. This is hosted on my github, so you can contribute too
- scikit-learn A popular python library for supervised and unsupervised learning algorithms
- MATLAB NN Toolbox The toolbox supports supervised learning with feedforward, radial basis, and dynamic networks and unsupervised learning with self-organizing maps and competitive layers.
- Murphy's MDP Toolbox for Matlab
- MATLAB Clustering Package By Frank Dellaert
- ICA Example

- UCI Machine Learning Repository An online repository of data sets that can be used for machine learning experiments.
- Stanford Large Network Dataset Dataset of large social and information networks.
- Vision Benchmark Suite Autonomous car dataset
- Other datasets

You can download Supplemental Materials, Lesson Videos and Transcripts from Downloadables (bottom right corner of the Classroom) or from the Dashboard (first option on the navigation bar on the left hand side).

- Classification and Regression overview
- Classification learning
- Example: Dating
- Representation
- Decision trees learning
- Decision tree expressiveness
- ID3 algorithm
- ID3 bias
- Decision trees and continuous attributes

- Regression and function approximation
- Linear regression and best fit
- Order of polynomial
- Polynomial regression
- Cross validation

- Artificial neural networks
- Perceptron units
- XOR as perceptron network
- Perceptron training
- Gradient descent
- Comparison of learning rules
- Sigmoid function
- Optimizing weights
- Restriction bias
- Preference bias

- Instance based learning before
- Instance based learning now
- K-NN algorithm
- Won’t you compute my neighbors?
- Domain K-NNowledge
- K-NN bias
- Curse of dimensionality

- Ensemble learning: Boosting
- Ensemble learning algorithm
- Ensemble learning outputs
- Weak learning
- Boosting in code
- When D agrees

- Support Vector Machines
- Optimal separator
- SVMs: Linearly married
- Kernel methods

- Computational Learning Theory
- Learning theory
- Resources in Machine Learning
- Defining inductive learning
- Teacher with constrained queries
- Learner with constrained queries
- Learner with mistake bounds
- Version spaces
- PAC learning
- Epsilon exhausted
- Haussler theorem

- Infinite hypothesis spaces
- Power of a hypothesis space
- What does VC stand for?
- Internal training
- Linear separators
- The ring
- Polygons
- Sampling complexity
- VC of finite H

- Bayes Rule
- Bayesian learning
- Bayesian learning in action!
- Noisy data
- Best hypothesis
- Minimum description length
- Bayesian classification

- Joint distribution
- Adding attributes
- Conditional independence
- Belief networks
- Sampling from the joint distribution
- Recovering the joint distribution
- Inferencing rules
- Naïve Bayes
- Why Naïve Bayes is cool

Follow this link to access the final project.