0%
Intro
1. High dimensional Feature
- what is the difference between low dimensional feature and high dimensional feature
- x = [x1,x2,…,xn] -> y
- infinite dimensional features
- select features based on the data
2. Regression vs Classification
- regression: if y is continuous variable,
e.g., price prediction
- classification: if the label is a discrete variable
e.g., the task of predicting the types of residence
3. Supervised learning in computer science
- Image Classification: x = raw pixels of the picture, and y = label
- Natural language processing
4. unsupervised learning
- only have data without labels
- goal is to find interesting structure in the data
5. Clustering & Other
- k-mean clustering, mixture of Gaussians
- clustering genes
- principal component analysis (tools used in
LSA)
- word emeddings(Represent words by vectors),eg Word2vec
- clustering words with similar meanings
6. Reinforcement Learning
- learning to walk to the right
- the algorithms can collect data interactively
- method: try the strategy and collect feedbacks(Data Collections & Training),to improve the strategy based on the feedbacks
SL:Setup
- Linear Regression
- x -> y
- gradient descent
Linear Algebra
Probability
Gaussian Discriminant Analysis
Support Vector Machines. Kernels.
Evaluation Metrics
Reference
- http://cs229.stanford.edu/notes2020spring/lecture1_slide.pdf
- http://cs229.stanford.edu/notes2020spring/cs229-notes1.pdf
- CS224N/CS231N
- Identifying Regulatory Mechanisms using Individual Variation Reveals Key Role for Chromatin Modification.
- Luo-Xu-Li-Tian-Darrell-M.’18