ml_big_picture-2.0.pptx

32
Machine Learning Big picture Francis Pieraut – Oct 2016

Upload: francis-pieraut

Post on 14-Apr-2017

111 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: ML_big_picture-2.0.pptx

Machine Learning Big pictureFrancis Pieraut – Oct 2016

Page 2: ML_big_picture-2.0.pptx
Page 3: ML_big_picture-2.0.pptx
Page 4: ML_big_picture-2.0.pptx
Page 5: ML_big_picture-2.0.pptx

My startup bias about efficiency

Page 6: ML_big_picture-2.0.pptx

Plan

1. My path in ML2. AI big picture (expert systems to ML)3. ML trends over time 1980-2008+4. Type of ML (supervised vs unsupervised)5. Relationship Data-mining vs ML6. Training process7. regularization technique8. ML research big picture9. What is this Deeplearning Revolution?

10. ML in practice -> feature engineering11. Importance of the cost function12. Data importance -> NIPS 2009 13. The tagging nightmare14. ML & Optimization 15. Adversarial Examples

Page 7: ML_big_picture-2.0.pptx

Francis Evolution in ML

• 1999 – Decision Tree expert (Samy Bengio)• 2001-2003 – Research with Bengio (huge networks) -> flayers• 2003 – Idilia -> Importance of good tagged dataset and features &

overfitting • 2005-2006 – Dakis -> KISS (ML not required & importance of

comprehensive knowledge) – Expert System• 2006-2009 – Data-Mining (Understand first & features extraction)…MLboost• 2010-2013 – QMining -> big-data mining • 2003-2016 – Nuance -> Data Maturity & Data-driven design

Page 8: ML_big_picture-2.0.pptx
Page 9: ML_big_picture-2.0.pptx

Data Maturity model reminder

Page 10: ML_big_picture-2.0.pptx

AI big picture

Page 11: ML_big_picture-2.0.pptx

Type of ML

Parametric Non-Parametric

Reinforcement

Page 12: ML_big_picture-2.0.pptx

ML trends over time 1980-2008+

http://fraka6.blogspot.com/2013/10/deep-learning-history-and-most.html

Page 13: ML_big_picture-2.0.pptx

10 main ML algo

• Naïve Bayes Classifier Algorithm• K Means Clustering Algorithm• Support Vector Machine Algorithm• Apriori Algorithm• Linear Regression• Logistic Regression• Artificial Neural Networks (gradient)• Random Forests• Decision Trees (info theory)• K Nearest Neighbors

***Machine learning dangerous hype****

Page 14: ML_big_picture-2.0.pptx

Traps and Pitfalls

Page 15: ML_big_picture-2.0.pptx

Data-Mining vs Machine Learning

Page 16: ML_big_picture-2.0.pptx

Traininig Process

Classification error over time

Training

Page 17: ML_big_picture-2.0.pptx

Regularisation technique

• Regularization is a technique used in an attempt to solve the overfitting [1] problem in statistical models.*

• Exemple:– Early stopping– Decrease constant– Dropout– Mini-batch– Better cost function (ex: margin vs MSE)

Page 18: ML_big_picture-2.0.pptx

What is tough about ML

• More parameters = more examples are required

• Tagged data is hard to create compare to untagged data

• There is no magic -> Feature engineering• Better features -> less examples -> less

capacity problem • Getting good example sampling (don’t

introduce bias)

Page 19: ML_big_picture-2.0.pptx

Example feature engineering

Page 20: ML_big_picture-2.0.pptx

Example feature engineering

Page 21: ML_big_picture-2.0.pptx

What is this Deeplearning Revolution?

• Deep architecture are more powerful then shallow architecture

• Before 2006 we couldn’t train deep architecture• Revolution

– Convolution NN – Train generative models (Auto-encoder) -> learn the

data constraints…..unsupervised learning… (better parameters initialization)

– STD Training

Page 22: ML_big_picture-2.0.pptx
Page 23: ML_big_picture-2.0.pptx
Page 24: ML_big_picture-2.0.pptx

Example of deep learning in images

Page 25: ML_big_picture-2.0.pptx

ML learning in practice

• Black box = recipe for a disaster• 90% feature engineering• ML = automatic tuning• Garbage in = Garbage out• Tagging is a pain….manual work

Page 26: ML_big_picture-2.0.pptx

Importance of the cost function

• Neural network cost functions (back prop)– MSE & Log soft max– Example NETFLIX & recommendation

• Optimization– SVM = Maximize Margin

Page 27: ML_big_picture-2.0.pptx

Data Importance NIPS 2009

• Google -> that is enough – Parameter optimization; tweaking kernels (SVM)

– More parameters then # examples

– Simpler model + more data = what works

Page 28: ML_big_picture-2.0.pptx

The tagging nightmare

• You still need tagged data• Tagged data is hard to automate + error

prone• Tagged data is error prone (garbage in

garbage out)– Idilia use case– Nuance use case

Page 29: ML_big_picture-2.0.pptx

The lie about ML

• Machine learning != Optimization• Machine learning != Statistics• Machine learning = Optimization problem

with constraints to generalize (regularization)

Page 30: ML_big_picture-2.0.pptx

Adversarial example - Ian Goodfellow (now at open.ai)

Page 31: ML_big_picture-2.0.pptx

Conclusion

• ML is a quite mature field • ML != Deeplearning

– Deeplearning = major breakthrough, hype phase, not mature

• NN = optimization problem with constraints• SP operates more like expert systems• Algo is as good as its inputs -> feature

engineering

Page 32: ML_big_picture-2.0.pptx

QUESTIONS

[email protected]

hum...