• Tidak ada hasil yang ditemukan

CS6464: Concepts In Statistical Learning Theory

N/A
N/A
Protected

Academic year: 2024

Membagikan "CS6464: Concepts In Statistical Learning Theory"

Copied!
2
0
0

Teks penuh

(1)

CS6464: Concepts In Statistical Learning Theory

Jan-May Semester 2020

‘G’ Slot; CS 36

Slots: Mon (1200–1250 HRS), Wed (1650-1740 HRS), Thu (1000-1050 HRS) and Fri (0900–0950 HRS)

Prof. Sukhendu Das, BSB 312; Phone: 4367 Email: [email protected], [email protected] TA(s:) Sonam Gupta ([email protected]),

Jasmeen Kaur ([email protected]), Devbrat Sharma([email protected])

Updated on January 24, 2020

Note: The course webpage is http://www.cse.iitm.ac.in/~vplab/statistical_learning_theory.html. A google group will be formed for course related communications; please check the e-mail regularly.

1 Course Objectives

In the recent past, algorithms of solving many illposed problems in the field of multi-dimensional signal pro- cessing and big data analytics have gained importance.

New methods of signal representation, modeling, opti- mization and learning have been formulated, which spans over various areas of Machine Learning, Pattern Recog- nition, Vision and Natural Language Processing. This course will provide an overview of the theories and cur- rent practices, required by students and scholars who intend to specialize in this field, to solve complex prob- lems in Machine Learning Applications for image, video, text and bioinformatics.

2 Learning Outcomes

• To learn existing statistical algorithms of Machine Learning (ML) and Pattern Recognition (PR).

• To understand the difference between Classifica- tion and Regression.

• To be aware of recent advances in the field of ML such as Online Learning, Transfer Learning etc.

• To have hands-on experience in implementing var- ious ML and PR techniques on different datasets.

• To learn how statistical distribution in datasets af- fect performance of ML and PR techniques.

• To learn to compare the performance of two learn- ing systems.

• To study few optimization methods used to esti- mate the parameters of a model during training.

3 Course prerequisite(s)

CS5011 or equivalent preferable.

4 Classroom Mode

Traditional Lectures (4 × [50 mins. slots]). Tutorials will be taken outside class hours (Tuesday or Saturday preferably). Tutorial problems have to be solved in class.

5 Textbooks

• T. Hastie, R.Tibshirani, J. Friedman, “The Ele- ments of Statistical Learning: Data Mining, Infer- ence and Prediction”, Springer Series in Statistics, 2009.

• V. N. Vapnik; “Statistical Learning Theory”, Wi- ley, 1998.

6 Reference Books

• Journal of the Royal Statistical Society: Series B (Statistical Methodology).

• Foundations and Trends in Machine Learning; Now Publishers Inc.

• Journal of Machine Learning Research; JMLR, Inc.

and Microtome Publishing (United States).

• Christopher M. Bishop, “Pattern recognition and machine learning”, Springer, 2006.

• Proceedings of ICML, NIPS, ICLR.

1

(2)

• Kevin Murphy, “Machine Learning, a Probabilistic Perspective”, MIT Press, 2012.

7 Course Requirements

You are required to attend all the lectures. If you miss any of them it is your responsibility to find out what went on during the classes and to collect any materials that may be handed out.

Class participation is strongly encouraged to demon- strate an appropriate level of understanding of the mate- rial being discussed in the class. Regular feedback from the class regarding the lectures will be very much appre- ciated.

8 Planed Syllabus

• Learning Problem, Risk functions, Statistical De- cision Theory.

• Ill posed and well posed problems.

• Least Square Regression, Bias Variance tradeoff.

• ERM (+Tikhonov Regularization), Iterative regu- larization by early stopping, SRM.

• Linear Models of Regression, Subset Selection methods, Shrinkage methods, Ridge regression.

• LASSO, LAR

• Bag-of-Words, Online Learning and Transfer Learning.

• SVM, Kernel, VC dimension, RKHS

• ADMM, Proximal gradient

9 Tentative Grading Policy

The following allocation of points is tentative. These may change during the semester.

End Sem 35-40

Mid Sem 15-20

Tutorials×[6] 15-20

Software Assignments×[2] 25-30

Total 100

10 Tentative Dates

The following allocation of points is tentative. These may also change during the semester.

Tutorial Dates 24/01/2020,

07/02/2020, 28/02/2020, 20/03/2020, 03/04/2020, 17/04/2020 Software Assignment 1 An-

nouncement

30/01/2020 Software Assignment 1 Deadline 25/02/2020 Software Assignment 2 An-

nouncement

21/02/2020 Software Assignment 2 Deadline 13/04/2020

Extra Classes 15/02/2020

(11:30-1:30) 21/03/2020 (11:30-1:30) 18/03/2020 (11:30 - 1:30) Mid Semester Exam 13/03/2020

(09:00-10:00) End Semester Exam 30/04/2020

(09:00-12:00)

11 Academic Honesty

Academic honesty is expected from each student partic- ipating in the course. NO sharing (willing, unwilling, knowing, unknowing) of assignment code between stu- dents, submission of downloaded code (from the Inter- net, or anywhere else) is allowed.

Academic violations will be handled by IITM Senate Discipline and Welfare (DISCO) Committee. In case of assignments, for the first instance of violation, an over- lap of greater than or equal to 75% will result in ZERO marks for all students involved. Overlap of 50% - 75%

will result in 50% deduction and overlap of 25% - 50%

will result in deduction of 20% of the marks originally awarded. In case of the second instance of code copy- ing, the DISCO Committee of will be intimated of the matter and will result in a drop of one-penalty level in final course grade. Please protect your Moodle account password. Do not share it with ANYONE. Do not share your academic disk drive space on the Campus LAN.

Attendance will strictly be maintained in all classes and monitored; Rules of academic section will be strictly followed for your eligibility in appearing for the End Sem exams and final grades will be released subject to such conditions as laid out by academic section rules of at- tendance. Each proxy in the attendance will be penalized by 5% of (absolute) marks. It becomes 2% each for the donor & beneficiary, if both accept the fault.

Tutorials are not exams. Occasional exchange of technical ideas is permitted, but not copying.

No complaints of copying will be entertained af- ter the tutorial. If you find anybody copying, immediately inform the TA’s present. If guilty, their copies will be taken away and they will ob- tain 0 marks for that tutorial.

2

Referensi

Dokumen terkait

Automatic Speaker Recognition System based on Optimised Machine Learning Algorithms 1st Tumisho Billson Mokgonyane Department of Computer Science University of Limpopo Polokwane,

4.1.1 Classifier Algorithms In our study, we used Machine Learning ML based classifiers like Logistic Regression LR, Random Forest RF, Gradient Boosting GB, and K-Nearest Neighbors

After preprocessing the data set, we apply nine conspicuous machine learning algorithms, namely k-nearest neighbors, logistic regression, SVM, naïve Bayes, classification, and

3.2.4 Machine Learning Algorithms: For their known accuracy rate Random Forest Algorithm, Logistic Regression, Support Vector Machine algorithm we use to applied on the training data

Result and Discussion The Differences in Learning Outcomes for Students Who Learn Using A Project-Based Learning Model than Those Who Learn Using Direct Learning Models Based on

Tiga model ML tradisional yang digunakan ialah Logistic Regression LR, Support Vector Machine SVM, dan Extreme Gradient Boosting XGB, manakala tiga lagi model deep learning yang

To represent the effect of predictor factors on categorical response variables, different machine learning classification algorithms are used, namely logistic regression, neural network

Machine learning in R I There are a huge number of packages available for machine learning in R I See CRAN Machine Learning task view I Rather than interacting with algorithms