== Beginners Workshop Machine Learning == || '''From:''' || 2018-09-03|| || '''To:''' || 2018-09-14|| || '''Lecture time:''' || 8:45 - 12:00, 13:30 - 17:00 (approx.) || || '''Exam:''' || 2018-09-24|| || '''Room:''' || MAR 4.063, Marchstr. 23 || || '''Organisation:''' || Seulki Yeom: yeom@tu-berlin.de, Philipp Seegerer: philipp.seegerer@tu-berlin.de, David Lassner: lassner@tu-berlin.de || ||'''Language'''|| English || ||'''Application deadline'''|| June 15th, 2018 || == Enrollment / Limited number of participants == If you intend to participate, please send an e-mail to lassner@tu-berlin.de with title "Beginners Workshop Enrollment" and this text: {{{ Name: Your name Matr.Nr: Your student ID (Matrikelnummer) Degree: The degree you are enrolled in and want to use this course for. TU student: Yes/No (Are you a enrolled as a regular student at TU Berlin?) Other student: If you are not a regular student, please write your status. ML1: Yes/No (Did you take the course Machine Learning 1 at TU Berlin?) Other ML course: If you did not take ML1 at TU Berlin, please write if you took any equivalent course. }}} Participation spots are mostly assigned on a random basis. Please keep in mind that auditing students and Nebenhörer can only participate if less than the maximum number of regular TU students register for the course (http://www.studsek.tu-berlin.de/menue/studierendenverwaltung/gast_und_nebenhoererschaft/parameter/en/). (temporary) Workshop Lecture topics are: 1. Clustering, mixtures, density estimation * Density estimation: kernel density estimation, Parzen windows, parametric density * K means clustering * Gaussian mixture models, EM algorithm * Curse of dimensionality 2. Manifold learning * LLE * Embeddings (RBF) * Multidimensional scaling * tSNE 3. Bayesian Methods * What is learning? * Frequentist vs Bayes * Bayes rule * Naive Bayes * Bayesian linear regression * Bayesian/Akaike information criterion, Occam's razor 4. Classical and linear methods * Matrix factorization * Logistic regression * Regularization, Lasso, Ridge regression * Fisher's Linear discriminant * Gradient descent * Decision boundaries 5. Support Vector Machine * Linear SVM * Linear separability, maximum margin and soft margin * Duality in optimization, KKT conditions * SVM for regression * Multi-class SVM * Applications 6. Kernels * Feature transformations * Kernel trick * Nadaraya-Watson kernel regression 7. Neural Networks * Rosenblatt's Perceptron * Multi layer perceptron * Motivation with logistic regression * Backpropagation, (Stochastic) (Minibatch) gradient descent * Convolutional NNs * Famous Conv net architectures: AlexNet, GoogleNet, ResNet etc. * Recurrent NNs * Applications * Practical recommendations for Training of DNNs, hyperparameter tuning