課程名稱 |
分散式機器學習系統 Distributed Machine-Learning System |
開課學期 |
108-2 |
授課對象 |
電機資訊學院 資訊網路與多媒體研究所 |
授課教師 |
周承復 |
課號 |
CSIE5319 |
課程識別碼 |
922 U4430 |
班次 |
|
學分 |
3.0 |
全/半年 |
半年 |
必/選修 |
選修 |
上課時間 |
星期二7,8,9(14:20~17:20) |
上課地點 |
資111 |
備註 |
總人數上限:30人 |
Ceiba 課程網頁 |
http://ceiba.ntu.edu.tw/1082CSIE5319_DMLS |
課程簡介影片 |
|
核心能力關聯 |
核心能力與課程規劃關聯圖 |
課程大綱
|
為確保您我的權利,請尊重智慧財產權及不得非法影印
|
課程概述 |
Distributed Machine-Learning System is an introduction to these system-focused aspects of machine learning, covering guiding principles and commonly used techniques for scaling up to large data sets. That is, we will cover the techniques that lie between a standard machine learning course and an efficient systems implementation. Topics will include stochastic gradient descent, acceleration, variance reduction, methods for choosing metaparameters, parallelization within a chip and across a cluster, popular ML frameworks, and innovations in hardware architectures. |
課程目標 |
We look at the performance as well as design issues of large-scale machine learning application that is deployed in practice. After taking this course, students, who basic models and the basic algorithms, are able to modify those models (or systems) in a bunch of different ways such that the systems could run faster and more efficiently. That is, these modifications are really important—they often are what make the system tractable to run on the data it needs to process. |
課程要求 |
待補 |
預期每週課後學習時數 |
|
Office Hours |
|
指定閱讀 |
待補 |
參考書目 |
待補 |
評量方式 (僅供參考) |
|
週次 |
日期 |
單元主題 |
第1週 |
03/03 |
Course introduction |
第2週 |
03/10 |
DNN Overview |
第3週 |
03/17 |
Techniques for SGD |
第4週 |
03/24 |
Techniques for SGD
課程影片
https://drive.google.com/open?id=1wwRTT_C4DdbSTZlhRfGnzRkE-7rfodHx |
第5週 |
03/31 |
Variance Reduction in SGD |
第6週 |
04/07 |
Markov Decision Process and Reinforcement Learning |
第8週 |
04/21 |
期中考 |
第9週 |
04/28 |
1. 期中考檢討
2. RL 上半部 |
第10週 |
05/05 |
RL下半部 |
第11週 |
05/12 |
Paper presentation
(1) Adam: A Method for Stochastic Optimization
(2) Random Search for Hyper-Parameter Optimization
(3) Lookahead Optimizer: k steps forward, 1 step back
(4) Random Features for Large-Scale Kernel Machines |
第12週 |
05/19 |
Paper presentation
(1) Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks
(2) A Survey on Transfer Learning
(3) Domain-Adversarial Training of Neural Networks
(4) Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding |
第13週 |
05/26 |
Paper presentation:
(1) The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks
(2) Human-level control through deep reinforcement learning
(3) Guided Cost Learning: Deep Inverse Optimal Control via Policy Optimization
(4) Mastering the game of Go with deep neural networks and tree search |
第14週 |
06/02 |
Paper presentation:
(1) Applications of Deep Learning and Reinforcement Learning to Biological Data
(2) Deep-Reinforcement Learning Multiple Access for Heterogeneous Wireless Networks
(3) Learning Transferable Architectures for Scalable Image Recognition
|
第15週 |
06/09 |
Paper presentation:
(1) Federated Learning for Ultra-Reliable Low-Latency V2V Communications
(2) Federated Learning via Over-the-Air Computation
(3) Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks
(4) HOGWILD!: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent |
第16週 |
06/16 |
Paper presentation and term project presentation
(1) When Edge Meets Learning: Adaptive Control for Resource-Constrained Distributed Machine Learning
(2) In-Datacenter Performance Analysis of a Tensor Processing Unit |
|