能夠可靠地執行算法計算的神經網絡可能對機器學習和理論計算機科學具有革命性的潛力。一方面,它們可以實現在深度學習模型中很少看到的那種外推泛化。另一方面,它們可以在以前認為不可訪問的輸入上運行經典算法。這兩個承諾都由神經算法推理藍圖指導,該藍圖最近在Petar Velickovic和Charles Blundell的立場論文中提出。理論上,這是一個非常優雅的對自然輸入進行推理的流程,它仔細地利用了深度神經網絡作為特征提取器的久經考驗的能力。實際上,我們到底走了多遠?在本教程中,我們旨在提供回答神經算法推理的三個關鍵問題所需的基礎:如何開發執行算法計算的神經網絡,如何在現實問題中部署這種神經網絡,以及如何深化其與經典算法的理論聯系。我們的教程將從頭開始,以一種具有基本計算機科學背景的任何人都可以訪問的方式。還將提供動手編碼片段,展示與會者如何在相關算法推理數據集(如CLRS)上直接發展他們在圖表示學習中的想法,然后在下游智能體中部署它們(如強化學習)。
參考文獻
Harris, TE. and Ross, FS. Fundamentals of a Method for Evaluating Rail Net Capacities. Project RAND Research Memorandum * Vlastelica, M., Paulus, A., Musil, V., Martius, G. and Rolínek, M. Differentiation of Blackbox Combinatorial Solvers. ICLR’20 * Hamrick, JB., Allen, KR., Bapst, V., Zhu, T., McKee, KR., Tenenbaum, JB. and Battaglia, PW. Relational inductive bias for physical construction in humans and machines. CogSci’18
Machine Learning (ML) techniques facilitate automating malicious software (malware for short) detection, but suffer from evasion attacks. Many researchers counter such attacks in heuristic manners short of both theoretical guarantees and defense effectiveness. We hence propose a new adversarial training framework, termed Principled Adversarial Malware Detection (PAD), which encourages convergence guarantees for robust optimization methods. PAD lays on a learnable convex measurement that quantifies distribution-wise discrete perturbations and protects the malware detector from adversaries, by which for smooth detectors, adversarial training can be performed heuristically with theoretical treatments. To promote defense effectiveness, we propose a new mixture of attacks to instantiate PAD for enhancing the deep neural network-based measurement and malware detector. Experimental results on two Android malware datasets demonstrate: (i) the proposed method significantly outperforms the state-of-the-art defenses; (ii) it can harden the ML-based malware detection against 27 evasion attacks with detection accuracies greater than 83.45%, while suffering an accuracy decrease smaller than 2.16% in the absence of attacks; (iii) it matches or outperforms many anti-malware scanners in VirusTotal service against realistic adversarial malware.
來自CBIO Chloé-Agathe Azencott講述的機器學習簡明教程,30分鐘內容快速了解機器學習。
強化學習是一種學習范式,它關注于如何學習控制一個系統,從而最大化表達一個長期目標的數值性能度量。強化學習與監督學習的區別在于,對于學習者的預測,只向學習者提供部分反饋。此外,預測還可能通過影響被控系統的未來狀態而產生長期影響。因此,時間起著特殊的作用。強化學習的目標是開發高效的學習算法,以及了解算法的優點和局限性。強化學習具有廣泛的實際應用價值,從人工智能到運籌學或控制工程等領域。在這本書中,我們重點關注那些基于強大的動態規劃理論的強化學習算法。我們給出了一個相當全面的學習問題目錄,描述了核心思想,關注大量的最新算法,然后討論了它們的理論性質和局限性。
Preface ix Acknowledgments xiii Markov Decision Processes 1 Value Prediction Problems 11 Control 37 For Further Exploration 63 Further reading 63 Applications 63 Software 64 Appendix: The Theory of Discounted Markovian Decision Processes 65 A.1 Contractions and Banach’s fixed-point theorem 65 A.2 Application to MDPs 69 Bibliography 73 Author's Biography 89
能夠解釋機器學習模型的預測在醫療診斷或自主系統等關鍵應用中是很重要的。深度非線性ML模型的興起,在預測方面取得了巨大的進展。然而,我們不希望如此高的準確性以犧牲可解釋性為代價。結果,可解釋AI (XAI)領域出現了,并產生了一系列能夠解釋復雜和多樣化的ML模型的方法。
在本教程中,我們結構化地概述了在深度神經網絡(DNNs)的背景下為XAI提出的基本方法。特別地,我們提出了這些方法的動機,它們的優點/缺點和它們的理論基礎。我們還展示了如何擴展和應用它們,使它們在現實場景中發揮最大的作用。
本教程針對的是核心和應用的ML研究人員。核心機器學習研究人員可能會有興趣了解不同解釋方法之間的聯系,以及廣泛的開放問題集,特別是如何將XAI擴展到新的ML算法。應用ML研究人員可能會發現,理解標準驗證程序背后的強大假設是很有趣的,以及為什么可解釋性對進一步驗證他們的模型是有用的。他們可能還會發現新的工具來分析他們的數據并從中提取見解。參與者將受益于技術背景(計算機科學或工程)和基本的ML訓練。
目錄內容:
Part 1: Introduction to XAI (WS) 可解釋人工智能
Part 2: Methods for Explaining DNNs (GM) 可解釋深度神經網絡方法
Part 3: Implementation, Theory, Evaluation, Extensions (GM) 實現,理論、評價
Part 4: Applications (WS) 應用
自2006年以來,神經網絡是引發深度學習革命的模型,但它們的基礎可以追溯到20世紀60年代。在這堂課中,DeepMind研究科學家Wojciech Czarnecki將介紹這些模型如何操作、學習和解決問題的基礎知識。他還介紹了各種術語/命名慣例,為與會者進一步、更高級的會談做準備。最后,他簡要介紹了神經網絡設計和開發的更多研究方向。
自監督學習(Self-Supervised Learning)是一種介于無監督和監督學習之間的一種新范式,旨在減少對大量帶注釋數據的挑戰性需求。它通過定義無注釋(annotation-free)的前置任務(pretext task),為特征學習提供代理監督信號。jason718整理了關于自監督學習最新的論文合集,非常值得查看!
地址: //github.com/jason718/awesome-self-supervised-learning
A curated list of awesome Self-Supervised Learning resources. Inspired by , , , and
Self-Supervised Learning has become an exciting direction in AI community.
Please help contribute this list by contacting or add
Markdown format:
- Paper Name.
[[pdf]](link)
[[code]](link)
- Author 1, Author 2, and Author 3. *Conference Year*
FAIR Self-Supervision Benchmark : various benchmark (and legacy) tasks for evaluating quality of visual representations learned by various self-supervision approaches.
Unsupervised Visual Representation Learning by Context Prediction.
Unsupervised Learning of Visual Representations using Videos.
Learning to See by Moving.
Learning image representations tied to ego-motion.
Joint Unsupervised Learning of Deep Representations and Image Clusters.
Unsupervised Deep Embedding for Clustering Analysis.
Slow and steady feature analysis: higher order temporal coherence in video.
Context Encoders: Feature Learning by Inpainting.
Colorful Image Colorization.
Unsupervised Learning of Visual Representations by Solving Jigsaw Puzzles.
Ambient Sound Provides Supervision for Visual Learning.
Learning Representations for Automatic Colorization.
Unsupervised Visual Representation Learning by Graph-based Consistent Constraints.
Adversarial Feature Learning.
Self-supervised learning of visual features through embedding images into text topic spaces.
Split-Brain Autoencoders: Unsupervised Learning by Cross-Channel Prediction.
Learning Features by Watching Objects Move.
Colorization as a Proxy Task for Visual Understanding.
DeepPermNet: Visual Permutation Learning.
Unsupervised Learning by Predicting Noise.
Multi-task Self-Supervised Visual Learning.
Representation Learning by Learning to Count.
Transitive Invariance for Self-supervised Visual Representation Learning.
Look, Listen and Learn.
Unsupervised Representation Learning by Sorting Sequences.
Unsupervised Feature Learning via Non-parameteric Instance Discrimination
Learning Image Representations by Completing Damaged Jigsaw Puzzles.
Unsupervised Representation Learning by Predicting Image Rotations.
Learning Latent Representations in Neural Networks for Clustering through Pseudo Supervision and Graph-based Activity Regularization.
Improvements to context based self-supervised learning.
Self-Supervised Feature Learning by Learning to Spot Artifacts.
Boosting Self-Supervised Learning via Knowledge Transfer.
Cross-domain Self-supervised Multi-task Feature Learning Using Synthetic Imagery.
ShapeCodes: Self-Supervised Feature Learning by Lifting Views to Viewgrids.
Deep Clustering for Unsupervised Learning of Visual Features
Cross Pixel Optical-Flow Similarity for Self-Supervised Learning.
Representation Learning with Contrastive Predictive Coding.
Self-Supervised Learning via Conditional Motion Propagation.
Self-Supervised Representation Learning by Rotation Feature Decoupling.
Revisiting Self-Supervised Visual Representation Learning.
AET vs. AED: Unsupervised Representation Learning by Auto-Encoding Transformations rather than Data.
Unsupervised Deep Learning by Neighbourhood Discovery. . .
Contrastive Multiview Coding.
Large Scale Adversarial Representation Learning.
Learning Representations by Maximizing Mutual Information Across Views.
Selfie: Self-supervised Pretraining for Image Embedding.
Data-Efficient Image Recognition with Contrastive Predictive Coding
Using Self-Supervised Learning Can Improve Model Robustness and Uncertainty
Boosting Few-Shot Visual Learning with Self-Supervision
Self-Supervised Generalisation with Meta Auxiliary Learning
Wasserstein Dependency Measure for Representation Learning
Scaling and Benchmarking Self-Supervised Visual Representation Learning
A critical analysis of self-supervision, or what we can learn from a single image
On Mutual Information Maximization for Representation Learning
Understanding the Limitations of Variational Mutual Information Estimators
Automatic Shortcut Removal for Self-Supervised Representation Learning
Momentum Contrast for Unsupervised Visual Representation Learning
A Simple Framework for Contrastive Learning of Visual Representations
ClusterFit: Improving Generalization of Visual Representations
Self-Supervised Learning of Pretext-Invariant Representations
Unsupervised Learning of Video Representations using LSTMs.
Shuffle and Learn: Unsupervised Learning using Temporal Order Verification.
LSTM Self-Supervision for Detailed Behavior Analysis
Self-Supervised Video Representation Learning With Odd-One-Out Networks.
Unsupervised Learning of Long-Term Motion Dynamics for Videos.
Geometry Guided Convolutional Neural Networks for Self-Supervised Video Representation Learning.
Improving Spatiotemporal Self-Supervision by Deep Reinforcement Learning.
Self-supervised learning of a facial attribute embedding from video.
Self-Supervised Video Representation Learning with Space-Time Cubic Puzzles.
Self-Supervised Spatio-Temporal Representation Learning for Videos by Predicting Motion and Appearance Statistics.
DynamoNet: Dynamic Action and Motion Network.
Learning Correspondence from the Cycle-consistency of Time.
Joint-task Self-supervised Learning for Temporal Correspondence.
Self-supervised Learning of Motion Capture.
Unsupervised Learning of Depth and Ego-Motion from Video.
Active Stereo Net: End-to-End Self-Supervised Learning for Active Stereo Systems.
Self-Supervised Relative Depth Learning for Urban Scene Understanding.
Geometry-Aware Learning of Maps for Camera Localization.
Self-supervised Learning of Geometrically Stable Features Through Probabilistic Introspection.
Self-Supervised Learning of 3D Human Pose Using Multi-View Geometry.
SelFlow: Self-Supervised Learning of Optical Flow.
Unsupervised Learning of Landmarks by Descriptor Vector Exchange.
Audio-Visual Scene Analysis with Self-Supervised Multisensory Features.
Objects that Sound.
Learning to Separate Object Sounds by Watching Unlabeled Video.
The Sound of Pixels.
Learnable PINs: Cross-Modal Embeddings for Person Identity.
Cooperative Learning of Audio and Video Models from Self-Supervised Synchronization.
Self-Supervised Generation of Spatial Audio for 360° Video.
TriCycle: Audio Representation Learning from Sensor Network Data Using Self-Supervision
Self-taught Learning: Transfer Learning from Unlabeled Data.
Representation Learning: A Review and New Perspectives.
Curiosity-driven Exploration by Self-supervised Prediction.
Large-Scale Study of Curiosity-Driven Learning.
Playing hard exploration games by watching YouTube.
Unsupervised State Representation Learning in Atari.
Improving Robot Navigation Through Self-Supervised Online Learning
Reverse Optical Flow for Self-Supervised Adaptive Autonomous Robot Navigation
Online self-supervised learning for dynamic object segmentation
Self-Supervised Online Learning of Basic Object Push Affordances
Self-supervised learning of grasp dependent tool affordances on the iCub Humanoid robot
Persistent self-supervised learning principle: from stereo to monocular vision for obstacle avoidance
The Curious Robot: Learning Visual Representations via Physical Interactions.
Learning to Poke by Poking: Experiential Learning of Intuitive Physics.
Supersizing Self-supervision: Learning to Grasp from 50K Tries and 700 Robot Hours.
Supervision via Competition: Robot Adversaries for Learning Tasks.
Multi-view Self-supervised Deep Learning for 6D Pose Estimation in the Amazon Picking Challenge.
Combining Self-Supervised Learning and Imitation for Vision-Based Rope Manipulation.
Learning to Fly by Crashing
Self-supervised learning as an enabling technology for future space exploration robots: ISS experiments on monocular distance learning
Unsupervised Perceptual Rewards for Imitation Learning.
Self-Supervised Visual Planning with Temporal Skip Connections.
CASSL: Curriculum Accelerated Self-Supervised Learning.
Time-Contrastive Networks: Self-Supervised Learning from Video.
Self-Supervised Deep Reinforcement Learning with Generalized Computation Graphs for Robot Navigation.
Learning Actionable Representations from Visual Observations.
Learning Synergies between Pushing and Grasping with Self-supervised Deep Reinforcement Learning.
Visual Reinforcement Learning with Imagined Goals.
Grasp2Vec: Learning Object Representations from Self-Supervised Grasping.
Robustness via Retrying: Closed-Loop Robotic Manipulation with Self-Supervised Learning.
Learning Long-Range Perception Using Self-Supervision from Short-Range Sensors and Odometry.
Learning Latent Plans from Play.
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.
Self-Supervised Dialogue Learning
Self-Supervised Learning for Contextualized Extractive Summarization
A Mutual Information Maximization Perspective of Language Representation Learning
VL-BERT: Pre-training of Generic Visual-Linguistic Representations
Learning Robust and Multilingual Speech Representations
Unsupervised pretraining transfers well across languages
wav2vec: Unsupervised Pre-Training for Speech Recognition
vq-wav2vec: Self-Supervised Learning of Discrete Speech Representations
Effectiveness of self-supervised pre-training for speech recognition
Towards Unsupervised Speech Recognition and Synthesis with Quantized Speech Representation Learning
Self-Training for End-to-End Speech Recognition
Generative Pre-Training for Speech with Autoregressive Predictive Coding
To the extent possible under law, has waived all copyright and related or neighboring rights to this work.
【導讀】2020新年伊始,多倫多大學Amir-massoud Farahmand和Emad A. M. Andrews博士開設了機器學習導論課程,介紹了機器學習的主要概念和思想,并概述了許多常用的機器學習算法。它還可以作為更高級的ML課程的基礎。
課程地址:
//amfarahmand.github.io/csc311/
機器學習(ML)是一組技術,它允許計算機從數據和經驗中學習,而不需要人工指定所需的行為。ML在人工智能作為一個學術領域和工業領域都變得越來越重要。本課程介紹了機器學習的主要概念和思想,并概述了許多常用的機器學習算法。它還可以作為更高級的ML課程的基礎。
本課程結束時,學生將學習(大致分類)
機器學習問題:監督(回歸和分類),非監督(聚類,降維),強化學習
模型:線性和非線性(基擴展和神經網絡)
損失函數:平方損失、交叉熵、鉸鏈、指數等。
Regularizers: l1和l2
概率觀點:最大似然估計,最大后驗,貝葉斯推理
偏差和方差的權衡
集成方法:Bagging 和 Boosting
ML中的優化技術: 梯度下降法和隨機梯度下降法
課程目錄:
參考資料:
(ESL) Trevor Hastie, Robert Tibshirani, and Jerome Friedman, The Elements of Statistical Learning, 2009.
(PRML) Christopher M. Bishop, Pattern Recognition and Machine Learning, 2006.
(RL) Richard S. Sutton and Andrew G. Barto Reinforcement Learning: An Introduction, 2018.
(DL) Ian Goodfellow, Yoshua Bengio and Aaron Courville (2016), Deep Learning
(MLPP) Kevin P. Murphy, Machine Learning: A Probabilistic Perspective, 2013.
(ISL) Gareth James, Daniela Witten, Trevor Hastie, and Robert Tibshirani, Introduction to Statistical Learning, 2017.
() Shai Shalev-Shwartz and Shai Ben-David Understanding Machine Learning: From Theory to Algorithms, 2014.
(ITIL) David MacKay, Information Theory, Inference, and Learning Algorithms, 2003.