In this seminar, we learn together about the trending data mining and machine learning (DaML) techniques. The seminar of fall 2019 is currently scheduled on Fridays on a monthly basis.
12/6 Towards Robust Machine Learning Models
Date: 11/8 Friday 2:30pm-5:00pm This seminar has been postponed to 12/6 Friday 2:30pm-5:00pm
Location: E205 Westgate Building
Outline:
- Adversarial Attack:
- Evasion attack (Yiwei Sun)
- White box attack
- Fast Gradient Sign Method (FGSM)
- Projected Gradient Method
- Deep Fool
- Jacobian-based Saliency Map Attack (JSMA)
- Carlini & Wanger’s Attack
- One pixel attack
- Black box attack
- Substitute model
- ZOO: Zeros-th order optimization attack
- Query-efficient attack
- White box attack
- Poisoning attack (Ren Pang)
- Poison Frog
- Stingray
- Watermark attack
- BadNet
- TrojanNN
- Evasion attack (Yiwei Sun)
- Defense:
- Gradient masking/obfuscation (Ren Pang)
- Distillation
- Shattered Gradients
- Stochastic/Randomized Gradients
- Exploding & Vanishing Gradients
- Robust Optimization (Tianxiang Zhao)
- Adversarial training
- Certificated defense
- Regularization
- Adversarial Sample Detection
- Heuristic statistics
- Auxiliary Detector
- Prediction consistency
- Gradient masking/obfuscation (Ren Pang)
- Application:
- Text Domain (Suhang Wang)
- Graph Domain
Slides: https://docs.google.com/presentation/d/1EpjL58RjswO-Yep06A0xLNSKq1nIwy4vI00A34MEyLc/edit?usp=sharing
Speaker:
- Suhang Wang, Assistant Professor, IST
- Yiwei Sun, PhD student, IST
- Ren Pang, PhD student , IST
- Tianxiang Zhao, PhD student, IST
10/11 Interpretable Machine Learning
Date: 10/11 Friday 2:30pm-5:00pm
Location: E205 Westgate Building
Abstract: Machine learning models have shown success in terms of accuracy in prediction or classification. But in order to use such models to make policies, it is important to interpret them first. In this seminar, we will talk about the traditional and trending techniques in interpretable ML models.
Outline:
- Intrinsically Interpretable Models
- Target: Model (Presented by Chacha Chen)
- Linear regression
- Logistic regression
- Decision tree
- Equation-based model [Schmidt, M., et al, Science 2009]
- Other traditional approaches
- Target: Sample (Presented by Fenglong Ma)
- KNN (Instance-based)
- Attention-based model[Bahdanau et al, ICLR 2015; Xu et al, ICML 2015; Ma et al, KDD 2017; Vaswani et al, NeurIPS 2017]
- Target: Model (Presented by Chacha Chen)
- Post Hoc Interpretable Models
- Target: Model (Presented by Fenglong Ma)
- Permutation feature importance [Altmann et al, Bioinformatics 2010]
- Representative instance generation [Nguyen et al, NeurIPS 2016]
- Target: Sample (Presented by Wenbo Guo)
- Model-specific explanation (treat the networks as white boxes)
- Perturbation-based important feature identification [Fong et al, ICCV 2017; Dabkowski et al, NeurIPS 2017]
- Gradient-based saliency maps [Sundararajan et al, ICML 2018; Zhang et al, USENIX Security 2020]
- Model-agnostic explanation (treat the networks as black boxes)
- Auxiliary model-based explanation[(Ribeiro et al, KDD 2016; Guo et al, NeurIPS 2018]
- Instance-based explanation (Presented by Xinyang) [Koh and Liang, ICML 2017; Yeh et al, NeurIPS 2018]
- Model-specific explanation (treat the networks as white boxes)
- Target: Model (Presented by Fenglong Ma)
- Evaluation (Presented by Xinyang Zhang)
- Qualitative evaluation [Murdoch et al, ICLR 2018; Simonyan et al, ICLR Workshop 2014]
- Quantitative evaluation [Dabkowski and Gal, NeurIPS 2017; Guo et al, CCS 2018; Yeh et al, NeurIPS 2019]
- Open Questions
Slides: https://docs.google.com/presentation/d/1KhAYjxee_bup281ro3hCmR2mKUtT5zdbiGOniSstIEQ/edit?usp=sharing
Speaker:
- Fenglong Ma, Assistant Professor, IST
- Wenbo Guo, PhD student, IST
- Chacha Chen, PhD student, IST
- Xinyang Zhang, PhD student, IST
Audience: This seminar is open to public, feel free to forward this information to interested people. The room has limited seats (54). Audience should be familiar with basic machine learning and data mining techniques since this seminar is targeted at advanced machine learning and data mining techniques.
Photos on 10/11 (credit: Jordan Ford, Penn State)
9/13 Learning with Small Data
Date: 9/13 Friday 2:30pm-5:00pm
Location: W219 Westgate Building (location change!) E205 Westgate Building
Abstract: Though being in the era of big data, we frequently face the real-world problems with only small (labeled) data. Could we still make machine learn from small data? In this seminar, we will cover the state-of-the-art machine learning techniques to handle small data issue.
Outline:
-
- Data: augmentation
- Augmentation by using labeled data (present by Guanjie Zheng)
- Hand-crafted rule based augmentation (application specific techniques)
- Feature space augmentation
- SMOTE (Chawla et al, 2002)
- Autoencoder (Bengio et al., 2007, Vincent et al., 2010)
- Adversarial model
s
- Adversarial training (Goodfellow et al., 2014)
- Generative adversarial networks (GAN)
- GAN for image (Goodfellow et al., 2014)
- GAN for NLP: seqGAN (Yu et al., 2016)
- GAN for RL: Generative adversarial imitation learning (GAIL) (Ho et al., 2016)
- Augmentation by using unlabeled data (present by Hua Wei)
- Semi-supervised learning
- Co-training (Avrim and Mitchell, 1998)
- Graph-based methods (Xiaojin et al., 2003)
- Self-training
- kNN propagation
- Pseudo-labeling (Dong-Hyun, 2013)
- Entropy minimization
- Consistency regularization (Samuli and Aila, 2016, Antti and Valpola, 2017)
- MixMatch: combination of self-training, entropy minimization, and consistency regularization (David et al., 2019)
- Active learning
- Semi-supervised learning
- Augmentation by using labeled data (present by Guanjie Zheng)
- Model: knowledge transfer
- Transfer knowledge from the model learned from similar datasets (present by Huaxiu Yao)
- Transfer learning
- Fine tuning (target domain with labeled data)
- Unsupervised transfer learning (target domain with no labeled data)
- Discrepancy-based method: Loss function considering the discrepancy between source and target (Long et al., 2016)
- Adversarial method (Tzeng et al., 2017)
- Multi-task Learning
- Meta-learning
- Gradient-based: MAML (Finn et al., 2017)
- Non-parametric: (Snell et al., 2017)
- Task heterogeneity
- Transfer learning
- Transfer knowledge from domain expert (present by Porter Jenkins)
- Enriching representations
using knowledge graph
- ConceptNet (Speer et al., 2017)
- Healthcare (Ma et al. 2018)
- Regularizing the loss function
- Adding prior in Bayesian model
- Adding prior in discriminative model (Ma et al. 2018)
- Enriching representations
using knowledge graph
- Transfer knowledge from the model learned from similar datasets (present by Huaxiu Yao)
- Data: augmentation
Slides: https://docs.google.com/presentation/d/1NtFw5YE2WK9xdaEytTyuy1WO3nhvR6O5wohvZVIZT18/edit?usp=sharing
Speaker:
- Jessie Li, Associate Professor, IST
- Guanjie Zheng, PhD Student, IST
- Hua Wei, PhD student, IST
- Huaxiu Yao, PhD student, IST
- Porter Jenkins, PhD student, IST
Audience: This seminar is open to public, feel free to forward this information to interested people. The room has limited seats (54). Audience should be familiar with basic machine learning and data mining techniques since this seminar is targeted at advanced machine learning and data mining techniques.
We had a full room in September seminar!
Interested in contributing or collaboration? Send an email to Prof. Jessie Li (jessieli@psu.edu).