Jiaxuan Wang
Me at Grand Canyon, 2014

Hobbies


basketball (20+ years); guitar (4 years); violin (20+ years);

Skills


C++ (4 years); python (10+ years); Java (1 year); PyTorch (6 years);
I'm a research scientist in machine learning at Meta. I'm interested in making AI interpretable and safe. I received PhD in computer science from the University of Michigan in April 2022, advised by professor Jenna Wiens.

My research focuses on model interpretability and model robustness to distribution shift. In addition, I'm interested in a wide range of topics including time series analysis, non-convex optimization, reinforcement learning, and sports analytics. My work is usually motivated by application in healthcare. Here's a copy of my PhD thesis.

I spent most of my pre-college years in Beijing. For the past 9 years, I've been living in Ann Arbor. I completed my undergrad at the Unisersity of Michigan as a computer science major and math minor in 2017, and started the PhD program in Fall 2017. In undergrad, I worked with professor Jia Deng to augment CNN with rotation invariant filters (2015-2016). I also worked as a machine learning intern at Bloomberg L.P. (2016) on parsing and ranking natural language query for financial charts, under the supervision of Dr. Konstantine Arkoudas and Dr. Srivas Prasad. Summer of 2020, I worked as a research intern in the Adaptive Systems and Interaction Group at Microsoft Research, mentored by Scott Lundberg, working on unifying Shapley value based model interpretation methods with a causal graph. After graudation (2022), I joined Meta as a research scientist to protect user privacy. As a team, we worked on building a reinforcement learning agent to prevent adversaries from misusing user data (that is the model decides actions to apply on users/IP addresses suspected to be malicious). The setting was formulated as a contextual bandit problem (context being user/ip level features and actions include block/no response/kill sessions and etc.) and was solved using thompson sampling with baysian linear regression. From scratch, I set up a feature attribution pipeline and a dashboard that enables the team to understand and defend the model’s logic. In addition, the tool makes it easy to track both covariate and conditional shifts, which are often the culprits of degrading model performances.

Publications

* denotes equal contribution
Learning Concept Credible Models for Mitigating Shortcuts
TL;DR: Mitigating shortcuts with partial knowledge on relevant concepts and extending credible model to the image domain
Jiaxuan Wang, Sarah Jabbour, Maggie Makar, Michael Sjoding, Jenna Wiens
Proceedings of the 36th Conference on Neural Information Processing Systems (NeurIPS), 2022
[paper][code][mobile friendly paper]

Shapley Flow: A Graph-based Approach to Interpreting Model Predictions
TL;DR: Don't choose between true to the model or true to the data: do both and more with a system level view.
Jiaxuan Wang, Jenna Wiens, Scott Lundberg
Proceedings of the 24th International Conference on Artificial Intelligence and Statistics (AISTATS), 2021
[paper][code]

AdaSGD: Bridging the gap between SGD and Adam
TL;DR: Speed of Adam and performance of SGD may be achieved by adapting a single learning rate.
Jiaxuan Wang, Jenna Wiens
arxiv preprint, 2020
[paper][code]

Relaxed Weight Sharing: Effectively Modeling Time-Varying Relationships in Clinical Time-Series
TL;DR: Reducing temporal conditional shift using multi task learning by treating each time step as a separate task.
Jeeheh Oh*, Jiaxuan Wang*, Shengpu Tang, Michael Sjoding, Jenna Wiens
Machine Learning for Healthcare, 2019
[paper][code]

Learning Credible Models
TL;DR: Learning models that are accurate and comply with human intuition so that they don't use proxy variables.
Jiaxuan Wang, Jeeheh Oh, Haozhu Wang, Jenna Wiens
ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2018
[paper][code]

The Advantage of Doubling: A Deep Reinforcement Learning Approach to Studying the Double Team in the NBA
TL;DR: It is a bad idea to double team star playes like LeBron who can pass and score.
Jiaxuan Wang*, Ian Fox*, Jonathan Skaza, Nick Linck, Satinder Singh, Jenna Wiens
MIT Sloan Sports Analytics Conference, 2018
[paper][code]

Learning to Exploit Invariances in Clinical Time-Series Data using Sequence Transformer Networks
TL;DR: A learnable data preprocessing module for time series data in healthcare.
Jeeheh Oh, Jiaxuan Wang, and Jenna Wiens
Machine Learning for Healthcare, 2018
[paper][code]

HICO: A Benchmark for Recognizing Human-Object Interactions in Images
TL;DR: A new image dataset focusing on who did what.
Yu-Wei Chao, Zhan Wang, Yugeng He, Jiaxuan Wang, Jia Deng
International Conference on Computer Vision (ICCV) 2015
[paper][data][code]

Technical reports

* denotes equal contribution
Using feature attribution to debug and monitor distribution shift for a production ML system
TL;DR: A case study to help debug and montior for distribution shift on a production ML system, demonstrating the use of feature attribution both with and without human in the loop.
Jiaxuan Wang
2023
[paper][slides]

© Jiaxuan Wang