I am currently a second-year Master’s student in Computational Science and Engineering at Harvard University, advised by Prof. Hima Lakkaraju. Previously, I graduated from UC Berkeley with a double major in Computer Science and Psychology, where I also collaborated with Prof. Bin Yu as part of the EECS Honors Program.
I’m broadly interested in the intersections of Trustworthy Machine Learning, Large Language Models, and Mechanistic Interpretability. The two overarching research questions I aim to address are:
(1) How could we obtain reliable interpretations of learning dynamics and explanations of observed model behaviors?
(2) How could we leverage such understanding to improve next-generation foundation models?
The trustworthiness of Large Language Models (LLMs) refers to the extent to which their outputs are reliable, safe, and ethically aligned, and it has become a crucial consideration alongside their cognitive performance. In practice, Reinforcement Learning From Human Feedback (RLHF) has been widely used to align LLMs with labeled human preferences, but its assumed effect on model trustworthiness hasn’t been rigorously evaluated. To bridge this knowledge gap, this study investigates how models aligned with general-purpose preference data perform across five trustworthiness verticals: toxicity, stereotypical bias, machine ethics, truthfulness, and privacy. Our results demonstrate that RLHF on human preferences doesn’t automatically guarantee trustworthiness, and reverse effects are often observed. Furthermore, we propose to adapt efficient influence function based data attribution methods to the RLHF setting to better understand the influence of fine-tuning data on individual trustworthiness benchmarks, and show its feasibility by providing our estimated attribution scores. Together, our results underscore the need for more nuanced approaches for model alignment from both the data and framework perspectives, and we hope this research will guide the community towards developing language models that are increasingly capable without sacrificing trustworthiness.
@inproceedings{li2024more,title={More RLHF, More Trust? On The Impact of Preference Alignment On Trustworthiness},author={Li, Aaron J. and Krishna, Satyapriya and Lakkaraju, Himabindu},booktitle={Proceedings of the International Conference on Learning Representations},year={2025},publisher={ICLR},venue={ICLR 2025},note={<span style="color:red;"> Oral Presentation, Top 1.8% </span>},}
ICML
Improving Prototypical Visual Explanations with Reward Reweighing, Reselection, and Retraining
Aaron J. Li, Robin Netzorg, Zhihan Cheng, Zhuoqin Zhang, and Bin Yu
In Proceedings of the International Conference on Machine Learning, 2024
In recent years, work has gone into developing deep interpretable methods for image classification that clearly attributes a model’s output to specific features of the data. One such of these methods is the Prototypical Part Network (ProtoPNet), which attempts to classify images based on meaningful parts of the input. While this architecture is able to produce visually interpretable classifications, it often learns to classify based on parts of the image that are not semantically meaningful. To address this problem, we propose the Reward Reweighing, Reselecting, and Retraining (R3) post-processing framework, which performs three additional corrective updates to a pretrained ProtoPNet in an offline and efficient manner. The first two steps involve learning a reward model based on collected human feedback and then aligning the prototypes with human preferences. The final step is retraining, which realigns the base features and the classifier layer of the original model with the updated prototypes. We find that our R3 framework consistently improves both the interpretability and the predictive accuracy of ProtoPNet and its variants.
@inproceedings{li2023improving,title={Improving Prototypical Visual Explanations with Reward Reweighing, Reselection, and Retraining},author={Li, Aaron J. and Netzorg, Robin and Cheng, Zhihan and Zhang, Zhuoqin and Yu, Bin},booktitle={Proceedings of the International Conference on Machine Learning},year={2024},publisher={PMLR},venue={ICML 2024},}