Shi, Taiwei

3 publications

ICLRW 2024 How Susceptible Are Large Language Models to Ideological Manipulation? Kai Chen, Zihao He, Jun Yan, Taiwei Shi, Kristina Lerman
ICLRW 2024 Safer-Instruct: Aligning Language Models with Automated Preference Data Taiwei Shi, Kai Chen, Jieyu Zhao
NeurIPSW 2024 WildFeedback: Aligning LLMs with In-Situ User Interactions and Feedback Taiwei Shi, Zhuoer Wang, Longqi Yang, Ying-Chun Lin, Zexue He, Mengting Wan, Pei Zhou, Sujay Kumar Jauhar, Xiaofeng Xu, Xia Song, Jennifer Neville