- About
- Courses
- Research
- Computational Social Science
- Critical Data Studies
- Data Science
- Economics and Information
- Education Technology
- Ethics, Law and Policy
- Human-Computer Interaction
- Human-Robot Interaction
- Incentives and Computation
- Infrastructure Studies
- Interface Design and Ubiquitous Computing
- Natural Language Processing
- Network Science
- Social Computing and Computer-supported Cooperative Work
- Technology and Equity
- People
- Career
- Undergraduate
- Info Sci Majors
- BA - Information Science (College of Arts & Sciences)
- BS - Information Science (CALS)
- BS - Information Science, Systems, and Technology
- Studying Abroad
- MPS Early Credit Option
- Independent Research
- CPT Procedures
- Student Associations
- Undergraduate Minor in Info Sci
- Our Students and Alumni
- Graduation Info
- Contact Us
- Info Sci Majors
- Masters
- PHD
- Prospective PhD Students
- Admissions
- Degree Requirements and Curriculum
- Grad Student Orgs
- For Current PhDs
- Diversity and Inclusion
- Our Students and Alumni
- Graduation Info
- Program Contacts and Student Advising
Sunnie S. Y. Kim is a Ph.D. candidate in Computer Science at Princeton University advised by Olga Russakovsky. She works on responsible AI and human-AI interaction — specifically, on improving the explainability and fairness of AI systems and helping people have appropriate understanding and trust in them. Her research has been published in both AI and HCI venues (e.g., CVPR, ECCV, CHI, FAccT), and she has organized multiple workshops connecting the two communities. She has been recognized by the NSF GRFP, Siebel Scholars, and Rising Stars in EECS, and has interned at Microsoft Research with the FATE group. Prior to graduate school, she received a BSc degree in Statistics and Data Science from Yale University and spent a year at Toyota Technological Institute at Chicago.
Talk: Advancing Responsible AI with Human-Centered Evaluation
Abstract: As AI technologies are increasingly transforming how we live, work, and communicate, AI evaluation must take a human-centered approach to realistically reflect real-world performance and impact. In this talk, I will discuss how to advance human-centered evaluation, and subsequently, responsible development of AI, by integrating knowledge and methods from AI and HCI. First, using explainable AI as an example, I will highlight the challenges and necessity of human (as opposed to automatic) evaluation. Second, I will illustrate the importance of contextualized evaluation with real users, revisiting key assumptions in explainable AI research. Finally, I will present empirical insights into human-AI interaction, demonstrating how users perceive and act upon common AI behaviors (e.g., LLMs providing explanations and sources). I will conclude by discussing the implications of these findings and future directions for responsible AI development.