Human-AI Tools for Aligning to Machine Representations
Currently, there is no effective method of meaningfully aligning abstract human conceptions into a representation that can be understood and operated upon reliably by a machine model. We propose a system that facilitates a human-machine feedback loop where designers can understand how their concepts are represented by the machine and use that information to refine their concept for better alignment. Through surfacing confounding context features and suggesting additional context features we help designers construct more precise definitions. Additionally, through simulating how their concepts are represented on different subsets of data we offer a method to help designers better understand their own conceptions of the situation as well as the “vocabulary” of their available data for better alignment. Through user testing, we found that simulation and prompting does help designer’s better align to their available machine representation, while giving them a better awareness of the tradeoffs between precision and recall as a result of refinement.
Team
Faculty
- None
Ph.D. Students
- None
Masters and Undergraduate Students
- 🎓 Harita Duggirala
- 🎓 Jiho Kwak