About
Ruth Rosenholtz is a Principal Research Scientist in MIT's Department of Brain and Cognitive Sciences, and a member of CSAIL. She joined MIT in 2003 after 7 years at the Palo Alto Research Center (formerly Xerox PARC). She has a B.S. in Engineering from Swarthmore College, and a Ph.D. from UC Berkeley in EECS. She brings her background in electrical engineering, specifically computer vision, to the study of human vision, including visual search, perceptual organization, visual clutter, and peripheral vision. Her work focuses on developing predictive computational models of visual processing.
Research
Human beings experience a rich visual percept. However, we often perform poorly when probed on the details of that percept, suggesting that vision is impoverished. Many such failures of vision have been interpreted in terms of limited capacity: our visual system cannot process everything with full fidelity, nor, in a given moment, perform all possible visual tasks. Rather, it must lose some information, and prioritize some tasks over others. But how, then, does real-world vision work as well as it does?
Currently my lab is interested in two important pieces of this puzzle. First, peripheral vision efficiently encodes its inputs using a scheme that preserves a great deal of useful information, while losing the information necessary to perform certain tasks used to probe the details of visual experience. We have developed the state-of-the-art model of peripheral encoding, and to date have tested this model on over 70 visual tasks. Because behavioral studies have often confounded peripheral vision with visual attention, our new understanding of peripheral vision necessitates rethinking what we know about attention from the ground up. Second, many tasks used to demonstrate impoverished vision are arguably inherently difficult; poor performance on these tasks may indicate general-purpose limits on the complexity of the tasks one can perform at a given moment. It seems promising that together these two components can make sense of a wide variety of phenomena, including vision’s marvelous successes, its quirky failures, and our rich percept of the visual world.
Past work in the lab includes study of perceptual organization, cues to shape perception, a measure of the "clutter" of a display, visual saliency, and visual search.
Publications
Rosenholtz, R. (2020). Demystifying visual awareness: Peripheral encoding plus limited decision complexity resolve the paradox of rich visual experience and curious perceptual failures. Attention, Perception, & Psychophysics, 82(3), 901-925.
Rosenholtz, R., Yu, D., & Keshvari, S. (2019). Challenges to pooling models of crowding: Implications for visual mechanisms. Journal of Vision, 19(7), 15.
Wijntjes, M. W. A. & Rosenholtz, R. (2018). Context mitigates crowding: Peripheral object recognition in real-world images. Cognition, 180, 158-164.
Ehinger, K. A., & Rosenholtz, R. (2016). A general account of peripheral encoding also predicts scene perception performance. Journal of Vision, 16(2), 13.
Keshvari, S., & Rosenholtz, R. (2016). Pooling of continuous features provides a unifying account of crowding. Journal of Vision, 16(3):39, 1-15.
X. Zhang, J. Huang, S. Yigit-Elliott, & R. Rosenholtz, “Cube search, revisited.” J. of Vision, 15(3):9, 1-18, 2015.