Why do the figures in the Anglo-American panel of the Orozco murals look like zombies? Whatâs going on in the landscape behind the throne in Peruginoâs painting of the ?
A new mobile appâdeveloped by Associate Professor of Russian , Associate Professor of Art History Mary Coffey, Associate Professor of Art History Nicola Camerlenghi, and their studentsâuses augmented reality technology to let viewers discover these and other hidden details in two of Dartmouthâs most prized works of art.
A beta version of the app, called Augmented Dartmouth, will be available for download from the and in time for homecoming this weekend, says Gronas. This version will provide an augmented experience to viewers of JosĂ© Clemente Orozcoâs murals in Dartmouth Libraryâs Baker-Berry Library as well as Peruginoâs masterpiece, now on view in the .
âThe analogy was a footnoteâthese are basically notes for an eye,â Gronas says. âI thought it would be great to allow paintings or murals or architectural objects to talk back to us, the way books can do.â
The app works a little bit like PokĂ©mon Goâwhen users aim their phones at the art, the app shows points on the work that have been digitally augmented. Click on one of those pointsâsay, on the white-robed figure of Quetzalcoatl in the Orozco murals, or one of the saints in the Peruginoâand youâll open a window with more information.
Eventually, says Gronas, these explanatory windows might include reference images and videos, all linked to specific details in the art. The team ultimately hopes to expand the project to include other works of art on campusâand to provide the open-source software, known as Eyenotes, to other museums and institutions to use with their own collections.
The idea grew out Gronasâ earlier experience collaborating on that let users register real-time approval for political candidates during the 2016 presidential debates.
âWeâd been developing technology that allows us to âglueâ a piece of augmented reality to part of a visual image, and I was thinking about ways to use this in educational and cultural contexts,â says Gronas, a specialist in Russian literature who calls app development a âside interestâ to his scholarly pursuits. âOne object I thought about was the Orozco murals.â
When Gronas first approached Coffeyâan expert on Mexican muralism who frequently leads tours of the muralsâshe was a little skeptical.
âWhen Mikhail first came to me, I didnât know much about augmented reality,â she says. âI found examples of how some museums use it, but they were all super-gimmickyâlike, make the Mona Lisa dance, or put yourself in with the dinosaurs. I wasnât interested in that. But I did see a few examples that helped me understand the potential.â
Coffey says sheâs excited about the potential of the Eyenotes technology to allow for data about artworks such as murals to be crowd-sourced. âMurals are everywhere, but theyâre poorly documented, and attempts to create resources to document them often fall apart because of the scale of the project,â she says. âAn Eyenotes wiki-like tool that people could crowdsource has the potential to create the kind of information that we have been wanting all these years for public art and murals. Itâs something that I think would greatly interest my colleagues in Mexico.â
Camerlenghi was interested in the potential of augmented reality to help annotate virtual-reality projectsâincluding his three-dimensional , an ancient church that burned to the ground in the 19th century.
âThe model was instrumental for publishing a book and other scholarly endeavors, but I realized that what I had created was not very user friendly if I wasnât there to tell people what they were looking at and how I arrived at my conclusions,â Camerlenghi says. âI want to show other scholars my thinking, and I want students and teachers to use this as a tool to understand what the space was like. That means annotations. So I was imaging some kind of augmented reality to the virtual realityâis there a word for that?â
âYesâfused reality,â Gronas says.
âSo in comes Mikhail with this proposal, and I saw the potential application for virtual reality,â Camerlenghi says. Thanks to funding from the Kress Foundation, this fused version of Eyenotes will be the projectâs next stage of development, once Augmented Dartmouth is launched.
In the spring term, students in Coffeyâs âMexican Muralismâ class drafted the first texts for the app, based on their research on the Orozco mural. One of these students, Grace Hanselman â20, stayed on campus over the summer to continue working on the project. Natalie Shteiman â21 helped test-drive and proofread the application prior to launch. Other contributors to the Orozco content include Yazmin Ochoa Flores â21, Jennifer Lopez â20, Alicia Massey â22, Jhon Ortiz â20, and Karla Rosas â20. A , Courtney McKee â21, helped Camerlenghi develop content for the Perugino part of the app.
âWorking on the content has provoked for me a whole bunch of bigger conceptual questions,â Coffey says. âWhen youâre thinking about a book, youâre always thinking about discrete chapters and some kind of through-line argument that runs across the text. When youâre doing a guided tour, you think about how to move people through the space. This was a new challenge, and one I feel like weâre still working on. Weâve had to think about where people get information about the whole panel before we elaborate on certain details. For me thatâs where the challenge lies, and Iâll be interested to see how people use it.â
Support for the project has come from the 250th Anniversary Committee, the Hood Museum of Art, the Leslie Center for the Humanities, the Dartmouth Center for the Advancement of Learning, and the Kress Foundation.
The Augmented Dartmouth app is available for download at the Apple and .
Hannah Silverstein can be reached at hannah.silverstein@dartmouth.edu.