Hey Model! 3D Printed Interactive & Modular Models Assist Blind & Limited Vision Users

Share this Article

Australian researchers Samuel Reinders, Matthew Butler, and Kim Marriott are exploring ways to improve 3D printed tools for individuals who are blind or have low vision (BLV). Releasing the details of their study in the recently published ‘Hey Model!” – Natural User Interactions and Agency in Accessible Interactive 3D Models,’ the researchers discuss issues that BLV users face, and what changes would help them—as well as exploring new features and preferred modes of interaction.

3D printing has played a role in a variety of different projects and programs for BLV people in the past few years—from allowing them to experience the joy of taking in famous art to creating interactive sculpture, and developing campaigns to help blind children learn. And, while some 3D printing programs have centered around making braille labels, in many cases such efforts are useless as so often blind people do not even know how to read braille fluently. Logically, audio labels would be more helpful, but the researchers realized a true need to speak with BLV people about their actual preferences for interaction.

In the first part of their study, the researchers attempted to understand which types of interaction strategies are ‘most natural,’ using a Wizard of Oz methodology—using a wizard to offer auditory resources for the model. The eight participants were able to interact with the 3D models without any technological limitations. Experiments lasted from one to two hours, with participants using their smartphones and either Siri or Google for completing tasks in reading, answering messages, making calls, searching for public transportation routes, and even requesting to hear jokes.

Participant demographic information

The six models used were chosen to vary in their application domain, the kinds of tasks they might support, and complexity. In all, they consisted of the following, with at least two participants exploring each:

  • Two bar charts with removable bars
  • Model of an animal cell
  • Map of a popular Melbourne public park
  • Thematic map of Australia
  • Frog dissection model with removable organs
  • Solar system model with removable planets

“These were intended to elicit desired interaction behaviors when components were removed or reassembled, as well as to determine whether participants removed components in order to compare them,” stated the researchers.

After interacting with the models, participants of the study were interviewed regarding their thoughts and comfort levels with the 3D-printed models, and they were asked if ‘differing agency capabilities’ would be useful. Interviews were stored on videotape, with audio footage—all of which was transcribed.

3D models used in Study 1 (left-to-right top-to-bottom): a) Two bar charts of average city temperatures with removable bars; b) Animal cell; c) Park map; d) Thematic map of the population of Australia; e) Dissected frog with removable organs; f) Solar system with removable planets

“The use of removable parts where appropriate proved to be a key design choice,” stated the researchers. The parts supported comparison (such as the size of planets), and also made for more compelling experiences, ‘Being able to pick them up? Yeah, I liked it … I like to be able to hold them in terms of the density and size, it is a bit hard to tell when you don’t pick them up.’

“Some design choices caused confusions. Most notable was the lack of an equator on the Earth in the solar system model, and the inclusion of Saturn’s ring, which was confused for the Earth’s equator.”

Interactive solar system model used in Study 2

The second study centered around validating a more functional I3M, integrating interaction techniques and agent functionality from the previous study. In choosing to prototype one model further, the solar system was chosen. Ultimately, as noted by the researchers, the model was ‘enjoyed immensely by all participants,’ and one individual gave more specific input stating that they had enjoyed such access to educational models greatly–and due to a lack of access to such resources at school.

“Indeed there seemed to be a knowledge gap regarding the solar system with most of the participants who used this model in Study 1,” stated the researchers.

Applicable modalities for each task and used by participants

The solar system model was chosen as the researchers realized it could be used for:

  • Tapping gestures to extract auditory information
  • Optional overview
  • On/Off functionality
  • Braille labelling
  • Conversational agent interface
  • Model intervention

Participants interfaced with the model, inspired by the Apple interface in saying ‘Hey Model!’ (as in ‘Hey Siri!’) before asking a question. Experiments were one to two hours long, depending on the enthusiasm of the participant.

“When talking to the model, participants treated it as a conversational agent and indicated that they preferred more intelligent models that support natural language and which, when appropriate, could provide guidance to the user,” concluded the researchers. “Participants wished to be as independent as possible and establish their own interpretations. They wanted to initiate interactions with the model and generally preferred lower model agency. However, they did want the model to intervene if they did something wrong such as placing a component in the wrong place.

“Such physically embodied conversational agents raise many interesting research questions, including their perceived agency, autonomy and acceptance by the end user. There are also many questions to be answered on how such agents can be implemented. A major focus of our future research will be to design and construct a fully functional prototype, conduct more extensive user evaluations with a variety of models, including maps, and to explore whether model agency preferences differ with age and environment.”

What do you think of this news? Let us know your thoughts! Join the discussion of this and other 3D printing topics at 3DPrintBoard.com.

[Source / Images: ‘Hey Model!” – Natural User Interactions and Agency in Accessible Interactive 3D Models’]

 

Share this Article


Recent News

3D Printing: the Future of Sticks

Intelligent and Automated Post-Processing for Resin 3D Printing Launched by Nexa3D



Categories

3D Design

3D Printed Art

3D Printed Food

3D Printed Guns


You May Also Like

3D Printing Webinar and Virtual Event Roundup, September 27, 2020

A range of topics will be covered in this week’s roundup of webinars and virtual events, starting with controlled nesting and increased productivity. Moving on, attendees can learn how to...

Featured

What Does the Siemens-Nexa3D Partnership Mean for 3D Printing?

3D printer manufacturer Nexa3D has announced a collaboration with technology company Siemens to automate its polymer laser sintering systems. Even during COVID-19, the two companies have remained committed to Industry...

3D Printing News Briefs, August 11, 2020: 3DEO, Nexa3D, AK Medical

In today’s 3D Printing News Briefs, 3DEO has won a design competition, and Nexa3D will be demonstrating its expanded line of ultra-fast polymers at this week’s AM Industry Summit. Finally,...

3D Printing Webinar and Virtual Event Roundup, August 9, 2020

We’ve only got four online events to tell you about this week—a summit and a few webinars, one of which is on-demand. Read on to learn more! AM Industry Virtual...


Shop

View our broad assortment of in house and third party products.