top of page

BirdBox - Mapping Data in different output modalities and measuring Interpretation and experience

Master Thesis 2021

Question

How can the output modality affect the user perception? 

Can we use User Experience Surveys to explore Data Physicalization and Tangible Interfaces? 

Objective

1: Measure user experience across three pairs of interactive data representations

2. Use AttrakDiff and PANAS response to identify their applicability

3. Measure how users interpret data and the output at cognitive, somatic and emotional level

The idea is to study the unique combination of two outputs and its influence on data perception and experience.

characterestic output.png
Group 71.png

Each side will have an information about a bird

Depending on the side user is facing there would an output

Group 69.png

The outputs were presented in 4 levels/intensities. For example if the user is exploring the category aggression:

Group 92.png

Gesse will yield the highest brightness/ audio frequency/haptic frequency

Research Method

Surveys (PANAS and AttrakDiff)
Observation
Think a loud
Interview

Pigeon will yield the lowest brightness/ audio frequency/haptic frequency

Participants

12 students (convenient sampling)
Age 24-50

Participants would be asked to participate in pairs to encourage a fluid ‘Think a loud’

Group 83.png

The Researcher simulates the output based on the side of the cubes participants can see

Two participants are asked to interact with the cube and interpret the output. Their interaction is recorded for later analysis

Study Design Considerations

Sequence of outputs were counter balanced to avoid the effect of task ordering.

Group 87.png

Pairs chosen were already living together and were taking necessary COVID 19 precautions. The study was conducted in their shared living space

Group 88.png

Urban Birds Dataset

Screenshot 2021-12-16 at 18.20 1.png

Cube Design (Pilot)

To simulate the Haptic, Audio and Display output, I used a Partner Vibrator, a bluetooth speaker and three remote operated cabinet lights

The audio and the display outputs are simulated from outside the box and the participants were asked to pretend that the lights would be inside the box

IMG_0566.JPG

Bluetooth Speaker

Cabinet lights with 4 levels of brightness control

9 cm App enabled Bluetooth Partner Vibrator

IMG_1114.JPG

The vibrator was placed snugly inside the box with foam peanuts to give a consistent output throughout the surface

Participants were provided with the taxonomy of the birds along with the pictures for them to familiarise or recall the said birds before and during the study

For the purpose of a Pilot I used a white cardboard box

Study Layout (Pilot)

IMG_3975.JPG

Changes in the final design

IMG_3862_edit 1.png

For the final study, a microcontroller and an accelerometer to automate the display output from within the box was used

Bluetooth Speaker

Adafruit LIS3DH Accelerometer

8.5cm App Enabled Bluetooth Vibrator

Adafruit ESP32 Microcontroller

Adafruit NeoPixel Fairy Lights

The vibrator and the microcontrollers were perched on the H shape support inside the cube.

IMG_1450.JPG
IMG_1039.JPG
IMG_0931.JPG

Study Layout (Final)

Analysis Method

Statistical Analysis (Quantitative data)
Thematic Analysis (Qualitative data)

Qualitative Analysis

Thematic Analysis - Compact analysis 4.jpg

Quantitative Analysis

PANAS and AttrakDiff surveys were used to measure the affect, hedonic and pragmatic values of the output.

Both surveys revealed that the data was either too scattered or was contradictory to what the participants said in the interview

PANAS_all.png

PANAS mean values

Fever graph word pair all.png

AttrakDiff Distribution

Summary

Participants learned, analyzed and compared data from memory

 

Audio Haptic output came off as more stimulating and generated a positive affect overall

The low vibration was associated with a living entity or may have subconsciously been treated as one

Haptic output was a metaphor for Aggression or size while Audio output was a metaphor for speed

Participants would focus on speed and repetition to search for complex patterns and more meaning

Display output was completely ignored

Audio communicated the data and Haptic communicated the sentiment of the data

Group 94.png
bottom of page