1 of 18

Privacy Leakage Study and Protection for Virtual Reality Devices

Dirk Catpo Risco, Brody Vallier, Emily Yao

7 August 2024

Project Advisor: Prof. Yingying (Jennifer) Chen

PhD Students as Mentors: Changming Li, Honglu Li, Tianfang Zhang

2 of 18

Introducing the Team Members

Students:

2

Dirk Catpo Risco

RU ECE MS

Changming Li

RU ECE PhD

Prof. Yingying (Jennifer) Chen

Honglu Li

RU ECE PhD

Tianfang Zhang

RU ECE PhD

Brody Vallier

RU ECE UG

Emily Yao

HTHS HS

Advisor:

Mentors:

3 of 18

Motivation

  • AR/VR devices have attracted millions of users and facilitate a broad array of emerging AR/VR applications
  • As a key component for motion tracking, Inertial Measurement Unit (IMU) consists of an accelerometer for measuring acceleration and a gyroscope for detecting rotations
  • Both sensors are present in each controller and the Head Mounted Display (HMD)

3

Gaming

Shopping

Banking

Education

4 of 18

Objectives

  • Data from zero-permission motion sensors encodes various types of the user’s private information, such as activity information and preferences

  • This project aims to study the sensor data management in commercial AR/VR headsets and analyze the potential of private information leakage

4

5 of 18

Methodologies

  • Investigate privacy leakage in Augmented Reality (AR)/Virtual Reality (VR) devices

  • Extract data from the IMU on AR/VR headset and controllers for Human Activity Recognition (HAR)

  • Use Support Vector Machine (SVM) and Large Language Model (LLM) to show how IMU data maliciously exposes activities of victim users

5

6 of 18

Attack Illustration

  • Utilize SVM as a baseline model to identify effective statistical features (e.g., mean, max, etc.) from motion data to recognize human activity

  • Design LLM prompts based on the effective statistical features

  • If LLM achieves comparable accuracy to SVM on motion prediction, it validates that adversaries could expose victim’s motion status without requiring data from victims

6

7 of 18

Motion Data Preprocessing

  • Denoise and smooth data to generate accurate waveforms
  • Compute 3D trajectories to visualize the motions

7

Example: Side Raise

8 of 18

Feature Extraction for SVM

8

Acc (m/s2)

Y-Axis Accelerometer

Acc (m/s2)

Acc (m/s2)

Front Raise #2

IQR = 0.21

Mean = -0.11

Peak-to-peak = 8.0

Front Raise #1

IQR = 0.16

Front Raise #3

IQR = 0.21

Time (s)

Peak-to-peak = 7.3

Mean = 0.03

Peak-to-peak = 7.4

Mean = 0.04

9 of 18

Feature Extraction for SVM

9

Y-Axis Gyroscope

Ang (deg/s)

Ang (deg/s)

Ang (deg/s)

Time (s)

Y-Axis Gyroscope

Mean = 0.35

Peak-to-peak = 21.4

Head Left

IQR = 1.09

Mean = 0.15

Head Right

IQR = 0.39

Peak-to-peak = 12.2

Head Down

IQR = 0.32

Mean = 0.16

Peak-to-peak = 2.1

10 of 18

Support Vector Machine (SVM)

  • An effective machine learning algorithm to find a hyperplane that separates classified data points

  • Works well on accurately classifying motion sensor data

  • Adversaries may require a huge amount of data from victim users during model training for accurate prediction

10

11 of 18

Large Language Model (LLM)

  • Works well on recognizing human language and other complex tasks

  • Can understand data and reproduce required outputs with designated prompts

  • Pre-trained on vast amounts of data, adversaries may require no training data from victim users to accurately expose human motions

11

12 of 18

Experimental Setup

  • Using Android Studio, we develop an application to extract data from the IMU sensors on Head-Mounted Display (HMD) and controllers of Meta Quest

12

13 of 18

Experimental Setup

  • We designed 6 activities for evaluation, including two hand-related activities and four head-related activities

13

Front Raise

Side Raise

Head Left

Head Right

Head up

Head Down

14 of 18

Activity Inference Using SVM

  • 3 statistic features (mean, peak-to-peak, and interquartile range) are extracted from the motion sensor data

  • The overall accuracy of exposing 6 types of activities using SVM achieves 99.33%

14

15 of 18

Activity Inference Using LLM

15

  • Developed a prompt for Gemini Advanced to understand the motion data
    • Explained the goal of the task and data types to be received
    • Asked LLM to extract features from the data and provided specific knowledge about how to utilize the features
    • Provided a response structure for results

Example prompt for specifying accelerometer readings

16 of 18

Activity Inference Using LLM

  • Using our prompt with Gemini Advanced, we achieve 90.6% accuracy

16

Gemini Advanced Accuracy

Trial #

Front Raise

Side Raise

Head Left

Head Right

Head Up

Head Down

1

H

2

H

H

3

L

4

H

H

5

H

H

6

H

H

7

H

H

H

8

H

R

9

H

10

H

Accuracy (%)

100

100

90

76.7

76.7

100

Key

Accurate (3/3)

Partial (2/3)

Inaccurate (1/3)

None (0/3)

Total (%)

90.6

*H = Head L = Left Hand R = Right Hand*

17 of 18

Conclusion and Future Work

  • With designated prompt, LLM achieves an accuracy similar to SVM, indicating the potential activity information leakage without training effort using LLM

  • With further prompt fine-tuning, the adversaries could realize stronger activity exposure attack using LLM

17

18 of 18

Thank You for Your Time

Questions?

Scan for

Project Website