Stanford GEDI Team META:
GolfVision

Meet the Students

Meet our Supervisors

Acknowledgments

We would like to express our sincere gratitude to Professor Khoo Eng Tat and our Teaching Assistant Graham Zhu for their invaluable guidance and support throughout this project. Their expertise and insights have been instrumental in shaping our approach and deepening our understanding of the challenges we aimed to address. We also extend our thanks to all the professors, social workers, elderly participants, and community organizers who generously shared their time and experiences with us. Their contributions were crucial in informing our research and ensuring that our solutions were grounded in real-world needs. Finally, we appreciate the support from the NUS-Stanford Global Engineering Design Innovation Programme for providing us with this opportunity to collaborate and innovate.

1. Introduction & Problem Clarification

1.1 Project Objective

1.1.1 Project Brief and Collaboration

This project is part of the NUS Innovation & Design Programme (iDP), undertaken as a collaboration between the National University of Singapore (NUS) and Stanford University, with Meta as the industry partner. Meta's project brief challenged the team to explore how smart glasses can be integrated into our daily activities. The core task was to discover a compelling market application for Meta's Ray-Ban smart glasses platform. We were encouraged to begin with deep user research, uncover genuine pain points, and then propose a solution where the smart glasses form factor delivers a meaningfully better experience than existing alternatives.

1.1.2 Market Opportunity and Form Factor Fit

We identified recreational golf as a uniquely high-value application for smart glasses technology. The global golf industry represents a substantial and growing market: The global golf industry is substantial and growing: the U.S. golf courses and country clubs sector alone reached US$41.0 billion in 2024 [1], the global golf equipment market is estimated at US$9.24 billion in 2025 [2], and the golf GPS device market — the segment most directly comparable to our product — was valued at US$3.2 billion in 2024, growing at 8.5% CAGR [3], confirming that golfers already pay for on-course technology. Golf is also an ideal form-factor fit: it is a hands-occupied, eyes-up sport where pulling out a phone or rangefinder disrupts the pre-shot routine, and where invisible environmental variables (wind, temperature, altitude) directly affect every shot. This combination of a large, tech-receptive market with well-documented performance pain points made golf the strongest candidate for an AR-assisted, AI-powered smart glasses application.

1.2 Introduction

"The adrenaline was pumping a little bit." [4] This was professional golfer Richard T. Lee's explanation after a stunning miss: a 2-foot putt on March 15, 2026, that cost him the Aramco LIV Golf Singapore title. If a seasoned pro can "choke" on a distance usually considered unmissable, it begs a fundamental question: why do golfers fail to consistently execute the shots they are physically capable of making?

The answer lies in a hidden battle between mental focus and environmental complexity. Golf demands instantaneous decisions based on invisible variables—wind gradients, altitude, and humidity—that drastically alter ball flight. While elite caddies help pros navigate these, most golfers are left to guess. More critically, sports psychology has identified the "Quiet Eye" (QE) phenomenon as the ultimate differentiator: elite golfers hold a steady gaze for 2.5 to 3 seconds before a putt, while amateurs typically fixate for less than one [5]. Research proves that training this specific attentional skill can shave 2 strokes per round [6], yet no consumer product currently exists to help golfers visualize and master this "pro" advantage in real-time.

Our project aims to change this. Working in conjunction with Stanford University, with Meta as our project partner, we present GolfVision: an Augmented Reality (AR) heads-up display (HUD) integrated with an AI caddie that provides an edge to golfers by allowing them to see what the elite golfers experience, with real-time, environment-adjusted playing yardage calculator and live focus feedback using a Pupils Lab Bare Metal module for measurements of quiet eye.

1.3 Primary research

Interviews were conducted with four golfers of varying skill levels and one coach. These sessions were guided by an outcome table comprising specific job steps and outcome statements synthesised from online sources [7][8].

1.3.1 Interviewing Golfers

Interviews were conducted with a cohort of five golfers to identify the specific performance outcomes they aimed to achieve through GolfVision. We presented a comprehensive list of outcome statements and asked respondents to rank them based on their perceived importance and current level of dissatisfaction with existing tools.

Format of outcome statement: Direction + Unit of Measure + outcome desired [9]

1.3.2 Jobs-To-Be-Done Framework

To systematically identify golfers' needs, we employed the Jobs-to-Be-Done (JTBD) framework, specifically the Job mapping technique outlined by Ulwick [9].

Through structured interviews with golfers of varying levels and online sources [7][8], we mapped the core Job steps of a golfer into three phases: Think, Feel, and Play. Each reflects the cognitive, emotional, and physical stages through which a golfer progresses for each shot.

Think Box: This phase covers the pre-shot decision-making process, from assessing the lie of the ball through to visualising the intended shot and managing emotions before addressing the ball. This phase accounts for half of the job steps, indicating that the majority of a golfer's cognitive workload occurs before the club is ever swung. This aligns with existing literature that identifies pre-shot routines and decision-making as critical determinants of shot quality [10].

Feel Box: This phase captures the transition from planning to execution, in which the golfer feels the intended swing, settles into a stance, fixates on the ball, and commits to the shot. This is where the quiet eye phenomenon is most active, as the golfer's final gaze fixation on the ball directly precedes motor execution.

Play Box: This phase encompasses post-shot actions, observing the ball's trajectory, reflecting on the shot, and locating the ball. Although shorter, this phase is significant because it feeds back into the Think phase for the next shot, creating a learning loop.

1.4 Analysis of Primary Research

Using the 5 Whys technique, we drilled down into the underlying reasons for performance inconsistency, specifically focusing on the 'Think Phase' and 'Feel Phase' identified through the Job mapping in Section 1.3.

1.4.1 Outcome-Driven Innovation Opportunity Analysis

To prioritise which outcomes to address, each outcome statement was scored using the Outcome-Driven Innovation (ODI) opportunity algorithm as defined by Ulwick [9]: Opportunity Score = Importance + MAX(Importance − Satisfaction, 0).

Focus-driven Execution improvement

Outcome Statement Adam Dom Nicholas Onn Lucas
Minimise the likelihood of hitting a shank 7 7 9 5 5
Minimize the likelihood of being nervous 5 - 8 6 7
Increase the likelihood of reproducing the desirable swing 7 - 9 6 7
Increase the likelihood of visualizing shot correctly 7 5 7 5 4
Minimize the risk of hitting the ball fat/thin 5 - 8 - -

Course Intelligence

Outcome Statement Adam Dom Nicholas ONN Lucas
Maximise the likelihood of calculating the best shot (temp, humidity, elevation, wind) 5 7 - 8 7
Maximise the accuracy of choosing the right club 7 6 8 6 6
Minimise the number of times the ball lands in hazard zones 7 8 8 5 6
Increase the likelihood of accurate ground yardage 6 6 7 5 5
Minimise yardage errors due to lack of experience 7 3 6 7 7
Misjudging elevation due to optical illusions 6 4 5 5 -

1.4.2 The 5 Whys

Considering the complicated nature of golf, we split the 5 whys into 2 different branches - the think phase and the feel phase.

Branch A: The “Think” Phase
(Decision-Making Routine)
Branch B: The “Feel” Phase
(Attentional State)
Problem: Golfers frequently make suboptimal tactical decisions on the course.
Why?
They must mentally synthesize multiple shifting environmental variables (wind, elevation, air density).
Why?
Data is gathered from fragmented, secondary sources and integrated under high pressure.
Why?
Current tools require the golfer to break their stance and line of sight to consult external displays.
Why?
Golfers rely on subjective “feel” over objective data because integrated data is not in their field of view.
Problem: Golfers often fail to execute shots they are physically and technically capable of performing.
Why?
Attentional state at the moment of address is inconsistent, preventing the transition from planning to action.
Why?
Focus is susceptible to pressure and distraction, and golfers cannot monitor their gaze-action coupling.
Why?
Golfers are unaware of instability because they cannot objectively self-assess “Quiet Eye” duration in real-time.
Why?
This biofeedback is missing from current training tools available to consumers.
Root Cause: No existing technology delivers integrated, real-time environmental intelligence directly within the golfer’s primary field of view during the critical decision-making window. Root Cause: No accessible consumer product provides real-time, gaze-based focus confirmation that allows a golfer to verify their cognitive readiness before initiating a swing.

The execution of both 5 Whys branches successfully converged on two fundamental, unaddressed root causes that define the scope of our proposed solution.

  1. Course Intelligence:
    1. Strategic Deficiency (Think): The inability to access integrated environmental data without disrupting the physical routine.
  2. Psychological Deficiency (Feel):
    1. The absence of objective indicators to validate and train the Quiet Eye phenomenon during active play.

1.5 Secondary Research

This literature review serves to validate the core problems and root causes established in the preceding primary research analysis, highlighting three overlapping challenges that set our product apart from competitors.

1.5.1 Inaccurate Playing Yardage

Inaccurate playing yardage estimation, crucial for cutting strokes, is complicated by environmental factors (wind, temperature, humidity, altitude) that most golfers ignore or vaguely estimate. While a common coaching approach suggests a simple wind-distance percentage [11], this method overlooks individual trajectory and spin sensitivity, as well as the asymmetrical effect of wind direction [12]. Research confirms that low confidence in yardage calculations leads to club selection hesitation and poor shot execution [13], highlighting the need for real-time, integrated environmental data.

1.5.2 Inaccessibility to Professionals

Access to professional guidance is expensive and scarce. Only 9% of U.S. golf courses offer caddie services, and only 5% of golfers use them [14], costing an average of $113 USD per bag—a recurring barrier for recreational players. Coaching sessions are similarly costly and time-constrained. This market gap confirms the salient unmet need for affordable, personalized feedback during actual play.

1.5.3 Unaddressed Mental Dimension

Lastly, the mental dimension of attentional control and focus is underserved by current tools. Traditional verbal coaching to "focus on the ball" is generalized and open-loop, offering no specific, quantifiable feedback. Research validates the Quiet Eye (QE) phenomenon as a reliable performance predictor [6], showing that QE training improves putting accuracy under pressure while reducing anxiety [15]. Despite evidence that QE training can shave 1.9 putts per round [6], no accessible product provides real-time QE feedback to golfers during play.

1.5.4 Competitor Analysis

The current market is fragmented across single-function devices that can be mapped along two axes: legacy versus emerging, and direct versus indirect competition. Legacy indirect competitors such as laser rangefinders (e.g. Bushnell Pro X3+) and GPS watches (e.g. Garmin Approach S70) address only raw distance, while legacy direct competitors like shot-tracking tags (e.g. Arccos Smart Sensors) and swing analysers (e.g. HackMotion) offer post-round or practice-only analytics with no real-time on-course utility. Emerging indirect competitors such as AI coaching apps (e.g. SWEE AI) provide personalised instruction but remain screen-bound and disconnected from live course conditions. The closest emerging direct competitors, such as CaddieVision and RANGEZ, consolidate distance data into AR glasses but still lack environmental sensing, eye-tracking, and AI coaching. PuttView X delivers AR green-reading via HoloLens but is restricted to putting practice. Across all four quadrants, a golfer seeking comprehensive support must purchase, carry, charge, and context-switch between multiple disconnected devices — and even then, no combination adjusts yardage for real-time environmental conditions, measures attentional state, or delivers strategic advice mid-round. Our product collapses this entire fragmented ecosystem into a single wearable, occupying a white space that no existing competitor — legacy or emerging — currently addresses.

1.6 Problem Statement

Recreational golfers face performance inconsistency because existing tools fail to provide two critical needs: integrated, real-time environmental data for accurate playing yardage, and objective, gaze-based feedback on attentional focus (Quiet Eye). This fragmentation and lack of a hands-free, heads-up solution disrupts the pre-shot routine, preventing amateurs from mastering the mental and environmental complexities of the game.

Our project develops GolfVision to solve performance inconsistency among recreational golfers by fusing two previously fragmented needs into a hands-free platform: delivering real-time, environment-adjusted playing yardage (Course Intelligence) and providing objective, gaze-based feedback on attentional control (Quiet Eye).

1.7 Value Proposition

Based on the synthesis of primary research, the 5 Whys analysis, and the market gaps validated by the secondary research and competitor analysis, we now formalize our solution's positioning. To validate that our products effectively address the root causes of the Think and Feel phases, we applied the Value Proposition Canvas framework. This methodology maps the customer profile's jobs, pains, and gains (right side) against our product’s value map (left side), which details the pain relievers and gain creators designed to deliver our distinct competitive advantage.

1.7.1 Customer Profile

Customer Jobs
  • Estimate playing yardage with higher accuracy, while accounting for environmental conditions
  • Select the appropriate clubs for each shot
  • Maintain focus and enhance attentional control before and during shot
  • Improve golf score by reducing strokes
  • Track performance over time to identify weaknesses and improve performance
  • Enjoy the game with minimal frustration
Pains
  • Inaccurate mental yardage calculations lead to over or under clubbing
  • Lack of access to caddie services; only 9% of courses offer caddie services
  • Coaching is expensive and time-limited
  • Unable to identify personal limitations, especially in attentional control, such as when quiet eye duration drops below optimal threshold
  • Decision fatigue and stress under pressure, which affects their performance, especially at critical shots
  • Bad shots might negatively affect mood and performance for the rest of the game
  • Existing tools (eg rangefinders, GPS watch) only give raw distances with no accounting for environment conditions
  • No objective feedback on mental training (eg focus quality)
  • Using phone apps or external devices break focus and flow during pre-shot, adding time and disruptions
  • Incompatibility with coach or caddie, which affects overall play
Gains
  • Confidence in club selection
  • Access to caddie-level advice while minimising cost
  • Real-time feedback on attentional control, which allows for correction before next shot
  • Reducing score variability and improving handicap
  • Better emotional regulation
  • Objective post-round data review and feedback
  • Maximise enjoyments and minimise frustration

1.7.2 Value Map

Product and Services
  • Provision of real time data
  • Playing yardage calculator that accounts for environment conditions
  • Eye tracking module measuring quiet eye duration before shot
  • Automatically updated weather panel
  • LLM-powered AI caddie that helps with club selection and shot strategy
  • Verbal input, providing hands-free experience
  • Personalised coaching feedback on-demand based on collected quiet eye and shot data from memory
Pain Relievers
  • Eliminates the need for mental calculation of playing yardage to account for environment conditions
  • Removes asymmetry of headwind and tailwind in calculations
  • Makes quiet eye data visible, allowing golfers to self-correct their fixation
  • Replaces the recurring fee of caddies and coaches, removing the barrier to access expert guidance
  • Reduces decision fatigue during play, such as club selection and shot strategy
  • Reduces stress under pressure, as quiet eye feedback helps with focus and AI provides guidance
Gain Creators
  • Makes tour-level environmental data more accessible, when it was previously only available to professionals with trained caddies
  • Trains attentional control and focus in the background while they play, acting as an additional mental coach
  • Enables golfers to build personalised distance and clubs used data over time
  • Creates hands-free experience, where information appears conveniently in the line of sight, without disrupting pre-shot routine
  • AI caddie receives input from play and corrects them or provides recommendations based on individual tendencies
  • Provide post-round analytics to help users improve their game
  • Bridges disconnect from lessons to course, allows coaching during actual play
  • Accelerates handicap improvement by optimising club selection, shot strategy and attentional control
  • Increases enjoyment and confidence

1.8 Project Boundary

The scope of this project is confined to the 'Think Box' and 'Feel Box' phases of the golfer's workflow, as defined in the primary research. Specifically, the technical development focuses on delivering 'Course Intelligence' within the Think Box and enhancing 'Attentional Control' within the Feel Box via gaze-based Quiet Eye feedback. While the 'Play Box' (post-shot analysis and learning loops) remains a critical component of the overall golfer journey, it is formally deferred to future iterations of the platform. Adhering to the collaborative mandate with Meta, the implementation is strictly constrained to the smart glasses form factor. Additionally, for the hardware prototyping, we will utilize commercial off-the-shelf technology to enable GolfVision’s functionalities, while relying on Meta’s specialized engineering expertise for the final industrial integration and hardware refinement.

1.9 Design Specifications

From the primary interview with the golfers, we are able to generate the user needs and system requirements.

User Need (Based on JTBD/Value Proposition) Operational Requirement Performance Requirements
UR1. The golfer needs consolidated, real-time course intelligence (yardage, wind, elevation) delivered within their natural line of sight. OR1. The system shall automatically inform the golfer of the environmental data (wind, elevation, temperature, humidity) and calculate an adjusted "Playing Yardage" based on the data. PR1. The system shall provide localised environmental data with a minimum sampling frequency of one every 1 minute.
PR2. The system shall provide an accurate yardage calculation within 5 meters.
UR2. The golfer needs real-time gaze-based focus feedback to confirm mental readiness before a shot. OR2. The system shall capture and process key gaze metrics, including Quiet Eye duration and fixation patterns. PR3. The system must support accurate Gaze tracking with an accuracy of up to 2°.
PR4. The sensor must provide access to raw eye-tracking data for AI integration.
OR3. The system shall provide feedback based on the gaze metrics of the golfer’s focus and readiness before a shot. PR5. The system shall provide feedback on the golfer’s focus within 5 seconds of taking a shot.
UR3. The golfer needs the system to minimize cognitive load and external distraction during the pre-shot routine to maintain a consistent attentional state. OR4. The system shall automatically remove the information when the golfer gets ready to take a shot. PR6. The system shall remove all information within 1 second of detecting the golfer ready to take a shot.
OR5. The system must be lightweight and comfortable for use over an extended period. PR7. The weight of the system shall be less than 200g.
OR6. The hardware must be compatible with standard golf attire and designed for on-course use. PR8. The system shall be IP67 rated.
PR9. The system shall operate continuously for a minimum of 3 hours without charging.
UR4. The golfer needs an accessible, integrated assistant that combines gaze training and course intelligence to provide insights to improve execution instantly. OR7. The system shall provide autonomous, real-time club selection and execution recommendations based on fused Quiet Eye and Course data. PR10. The system shall provide recommended club selection and execution within 5 seconds of inquiry by the golfer.

2. Concept Development

Based on the system requirements mentioned in Section 1.9, we formed three major subsystems to enable our GolfVision: Hardware Subsystem, Application Subsystem and Backend Software Subsystem. The Hardware Subsystem will focus on collecting the gaze information from the user and relaying the information to the user. The Application Subsystem will focus on the user interface. Finally, the Backend Software Subsystem will focus on creating an AI caddie that will integrate the environmental data and gaze data to provide relevant insights that the golfers can use to improve their putting experience.

2.1 Design Selection for Information Relay

The first major design decision for the Hardware Subsystem was selecting the mode of feedback: Visual or Audio. We chose a visual Heads-Up Display (HUD) approach based on the comparative analysis below.

Design Pros Cons
Visual
  1. Information is always available
  2. Golfers can refer to the information quickly
  3. Information can be presented concisely
  1. Risk of visual obstruction or distraction during the critical execution phase
  2. Visibility concerns under direct, harsh sunlight
Audio
  1. Minimal interference with the golfer's vision
  2. Maintains gaze-action coupling by eliminating visual distraction during the swing
  1. Transient information that cannot be easily referred back to
  2. Not suitable for continuous monitoring of complex numerical data (e.g., detailed yardage tables)
  3. Potential for auditory distraction to the golfer or other players

The visual HUD was selected as the optimal relay method based on human factors principles of cognitive workload and attentional focus. Unlike audio feedback, which is transient and requires high levels of auditory processing that can conflict with a golfer's internal dialogue, the HUD integrates complex, multi-modal data—including shifting environmental variables and millisecond-level Quiet Eye metrics—directly into the user's natural line of sight. This design alignment fulfills UR3 by minimizing the cognitive load required to access and mentally integrate disparate data points. By placing actionable intelligence within the primary visual field, the system ensures that golfers can verify their attentional readiness and course conditions without disrupting the flow of their critical pre-shot routine or breaking gaze-action coupling.

2.2. Technology Selection for the Hardware Subsystem

Following the selection of a visual design for information relay, the next step was to select the specific hardware components for the display and the eye tracking system. The goal was to ensure a high sampling rate for accurate gaze measurement and a lightweight, bright display suitable for outdoor use.

Comparative Analysis of Hardware

The table below compares the leading options for the two critical components: the Eye Tracking Sensor and the Visual Display.

Components Option 1 Option 2 Option 3
Eye Tracking Module Pupil Labs Bare Metal [16]
  • Sampling Rate: 200Hz
  • Accuracy: 1.2°
  • Weight: 7.3g
  • Data Collected: Gaze point, head pose, pupillometry data, eye state, eye position, eye orientation, pupil diameter, fixation, blink
HoloLens 2 [17, 18]
  • Sampling Rate: up to 90Hz
  • Accuracy: 1.5°
  • Weight: 566g
  • Data Collected: Gaze
Tobii Pro Spark [19]
  • Sampling Rate: 60Hz
  • Accuracy: 0.45°
  • Weight: 97g
  • Data Collected: Gaze, fixation, Saccade, Blink
Visual Display Retrofitted glasses with a transparent OLED screen [20]
  • Weight: 13g
  • Power source: External power source
  • Display: Transparent OLED
HoloLens 2 [17]
  • Weight: 566g
  • Power source: In-built battery pack
  • Display: See-through holographic lenses
Xreal Air 2 Ultra [21]
  • Weight: 83g
  • Power source: Wired connection with the phone
  • Display: Micro OLED panel with brightness up to 500 nits

Eye Tracker Module: The Pupil Labs Bare Metal was selected. It provides the necessary high sampling rate (200Hz vs. HoloLens 2's 90Hz and Tobii's 60Hz), superior accuracy (1.2° vs. 1.5° for HoloLens 2), and crucial full access to raw eye-tracking data, which is essential for integration with our AI coach. The Tobii Pro Spark was dismissed due to its requirement for off-head placement (45cm away), making it unsuitable for mobile, outdoor golf use.

Visual Display: The Xreal Air 2 Ultra was chosen. Key selection criteria included outdoor functionality, spatial mapping, AR capabilities, and comfort. The Xreal Air 2 Ultra features a built-in screen, electrochromic lenses for direct sunlight visibility, and a lightweight design (83g vs. HoloLens 2's 566g). This lightweight form factor is critical for comfort over extended periods, as stated in PR7. The HoloLens 2 and transparent OLED options were rejected due to brightness issues in direct sunlight, the difficulty of integrating the Pupil Labs Bare Metal with the HoloLens 2's shape, and the HoloLens 2's weight.

2.3. Interfacing the Xreal Glasses with Bare Metal

Integrating the Pupil Labs Bare Metal eye-tracking module with the Xreal Air 2 Ultra display was a core engineering challenge necessary to create a cohesive prototype for real-time gaze tracking and HUD. This required overcoming several physical and optical constraints to ensure the system remained functional and comfortable. The integration effort focused on solving three key challenges:

  1. Accurate Positioning: Ensuring the Bare Metal was positioned at the correct eye level and angle to capture precise eye movement.
  2. Secure Attachment: Developing a snug attachment mechanism to prevent movement and sustain fast head movements during a golf swing.
  3. Display Visibility: Maintaining the user’s clear view of the Xreal glasses’ augmented reality display.

The iterative design process is presented in Section 3.1.

2.4. Logical Architecture

A high-level architecture diagram of our system is as shown. We were required to do hardware and software integrations to achieve the operational and performance requirements as set out by our design specifications.

3. Prototyping

Drawing from the design specifications and conceptual frameworks established in Section 2, the project proceeded to the development of a functional prototype. This phase was characterized by an iterative design approach across the hardware, application, and backend software layers to ensure system integration and performance.

3.1. Iterative Design of Hardware Layer

The hardware integration phase focused on securing the Pupil Labs sensor to the Xreal glasses while maintaining precise optical alignment and user comfort. Through five distinct design iterations, the mounting mechanism was refined to achieve high-accuracy gaze tracking (PR3) and an ergonomic, lightweight form factor (OR5, PR7) suitable for active on-course play.

Iteration 1: Proof of Concept Iteration 2: Rigid Integration Iteration 3: Front-Mounted Stability
Design The attachment arms secured the Bare Metal sensor by hooking around the sides of the glasses and provided necessary structural support, as the glasses' nose bridge had to be removed for mounting. In the second iteration, we implemented a slide-in mechanism designed to provide a more secure, rigid fit of the mounting arms to the sides of the glasses, effectively eliminating structural flexing. Additionally, the vertical positioning of the Bare Metal sensor was lowered to minimize visual occlusion of the camera module and improve the field of view. In the third iteration, we pivoted to a front-mounted attachment for the Bare Metal sensor. This structural shift eliminated the flexibility issues associated with the glasses' side arms in previous designs. By securing the eye-tracking module directly to the frontal frame, the prototype achieved superior stability and alignment, providing a more consistent and reliable gaze-tracking environment for the user.
Problems faced In the initial iteration, the mounting arm required additional length to wrap around the side of the Xreal glasses. However, this extra length introduced structural flexibility, causing the bracket to flex upwards. This displacement shifted the Bare Metal sensor above its intended height, resulting in the glasses' frame obstructing the eye-tracking camera's field of view. The design restricted the side arms from flexing beyond a certain angle, which significantly reduced the wearability and comfort for users with different head shapes. Two primary concerns persisted: minor visual interference from the front-mounted hardware and the exposure of sensitive electronics to environmental elements.
Iteration 4: Environmental Housing Iteration 5: Modular Locking System
Design The fourth iteration focused on protecting the critical electronic circuitry from rain and airborne debris common during on-course gameplay. We developed a custom, wing-like housing that follows the contour of the Xreal frame, providing a protective shell. For the final iteration, a two-part modular attachment was engineered. This design allows the protective housing to be securely snapped over the Xreal frame, while a separate, slide-in nose bridge component locks the entire assembly into place. This improved both structural rigidity and the ease of mounting the sensors for different users.
Problems faced The wing-shaped design, while aesthetically integrated, created significant friction during the assembly process. Physical interference between the custom housing and the Xreal frame required excessive force to latch the components together, posing a risk of structural damage. While the sliding mechanism can occasionally allow the nose bridge to detach at specific angles, this remains a marginal issue in practice. The natural mechanics of donning and removing the glasses consistently force the bridge toward the frame, effectively locking it in place and preventing accidental detachment during use.

3.2. Iterative Design of Application Layer

The application layer serves as the critical interface between hardware telemetry and user visualization. Developed using Unity, this layer enables the rendering of AR features and real-time data overlays (OR1).

Iteration 1: Backend Integration Iteration 2: HUD Visualization
Design Initial development established a reliable communication link between the Unity environment and backend servers, enabling the transmission and rendering of raw eye-tracking data. Hardware telemetry was successfully bridged to the AR environment, allowing for the first implementation of a floating Heads-Up Display (HUD) within the user's field of view.
Problem faced The application was limited to PC execution, preventing true mobile AR deployment on Android-based hardware required for on-course testing. The interface suffered from information overload. A lack of visual hierarchy and signifiers made it difficult for users to quickly process key data points.
Iteration 3: Modular Information Hierarchy Iteration 4: Context-Aware HUD Gating
Design Visual segmentation was introduced to organize data modules. By adding clear boundaries, golfers could more effectively distinguish between weather, yardage, and gaze metrics. To prevent distraction during execution, the HUD was locked above the primary eye level. The display is gated by head tilt, disappearing when the user looks down to address the ball.
Problem faced Persistent AR overlays were found to obstruct the golfer's view of the ball during the address and swing phases, negatively impacting focus.

3.3. Iterative Design of Backend Software Layer

The backend software architecture provides the computational logic required to fulfill operational requirements, divided into five specialized modules.

The backend software layer can be split into 5 main modules.

  1. Remote control module
  2. Eye tracking module
  3. AI caddie module
  4. Weather data module
  5. Yardage calculation module

3.3.1. Remote Control Module

This module facilitated manual data override for testing. Using a "Wizard of Oz" methodology, the team could simulate automated responses to validate the user experience prior to full algorithm automation.

Iteration 1: Manual Control Interface
Design

The control interface included the following diagnostic capabilities:

  1. Active player button to load player data based on which player is wearing the glasses.
  2. Distance input field and send to phone button for us to manually input the actual distance between the golfer and their target, and display it on the Xreal glasses connected to the phone.
  3. Trigger glasses recording button which allowed users to communicate with our AI module via speech.
  4. Trigger computer vision (QE) button which allowed us to extract the golfer fixation duration on ball at the time of shot.
Problems faced High operational overhead; the system required a dedicated human operator to manually trigger data updates and capture events, which is not scalable for independent use.

3.3.2 Eye Tracking Module

The Eye Tracking Module processes high-frequency gaze data to provide actionable feedback on attentional stability, specifically targeting the Quiet Eye (QE) duration.

The following iterations fulfilled OR2, OR3, PR3, PR4 and PR5 of our requirements.

Iteration 1: Real-Time Gaze Streaming Iteration 2: Automated Analytics Pipeline
Design Wizard of oz using pupil labs software. The initial proof of concept utilized Pupil Labs' native streaming to allow a third party to monitor gaze patterns and provide verbal coaching based on visual feedback. This iteration only allowed third parties to monitor the user’s real time gaze through a red circle which traced their gaze. We used the pupil lab’s remote streaming capabilities to capture performance manually, and advised golfers on the spot to improve their gaze patterns and fixations [22]. This iteration improved on the previous iteration by empowering golfers with access to more eye data metrics - pupil diameter, eye focus duration etc. A custom data pipeline was developed to automate the extraction of raw gaze metrics. This transition empowered the system to independently calculate QE duration and fixation patterns without manual interpretation.
Problems faced Heavy reliance on external observers to translate raw visual gaze markers into coaching advice for the golfer. The module lacked other eye data metrics such as eye focus duration on the ball which is crucial. While data was captured automatically, the relay of these insights to the golfer's HUD remained a manual bottleneck.
Iteration 3: Integrated Feedback Loop
Design The final iteration closed the feedback loop by delivering QE duration metrics directly to the HUD immediately following a shot, providing the golfer with instant objective focus data. This was done by showing users their fixation duration right after the shot. Not only that, the shot number and fixation duration would be sent to the AI module, where actionable insights on how to improve user fixation would be provided back to the user.
Problem faced

3.3.3. AI Caddie Module

The AI Caddie Module utilizes Large Language Models (LLMs) to provide contextualized recommendations and coaching based on environmental data and historical player performance.

The AI caddie module fulfilled OR7 and PR10 as part of our design specifications.

Iteration 1: General LLM Interaction Iteration 2: Context-Aware Personalization
Design Unity app working on the Windows interface Unity app working on the Xreal glasses The first implementation established basic communication between the HUD and the AI server, allowing the user to query a general golf knowledge base via text/voice. Data collection from Foresight Indoor golfing Data from Foresight indoor golfing. The module was upgraded to ingest player-specific club yardage reports. This allowed the AI to provide personalized club recommendations based on the user's actual hitting profile. Besides club selection, they were also given actionable insights based on golf course conditions as well as their eye focus duration. For testing purposes, we even added multiple profiles with different clubs that they own and the average distances.
Problem faced
  1. General responses provided limited tactical value as they did not account for the specific golfer's skill level or equipment.
  2. The AI did not factor in the unique context or statistical baselines of the individual golfers, overlooking crucial variables such as the fact that club yardages for a 9 iron versus a 6 iron vary drastically from player to player.
The UI/UX was a problem as it was too cluttered with information which was addressed with the other features on the later part.

3.3.4. Weather Data Module

The Weather Data Module automates the retrieval of environmental variables, fulfilling OR1 and PR1 to inform tactical decision-making.

We first focused on ensuring that the Xreal glasses could pull data from API and display crucial metrics such as temperature, wind, humidity etc.
Iteration 1: Bridging the weather API to the xreal glasses Iteration 2: Displaying golf-specific weather data
Design The initial iteration focused on establishing a reliable data pipeline to fetch real-time weather information from an external API and display it on the Xreal glasses. This involved integrating the API with the Unity application and ensuring that key meteorological metrics such as temperature, wind speed, and humidity were accurately rendered in the user's field of view. Based on our interviews we found that the top 3 concerns are wind speed, temperature and humidity. Hence we reduced the number of weather data shown to these 3.
  1. Wind speed with respect to where user is facing
  2. Temperature data
  3. Humidity data
Problem faced Early versions displayed excessive meteorological data, overwhelming the user and making it difficult to find actionable information quickly.

3.3.5 Yardage Calculator Module

This module calculates "Playing Yardage" by applying ball-flight physics to raw distance, factoring in wind, temperature, humidity, and altitude to eliminate mental estimation errors (OR1, PR2).

Iteration 1: Displaying calculated playing yardage
Design Calculated playing yardage was displayed directly on the HUD. Verification was performed by manually inputting rangefinder values to ensure algorithmic accuracy against professional estimates.
Problem faced Distance input was not yet autonomous; a technician was required to provide raw yardage values to the system for the calculation to trigger.

4. Testing & Validation

We verified our prototype by mapping it to the list of operational requirements and performance requirements in our design specifications, in line with the verification and validation framework [4]. A checklist of operational requirements and performance requirements was used to verify if the system built met the requirements that were set out.

4.1 Test Protocols

Two golfers participated in the on-course evaluation: a sponsored professional athlete with an official handicap of -2 and an amateur-level player with an approximate handicap of 20. This pairing was intentional, as it allowed the prototype to be assessed across distinctly different player profiles and skill levels.

To simulate a seamless, hands-free user experience, a pouch was used to store the phones on the golfers' persons, keeping the hardware out of sight during play. Additionally, the AI interaction was Wizard-of-Oz's - the golfer triggered the system by saying "Hey Meta," replicating the voice-activation flow of the Meta Ray-Ban glasses to ensure the testing experience closely mirrored the intended final product interaction.

Testing was conducted on Holes 3, 6, and 8 of the Mandai Executive Golf Course. These holes were specifically selected because they are all Par 3 holes, meaning the golfer's objective on each tee shot is to reach the green directly. This eliminates the variable of course management strategy. There is no need for the golfer to decide between laying up, shaping a shot around a dogleg, or targeting an intermediate position. Instead, the target is fixed (the pin), and the distance from white tee to green is known. This made Par 3 holes ideal for evaluating the system's core functionality since all key variables such as distance, target, and shot intent are clearly defined from the outset.

4.2 Data Collection & Observations

Requirement Description Status Evidence/ Remarks
PR1 The system shall provide localised environmental data with a minimum sampling frequency of one every 1 minute. Met The API provided updated localised weather data well within the 1 minute threshold.
PR2 The system shall provide an accurate yardage calculation within 5 metres. Met Verified by the professional golfer's independent cross-check during on-course play. His mental calculation, informed by years of competitive experience, aligned within 5 metres of the system's computed playing yardage. While this represents a single-user validation rather than a controlled measurement, it provides strong preliminary evidence given the golfer's expertise.
PR3 The system must support accurate gaze tracking with an accuracy of up to 2°. Met The Pupil Labs Bare Metal has a manufacturer-specified accuracy of 1.2°, which exceeds the 2° requirement. Standard calibration according to pupil labs guidelines was performed before each session.
PR4 The sensor must provide access to raw eye-tracking data for AI integration. Met The Pupil Labs Bare Metal provides full access to raw gaze data via its API, including gaze point, fixation events, pupil diameter, blink data, and head pose. This data was successfully extracted and piped into the AI coaching pipeline during the on-course evaluation.
PR5 The system shall provide feedback on the golfer's focus within 5 seconds of taking a shot. Met From QE data capture, to processing, to GPT-4 inference with feedback within 5 seconds during the Mandai evaluation.
PR6 The system shall remove all information within 1 second of detecting the golfer ready to take a shot. Not Met Gaze-triggered HUD removal was not implemented in the current prototype. This is identified as a key area for future development; implementation could leverage the existing Pupil Labs fixation data to detect the onset of the pre-shot routine and trigger HUD dismissal.
PR7 The weight of the system shall be less than 200g Met The overall device weighs less than 200g.
PR8 The system shall be IP67 rated. Not met Waterproofing and dust ingress protection were not addressed in the current prototype
PR9 The system shall operate continuously for a minimum of 3 hours without charging. Met The system operated for the full duration of the on-course evaluation of 3 hours without requiring a recharge, meeting the minimum operational endurance requirement.
PR10 The system shall provide recommended club selection and execution within 5 seconds of inquiry by the golfer. Met AI response latency was consistently under 5 seconds during the Mandai evaluation. Both testers noted that the response speed was practical for on-course use without disrupting their routine.
OR1 The system shall automatically integrate environmental data (wind, elevation, temperature, humidity) to calculate an adjusted "Playing Yardage." Met The system retrieved real-time weather data via the Openmeteo api and computed an adjusted playing yardage, factoring in wind speed, wind direction, temperature, and humidity. Wind-adjusted club recommendations directly influenced his club selection on at least one hole.
OR2 The system shall capture and process key gaze metrics, including Quiet Eye duration and fixation patterns. Met The Pupil Labs Bare Metal eye tracker successfully captured raw gaze data during the Mandai evaluation. The data pipeline processed this to extract Quiet Eye duration, which was then fed into the GPT-4o coaching model as contextual input for personalised feedback.
OR3 The system shall provide feedback based on the golfer's gaze metrics of focus and readiness before a shot. Met Quiet Eye duration was measured per shot and injected into the GPT-4o prompt context. The AI coach provided personalised feedback referencing the golfer's gaze behaviour, including whether their QE duration was within the optimal range established in the literature (2–3.5 seconds).
OR4 The system shall automatically remove information when the golfer gets ready to take a shot. Partially Met Auto-removal of the HUD was not implemented. However, the AR overlay was positioned in the upper field of view and did not track the golfer's downward gaze when addressing the ball. Both testers confirmed the overlay did not obstruct their view during setup and execution, achieving the intent of the requirement through spatial placement rather than automatic toggling.
OR5 The system must be lightweight and comfortable for use over an extended period. Partially met Testers preferred a wireless design. Testers reported the combined assembly felt front-heavy due to the forward-mounted sensor and 3D-printed bracket, and the tethered cable setup added bulk.
OR6 The hardware must be compatible with standard golf attire and designed for on-course use. Partially Met The glasses' form factor was compatible with standard golf caps — testers were able to hook the glasses onto the cap brim as they would with regular sunglasses. Depth perception was an issue for golfers when they took putts due to the nature of the HUD.
OR7 The system shall provide autonomous, real-time club selection and execution recommendations based on fused Quiet Eye and Course data. Met The AI coaching module combined environmental data (wind-adjusted playing yardage) with player-specific club distance profiles (loaded from historical Foresight launch monitor data) and Quiet Eye metrics to generate personalized club recommendations and execution tips.

Out of 17 requirements, 12 were fully met, 3 were partially met, and 2 were not met. The two unmet requirements, PR6 (auto-HUD removal) and PR8 (IP67 rating), were bound by time and hardware constraints rather than fundamental technical limitations. Importantly, every requirement tied to our primary value proposition of real-time Quiet Eye feedback and AI-powered coaching was fully met during the on-course evaluation.

5. Results & Discussion

5.1 Analysis of Results

5.1.1 Quiet Eye Training (QET) Validation

To validate the QET component of GolfVision, testing was conducted across two phases: a controlled putting experiment (Phase 1) and an on-course system evaluation at Mandai Executive Golf Course (Phase 2).

Phase 1: Controlled Experimental Validation

Controlled 8ft putting trials established that QET intervention produced statistically meaningful improvements in proximal performance metrics. Precision and distance control improved by 10-15cm. This divergence validated the necessity of GolfVision as a continuous, on-course feedback architecture rather than a discrete coaching intervention, directly addressing the requirement for real-time gaze-action coupling stabilization.

As seen in the figure below the introduction of QET helped the golfer to putt nearer to the hole, which is optimal as golfers usually aim to get the ball as close to the hole as possible in order to finish it off on the second putt.

Phase 2: On-Course System Integration (Mandai)

On-course deployment confirmed the system's ability to extract and process high-frequency QE data via the Pupil Labs module within the required 5-second latency threshold (PR5). Gaze patterns aligned with skill-differentiated benchmarks (2.0-3.5s for the professional; ~1.5s for the amateur), with the amateur exhibiting an adaptive upward trend in fixation duration following real-time feedback. This serves as a functional proof-of-concept for the primary value proposition: transforming an internal cognitive state into a quantifiable, actionable HUD metric to improve execution consistency during live play.

5.1.2 Playing Yardage Calculator — Accuracy and Decision Impact

During the on-course evaluation across the three selected Par 3 holes, the system dynamically sourced localized environmental parameters via the OpenMeteo API to compute an adjusted "Playing Yardage," factoring in wind velocity, direction relative to the golfer, temperature, and humidity. Baseline raw distances were obtained using the hardcoded values of the holes from the mandai golf data sheet. Crucially, the final computed Adjusted Playing Yardage values across all three holes closely approximated the professional golfer's experienced mental assessment, providing strong initial validation for the algorithm’s real-world accuracy.

Hole Raw Yardage Wind wrt to the player Temp Humidity Adjusted Playing Yardage (UNITY APP) Pro's Mental Estimate
3 110 5 Mph (NE) 30 84% 112 115
6 83 5 Mph (SE) 30 84% 80 75
8 90 5 Mph (NE) 30 84% 94 100

5.1.3 Hardware Ergonomics — Human Factors Analysis

While the system weight met the 200g target (PR7), testers reported the assembly felt front-heavy due to the forward-mounted sensor and 3D-printed bracket. This created pressure points and fatigue over time. Future designs should redistribute mass, potentially by moving companion electronics to the rear of the head strap.

Testers noted depth perception issues during putting, where objects appeared closer than their actual distance. This results from the Xreal display's fixed focal plane conflicting with the 1–2 meter convergence required for putting—a conflict less noticeable at longer distances. Future designs must mitigate this optical constraint by dynamically adjusting the virtual image plane or disabling the HUD during the putting routine.

Both testers preferred a wireless system, as the current tethered setup felt bulky and restricted movement (OR5 partially met). While the USB-C connection was necessary for high-bandwidth raw data streaming from the Pupil Labs sensor, moving to a wireless solution introduces challenges in balancing power consumption with the low latency required for real-time feedback (PR5). Future development will focus on integrated, power-efficient wireless modules.

5.1.4 AI Caddie Module — Recommendation Quality and Decision Impact

During the Mandai evaluation, the AI Caddie's wind-adjusted club recommendation directly caused the professional golfer to change his club selection on at least one hole — a moment captured on video. This demonstrates that the system's output was actionable: an experienced professional trusted the AI's recommendation over his initial instinct, indicating that the fused data (wind, playing yardage, and personal club profiles) provided value beyond what even a trained golfer could mentally compute.

A limitation was observed in context sensitivity. The professional golfer expressed interest in receiving Quiet Eye feedback specifically for putting, but considered it less relevant for full shots. The current system does not distinguish between shot types when delivering gaze-based coaching — it provides the same feedback for putts, approach shots, and tee shots. This lack of shot-context awareness reduces personalisation and represents a gap relative to the nuanced, situation-dependent advice a human caddie would provide.

5.1.5 Weather Data Module and HUD User Experience

The professional golfer confirmed that all three displayed weather variables — temperature, humidity, and wind speed — were relevant to his decision-making, validating the design decision in Section 3.3.4 to limit the weather display to these three metrics.

The pro golfer noted that having real time wind data with respect to his gaze direction was very helpful in helping him decide how to take his shot.

A practical usability finding emerged regarding units of measurement. The professional expressed a preference for mixed units depending on context: wind speed in miles per hour, distance in metres for approach shots, but feet for close-range putting (e.g., "a 10-foot putt" versus "a 20-metre approach"). The current system uses fixed units, requiring the golfer to mentally convert — a minor but unnecessary cognitive overhead that works against the intent of UR3.

Testers also noted that while they appreciated the current display layout and information hierarchy, the overall UI required further visual refinement before it would be suitable for a broader user cohort.

5.1.6 Spatial Mapping and Feature Requests

Both testers provided unsolicited feedback on features beyond the current prototype scope. The most requested feature was a minimap-style hazard overlay — described by the testers as a "video game-like map" - displaying straight-line distances to bunker edges (front and back) and green boundaries. Testers emphasised that understanding the size of their landing zone relative to surrounding hazards was critical to course management, and that an in-HUD overlay would eliminate the need for separate course maps.

Additional requests included green scanning to overlay the intended shot line and ball tracking post-impact. While outside the current project scope, this feedback validates the broader market demand identified in Section 1.1.2 and informs the Stanford development roadmap (Section 6.2).

5.2 System-Level Validation Against User Requirements

Sections 4.2 and 5.1 verified that individual components met their specifications and analysed their on-course performance. This section maps those findings back to the four User Requirements (UR1–UR4) from Section 1.9, assessing whether the integrated system addresses the golfer's actual needs as identified through the JTBD framework.

User Need Linked Reqs Key Evidence Status
UR1. Consolidated, real-time course intelligence within the golfer's line of sight OR1, PR1, PR2
  • Adjusted yardage within 2–10 yards of pro's estimate across 3 holes (Section 5.1.2)
  • Pro confirmed wind, temperature, humidity were all decision-relevant
  • All data consolidated on HUD - no secondary device needed
  • Addresses Think Branch root cause (Section 1.4.1)
Validated
UR2. Real-time gaze-based focus feedback to confirm mental readiness OR2, OR3, PR3, PR4, PR5
  • Phase 1: putting precision improved 10-15 cm with QE intervention
  • Phase 2: amateur QE duration trended upward after real-time feedback
  • Gaze data captured at 1.2° accuracy; feedback within 5s threshold
  • Addresses Feel Branch root cause (Section 1.4.1)
Validated
UR3. Minimise cognitive load and distraction during pre-shot routine OR4, OR5, OR6, PR6, PR7, PR8, PR9
  • HUD spatial anchoring avoided gaze obstruction during shot execution
  • Weight under 200g; 3-hour battery life; golf cap compatible
  • Gaps: Auto-HUD removal not implemented (PR6); front-heavy discomfort (OR5); depth perception error during putting (OR6); IP67 not addressed (PR8); fixed units add cognitive overhead
Partially Validated
UR4. Accessible, integrated assistant combining gaze training and course intelligence OR7, PR10
  • AI Caddie fused wind, club profiles, and QE metrics into recommendations
  • Pro changed club selection based on AI advice (video evidence)
  • Response latency consistently under 5s
  • Gap: No shot-type distinction in QE feedback delivery
Validated

Three of four user requirements (UR1, UR2, UR4) were validated. UR3 was partially validated - notably, both fully unmet performance requirements (PR6, PR8) fall under this user need, confirming that minimising cognitive load remains the primary development gap. The validated requirements correspond to the system's core value proposition (course intelligence and gaze coaching), while UR3 relates to ergonomic and interaction design - suggesting the concept is sound but the form factor needs further iteration.

6. Recommendations & Reflection

The development of GolfVision posed significant engineering challenges, requiring the seamless integration of high-frequency eye-tracking hardware with real-time AR visualization and AI-driven analytics. This project demonstrates a high-value application for smart glass technology, particularly as Meta integrates eye-tracking sensors into future consumer hardware. By providing golfers with previously invisible environmental data and elite-level gaze training, GolfVision transforms a standard accessory into a powerful performance-enhancing tool.

6.1 Identification of Shortcomings

Software Compatibility and Stability: The most significant challenge encountered was the shifting software ecosystem for the Xreal Air 2 Ultra. During development, Xreal transitioned from the Nebula OS to AndroidXR, deprecating support for legacy Unity integrations. This transition required sourcing third-party middleware to run the custom Unity application, which introduced periodic system glitches and limited the ability to leverage native features like high-fidelity hand tracking, ultimately impacting the fluidity of the user interface.

Hardware Complexity and Portability: The current prototype requires a fragmented ecosystem of five distinct devices to function—including two smartphones, a laptop for backend processing, the eye-tracking module, and the AR glasses. This multi-device dependency creates significant friction for on-course play, increasing bulk and reducing the portability expected of a consumer-grade sports accessory. A streamlined, wireless architecture that integrates processing into a single companion device is necessary for market viability.

Optical Convergence and Depth Perception: Testing revealed a persistent depth perception error caused by the prism optics of the AR display. Golfers reported difficulty gauging the exact spatial relationship between the club head and the ball during the putting stroke, as the virtual overlay's fixed focal plane competed with the close-range physical target. This optical conflict necessitates frequent recalibration and suggests that the design of the AR glasses must maintain the natural visual focus of the golfer.

6.2 Future works to be done at Stanford

While the current prototype demonstrates core functionality, several advanced features remain for future development in collaboration with Stanford University. These include hazard detection, stress management, and course management strategy—capabilities intended to transition the system from Par 3 evaluation to comprehensive full-course decision support. This evolution aims to provide a robust digital alternative to traditional caddie services by addressing a broader spectrum of on-course challenges. Due to the limited duration of initial testing in Singapore, a subsequent phase of rigorous user validation is scheduled in the United States. This phase will involve a larger cohort of athletes to gather diverse technical feedback and refine the system for global market requirements.

Based on hardware ergonomics feedback (Section 5.1.3), future work will also focus on redistributing the mass of the glasses to address the front-heavy feel. Additionally, we will explore ways to dynamically disable the lens or move the HUD out of the way to mitigate the optical convergence error observed during putting.

Other sports application

7. Impact & Visibility (A9)

7.1 Cross-Sport Applicability

GolfVision represents a pioneering application of real-time gaze-action coupling training in a mobile sporting environment. The underlying architecture—fusing high-frequency eye-tracking data with contextual environment analytics—possesses significant transferability to any target-oriented sport where attentional control is a primary performance determinant. In disciplines such as hockey, archery, and soccer, the system can provide coaches with a precise, objective quantification of an athlete’s visual search strategy and fixation stability. By identifying suboptimal gaze patterns during high-pressure execution, the platform enables targeted interventions to enhance team dynamics and individual precision.

To validate the transferability, testing was done with field hockey goalkeepers from the Singapore Women’s National Hockey Team as well as a Premier League goalkeeper. The same Pupil Labs eye tracker used on GolfVision was deployed and we found that elite goal keepers have a longer fixation duration and more structured gaze patterns compared to non-elite ones. A follow-up interview with the Singapore Women’s goalkeeping coach confirmed that eye tracking data would be highly valuable for coaching as it will allow coaches to quantify goalkeeper attentional focus which is currently assessed through subjective observation. These findings suggest that the sensing and feedback architecture we developed for golf can be applied on sports such as hockey as well.

7.2 Non-Sport Training and Education

Beyond the athletic arena, the technology’s capacity to track and guide visual focus offers transformative potential for high-stakes professional training. Empirical research has already demonstrated that Quiet Eye training significantly improves the performance of junior surgeons under pressure, particularly in precision tasks like surgical knot-tying [23]. By adapting this system to educational contexts, we can accelerate the motor-skill acquisition phase for complex manual tasks, providing learners with real-time feedback on where and when to focus their visual attention to mirror expert-level performance.

7.3 Market Potential and Industrial Scale

The integration of advanced biometric tracking within a lightweight AR form factor positions GolfVision at the forefront of the global sports technology market, currently valued at approximately USD $93.81 billion [24]. As hardware manufacturers like Meta continue to embed eye-tracking sensors into consumer-grade smart glasses, the modular software framework developed in this project can be scaled across various industrial sectors to enhance productivity, safety, and operational precision.

8. References

[1] “2025 Golf Courses and Country Clubs Industry Market Research Report,” Kentley Insights. [Online]. Available: https://www.kentleyinsights.com/golf-courses-and-country-clubs-industry-market-research-report/. [Accessed: 03-Apr-2026].

[2] MordorIntel, “Golf Equipment Market Size, Share & Industry Report, 2031,” Mordor Intelligence. [Online]. Available: https://www.mordorintelligence.com/industry-reports/golf-equipment-market. [Accessed: 03-Apr-2026].

[3] “Golf GPS Watches Market Size & Industry Growth 2030,” FutureDataStats. [Online]. Available: https://www.futuredatastats.com/golf-gps-watches-market. [Accessed: 03-Apr-2026].

[4] R. Camacho, “Requirements Traceability Matrix Templates & Examples,” Parasoft, Apr. 9, 2020. [Online]. Available: https://www.parasoft.com/blog/requirements-management-and-the-traceability-matrix/. [Accessed: 31-Mar-2026].

[5] J. N. Vickers, “Neuroscience of the Quiet Eye in Golf Putting,” International Journal of Golf Science, vol. 1, no. 1, pp. 2–19, 2012.

[6] S. J. Vine, L. J. Moore, and M. R. Wilson, “Quiet eye training facilitates competitive putting performance in elite golfers,” Frontiers in Psychology, vol. 2, article 8, 2011.

[7] Rick Shiels Golf, “THE PERFECT GOLF PRE SHOT ROUTINE = More Consistency,” YouTube, Feb. 22, 2021. [Online]. Available: https://www.youtube.com/watch?v=kdR6aF2CxtQ. [Accessed: 20-Mar-2026].

[8] The Scoring Method, “Proven Pre-Shot Routine to BREAK 80!,” YouTube, Apr. 4, 2024. [Online]. Available: https://www.youtube.com/watch?v=aNQBTKTZMes. [Accessed: 20-Mar-2026].

[9] A. W. Ulwick, Jobs to Be Done: Theory to Practice. Houston, TX: Idea Bite Press, 2016. [Online]. Available: https://jobs-to-be-done-book.com/thank-you/.

[10] J. N. Vickers, Perception, Cognition, and Decision Training: The Quiet Eye in Action. Champaign, IL: Human Kinetics, 2007.

[11] Golf Monthly, “How to Calculate Distance in the Wind,” GolfMonthly.com, 2025.

[12] S. Malik et al., “Effects of wind on ball flight,” as cited in Golf Science Journal, 2018.

[13] “Trust in Distance Measuring Devices (DMDs) Automation in Golf,” International Journal of Golf Science, 2025.

[14] “Don’t Underestimate the Demand for Caddies,” Golf Digest, 2025.

[15] “The effect of quiet eye training on golf putting performance in pressure situation,” Scientific Reports, vol. 14, article 55716, 2024.

[16] “Neon - Technical Specifications - Neon eye tracking module and frames,” Pupil-labs.com. [Online]. Available: https://pupil-labs.com/products/neon/specs. [Accessed: 17-Mar-2026].

[17] MSBryan, “HoloLens 2 hardware,” Microsoft.com. [Online]. Available: https://learn.microsoft.com/en-us/hololens/hololens2-hardware. [Accessed: 17-Mar-2026].

[18] MSBryan, “Eye tracking overview - Mixed Reality,” Microsoft.com. [Online]. Available: https://learn.microsoft.com/en-us/windows/mixed-reality/design/eye-tracking. [Accessed: 17-Mar-2026].

[19] Tobii.com. [Online]. Available: https://go.tobii.com/tobii-pro-spark-brochure. [Accessed: 01-Apr-2026].

[20] “Blue 1.51" Transparent OLED Display 128x56 Controller SSD1309 SPI,” Buydisplay.com. [Online]. Available: https://www.buydisplay.com/blue-1-51-inch-transparent-oled-display-128x56-controller-ssd1309-spi. [Accessed: 02-Apr-2026].

[21] “XREAL - Building Augmented Reality for Everyone,” XREAL Ltd. [Online]. Available: https://www.xreal.com/. [Accessed: 17-Mar-2026].

[22] K. Dalton, “The Vision Strategy of Golf Putting,” Aston Publications Explorer (Aston University), Jan. 1, 2013. [Online]. Available: https://doi.org/10.48780/publications.aston.ac.uk.00019543. [Accessed: 10-Jan-2026].

[23] J. Causer, J. N. Vickers, R. Snelgrove, G. Arsenault, and A. Harvey, “Performing under pressure: quiet eye training improves surgical knot-tying performance,” Surgery, vol. 156, no. 5, pp. 1089–1096, 2014.

[24] “Wearable Devices Technology in Sport Market - Companies & Trends,” Mordor Intelligence. [Online]. Available: https://www.mordorintelligence.com/industry-reports/wearable-devices-in-sports-market. [Accessed: 02-Apr-2026].

Annex A

Jobs to be done steps

THINK BOX

  1. Figure out the lie of the ball
  2. Decide the direction I want to be aiming
  3. Determine the Ground Yardage
  4. Find the wind Direction
  5. Determine the Elevation
  6. Determine the Playing Yardage
  7. Decide how to play the shot
  8. Decide what club to hit with
  9. Decide the shot shape
  10. Visualise the golf shot
  11. Keeping yourself emotionally steady before addressing the ball
  12. Get comfortable

FEEL BOX

  1. Feel the swing, I want to swing
  2. Get into a stance
  3. Focus on the ball
  4. Just react and pull the trigger

PLAY BOX

  1. Observe the trajectory of the ball
  2. Reflect on the previous shot
  3. Proceed to find the ball
  4. Just react and pull the trigger

Comparison of 4 Sports Vision Techniques

Training method Strengths Limitations
Quiet Eye Training Strongest golf-specific evidence QE-trained golfers made 1.9 fewer putts compared to pre-training QE duration is a good predictor of variance of putting performance [m5] Showed to maintain performance under stress [m10] Without training, golfers might fall back below threshold
Stroboscopic Visual Training Improves short-term visual memory (up to 24 hours) Enhances motion perception and transient attention in labs No golf specific studies Requires opaque-lens flicker glasses that periodically block vision, incompatible with our application of on-course play More beneficial to sports with interruptions that require tracking of faster movements No reported benefit on stress
Perceptual-cognitive Training Improves multi object tracking No golf specific evidence, limited largely to hockey, soccer and basketball Requires screen-based set-up, not suitable for on-course play No reported benefits on stress
Simulation Training Simulated pressure scenarios provide exposure training Designed for practice settings, not for actual play