GazeProphetV2: Head-Movement-Based Gaze Prediction Enabling Efficient Foveated Rendering on Mobile VR

arXiv — cs.CVWednesday, November 26, 2025 at 5:00:00 AM
  • The introduction of GazeProphetV2 marks a significant advancement in gaze prediction for virtual reality (VR) environments, utilizing a multimodal approach that integrates head movement data, gaze history, and visual scene information. This method employs a gated fusion mechanism with cross-modal attention to enhance predictive accuracy across multiple future frames, achieving a validation accuracy of 93.1% across 22 VR scenes with over 5.3 million gaze samples.
  • This development is crucial for optimizing rendering processes in mobile VR, as accurate gaze prediction can lead to more efficient foveated rendering techniques. By improving how VR systems anticipate user focus, GazeProphetV2 can enhance user experience and reduce computational load, making VR applications more accessible and responsive.
  • The advancement in gaze prediction technology reflects a broader trend in AI and VR, where integrating various data modalities is becoming essential for improving user interaction and experience. Similar innovations, such as those focusing on egocentric visual span and hands-free input methods, highlight the ongoing exploration of how AI can enhance human-computer interaction across different contexts.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
Restora-Flow: Mask-Guided Image Restoration with Flow Matching
PositiveArtificial Intelligence
Restora-Flow has been introduced as a training-free method for image restoration that utilizes flow matching sampling guided by a degradation mask. This innovative approach aims to enhance the quality of image restoration tasks such as inpainting, super-resolution, and denoising while addressing the long processing times and over-smoothing issues faced by existing methods.
RobustMerge: Parameter-Efficient Model Merging for MLLMs with Direction Robustness
PositiveArtificial Intelligence
RobustMerge has been introduced as a parameter-efficient model merging method designed for multi-task learning in machine learning language models (MLLMs), emphasizing direction robustness during the merging process. This approach addresses the challenges of merging expert models without data leakage, which has become increasingly important as model sizes and data complexity grow.
EmoFeedback$^2$: Reinforcement of Continuous Emotional Image Generation via LVLM-based Reward and Textual Feedback
PositiveArtificial Intelligence
The recent introduction of EmoFeedback$^2$ aims to enhance continuous emotional image generation (C-EICG) by utilizing a large vision-language model (LVLM) to provide reward and textual feedback, addressing the limitations of existing methods that struggle with emotional continuity and fidelity. This paradigm allows for better alignment of generated images with user emotional descriptions.
From Inpainting to Layer Decomposition: Repurposing Generative Inpainting Models for Image Layer Decomposition
PositiveArtificial Intelligence
A new study has introduced a diffusion-based inpainting model adapted for image layer decomposition, addressing the challenges of separating images into distinct layers for independent editing. This model employs lightweight finetuning and a multi-modal context fusion module to enhance detail preservation in the latent space, achieving superior results in object removal and occlusion recovery using a synthetic dataset.
CaptionQA: Is Your Caption as Useful as the Image Itself?
PositiveArtificial Intelligence
A new benchmark called CaptionQA has been introduced to evaluate the utility of model-generated captions in supporting downstream tasks across various domains, including Natural, Document, E-commerce, and Embodied AI. This benchmark consists of 33,027 annotated multiple-choice questions that require visual information to answer, aiming to assess whether captions can effectively replace images in multimodal systems.
Structure-Aware Prototype Guided Trusted Multi-View Classification
PositiveArtificial Intelligence
A novel framework for Trustworthy Multi-View Classification (TMVC) has been proposed, addressing the challenges of reliable decision-making in scenarios with heterogeneous and conflicting multi-source information. This framework introduces prototypes to represent neighbor structures of each view, simplifying the learning of intra-view relations and enhancing consistency across inter-view relationships.
PG-ControlNet: A Physics-Guided ControlNet for Generative Spatially Varying Image Deblurring
PositiveArtificial Intelligence
PG-ControlNet has been introduced as a novel framework for spatially varying image deblurring, addressing the challenges posed by complex motion and noise. This approach reconciles model-based deep unrolling methods with generative models, capturing minute variations in degradation patterns through a dense continuum of high-dimensional compressed kernels.
Long-Term Alzheimers Disease Prediction: A Novel Image Generation Method Using Temporal Parameter Estimation with Normal Inverse Gamma Distribution on Uneven Time Series
PositiveArtificial Intelligence
A novel image generation method has been developed for long-term prediction of Alzheimer's Disease (AD), utilizing a temporal parameter estimation model based on the Normal Inverse Gamma Distribution. This approach addresses challenges in maintaining disease-related characteristics in sequential data with irregular time intervals, allowing for the generation of intermediate and future brain images to aid in diagnosis.