FlatTrack: Eye-tracking with ultra-thin lensless cameras
Abstract
Existing eye trackers use cameras based on thick compound optical elements, necessitating the cameras to be placed at focusing distance from the eyes. This results in the overall bulk of wearable eye trackers, especially for augmented and virtual reality (AR/VR) headsets. We overcome this limitation by building a compact flat eye gaze tracker using mask-based lensless cameras. These cameras, in combination with co-designed lightweight deep neural network algorithm, can be placed in extreme close proximity to the eye, within the eyeglasses frame, resulting in ultra-flat and lightweight eye gaze tracker system. We collect a large dataset of near-eye lensless camera measurements along with their calibrated gaze directions for training the gaze tracking network. Through real and simulation experiments, we show that the proposed gaze tracking system performs on par with conventional lens-based trackers while maintaining a significantly flatter and more compact form-factor. Moreover, our gaze regressor boasts real-time (125 fps) performance for gaze tracking.
![[Uncaptioned image]](https://cdn.awesomepapers.org/papers/142255c0-d6b3-4ab1-830f-faf354762ad9/firstfig2.png)
1 Introduction
Eye tracking plays a crucial role in Augmented and Virtual Reality (AR/VR) systems, whether it be augmenting the realism and interactivity of the immersive experience or rendering efficient display imagery. For example, by precisely tracking where the user is looking at any given time, eye tracking facilitates more natural and intuitive interactions, such as gaze-based interfaces for selection and control, which can enhance the user engagement with the virtual environment without the need for physical controllers [20, 24]. Eye tracking also facilitates optimized usage of system resources by rendering high-resolution graphics only in the regions of the visual field where the user’s gaze is well-focused, and there is less detail in the periphery, a technique called foveated rendering that is shown to dramatically improve the display performance[12, 1, 19, 9]. Apart from improving the overall graphical performance through selective on-demand rendering, foveated rendering significantly reduces computational load, allowing for smoother and more immersive AR/VR experiences.
Existing eye-tracking technologies (Tobii Pro, Pupil Labs) face several significant challenges that impact their practicality across various applications despite their rapid developments in recent years. Apart from the need for more accuracy and reliability, the physical size of eye-tracking units often limits their integration into wearable AR/VR headsets, posing significant design issues. Specifically, the optical components like lenses, filters and their housing directly affect the overall size of the device. The structure used to mount the sensors and associated electronics, including the enclosure that protects these components, further adds to the bulk of the system. Furthermore, privacy concerns also arise as eye tracking can reveal sensitive personal information. As demonstrated in earlier works, eye images can possibly be misused to reconstruct the scenes being observed, which pose significant threats and privacy concerns[2]. Finally, eye-trackers require low latency to ensure responsive and immersive user experiences in interactive applications. These obstacles necessitate continuous advancements in hardware and software to enhance the versatility and user-friendliness of eye-tracking technology. Hence, we propose a lensless gaze-tracking prototype that has a small form factor, inexpensive, lightweight, and privacy-preserving[14, 7].
In lensless imaging, the focusing lens is replaced with a thin, lightweight, and potentially inexpensive optical encoder, resulting in flat and lightweight miniaturized cameras. A lensless camera uses advanced computational reconstruction algorithms to recover the scene from captured sensor measurements, which no longer resemble the imaged scene [7, 4, 5]. However, with a well-designed encoder, the measurements contain sufficient information to recover an image of the scene using post-processing algorithms that can demultiplex the sensor measurements to reconstruct a sharp image of the scene. While digital post-processing of images is standard for tasks such as distortion correction, synthetic depth of field, and denoising, lensless cameras are fundamentally different in that the post-processing is a part of the imaging system design where the optical imaging hardware, and the algorithmic software is designed together. Moreover, since lensless cameras encode information indirectly in the measurement, then computationally extract it by solving an inverse problem, they also provide promise for enabling privacy in the eye tracking measurements[14]. Additionally, lensless cameras inherently capture the 3D information, which ensures a more accurate prediction[6].
In the current work, we employ a lensless camera design accompanied by a lightweight convolutional neural network to achieve compact eye tracking. Specifically, we use a Near-Infrared (NIR) PhlatCam[7] to develop an ultra-thin near-eye tracker. Using our system, we collect the first-of-its-kind large dataset of nearly 20000 paired lensless captures and calibrated gaze vectors. We propose a two-stage approach for estimating gaze from lensless capture and evaluate the efficacy of such an approach on our proposed dataset. Our experiments show that lensless cameras allow us to build compact gaze trackers with high-fidelity gaze estimates. We plan to release the dataset and gaze tracking algorithm upon acceptance of the paper.
In summary, our contributions in this paper are:
-
•
We build a system using NIR PhlatCam prototype for gaze tracking. Existing works on lensless gaze estimation have mostly shown results on simulated data that don’t capture the real data’s non-idealities.
-
•
We collect a large dataset of 20475 paired lensless captures and gaze directions from a near-eye setup. To the best of our knowledge, we are the first to collect such a dataset for lensless gaze tracking.
-
•
We build a two-stage gaze estimator to evaluate the efficacy of lensless gaze estimation and demonstrate that lensless cameras allow high-fidelity gaze recovery with a much smaller form factor. Our gaze regressor performs real-time (125 fps) gaze estimation.
2 Related Works
Mask-based Lensless Imaging. Mask-based lensless cameras replace the lens of conventional cameras with a thin optical mask. This mask has the ability to modulate the incoming light and can be placed extremely close to the sensor, leading to a flat form factor. Moreover, the ultra-small mask-to-sensor distance allows these cameras to have a large depth of field, thereby making it possible to image from a very close range - a property that will allow lensless camera-based gaze trackers to have a much smaller form factor. Multiple lensless cameras have been developed in the literature. FlatCam [5] is a lensless camera that places a separable coded amplitude mask above a bare sensor array to enable a thin and flat form-factor imaging device, which can simulate a conventional camera by reconstructing conventional images from coded measurements. DiffuserCam [4] and PhlatCam [7] replace the coded amplitude mask from FlatCam [5] with a coded phase mask for improved light efficiency and reconstruction quality. Spectral DiffuserCam[17] exploits the multiplexing ability of lensless imagers to do hyper-spectral imaging. The primary differences in the above design lie in the pattern of the mask used and the mask-to-sensor distance. In this work, we will primarily focus on the PhlatCam [7] lensless camera. Previous works like [6, 14] have shown the application of this camera design for 3D imaging and optical encryption. Previously, FlatCam [5] was shown in [23] to perform at par with lensed cameras for gaze-tracking. However, the authors only reported the performance on simulated data, and ensuring the same performance on real data is hard. We take a step further and develop the first-of-its-kind real lensless gaze dataset using PhlatCam. We run extensive experiments on this real dataset to highlight the efficacy of our system. We plan to release this dataset upon acceptance of the paper.
Gaze Tracking Datasets. Since learning-based gaze trackers need large datasets, many such datasets have been proposed in the literature for both remote and near-eye gaze tracking. Among the remote gaze tracking datasets, MPII-Gaze[25] contains 214K face images captured using a laptop camera from 15 subjects along with their gaze vectors. Columbia Gaze[22] contains 5880 images from 56 people and focused on gaze locking. GazeCapture[16] contains 2.5M frames collected from over 1450 people. Among the near-eye gaze tracking datasets, NVGaze[15] contains both real and synthetic data of 2M infrared images of eyes at 1280 × 960 resolution. OpenEDS[18] contains more than 500K eye-images with the corresponding 3D gaze vectors collected from 80 participants. Recently, [3] proposed a dataset captured using a DAVIS event camera. This dataset contains both event and grayscale frames along with the corresponding gaze. Although the existing datasets contain a wide distribution of eye images and environment, none of them are suited for lensless cameras, which have a very distinct imaging model. Thus, there’s an urgent need for a dataset developed specifically for lensless cameras.
3 Lensless Imaging Background
In a mask-based lensless camera[5, 7] like the one used in this work, the conventional lens is replaced by a thin optical mask placed at some distance from the sensor that modulates the incoming light. For a scene , the measurement recorded by a sufficiently large sensor is given by:
(1) |
where is the full-size convolutional operator (no cropping due to finite sensor size), is the point spread function (PSF), and is additive noise. The PSF is the response of the camera to a point source. The optical mask can be implemented using an amplitude mask that attenuates the incoming light [5] or a phase mask that modulates based on diffraction [7]. In this work, we use a phase mask. The lensless PSF is a function of the mask pattern, wavelength of light, and the mask-sensor distance. Our PhlatCam uses a phase mask designed for NIR wavelength of 700nm and is placed slightly less than 1.5mm from the sensor.
Due to the lack of a focusing element like a lens, a lensless camera capture is a multiplexed representation of the scene. To recover the original scene, one needs to computationally solve an inverse problem of estimating the scene given the measurement and PSF , which is challenging due to properties like large PSF support (which can be bigger than the scene projection ), finite sensor size, noise, etc. Although there exist different methods of lensless scene reconstruction in the literature, we will primarily focus on the two most popular methods – FlatNet[13] and Wiener deconvolution. Please see Fig. 2 for the lensless imaging and reconstruction process.

4 FlatTrack:Lensless Gaze Estimation Dataset

We use an NIR-PhlatCam designed for 700nm to capture the FlatTrack Gaze Dataset. We collected a comprehensive dataset comprising 20,475 lensless measurements meticulously gathered from 13 distinct subjects. The importance of user-specific data lies in the practical application of eye-tracking systems, where such systems are typically fine-tuned or calibrated for individual users using a small amount of personalized data. This calibration process is essential in real-world scenarios, such as AR/VR applications, where precise gaze estimation requires personalization.
Each subject participated in a series of experiments in which they were tasked with focusing on stimuli presented on a computer screen. The stimuli moved along a 15x15 grid pattern as in Fig. 3 within a 1080x1920 pixel area of the monitor (DELL S2421HN), and their eye images were captured by the lensless camera. To ensure accuracy, subjects were seated 50 cm away from the screen, and before each round of data collection, calibration was performed to align the center of the eye with the origin on the screen. This process was repeated for each round of data collection for every subject. We also ensured that the distance between the eye of the subject and PhlatCam is approximately 4cm. We collected 4 to 5 rounds of data for each subject, each round comprising either 225 or 675 images (multiple measurements) corresponding to the 15 x 15 grid pattern. For each measurement, we also note the gaze point in pixels on the monitor and convert it to a unit vector in the gaze direction using the known monitor distance.
The distance between two consecutive stimulus points along the x-direction in the grid pattern measures 121.3 pixels, while along the y-axis, it is 66.3 pixels. A margin of 50 pixels was also uniformly removed from all four sides of the grid to mitigate any potential edge effects. The field of view (FoV) subtended by the eye along the x-axis, spans 53.03 degrees. Similarly, the FoV along the y-axis extends to 29.6 degrees. Our capture setup, grid pattern, and lensless reconstruction of captured data are shown in Fig. 3.
Analysis of the angular differences between two consecutive grid points along the x-axis reveals a distinct trend characterized by an initial increase, peaking, and subsequent reduction. The minimum angle along the x-axis and y-axis is observed to be 3.21 degrees and 1.77 degrees, respectively. These trends are observed along the edges as the angular distance reduces with distance from the eye.
5 Lensless Gaze Estimation Approach

Since lensless measurements are globally multiplexed and do not resemble the scene, we use a two-stage approach: in the first stage, we use a reconstruction algorithm to obtain the scene estimate from the lensless captures; in the second stage, we use a gaze-regressor neural network to obtain the gaze direction. We discuss the approach in detail below.
Scene Reconstruction. Given the lensless forward model described in Eq. 1 Scene reconstruction from lensless capture involves solving an inverse problem of estimating the scene , given the measurement and PSF , i.e.,
(2) |
where is the scene estimate and is the reconstruction algorithm. In this work, we experimented with two different approaches for – Tikhonov-regularized least-squares (also known as Wiener deconvolution) and a learning-based algorithm called FlatNet[13]. Tikhonov Regularization involves solving the following equation, where Y is the lensless measurement, P is the PSF, and is the scene estimate.
(3) |
. This equation has the following closed-form solution, where is the Fourier transform and gamma is set to 1e-5.
(4) |
For the learning-based FlatNet, we use pre-trained weights obtained by training it on a natural image dataset. FlatNet consists of two stages: (1) an inversion stage that maps the measurement into a space of intermediate reconstruction by learning parameters within the forward model formulation, and (2) a perceptual enhancement stage that improves the perceptual quality of this intermediate reconstruction. These stages are trained together in an end-to-end manner.
Gaze Regressor. Once the scene estimate is obtained, we use a gaze regression network to finally predict the gaze direction. Our network first predicts the unit vector corresponding to the gaze direction. We then project this 3D gaze direction to a 2D gaze point (in pixels) on the monitor using the known distance between the monitor and the eye.
Loss Function. Once the gaze points on the monitor are estimated, we then use L1 loss on the predicted and groundtruth gaze points. More specifically, we use,
(5) |
where and are the ground-truth gaze point and predicted gaze point, respectively. Finally, we use this loss to update the gaze regression stage while keeping the reconstruction stage frozen. We show our gaze estimation pipeline in Fig. 4.
6 Experiments
In this section, we provide details of the implementation of our training strategy and a comprehensive evaluation of the gaze estimator. We benchmark various eye-tracking algorithms on our FlatTrack dataset and evaluate the performance on angular error and inference times. We also perform a quantitative comparison of lens-based v/s lensless imaging systems. We also assess the different lensless reconstruction methods: Wiener and FlatNet. Finally, we analyze the average per-pixel error of gaze estimation.
6.1 Implementation Details
The FlatNet model is pre-trained on measurements simulated using Eq. 1 from the natural image MIRFLICKR dataset[11]. We freeze this pre-trained FlatNet model for the reconstruction of our captured measurements.
We evaluate the performance of the following gaze estimation models: ResNet-18[10], EyeCOD [23], and MobileNetv2[21]. Of the 4 or 5 rounds of images captured for each subject, we keep one hold-out test round for each subject, which is later utilized to evaluate our models. We split the remaining dataset into training and validation sets (80:20). For each subject, we crop out the eye region from the reconstruction. We also perform random affine transformations on the reconstructed eye images to augment our data and to deal with small misalignments among the different sets.
Our models predict a 3D unit gaze vector corresponding to a measurement. We follow a common training strategy for all models, which involves pre-training the models on the training dataset prepared above, followed by a subject-specific fine-tuning in which we freeze all the model parameters except the last 2 layers and fine-tune for the specific subject’s training data. For training the models, 3D gaze vector outputs from the models are projected into 2D gaze positions using the known dimensions of the monitor and distance from the screen. Following this, L1 loss is calculated on these 2D labels and the 2D predictions for training the models. The common training regimen spans 50 epochs, employing the Adam optimizer with a weight decay of , an initial learning rate of , and a StepLR scheduler with a decay factor of 0.5 every 5 epochs. These hyperparameters remain consistent across all models for pre-training and fine-tuning. Finally, for each subject we test the fine-tuned model on the held-out set.
6.2 Comparison of Lensless Reconstruction Methods
In this section, we conduct a comparative analysis of the latency and performance of the ResNet-18 model using different lensless reconstruction methods, namely Weiner Deconvolution and FlatNet, on our captured dataset. Our primary focus is to evaluate the trade-off between accuracy and inference time. The results of our comparison are summarized in Table 1.

Reconstruction | Average | Inference Time |
Method | Error (in deg) | (in ms) |
Weiner | 2.08∘ | 3.81 |
FlatNet | 1.92∘ | 7.81 |
Fig 5 shows the example of a captured image reconstructed through both methods. The FlatNet reconstruction looks much cleaner and has less noise, which can be attributed to its better gaze estimation performance.
We observe minimal latency variation across both methods with a clear performance improvement in the case of FlatNet reconstruction; hence, we utilize the FlatNet method for our experiments. Note that these latency metrics are evaluated on an NVIDIA GeForce RTX3080 Ti graphics card and Intel Core i9-10900KF processor.
6.3 Gaze Estimator Comparison
In this section, we present the comparisons of the gaze estimation stage on the FlatTrack dataset. For this experiment, we fix the reconstruction stage as FlatNet and evaluate the performance of three different gaze estimators: ResNet-18[10], EyeCOD [23], and MobileNetv2[21]. We initialize ResNet-18 with ImageNet weights and replace the last layer to predict a 3D gaze direction. EyeCOD is a gaze estimation model proposed by You et al. [23] for FlatCam captures. EyeCOD incorporates RITnet[8] for semantic segmentation to crop out the Region of Interest(ROI) around the pupil consisting of the pupil, iris, and sclera, which is then passed on to ResNet-18 for gaze estimation. Since MobileNetv2 is specifically designed for on-the-edge devices and embedded vision applications where computational resources are limited, we also experiment with it.
We evaluate the performance of these gaze estimation architectures on the held-out test set for each subject based on the average angular error on the 3D gaze vectors and the inference times. We report two different errors in Table 2. The best-case error corresponds to the error on the held-out set for the subject with a minimum average error, while the average error is calculated across all the subjects. We also report the inference time evaluated on a single NVIDIA GeForce RTX3080 Ti graphics card.
Gaze | Best-Case | Average | Inference |
Estimator | Error | Error | time (in ms) |
ResNet-18 | 0.91∘ | 1.92∘ | 7.81 |
EyeCOD | 0.95∘ | 1.82∘ | 22.72 |
MobileNet | 1.08∘ | 2.43∘ | 7.67 |
We observe only improvement in performance using the EyeCOD methodology, although it bears a increase in inference time. Clearly, Resnet-18 proves to be a better trade-off, so we perform further experiments using Resnet.
6.4 Comparison of Lensed v/s Lensless Imaging Systems
In this experiment, we assess the performance disparity between images captured by lensed and lensless cameras. As previously noted, there is no existing real-world dataset specifically for lensless eye tracking, which makes direct comparison with conventional datasets challenging. However, we have addressed this by simulating lensless measurements using the Davis-GS dataset[3], which includes user-specific data. This is a crucial aspect, as our collected dataset also includes user-specific labels, making Davis-GS the most relevant benchmark for comparison. In contrast, datasets like OpenEDS2020[18] do not provide user-specific labels, which makes them less suited for evaluating fine-tuned eye-tracking systems. More specifically, our evaluation focuses on comparing the performance of ResNet-18 on original lens images and lensless simulated images. The simulation process involves convolving the images with the point spread function (PSF) of our camera with some additive noise as described in Eq. 1 and subsequently reconstructing them using FlatNet [13].
There are 27 subjects in the Davis-GS dataset, out of which we use 21 subjects’ data for pre-training our gaze estimator and the remaining 3 subjects’ data for fine-tuning and evaluation. Since the Davis-GS dataset consists of multiple images for random grid locations, we take one image for each grid pixel location (to avoid data imbalance) per subject and simulate lensless measurement. For fine-tuning the gaze estimator, we partition the 3 subject’s data into train, validation, and test sets. After fine-tuning the model for respective subjects, we evaluate it on the test data for each subject.
For pre-training and fine-tuning, we follow the same methodology as explained in Section 6.1. However, here, due to the unavailability of 3D gaze vectors, we directly train the models on 2D gaze positions, which corresponded to the pixel location of the stimulus on the screen; hence, we replaced the last layer with a fully connected layer with the output size of 2 and borrow the other hyperparameters as mentioned in Section 6.1. For fine-tuning, we unfreeze the last 2 layers of the pre-trained model and train on the test subject data. For the calculation of angular error, we first convert the predictions and labels to 3D gaze vectors using the known dimensions of the monitor and the distance of the subject from the screen. We report our findings on the test set using the average gaze angular error in degrees for the test set for each subject as summarized in Table 3.
Subject | Lensed | Lensless |
---|---|---|
(in deg) | (in deg) | |
Subject25 | 1.79∘ | 1.84∘ |
Subject26 | 1.72∘ | 1.81∘ |
Subject27 | 1.67∘ | 1.62∘ |
From Table 3, we can observe minimal difference in the performance of the two imaging systems, indicating that the lensless imaging system doesn’t introduce a significant performance loss while allowing a compact form factor.
6.5 Lensless Gaze Estimation Analysis

In this section, we analyze the errors observed in our experiments. To start with, the average error (across all subjects) at each grid pixel location is shown in Fig. 6. The radius of the circle in the plot is proportional to the gaze prediction error. A circle with a higher radius implies more error. Six points are circled randomly from the grid, and their corresponding eye image reconstruction is also displayed on the side.
We observe that the grid locations with lesser error correspond to better lensless reconstructed images, with pupils properly visible. Other points with high error are relatively dark images that lack pupil resolution. This is due to a lack of proper lighting as some of these points are farther away from the light source. A fix for these points with higher error would be to place additional light sources near them.
7 Conclusion
We propose a gaze-tracking framework for thin, lensless cameras. Lensless cameras, due to their ultra-thin form factor and lightweight nature, are attractive for applications requiring eye gaze tracking, such as Augmented/Virtual Reality. However, existing works on gaze estimations have not fully exploited this advantage. One of the bottlenecks for fully exploiting the benefits of lensless cameras for gaze tracking is the lack of a lensless gaze dataset. To deal with this, we propose the FlatTrack dataset - a first-of-its-kind NIR PhlatCam gaze dataset collected from 13 subjects and contains around 20K paired lensless measurements and gaze ground truth. Using a custom lensless gaze estimation pipeline evaluated on the FlatTrack dataset, we show that lensless gaze estimates can be of high fidelity. Furthermore, these gaze estimation algorithms can run at very high speeds (125 fps) on a typical GPU. We hope the proposed dataset and the pipeline, when released, will drive more innovation in this promising space.
Acknowledgments: KM acknowledges funding support from IITM Pravartak Technologies Foundation, Department of Science and Technology, India (CRG/2023/007358) and Qualcomm Faculty Award 2024.
References
- [1] Kaan Akşit, Praneeth Chakravarthula, Kishore Rathinavel, Youngmo Jeong, Rachel Albert, Henry Fuchs, and David Luebke. Manufacturing application-driven foveated near-eye displays. IEEE transactions on visualization and computer graphics, 25(5):1928–1939, 2019.
- [2] Hadi Alzayer, Kevin Zhang, Brandon Feng, Christopher A. Metzler, and Jia-Bin Huang. Seeing the world through your eyes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4864–4873, June 2024.
- [3] Anastasios N. Angelopoulos, Julien N.P. Martel, Amit P. Kohli, Jorg Conradt, and Gordon Wetzstein. Event-based near-eye gaze tracking beyond 10,000 hz. IEEE Transactions on Visualization and Computer Graphics, 27(5):2577–2586, May 2021.
- [4] Nick Antipa, Grace Kuo, Reinhard Heckel, Ben Mildenhall, Emrah Bostan, Ren Ng, and Laura Waller. Diffusercam: lensless single-exposure 3d imaging. Optica, 5(1):1, Dec. 2017.
- [5] M. Salman Asif, Ali Ayremlou, Aswin Sankaranarayanan, Ashok Veeraraghavan, and Richard G. Baraniuk. Flatcam: Thin, lensless cameras using coded aperture and computation. IEEE Transactions on Computational Imaging, 3(3):384–397, 2017.
- [6] Dhruvjyoti Bagadthey, Sanjana Prabhu, Salman S Khan, D Tony Fredrick, Vivek Boominathan, Ashok Veeraraghavan, and Kaushik Mitra. Flatnet3d: intensity and absolute depth from single-shot lensless capture. JOSA A, 39(10):1903–1912, 2022.
- [7] Vivek Boominathan, Jesse K Adams, Jacob T Robinson, and Ashok Veeraraghavan. Phlatcam: Designed phase-mask based thin lensless camera. IEEE transactions on pattern analysis and machine intelligence, 42(7):1618–1629, 2020.
- [8] Aayush K. Chaudhary, Rakshit Kothari, Manoj Acharya, Shusil Dangi, Nitinraj Nair, Reynold Bailey, Christopher Kanan, Gabriel Diaz, and Jeff B. Pelz. Ritnet: Real-time semantic segmentation of the eye for gaze tracking. In 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW). IEEE, Oct. 2019.
- [9] Brian Guenter, Mark Finch, Steven Drucker, Desney Tan, and John Snyder. Foveated 3d graphics. ACM Trans. Graph., 31(6), nov 2012.
- [10] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, June 2016.
- [11] Mark J Huiskes and Michael S Lew. The mir flickr retrieval evaluation. In Proceedings of the 1st ACM international conference on Multimedia information retrieval, pages 39–43, 2008.
- [12] Susmija Jabbireddy, Xuetong Sun, Xiaoxu Meng, and Amitabh Varshney. Foveated rendering: Motivation, taxonomy, and research directions, 2022.
- [13] Salman Siddique Khan, Varun Sundar, Vivek Boominathan, Ashok Veeraraghavan, and Kaushik Mitra. Flatnet: Towards photorealistic scene reconstruction from lensless measurements. IEEE Transactions on Pattern Analysis and Machine Intelligence, page 1–1, 2020.
- [14] Salman S Khan, Xiang Yu, Kaushik Mitra, Manmohan Chandraker, and Francesco Pittaluga. Opencam: Lensless optical encryption camera. arXiv preprint arXiv:2312.01077, 2023.
- [15] Joohwan Kim, Michael Stengel, Alexander Majercik, Shalini De Mello, David Dunn, Samuli Laine, Morgan McGuire, and David Luebke. Nvgaze: An anatomically-informed dataset for low-latency, near-eye gaze estimation. In Proceedings of the 2019 CHI conference on human factors in computing systems, pages 1–12, 2019.
- [16] Kyle Krafka, Aditya Khosla, Petr Kellnhofer, Harini Kannan, Suchendra Bhandarkar, Wojciech Matusik, and Antonio Torralba. Eye tracking for everyone. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2176–2184, 2016.
- [17] Kristina Monakhova, Kyrollos Yanny, Neerja Aggarwal, and Laura Waller. Spectral diffusercam: lensless snapshot hyperspectral imaging with a spectral filter array. Optica, 7(10):1298–1307, Oct 2020.
- [18] Cristina Palmero, Abhishek Sharma, Karsten Behrendt, Kapil Krishnakumar, Oleg V Komogortsev, and Sachin S Talathi. Openeds2020: Open eyes dataset. arXiv preprint arXiv:2005.03876, 2020.
- [19] Anjul Patney, Marco Salvi, Joohwan Kim, Anton Kaplanyan, Chris Wyman, Nir Benty, David Luebke, and Aaron Lefohn. Towards foveated rendering for gaze-tracked virtual reality. ACM Transactions on Graphics (TOG), 35(6):1–12, 2016.
- [20] Thammathip Piumsomboon, Gun Lee, Robert W. Lindeman, and Mark Billinghurst. Exploring natural eye-gaze-based interaction for immersive virtual reality. In 2017 IEEE Symposium on 3D User Interfaces (3DUI), pages 36–39, 2017.
- [21] Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, and Liang-Chieh Chen. Mobilenetv2: Inverted residuals and linear bottlenecks. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE, June 2018.
- [22] Brian A Smith, Qi Yin, Steven K Feiner, and Shree K Nayar. Gaze locking: passive eye contact detection for human-object interaction. In Proceedings of the 26th annual ACM symposium on User interface software and technology, pages 271–280, 2013.
- [23] Haoran You, Cheng Wan, Yang Zhao, Zhongzhi Yu, Yonggan Fu, Jiayi Yuan, Shang Wu, Shunyao Zhang, Yongan Zhang, Chaojian Li, Vivek Boominathan, Ashok Veeraraghavan, Ziyun Li, and Yingyan Lin. Eyecod: eye tracking system acceleration via flatcam-based algorithm and accelerator co-design. In Proceedings of the 49th Annual International Symposium on Computer Architecture, ISCA ’22. ACM, June 2022.
- [24] Xuebai Zhang, Xiaolong Liu, Shyan-Ming Yuan, and Shu-Fan Lin. Eye tracking based control system for natural human-computer interaction. Computational intelligence and neuroscience, 2017, 2017.
- [25] Xucong Zhang, Yusuke Sugano, Mario Fritz, and Andreas Bulling. Appearance-based gaze estimation in the wild. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4511–4520, 2015.