13 December, 2025
Room 602, Student Commons 6/F, Pacific Plaza (Off-campus), Hong Kong SAR
Navigate Here
Yuhui Liu is a PhD researcher in the Department of Electrical and Electronic Engineering (EEE) at The University of Hong Kong (HKU). She received her bachelor's degree in Optics from Nankai University and master's degree in EEE from HKU. Her research focuses on learned optics, computational imaging, and camera system design through hardware-software joint optimization.
Ms. Najia Sharmin is a PhD candidate in the Photonic Systems Research Lab at The University of Hong Kong, and a recipient of the Hong Kong PhD Fellowship and the HKU Presidential PhD Scholarship. Her research focuses on advancing photoacoustic microscopy for biomedical applications, combining laser system design, optical modeling, and image processing to improve diagnostic and translational capabilities, with a particular focus on non-invasive, label-free detection of lipid distributions in human liver tissues within the 1.7-µm optical window.
Dr. Qiang Fu is a Research Scientist in the Computational Imaging Group at KAUST. He received his Ph.D. in Optical Engineering at University of Chinese Academy of Sciences in 2012 and bachelor's degree in Mechanical Engineering at University of Science and Technology of China in 2007. Before joining KAUST in 2017, he was a Research Associate Professor at ShanghaiTech University, a Postdoc at KAUST, and an Assistant Research Fellow at the Academy of Opto-Electronics. His research interests include computational imaging, optical design, diffractive optics, nanofabrication, wavefront sensing, and hyperspectral imaging.
Dr. Erqian Dong is a Postdoctoral Fellow in Mechanical Engineering at The University of Hong Kong, specializing in ultrasound imaging, underwater acoustic metamaterials, and soft-material metalens technologies. He holds a PhD in Oceanography from Xiamen University and has led multiple Hong Kong-based research projects, including RGC Healthy Longevity Catalyst Award and NSFC/RGC JRS projects as Co-Is. Dr. Dong has published extensively in top journals such as Nature Communications, Science Advances, and National Science Review, and has received numerous awards, including Outstanding Young Scholar (2025) at the National Metamaterials Conference and multiple national scholarships.
Chutian Wang received the B.S. degree from the University of Science & Technology Beijing in 2020, and the M.S. degree with distinction at Imperial College London in 2021. He is currently working towards his Ph.D. degree with the Department of Electrical and Electronic Engineering, the University of Hong Kong. His research interests include computational neuromorphic imaging, wavefront sensing and digital holography.
Jiankai Xing is a 5th-year PhD student in the Department of Computer Science and Technology, Tsinghua University, supervised by Professor Xu Kun. He previously obtained his bachelor's degree from the same department. His current research focuses on differentiable rendering and inverse rendering.
Kaixuan Wei is a Ph.D. student at KAUST, supervised by Prof. Wolfgang Heidrich. His research bridges machine learning, optimization, and statistical modeling with applications in computational photography, optics, and computer vision. He received his Bachelor and Master degrees from Beijing Institute of Technology in 2018 and 2021, respectively. Mr. Wei is a recipient of several awards, including the Best Paper Award at ICML 2020 and recognition among the Global Top 100 Chinese Rising Stars in AI. His work has been published in top-tier venues across computer vision, machine learning, and graphics, such as Science Advances, TOG, IEEE TPAMI, ICML, and CVPR.
Zhenyang Li is a Ph.D. student in Electrical and Electronic Engineering at The University of Hong Kong (HKU). His research spans computer vision, computer graphics, VR/AR/MR, and holographic imaging. He focuses on 3D/4D scene and object reconstruction and generation, developing physics-informed rendering and dynamic scene modeling methods for high-fidelity reconstruction and simulation, and exploring bridges between RGB/video data and event-based or physics-aware modalities for realistic scene understanding.
Shi Mao is a Ph.D. student in the Computational Imaging Group at KAUST, advised by Prof. Wolfgang Heidrich. His research interests span computational imaging and 3D vision. Before joining KAUST, Shi served as a Research Consultant at Baidu Research. He received his M.S. from Tsinghua University (TBSI) and his B.E. from South China University of Technology, and was also a visiting student at UC Berkeley during his undergraduate studies.
Dr. Yanmin Zhu is a Research Assistant Professor in the EEE Department at HKU. She obtained her BSc at KU Leuven (Belgium), MSc at Imperial College London, PhD at The University of Hong Kong, and conducted postdoctoral research at MIT. Her research expertise spans intelligent optical metrology, optical diagnosis and material characterization, and wearable and portable devices.
Wenbin Zhou is currently a third-year Ph.D. student at HKU WeLight Lab led by Dr. Yifan (Evan) Peng. He completed his bachelor's degree in Physics, under the guidance of Prof. Ligang Liu, at the University of Science and Technology of China. He has served as a Research Assistant in Prof. Bedrich Benes's group at Purdue University and in Prof. Brian A. Barsky's group at UC Berkeley, contributing to research at the intersection of displays and visual computing. His current academic interests revolve around computational holographic displays, computer graphics, and virtual/augmented reality.
This report presents recent advances in learned optics for RGB-D imaging, focusing on expanding information bandwidth and enhancing optical modulation through hardware–software co-design. By jointly optimizing optical elements and neural reconstruction networks, the proposed systems break the conventional separation between lens design and computational processing. Two main directions are explored: learned stereo imaging and learned wide field-of-view (FoV) imaging. For stereo RGB-D, a rank-2 coded diffractive optical element (DOE) is introduced to encode complementary stereo and focus cues, enabling high-fidelity depth estimation and RGB reconstruction with improved robustness and flexibility. For wide-FoV imaging, off-aperture optical encoding provides additional degrees of freedom and localized wavefront control, overcoming angular trade-offs while maintaining compact system structures. These approaches are validated through differentiable image formation modeling and real-world camera prototypes, demonstrating superior performance in depth accuracy, image quality, and system efficiency. The work highlights the potential of intelligent optical systems for next-generation imaging in applications such as robotics, augmented reality, and computational photography, paving the way toward fully end-to-end learned camera systems.
Lipids are among the most ubiquitous components in living organisms, and changes in lipid levels can indicate a wide range of physiological and pathological conditions, including non-alcoholic fatty liver disease (NAFLD), atherosclerosis, and neurodegenerative disorders. Photoacoustic imaging (PAI) in the 1.7-μm absorption window offers strong lipid contrast, but its quantitative performance has been limited by the capabilities of existing laser sources. We introduce a 1725-nm hybrid optical parametric oscillator emitter (HOPE) tailored for lipid-based PAI, which outperforms existing sources in this wavelength band for photoacoustic applications by virtue of a narrow emission bandwidth, high optical signal-to-noise ratio, and enhanced spectral energy density parameters. By substantially increasing the efficiency of photoacoustic signal generation, the HOPE enables high-contrast lipid imaging that requires no pre-processing before segmentation. This allows us to implement a streamlined, custom segmentation pipeline for quantifying lipid infiltration in human liver tissues imaged with the 1725-nm PAM platform. The resulting steatosis measurements show strong agreement with clinical assessments, underscoring the potential of this approach to complement existing diagnostic procedures through accurate, label-free lipid evaluation. In addition, we also investigate a physics-informed deconvolution framework to mitigate depth-dependent resolution loss caused by optical scattering. This approach can potentially enable significant resolution recovery in the diffuse optical regime and extend the effective imaging depth of our 1725-nm lipid-specific PAM platform, supporting high-fidelity visualisation of lipid-rich tissues across greater depths.
Recent efforts in data-driven spectral reconstruction aim at extracting spectral information from RGB images. Published work reports exceedingly high numerical scores for this reconstruction task, yet real-world performance lags substantially behind. In this paper, we systematically analyze the performance of such methods with three groups of dedicated experiments. First, we evaluate the practical overfitting limitations with respect to current datasets. Second, we reveal fundamental limitations in the ability of RGB to spectral methods to deal with metameric or near-metameric conditions, which have so far gone largely unnoticed due to the insufficiencies of existing datasets. Finally, we analyze the potential for modifying the problem setting to achieve better performance by exploiting optical encoding and metameric augmentation. Our experiments show that the overall performance of existing methods is primarily limited by the dataset issues. Future progress on snapshot spectral imaging will heavily depend on the generation of improved datasets for the design of effective optical encoding strategies.
We present an inverse-designed acoustic metalens for holographic wavefront control, enabled by a holographic iterative angular spectrum algorithm that derives spatially en-coded phase profiles. By unifying the high-frequency eikonal equation with long-wavelength homogenization theory, we establish a frequency-independent, one-to-one correspondence between refractive index distributions and transmission phases—bridging microscale particle design and macroscale wave engineering. Experimentally, the fabricated metalens achieves diffraction-limited acoustic focusing, demonstrating broadband performance with minimal aberration.
High-throughput sensing in manufacturing and optical systems requires the ability to recover fine structural or wavefront details under rapid motion and challenging illumination, where conventional frame-based imaging and classical Shack–Hartmann wavefront sensing often fail due to motion blur, limited dynamic range, and restricted tolerance to large-gradient variations. This seminar introduces a unified approach based on Computational Neuromorphic Imaging (CNI), which leverages the asynchronous, high-dynamic-range nature of event-based sensors. For industrial inspection, CNI enables reliable detection of dynamic surface defects with strong robustness to vibration and rapidly changing conditions. For wavefront analysis, a neuromorphic wave-normal inference framework directly reconstructs high-dimensional wave normals and overcomes the spot-overlapping constraints faced by traditional angle-based sensors, pushing the wavefront sensor dynamic range to Type III. Together, these developments illustrate how CNI provides a coherent foundation for ultrafast optical field inspection beyond the limits of conventional imaging systems.
Differentiable rendering is a key technique for inverse rendering, but existing methods have some limitations: Conventional color derivatives fail to fully capture how scene parameters affect rendering results; traditional loss functions suffer from locality and gradient sparsity, which hinder optimization; and overall pipelines lack robustness in complex scenarios. To solve these problems, this presentation proposes novel path geometry derivatives. These derivatives establish a direct, constraint-driven relationship between scene parameters and light path geometry based on the implicit function theorem, moving beyond over-reliance on color derivatives. Starting with simple paths in rasterization, the method extends to complex paths in path tracing and uses tailored constraints to ensure smooth path contributions and valid derivative computation. It further applies this framework to differentiable photon mapping and combines it with a matching loss function based on optimal transport to achieve global, dense gradients. Experiments show the method outperforms other differentiable rendering method, providing smoother gradients and more robust optimization.
Differentiable optics enables innovative designs using diffractive optical elements, yet a critical gap between simulation and manufacturing limits their practical application. We propose a fabrication-aware design pipeline tailored for direct-write grayscale lithography and nano-imprinting to enable inexpensive, large-area mass production. We introduce a super-resolved neural lithography model that predicts fabricated 3D geometry, enabling fabrication-aware, end-to-end optimization. To address computational challenges, we implement a tensor-parallel framework distributing FFTs across multiple GPUs, facilitating designs up to 32.16×21.44 mm, simulated on grids of up to 128,640 by 85,760 feature points. Our fabricated prototypes demonstrate strong agreement with simulations in holography and PSF engineering. Additionally, we achieve high-quality imaging using a single DOE paired with Wiener filtering. These contributions effectively lower the technical barriers to the real-world deployment of computational diffractive optics.
The work Enhanced Velocity Field Modeling for Gaussian Video Reconstruction improves 3D Gaussian-splatting video reconstruction for dynamic scenes. The authors identify that deformation-field based approaches often overfit in cases of complex motion or scale variation, producing irregular Gaussian trajectories and sub-optimal visual quality. They propose FlowGaussian-VR, which introduces a velocity-field rendering (VFR) pipeline to optimize per-Gaussian velocities under optical-flow supervision, and a flow-assisted adaptive densification (FAD) strategy to increase Gaussian density in dynamic regions. The result is significantly improved temporal coherence, reduced blur on challenging multi-view dynamic datasets, with physically plausible, trackable Gaussian motions. Building on these high-fidelity dynamic reconstructions, EventTracer takes the reconstructed 3D scenes and renders high-temporal-resolution event streams. By combining low-sample-per-pixel path tracing with a lightweight spiking network using BiLIF units and trained via bidirectional EMD loss, EventTracer efficiently generates realistic event sequences that preserve fine spatiotemporal details, more closely matching real event-camera data than prior simulators, while running at roughly 4 minutes per second of 720p video. Together, these works present an end-to-end, physics-informed pipeline: first reconstructing dynamic 3D scenes with motion-aware Gaussian modeling, then producing realistic, high-frequency event data — which enhances dynamic scene modeling, rendering, and event-based vision for VR/AR, robotics, and beyond.
Traditional cameras require complex, heavy lens stacks to ensure corner-to-corner sharpness. Nature, however, found a more efficient solution: the Fovea. The human eye doesn't try to be sharp everywhere at once; it directs high-resolution focus exactly where it's needed. We have engineered a camera system that adopts this biological strategy. Using a Deformable Phase Plate (DPP), our system dynamically performs localized aberration correction. Just as your eye darts around a scene to build a picture, our system captures multiple "foveated" images—sharp at the fixation point, softer at the periphery. By stacking these dynamic glances, we achieve a composite image with full-field clarity, eliminating the need for bulky corrective optics. This offers a glimpse into a future of miniaturized, dynamic optics for VR and tracking applications. To efficiently cover the entire field of view, we propose joint optimization of DPP deformations under imaging budget constraints. Due to the DPP device's non-linear behavior, we introduce a neural network-based control model for improved alignment between simulation-hardware performance. Further applications on extended depth-of-field imaging and object tracking is discussed.
Artificial intelligence-driven materials discovery offers rapid design of novel material compositions, yet synthesis and characterization lag behind. Characterization, in particular, remains bottlenecked by labor-intensive experiments using expert-operated instruments that typically rely on electromagnetic spectroscopy. We introduce SpectroGen, a generative AI model for transmodality spectral generation, designed to accelerate materials characterization. SpectroGen generates high-resolution, high-signal-to-noise ratio spectra with 99% correlation to ground truth and a root-mean-square error of 0.01 a.u. Its performance is driven by two key innovations: (1) a novel distribution-based physical prior and (2) a variational autoencoder (VAE) architecture. The prior simplifies complex structural inputs into interpretable Gaussian or Lorentzian distributions, while the VAE maps them into a physically grounded latent space for accurate spectral transformation. SpectroGen generalizes across spectral domains and promises rapid, accurate spectral predictions, potentially transforming high-throughput discovery in domains such as battery materials, catalysts, superconductors, and pharmaceuticals.
Holographic techniques are increasingly recognized for their potential in the development of display systems for virtual reality (VR) and augmented reality (AR). The generation of realistic 3D visual representations is crucial for facilitating user immersion. However, current computer-generated holography (CGH) methods utilized in these applications are often constrained by limitations in computational efficiency, substantial data processing demands, and limitations in display form factor. This manuscript explores recent holographic display technologies employing system-aware CGH algorithms designed to mitigate these inherent challenges.