events

Workshop on

Frontiers of Image Science and Visual Computing 2024

HKU

Let there be light

8 April, 2024 PM

Rayson Huang Lecture Theatre, The University of Hong Kong, HONG KONG SAR

Workshop Objectives

Encourage Innovative Spirit

Promote Excellence and Sustain Quality

Strive for Improvement

Connect Communities

Schedule

Register Now (Free)
14:15 - 14:30
Check-in & Welcome Preview
14:30 - 14:40
Opening Remarks
David Srolovitz, NAS, Dean of Engineering, HKU
Xu Liu, Director of SKL of Extreme Photonics and Instrumentation, Zhejiang University
14:40 - 15:30
William Mong Distinguished Lecture: Simulation Technologies for Image Systems
Brian A. Wandell, NAS, Stanford University
15:30 - 15:35
Pitching for Stanford Center for Image Systems Engineering (SCIEN)
Joyce Farrell, Executive Director of SCIEN
15:35 - 16:05
Tea Reception
16:05 - 16:30
Self-Calibrating, Fully Differentiable NLOS Inverse Rendering
Min H. Kim, Korea Advanced Institute of Science and Technology (KAIST)
16:30 - 16:55
High-speed Photometric Analysis using an Event Camera
Boxin Shi, Peking University (PKU)
16:55 - 17:15
When No-Reference Image Quality Models Meet MAP Estimation
Kede Ma, City University of Hong Kong (CityU)
17:15 - 17:35
Reshaping the Soundscape with User Defined Metamaterials for Imaging
Nicholas X. Fang, HKU
17:35 - 17:40
Closing Remarks
Head of EEE Department, HKU

Speakers/Guests

Speaker Name

Brian A. Wandell

Founding Director, Stanford’s Center for Neurobiological Imaging, Stanford University

Brian A. Wandell, the Isaac and Madeline Stein Family Professor, joined the Stanford Psychology faculty in 1979. He is a member, by courtesy, of Electrical Engineering, Ophthalmology, and the Graduate School of Education. Wandell is the Founding Director of Stanford’s Center for Cognitive and Neurobiological Imaging, and he served as a Deputy Director of the Wu Tsai Neuroscience Institute from 2013-2021. His research centers on vision science, spanning visual neuroscience to digital imaging systems (cameras, displays). He was elected to the American Academy of Arts and Sciences in 2011, received the highest honor of the Society for Imaging Science and Technology in 2014, was awarded the George A. Miller prize of the Cognitive Neuroscience Society in 2016, and Proctor Medal from the Association for Research in Vision and Ophthalmology in 2021. Wandell was elected to the US National Academy of Sciences in 2003.

Speaker Name

Joyce Farrell

Executive Director, Stanford Center for Image Systems Engineering, Stanford University

Joyce Farrell is the Executive Director of the Stanford Center for Image Systems Engineering. She has more than 20 years of research and professional experience working at a variety of companies and institutions, including the NASA Ames Research Center, New York University, the Xerox Palo Alto Research Center, Hewlett Packard Laboratories and Shutterfly (a startup company specializing in online digital photo-finishing). She is also the CEO and founder of ImagEval Consulting, LLC.

Speaker Name

Min H. Kim

Chair Professor, Korea Advanced Institute of Science and Technology (KAIST)

Min H. Kim is a Professor of Computer Science at the KAIST School of Computing, where he directs the Visual Computing Lab (VCLAB). Prior to joining KAIST, he was a postdoctoral researcher at Yale University. He holds a Ph.D. in Computer Science from University College London (UCL). He has received numerous awards, including the SIGGRAPH Technical Paper Award Honorable Mention in 2022. His main research areas are computational imaging, computational photography, 3D imaging, BRDF acquisition, and 3D reconstruction. He has served as Technical Paper Chair for Eurographics 2022, Course Chair for SIGGRAPH Asia 2022, and on many computer graphics and computer vision program committees, including SIGGRAPH, CVPR, ICCV, and Eurographics. He has served as an Associate Editor for top CS journals, including ACM Transactions on Graphics (TOG) and IEEE Transactions on Visualization and Computer Graphics (TVCG).

Speaker Name

Xu Liu

Director of State Key Laboratory of Extreme Photonics and Instrumentation, Zhejiang University

Xu Liu is a Distinguished Professor of Optical Science and Engineering at Zhejiang University. Serving as the Director of the State Key Laboratory of Extreme Photonics and Instrumentation, he holds fellowships with OPTICA and SPIE, and is the Secretary General of the Chinese Optical Society (COS). His research primarily focuses on optical thin films coatings and metasurfaces, optical imaging techniques—including super-resolution microscopy and computational imaging—and display technologies such as 3D light field displays, projection displays, and AR/VR.

Speaker Name

Boxin Shi

Associate Professor, Peking University

Boxin Shi is currently a Boya Young Fellow Associate Professor (with tenure) and Research Professor at Peking University, where he leads the Camera Intelligence Lab. He has also been a Young Scientist at Beijing Academy of Artificial Intelligence. He received the PhD degree from the University of Tokyo in 2013. From 2013 to 2017, he did research at MIT Media Lab, Singapore University of Technology and Design, Nanyang Technological University, and National Institute of Advanced Industrial Science and Technology. His research interests are computational photography and computer vision. He received the Okawa Foundation Research Grant in 2021. He has served as an associate editor of TAPMI/IJCV and an area chair of CVPR/ICCV/ECCV. He is a Senior Member of the IEEE/CCF/CSIG, and a Distinguished Lecturer of APSIPA.

Speaker Name

Kede Ma

Assistant Professor, City University of Hong Kong

Kede Ma is an Assistant Professor with the Department of Computer Science at City University of Hong Kong (CityU). He received the B.E. degree from the University of Science and Technology of China (USTC) in 2012, the MASc. and Ph.D. degrees from the University of Waterloo, in 2014 and 2017, respectively. Prior to joining CityU, he was a Research Associate with Howard Hughes Medical Institute and New York University, from 2018 to 2019. Dr. Ma leads the Multimedia Analytics (MA) Laboratory, which is an interdisciplinary research group. We are interested in Computational Vision, Computational Photography, Multimedia Forensics and Security.

Speaker Name

Nicholas X. Fang

Professor, The University of Hong Kong

Professor Fang joined the University of Hong Kong in 2022 as Professor of Mechanical Engineering. He was full professor with tenure Mechanical Engineering at MIT. His research programs have focused on the scalable manufacturing processes for wave functional materials. Professor Fang earned his B.S. and M.S. degrees in Physics from Nanjing University, China; and Ph.D. degree in Mechanical Engineering from the University of California at Los Angeles. Professor Fang’s His research on nanoarchitectured metamaterials was highlighted among the top 10 Emerging breakthrough technologies of the year 2015. His recognitions also include the OSA Fellow (2021); ASME Chao and Trigger Young Manufacturing Engineer Award (2013); the ICO prize from the International Commission of Optics (2011); the NSF CAREER Award (2009) and MIT Technology Review Magazine’s 35 Young Innovators Award (2008).

Workshop Materials

William Mong Distinguished Lecture: Simulation Technologies for Image Systems

Brian A. Wandell

The number and type of imaging systems has grown enormously over the last several decades; these systems are an essential component in mobile communication, medicine, automotive and drone applications. Imaging systems are also increasingly used with deep learning systems that require large amounts of training data. For these reasons software prototyping has become an essential tool for the design, evaluation and training of modern image systems. I will describe three closely related open-source and freely available image systems toolboxes - ISETCam, ISETBio, and ISET3d - that support design and evaluation of whole image systems. The presentation will include examples of how we model the three-dimensional scene spectral radiance, retinal encoding (physiological optics and cone sampling), and image systems hardware (multi-element lenses, image sensors). I will also provide two examples of how we used end-to-end image system simulation to aid understanding and design.


Related publications
  1. ISETAuto: Detecting vehicles with depth and radiance information (2021) . Zhenyi Liu, Joyece Farrell, Brian Wandell, IEEE Access ACCESS.2021.3063692
  2. A system for generating complex physically accurate sensor images for automotive applications (2019). Zhenyi Liu, Minghao Shen, Jiaqi Zhang, Shuangting Liu, Henryk Blasinski, Trisha Lian, Brian Wandell. IS&T Electronic Imaging Conference, San Francisco.
  3. Ray tracing 3D spectral scenes through human optics models (2019). Trish Lian, Kevin McKenzie, David Brainard, Nicolas Cottaris, Brian Wandell. Journal of Vision October 2019, Vol.19, 23. doi:https://doi.org/10.1167/19.12.23
  4. A computational observer model of spatial contrast sensitivity: Effects of photocurrent encoding, fixational eye movements and inference engine (2020). Nicolas P. Cottaris, Brian A. Wandell, Fred Rieke, David H. Brainard Journal of Vision doi:https://doi.org/10.1167/jov.20.7.17

See the wiki pages of the repositories at: https://github.com/ISET

Self-Calibrating, Fully Differentiable NLOS Inverse Rendering

Min H. Kim

Existing time-resolved non-line-of-sight (NLOS) imaging methods reconstruct hidden scenes by inverting the optical paths of indirect illumination measured at visible relay surfaces. These methods are prone to reconstruction artifacts due to inversion ambiguities and capture noise, which are typically mitigated through the manual selection of filtering functions and parameters. This talk introduces a fully-differentiable end-to-end NLOS inverse rendering pipeline that self-calibrates the imaging parameters during the reconstruction of hidden scenes, using as input only the measured illumination while working both in the time and frequency domains. The proposed pipeline extracts a geometric representation of the hidden scene from NLOS volumetric intensities and estimates the time-resolved illumination at the relay wall produced by such geometric information using differentiable transient rendering. It then uses gradient descent to optimize imaging parameters by minimizing the error between our simulated time-resolved illumination and the measured illumination. This end-to-end differentiable pipeline couples diffraction-based volumetric NLOS reconstruction with path-space light transport and a simple ray marching technique to extract detailed, dense sets of surface points and normals of hidden scenes. This talk demonstrates the robustness of the proposed method to consistently reconstruct geometry and albedo, even under significant noise levels.

High-speed Photometric Analysis using an Event Camera

Boxin Shi

Compared with conventional frame-based cameras, the event camera has unique advantages, especially in their ability to perceive high-speed moving objects and scenes with high dynamic range. Existing research has fully demonstrated the advantages of event cameras in computer vision tasks such as image deblurring, high dynamic range imaging, and high-speed object detection and recognition. However, the photometric image formation model of event cameras has not been fully analyzed. This talk will share a series of research progress on modeling and analyzing the photometric image formation model of an event camera that records and responses to high-speed radiance changes: obtaining high-fidelity scene radiance estimation via analyzing the transient event frequency, constructing real-time photometric stereo with fast-moving light sources to estimate surface normals, and conducting direct-global illumination separation by capturing shadows of a line occlude swiftly sweeping over the scene.

When No-Reference Image Quality Models Meet MAP Estimation

Kede Ma

Contemporary no-reference image quality assessment (NR-IQA) models can effectively quantify the perceived image quality, with high correlations between model predictions and human perceptual scores on fixed test sets. However, little progress has been made in comparing NR-IQA models from a perceptual optimization perspective. However, little progress has been made in leveraging NR-IQA models as natural image priors for use in real-world image enhancement. Here, for the first time, we demonstrate that NR-IQA models can be plugged into the maximum a posteriori (MAP) estimation framework for image enhancement. This is achieved by taking the gradients differentiable and bijective diffusion latents rather than in the raw pixel domain. Different NR-IQA models are likely to induce different enhanced images, which are ultimately subject to psychophysical testing. This leads to a new computational method for comparing NR-IQA models within the analysis-by-synthesis framework. Compared to conventional correlation-based metrics, our method provides complementary insights into the relative strengths and weaknesses of the competing NR-IQA models in the context of perceptual optimization. Taking a step further, we finetune the best-performing NR-IQA model on the combination of previously trained datasets and newly MAP-enhanced images, showcasing the value of MAP-enhanced images for NR-IQA model rectification.

Reshaping the Soundscape with User Defined Metamaterials for Imaging

Nicholas X. Fang

In recent years, remarkable progress has been made in the fields of spatial wave modulation and nanomaterial fabrication, unlocking numerous possibilities for engineering wave-matter interactions. Through systematic exploration of the design space, promising applications have been found in opto-mechanics, nanophotonics, and 3D imaging. Consequently, large-scale computational optimization has become an invaluable tool for discovering novel optical structures. This raises the question: can we harness the power of inverse design to develop acoustic metamaterials with tailored wave-filtering properties for next-generation ultrasound imaging devices?

This talk presents the recent exploration of metasurfaces for acoustic waves enabled by advanced design and micro/nanofabrication techniques. An example is the recent effort of ultrasonic metalenses over broadband frequencies, ranging from 38 kHz to 0.5 MHz. Through simulations and experimental validation, we show that ultrasonic metalens offer remarkable advantages in intensity transmission and enhancement in spatial resolution. We will also discuss the potential impact of user defined metamaterials for emerging applications such as underwater acoustic navigation, non-destructive evaluation, and high-resolution medical ultrasound imaging.