Physics World Model — Modality Catalog

3 imaging modalities with descriptions, experimental setups, and reconstruction guidance.

Integral Photography

integral Computational

Integral photography (IP), originally proposed by Lippmann in 1908, captures a light field using a fly-eye lens array (matrix of small lenses) where each lenslet records a small elemental image from a slightly different perspective. The array of elemental images encodes 3D scene information, enabling computational refocusing, depth estimation, and autostereoscopic 3D display. Compared to microlens-based plenoptic cameras, IP typically uses larger lenslets with correspondingly more pixels per lens. Reconstruction includes depth-from-correspondence between elemental images and 3D focal stack computation.

Physics: light field
Solver: depth_estimation
Noise: gaussian
#computational #integral #multi_view #3d_display #depth
View details →

Light Field Imaging

light_field Computational

Light field imaging captures the full 4D radiance function L(x,y,u,v) describing both spatial position (x,y) and angular direction (u,v) of light rays. A microlens array placed before the sensor captures multiple sub-aperture views simultaneously, enabling post-capture refocusing, depth estimation, and perspective shifts. Each microlens images the objective's exit pupil, trading spatial resolution for angular resolution. The 4D light field can be processed with shift-and-sum for refocusing, disparity estimation for depth, or epipolar-plane image (EPI) analysis. Primary challenges include the inherent spatial-angular resolution tradeoff and microlens aberrations.

Physics: light field
Solver: shift_and_sum
Noise: gaussian
#computational #light_field #plenoptic #depth #refocusing
View details →

Panorama Multi-Focus Fusion

panorama Computational

Multi-focus panoramic fusion combines images captured at different focal planes and/or different spatial positions to produce an all-in-focus image with extended depth of field and wide field of view. Focus stacking selects the sharpest regions from each focal plane using local contrast measures, then blends them via Laplacian pyramid fusion or wavelet-based methods. Panoramic stitching aligns overlapping images using feature matching (SIFT/SURF) and blends seams. Primary challenges include parallax at scene edges and focus measure ambiguity in low-texture regions.

Physics: multi focus
Solver: laplacian_pyramid_fusion
Noise: gaussian
#computational #panorama #fusion #focus_stacking #extended_dof
View details →