The optical sectioning problem: Pawley explains that confocal microscopy acquires true optical sections — each Z-plane captures only in-focus information. But widefield microscopy integrates light from all depths simultaneously. For widefield Z-stacks, each slice contains in-focus signal (from structures at that plane) mixed with out-of-focus haze (from structures above and below). Projection methods must distinguish in-focus from out-of-focus content.
Wavelet fusion — the mathematical basis: Gonzalez & Woods describe wavelet decomposition as multi-resolution analysis: images are decomposed into approximation (low-frequency, smooth) and detail (high-frequency, edges) coefficients at multiple scales. Vetterli's nonlinear approximation theorem shows that keeping the largest-magnitude wavelet coefficients (regardless of which Z-slice they come from) produces the optimal sparse representation — the image that preserves the most important structures (edges, features) while suppressing noise. This is exactly what wavelet fusion does: at each spatial location and frequency scale, it selects the coefficient with the largest magnitude from the Z-stack.
Why wavelet outperforms MIP: Maximum Intensity Projection selects the brightest pixel, which is not necessarily the most focused. A noisy out-of-focus plane can have bright noise spikes that "win" the MIP competition at scattered pixels, introducing salt-and-pepper noise artifacts. Wavelet fusion operates in the frequency domain where in-focus content produces large high-frequency coefficients (sharp edges) while noise produces small coefficients at random positions. By selecting the largest coefficients, wavelet fusion inherently favors signal over noise.
Hanrahan's signal processing perspective: Each Z-slice is a sample of the 3D specimen at a different depth. Projection is reconstruction — combining these samples into a single representation. The choice of reconstruction method (pixel-wise max, mean, or wavelet) determines how faithfully the 3D information is preserved in 2D. Wavelet methods provide the best reconstruction because they respect the multi-scale structure of biological images.
When you capture images at multiple focus depths, some planes are sharp for certain structures and blurry for others. Wavelet fusion analyzes each plane for sharp details (strong edges, clear textures) using frequency decomposition, then assembles the sharpest details from all planes into one composite. This works better than simply taking the brightest pixel (which can pick up noise) or averaging (which blurs everything).