The Quantum Nature of Observation
I watched the memorial service on January 27 at Auschwitz-Birkenau Memorial and Museum.
Words fail.
Annelies Marie Frank clutched her "escape bag," more because she needed something to hold onto than because escape was possible. The sirens wailed—not just air-raid warnings, but constant, day-shattering interruptions that blended into the deeper terror she documented: the sudden knocks at the door that could end everything, the clutching fear in trembling hands, the anxiety of discovery in every ordinary moment. That dread was not abstract. It lived in neighborhoods, enforced street by street.
Two officers stood in front of Buildings 10 and 11—the former medical-experiment and punishment blocks—their windows sealed shut. That image was surreal. It demanded absolute human respect and dignity. They were not enforcing authority. They stood as witnesses.
The red lights along the perimeter brought the full machinery into brutal focus: city authorities, fire departments, police, military, schools, hospitals, universities, businesses—all woven into the system. Under Himmler, local police and the incorporated fire brigades (Feuerschutzpolizei within the Ordnungspolizei) became essential gears. They patrolled, arrested, sustained the regime's grip at the everyday level. They created immediate, neighborhood-level fear—the kind that punctuates meals, shatters appetites, turns every knock into potential annihilation.
This was no distant machinery. It was sustained by ordinary institutions, ordinary people following orders, ordinary routines made lethal. The weight of that complicity presses down without mercy.
That is what the sealed windows, the standing witnesses, the red perimeter lights forced me to see. Words still fail.
They lived it: the ever-present threat that any moment might force them to run or face seizure, the bogus pretexts for searches and arrests that tore lives and families apart without warning or real cause. This is how it worked. All part of the machinery.
Today the parallels are quieter but no less pervasive: digital red lights—algorithmic alerts, data flags, compliance pings—sound daily across the same web of institutions. Police pull from private databases; fire/EMS systems feed into shared tracking networks; schools deploy behavioral monitoring and share records; hospitals exchange patient data for "safety" or billing; universities partner on AI surveillance research; businesses track every interaction under the guise of efficiency. The machinery hums on, normalizing fear of exclusion through routine, interconnected oversight.
Extraordinary measures were taken to engineer it and then to lay a blanket over it. Not accountability. Not repair. A kind of institutional prayer. Comfort without reckoning. Responsibility was distributed. Identity reduced to workflow.
And the red itself felt intentional. It read as blood—a quiet, unavoidable reminder of what this system consumes.
This must never be repeated. And yet both young and old in our society increasingly behave as if historical memory were optional.
From the arrest of innocent families, their renaming and systematic indoctrination, to consequences that persist across generations, we must understand the full depth of what happened—during and after. This was not confined to a handful of sites. It unfolded across more than 44,000 camps, ghettos, and detention facilities. An estimated 17 million people or more.
The accounting remains incomplete, bound up with the medical, scientific, engineering, and social machinery that was absorbed into civilian institutions after the war. This was a distributed system, not isolated evil.
Germany in the 1940s produced world-class engineering. Less understood is that its social machinery was also highly engineered—optimized for hierarchy, compliance, and administrative efficiency. That architecture did not disappear. It was absorbed.
While so many people seem to have forgotten, perhaps they never learned history—or maybe they didn't learn the real history. If they did, would they remember?
I don’t usually write outside technical documentation.
My world is syntax, signal paths, kernel logs, and build systems, not essays about power or society. But visibility changes responsibility.
Technology is powerful. Distributed systems shape reality at scale. With that in mind, I want to pivot back to the technical work.
The system is built around MST++, a transformer based model trained to take a standard three channel RGB photograph and mathematically reconstruct it into a 31 band hyperspectral cube covering roughly 400 to 1000 nm. The process involves tiling the image into overlapping patches, running inference on each in batches, then blending the results with weighted overlap to avoid edge artifacts. The output cube is validated for basic quality (dynamic range, spectral smoothness, non zero values) before any further analysis. At its core, the pipeline does not see more colors, it learns to infer hidden correlations across wavelengths that the human eye and ordinary cameras discard. It is highly accurate on benchmark datasets (e.g., PSNR ~34 dB) and has shown strong results in medical imaging tasks such as tissue classification and pathology detection, where subtle spectral differences can reveal conditions invisible to standard photography — much like the hidden quantum correlations Einstein once called "spooky."
Quantum entanglement, first highlighted in the 1935 EPR paper by Einstein, Podolsky, and Rosen, describes how two particles can be born together from a single quantum event and remain perfectly related — so that measuring one instantly reveals the state of the other, no matter the distance. Einstein called this "spooky action at a distance" because it seemed to violate locality, yet experiments confirmed it is real. The correlation is hidden: you can not see it by looking at one particle alone; only when you compare both do the non local links appear.
Hyperspectral imaging shares a similar logic, though purely classical. A standard RGB photo captures only three broad wavelength bands, discarding the fine grained spectral fingerprint of materials. The MST++ model reconstructs the full 31 band cube by inferring hidden correlations across wavelengths — correlations that are real but invisible to trichromatic vision. Just as entanglement reveals non local quantum reality only through joint measurement, spectral reconstruction reveals non visible material properties only through mathematical inference across the entire spectrum. In both cases, direct local observation is incomplete; deeper truth requires reconstruction — whether by comparing entangled particles or blending the tiled results.
Building a convolutional neural network or a multi-stage inference pipeline isn’t magic. It’s engineering. Difficult, yes, but not inaccessible. Most doctorate level research today focuses on algorithmic refinement, not on writing efficient compiled systems or production grade pipelines. If you can code, understand semantics, and reason about data flow, you can build perception systems.
So the real question isn’t can we build them.
It’s why, and for whom.
Modern municipal vision systems combine thermal camera feeds with object detection overlays. They identify “subjects” as they move, run, or flee across a landscape. Heat signatures become tracks. Tracks become classifications.
Radiological and infrared sensing are often framed as technical achievements.
They aren’t.
This is old technology.
Thermal imaging based on radiological and infrared principles has existed for decades. The physics are well understood. The sensors are mature. What has changed is not capability, but scale, automation, and framing.
These systems now convert light into heat maps, heat maps into bounding boxes, and bounding boxes into behavioral metadata automatically, continuously, and without context.
My own use of these tools is microscopic, confined to a home lab environment. It is for imaging and algorithmic research. Not surveillance. Not tracking. Not classification of people.
What matters isn’t the technology.
What matters is how casually it is framed.
People reduced to vectors.
Human movement including fleeing reduced to metadata.
Lives abstracted into pipelines.
This framing is grotesquely offensive.
Once raw physics becomes structured data, it stops being imagery. It becomes inference.
And inference, at scale, becomes power.
The kind of data that shows you exactly how perception is transformed into classification, signals into vectors, people into metadata.
And it forces a harder reflection.
Where we were in 1935, and where we are today.
Back then, people were cataloged by hand. Paper forms. Filing cabinets. Human clerks deciding who belonged and who did not. The machinery was slower, but the intent was familiar. Classification first. Movement second. Consequences later.
Today, it’s automated.
Bounding boxes replace clipboards. Vectors replace names. Dashboards replace offices.
The velocity has changed. The abstraction has changed.
The danger has not.
Today that power increasingly lives behind closed source cameras, closed firmware, proprietary models, and vendor controlled pipelines. These systems are marketed as convenience and safety, but they operate as black boxes. Communities do not get to inspect the models, while private individuals increasingly monitor public feeds inside everyday commercial spaces. People cannot audit the datasets. There is no meaningful public accountability for how identity, motion, or behavior are classified.
These tools are now embedded in apartment complexes, homeowners associations, parking operators, retail centers, restaurants, and private property networks. Surveillance infrastructure is quietly being absorbed into the private sector, where accountability is weaker and oversight is minimal.
You no longer need formal authority to deploy perception at scale.
You just need access.
When vision pipelines move into restaurants and storefronts, when license plate readers sit beside drive-through lanes, when facial classifiers operate in spaces meant for ordinary life, the line between civic safety and commercial monitoring disappears.
As perception systems migrate outside institutional boundaries, the risk multiplies.
There are no transparency requirements. No due process guarantees. No public records requests. Just automated identification feeding proprietary dashboards.
That is how monitoring becomes rounding up.
Not through dramatic announcements, but through normalized pipelines. License plate hits. Face matches. Movement graphs. Quiet coordination between private entities using tools originally justified for safety and efficiency.
This is where remembrance matters.
We remember because forgetting makes repetition easy.
Entire populations have been cataloged before. Movement has been tracked before. Human beings have been reduced to identifiers before. History shows us exactly what happens when classification replaces compassion and automation replaces accountability.
Remembrance is not symbolic. It is technical. It is cultural. It is defensive.
It reminds us what centralized perception can become when left unchecked.
These systems are still owned by the same institutions, built on closed hardware, closed firmware, closed datasets, and export restricted sensors. The stack is opaque by design. Not because it has to be, but because opacity preserves control.
Software systems that perform hyperspectral transformation, thermal fusion, or automated classification aren’t rare anymore. What’s rare is transparency.
Open standards matter because they decentralize understanding. They allow inspection. Reproduction. Audit. They make it harder for surveillance to quietly become infrastructure.
Technology does not drift toward ethics.
It drifts toward whoever funds it.
And that brings us to where we were as a country, and where we are going.
Entire generations fought wars so that power would not be centralized, so that citizens could question authority, so that freedom of movement, thought, and expression were not mediated by unseen systems. Those sacrifices were not made so that people could eventually be reduced to vectors in proprietary dashboards.
We did not fight for independence so that perception itself could become privatized.
The same neural networks used to track movement across terrain can inspect bridges, analyze crops, preserve artwork, or study climate systems. The code is identical. The difference is governance.
Engineers like to believe we’re neutral.
We aren’t.
Every dataset encodes priorities. Every optimization implies intent. Every deployment answers a simple question.
Who benefits?
I don’t write this as a theorist.
I write this as someone who builds systems.
We are now creating machines that see.
If we don’t also build systems that explain, verify, and democratize that vision, we shouldn’t be surprised when it only looks in one direction.
If I could show you photos of the street on Merwedeplein in 1944, I would. The people who live there understand exactly what their city endured. They walk past it. They live beside it. This history is not abstract to them. It is embedded in place. And I know, with certainty, that they have not forgotten. Lately, it feels like the distance between there and here is shrinking.
They lived.
They loved.
They laughed.
They smiled.
I saw the unseen details.
I’m eternally grateful.
And I saw what followed.
I lived part of it.
I’ve spent about 30 years writing C, and I enjoy algorithmic-level challenges. That’s what led me to build a hyperspectral imaging system. What follows is a semi high level overview.
The interrogation of physical reality through the medium of light remains one of the most profound endeavors of scientific inquiry. This pursuit traces its modern theoretical roots to the mid-20th century, a pivotal era for physics.
In 1935, Albert Einstein and his colleagues Boris Podolsky and Nathan Rosen published a seminal paper that challenged the completeness of quantum mechanics.1 They introduced the concept of EPR pairs to describe quantum entanglement, where particles remain inextricably linked, their states correlated regardless of spatial separation.
It is the quintessential example of quantum entanglement. An EPR pair is created when two particles are born from a single, indivisible quantum event, like the decay of a parent particle.
This process "bakes in" a shared quantum reality where only the joint state of the pair is defined, governed by conservation laws such as spin summing to zero. As a result, the individual state of each particle is indeterminate, yet their fates are perfectly correlated.
Measuring one particle (e.g., finding its spin "up") instantaneously determines the state of its partner (spin "down"), regardless of the distance separating them. This "spooky action at a distance," as Einstein called it, revealed that particles could share hidden correlations across space that are invisible to any local measurement of one particle alone. While Einstein used this idea to argue quantum theory was incomplete, later work by John Bell2 and experiments by Alain Aspect3 confirmed this entanglement as a fundamental, non-classical feature of nature.
The interrogation of physical reality through the medium of light remains one of the most profound endeavors of scientific inquiry. This pursuit traces its modern theoretical roots to the mid-20th century, a pivotal era for physics.
In 1935, Albert Einstein and his colleagues Boris Podolsky and Nathan Rosen published a seminal paper that challenged the completeness of quantum mechanics.1 They introduced the concept of EPR pairs to describe quantum entanglement, where particles remain inextricably linked, their states correlated regardless of spatial separation.
It is the quintessential example of quantum entanglement. An EPR pair is created when two particles are born from a single, indivisible quantum event, like the decay of a parent particle.
This process "bakes in" a shared quantum reality where only the joint state of the pair is defined, governed by conservation laws such as spin summing to zero. As a result, the individual state of each particle is indeterminate, yet their fates are perfectly correlated.
The EPR-Spectral Analogy: Hidden Correlations
|
Quantum Physics (1935)
EPR Pairs: Particles share non-local entanglement.
Measuring one particle gives random results; correlation only appears when comparing both
|
Spectral Imaging (Today)
Spectral Pairs: Materials share spectral signatures.
The correlation is invisible to trichromatic (RGB) vision
|
|
↓
Mathematical
Reconstruction ↓
Reveals Hidden
Correlations |
|
While the EPR debate centered on the foundations of quantum mechanics, its core philosophy, that direct observation can miss profound hidden relationships, resonates deeply with modern imaging. Just as the naked eye perceives only a fraction of the electromagnetic spectrum, standard RGB sensors discard the high-dimensional "fingerprint" that defines the chemical and physical properties of a subject. Today, we resolve this limitation through multispectral imaging. By capturing the full spectral power distribution of light, we can mathematically reconstruct the invisible data that exists between the visible bands, revealing hidden correlations across wavelength, just as the analysis of EPR pairs revealed hidden correlations across space.
Silicon Photonic Architecture: The 48MP Foundation
The realization of this physics in modern hardware is constrained by the physical dimensions of the semiconductor used to capture it. The interaction of incident photons with the silicon lattice, generating electron-hole pairs, is the primary data acquisition step for any spectral analysis.
Sensor Architecture: Sony IMX803
The core of this pipeline is the Sony IMX803 sensor. Contrary to persistent rumors of a 1-inch sensor, this is a 1/1.28-inch type architecture, optimized for high-resolution radiometry.
- Active Sensing Area: Approximately \(9.8 \text{ mm} \times 7.3 \text{ mm}\). This physical limitation is paramount, as the sensor area is directly proportional to the total photon flux the device can integrate, setting the fundamental Signal-to-Noise Ratio (SNR) limit.
- Pixel Pitch: The native photodiode size is \(1.22 \, \mu\text{m}\). In standard operation, the sensor utilizes a Quad-Bayer color filter array to perform pixel binning, resulting in an effective pixel pitch of \(2.44 \, \mu\text{m}\).
Mode Selection
The choice between binned and unbinned modes depends on the analysis requirements:
- Binned mode (12MP, 2.44 µm effective pitch): Superior for low-light conditions and spectral estimation accuracy. By summing the charge from four photodiodes, the signal increases by a factor of 4, while read noise increases only by a factor of 2, significantly boosting the SNR required for accurate spectral estimation.
- Unbinned mode (48MP, 1.22 µm native pitch): Optimal for high-detail texture correlation where spatial resolution drives the analysis, such as resolving fine fiber patterns in historical documents or detecting micro-scale material boundaries.
The Optical Path
The light reaching the sensor passes through a 7-element lens assembly with an aperture of ƒ/1.78. It is critical to note that "Spectral Fingerprinting" measures the product of the material's reflectance \(R(\lambda)\) and the lens's transmittance \(T(\lambda)\). Modern high-refractive-index glass absorbs specific wavelengths in the near-UV (less than 400nm), which must be accounted for during calibration.
The Digital Container: DNG 1.7 and Linearity
The accuracy of computational physics depends entirely on the integrity of the input data. The Adobe DNG 1.7 specification provides the necessary framework for scientific mobile photography by strictly preserving signal linearity.
Scene-Referred Linearity
Apple ProRAW utilizes the Linear DNG pathway. Unlike standard RAW files, which store unprocessed mosaic data, ProRAW stores pixel values after demosaicing but before non-linear tone mapping. The data remains scene-referred linear, meaning the digital number stored is linearly proportional to the number of photons collected (\(DN \propto N_{photons}\)). This linearity is a prerequisite for the mathematical rigor of Wiener estimation and spectral reconstruction.
The ProfileGainTableMap
A key innovation in DNG 1.7 is the ProfileGainTableMap (Tag 0xCD2D). This tag stores a spatially varying map of gain values that represents the local tone mapping intended for display.
- Scientific Stewardship: By decoupling the "aesthetic" gain map from the "scientific" linear data, the pipeline can discard the gain map entirely. This ensures that the spectral reconstruction algorithms operate on pure, linear photon counts, free from the spatially variant distortions introduced by computational photography.
Algorithmic Inversion: From 3 Channels to 16 Bands
Recovering a high-dimensional spectral curve \(S(\lambda)\) (e.g., 16 channels from 400nm to 700nm) from a low-dimensional RGB input is an ill-posed inverse problem. While traditional methods like Wiener Estimation provide a baseline, modern high-end hardware enables the use of advanced Deep Learning architectures.
Wiener Estimation (The Linear Baseline)
The classical approach utilizes Wiener Estimation to minimize the mean square error between the estimated and actual spectra:
This method generates the initial 16-band approximation from the 3-channel input.
State-of-the-Art: Transformers and Mamba
For high-end hardware environments, we can utilize predictive neural architectures that leverage spectral-spatial correlations to resolve ambiguities.
- MST++ (Spectral-wise Transformer): The MST++ (Multi-stage Spectral-wise Transformer) architecture represents a significant leap in accuracy. Unlike global matrix methods, MST++ utilizes Spectral-wise Multi-head Self-Attention (S-MSA). It calculates attention maps across the spectral channel dimension, allowing the model to learn complex non-linear correlations between texture and spectrum. Hardware Demand: The attention mechanism scales quadratically \(O(N^2)\), requiring significant GPU memory (VRAM) for high-resolution images. This computational intensity necessitates powerful dedicated hardware to process the full data arrays.
- MSS-Mamba (Linear Complexity): The MSS-Mamba (Multi-Scale Spectral-Spatial Mamba) model introduces Selective State Space Models (SSM) to the domain. It discretizes the continuous state space equation into a recurrent form that can be computed with linear complexity \(O(N)\). The Continuous Spectral-Spatial Scan (CS3) strategy integrates spatial neighbors and spectral channels simultaneously, effectively "reading" the molecular composition in a continuous stream.
Computational Architecture: The Linux Python Stack
Achieving multispectral precision requires a robust, modular architecture capable of handling massive arrays across 16 dimensions. The implementation relies on a heavy Linux-based Python stack designed to run on high-end hardware.
- Ingestion and Processing: We can utilize rawpy (a LibRaw wrapper) for the low-level ingestion of ProRAW DNG files, bypassing OS-level gamma correction to access the linear 12-bit data directly. NumPy engines handle the high-performance matrix algebra required to expand 3-channel RGB data into 16-band spectral cubes.
- Scientific Analysis: Scikit-image and SciPy are employed for geometric transforms, image restoration, and advanced spatial filtering. Matplotlib provides the visualization layer for generating spectral signature graphs and false-color composites.
- Data Footprint: The scale of this operation is significant. A single 48.8MP image converted to floating-point precision results in massive file sizes. Intermediate processing files often exceed 600MB for a single 3-band layer. When expanded to a full 16-band multispectral cube, the storage and I/O requirements scale proportionally, necessitating the stability and memory management capabilities of a Linux environment.
The Spectral Solution
When analyzed through the 16 band multispectral pipeline:
| Spectral Feature | Ultramarine (Lapis Lazuli) | Azurite (Copper Carbonate) |
|---|---|---|
| Primary Reflectance Peak | Approximately 450 480nm (blue violet region) | Approximately 470 500nm with secondary green peak at 550 580nm |
| UV Response (below 420nm) | Minimal reflectance, strong absorption | Moderate reflectance, characteristic of copper minerals |
| Red Absorption (600 700nm) | Moderate to strong absorption | Strong absorption, typical of blue pigments |
| Characteristic Features | Sharp reflectance increase at 400 420nm (violet edge) | Broader reflectance curve with copper signature absorption bands |
Note: Spectral values are approximate and can vary based on particle size, binding medium, and aging.
Completing the Picture
The successful analysis of complex material properties relies on a convergence of rigorous physics and advanced computation.
- Photonic Foundation: The Sony IMX803 provides the necessary high-SNR photonic capture, with mode selection (binned vs. unbinned) driven by the specific analytical requirements of each examination.
- Data Integrity: DNG 1.7 is the critical enabler, preserving the linear relationship between photon flux and digital value while sequestering non-linear aesthetic adjustments in metadata.
- Algorithmic Precision: While Wiener estimation serves as a fast approximation, the highest fidelity is achieved through Transformer (MST++) and Mamba-based architectures. These models disentangle the complex non-linear relationships between visible light and material properties, effectively generating 16 distinct spectral bands from 3 initial channels.
- Historical Continuity: The EPR paradox of 1935 revealed that quantum particles share hidden correlations across space, correlations invisible to local measurement but real nonetheless. Modern spectral imaging reveals an analogous truth: materials possess hidden correlations across wavelength, invisible to trichromatic vision but accessible through mathematical reconstruction. In both cases, completeness requires looking beyond what direct observation provides.
This synthesis of hardware specification, file format stewardship, and deep learning reconstruction defines the modern standard for non-destructive material analysis — a spectral witness to what light alone cannot tell us.
And what about the paint? Here is a physical sample: pigment, substrate, history compressed into matter. Light passes through it, scatters from it, carries fragments of its story — yet the full truth remains hidden until we choose to look deeper. Every layer, every faded stroke, every chemical trace is a silent archive. We are not just observers; we are custodians of that archive. When we build tools to see beyond the visible, we are not merely extending sight — we are accepting a quiet responsibility: to bear witness honestly, to preserve what time would erase, to honor what has been made and endured. Light can expose structure. It cannot carry history. That part is on us. We can choose to let the machines we build serve memory rather than erasure, dignity rather than classification, truth rather than convenience. The past does not ask for perfection — it asks only that we refuse to let it be forgotten. In every reconstruction, in every layer we uncover, we have the chance to listen again to what was silenced. That is not just engineering. That is the work of being human.
References
- Einstein, A., Podolsky, B., & Rosen, N. (1935). Can Quantum-Mechanical Description of Physical Reality Be Considered Complete? Physical Review, 47(10), 777–780. ↑
- Bell, J. S. (1964). On the Einstein Podolsky Rosen paradox. Physics Physique Физика, 1(3), 195–200. ↑
- Aspect, A., Dalibard, J., & Roger, G. (1982). Experimental Test of Bell's Inequalities Using Time-Varying Analyzers. Physical Review Letters, 49(25), 1804–1807. ↑
Bryan R Hinton
bryan (at) bryanhinton.com

