Encoding interface
Input encoders map digital vectors to optical excitation envelopes through calibrated couplers. Launch conditions are version-controlled to preserve reproducibility.
Documentation
This page consolidates the full explanatory content for CommonAccess: thesis, computation model, training pipeline, materials, verification, interoperability, and collaboration pathways.
CommonAccess builds a computing layer where machine learning inference is executed by physical dynamics rather than repeated digital instruction execution.
Core thesis: digital inference repeatedly reconstructs state; CommonAccess persists state physically and performs inference as response to perturbation.
Mechanism
Boundary
System topology: model definitions are compiled into physical operators, then exposed through runtime interfaces that preserve standard ML integration.
A Photonic Inference Plate (PiP) is a fabricated operator that maps input tensors to output tensors by wave propagation through learned material structure.
Model parameters are encoded as refractive distribution, scattering coupling, and threshold behavior. Inference is physical transformation plus readout.
Mechanism
PiP cross-section: model parameters are represented as refractive distribution, coupling geometry, and threshold behavior. Weight storage and transform behavior are unified in structure.
PiPs do not remove machine learning blocks; they relocate where they are executed. Tokens and encoded activations enter as optical boundary conditions, structural equivalents of weights are embedded in material geometry, and heavy transforms are executed by propagation.
Forward and backward roles are separated by phase: forward inference is physical on fabricated PiPs, while backward optimization and gradient updates are handled in differentiable simulation before fabrication. Softmax, temperature control, and some normalization remain digital at the readout interface in early releases.
Mechanism
Training still uses gradients and loss optimization in differentiable simulation. At inference time, heavy transforms are executed by fabricated structure and only normalization, control, and system orchestration remain digital.
Beyond weights, tokens, softmax, gradients, forward/backward, loss, and temperature, the primitives below define full-stack coverage and hybrid boundaries for PiP deployments.
| Primitive | Substrate mapping | Hybrid boundary | Primary validation metric |
|---|---|---|---|
| Attention (Q/K/V, masks) | Projection and score-forming transforms mapped to PiP propagation paths | Mask enforcement and final score normalization remain runtime-controlled | Attention quality parity and context-length stability |
| KV-cache semantics | Persistent optical-state and readout-assisted memory traces | Cache policy and compaction logic stay digital | Update latency and continuity error vs baseline caches |
| Positional encoding | Phase/amplitude modulation at encoder boundary conditions | Position scheme switching and runtime policy remain software-driven | Long-context degradation slope |
| Normalization (LayerNorm/RMSNorm) | Preconditioned physical response with calibrated readout statistics | Final normalization constants applied digitally in early releases | Distribution drift and output quality parity |
| Activations (GELU/SiLU/ReLU) | Nonlinear response regions and thresholded transfer bands in material design | Fallback nonlinear shaping available in runtime decode path | Approximation error and robustness under perturbations |
| Residual/skip connections | Interference-based merging channels and parallel propagation paths | Safety combine path at readout for controlled blending | Stability of deep-stack composition |
| MLP/FFN blocks | Expansion and projection transforms partitioned across PiP operator stages | Block orchestration and fallback routing remain in runtime | Throughput per token and quality retention |
| Optimizer states (Adam moments, clipping) | Design-time differentiable simulation and compiler parameter updates | Not in deployed PiP inference loop | Convergence speed and fabrication-ready candidate quality |
| Quantization and precision policy | Readout dynamic-range design with calibrated sampling points | ADC/DAC precision policy controlled in runtime stack | Accuracy vs power and thermal envelope |
| Sampling policy (top-k/top-p, penalties) | PiP provides logits and stateful score trajectories | Sampling and policy logic remain digital by design | Response quality and controllability metrics |
| Batching and sequence scheduling | Shared propagation windows and readout cadence planning | Scheduler remains software-controlled for QoS guarantees | Latency distribution under mixed-load scenarios |
| Sparse routing / MoE and specialty blocks | Route-specific operator banks and selective mode activation | Router policy and expert arbitration remain runtime-managed | Routing overhead and expert quality parity |
Device-level understanding is managed through multiple synchronized views: top routing, cross-sectional material stack, and readout planes. Each view contributes independent constraints to fabrication readiness.
This section tracks how encoding and decoding hardware interfaces are physically realized, where mode coupling is intentional, and where energy and thermal controls are implemented.
Device-level representations include top routing geometry, material stack cross-section, and readout plane instrumentation. Each view is used to track fabrication readiness and runtime calibration reliability.
Input encoders map digital vectors to optical excitation envelopes through calibrated couplers. Launch conditions are version-controlled to preserve reproducibility.
Readout taps are sampled by sensing arrays, then passed through calibrated ADC and decode software to produce logits or task outputs.
Sparse readout cadence, bounded perturbation updates, and thermal-aware scheduling reduce peak energy spikes and protect latency stability.
Runtime guardrails detect drift, crosstalk growth, or calibration decay and automatically shift to conservative sampling profiles before quality is affected.
PiP rollout is staged by mode complexity and control maturity. Early releases prioritize deterministic single-mode behavior, while later releases expand to multi-mode fabrics only after crosstalk and thermal bounds are consistently controlled.
Full-scale production is defined by stable yield, calibration repeatability, and benchmark commitments under deployment envelopes, not by maximum mode count.
Mode expansion is staged. Production-scale rollouts are gated by crosstalk bounds, calibration stability, yield, and thermal resilience rather than mode count alone.
Each release class defines allowed mode families, coupling bounds, and crosstalk ceilings. Promotion requires all stability metrics to pass.
Rollout specs include mode isolation error, calibration retention window, thermal drift tolerance, and minimum manufacturing yield threshold.
Throughput scaling is accepted only when latency distribution remains bounded across context growth and multi-mode operation.
Energy-per-update slope and peak transient limits are tracked per mode class; noncompliant profiles block release escalation.
The objective is not novelty for its own sake. PiPs are pursued because several inference bottlenecks in conventional digital execution are structural and persist even with faster GPU or NPU generations.
Many modern workloads repeatedly rebuild context at each step. This creates avoidable compute and memory traffic overhead in long-running inference paths.
Even optimized digital stacks spend significant budget moving activations and parameters. PiPs target reduced movement by embedding transform behavior in physical structure.
Robotics, edge sensing, and long-context agents operate continuously. A substrate with embodied state is better aligned with these always-on operating conditions.
The next efficiency gains are unlikely to come from execution speed alone. This program explores a compute path where part of inference is delegated to material dynamics.
Hardware is not trained directly. A differentiable physical simulation is trained and optimized, then transformed into fabrication-ready structure.
Training path: dataset to simulation to optimized material geometry to fabrication.
Mechanism
Training complexity remains at design time. Fabricated operators contain learned mappings without retaining datasets or digital weight files at inference time.
CommonAccess competes on manufacturing class and stable physical response, not transistor density or clock speed.
Material and geometry choices are advanced only when repeatability, separability, and drift tolerances pass validation gates.
Development proceeds only when validity gates are satisfied: deterministic mapping, high-rank dimensionality, linear separability, and temporal stability.
Workload fit: continuous perception, persistent agents, long-context reasoning, robotic cognition, and always-on inference systems.
Evidence
Boundary
Test method
repeatability run set
Gate
Input class variance < 1.5% across 1000 cycles
Test method
SVD on output state matrix
Gate
Effective rank exceeds benchmark floor per task family
Test method
readout classifier probe
Gate
Probe accuracy above digital baseline threshold
Test method
drift and hysteresis monitoring
Gate
Stability window remains within calibration band
PiPs are positioned to outperform conventional inference pipelines under persistent, context-heavy workloads. Commitments below define how those claims are tested against GPU, NPU, and digital accelerator baselines.
Claim: PiP token throughput degrades materially less with context growth than conventional GPU/NPU inference loops.
Benchmark: fixed model family, progressive context ladder, compare slope of tokens-per-second decay.
Claim: PiP inference energy tracks perturbation magnitude, delivering lower cost than repeated full recomputation on digital accelerators.
Benchmark: joules per update across identical workload traces on PiP, GPU, and NPU paths.
Claim: PiPs maintain usable response continuity between inputs without full hidden-state rebuild at each step.
Benchmark: multi-step agent/perception traces, measure recovery cost and continuity error vs conventional pipelines.
Claim: PiP systems sustain useful inference under tighter power and thermal envelopes than comparable GPU/NPU deployments.
Benchmark: constrained edge environments, compare sustained performance at fixed thermal and power ceilings.
Baselines: representative GPU, NPU, and digital inference configurations for the same task class.
Inputs: shared datasets and workload traces, identical output quality targets.
Metrics: tokens/sec, energy per update, latency distribution, continuity error, thermal stability.
Publication: each promise reported with reproducible setup and baseline-relative delta.
CommonAccess is not aimed at all computing tasks. The strongest fit is in persistent, context-heavy systems where repeated recomputation dominates energy and latency budgets.
Applied use case: always-on perception and control loops with persistent scene context.
Disruption vector: shift from repeated frame-level context reconstruction to stateful response updates.
Applied use case: persistent multi-turn reasoning agents with large evolving context windows.
Disruption vector: reduce token-by-token full-context recomputation pressure and context-length sensitivity.
Applied use case: continuous anomaly detection on streaming sensor fields.
Disruption vector: transition from batch-like periodic inferencing to continuous physical state tracking.
Applied use case: local inference nodes under strict power and thermal limits.
Disruption vector: replace repeated digital heavy transforms with fabricated physical operators for persistent workloads.
Developers should use a standard ML workflow: model to compile to run. CommonAccess is a new backend, not a new programming model.
Required software layers: model compiler, simulation SDK, hardware runtime driver, and framework integration adapters.
Open to research labs, robotics developers, model developers, and hardware partners.
Contact: agni@comac.network
Include research focus, interface requirements, and expected collaboration window.
Stage 1
Validate operator fidelity across benchmarked transforms and stability windows.
ExitDeterministic mapping and rank threshold confirmed
Statusin validation
Stage 2
Establish repeatable fabrication pathways for tunable and task-specific structures.
ExitFabrication variance remains within acceptable tolerance band
Statusplanned
Stage 3
Expose compiler interfaces for mapping model structure to physical geometry.
ExitCompiler to runtime handoff validated across test suites
Statusplanned
Stage 4
Integrate substrate, readout, and software interfaces into deployable systems.
ExitEnd-to-end integration passes persistent workload trials
Statusplanned