A ground-up probability series for engineers and practitioners working with real-world signals — vibration sensors, audio streams, network telemetry, camera pixels, time series. The distribution chain — Bernoulli → Binomial → Poisson → Normal → CLT — is built step by step, with every part led by a concrete applied scenario (a vibration sample crossing an alarm threshold, bearing-fault counts per hour, a moving-average filter on a noisy signal) before the math arrives. The camera-sensor noise model is then worked through end-to-end as the cleanest physical instance of the chain — every link maps to a separate piece of silicon — and other sensors are shown to plug into the same template.
Parts¶
| File pair | Description |
|---|---|
part0_what_is_a_distribution | The full conceptual ladder — random process → sample space → event → random variable → distribution → parameters — built with two running examples (a synthetic word source and a vibration sensor stream). Signal-generic, stays in . |
part1_bernoulli | One vibration sample crosses the alarm threshold; the Bernoulli trial as the atom of randomness across signal-processing domains. |
part2_binomial | Counting threshold crossings in a 1-second vibration window; PMF built term by term; shape as a function of and ; Monte Carlo validation. |
part3_poisson | Bearing shock pulses per hour as the canonical rare-event count; Binomial limit as , ; the rule. |
part4_normal | The silent-sensor noise floor as the entry point to the bell curve; two parameters, Poisson → Normal convergence as grows. |
part5_clt | Moving-average filtering on a noisy stream as the entry point to the CLT; any distribution becomes Gaussian when summed; noise reduction; KS-distance convergence rate. |
part6_putting_it_together | The generic measurement chain (transduction → accumulation → electronics → ADC) worked end-to-end for the camera sensor; vibration / audio / network sensors plug into the same template. |
exercises | Four practice problems: Binomial/Poisson convergence rate, Anscombe transform, noise budget, CLT skewness. |
aside_high_dim_distributions | Forward-reference, optional. High-dimensional distributions, manifolds, what neural networks learn in feature space, distribution shift, adversarial examples, generative models. Read after parts 1–6; will be relocated to a Part IV chapter once the CNN material is in place. |
stochastic_processes | Placeholder. Time-indexed randomness — stationarity, autocorrelation, power spectral density, ergodicity, named processes (white noise, Wiener, Markov, Poisson). Outline only; will be filled in when Part II (Signals and Measurement) needs it. |
Running¶
Every .py file is standalone:
# from project root
source .venv/bin/activate
python math/probability/part0_what_is_a_distribution.py
python math/probability/part1_bernoulli.py
python math/probability/part2_binomial.py
python math/probability/part3_poisson.py
python math/probability/part4_normal.py
python math/probability/part5_clt.py
python math/probability/part6_putting_it_together.py
python math/probability/exercises.py # stub — complete the exercises firstWho links here¶
applied_sensors.md— the applied capstone of this series; uses the Bernoulli → Binomial → Poisson → Normal → CLT chain to derive the sensor noise model.nn-basics/fundamentals/math_concepts.ipynb— links to specific parts from §1 (random variables), §2 (Gaussian distribution), §5 (CLT), §6 (law of large numbers).
If you add a new downstream reference, list it here and link back to this directory so the cross-reference graph stays discoverable.