EmbeddedRelated.com
The 2026 Embedded Online Conference

Ten Little Algorithms, Part 2: The Single-Pole Low-Pass Filter

Jason SachsJason Sachs April 27, 201517 comments

Jason Sachs shows how a single-pole IIR low-pass filter, implementable in one line y += alpha * (x - y), tames noise in embedded signals without floating point. The post explains how to compute alpha from tau and delta-t, practical tradeoffs like phase lag and oversampling, and fixed-point pitfalls including how many extra state bits you need to avoid quantization. Short, practical, and code-ready.


Ten Little Algorithms, Part 1: Russian Peasant Multiplication

Jason SachsJason Sachs March 21, 20156 comments

Jason Sachs revisits a centuries-old multiplication trick and shows why it still matters. He lays out Russian Peasant Multiplication with simple Python code, then reveals how the same shift-and-add pattern maps to GF(2) polynomial arithmetic and to exponentiation by squaring. The post mixes historical context with practical bitwise techniques that are useful for embedded and low-level math work.


Unraveling the Enigma: Object Detection in the World of Pixels

Charu PandeCharu Pande February 8, 2024

Exploring the realm of embedded systems co-design for object recognition, this blog navigates the convergence of hardware and software in revolutionizing industries. Delving into real-time image analysis and environmental sensing, the discussion highlights advanced object detection and image segmentation techniques. With insights into Convolutional Neural Networks (CNNs) decoding pixel data and autonomously extracting features, the blog emphasizes their pivotal role in modern computer vision. Practical examples, including digit classification using TensorFlow and Keras on the MNIST dataset, underscore the power of CNNs. Through industry insights and visualization aids, the blog unveils a tapestry of innovation, charting a course towards seamless interaction between intelligent embedded systems and the world.


Monte Carlo Integration

Jason SachsJason Sachs March 16, 2026

Monte Carlo integration looks deceptively simple, estimate an area by throwing random points at it and counting hits. Jason Sachs uses that idea to approximate pi, compare error scaling, and then show why the same approach becomes far more useful in higher dimensions. He also demonstrates a stratified sampling trick that improves accuracy by spending samples where they matter most.


The 2026 Embedded Online Conference