Українською
  In English
Збирач потоків
The story of 10 years, 10 PCBs, and everything I got wrong building a WiFi sub-PPB clock sync device
| Story time submitted by /u/johny1281 Let me tell you a story about how a simple idea turned into a 10 year obsession. The end result is a tiny (4cm x 4cm) battery-powered node that syncs its clock to other nodes over the air with sub-PPB accuracy. No cables between them. You drop them wherever you want and they self-synchronize. I use it for phase-coherent Wi-Fi measurements across multiple receivers, which lets you do things like angle-of-arrival estimation and indoor localization. But getting here was not pretty. Board 0: The Aliexpress dev kit. I just finished my master's thesis and I've never made a PCB. I grab two ESP32 dev kits, learn how to flash them, learn how to capture Wi-Fi phase data. The data is pure noise. I spend months staring at random numbers before I understand why. Turns out a 10 ppm crystal gives you about 24 full phase rotations between consecutive Wi-Fi frames. Indistinguishable from random. Cool. Board 1: The Chinese flasher board. Before spending real money on a custom PCB I want to make sure I can flash bare ESP32 modules. Got this little Chinese jig, drop the chip in, flash over UART. Works first try. Good confidence boost. Still a garbage clock though. Board 2: First custom PCB ever. This is the big jump. Real money, real components, real chance of screwing up. EasyEDA, auto-router, fingers crossed. I try hand soldering the first batch and destroy every single one. Switched to solder paste and a $30 hot plate. Suddenly everything works beautifully. Same garbage data, but at least I stopped burning money on dead boards. Baby steps. Board 3: The "just share the clock" idea. Upgraded to a 0.5 ppm TCXO and tried to share it between two chips via jumper wires. Seemed so obvious. The parasitic capacitance of even short wires killed the signal dead. Touching a finger near the wire did the same thing. On the plus side, I discovered that the ESP32-C3 has hidden nanosecond RX timestamps buried in the firmware structs. That discovery ended up being the foundation of everything later. Board 4: The SMA cable attempt. Proper 50 ohm coax should fix the clock distribution problem, right? Somehow worse than bare wire. Also picked a clock buffer where the oscillator output was below the CMOS input threshold, so the buffer did literally nothing. Most expensive useless board of the project. Board 5: Two chips, one PCB, as close as physically possible. If cables don't work, just put the oscillator millimeters from both chips. No wires, no connectors, just traces. And it worked! First time I ever saw coherent phase. But the PCB antenna couldn't transmit (2-layer board, matching was completely wrong), and I measured about 1 ppb drift between two chips sitting 5mm apart. Thermal gradients. They're not at exactly the same temperature even when they're neighbors. Board 6: Scale to four chips. Got ambitious. Shared the voltage regulators because I didn't know you can't parallel LDOs. Only 2 out of 4 would boot. External SMA antennas made it the size of a shoebox. Back to the drawing board. Board 7: Remove the ground plane under the clock. Read somewhere online that ground pour causes interference near clock lines. Removed it. Everything got worse. Missing edges on the scope. Noise everywhere. Put it back. Don't believe everything you read. Board 8: Four layers, proper matching. Finally understood why every app note says 4 layers. On a 2-layer board the signal-to-ground distance is too large, coupling is loose, trace dimensions make no sense. On 4 layers everything behaves like the textbook. All 4 chips synced. But all 4 PCB antennas were coupled through the shared ground plane. PCB antennas use the ground as part of the radiating structure. Shared ground = shared antenna. Touching one killed the others. 20 dB down from a reference module. Board 9: Stop sharing clocks entirely. The breakthrough. Give each node its own voltage-controlled oscillator. Measure drift over the air using Wi-Fi timing exchanges. Correct with a DAC on the oscillator's tuning pin. One DAC got me to 10 ppb but each step was too coarse. Added a second DAC in a 1:30 ratio, coarse to get close, fine to hold steady. Sub-PPB. No shared ground, no coupled antennas, no cables. Each node is 4cm x 4cm and battery powered. Board 10: ESP-PPB. A few more boards in between with minor tweaks, but the big addition was the dual-DAC setup. 1 ppb typical in the open. 0.1 ppb in a stable enclosure, which is the measurement floor of the hardware. Oh and one more fun discovery: the radio silently compensates for frequency mismatch between sender and receiver internally. If two boards don't land on the same correction value, your data is garbage and you won't know why. With synced clocks they always agree. With unsynced clocks it's a coin flip. That one cost me months. Everything is open sourceEverything is open source. Firmware, schematics, Gerbers, BOM, 3D model. There's a story.md in the repo with photos of every board and what went wrong each time: https://github.com/jonathanmuller/esp-ppb Ask me anythingWhat was hardest, what was easiest, what I'd redo completely. This has been my side project for a decade and I'm happy to talk about any of it. [link] [comments] |
India to Boost Local Chip-making with a 1 Trillion Rupees Funding
India is expected to unveil a funding of more than a trillion rupees, nearly $10.8 billion, to fuel domestic chip-making. With 10 projects already approved, India aims to become a global manufacturing hub in the near future.
This new set of funds is expected to provide subsidies for chip designing projects, supply chain developments, and manufacturing equipment. Currently, the plan is under evaluation and may see the green light in a couple of months.
With the accelerated demand for chips driven by the rise of AI and electronics, the market is evolving at an exponential rate. Under the current leadership, India aims to position itself at the top to meet global demand.
With companies like Micron, TATA, and Foxconn already building India’s chip ecosystem, the country is expected to be close to industry leaders such as Taiwan and the US by 2032.
The post India to Boost Local Chip-making with a 1 Trillion Rupees Funding appeared first on ELE Times.
OK, this book is awesome!
| Every connector under the sun is here. Plus it has IC interconnects so this post is technically not breaking the rules :) Thanks Davide for this great resource! [link] [comments] |
Prototype HV DC buck converter running on a PCB I fabricated with a fiber laser
| This is a quick prototype HV DC buck board I built using the fiber-laser PCB process I posted earlier. Still experimenting with trace limits and thermal performance, but it's working surprisingly well so far. [link] [comments] |
E-ink mp3 player
| This is V2 of my e-ink DAP project, it has :
V1 horribly failed, here is what changed since then:
The firmware is still in very early stages, I still haven't implemented a ton of features that the hardware is capable of, like DSP, Bluetooth, etc. I also need 3D print the case in resin, so it doesn't look like this, I want to use transparent resin The whole project is open source: GitHub [link] [comments] |
SpiceCrypt: open-source decryption tool for LTspice-encrypted .CIR/.SUB model files
| submitted by /u/jtsylve [link] [comments] |
USBpwrME
| Every time i want to do an experiment in the lab and use USB power to my DUT i need to find a cabler with correct connector and thick wires enough for the purpose and then cut it :(:( to be able to connect it to my bench power supply. So finally i decided to solve this reoccurring issue with a universal adaptor that will solve all my challenges and stopping me cutting cable after cable. This led up to designing the small adaptor that fits most power boxes since it has moveable banana binding posts. I have added polarity protection and over voltage protection that can be disabled to make it flexible and pass thru voltages from 3-20V out to the USB-A and USB-C connector. I have also added charging negotiation circuits for both USB-A (up to 10W @ 5V) and USB-C (up to 15W@ 5V). The adaptor can handle up to 6A so it will work for most application!! I have worked a lot with heat managment and tried to keep low resistance in the current paths. When loading max the hottest component reaches around 85 degrees C in room temp [link] [comments] |
Digi-key; A small U.S. town grew a big company. Can it weather the tariff blizzard?
| submitted by /u/1Davide [link] [comments] |
Weekly discussion, complaint, and rant thread
Open to anything, including discussions, complaints, and rants.
Sub rules do not apply, so don't bother reporting incivility, off-topic, or spam.
Reddit-wide rules do apply.
To see the newest posts, sort the comments by "new" (instead of "best" or "top").
[link] [comments]
Spent hours troubleshooting to find out I got my PFETs backwards qnq
| I’m attempting to make an LED scoreboard for my cricket team using large 7‑segment LED displays. I want it to be battery powered, so I’m trying to reduce the power needed to run 6+ digits at once by using multiplexing. Each segment is connected to a high‑side switch, and the digits to the low‑side. That way I can turn on each digit by pulling it low, and only the segments held high will activate. The code I’m using runs on an Arduino, which talks to a cheap PCA8695 PWM board. That board connects to a custom MOSFET driver board that handles the high‑ and low‑side switching. Running code that worked fine in my prototype setup just gave me an epileptic strobing effect on all segments, which completely threw me. I spent hours probing with a multimeter, using the oscilloscope at work, and eventually started cutting “non‑essential” components off the board. Instead of getting an inverted 12 V PWM signal like I expected, I was constantly getting a square wave oscillating between 12 V and 11.5 V no matter what I did. I was about to post on r/AskElectronics for help, but I wanted to be 110% sure I wasn’t missing something obvious. So I went to falstad.com and built the circuit in the simulator. Sure enough, it behaved exactly how I expected. Then I noticed a little checkbox for “Swap D/S,” and out of curiosity I clicked it… bingo. For testing, I’m going to desolder the PFETs I’ve got and jankily wire them in upside‑down just to confirm that’s the issue before ordering new ones. Moral of the story: make sure you’re using the right datasheet for your parts, because manufacturers love reusing part numbers even when the pinouts are completely different. (p.s. pls don't be too mean about diagram conventions, signal noise, etc. cos this is a self-taught learning exercise and I'm trying my best) [link] [comments] |
30-minute PCB fabrication with a fiber laser (double-sided boards)
| I've been experimenting with using a fiber laser to fabricate prototype PCBs. Current workflow: - design PCB - laser isolate traces - drill vias - clean - solder Total time from design to board is about 30 minutes. Trace pitch so far is around ___ mil and I've been able to do reliable double-sided boards. I made a video showing the full process and the relaxation oscillator circuit I designed for it: [link] [comments] |
Exploration Alternatives of Component Marketplaces
| The goal was to find where to buy electronics that i need(STM32F103C8T6 and STM32F401RET6), but figured it will be cool if i put all that in one post. Maybe someone finds it interesting. [link] [comments] |
IFW Dresden selects Agnitron Agilis 100 MOCVD platform for precursor chemistry and ultra-wide-bandgap materials development
Фінансово-бюджетний звіт за 2025 рік (МОН)
📰 Газета "Київський політехнік" № 9-10 за 2026 (.pdf)
Вийшов 9-10 номер газети "Київський політехнік" за 2026 рік
TNO and High Tech Campus Eindhoven begin construction of first 6-inch indium phosphide photonic chip foundry
Balun transformers: Linking balanced to unbalanced

Balun transformers remain indispensable in RF and high-frequency design, serving as the quiet interface between balanced transmission lines and unbalanced circuits. By enabling impedance matching, minimizing signal distortion, and suppressing common-mode noise, they provide the foundation for reliable connectivity in applications ranging from antennas to amplifiers to broadband communication systems.
As wireless technologies push toward higher frequencies and tighter integration, understanding the principles and practical nuances of balun transformers is key to optimizing performance and ensuring design resilience.
The term “balun” itself comes from balanced to unbalanced. While many implementations use transformer coupling, not all baluns are transformer-based—some rely on transmission line techniques. Using “balun transformer” specifies the transformer-type design, distinguishing it from coaxial sleeve or other non-transformer baluns.
Historic note: The iconic TV balun adapter
Before digital tuners and streaming boxes took over, this compact 300 Ω to 75 Ω matching transformer was a fixture in analog television setups. Designed to reconcile the impedance and mode mismatch between twin-lead ribbon antennas and coaxial inputs, it featured screw terminals for the antenna wire and a standard coaxial plug for the TV’s antenna input socket.
Connected at the final stage of the antenna lead and plugged directly into the tuner, it quietly performed its dual role—impedance transformation and balanced-to-unbalanced conversion. This ensured that rooftop signals reached living rooms with minimal distortion. In the analog broadcast era, this unassuming adapter was the last link in the RF chain, faithfully bridging generations of antenna technology.

Figure 1 Screwing the 300 Ω ribbon cable into the balun terminals and plugging its coaxial end into the TV’s antenna input socket completes the balanced-to-unbalanced transition. Source: Author
Video balun transformers: Bridging coax and twisted pair
Video balun transformers—more commonly referred to simply as video baluns in industry parlance—extend the utility of balun technology beyond RF and audio domains into the realm of video signal transmission. These devices convert unbalanced coaxial signals (such as composite video) into balanced signals suitable for twisted-pair cabling, and vice versa.
This conversion not only reduces susceptibility to electromagnetic interference (EMI) but also enables cost-effective long-distance video distribution using standard Cat5/Cat6 cabling. Passive video baluns rely on transformer coupling to maintain signal integrity without external power, while active baluns incorporate amplification and equalization to support higher resolutions or longer cable runs.
In surveillance and broadcast applications, video baluns have become indispensable for bridging legacy coaxial infrastructure with modern structured cabling, ensuring clean signal delivery and simplified installation.

Figure 2 Video baluns connect coaxial BNC interfaces to twisted-pair cabling and deliver HD CCTV signals over long distances with reduced interference. Source: Author
As a quick aside, it’s worth noting that the K and MP ratings of a video balun both denote its supported resolution class. The MP rating specifies the maximum camera resolution in megapixels, while the K rating expresses the same capability in terms of horizontal pixel count.
In practice, both ratings reflect the balun’s bandwidth and signal-handling capacity for HD CCTV. For example, a 4K balun supports roughly 8 megapixels of resolution, since 3840 × 2160 pixels equals about 8.3MP (8.3 million pixels).
Baluns in practice: Theory meets application
Balun transformers are invaluable not only for converting between balanced and unbalanced signals but also for performing impedance transformations with minimal loss. Unlike LC circuits, many balun designs can operate effectively across very wide frequency ranges.
In RF applications, baluns are commonly used to interface antennas with transmitters and receivers, ensuring that as much power as practically possible is delivered. This session blends accessible theory—without heavy mathematics—with a few practical pointers and real-world implementations.
Among the fundamental designs, the balun transformer is the most widely recognized. Using magnetic coupling, it converts between balanced and unbalanced signals while providing excellent isolation and impedance matching. Transmission-line baluns achieve balance through carefully arranged lengths of coaxial or twisted-pair lines, making them well-suited for wideband RF applications.
Hybrid baluns combine transformers and transmission-line techniques, offering flexibility across frequency ranges. Together, these basic types form the foundation for more advanced designs, and understanding their principles helps engineers and experimenters select the right balun for applications ranging from antenna systems to CCTV.
In practice, the terms “balun transformer” and “transformer balun” both refer to the same device: a balun realized through transformer coupling. The difference is mostly in emphasis. Balun transformer highlights the function first—balanced-to-unbalanced conversion—while noting that it’s implemented as a transformer.
Transformer balun highlights the construction first, pointing out that it’s a transformer adapted to serve as a balun. Both usages are common, but in technical writing “balun transformer” is often preferred because it stresses the primary role of the device.
A further distinction often made is between voltage baluns and current baluns. A voltage balun enforces equal voltages on the balanced output terminals, which can work well in many cases but may allow unequal currents if the load is not perfectly symmetrical. In contrast, a current balun enforces equal and opposite currents in the balanced lines, often providing better suppression of common-mode currents on antenna feedlines.
Both approaches have their place: voltage baluns are straightforward and widely used, while current baluns are often preferred in RF antenna systems where minimizing feedline radiation and maintaining balance are critical.
Also essential to audio systems, baluns form the core of passive direct injection (DI) boxes. A passive DI employs a transformer—acting as a voltage balun—to convert an unbalanced, high-impedance instrument signal into a balanced, low-impedance output. This conversion is vital for interfacing high-Z sources such as electric guitars with low-Z mixing console inputs over long cable runs.
By enforcing equal and opposite voltages on the balanced lines, the transformer achieves high common-mode rejection, suppressing noise and ensuring transparent signal transfer. This application demonstrates how the balancing principles fundamental to RF and CCTV extend seamlessly into professional audio, underscoring the cross-domain versatility of balun technology.

Figure 3 A passive DI box handles extreme signal levels without introducing any distortion. Source: Radial Engineering
Seemingly, instead of diving straight into balun transformer–based RF or video projects, makers may find it easier—and just as rewarding—to begin with a closely related audio build: the passive DI box. Ready-to-use direct box transformers are widely available, and their simplicity makes them an ideal starting point for a fun and accessible DIY project.
Notable part numbers include JT-DB-EPC and A187A10C, both excellent examples of components that make this project approachable for beginners. The Hammond 1140-DB-A is another great catch, offering a versatile option for those eager to experiment with high-quality audio designs.

Figure 4 The 1140-DB-A direct box transformer delivers a balanced microphone output from an unbalanced line-level signal, enabling long cable runs with minimal high-frequency loss. Source: Hammond
From first steps to deeper layers
As is often the case, we have only just wetted our feet—there is still a vast ocean of balun transformer theory, design variations, and application nuances left to explore. From specialized wideband implementations to creative DIY builds, each path opens new insights into how these deceptively simple devices shape signal integrity across RF, audio, and video domains.
This overview is meant as a starting point, a foundation for deeper dives into the many layers of balun transformer technology that await.
Your turn: If this sparked your curiosity, take the next step—experiment with a simple antenna balun build, revisit your audio gear with fresh eyes, or explore advanced designs in RF literature. Share your experiences, questions, or even your own schematics, because the best way to deepen understanding is to connect theory with practice.
T. K. Hareendran is a self-taught electronics enthusiast with a strong passion for innovative circuit design and hands-on technology. He develops both experimental and practical electronic projects, documenting and sharing his work to support fellow tinkerers and learners. Beyond the workbench, he dedicates time to technical writing and hardware evaluations to contribute meaningfully to the maker community.
Related Content
- Sevick’s Transmission Line Transformers, Baluns
- Delicate balancing acts ensure balun performance
- Understand baluns for highly integrated RF modules
- Harmonic balance simulation speeds RF mixer design
- Using Baluns and RF Components for Impedance Matching
The post Balun transformers: Linking balanced to unbalanced appeared first on EDN.
🚀 Інженерні тижні «KPISchool» для учнів 9–11 класів у КПІ ім. Ігоря Сікорського
З 23 березня по 28 березня 2026 року в Національному технічному університеті України «Київський політехнічний інститут імені Ігоря Сікорського» відбудуться Інженерні тижні «KPISchool» — освітній профорієнтаційний захід у межах проєкту «Майбутній КПІшник».
Designing the voice AI stack: Integrating spatial hearing AI with edge-based intent gating

We’re past the point where voice can be treated as just another feature.
For more than a decade, the smart home has operated under a flawed assumption: that voice is optional. It’s not. As homes grow more complex and connected, voice is the only interface that aligns with how people actually live.
Traditional interfaces don’t scale: touchscreens fail when your hands are full, apps demand too much attention, and remotes are always missing when you need them. Voice is the only input that works across rooms, contexts, and users, if it works reliably.
And yet, we’re still tethered to physical buttons and remote controls, because we don’t fully trust voice interfaces. They miss commands, struggle in noisy environments, and break the moment connectivity becomes unstable. That’s not a UI flaw. It’s an architectural one.
To replace the light switch, voice needs to be always available, always accurate, and always in context. That means rethinking where intelligence lives and how decisions are made.
Hybrid Voice AI architecture is not an incremental upgrade, it’s an engineering breakthrough that transforms the smart home from a scattered set of reactive gadgets into a cohesive, proactive system. By separating real-time, on-device reflexes from deep, cloud-based reasoning, this architecture is designed to make voice a trusted, primary interface, every time, in every room.
Making voice work in the real worldThe flaw in current voice technology isn’t a lack of data; it’s a lack of clarity
Real homes are acoustically chaotic. They’re full of overlapping conversations, background music, household noise, and hard surfaces that introduce echo and reverb. Users speak from different rooms, distances, and angles. Commands are often ambiguous or incomplete. These aren’t edge cases. They’re the default operating conditions.
Current cloud-only models are powerful but slow, while legacy on-device models are fast but dim-witted. Neither alone can deliver the “Star Trek” experience users crave. To achieve the non-negotiable standard of 100% reliability, we need a system that mimics the human brain’s ability to process reflexes locally and complex thoughts deeply
In that context, today’s voice interfaces consistently fall short. Not because of a lack of data or model size, but because of fundamental architecture-level decisions about where processing happens, how quickly systems respond, and how they handle failure.
A symbiotic two-tier architectureThe innovation lies in splitting the intelligence. By decoupling immediate execution from deep reasoning, we create a system that is both instant and intelligent.
- The Reflex Layer – Edge AI (Supports Instant Response):
- Definition: Think of this as the smart home’s autonomic nervous system.
- Innovation: High-performance, always-on SLM embedded directly on the device’s silicon.
- Function: Handles the “here and now.” Commands like “Lights on” or “Volume down” are processed locally with near-zero latency.
- Impact: Delivers absolute privacy and instant responsiveness. No data leaves the room, and the experience feels as immediate as flipping a physical switch.
- The Reasoning Layer – Cloud AI (Intelligent Coordination):
- Definition: This acts as the system’s prefrontal cortex—responsible for reasoning.
- Innovation: Leverages large language models (LLMs) to manage long-term state, memory, and complex logic across devices and use cases.
- Function: Handles the “what if” and “what next.” It manages household routines, coordinates multiple devices, and draws inferences from incomplete inputs (e.g., “Order dinner for whoever is home tonight.”)
- Impact: Enables devices to go beyond command execution—they begin to understand intent, anticipate user needs, and adapt over time (Figure 1).

Figure 1 A hybrid voice stack routes audio through on-device perception (AEC, spatial analysis, separation, intent gating) and escalates only complex requests to cloud reasoning. (Source: Kardome)
Differentiation for the decade aheadFor OEMs and Tier 1 suppliers, architecture, not features, is emerging as the defining battleground for the next generation of smart home systems.
The market is saturated with devices that can set timers, play music, or toggle lights. These capabilities are now commodity. What will set future systems apart is their ability to demonstrate true Auditory Intelligence—to perceive, localize, and interpret human speech reliably, even in noisy, multi-speaker, real-world environments.
By integrating spatial hearing AI and cognition technologies into a hybrid architecture, manufacturers can go beyond individual product features and instead build the auditory nervous system of the modern home.
We are past the era of voice assistants that require users to repeat themselves or speak in rigid syntax. Hybrid Voice AI enables a different class of experience—one where technology is felt, but rarely seen.

Figure 2 Spatial processing turns a mixed audio scene (TV + two speakers + reverb) into separated target streams suitable for intent detection and command execution. (Source: Kardome)
What “reflex vs. reasoning” meansIn a production voice system, “hybrid” isn’t simply “ASR on-device and an LLM in the cloud.” It’s a routing architecture with a continuously running perception pipeline that decides:
- Is anyone speaking?
- Who is speaking (and where)?
- Is it directed at the device?
- Can we execute locally, or do we need cloud reasoning?
A practical edge “reflex” stack typically includes:
- Acoustic front end (always-on): microphone capture → gain control / denoise → echo cancellation (to remove the device’s own playback).
- Spatial scene analysis: estimate how many sources exist and where they are relative to the device (near/far, left/right, different rooms).
- Source separation + target selection: isolate the intended speaker stream(s) and suppress competing sources (TV, music, second speaker).
- Speech activity detection + endpointing: stable detection of speech start/stop to avoid clipped commands and reduce false triggers.
- Device-directed intent gating (SLM): a lightweight model answers: “Is this speech for the device?” using spatial cues + conversational flow + linguistic signals.
- Execution vs. escalation:
- Local path: deterministic actions and short commands (“lights on,” “stop,” “volume down”) with minimal latency.
- Cloud path: long-horizon reasoning, multi-device planning, and tasks requiring external knowledge—only when needed.
The engineering advantage is that the system can stay fast and predictable for everyday commands while still enabling deeper capabilities when appropriate.
Why spatial audio is the “make or break” layerMost failures in today’s voice assistants begin before language: the system is fed garbage audio (mixed speakers, reverberation, background media), then asked to “understand” it. Hybrid architectures push the hard work earlier: fix the audio scene first, then do language.
Spatial processing matters because it enables three foundational capabilities:
- Localization: determine where speech is coming from and whether it’s in the same room.
- Separation: isolate a voice even with overlapping speakers and media noise.
- Attribution: reduce wrong-room actions and improve “who said what” reliability.
This is also where direction of arrival (DOA)-only approaches struggle in real homes: reflective surfaces create strong echoes and multiple delayed arrivals. A “flat” directional estimate can become unstable under reverb, causing separation and attribution errors. A more robust approach treats each source as having a unique spatial signature (an “acoustic fingerprint”) and uses that signature to stabilize separation and tracking over time.
Latency, offline behavior, failure modesIf voice is going to replace physical controls, reliability can’t be an aspiration—it has to be engineered with explicit budgets and test matrices.
Latency budgetHumans pause roughly ~200ms between conversational turns, while cloud round trips often land in the 1–3 second range—good enough for Q&A, not good enough for control.
The reflex path should therefore be designed so the most common commands complete without waiting on the network.
Offline and “brownout” modesDefine tiers of capability that remain functional without connectivity:
- Tier A (must work offline): lights, volume, stop/quiet, timers, basic routines.
- Tier B (cloud-required): deep reasoning, external services.
This avoids a binary “voice works / voice is dead” experience and increases user trust.
Failure modes that must be tested (not treated as edge cases)
- overlapping speakers (barge-in, crosstalk)
- competing media (TV/music)
- far-field speech + occlusion (speaker in hallway / adjacent room)
- changing echo paths (content and volume changes)
- reverberant rooms (kitchen tile, open-plan living spaces)
Metrics that map to trust (beyond WER):
- end-to-end command success rate by scenario class
- false accept / false reject rates for device-directed intent gating
- speaker attribution / room attribution accuracy
- P95 latency (not just average) for Tier A commands
- recovery time after connectivity loss
A counterintuitive benefit of edge-first reflex layers is that they can be more private and more cost-stable than cloud-streaming approaches—because a large fraction of everyday interactions can be processed locally, and the cloud is invoked only when deeper reasoning is necessary.
On the economics side, cloud inference costs scale with usage, while edge compute is amortized with silicon volume and can reduce the need for continuous cloud processing for trivial requests.
One example of this architectural direction is Kardome, which focuses on combining spatial hearing (to separate and localize voices) with an on-device context-aware SLM (to decide whether speech is directed at the system), escalating to the cloud only when deeper reasoning is needed.

Dr. Alon Slapak is the co-founder and CTO of Kardome, a voice AI startup pioneering Spatial Hearing and Cognition AI technology that enables seamless, natural voice interaction in real-world noisy environments. He holds a Ph.D. from Tel Aviv University and brings deep expertise in acoustics, signal processing, and machine learning. Alon and co-founder and CEO Dr. Dani Cherkassky launched Kardome out of a shared passion for solving end-user frustrations with voice devices, combining their expertise in acoustics and advanced machine learning to build leading-edge voice user interface technology. Kardome has raised $10M in Series A funding.
Related Content
- Sparse AI MCU facilitates voice processing and cleanup at edge
- IoT: GenAI voice helps generate speech recognition models
- AI noise suppression for a better listening experience
The post Designing the voice AI stack: Integrating spatial hearing AI with edge-based intent gating appeared first on EDN.



