Українською
  In English
Збирач потоків
Microchip Introduces 600V Gate Driver Family for High-Voltage Power Management Applications
The post Microchip Introduces 600V Gate Driver Family for High-Voltage Power Management Applications appeared first on ELE Times.
⛩️ Запрошуємо на відкритий турнір Кубка Посла Японії з шьоґі 2026!
7-8 лютого відбудеться відкритий турнір Кубка Посла Японії з шьоґі 2026! Учасником турніру можуть стати усі охочі, які засвоїли правила гри.
У КПІ відбувся фінал V Всеукраїнського інженерного хакатону SmaRTF
Ювілейний хакатон уже традиційно зібрав молодих, амбітних і креативних винахідників, які цього року проєктували рішення у сфері smart military electronics для посилення стійкості держави.
From Power Grids to EV Motors: Industry Flags Key Budget 2026 Priorities for India’s Next Growth Phase
As India approaches Union Budget 2026–27, multiple industrial sectors—from power and automation to digital infrastructure and electric mobility—find themselves at a critical inflexion point. With the country balancing rapid industrialisation alongside sustainability and energy-transition goals, industry leaders are calling for continued capital expenditure, targeted incentives, and policy stability to strengthen infrastructure depth and global competitiveness.
At the core of these recommendations is the need to reinforce India’s power and grid ecosystem. According to Meenu Singhal, Regional Managing Director, Socomec Group, Greater India, sustained capex allocation, grid modernisation, and deeper indigenisation of critical power equipment will be essential to support rising industrial and digital demand. Industry stakeholders are urging the government to prioritise scalable manufacturing clusters, digitally enabled grid infrastructure, and structural reforms that improve reliability and execution efficiency.
Strategic schemes such as capex support mechanisms, fiscal incentives for local manufacturing, and policies favouring large-scale infrastructure implementation are seen as vital to closing capability gaps across transmission and distribution networks. Equally important, experts stress, is policy consistency and an enabling tax framework that continues to attract both domestic and global capital into the power sector, reinforcing India’s long-term vision of energy security and sustainable growth.
Automation as a Manufacturing MultiplierBeyond core infrastructure, industrial automation has emerged as a key lever for enhancing India’s manufacturing competitiveness as the economy advances towards the $5-trillion milestone. Sanjeev Srivastava, Business Head – Industrial Automation SBP at Delta Electronics India, highlights that smart factories, AI-driven automation, and closer human–machine collaboration will define the next phase of industrial transformation.
Industry players believe that stronger Budget support in the form of smart manufacturing incentives, R&D-linked tax benefits, and skill-development programmes can significantly accelerate the adoption of next-generation automation technologies. Such measures would help manufacturers improve productivity, reduce operating costs, and strengthen India’s position on the global manufacturing and automation curve.
Also read industry’s recommendations on the Union Budget 2026 at: PCB Duty Cuts to Manufacturing Zones: Top Industry Recommendations for Budget 2026
Digital Infrastructure and Data CentresAs India moves deeper into the 5G, cloud, and AI era, mission-critical digital infrastructure is increasingly being viewed as the backbone of every industry. Pankaj Singh, Head – Data Centre & Telecom Business Solutions at Delta Electronics India, notes that the upcoming Budget presents an opportunity to prioritise energy-efficient and resilient data-centre ecosystems.
Industry recommendations include stronger incentives for modular and containerised data-centre deployments to enable faster rollout of scalable core and edge facilities. There is also a growing emphasis on supporting advanced cooling technologies—such as liquid-to-liquid and liquid-to-air coolant distribution systems—to manage the high thermal loads associated with AI-driven workloads. When complemented with sustainability-linked benefits and Make-in-India incentives for locally manufactured power, cooling, and automation equipment, these measures could encourage OEMs to invest with greater confidence in building a future-ready, low-carbon digital backbone.
Strengthening the EV Manufacturing BaseMeanwhile, India’s electric mobility ecosystem is entering a decisive phase, where long-term resilience and supply-chain stability are becoming as critical as adoption numbers. Bhaktha Keshavachar, Co-Founder & CEO of Chara Technologies, points out that while policy efforts have successfully focused on vehicle adoption and battery localisation, recent global disruptions have exposed vulnerabilities stemming from India’s dependence on imported rare-earth magnet motors.
As Budget 2026 approaches, industry voices are calling for formal recognition and fiscal support for magnet-free motor technologies within existing incentive frameworks. These solutions offer predictable costs, reduced supply-chain risk, and the development of indigenous intellectual property—particularly for high-volume segments such as two-wheelers, three-wheelers, and commercial fleets.
Targeted incentives for rare-earth-free motor manufacturing, stakeholders argue, would not only de-risk India’s EV ambitions but also position the country as a global hub for affordable, resilient, and export-ready EV powertrain solutions.
The Road AheadTaken together, these pre-Budget recommendations underline a shared industry priority: building resilient, scalable, and future-ready industrial ecosystems through focused policy support. Whether in power infrastructure, automation, digital systems, or electric mobility, Budget 2026 is widely seen as a pivotal opportunity to reinforce India’s transition towards sustainable growth, technological leadership, and global manufacturing competitiveness.
The post From Power Grids to EV Motors: Industry Flags Key Budget 2026 Priorities for India’s Next Growth Phase appeared first on ELE Times.
So cool to actually be using all this gear for real work
| On the bench is a Behringer EP2500 pro audio amplifier. It's having a blown output stage and a shorted rectifier diagnosed and repaired. In play is a TTI signal generator and a Tek 468 scope, as well as a DIY dim bulb tester. I've been slowly acquiring all this gear over the past few years. Recently got hold of a proper electronics work bench with shelf a I've for the instruments. This has made life so much easier with all of the extra space it's freed up. It's great to be using all this stuff for real work, not just playing around! [link] [comments] |
Polar Light achieves nano-scale LED, paving way to next-gen micro-LED/nano-LED devices
Polar Light raises €5m+ funding to accelerate micro-LED commercialization
Some PCBs I've made for my 8 bit computer
| | Here are some of the PCBs I've made myself for an 8 bit computer project I'm working on. The boards, except the A register board, are double sided. Unfortunately no plated throughholes but there are functional vias with a piece of wire. Will definitely be posting more update about the entire project as I'm slowly finishing it. [link] [comments] |
If it works it's not stupid
| | submitted by /u/_binda77a [link] [comments] |
Coherent and Quside demo verifiable entropy for quantum-safe encryption
India’s Next Big Concern in the AI Era: Cybersecurity for Budget 2026
Artificial Intelligence (AI), like any other technology, comes with its own set of boons and banes. According to the Stanford AI Index 2024, India ranks first globally in AI skill penetration with a score of 2.8, ahead of the US (2.2) and Germany (1.9). AI talent concentration in India has grown by 263% since 2016, positioning the country as a major AI hub. India also leads in AI Skill Penetration for Women, with a score of 1.7, surpassing the US (1.2) and Israel (0.9).
India in its AI Era
According to a PIB report, India is one of the top five fastest-growing AI talent hubs, alongside Singapore, Finland, Ireland, and Canada. The demand for AI professionals in India is projected to reach 1 million by 2026. Taking this advancement into mind, all eyes will be on the Budget for 2026, judging what the government intends to propose to boost the AI landscape in India.
“India’s rapid advancement in the AI era places the spotlight on the upcoming Union Budget as a decisive moment for building a future-ready workforce. With over 40% of India’s IT and gig workforce already utilising AI tools, and India projected to account for the world’s AI talent by 2027, there is clear momentum; yet, significant gaps remain.
Although the employability of the workforce has improved, some part of the young workforce possesses deep AI skills, and many companies complain about the difficulties in hiring the right people. The provision of more funds for AI workforce training programs, along with the National Education Policy’s emphasis on the incorporation of applied AI, data science, and digital technologies into the curriculum, is important. We anticipate policies that foster close cooperation between the industry and the universities, provide incentives for certifying the basic knowledge gained through practical training, and allow more students to have access to hands-on labs and internships. Not only will these measures lift the entry-level skills of the labour force, but they will also make it certain that the young population of India is capable of turning to the global market as the main supplier of leaders in the coming years,” says Tarun Anand, Founder & Chancellor, Universal AI University.
Shortcomings of AI: Cybersecurity Threats
While AI has great potential and proposed advanced opportunities in various sectors, it comes with its own set of shortcomings. The issues of cybersecurity have escalated significantly in the AI era. Subsequently, it will be important to note what the new budget has in store to build on cybersecurity, as AI will continue to dominate the Indian landscape.
“As India approaches the 2026 Union Budget, the cybersecurity sector does so with clarity: compliance is no longer optional, and policy must now accelerate infrastructure transitions that enterprises cannot manage alone. In 2025, India faced nearly 265 million cyberattacks, with AI-driven ransomware democratizing threats at an unprecedented scale.
First, cybersecurity data centre infrastructure must be formally recognised as a critical national asset. Expanding the PLI framework to include cybersecurity data centres would strengthen India’s cyber sovereignty and reduce reliance on offshore infrastructure.
Second, the Budget should enable public–private partnerships to bolster SME cyber resilience. Manufacturing and mid-market enterprises are increasingly targeted by ransomware-as-a-service, yet lack access to enterprise-grade security. Government-backed subsidies routed through certified MSP networks would protect the Make in India ecosystem while democratizing DPDP compliance at scale.
Third, India must invest decisively in cybersecurity talent infrastructure. With a shortage of over 80,000 professionals, Budget 2026 should fund structured partnerships between government, academic institutions, and industry certifiers. This would create a domestic talent pipeline comparable to Singapore’s model. While we currently train over 2,000 professionals annually, government backing could scale this to more than 10,000 within three years.
The DPDP execution phase, starting in November 2026, will ultimately determine whether cyber resilience scales equitably across the country or remains concentrated in metro markets. Through targeted investments in infrastructure, partnerships, and education, the 2026 Budget has the opportunity to shape that outcome decisively,” says Rajesh Chhabra, General Manager, India and South East Asia, Acronis
By: Shreya Bansal, Sub-Editor
The post India’s Next Big Concern in the AI Era: Cybersecurity for Budget 2026 appeared first on ELE Times.
The AI-tuned DRAM solutions for edge AI workloads

As high-performance computing (HPC) workloads become increasingly complex, generative artificial intelligence (AI) is being progressively integrated into modern systems, thereby driving the demand for advanced memory solutions. To meet these evolving requirements, the industry is developing next-generation memory architectures that maximize bandwidth, minimize latency, and enhance power efficiency.
Technology advances in DRAM, LPDDR, and specialized memory solutions are redefining computing performance, with AI-optimized memory playing a pivotal role in driving efficiency and scalability. This article examines the latest breakthroughs in memory technology and the growing impact of AI applications on memory designs.
Advanced memory architectures
Memory technology is advancing to meet the stringent performance requirements of AI, AIoT, and 5G systems. The industry is witnessing a paradigm shift with the widespread adoption of DDR5 and HBM3E, offering higher bandwidth and improved energy efficiency.
DDR5, with a per-pin data rate of up to 6.4 Gbps, delivers 51.2 GB/s per module, nearly doubling DDR4’s performance while reducing the voltage from 1.2 V to 1.1 V for improved power efficiency. HBM3E extends bandwidth scaling, exceeding 1.2 TB/s per stack, making it a compelling solution for data-intensive AI training models. However, it’s impractical for mobile and edge deployments due to excessive power requirements.

Figure 1 The above diagram chronicles memory scaling from MCU-based embedded systems to AI accelerators serving high-end applications. Source: Winbond
With LPDDR6 projected to exceed 150 GB/s by 2026, low-power DRAM is evolving toward higher throughput and energy efficiency, addressing the challenges of AI smartphones and embedded AI accelerators. Winbond is actively developing small-capacity DDR5 and LPDDR4 solutions optimized for power-sensitive applications around its CUBE memory platform, which achieves over 1 TB/s bandwidth with a significant reduction in thermal dissipation.
With anticipated capacity scaling up to 8 GB per set or even higher, such as 4Hi WoW, based on one reticle size, which can achieve >70 GB density and bandwidth of 40TB/s, CUBE is positioned as a viable alternative to traditional DRAM architectures for AI-driven edge computing.
In addition, the CUBE sub-series, known as CUBE-Lite, offers bandwidth ranging from 8 to 16 GB/s (equivalent to LPDDR4x x16/x32), while operating at only 30% of the power consumption of LPDDR4x. Without requiring an LPDDR4 PHY, system-on-chips (SoCs) only need to integrate the CUBE-Lite controller to achieve bandwidth performance comparable to full-speed LPDDR4x. This not only eliminates the high cost of PHY licensing but also allows the use of mature process nodes such as 28 nm or even 40 nm, achieving performance levels of 12-nm node.
This architecture is particularly suitable for AI SoCs or AI MCUs that come integrated with NPUs, enabling battery-powered TinyML edge devices. Combined with Micro Linux operating systems and AI model execution, it can be applied to low-power AI image sensor processor (ISP) edge scenarios such as IP cameras, AI glasses, and wearable devices, effectively achieving both system power optimization and chip area reduction.
Furthermore, SoCs without LPDDR4 PHY and only CUBE-light controller can achieve smaller die sizes and improved system power efficiency.
The architecture is highly suitable for AI SoCs—MCUs, MPUs, and NPUs—and TinyML endpoint AI devices designed for battery operation. The operating system is Micro Linux combined with an AI model for AI SoCs. The end applications include AI ISP for IP cameras, AI glasses, and wearable devices.

Figure 2 The above diagram chronicles the evolution of memory bandwidth with DRAM power usage. Source: Winbond
Memory bottlenecks in generative AI deployment
The exponential growth of generative AI models has created unprecedented constraints on memory bandwidth and latency. AI workloads, particularly those relying on transformer-based architectures, require extensive computational throughput and high-speed data retrieval.
For instance, deploying LLamA2 7B in INT8 mode requires at least 7 GB of DRAM or 3.5 GB in INT4 mode, which highlights the limitations of conventional mobile memory capacities. Current AI smartphones utilizing LPDDR5 (68 GB/s bandwidth) face significant bottlenecks, necessitating a transition to LPDDR6. However, interim solutions are required to bridge the bandwidth gap until LPDDR6 commercialization.
At the system level, AI edge applications in robotics, autonomous vehicles, and smart sensors impose additional constraints on power efficiency and heat dissipation. While JEDEC standards continue to evolve toward DDR6 and HBM4 to improve bandwidth utilization, custom memory architectures provide scalable, high-performance alternatives that align with AI SoC requirements.
Thermal management and energy efficiency constraints
Deploying large-scale AI models on end devices introduces significant thermal management and energy efficiency challenges. AI-driven workloads inherently consume substantial power, generating excessive heat that can degrade system stability and performance.
- On-device memory expansion: Mobile devices must integrate higher-capacity memory solutions to minimize reliance on cloud-based AI processing and reduce latency. Traditional DRAM scaling is approaching physical limits, necessitating hybrid architectures integrating high-bandwidth and low-power memory.
- HBM3E vs CUBE for AI SoCs: While HBM3E achieves high throughput, its power requirements exceed 30 W per stack, making it unsuitable for mobile and edge applications. Here, memory solutions like CUBE can serve as an alternative last level cache (LLC), reducing on-chip SRAM dependency while maintaining high-speed data access. The shift toward sub-7-nm logic processes exacerbates SRAM scaling limitations, emphasizing the need for new cache solutions.
- Thermal optimization strategies: As AI processing generates heat loads exceeding 15 W per chip, effective power distribution and dissipation mechanisms are critical. Custom DRAM solutions that optimize refresh cycles and employ TSV-based packaging techniques contribute to power-efficient AI execution in compact form factors.
DDR5 and DDR6: Accelerating AI compute performance
The evolution of DDR5 and DDR6 represents a significant inflexion point in AI system architecture, delivering enhanced memory bandwidth, lower latency, and greater scalability.
DDR5, with 8-bank group architecture and on-die error correction code (ECC), provides superior data integrity and efficiency, making it well-suited for AI-enhanced PCs and high-performance laptops. With an effective peak transfer rate of 51.2 GB/s per module, DDR5 enables real-time AI inference, seamless multitasking, and high-speed data processing.
DDR6, still in development, is expected to introduce bandwidth exceeding 200 GB/s per module, a 20% reduction in power consumption along with optimized AI accelerator support, further pushing AI compute capabilities to new limits.

Figure 3 CUBE, an AI-optimized memory solution, leverages through-silicon via (TSV) interconnects to integrate high-bandwidth memory characteristics with a low-power profile. Source: Winbond
The convergence of AI-driven workloads, performance scaling constraints, and the need for power-efficient memory solutions is shaping the transformation of the memory market. Generative AI continues to accelerate the demand for low-latency, high-bandwidth memory architectures, leading to innovation across DRAM and custom memory solutions.
As AI models become increasingly complex, the need for optimized, power-efficient memory architectures will become increasingly critical. Here, technological innovation will ensure commercial realization of cutting edge of AI memory solutions, bridging the gap between high-performance computing and sustainable, scalable memory devices.
Jacky Tseng is deputy director of CMS CUBE product line at Winbond. Prior to joining Winbond in 2011, he served as a senior engineer at Hon-Hai.
Special Section: AI Design
- The AI design world in 2026: What you need to know
- AI workloads demand smarter SoC interconnect design
- AI’s insatiable appetite for memory
The post The AI-tuned DRAM solutions for edge AI workloads appeared first on EDN.
Cree LED and Blizzard Lighting settle patent infringement dispute
Про цікавий воркшоп у Політехнічному музеї
Спільний навчальний проєкт Державного політехнічного музею ім. Бориса Патона та доцентки кафедри української мови, літератури та культури КПІ ім. Ігоря Сікорського Антоніни Березовенко було присвячено розвиткові і застосуванню методів музейної педагогіки під час вивчення курсу "Культура усного професійного мовлення (риторика)" студентами-першокурсниками НН ІПСА, які навчаються за спеціальністю "Штучний інтелект".
Work in progress workbench
| submitted by /u/john_galt_42069 [link] [comments] |
Досягнуто прогресу у створенні гнучкого OLED-дисплея
Член наглядової ради КПІ ім. Ігоря Сікорського, професор Університету Дрекселя Юрій Гогоці, став ключовою фігурою у створенні нового покоління розтяжних OLED-дисплеїв. Він є одним із відкривачів матеріалів максенів (MXenes), які лягли в основу проривної технології, та очолив дослідницьку групу з Університету Дрекселя у США та Сеульського національного університету.
Made some simple kelvin clamps
| Used some nickel plated 3x10mm copper, cheap wire, and some banana connector from work [link] [comments] |
Don’t buy mini fridges I guess!
| submitted by /u/Fit-Spring1143 [link] [comments] |



