Українською
  In English
Збирач потоків
From 10 to 1000 TOPS: Why Automotive Chips Need a New Architecture
Speaking at the Auto EV Tech Vision Summit 2025, Namrta Sharma, Technical Director at Aritrak Technologies, highlighted how chiplet architectures are emerging as a crucial enabler for the next generation of automotive semiconductors. As vehicles transition toward software-defined platforms, Sharma emphasized that semiconductor design must evolve to meet unprecedented computational demands while balancing cost, scalability, and time-to-market.
Setting the context, Sharma described the ongoing transformation in the automotive sector. “We should all agree that we are actually going through a metamorphosis,” she said, adding that the automotive industry is currently experiencing “a big transformation of the century.” Vehicles are no longer defined solely by mechanical and electrical components. Instead, the modern car is increasingly becoming a software-driven platform.
Computational Requirements at Rise in Automotives
“The car is no longer just mechanical and electrical,” Sharma noted. “The car is a software-defined vehicle. So, the role of the semiconductors in the car is also changing. It is no longer just adding to some features. It actually defines the car. It is the core intelligence of the car.” This shift is dramatically expanding the semiconductor requirements inside vehicles. Modern automotive electronics must support electrification, battery management, in-vehicle infotainment, constant connectivity, and advanced driver assistance systems (ADAS), while also laying the groundwork for autonomous driving.
Sharma highlighted how the computational requirements behind these capabilities have grown exponentially. “If you see the numbers for the compute, Level-2 ADAS required just 10 TOPS, which is 10 trillion operations per second, which was just a decade ago,” she explained. “But now the requirement is about 1,000 TOPS for full autonomy. So it is like a 100x gap.”
Limitations at Play
Meeting such performance requirements using conventional monolithic chip design is becoming increasingly difficult. Integrating CPUs, GPUs, communication circuits, and power management components into a single system-on-chip (SoC) results in extremely large semiconductor dies. However, manufacturing such large chips is constrained by physical and economic limits. “There is a limit. It is called the reticle limit,” Sharma explained. “That is the biggest size that we can manufacture in a foundry, and it is about 850 mm² as of now.”
Large monolithic chips also face yield challenges. A defect in even a small portion of a large die can render the entire chip unusable, leading to significant losses in manufacturing yield and rising costs. To address these issues, the semiconductor industry has increasingly turned to chiplet architectures. “The solution was simple,” Sharma said. “Just cut this big die into multiple dies. These are all small functioning blocks, and these are called chiplets.”
Chiplets: A Yield Optimisation Solution
By breaking down a large chip into smaller modular dies, chiplet architectures help overcome reticle limitations and improve yield. If a defect occurs, only the affected chiplet is discarded rather than the entire system. This modular approach has already gained traction in high-performance computing systems and is now finding relevance in automotive electronics.
Sharma also pointed out that chiplets enable a shift toward heterogeneous integration. Instead of manufacturing every component using the most advanced—and expensive—process node, different chiplets can be fabricated using technology nodes optimized for their specific functions. “For example, the CPU needs the highest compute, so it is in the latest technology node,” she explained. “But the other things, like sensors or memory engines, need not be.”
Chiplets Optimised Time to Market
Beyond cost and yield advantages, chiplets significantly improve time-to-market. Sharma emphasized that modular architectures allow semiconductor companies to reuse proven components and focus their efforts on product differentiation. “You need not design the whole chip, you need not design the whole SoC,” she said. “You can just design your differentiating chiplet.” This flexibility also allows manufacturers to scale systems quickly for different vehicle segments.
Performance levels can be adjusted simply by replacing or modifying individual chiplets, enabling rapid customization without redesigning the entire architecture. However, Sharma cautioned that the transition to chiplet-based systems introduces new challenges. With multiple chiplets integrated into a single package, design complexity shifts from the chip level to the system level. This requires advanced electronic design automation (EDA) tools capable of co-optimizing silicon, packaging, and interconnect technologies.
Testing & Standards
With this, testing and validation also become more complex. The success of chiplet integration depends on ensuring that each component integrated into the system is a “known good die.” As Sharma noted, this requires new testing methodologies and infrastructure capable of validating chiplets both individually and within the larger system.
Another key factor in the long-term success of chiplets is the development of industry-wide standards. Sharma highlighted the importance of emerging interconnect standards such as Universal Chiplet Interconnect Express (UCIe), which aim to enable interoperability between chiplets from different vendors. As the ecosystem evolves, Sharma believes collaboration across the semiconductor value chain will play a critical role. Foundries, design houses, EDA companies, substrate providers, and industry consortia are already working together to establish the standards and infrastructure needed to support chiplet-based systems.
Conclusion
Summarizing her key message, Sharma emphasized that chiplet architectures are not just about cost optimization. Instead, they represent a fundamental shift in how semiconductor systems are designed for rapidly evolving markets like automotive. “Chiplet heterogeneous integration provides not only the cost benefit,” she concluded, “it also provides the speed of execution—the speed to make changes fast and react to innovation.”
As automotive electronics continue to grow in complexity and performance requirements, chiplets may well become the architectural foundation enabling the next wave of innovation in software-defined vehicles.
The post From 10 to 1000 TOPS: Why Automotive Chips Need a New Architecture appeared first on ELE Times.
Evolution of Technology Calls for Continuous Upskilling of Industry Professionals: Arindam Lahiri, ASDC
As India’s mobility ecosystem undergoes rapid technological transformation, the importance of a skilled workforce has become more critical than ever. Electrification, connected vehicles, advanced electronics and digital technologies are reshaping the automotive landscape, creating both opportunities and challenges for the industry. In an exclusive conversation with ELETimes, Arindam Lahiri, CEO, ASDC, shares his insights on the opportunities and gaps that India possesses as it unravels the electronics revolution.
At the centre of this transformation is the Automotive Skills Development Council (ASDC), which plays a key role in developing industry-relevant skills and building a future-ready workforce. Established to align vocational training with the evolving needs of the automotive sector, ASDC works closely with industry bodies, government institutions and training partners to create a robust skill ecosystem.
The organisation is playing a key role in the evolution of India’s skill development ecosystem, emerging technological trends in the automotive sector and the initiatives being taken to prepare the next generation of professionals.
Here are the excerpts from the interview:
ELETimes: Could you briefly explain ASDC and its key objectives?
Arindam Lahiri: Automotive Skills Development Council is one of the 36 sector skill councils operating in India. Sector skill councils were created to bridge the gap between industry requirements and workforce capabilities by developing structured training frameworks aligned with specific industries. The council is promoted by three major industry associations: the Society of Indian Automobile Manufacturers (SIAM), the Automotive Component Manufacturers Association of India (ACMA), and the Federation of Automobile Dealers Associations (FADA). These organisations represent vehicle manufacturers, component manufacturers and dealerships respectively, ensuring that the entire automotive value chain is represented within ASDC’s governance framework. The council also works in close coordination with the Ministry of Skill Development and Entrepreneurship (MSDE). In addition, the Ministry of Heavy Industries (MHI) and the Ministry of Road Transport and Highways (MoRTH) serve as line ministries for the automotive sector.
More recently, ASDC has also become an affiliated awarding body under the National Council for Vocational Education and Training (NCVET), which is the apex regulator for skill education in India.
Our core objective is to create a national platform for skill development in the automotive sector. This involves developing industry-aligned training curricula, establishing rigorous assessment and certification processes and ensuring that training programmes meet global standards. Ultimately, skill development contributes to higher productivity within the industry, which leads to capital creation and economic growth. As businesses expand and reinvest their resources, this generates further employment opportunities. Another key focus area for us is integrating skill-based education within mainstream academic pathways, a vision that has been strongly reinforced by the National Education Policy (NEP 2020).
ELETimes: Over the last decade, how do you assess the progress made in India’s skill development ecosystem?
Arindam Lahiri: India has a long history of vocational training through institutions such as Industrial Training Institutes (ITIs), which were established soon after independence. These institutions have played an important role in producing skilled manpower for the engineering and manufacturing sectors.
However, the last decade has been particularly significant for the skill development landscape in India. With the launch of the national Skill India mission, the government brought skill development into mission mode. Under the leadership of the Prime Minister, multiple ministries, agencies and training institutions have come together under a common framework to strengthen vocational education.
This coordinated effort has created a strong momentum in the skill development ecosystem. At the same time, it is important to recognise that India is still in the early stages of building a fully mature vocational training system.
Within the automotive sector specifically, we see a large number of training initiatives being undertaken by vehicle manufacturers, component manufacturers and dealership networks. While these initiatives are valuable, there is significant scope to scale them further and create more structured pathways for skill development.
India’s automotive industry is globally competitive, with many international brands operating in the country. These companies maintain very high standards when it comes to workforce capabilities. One of the concerns often raised by industry is that graduates from engineering colleges, polytechnics and ITIs sometimes lack the practical skills required for immediate employment.
This is precisely where organisations like ASDC play a critical role. By working closely with industry, we identify skill gaps and develop training programmes that address those gaps. Our work includes curriculum development, training of trainers, development of learning content, candidate assessment and certification.
ELETimes: What initiatives is ASDC taking to ensure that its curriculum remains contemporary and globally relevant?
Arindam Lahiri: Our curriculum framework is built upon National Occupational Standards (NOS), which define the competencies required for specific job roles in the automotive sector. These standards form the foundation for developing qualification packs and training curricula.
The qualification framework and associated curricula are reviewed and approved by the National Skill Qualifications Committee under NCVET. This ensures that training programmes meet national standards and remain aligned with industry expectations.
At regular intervals, we undertake comprehensive reviews of our curriculum with the support of industry experts. These experts come from various segments of the automotive ecosystem, including manufacturing, servicing, supply chain and emerging technology domains. Their inputs help ensure that every qualification and training programme reflects real-world industry requirements.
In recent years, we have incorporated several emerging areas into our curriculum. These include electric mobility, Industry 4.0 technologies, sustainability practices, vehicle diagnostics and safety systems.
In addition to entry-level training programmes, we are increasingly focusing on upskilling and reskilling the existing workforce. As technology evolves, professionals who are already working in the industry must continuously update their skills to remain relevant.
Given the pace at which automotive technologies are advancing, curriculum development is not a one-time exercise. It is a continuous process that evolves alongside the industry.
ELETimes: Safety remains a major concern within the automotive ecosystem. How can the industry address safety challenges more effectively?
Arindam Lahiri: Safety is a critical aspect of the automotive sector, and it can broadly be viewed in two dimensions: industrial safety and road safety. Industrial safety relates to the safety of workers within manufacturing plants, workshops and service facilities. Employees involved in vehicle manufacturing, component production or vehicle servicing must follow strict safety protocols.
This aspect of safety has become even more important with the increasing electrification of vehicles. Electric vehicles operate on high-voltage systems, which can pose serious risks if proper safety procedures are not followed. Therefore, all our training programmes place significant emphasis on safety standards and best practices related to specific job roles.
The second dimension is road safety. Road safety involves both preventive measures and post-accident response mechanisms. Preventive measures include awareness programmes that educate people about responsible driving behaviour, traffic regulations and safe mobility practices.
ASDC has been actively working on awareness initiatives at colleges and educational institutions, where young drivers can be sensitised to road safety principles. In addition, we are exploring programmes related to first responder training so that individuals can respond effectively in the event of a road accident and help reduce fatalities.
ELETimes: What major technological trends do you expect to shape the automotive sector by 2026 and beyond?
Arindam Lahiri: Today’s vehicles are increasingly becoming sophisticated digital platforms. Modern automobiles contain a significant amount of electronic hardware and software systems, whether they are two-wheelers, passenger vehicles or commercial vehicles.
In recent years, the automotive industry has experienced supply challenges due to semiconductor shortages. While the situation has improved, the increasing reliance on electronic components highlights how deeply technology is embedded in modern vehicles.
Another major trend is the rapid advancement of software and computing capabilities within vehicles. Improvements in computing power, communication speeds and the adoption of 5G connectivity are enabling new forms of vehicle intelligence.
Connected vehicles are already becoming a reality. Many manufacturers can now monitor the performance and location of vehicles through remote systems. These systems can even predict maintenance requirements and detect potential issues before they escalate.
Technology is also transforming logistics and transportation. Road transport remains the backbone of India’s logistics sector, and digital tools are helping drivers, fleet owners, and logistics managers optimise operations and respond quickly to unexpected situations.
Electric mobility will continue to expand as battery technologies improve and charging infrastructure becomes more widespread. At the same time, research is progressing in areas such as hydrogen fuel cells and solar-powered mobility solutions.
For young professionals, this presents tremendous opportunities. Individuals who continuously upgrade their skills will find exciting career prospects in the automotive sector, both in India and globally.
ELETimes: If a student wants to pursue training in the automotive sector through ASDC, what is the process?
Arindam Lahiri: ASDC functions as an awarding body and does not directly operate training centres. Instead, we affiliate training institutions that deliver automotive skill programmes based on our curriculum framework.
Currently, more than 300 active training institutions across India are affiliated with ASDC. These institutions offer training in various automotive domains, ranging from vehicle servicing and diagnostics to emerging technologies.
Students who are interested in exploring career opportunities in the automotive sector can visit our dedicated career guidance platform at careerguide.asdc.org.in. This platform provides detailed information about different job roles, career pathways and training opportunities available within the automotive industry.
The platform also helps students understand how specific training programmes can lead to long-term career progression. It provides insights into training centres located across different regions of the country.
Students can also submit queries through the contact section of the ASDC website. Our team then guides them in selecting the appropriate training programme and identifying the nearest training centre offering that programme.
The post Evolution of Technology Calls for Continuous Upskilling of Industry Professionals: Arindam Lahiri, ASDC appeared first on ELE Times.
Complex System Design Is Key to Navigating the Evolving Tech Ecosystem: Sadaf Arif Siddiqui, Keysight
In a conversation with Kumar Harshit, Technology Correspondent, ELE Times, Sadaf Arif Siddiqui, Director Marketing, Keysight Technologies India, shares his perspective on the latest developments in India’s electronics and telecom sectors. He discusses how advanced design validation and other technologies are gaining traction in product development. Further, the conversation also touches upon India’s position in the upcoming 6G technology and how unprecedented challenges are awaiting us.
The conversation moves further to touch upon certain technical aspects of testing in aerospace and defense, along with the simultaneous growth being exhibited by various industries in India.
Here are the excerpts from the interview:
ELE Times: India is increasingly positioning itself as a design-led electronics ecosystem rather than a manufacturing-only hub. From your perspective, how is this shift changing the role of advanced design validation and emulation in development cycles?
S.A Siddiqui: This shift is reshaping how companies approach product development from the ground up. Today, the focus is firmly on establishing a fully integrated, end-to-end workflow that seamlessly spans the entire product lifecycle — from initial design and prototyping to rigorous validation and ultimately, full-scale manufacturing. In addition, strong policy support led by the government and a deep pool of engineering talent are collectively enabling the country to move beyond pure manufacturing and into high-value innovation and product ownership. A key example of this momentum is the government’s push toward self-reliance under initiatives such as Make in India. This effort is encouraging growth across sectors, including telecom, automotive, aerospace and defense, semiconductors, and electronics.
With these shifts, the role of design validation and emulation has become extremely critical. The technologies driving this innovation are only getting more complicated, and at the same time, there is growing pressure to launch products faster. In this environment, adopting advanced design validation as early as possible in the development cycle will give companies a competitive edge.
ELE Times: India has begun early conversations around 6G and non-terrestrial networks. What new testing challenges do these next-generation technologies introduce compared to 5G?
S.A Siddiqui India is positioning itself as an active contributor to 6G research and standardization, with several academic institutions, government research bodies, and the Ministry of Communications already engaged in exploratory programs to define India’s role in shaping 6G and non-terrestrial network (NTN) technologies.
5G primarily focused on enhanced mobile broadband, ultra-low latency, and high throughput. Whereas 6G and NTNs significantly expand the scope and are leveraging AI and communication to extend connectivity beyond terrestrial boundaries and into space through satellite networks. In terms of spectrum, 6G is projected to move beyond millimeter wave and into sub-terahertz (sub-THz) frequency bands.
This entirely changes the concept of testing in 6G. NTNs present challenges, the first being that satellite-based equipment, once deployed, is largely inaccessible. The technology also needs to operate in extremely harsh environments with temperature fluctuations and radiation. There are also challenges around frequency conditions, latency and doppler shifts. At the same time, energy efficiency is also a central pillar of 6G system design. Together, these requirements make 6G testing far more multidimensional than previous generations, demanding greater precision and accuracy under harsher conditions. As a result, over-the-air testing is growing in importance so developers can validate systems with the accuracy they need.
ELE Times: With vehicles becoming increasingly software-defined, how important is cross-domain validation — combining RF, digital, power electronics, and cybersecurity testing?
S.A Siddiqui: Cross-domain validation has become absolutely critical as the automotive industry evolves. Modern vehicles integrate a dense network of electronic control units (ECUs), sensors, high-speed digital interfaces, connectivity modules, and power electronics. This convergence of multiple domains means that validation can no longer occur in isolated silos. Instead, it must address the full system-level interaction between RF, digital, power, and software layers—particularly given the stringent safety, reliability, and regulatory requirements governing the automotive sector.
ELE Times: In aerospace and defence applications, testing is often mission-critical. How is the need for high-fidelity emulation and secure communications testing evolving in India?
S.A Siddiqui: Aerospace and Defense is one of the cornerstones of India’s ‘Make in India’ initiatives. This means all components should be developed in the country and fully integrated into the platform. As a result, validation and emulation are critical as every element needs to be deployed with extreme precision and accuracy. As India continues to push technology boundaries, testing needs are evolving, and this is a clear example. Every element, right from designing the first component through to the prototype and final product, must be tested in order to ensure precision. This also means testing under a wide set of conditions and environmental requirements.
ELE Times: India is one of the few markets where telecom, automotive, semiconductors, defence, and AI are all scaling simultaneously. How does this convergence influence the demand for advanced testing ecosystems?
S.A Siddiqui: India’s simultaneous scaling across multiple high-tech sectors creates a uniquely demanding environment for testing and validation. With the rapid development of technologies such as 5G/6G, artificial intelligence, semiconductors, quantum initiatives, and defence systems, the country is not only meeting domestic requirements but also positioning itself to compete on a global stage. This simultaneous growth across diverse sectors intensifies the need for integrated, high-performance testing ecosystems that can handle increasingly complex and interdependent technologies.
The convergence of RF, digital, software, and security components across these domains means that testing can no longer be confined to individual components or isolated subsystems. Instead, advanced testing must address full system-level interactions, integrating hardware, software, and cybersecurity layers to ensure reliable performance under real-world conditions. Dynamic, repetitive testing platforms are now essential, enabling engineers to validate interoperability, functionality, and robustness before deployment. Technologies such as digital twins are playing a transformative role in this context. By creating virtual replicas of systems, engineers can simulate and validate complex behaviors in the lab, reducing risk and accelerating development cycles before field testing.
ELE Times: Looking ahead 3–5 years, what will differentiate engineering organizations that succeed in complex system design from those that struggle?
S.A Siddiqui Over the next couple of years, the organizations that will excel in complex system design are those that can navigate a dynamic, rapidly evolving ecosystem while balancing immediate needs with future-facing innovation. Success will hinge on the ability to test as early as possible and throughout the entire development lifecycle – allowing organizations to improve speed to market, while still delivering on quality and meeting modern engineering challenges.
| Brief profile of the speaker:
Sadaf Arif Siddiqui is the Director of Marketing – India at Keysight Technologies, where he leads strategic marketing initiatives to support the company’s growth across telecommunications, automotive, aerospace & defense, and emerging technology segments. With over 23 years of experience in test, measurement, embedded systems, and technology solutions, Sadaf has driven customer engagement, industry partnerships, and market expansion throughout his career. He joined Keysight (formerly part of Agilent Technologies) in 2008 and has held multiple leadership roles spanning industry marketing, business development, and global program management. Sadaf holds a Bachelor’s degree in Electronics Engineering from Aligarh Muslim University and a Post Graduate Diploma in Business Management (Marketing) from Symbiosis Institute, Pune. |
The post Complex System Design Is Key to Navigating the Evolving Tech Ecosystem: Sadaf Arif Siddiqui, Keysight appeared first on ELE Times.
Пам'яті Марії Бігун
🕯 На війні загинула випускниця Київської політехніки Марія Бігун (Леут) (28.08.2000 — 21.02.2026)
Elite level soldering
| submitted by /u/YahudiKundakcisi [link] [comments] |
Building I2C-PPS. Part 5 - BOM
| This is the next update on the programmable power supply project (you can find previous posts and more details in its umbrella repository condevtion/i2c-pps, while schematics itself is in condevtion/i2c-pps-hw). During the past week I managed to select exact market available components for the device and create detailed BOM. I need parts for 3 copies of the power supply - two sets to actually build devices and one on standby just in case. Honestly, I expected the BOM to be 3 times cheaper (or at least 2) but costs for hundred components quickly add up. In the first picture above (left chart) you can see average unit price of a part per its type with quite expected the BQ25758S controller being the most expensive thing. However, as the right chart shows total amount of capacitors easily gives them the lead in final cost, which is $108.88 (or $36.29 per set). For just one set the total is $48.99 making it almost buy two get one for free. The next picture shows quantities of parts per device and totals for 3 devices with additional components (marked green) per part type (total here is 393 for all 3 sets). The latest allows to reduce cost even further due to substantially lower prices on bigger quantities. Now, knowing exact parts and their footprints I can start designing PCB itself. [link] [comments] |
Weekly discussion, complaint, and rant thread
Open to anything, including discussions, complaints, and rants.
Sub rules do not apply, so don't bother reporting incivility, off-topic, or spam.
Reddit-wide rules do apply.
To see the newest posts, sort the comments by "new" (instead of "best" or "top").
[link] [comments]
На війні загинула студентка Дарья Лопатіна
🕯На війні загинула студентка Київської політехніки Дарья Лопатіна (08.07.2006 — 02.09.2025).
My first two PCBs created while I try to teach myself electronics!
| The first started as a way to test ADCs and parallel I/O, and I turned it into a toy oscilloscope using some software I wrote for my Raspberry Pi. I didn't really understand op-amp input bias current and so it doesn't really work properly with the probe in 10x mode. The offset is huge, but I now understand the mistake. I also used one more op-amp than I really needed, and could've gotten away with cheaper ones, but it works up to 50MS/s! The second board is a buffered variable-gain amplifier test with voltage-variable gain and bias. I fell down a rabbit hole w/oscilloscopes and am working on making an improved 2-channel one with modern components, so I broke out some of the front end into a test board and just finished building it. It's a miracle the QFN op-amp works, I was sure I'd bridge something underneath it. There's a subtle crucial mistake in the second design, all you need to know to spot it is that the second amp is an LMH6505. It somehow does partially function still! [link] [comments] |
Took apart this broken digital scale just for fun
| submitted by /u/netsurfer79 [link] [comments] |
EEVblog 1738 - Solar Freakin' Roadways NEW FUNDING!
Risk assessment in the workplace

Risks come in more than one form. There are risks that arise from science and technology, and there are risks that arise from human motivations, which are not always of an obvious sort. This is about the latter.
I had a client company that was owned by a husband and wife for whom I had once solved a power supply thermal runaway problem. I had measured temperature rise versus time and temperature fall versus time, and of course, the two were not exactly the same. Their difference was quite pronounced when I first looked at the issue, but they were almost identical to each other after I had solved their problem. If you’re curious about that, please see the How2Power article here.
A couple of years went by, and I got a call from that same company about a different power supply that also seemed to have a thermal runaway problem. By then, sadly, the husband had passed away, and only the wife remained to run the business.
During the first time frame, the wife had displayed a hair-trigger temper. Any moment of uncertainty as events unfolded would result in a raging torrent from her, to which her husband would make great efforts to calm her down. I would hear lines like “It’s okay. It’s oh-kay! Please relax. Things are going well.” to which she would then go silent, but now she didn’t have anyone to give her any assurance when it was needed.
An employee who had been promoted to Chief Engineer was my new point of contact. I explained to him that I would examine the thermal rise and thermal fall traits of this new power supply to see if indeed the same situation pertained as it did in the first case or not.
“There’s no need for that. I’ve already made those measurements.” He handed me a sheet of paper with columns of numbers, purportedly the data I had planned to acquire. That night, I examined those numbers and discovered that if you plotted the thermal rise and inverted a plot of the thermal fall, the two curves precisely lined up and were EXACTLY the same!! There was absolutely zero difference. They were totally spot on, no ifs, ands, buts, hows, whys, or wherefores, exactly the same, which meant that the rising and falling curves given to me were not the results of actual testing. They were false.
Confronted with a Chief Engineer whom I then knew to be dishonest and confronted with the woman whom I knew to be extremely volatile and prone to bursts of rage, I assessed the risk of dealing with it all to be unacceptable.
I made up some excuse (I don’t remember what it was) and declined to offer my services.
John Dunn is an electronics consultant and a graduate of The Polytechnic Institute of Brooklyn (BSEE) and of New York University (MSEE).
Related Content
The post Risk assessment in the workplace appeared first on EDN.
New LX4580 – Highly Integrated 24‑Channel Mixed‑Signal IC for Aviation & Defence Actuation Systems
The post New LX4580 – Highly Integrated 24‑Channel Mixed‑Signal IC for Aviation & Defence Actuation Systems appeared first on ELE Times.
Last-level cache has become a critical SoC design element

As AI workloads extend across nearly every technology sector, systems must move more data, use memory more efficiently, and respond more predictably than traditional design methodologies allow. These pressures are exposing limitations in conventional system-on-chip (SoC) architectures as compute becomes increasingly heterogeneous and traffic patterns become more complex.
Modern SoCs integrate CPUs, GPUs, NPUs, and specialized accelerators that must operate concurrently, placing unprecedented strain on memory hierarchies and interconnects. To Keep processing units fully utilized requires high-bandwidth, low-latency access to data, making the memory hierarchy as critical to overall system effectiveness as raw performance.
On-chip interconnects move data quickly and predictably, but once requests reach external memory, latency increases, and timing becomes less consistent. As more data accesses go off chip, the gap between compute throughput and data availability widens. In these conditions, processing engines stall while waiting for memory transactions to complete, creating data starvation.
The role of last-level cache
To mitigate this imbalance, SoC designers are increasingly turning to last-level cache (LLC). Positioned between external memory and internal subsystems, LLC stores frequently accessed data close to compute resources, allowing requests to be served with significantly lower latency.
Unlike static buffers, an LLC dynamically fetches and evicts cache lines based on runtime behavior without direct CPU intervention. When deployed effectively, this architectural layer delivers measurable benefits, including substantial reductions in external memory traffic and power consumption.
Simply including an LLC does not guarantee improved performance. Configuring the cache correctly is a complex task that must account for workload characteristics, compute-unit behavior, and real-time constraints. Poorly chosen parameters can waste area without meaningful gains, while under-provisioned configurations may fail to alleviate memory bottlenecks.
Architects must carefully determine cache capacity, the number of cache instances, and internal banking structures to support sufficient parallelism. Partitioning strategies must also be defined to ensure that individual IP blocks receive the bandwidth and predictability they require. While some settings can be adjusted later through software, foundational decisions on cache size, banking, and associativity must be finalized early in the development cycle.

The role of last-level cache is shown in successful designs. Source: Arteris
Factors influencing cache behavior
Banking configuration illustrates this trade-off clearly. Increasing the number of cache banks improves internal parallelism and throughput, but it also increases silicon area. Workloads with largely sequential access patterns may see limited benefit from aggressive banking.
In contrast, highly parallel workloads, especially those driven by AI accelerators or GPUs, require substantial internal concurrency to maintain utilization. Because these characteristics vary by application, banking decisions must be informed by realistic workload analysis during the architectural phase.
Cache capacity is just as important. A cache that is too small struggles to achieve acceptable hit rates, pushing excessive traffic to external memory. Conversely, oversizing the cache often yields diminishing returns relative to the additional area consumed. The optimal balance depends on actual runtime behavior rather than theoretical assumptions.
In practice, acceptable hit rates vary widely. Some systems can tolerate moderate miss rates if latency and power reductions outweigh the cost, while real-time applications demand consistently high hit rates to maintain deterministic behavior.
This variability underscores why no single LLC configuration is universally optimal. Mobile devices may require only a few megabytes of cache to balance power efficiency and responsiveness. At the same time, servers and HPC platforms often deploy tens or hundreds of megabytes to reduce DRAM pressure. Despite these differences, successful designs rely on a common principle in which cache parameters are derived from the workloads the system will actually execute.
Managing shared caches
Diversity in system demands further complicates how an LLC must be structured. Automotive chips built around concurrent vision processing and strict timing requirements operate under very different constraints than data-center platforms optimized for accelerator-heavy inference at scale. Even within a single chip, CPUs, accelerators, and I/O subsystems generate distinct access patterns with different latency sensitivities.
The LLC must accommodate all of them without allowing one workload to interfere with another’s real-time guarantees. This makes early understanding of system-level access behavior essential, since cache configuration otherwise becomes speculative at best.
Partitioning provides a powerful mechanism for preserving determinism in such environments. By allocating portions of cache capacity to specific clients, architects can prevent high-bandwidth workloads from starving latency-sensitive subsystems. This capability is particularly critical in environments that must meet strict timing guarantees. Partition sizes must be tuned carefully, as oversizing wastes area while undersizing risks violating latency requirements.
Configuring a last-level cache is ultimately a multidimensional challenge shaped by workload demands, compute topology, latency requirements, and silicon constraints. Achieving the right balance between performance, determinism, power, and area depends on understanding how an SoC behaves under real operating conditions.
To address this, SoC teams increasingly rely on system-level simulation using realistic data flow profiles generated by multiple on-chip request sources. This approach allows teams to evaluate cache behavior before key architectural decisions are finalized. It helps identify bottlenecks, validate cache sizing, and determine when isolation mechanisms such as partitioning are required to preserve real-time guarantees.
Arteris developed its CodaCache IP, which operates as a configurable last-level cache between on-chip initiators and different types of external memories such as DDR-DRAM, HBM and even NVM for execution in place (EIP) use cases. With CodaCache, architects can equip their SoC fabric with the optimal configuration to address intelligent, scalable, and automated data management in a wide range of applications.
Andre Bonnardot is product marketing manager at Arteris.
Related Content
- Understanding cache placement
- Optimizing for instruction caches
- How to Turbo Charge Your SoC’s CPU(s)
- Bringing SOT-MRAM Tech Closer to Cache Memory
- SoC design: When a network-on-chip meets cache coherency
The post Last-level cache has become a critical SoC design element appeared first on EDN.
Marktech adds 230nm and 265nm deep UV, 310nm UVB and 340nm UVA LEDs
TI redoubles advancement of next-gen physical AI with NVIDIA
Texas Instruments announced accelerating the safe deployment of humanoid robots into the real world with NVIDIA. By combining TI’s real-time motor control, sensing, radar and power technologies with NVIDIA’s advanced robotics compute, Ethernet-based sensing and simulation technologies, robotics developers can validate perception, actuation and safety earlier and more accurately. TI connects NVIDIA physical AI compute to real-world applications with deterministic control, sensing, power, and safety at every joint and subsystem. This partnership will help developers move faster from virtual development to production-ready, scalable and safety-compliant systems.
As part of this collaboration, TI designed a sensor fusion solution by integrating its mmWave radar technology with NVIDIA Jetson Thor using NVIDIA Holoscan Sensor Bridge to enable low-latency, 3D perception and safety awareness for humanoid robots. TI will showcase the solution at NVIDIA GTC, March 16–19, 2026, in San Jose, California.
“The next generation of physical AI requires more than just advanced compute – it demands seamless integration between sensing, control, power and safety systems,” said Giovanni Campanella, general manager of industrial automation and robotics at TI. “TI’s comprehensive portfolio bridges the gap between NVIDIA’s powerful AI compute and real-world applications, enabling developers to validate complete humanoid systems earlier in development. This integrated approach will help accelerate the evolution from prototypes to commercially viable humanoid robots operating safely alongside humans.”
“The safe operation of humanoid robots in unpredictable environments requires a massive leap in processing power to synchronise complex AI models with real-time sensor data and motor controls,” said Deepu Talla, vice president of robotics and edge AI at NVIDIA. “The integration of Texas Instruments’ sensing and power management technologies with the NVIDIA Jetson Thor platform provides developers with a functional safety-capable foundation to accelerate the deployment of next-generation physical AI.”
Enabling safer humanoid robots with real-time sensor fusion technology
TI’s mmWave radar sensor, IWR6243, connected via Ethernet to NVIDIA Jetson Thor, enables scalable low-latency, 3D perception and safety awareness for physical AI applications. By fusing camera and radar data, the solution improves object detection, localisation, and tracking while reducing false positives for confident, real-time decision-making in humanoid robots.
This solution enables human-like perception that works reliably in challenging conditions – from low light and bright glare to fog and dust indoors and outdoors – and addresses a critical safety gap that has limited real-world deployment of humanoid robots. For example, while cameras may not reliably detect glass doors or reflective surfaces, radar provides consistent detection of these transparent obstacles, enabling smooth navigation in places like office buildings, hospitals and retail environments.
TI at NVIDIA GTC
TI will present its technologies at NVIDIA GTC in booth 169 at the San Jose McEnery Convention Centre. TI and D3 Embedded’s live demonstration, “Real-time sensor fusion for reliable robotic perception with Holoscan,” showcases how TI’s mmWave radar technology integrates with NVIDIA’s Jetson Thor and Holoscan ecosystem using an end-to-end software processing chain and visualisation from D3 Embedded.
On Wednesday, March 18, from 3:00-3:40 p.m. PT, TI’s Giovanni Campanella will participate in a lightning talk, “The Edge of the Edge: Redefining GPU-Enabled AI Sensor Processing.” Campanella will discuss how the tight integration of sensing, networking and GPUs is enabling real-time physical AI at the edge of industrial systems.
The post TI redoubles advancement of next-gen physical AI with NVIDIA appeared first on ELE Times.
Everspin Advances High-Reliability xSPI MRAM Portfolio With Complete Production Qualification for 64Mb MRAM
Everspin Technologies, the world’s leading developer and manufacturer of magnetoresistive random access memory (MRAM) persistent memory solutions, announced continued progress across its high-reliability (HR) PERSYST xSPI STT-MRAM portfolio, including the completion of full production qualification for its 64Mb MRAM and the expansion of the family to a new 256Mb density.
The HR 64Mb xSPI STT-MRAM has now completed full production qualification for the AEC-Q100 Grade 1 specification. It is currently available for customer orders and supports high-volume production programs, with inventory available through Everspin’s authorised distributors worldwide.
The 128Mb xSPI STT-MRAM is expected to complete production qualification in May 2026, and a new 256Mb option is scheduled to complete full production qualification in July 2026, with volume availability expected in the second half of 2026.
“Advancing our high-reliability product family through production qualification and expanding density options reflects steady progress against our technology roadmap,” said Sanjeev Aggarwal, president and CEO of Everspin Technologies. “Customers designing long-lifecycle systems require validated memory solutions with predictable performance, and we are extending the PERSYST platform to meet those needs across a wider range of densities.”
The addition of the 256Mb density enables higher-capacity persistent memory designs within the same xSPI-based architecture. Together with the 64Mb and 128Mb xSPI STT-MRAM products, the expanded Hi-Rel portfolio provides scalable options for applications operating across extended temperature ranges and demanding reliability environments.
“Production qualification provides the level of confidence required for space and satellite programs moving into long-term deployment,” said Billy Wahng, Chief Technology Officer at Astro Digital. “Everspin’s focus on endurance, data integrity and radiation tolerance addresses the challenges of operating in unpredictable environments.”
These milestones represent continued execution of Everspin’s roadmap to broaden its HR MRAM portfolio for aerospace, defence, automotive, industrial and other mission-critical applications.
The post Everspin Advances High-Reliability xSPI MRAM Portfolio With Complete Production Qualification for 64Mb MRAM appeared first on ELE Times.
Rohde & Schwarz enables rapid validation of next-gen Wi-Fi 8 networking platforms, including 5×5 MIMO capabilities
Qualcomm Technologies has used the CMP180 radio communication tester from Rohde & Schwarz to validate advanced multi-antenna capabilities that are designed into its next-generation Wi-Fi 8 platforms, including support for 5×5 MIMO in the 2.4, 5, and 6 GHz bands. Advanced 5×5 MIMO architectures help Wi‑Fi 8 platforms deliver higher capacity and more reliable connectivity across a wider range of real‑world deployment scenarios.
The industry‑leading CMP180 delivers full bandwidth and seamless scalability for testing leading Wi‑Fi 8 chipsets across the entire device lifecycle — from development to production. As a result of this collaboration, Rohde & Schwarz now offers pre‑built test routines and early access to key resources, enabling device manufacturers to accelerate the time‑to‑market of their products.
Wi-Fi 8, based on the IEEE 802.11bn specification, builds on the foundation of Wi-Fi 7 to deliver next-level reliability, efficiency, and seamless mobility. New PHY and MAC layer technologies work together to extend range, improve spectrum utilization, reduce latency, and enable coordinated access across dense environments, setting the stage for ultra-high reliability (UHR) performance. Advanced antenna architectures such as 5×5 MIMO help enhance spatial efficiency and link robustness and provide a more consistent performance in real-world environments.
This new feature set of Wi-Fi 8 will accelerate the wireless LAN performance at home, in offices, venues, and factories, and enable applications like extended reality (XR), AI-assisted applications, real-time cloud gaming, and ultra-high-definition content streaming. To realize these benefits, test equipment must support all bands, full channel bandwidths, multi-antenna operation (MIMO), and deliver best-in-class measurement accuracy at benchmarking test efficiency. Rohde & Schwarz has designed the CMP180 radio communication tester with these capabilities in mind.
The CMP180 enables Qualcomm Technologies to validate essential features of its latest Wi-Fi innovation, including:
- 5×5 MIMO performance to further improve maximum data throughput per link
- Advanced modulation and coding schemes that enable fine‑grained adaptation to real‑time radio conditions.
- Distributed-tone resource units to improve uplink performance under regulatory limits.
Goce Talaganov, Vice President Mobile Radio Testers at Rohde & Schwarz, said: “We are excited to strengthen our long-time collaboration with Qualcomm Technologies to provide a unique testing solution for the next area of Wi-Fi innovations. The CMP180’s advanced features and our close collaboration will empower device manufacturers to bring innovative Wi-Fi 8 products to market quickly and confidently.”
Ganesh Swaminathan, Vice President and General Manager, Wireless Infrastructure and Networking, Qualcomm Technologies, Inc., said: “Qualcomm Technologies’ Wi-Fi 8 portfolio is engineered to deliver next-level performance, reliability, and scalability across a broad range of networking use cases. As part of this portfolio approach, we are advancing innovations such as higher-order MIMO to help increase performance in real-world environments. Our collaboration with Rohde & Schwarz highlights the progress of these capabilities as the Wi-Fi 8 ecosystem builds momentum.”
The post Rohde & Schwarz enables rapid validation of next-gen Wi-Fi 8 networking platforms, including 5×5 MIMO capabilities appeared first on ELE Times.
R&S acquires SRS, specialists in SDR communications solutions
Rohde & Schwarz acquired Software Radio Systems (SRS), a specialist in 5G software-defined radio systems. This acquisition further strengthens the position of Rohde & Schwarz in the cellular and wireless communications software market and accelerates development in AI-based test solutions for satellite and next-generation 6G wireless technologies. SRS will continue to operate under its own name as a Rohde & Schwarz group company, maintaining its product roadmap, leadership team and strategic focus.
The management team at Software Radio Systems, Paul Sutton, Ismael Gomez and Andre Puschmann, will remain. Together with their team at SRS and their new colleagues at Rohde & Schwarz, they will drive innovation in software-defined mobile communications and contribute to the long-term success of the group.
With the acquisition of Software Radio Systems (SRS), Rohde & Schwarz is expanding its software-defined radio (SDR) technology portfolio, strengthening its position in the market for mobile and radiocommunications software, especially in the emerging fields of non-terrestrial networks (NTN) and 6G. The acquisition became effective as of March 5, 2026. Founded in 2012 and headquartered in Ireland, with branches in Barcelona and the USA, SRS has grown from a startup into a globally recognised innovator in mobile and radiocommunications software. The company has established itself as a key player in a highly competitive market through deep technical expertise, product innovation and a strong commitment to open and interoperable network architectures.
By integrating SRS, Rohde & Schwarz decisively expands its capabilities in software-defined radio technology to better serve existing market segments and tap new business opportunities. At the same time, SRS enters a new phase of accelerated growth. With access to deep technical resources from Rohde & Schwarz, global reach and long-term strategic stability, SRS is ideally positioned to scale its technology, expand internationally and execute its long-term mission. The combination of Rohde & Schwarz leadership in test and measurement and RAN expertise from SRS creates powerful synergies that will further advance software-defined mobile network solutions.
Goce Talaganov, Vice President Mobile Radio Testers at Rohde & Schwarz, explains: “I am very pleased that Software Radio Systems is becoming part of the Rohde & Schwarz group. SRS combines extensive telecommunications and wireless expertise with agile software development, making it the ideal complement to our mobile radio testing portfolio, which will benefit the customers of both companies. What SRS and its employees have built up over the past few years is a real success story that you don’t often see. From now on, we will work together to push the boundaries of the technically possible even further, to leverage new synergies and to make an impact on the market.”
Paul Sutton, Chief Executive Officer and co-founder of Software Radio Systems, is ready for the future. “Joining the Rohde & Schwarz group is a proud moment for everyone at SRS. We will continue to operate as SRS with our existing leadership team and a clear commitment to our strategic priorities. What changes is our ability to scale with speed and accelerate the execution of our roadmap, including key initiatives such as the OCUDU project. With Rohde & Schwarz, we gain the strength of a global technology leader whose deep technical expertise, worldwide presence and long-term perspective enable us to grow faster and think bigger. Together, we will expand the impact of our software-defined solutions and deliver innovation that truly makes a difference for our customers.”
The post R&S acquires SRS, specialists in SDR communications solutions appeared first on ELE Times.
Differentiating Between LPDDR6, LPDDR5, and LPDDR5X
Courtesy: Synopsys
Advances in memory standards are driving faster and more power-efficient mobile and connected devices, from smartphones and tablets to ultra-thin laptops and wearables.
One such standard is Low Power Double Data Rate (LPDDR), which plays a crucial role in balancing high performance with energy efficiency. The latest iteration of the standard, LPDDR6, represents a big step forward in memory management. Comparing LPDDR6 to its predecessors, LPDDR5 and LPDDR5X, reveals just how quickly mobile memory technology is evolving — and what these advances mean for next-generation devices.
The role of LPDDR memory
LPDDR acts as the main system memory inside electronic devices. Working hand-in-hand with device processors and other components to store and access frequently used data, it helps keep applications, media, and multitasking features running smoothly. LPDDR is optimised for low power usage, compact footprint, and fast data transfer, making it ideal for portable, battery-powered devices.
LPDDR can integrate with Inline Memory Encryption (IME) modules to ensure data confidentiality — both in-use and when stored in off-chip memory. This is achieved through standards-compliant independent cryptographic support for read and write operations, providing robust protection against unauthorised access.
LPDDR memory is also available as automotive-grade Synchronous Dynamic Random-Access Memory (SDRAM), making it the preferred DRAM solution for automotive applications that require strict compliance with automotive standards.
LPDDR5 and LPDDR5X: the previous benchmarks
LPDDR5 marked a big step up in mobile memory when it was introduced in 2019. It delivered data rates up to 6.4 Gbps with improved energy efficiency (through features such as Dynamic Voltage Scaling) and smarter data handling. These upgrades led to longer battery life and better support for demanding applications like 5G connectivity, high-resolution media, and the initial wave of artificial intelligence (AI).
LPDDR5 also added new reliability features and smarter error handling, helping stabilise performance under complex workloads. As a result, devices using LPDDR5 delivered noticeable gains in both speed and overall user experience compared to devices using previous generations of LPDDR SDRAMs.
Introduced in 2021, LPDDR5X offered increased performance (up to 10.67 Gbps) and minor enhancements to LPDDR5’s features. LPDDR5X SDRAMs represent the vast majority of LPDDR SDRAMs shipping today.
LPDDR6: the next generation

Published in July 2025, the new LPDDR6 specification and compliant SDRAMs deliver even more performance, efficiency, and features — all designed to meet the growing demands of next-generation mobile and connected devices. LPDDR6 offers:
- Faster data rates. LPDDR6 is expected to reach up to 14.4 Gbps, a significant increase from LPDDR5X. This extra speed is essential for power-hungry applications like augmented reality, ultra-high-definition video streaming, advanced AI, and automotive electronics, all of which depend on rapid data processing.
- Wider bandwidth. Using 24-bit channels (up to 96 bits per package with 4 channels total), LPDDR6 effectively doubles LPDDR5X’s bandwidth per package. In addition, two 12-bit sub-channels in each channel help improve latency and access.
- Enhanced power management. LPDDR6 introduces more precise control over voltage and power states. This upgrade helps devices run more efficiently and extends their battery life as a result.
- Improved reliability and error correction. As the speed and footprint of LPDDR rise, so too does the risk of data errors — especially in data centres. LPDDR6 addresses this challenge with enhanced RAS (Reliability, Availability, and Serviceability) capabilities, providing robust error correction via Metadata, Advanced ECC, and Link ECC features. These improvements help minimise system glitches and stabilise device performance.
While LPDDR6 builds on LPDDR5 and LPDDR5X’s foundations, some legacy mechanisms were streamlined or replaced to support higher speeds and tighter power control. For example, earlier voltage scaling and command encoding schemes have been reworked to enable more granular power states and improved signal integrity. These changes mean LPDDR6 prioritises advanced efficiency and reliability features over older approaches that were optimised for lower data rates.

The implications for memory design and mobile devices
The improved performance, efficiency, and features of LPDDR6 will have wide-ranging impacts. From a technical perspective, LPDDR6 introduces a variety of upgrades to memory architecture:
- Signal integrity and bank management. Smarter signalling and improved memory bank management reduce latency and maximise data throughput.
- Ultra-low power modes. New power-saving states allow devices to conserve energy when idle, a big advantage for wearables and Internet of Things (IoT) products that run on small batteries.
- The new specification is engineered for seamless integration with the latest processors and chipsets, making it easier for manufacturers to integrate LPDDR6 into their next-generation devices.
These upgrades will enable the creation of mobile devices that offer:
- Faster, smoother performance. Higher data rates mean apps open quicker, multitasking is more efficient, and device operation is smoother.
- Better battery life. Improved power management reduces energy consumption, allowing devices to run longer between charges.
- Greater system stability. Stronger error correction improves reliability and reduces the risk of crashes and data loss.
- Future-proofing. LPDDR6 enables devices to support future advances in mobile computing, connectivity, and multimedia.
The Impact of LPDDR6 on smartphones, laptops, and wearables
LPDDR6 represents a significant step forward in mobile memory technology, delivering faster speeds, increased capacity, improved reliability, and better energy efficiency.
Leveraging silicon-proven interface IP and verification IP solutions — which have also been successfully validated at 10.667 Gb/s for SDRAM — device manufacturers are already upgrading their flagship smartphones, high-end laptops, and innovative wearables with LPDDR6-based memory.
But the transition from LPDDR5X/5 to LPDDR6 is more than just a technical upgrade — it enables new possibilities in mobile computing. As manufacturers adopt the new standard, users can expect devices that are faster, more reliable, and ready to support the next wave of on-device and cloud-connected experiences.
The post Differentiating Between LPDDR6, LPDDR5, and LPDDR5X appeared first on ELE Times.



