Pages Menu
TwitterRSSFacebook
Categories Menu

Posted in Q&A

Test Challenges Grow for DRAMs and HBM – Q&A with Jin Yokoyama

This article is adapted with permission from a Q&A with Advantest’s Jin Yokoyama, senior director and memory product marketing manager, with Mark LaPedus from Semiecosystem. The original article can be viewed here.

Q. In general, how do we test the latest DRAMs? Isn’t there a test flow involved here? Can you briefly describe the basic test flow for DRAMs and each step?

A. Typically, for DRAM devices, the test flow starts with wafer-level test. At the wafer level, our test systems apply memory test algorithms at varying speeds to test DRAM performance. The DRAM device is then tested again after it is assembled into a package. The final test stage runs tests at operational speed to confirm whether the device processes data correctly and meets end-use requirements.

 

Q. Over the years, DRAMs have become more complex. What are some of the technology trends that you see with current and future leading-edge DRAMs?

A. We see myriad trends emerging as DRAM technology evolves. DRAM devices are playing a critical role across a variety of applications, including artificial intelligence (AI)/high-performance computing (HPC), data centers, smartphones, AI PCs, and video game consoles.

The growth of AI, specifically, has led to an increase in demand for DRAM devices with higher per-pin speed, higher bandwidth, and low latency. Meanwhile, data centers and AI edge applications require devices to integrate new industry standards for speed and power consumption, including GDDR7, LPDDR6, and HBM4. These devices must also have lower output levels to maintain accuracy and higher operational efficiency to reduce cost.

We also see the trend of miniaturization and the shrinking of bit cells. There is a shift in size to 3,000 devices under test (DUTs) per 300mm wafer, leading to higher density that makes DRAM devices more vulnerable to defects. This means that these devices will likely require more tests and screening to meet demands for quality and reliability. Further densification of tester resources and technology will be necessary to accommodate device miniaturization trends.

Q. What are some of the test challenges with the latest DRAMs?

A. The increasing speed of these devices certainly presents a challenge in the test process. Speed performance requirements for at-speed interface testing on DRAM devices are increasing from 4.5 Gbps to over 10 Gbps, requiring high-performance test equipment to run test algorithms at high operational speeds.

We also need test systems that can test more devices in parallel to maintain lower costs. This requires highly specialized probe card technology that can handle high-speed testing. Massive levels of parallelism will be vital, as the industry trends toward requiring only one touchdown per wafer. Miniaturizing DRAM devices also makes them more susceptible to defects, requiring more thorough testing in the form of burn-in test at the wafer and package levels. Moreover, thermal requirements for DRAM devices continue to rise alongside demands for higher power and bandwidth capacity, and the complexity of integrating DRAM in 2.5D/3D packaging poses its own set of unique test challenges.

In the DRAM market, where high bandwidth and high performance continue to increase, Advantest recently introduced a next-generation ultra-high-speed solution, the T5801. This cutting-edge platform is engineered to support the latest advancements in high-speed memory technologies – including GDDR7, LPDDR6, and DDR6 – critical to meeting the growing demands of AI, HPC, and edge applications.

Q. What are some of the technology trends and challenges that you are seeing with HBM?

A. AI and HPC devices are demanding higher and higher processing speeds to power advanced applications like large-language models (LLMs). HBM consists of multiple DRAM chips stacked vertically. This structure allows for faster and more efficient data transfer with a smaller footprint, which is why HBM plays a key role in data centers. Of course, these HBM devices are incredibly complex and pose various challenges during testing. With multiple DRAM ICs stacked on top of each other, these devices are extremely dense and produce a lot of heat, which poses a risk to the structure of the device. Test systems must be equipped with the proper thermal-management capabilities and refined handler/probe equipment needed to monitor potential hotspots and dissipate heat.

The density of device circuitry and the number of DRAM devices stacked is increasing with each generation of HBM, moving from 8 stacks to 12, 16 and even 24. This has led to longer test times, especially as HBM migrates much faster than traditional memory technology transitions by generation. Higher interface speeds and higher bandwidth of HBM devices also mean high test speed and an increased load on peripheral circuits, requiring high-speed probing technology. Moreover, increases in power supply and current capacity require scalability.

The 2.5D/3D packaged structure of HBM devices also poses significant challenges, requiring thorough testing before the devices are packaged together. More and more manufacturers are looking to utilize die-level test, i.e., testing individual die after the wafer has been diced, to ensure that the DRAM ICs function properly before they are stacked into an HBM device and packaged, increasing overall yield and reliability. “Known-good die” or KGD testing, as it’s sometimes called, helps to prevent manufacturers from assembling packages containing defective die. If even one die in a stack is defective, then the entire package must be disposed of at great cost to the manufacturer.

Q. Are there any test challenges when the industry migrates to HBM4?

A. The greatest challenge we will see is the rising complexity of HBM4 devices. HBM4-based logic wafers are being made at leading-edge logic foundries for use in high-performance data centers to power AI edge applications. This will make the supply chain more complex, and the requirements and manufacturing and testing processes for DRAM manufacturers, foundries, and SoCs will also become more complex as a result. So, more advancements will have to be made in test process optimization in the future to support the next generation of HBM.

Read More

Posted in Top Stories

Rethinking Security in Semiconductor Testing: Why Containment Is the New Imperative

This article is adapted with permission from a recent Advantest blog post.

By Arik Peltz, Director, Technical Product Marketing, Advantest Cloud Solutions (ACS)

It’s nearly impossible to keep up with the headlines without stumbling upon another major cybersecurity incident. According to recent reports, 2024 witnessed a staggering 5.5 billion breaches globally. In the United States alone, the average cost of a single data breach clocked in at $9.36 million—slightly lower than 2023’s figure, but still a significant hit for any organization. On a global scale, the average breach cost was $4.88 million, reinforcing that the threat landscape is both vast and expensive.

Perimeter defenses are no longer enough

Traditionally, cybersecurity in semiconductor manufacturing and test environments has centered around perimeter defense—blocking threats before they enter. These measures include timely software updates, closing unused network ports, deploying firewalls, running IDS/IPS tools, and enforcing strict password protocols. While foundational, these practices fall short in the face of today’s diverse and evolving threats.

Why? Because the scope of potential vulnerabilities continues to grow. It’s not just about outdated software anymore. Misconfigured devices, overlooked network components, and even human error—like someone plugging in an unknown USB drive—can expose a facility. It’s clear: breaches are no longer rare events. They are inevitable. The challenge now lies in what happens after a breach occurs.

The semiconductor test floor: complex and exposed

A semiconductor test facility is far more intricate than it might appear. The test cell may be the focal point, but it operates within a larger network of systems—ranging from HVAC units and network switches to printers and third-party software tools. These are all potential gateways for exploitation.

Further complicating matters is the need for seamless cloud connectivity. Many test operations rely on real-time data sharing with external stakeholders for performance analytics and optimization. That constant network exposure, while essential for business operations, makes the environment even more susceptible to attacks.

The evolving nature of cyber threats

Today’s cyberattacks are smarter, stealthier, and increasingly driven by artificial intelligence (AI). Many use social engineering to build trust and then employ malware that mimics normal system behavior. These “slow burn” attacks can run undetected for extended periods, slowly extracting data or altering system operations. Conventional security systems often can’t spot these subtle anomalies in time.

The rise in customer-facing data sharing exacerbates this issue. As test facilities transmit more data to clients, they inadvertently widen their attack surface. Without a robust containment strategy, these vulnerabilities can become entry points for more sophisticated breaches.

Introducing a containment-first approach: True Zero Trust™

To address these challenges, Advantest advocates for a containment-first strategy, embodied in its True Zero Trust™ Environment. This model borrows its philosophy from high-risk environments like the International Space Station, where systems are designed to automatically isolate compromised sections to prevent broader damage.

Here’s how it works:

  • Automated containment gates on the test floor that activate in response to suspicious activity
  • Built-in assumption that threats can move both upward and downward across the facility
  • Shared accountability between OSATs and fabless companies for maintaining a secure test floor
  • Full traffic visibility via packet inspection and detailed logging
  • Encrypted communications between equipment and central containment servers
  • Verification of third-party software through mandatory software bills of materials (SBOMs)

This model doesn’t just focus on keeping threats out—it plans for when, not if, they get in. Every node, transaction, and data packet is treated as untrusted until proven otherwise.

The power of ACS RTDI™

At the heart of Advantest’s containment strategy is the ACS Real-Time Data Infrastructure (ACS RTDI™). This platform provides the foundation for building a secure, flexible, and data-rich test environment.

ACS RTDI™ includes:

  • A communication backplane spanning the entire test floor
  • Edge computing nodes that allow local processing and reduce data exposure
  • Advantest Unified Server for orchestration and policy enforcement
  • A containerized application hub that supports isolated, scalable software deployment

ACS RTDI™ not only facilitates secure operations but also unlocks advanced capabilities such as adaptive test, data feedforward and backward, outlier detection, and AI-driven predictive modeling. It’s security and innovation, rolled into one powerful solution. 

Looking Ahead: Secure Innovation as a Competitive Edge

The future of semiconductor testing is moving toward intelligent, semi-autonomous systems powered by AI. These applications demand real-time responsiveness and uninterrupted data flow—both of which are incompatible with outdated security models.

Containment isn’t just a cybersecurity necessity—it’s a foundation for innovation. Through the True Zero Trust™ framework and ACS RTDI™, organizations can safeguard their intellectual property while enabling breakthrough performance enhancements.

To further support this vision, Advantest offers the ACS Solution Store—a digital marketplace featuring software modules tailored to AI/ML, big data analytics, and secure test operations. These tools are designed to be customizable, scalable, and ready to meet the evolving demands of the semiconductor industry.

Read More

Posted in Top Stories

How AI Enables Adaptive Probe Cleaning

This article is adapted with permission from Electronics360. The original article can be viewed here.

By Nunzio Renzella, Specialist System Engineer, Advantest Europe GmbH

Test cost reduction is a key goal of semiconductor ATE customers, with major cost factors being test time, the test cell, and—for wafer-probing process—the probe card. Maintaining probe card cleanliness is a key aspect of a successful test process. Lack of cleanliness can result in good devices being rejected, but frequent probe card cleaning is time-consuming and increases the cost of test.

To optimize the cleaning interval, a new technique called adaptive probe card cleaning (APC) employs artificial intelligence (AI) and machine learning (ML) to provide real-time monitoring of probe card performance, initiating probe card cleaning only when necessary. The technique improves test efficiency and extends probe card life, resulting in cost savings and consistent performance while providing valuable insights that provide a competitive advantage.

Probe card cost breakdown

Probe card costs include the original purchase price. Probe cards are generally considered consumable products that need to be replaced, but during their useful life, they also incur online and offline maintenance costs. Online maintenance is performed inside the probe station, while offline maintenance requires the removal of the probe card for needle adjustment and replacement, tasks usually performed by the probe-card manufacturer.

In addition, probe cards incur costs related to the cleaning sheets required for the online cleaning process. One estimate puts the initial probe card purchase cost at 40% of the total probe card cost, with maintenance accounting for 55% of the total cost and cleaning sheets adding another 5%.

These figures point to maintenance cost as a key target for overall probe card cost reduction, and online maintenance costs are greatly influenced by the cleaning frequency. Currently, as illustrated in Figure 1, probe cleaning occurs on a fixed cleaning cycle based on a tradeoff between cleaning costs and yield loss. Complicating the estimation of the optimum cleaning cycle is the fact that yield loss varies from product to product and lot to lot. In the figure, for example, probing yield loss degrades much more quickly for lot B than for lot A. Taking into account cleaning costs and the wide range of yield losses, the customer in this example has estimated, based on practical experience and experimentation, that the “best” cleaning interval is at the 100-shot point.

 

Figure 1. Customers trade off cleaning costs (red trace) against yield losses (black traces) to determine the “best” cleaning interval (green bar).

Shifting the cleaning interval

In an effort to better optimize the timing of the cleaning cycle, Advantest has developed the software-based APC technique that applies AI algorithms to assess the condition of the probe needles in real time and to perform cleaning only when dirty needles begin affecting yield. In the Figure 2 example, APC enables probe cleaning to shift right to the 200-shot point, providing significant cost savings. 

Figure 2. APC can allow cleaning to be shifted to the right (to the yellow bar), saving overall cleaning time and costs.

The APC AI algorithm functions by observing failure trends among different sites. If trends differ between two sites, the algorithm judges that one or more needles are dirty on the site exhibiting the most failures, and it sends a general-purpose interface bus (GPIB) command to the prober to initiate the online cleaning cycle.

Figure 3 provides more detail on how to determine if the needle tips are dirty. For case 1 on top, the failure balance has significantly broken down, with site 1 showing many more failures than site 2. Consequently, the APC algorithm judges that a cleaning is required for site 1. For case 2 on the bottom, the failure balance remains relatively constant, and APC judges that cleaning is unnecessary; the observed failures may result from a wafer production issue. 

Figure 3. For case 1, the failure rate is out of balance, indicating a need for cleaning, while the relative balance of case 2 indicates that cleaning is not necessary.

A potential issue with APC arises when a continuous failure at a specific site results in APC triggering too many cleaning actions. To prevent this situation, APC can observe whether a cleaning has resolved an issue. If not, it concludes that the issue has a root cause other than a dirty needle, and it activates an Auto Bin Disabling function, which disables monitoring of the problematic site and bin. The customer can choose whether or not to use this function.

APC flow

The APC implementation flow (Figure 4) begins with a learning phase with a pre-fixed cleaning cycle (equal to the cycle used by the customer in the conventional probing approach) for the first wafer in a lot. The APC AI algorithm employs ML to determine the pass/fail tendency of each site. If the ML from the first wafer is successful, the adaptive-cleaning phase begins with the second wafer and continues with subsequent wafers, with cleaning performed whenever contamination is detected. If the ML is not successful on the first wafer due to low yield, the learning phase continues with the second wafer.

 

Figure 4. An AI algorithm employs ML on the first wafer and applies APC to other wafers in a lot.

Tester and prober independent

APC is adaptable and flexible. It works with Advantest’s T2000, V93000, and T6000 SoC test systems, as well as with other companies’ ATE systems. It also works with a variety of probers, including those from Accretech, SEMICS, and TEL, and with a variety of cantilever, vertical, and other probe-card types. APC supports parallel counts from two to 256 sites.

In addition, the APC algorithm executes in less than 1 ms per shot and will have a negligible impact on test throughput. Customers do not need to modify their test programs, and APC can run on the tester controller, so they do not need to add any specific hardware. Also, APC does not require access to Standard Test Data Format (STDF) data or any specific parametric data. Customers can use a standard prober driver, but they will need to add a GPIB command for initiating the cleaning process, and they will need to modify the prober recipe by adjusting the prober parameter setting.

Figure 5 summarizes the differences between the traditional fixed cleaning cycle (top) and the APC cycle (bottom), with the horizontal axis showing the elapsed time. As shown for the fixed cleaning cycle, a probe can become contaminated well before the 100-shot fixed cleaning interval, potentially resulting in many good devices being rejected (during the times indicated by the solid red arrows).

 

Figure 5. Fixed-cycle cleaning (top) can result in the rejection of good parts, while APC (bottom) can save these devices.

However, APC can detect contamination in real time, minimizing yield loss due to dirty needles and saving devices tested during the time indicated by the outlined red arrows.

APC results

APC has yielded favorable production line results. Based on one year of data, one customer reported that on one fab testing 37,000 wafers, it saw a 75% average cleaning reduction, and on another fab testing 10,000 wafers, it found a 50% to 80% reduction, both while maintaining the same yields. The same customer reported that probe card lifetimes were extended by 35% to 100%.

A second customer, using four different vertical probe card types, reported an online cleaning reduction of 65%, a yield improvement of 0.6%, and a lot test-time reduction of 7.4%. This customer also found that probe card life doubled and probe card maintenance costs were reduced by 50% per year.

To illustrate how APC can help prospective customers, Advantest offers an APC offline simulation capability. These customers can send Advantest STDF files for several wafer lots as well as information on the current probe-cleaning interval setting. Advantest will return an APC simulation report stating the achievable cleaning reduction ratio and including HBIN sequencing maps showing when and where the cleaning occurred.

Conclusion

Several factors can lead to probe tips becoming dirty, but many are unpredictable, and test managers must expect that the probe tips can become dirty at any time. APC can determine when this condition occurs, and it can take the appropriate cleaning action at the appropriate time. 

APC offers several specific benefits. It prolongs the life of probe needles, reduces probe-card maintenance costs, reduces yield loss, and shortens lot inspection times. In addition, customers need not manually adjust the cleaning cycle in response to new observations, and the optional Auto Bin Disabling function can automatically avoid continuous failures. Finally, APC reduces test time by eliminating extraneous probe-card cleaning and realignment after cleaning. 

Read More

Posted in Top Stories

AI Memory: Enabling the Next Era of High-Performance Computing

By Tadashi Oda, Senior Director of Memory and System Engineering, Advantest America

The rapid advancement of artificial intelligence (AI) is driving unprecedented demand for high-performance memory solutions. According to TechInsights, AI-driven applications are fueling a 70% year-over-year growth in high-bandwidth memory (HBM). However, as AI models grow in complexity—from large language models (LLMs) to real-time inference applications—the need for faster, higher-bandwidth, and energy-efficient memory architectures has become critical.

HBM devices consist of multiple stacked DRAM chips. The reduced mounting area and heightened wiring density of the stacked structure improve speed, capacity, and power consumption. Memory-intensive applications, such as high-performance computing (HPC), graphics processing units (GPUs), and AI accelerators, areas once dominated by DDR and GDDR memory, rely on HBM to meet the data throughput requirements of training and inference. At the same time, power-efficient alternatives such as low-power double data rate (LPDDR) are gaining traction in AI servers, where energy consumption is a primary concern. These trends, alongside the increasing integration of AI in mobile and edge applications, are shaping the future of AI memory technologies. However, these advancements bring significant test and validation challenges, requiring innovative solutions to ensure performance, reliability, and cost efficiency.

Market evolution and growth drivers

AI has become the dominant driver of innovation in the semiconductor industry, creating unprecedented demand for memory technologies. The computational complexity of AI models—particularly large language models and generative AI applications—has pushed GPUs and AI accelerators to their limits, requiring memory solutions that can sustain high-speed data access and processing; see Figure 1. HBM has emerged as the preferred solution for AI training workloads, offering multi-die stacking architectures that significantly increase memory bandwidth while minimizing power consumption.

Figure 1. AI data: near-memory compute for energy-efficient systems.

Beyond training, AI inference workloads are expanding into new applications, including mobile devices, automotive systems, and consumer electronics. These deployments require memory solutions optimized for efficiency, low power consumption, and cost-effectiveness. LPDDR, originally designed for mobile applications, is now gaining traction in AI servers due to its ability to deliver high performance with reduced power usage. As AI continues to move from centralized data centers to distributed computing environments, the demand for diverse memory architectures will continue to grow.

Investment in AI infrastructure is accelerating, with hyperscale data centers expanding their AI server capacity to accommodate increasingly complex models. AI workloads require not only high-performance memory but also specialized storage solutions that can handle vast datasets efficiently. The expansion of inference applications beyond data centers is also reshaping memory demand, as AI capabilities are integrated into smartphones, autonomous vehicles, and edge computing devices.

While AI-related markets are seeing rapid growth, traditional memory markets—including automotive, PCs, and mainstream mobile devices—are experiencing soft expansion. This shift reflects the broader industry transition from general-purpose computing to AI-driven architectures. As more AI workloads transition from training to inference, the balance of memory demand will continue to evolve.

AI memory architectures rely on advanced packaging techniques to optimize performance and power efficiency. The adoption of 2.5D / 3D stacking, heterogeneous integration, and silicon interposers enables higher memory densities and faster communication between memory and processors, as shown in Figure 2. These innovations improve AI system performance but introduce new challenges in manufacturing, validation, and test.

Figure 2. A diagram showing a typical HBM DRAM stack.

Ecosystem collaboration is also essential to advancing memory packaging technologies. Memory vendors, semiconductor foundries, and AI chip manufacturers must work closely to ensure seamless integration of next-generation memory architectures. Multi-die stacking and interposer-based designs require sophisticated testing and validation processes to detect defects early and optimize yield. As AI-driven memory solutions become more complex, new test methodologies will be required to ensure reliability and scalability.

The AI memory supply chain is undergoing a significant transformation, with increased vertical integration among semiconductor companies. Leading foundries are expanding their role in memory packaging and testing, consolidating manufacturing processes to improve efficiency. This shift requires closer collaboration between memory suppliers, GPU manufacturers, and semiconductor fabs to optimize design and production workflows.

At the same time, supply chain constraints—particularly in high-performance silicon interposers and through-silicon via (TSV) technologies—are impacting memory production. AI-driven demand for HBM and LPDDR is straining existing manufacturing capacity, making supply chain coordination more critical than ever. Companies must navigate these challenges while maintaining production efficiency and meeting the growing needs of AI applications.

Technology advancements in AI memory

HBM has become the foundation of AI computing, evolving from traditional DDR and GDDR memories to multi-die stacking architectures that deliver extreme bandwidth and low latency. Figure 3 depicts this evolution. The ability to stack multiple memory dies vertically and integrate them with high-speed interconnects allows AI accelerators to process massive datasets more efficiently. However, these advanced designs present unique challenges in manufacturing and test.

Figure 3. Descriptions of commonly used types of memory devices.

As AI workloads expand beyond data centers, power efficiency is becoming a key consideration in memory design. LPDDR, originally developed for mobile devices, is now being adopted in AI servers to reduce power consumption while maintaining high performance. Although LPDDR carries a higher upfront cost than traditional DDR memory, its long-term energy savings make it an attractive option for large-scale AI deployments.

Balancing performance, cost, and power efficiency is a major challenge for AI memory architects. LPDDR must be rigorously tested to ensure it meets the performance demands of AI applications while maintaining the power-saving benefits that make it viable. As AI adoption grows in power-sensitive environments such as mobile and edge computing, LPDDR is expected to play an increasingly important role in AI memory solutions.

AI workloads require specialized storage solutions that can handle the massive data volumes associated with model training and inference. Enterprise AI storage relies on high-speed solid-state drives (SSDs) designed to support AI-driven workloads, enabling fast data retrieval and processing in hyperscale data centers. Meanwhile, edge and on-device AI applications depend on Universal Flash Storage (UFS), a high-speed, low-power interface optimized for flash memory in mobile devices.

Ensuring the performance and reliability of AI storage solutions requires advanced testing methodologies. Both enterprise SSDs and mobile UFS solutions must be validated under AI-specific workloads to ensure they can handle the demands of real-time AI processing. As AI applications continue to diversify, memory and storage technologies will need to evolve accordingly.

Another AI memory function advancement is processing-in-memory (PIM) technology. AI training and inference require significant computational or mathematical resources to operate a combination of matrix multiplications, vector operations, activation functions, or gradient calculations. Figure 4 illustrates the benefits of implementing PIM in conjunction with LPDDR.

The purpose of PIM is to implement processing features inside the memory chip to reduce data movement between the memory and the processor. Because it reduces power consumption and increases performance, PIM is considered particularly effective in the mobile space for enabling a range of AI-powered applications. PIM is one example of how semiconductor memory is a critical component for the AI industry and illustrates how technology continues to evolve to advance future AI capabilities.

Figure 4. Memory in the AI/ML and data era.

Key market and test challenges for HBM

Ensuring the reliability of HBM stacks requires multiple test insertions throughout the production process. Each die within the stack must be validated for performance and integrity before final assembly, as shown in Figure 5. Before stacking, manufacturers perform wafer test that consists of burn-in stress, trimming, failure analysis and repair processes to ensure all memory cells work properly. The logic base die is also tested at wafer-level through SCAN test, IEEE1500 test, and logic test. Additional tests are performed after the DRAM die are stacked on top of the logic base die wafer, including burn-in stress test, DRAM array test, logic die test and high-speed operation test to ensure proper TSV connectivity within the DRAM stack. Advantest has been collaborating with DRAM vendors to develop HBM tests since before volume production began. As a result, the company now supports all test processes essential for the mass production of AI devices.

Thermal management is also a critical consideration, as the increased power density of HBM configurations can lead to overheating and performance degradation. As AI workloads continue to push memory technology forward, new innovations in stacked memory design and thermal control will be essential. Relating to thermal management, the high-power consumption of AI memory solutions is also a concern. Enhanced thermal control solutions, including precision cooling mechanisms, are necessary to address heat dissipation challenges. 

Increasing AI memory complexity heightens the challenges associated with the testing and handling of these devices. Ensuring die integrity during stacking and transportation requires the development of specialized die carriers to mitigate physical damage. Advances in automated handling systems improve secure stacking and assembly, reducing manufacturing defects.

The memory supply chain is also evolving, with vendors collaborating more closely with fabs and GPU manufacturers. The shift from siloed production models to integrated development ecosystems aims to overcome supply constraints and streamline production. However, challenges remain, particularly in securing high-performance silicon interposers and TSV technologies, which are essential for advanced memory integration.

Figure 5. Test and assembly process for advanced memory devices.

Advantest’s solutions and market leadership

Advantest is at the forefront of AI memory testing, offering comprehensive solutions that address industry challenges. Various products in the company’s T5000 series, such as the T5833, provide high-speed, high-accuracy testing for HBM configurations, incorporating multi-step validation processes to ensure die integrity. Its modular architecture provides flexibility to scale with evolving test requirements, optimizing cost and efficiency for high-throughput production environments. This means that the platform is prepared for future testing of LPDDR6 devices as well as mobile storage protocols like UFS. 

Moving forward in this space, HBM4/E increases power consumption, doubling the DQ pin count in the DDR bus to 2048 DQs. The industry is also considering the development of “custom HBM” to tune specific areas of the AI workload. Advantest is working with industry-leading GPU and ASIC companies and hyperscalers to prepare for custom HBM testing.

Advantest’s proprietary die carriers ensure secure handling, minimizing physical stress on stacked memory dies and preventing defects during high-volume production. The company recognizes the need for handlers that provide superior thermal capabilities to address the growing demand for precision temperature management in AI memory applications.

The latest addition to the T5000 series, the T5801 ultra high-speed memory tester, is engineered to support the latest advancements in high-speed memory technologies —including GDDR7, LPDDR6, and DDR6— critical to meeting the growing demands of AI, HPC, and edge applications. Featuring an innovative front-end unit (FEU) architecture, the system is uniquely equipped to handle the rigorous requirements of next-generation DRAM modules, delivering industry-leading performance of up to 36Gbps PAM3 and 18Gbps NRZ.

The modular and scalable T5851 platform reflects Advantest’s deep collaboration with industry leaders to meet the evolving needs of mobile storage. Designed in close partnership with leading mobile device companies and NAND manufacturers, the system has been supporting multiple generations of PCIe protocols. Its system-level test capabilities enable realistic workloads, ensuring read/write performance, link stability and storage reliability under actual operating conditions. With support for emerging standards like PCIe Gen 6 and UFS M-PHY6.0 using PAM4 signaling, the T5851 showcases Advantest’s commitment to co-developing future-ready solutions for next-generation AI devices.

Conclusion

As AI memory technologies continue to evolve to support increasingly complex workloads, Advantest remains a trusted partner in delivering reliable, scalable, and high-performance memory test solutions. With a vertically integrated strategy, Advantest distinguishes itself as a leader in testers, handlers, and load boards, offering broad support across the memory test ecosystem. Trusted by leading semiconductor manufacturers, Advantest works closely with industry partners to innovate and develop unique solutions that meet their needs. 

Moreover, Advantest ensures its platforms are aligned with emerging standards and practical requirements. The T5000 series—including the T5833, T5835, T5503HS2, T5801, and T5851—reflects this commitment, offering modular, high-speed, and flexible solutions for a wide range of memory and storage technologies, from AI-critical HBM to LPDDR, GDDR, DDR, NAND, NOR, and beyond.

Advantest’s continued innovation in areas such as die carriers and thermal management helps address the physical and operational challenges of stacked memory and high-power AI applications. As AI workloads expand across data centers, mobile, and edge environments, the company remains focused on advancing test methodologies that support performance, reliability, and efficiency.

With a forward-looking roadmap and strong industry partnerships, Advantest is well-positioned to support the next generation of AI memory architectures, helping customers navigate complexity and drive innovation in the AI era.

Read More

Posted in Top Stories

Tackling Chip Complexity with Integrated System-Level Test Solutions

By Davette Berry, Senior Director of Business Development, Advantest

As the sophistication of semiconductors continues to grow, so does the need for system-level test (SLT) in production to ensure that high-performance processors, chiplets, and other advanced devices function as expected in real-world environments. Once seen primarily as a fallback to catch what traditional automated test equipment (ATE) missed, SLT has now become a mission-critical step for validating AI accelerators and central, graphical and application processing units (CPUs, GPUs, and APUs) before market release. It’s increasingly being evaluated for use in production of devices like network and automotive processors—where reliability is non-negotiable. However, implementing SLT at production scale introduces new complexities in balancing cost, throughput, and test coverage.

While ATE works by driving known test patterns to stimulate in-chip circuitry and observe expected internal responses, SLT, in contrast, looks at how the chip behaves as part of a larger system, focusing on interactions across cores and peripherals, including aspects like power regulation and sensor behavior. This means that SLT platforms must support a broad mix of application-level use cases and interfaces, especially when testing cutting-edge semiconductors.

This requirement grows even more pressing with the rise of chiplet-based design strategies. Instead of assessing a single chip in isolation, SLT can evaluate the communication between multiple dies in a single package. This ability to help validate cross-chip data paths and their impact on power, performance, and reliability is a key advantage of SLT. However, many SLT deployments still rely on manual methods of creating the test content, which limits their scalability.

Overcoming coverage gaps

Electronic design automation (EDA) companies have helped chip designers automatically generate structural test patterns to achieve close to 99% transistor coverage. With today’s 100-billion-transistor AI processors, this coverage still leaves a billion transistors unchecked. Closing this last 1% gap using only ATE is an expensive, time-consuming effort, often requiring months of development time.

Furthermore, chiplet integration introduces new mechanical and thermal challenges. With fewer external access points, test engineers must route signals through complex, multi-die pathways. These large packages can warp, complicating socket alignment and risking poor (device-under-test) DUT connectivity. Under load, heat spots form across tightly integrated dies—an issue that demands close coordination between thermal control, actuation mechanics, and power sequencing in the SLT system.

SLT implementation also requires cross-disciplinary cooperation. Test content needs to be co-developed and validated by stakeholders across the ecosystem—equipment and socket vendors, silicon designers, OSATs, and application board manufacturers, along with end-users like hyperscale data center operators or smartphone OEMs. The SLT test station must authentically replicate the device’s actual operating environment.

As power demands rise, testing infrastructure must keep pace. SLT test times can last 30 minutes or more, pushing facilities to deploy dense configurations of testers within space- and energy-efficient layouts. At the same time, the DUTs and their supporting hardware continue to grow in size and complexity.

Advances in test methodology

The chip industry has responded with new design for test (DFT) methods that deliver test data over high-speed serial interfaces—like USB or PCIe—rather than parallel pin scan chains. During SLT, once these ports are activated and enumerated, test programs can operate through them to trigger built-in tests or deliver packetized patterns, minimizing pin requirements while maximizing coverage.

Once validated, this structural test content can be correlated across various platforms—ATE, SLT, and post-silicon validation—improving debug efficiency and accelerating time to market. Advantest supports this with platforms like Link Scale and our new SiConic™ system, helping unify the test and validation landscape.

Thermal management remains a universal concern. Testing power-hungry processors in aggressive workloads stresses both the DUT and the test infrastructure. Solutions today span from traditional air cooling to advanced liquid and refrigerant-based systems, all while emphasizing environmental sustainability. SLT handlers must support repeated thermal cycling and actuation without compromising electro-mechanical stability or DUT safety.

Machine learning and AI bring fresh opportunities to optimize test operations. Advantest’s platforms, now equipped with Real-Time Data Infrastructure (RTDI™), deliver fast, secure access to test data, empowering AI tools to enhance yield and resource utilization.

Advantest’s integrated SLT offering

Advantest delivers a cohesive ecosystem of SLT solutions tailored to modern semiconductor demands. Our lineup includes ATE handlers, lab-grade engineering testers, and full-scale SLT systems, all designed with shared physical and thermal interfaces to simplify correlation and integration.

Our thermal systems will support up to 5000 W per DUT and enable multizone thermal management, with programmable set points that adapt dynamically during test cycles. These capabilities are essential for maintaining test fidelity in the face of rising power levels and integration density.

In the past, some chipmakers had no choice but to build their own SLT test setups from scratch. Advantest now offers turnkey SLT test cells that integrate all critical components—test execution, thermal and power control, mechanical handling, and test content delivery—backed by our global service infrastructure. These systems ensure every test cell remains consistent, controlled, and up to date.

We work directly with customers to simulate how packages will behave thermally and mechanically during test. This includes modeling warpage, compression, and power dynamics—all of which are crucial for validating new packaging formats, including future designs like co-packaged optics.

A Software Suite for End-to-End Device Control and Holistic Test Cell Management

To orchestrate all this complexity, Advantest offers ActivATE360™, our integrated software suite for SLT and burn-in systems. The platform includes:

  • Studio360 – a complete integrated development environment and software development kit for test program development and test hardware control.
  • Device360 – for managing DUT communication and executing binaries, test contents and test flows.
  • Cluster360 – enables real-time, cross-platform messaging with support for multiple languages and application interfaces.
  • Cell360 – to manage distributed test cells and process lots and generate operator instructions.
  • Facility360 – for monitoring and optimizing test operations at the facility level.

ActivATE360 seamlessly communicates with Advantest Cloud Solutions (ACS) and SmarTest 8 software for ATE, enabling real-time data sharing, unified debug, and cross-insertion test correlation. Used with hardware like Link Scale and SiConic, these tools help accelerate validation and close the loop from silicon design to high-volume production.

Delivering SLT at Scale

With decades of expertise in building high-throughput test systems, Advantest is well-positioned to meet the demands of system-level testing. Our robust handlers and SLT systems are built for longevity, mechanical precision, and thermal performance—while maintaining uptime in high-volume environments. Whether you’re testing massive AI chips or future-ready chiplet systems, Advantest ensures that every SLT investment delivers maximum value, scalability, and support.

 

 

 

Read More