As semiconductor manufacturers build ever-smaller components, circuits and chips at the nanoscale become less reliable and more expensive to produce. The variability in their behavior from device to device and over their lifetimes — due to manufacturing, aging-related wear-out, and varying operating environments — is largely ignored by today’s mainstream computer systems.
Now a visionary team of computer scientists and electrical engineers from six universities, including UCLA, is proposing to deal with the downside of nanoscale computer components by rethinking and enhancing the role that software can play in a new class of computing machines that are adaptive and highly energy efficient.
The National Science Foundation today awarded a $10 million, five-year grant to these researchers for their research initiative “Variability-Aware Software for Efficient Computing with Nanoscale Devices.” The grant is part of the funding agency’s Expeditions in Computing program, which rewards far-reaching agendas that “promise significant advances in the computing frontier and great benefit to society.”
“We envision a world where system components — led by proactive software — routinely monitor, predict and adapt to the variability in manufactured computing systems,” said Rajesh Gupta, director of the Variability Expedition and a professor of computer science and engineering at the Jacobs School of Engineering at UC San Diego. “Changing the way software interacts with hardware offers the best hope for perpetuating the fundamental gains in computing performance at lower cost of the past 40 years.”
Joining Gupta in this effort is the expedition’s deputy director, Mani Srivastava, an electrical engineering professor at UCLA’s Henry Samueli School of Engineering and Applied Science, and a team of eleven other computer scientists and electrical engineers from University of Michigan, Stanford University, UC Irvine and the University of Illinois at Urbana-Champaign. The project is based in the UCSD division of the California Institute for Telecommunications and Information Technology (Calit2).
Researchers on the team include: Ranjit Jhala, Sorin Lerner, Tajana Simunic Rosing, Steve Swanson and Yuanyuan “YY” Zhou, from UC San Diego; Lara Dolecek and Puneet Gupta, from UCLA Engineering; Subhasish Mitra, from Stanford; Dennis Sylvester, from Michigan; Rakesh Kumar, from the University of Illinois at Urbana-Champaign’s Coordinated Science Laboratory; and Nikil Dutt and Alex Nicolau, from UC Irvine’s Center for Embedded Computer Systems.
These researchers bring to the project broad expertise in a wide range of relevant topics. Their specialties include operating systems, compilers, languages, embedded and mobile systems, data-center computing, integrated circuits, storage, statistical techniques, design automation and testing, and cross-cutting issues spanning reliability, energy efficiency and robust system design.
The research team seeks to develop computing systems that will be able to sense the nature and extent of variations in their hardware circuits and expose these variations to compilers, operating systems and applications to drive adaptations in the software stack.
“As the transistors on their chips get smaller, semiconductor makers are experiencing lower yields and more variability — in other words, more circuits have to be thrown away because they don’t meet the timing-, power- and lifetime-related specifications,” said Michigan’s Sylvester, an expert in designing computer circuits in nanoscale technologies. If left unaddressed, added UCLA’s Puneet Gupta, “this trend toward parts that scale in neither capability nor cost will cripple the computing and information technology industries. So we need to find a solution to the variability problem.”
Software experts on the team will develop models and abstractions to expose the hardware’s variable specifications accurately and compactly and to create mechanisms for the software to react to variable hardware specifications. Hardware researchers will be focused on more efficient design and test methods to ensure that device designs exhibit well-behaved variability characteristics — ones that a well-configured software stack can easily communicate with and influence.
“The resulting computer systems will work while using components that vary in performance or grow less reliable over time,” said UCLA’s Srivastava. “A fluid software-hardware interface will mitigate the variability of manufactured systems and make them robust, reliable and responsive to the changing operating conditions.”
Added professor Rakesh Kumar, who will lead the expedition efforts at Urbana-Champaign: “Steering the effects of the variability will be particularly important.”
Variability-aware computing systems would benefit the entire spectrum of embedded, mobile, desktop and server-class applications by dramatically reducing hardware design and test costs for computing systems while enhancing their performance and energy efficiency. Many in-demand applications — from search engines to medical imaging — would also benefit, but the project’s initial focus will be on wireless sensing, software radio and mobile platforms of all kinds, with plans to transfer advances in these early areas to the marketplace.
To ensure that the project reflects real-world challenges in the computing industry, organizers have recruited a high-powered technical advisory board that initially includes top industry executives from HP, ARM, IBM and Intel. The board also includes two senior academic researchers with expertise in modeling and manufacturing of nanoscale devices and circuits: Robert Dutton, a professor at Stanford, and Andrew Kahng, a professor UC San Diego.
“If this project is successful and the breakthroughs are transferred to industry,” said Stanford’s Subhasish Mitra, “we will have contributed to the continued expansion and reach of the semiconductor and computing industries.”