|
||||||
Reconfigurable illogic
Reconfigurable illogic Some say the microprocessor is dead. Not so fast, says industry expert Rich Belgard. The microprocessor isn't going away. As with the laws of motion, you can't just shift energy or complexity around without paying the price. Here's a more optimistic prognosis for the microprocessor. For at least a decade, some people have been predicting that reconfigurable logic will replace the microprocessor. That hasn't happened yet, and I'm convinced it won't. Although reconfigurable logic has its place, I believe it's in addition to or as an extension of the microproc Nick Tredennick and Brian Shimamoto conclude in their article ("The Death of Microprocessors") that microprocessors will no longer be capable of handling the new age of applicationswhat the authors call untethered systems. They define an untethered system as battery operated and not plugged into a wall. Fittingly, I'm writing this counterpoint on a new Intel ULV-Pentium M laptop, somewhere over Kansas. Weighing around three pounds, my laptop says I have about four and half hours of battery left. On this particular trip, my Blackberry has been powered-on for the past three days, has received my e-mail, and still has 85% battery life. I have my trusty cell phone with me, and its battery lasts about a week, if I'm circumspect. Each of these untethered devices uses at least one microprocessor, and they all work pretty well. So, I find it quite strange that Tredennick and Shimamoto suggest I'll need to throw out everything I know and our industry knows and switch the entire model to reconfigurable systems. Defining a reconfigurable system Apparently, it's much more easily described by what it is not. Tredennick and Shimamoto claim it's not a PLD nor FPGA as we know them. They say these are too expensive, slow, and power hungry. It's not an extensible microprocessor, such as those ARC and Tensilica have created, because those are configured at design time. And, clearly, it's not a standard microprocessor. So exactly what is it? First, a reconfigurable system sounds to me as if it's something that relies on or is enabled by a new, nonvolatile memory. Again, the authors tell us what this new enabling memory is not, but they're short on what they claim it is. The character of this new memory is that it's fast, cheap, dense, nonvolatile, and, processwise, compatible with logic. Tell me again where this comes from? Given this miracle memory, we'll page the hardware configurations into and out of a programmable-logic area on the chip. We can do this on application boundaries, and when we get proficient as Tredennick and Shimamoto say we will, we'll even do it cycle by cycle. Well, I can't speak to reconfiguring logic on a cycle-by-cycle basis, but as far as reconfiguring logic: I think we've already been there. Been there, done that This machine, although hardly a microprocessor, had different instruction sets for each of a system development language (SDL), and the application languages of the time: Fortran, Cobol, and others. The microarchitecture of the machine was generic, in that the programmers' model, for example, could assume any word length; had no predefined condition codes; had a variable arithmetic logic-unit width; allowed stack and flat memory; and, through microprogramming, supported soft instruction sets (called S-languages) for the interface to each of the programming languages. Some of the experimentation we did at Burroughs and that was carried out at a few universities that had B1700s was to build microprograms dedicated to applications, not just S-languages. Not surprisingly (at least to computer scientists older than 30), these applications ran an order of magnitudesometimes multiple orders of magnitudefaster than the S-language versions of the programs to which the applications were compiled. Why? Because we didn't have to go to memory to pick up an instruction, decode it, get the operands, vector to its implementation in microcode, perform its steps, check for interrupts, and do it all over again. This experimentation led to conference papers on subjects like "vertical migration," where experimenters moved sections of applications into microcode and back to decide on the "proper" level of performance versus complexity of implementation, since this microcode was difficult to write. Some people even suggested migrating the whole problem into in-line microcode, but the microcode was difficult to write. Remember this. Of course, the B1700 reconfigured itself at the start of each program and again at task swaps. At the time, we didn't call it reconfigurable logic, but I think that's what it was. We've had reconfigurable systems, such as the Burroughs B1700 and others. We've had them before, and if we built them today, in today's technology, we'd call them microprocessorseven though they would include reconfigurable logic. Cause of death Another premise is that we, as an industry, will be unable to make smaller, faster transistors for microprocessors. This has nothing to do with replacing microprocessors with reconfigurable systems. Whether we use transistors to build microprocessors or reconfigurable logic, we still need fast, small transistors. To achieve the same densities we have (or will have) in microprocessors, we'll also need to put these smaller, faster transistors into reconfigurable logic. The ultimate technology that makes reconfigurable logic work will also make microprocessors work. Many Japanese semiconductor companies have experimented for almost a decade with reducing leakage current in transistors by biasing the substrates of the transistors. And, what of the newly announced LongRun 2 system Transmeta has developed for microprocessors? As I understand it, Transmeta's LongRun 2 reduces leakage current by programmatically adjusting the bias voltage of the transistor substrate to effect a larger threshold voltage in the transistors when the transistors aren't required to run at their fastest. I understand Transmeta can do this across different parts of the substrate. Therefore, total leakage current can be substantially reduced. Today, virtually every semiconductor manufacturer is working on new materials, new techniques, and new processes to solve the leakage problem. Finally, a key premise of the article is that reconfigurable logic will replace microprocessors with the advent of memory that is fast, cheap, dense, nonvolatile, and logic-process compatible. It's difficult to argue that such a memory won't change the industry. But I don't believe this memory necessarily argues for replacing microprocessors with reconfigurable logic. I can imagine using this memory in microprocessors as easily as I can imagine using it in reconfigurable logic. No free lunch We can move the inherent complexity of a problem's solution from microprocessors into reconfigurable logic, but where does the complexity we see in contemporary microprocessors go? Does it go into the design tools? Does it go into the control of the devices? Does it go into the interconnects? I don't know. But my experience tells me it goes somewhere. Microprocessors take much of the complexity of the problem domain and encapsulate it into the microprocessor design. Those hundreds of designers and engineers in the engineering groups of the microprocessor companies shoulder the weight of much of the complexity. Thousands of person-years of compiler and other software design absorb more of the complexity. Ultimately, the tools and the microprocessors give a platform to the customer, with a lot of the complexity abstracted out by experts in microarchitecture, programmers' model, and tools. ARC and Tensilica are creating microprocessors that allow a degree of configurability by enabling extensions to the instruction sets of their microprocessors. I would imagine that the companies creating these extensions do so to increase the performance of operations specific to these companies' specific problem domain. These companies factor specific complexities out of the programs that run on the microprocessors and into the microprocessors themselves. DSP-like extensions, SIMD extensions, and other extensions important to specific customers are created to increase the performance of the algorithms that are important to these customers. General-purpose microprocessors can't anticipate these specific algorithms and therefore leave the complexity to the user. To move application-specific complexities into the design of the underlying microprocessor once, as these companies allow, makes sense. Altera, Atmel, Xilinx, and others have begun embedding microprocessor cores on their programmable logic devices. They come at the microprocessor-extensions solution from the opposite end: start with application-specific portions and let the designers use microprocessor software for the rest. This solution is similar to the ones from ARC and Tensilica, but it comes from a different worldview. This solution also makes sense. Tools and infrastructure When there were user-microprogrammable systems, such as the B1700, few people actually microprogrammed these systems because the complexity was too hard to handle. The tools needed were just too primitive. Implementing reconfigurable systems will require sophisticated tools and an infrastructure that doesn't yet exist. Some weeks ago, I attended a panel where one of the speakers said something like "I couldn't imagine having to write Microsoft Word in VHDL." Neither can I. MPU lives on Sure, we trade off efficiency by using microprocessors instead of full-custom reconfigurable systems. But we trade it for time to market, for economy of scale, for cost, and for abstracting out implementation complexity, so that the problem solvers can think in their domain and not in the domain of the implementation. We now see microprocessors that have custom extensions, and we see traditional PLDs that incorporate standard microprocessor cores. These hybrid approaches, combining reconfigurable (or reconfigured) logic as an adjunct to standard microprocessor cores, make sense. Are fully reconfigurable systems the completely wrong way to go? No, certainly not. In the future, there will be a niche for systems that can use a reconfigurable logic, but reconfigurable systems will never replace the microprocessor. Rich Belgard is a contributing editor of The Microprocessor Report and a consultant who has been active in the computer industry for more than 30 years. He designed and managed the development of computer architectures at Burroughs, Data General, Tandem Computers, and Rational Software, including hardware, software, and microarchitecture and served as chairman and vice chairman of the Association for Computing Machinery's Special Interest Group on Microarchitectures and vice chairman of the Institute of Electrical and Electronic Engineers' Technical Committee on Microprogramming and Microarchitectures. Copyright 2005 © CMP Media LLC |
Home | Feedback | Register | Site Map |
All material on this site Copyright © 2017 Design And Reuse S.A. All rights reserved. |