|
||||||||||
High Performance Connectivity IP -- Avoiding Pitfalls When Selecting An IP Vendor
By Navraj S. Nandra Synopsys
Mountain View, California USA Abstract: The demand for connectivity IP for high speed serial busses such as USB 2.0, PCI Express®, SATA, DDR2 and HDMI is increasing as standard interfaces in applications such as single chip recordable DVD CODEC’s and MP3 players. In order to stretch battery life of these chips, the semiconductor technologies require ultra-low power derivatives of high-performance logic manufacturing processes, enabling production of very low-power chips for these mobile platforms and small-form factor devices. Today many of these chips are manufactured in 90 nm, and the ramp for 65 nm design starts has been more aggressive than expected. 45 nm is following close behind, with early versions of design rules and process parameters already available. Since connectivity IP requires both a digital controller (MAC) and mixed-signal circuitry (PHY), the challenge is to meet analog performance in a technology that has been targeted for densely packed digital logic. IP is the critical enabler to meet the time to market demands for these SoC’s and an industry of IP providers has mushroomed to service this demand. The key question for the SoC developer, and the goal of this paper, is to highlight the importance of selecting the right IP vendor – how to avoid the pitfalls. In this paper, we will discuss how to select a third party IP vendor, how to verify third party IP, and some of the gotcha's when integrating third party IP, with a special focus on the SerDes-based PHYs for PCI Express and SATA as well as PHYs for USB and DDR2. In addition, we will discuss the impact of 65 nm and 45 nm process effects on yield, we will review the advantages and disadvantages of moving to serial links, we will propose complete vertical integrated solutions, and we will present production testing techniques. As the number of third party IP vendors increase, the quality of the IP available varies. This paper discusses IP selection criteria to ensure that the right choices are made. Figure 1 highlights the essential elements of IP selection, in terms of functional correctness, integration, usability and support. Figure 1 Essential elements of IP Selection Obviously, the IP must be correct functionally, but other factors are important such as ease of integration and support by experienced protocol experts – with design teams located world-wide, the IP vendor must be able to provide this support round-the-clock, every day. 1.0 The Gotcha’s – The Ten Questions Figure 1 described a high-level view of IP vendor selection, but due to the complex nature of connectivity IP, further investigation is needed. The following are the ten most important considerations:
Figure 2 shows the seven layer OSI model. For connectivity IP, this is divided into levels L1 to L4. Ideally, an interface standard (e.g., USB and PCI Express) IP vendor should provide a complete solution including all of these, essentially circumventing interoperability issues. Figure 2 Seven layer OSI model The second step in the process is to determine how thoroughly the IP was verified. For mature IP (which has undergone multiple successful tape-outs), this effort is minimal. However, for new or emerging IP, the certification and verification effort can be challenging. This effort is usually underestimated by the SoC implementer, and in some cases ignored by the IP vendor. The key is to ensure that the IP has gone through a certification process, for example with PCIe SIG. The third step is to get written commitment from the vendor on issues that you may find during product development and IP integration. The next section discusses the above challenges in the context of the commonly used and emerging IP. 2.0 High Speed Serial Links Recently, high-speed serial links have become increasingly common place in the industry. There are a number of advantages to using high-speed serial links instead of source-synchronous parallel I/O currently deployed. Because of their advantages, the industry will be moving toward high-speed serial links for those applications where performance, footprint, and cost are paramount. The popular protocols are the SerDes based (serializer-deserializer) PHY’s for PCI Express and SATA. Figure 3 clearly shows the advantage of using the serial based approach compared to the parallel cable. Figure 3 ATA cable (below) and SATA cable 2.1 High Speed Serial Link Advantages The following summarizes the advantages of serial links.
2.2 High Speed Serial Link Disadvantages There are some disadvantages and these should be discussed with the IP vendor.
2.3 Choosing Serial Link IP As with any new technology, the trade-offs between developing the high-speed serial link interface internally or licensing the IP through a third party need to be thoroughly examined. The high speed serial link consists of two layers: the physical layer (PHY) and the media-access layer (MAC). The PHY is responsible for merging slow incoming data (16 bits at 125 MHz for PCI Express) and clock (125 MHz for PCI Express) and generating two differential high-speed outputs (2.5 Gbps). The PHY is also responsible for decoding the incoming high-speed serial data stream. The MAC handles all high-level functionality, such as flow control, transaction generation, routing, etc. In PCI Express and SATA, the MAC layer includes the transaction layer and the data link layer. Figure 4 PCI Express, complete solution Since the PCI Express and SATA PHY’s are complex (involving high-speed analog/mixed-signal design), it is usually licensed from a third party. Few companies are capable of developing the PHY internally or in sufficient time. The MAC can either be developed internally or licensed as well. This will largely depend on the time to market requirements, availability if suitable IP, and verification requirements. For PCI Express and SATA, with strict requirements in compatibility, verification, and testing, the development time for the MAC layer can be quite long. For this reason, licensing the MAC as well as the PHY makes sense. When licensing both the PHY and MAC layers from the same third party, the two layers should have been tested together. It is important to know what configurations were tested together and what methods of verification were utilized. Figure 4 PCI Express, complete solution shows the PCI Express complete solution with various configurations of MAC that are available and also the verification IP. If the two layers are licensed from two different vendors, then the verification challenge is much greater. For example, in PCI Express, the majority of PHY and MAC layers communicate to each other over the PIPE interface (as defined by Intel). Although PIPE is a standard, it can be interpreted differently by different vendors. In addition, low-power functionality is not very clearly specified. These issues can lead to bugs. Thus, it is very important to understand not only the PIPE interface spec – but the implementation assumptions made by both vendors. 3.0 My Chip Needs USB 2.0 Almost all mobile devices require USB functionality and in addition to the points made above the target for USB is to be low in power and area. Therefore the IP vendor must provide a core that is competitive in area and power without sacrificing performance. Significant area reduction can be achieved through a combination of architectural and implementation optimizations. For example, the PLL/DLL architecture should be targeted to support the 12/24/48 MHz reference frequencies with 480 Mbs output. Also products requiring USB IP tend to follow the aggressive feature size reductions and therefore the PHY should be floor-planned in such a way that as the digital block scales with the smaller process geometries, allowing the overall macro area to shrink -- this is not always the case with analog/mixed-signal designs. Another important consideration is the pin count design – low pin count is a major advantage in terms of enabling the use of lower-cost packaging. The need for fewer pins also reduces the cost of production test, as well as considerably easing SoC integration. With the transition from 90 nm to 65 nm and 45 nm process geometries, yield has assumed a much higher priority. With USB, chip yield is linked to the performance of key specification parameters, such as PLL jitter performance and band-gap variation, as well as being dependent on chip area. A lower chip yield, even by as little as two or three per cent, can cause manufacturing cost increases which may overwhelm any savings that are achieved with a smaller die area. 4.0 Why Should I Ask For A Complete DDR2 Memory Interface Solution? The benefits of high performance DDR2 SDRAM memories require systems-on-chip (SoC) interface designers to approach memory subsystem integration with attention to detail. As data rates have progressively increased from DDR2 400, DDR2 533, DDR2 667, and now DDR2 800, the complexities associated with the timing and signal integrity of the memory interface has become increasing difficult. Migrating from 400 Mbps to 800 Mbps DDR2 requires additional engineering effort. Ideally, this migration was planned for when the 400 Mbps application was first implemented. For DDR signaling, increasing frequency reduces total bit time from 2.5 ns to 1.25 ns. This bit time is then evenly divided into a setup-and-hold budget of 625 ps each. Source synchronous timing depends on the uncertainty of the placement of the DQ data edge relative to the DQS strobe edge. Any skew, jitter or uncertainty component will erode the setup and hold margins. Total timing is composed of three budgets: Transmitter, Interconnect, and Receiver. Nominally, each of these three budgets account for about 33% of the total timing budget. JEDEC managed to scale down the DRAM contributors to the transmitter budget (during Reads) and the receiver budget (during Writes) accordingly with the increase in frequency. Unfortunately, the scaling is not applied proportional to the bit period as data rates increase. For example, the uncertainty of when the DRAM will generate DQS relative to CK is +/- tDQSCK. For DDR2 400, this is +/- 500 ps, or 40% of the 2500 ps bit time. For DDR2 800, this is +/- 350ps, or 56% of the 1250 ps bit time. Assume also, that the system designer planned for this migration and specified the controller, PHY, and I/O cell to meet 800 Mbps timing at the beginning of the project. The remaining timing budget is consumed by the interconnect between the PHY and the DRAM. Three items in particular in the interconnect budget need to be addressed. 4.1 PCB and Package Skew The electrical length differences between the DQS and the DQ of a particular byte must be reduced to meet the now reduced timing budget. Where a 35 ps skew budget may have been adequate for 400 Mbps, less than 20 ps may be required at 800 Mbps. 4.2 Inter-Symbol Interference (ISI) This effect is the overlap of random signal bits at the receiver. ISI is exacerbated by capacitive loading of the net and frequency dependent losses in the channel routing. The impact is to increase the data-dependent jitter at the receiver thresholds and reduce the minimum amplitude of the received signal. Both of these effects can be captured in eye-patterns. When the bit rate doubles, these effects will increase since the signal now has 50% less time to reach the required threshold levels. DDR2 has lower AC thresholds at 800 Mbps that addresses the amplitude issue; however, increased data-dependent jitter will still be a problem. Fortunately, the capacitive loading can be reduced by decreasing the number of ranks of memory of each DQ and DQS, consequently reducing the roll-off of the received signal and reducing ISI. The PCB losses can be reduced by shortening the overall route length, using a lower loss dielectric material and/ or increasing the trace width (watch out for crosstalk). 4.3 SSO pushout During write operations, the DQS is launched 90 degrees out of phase with the eight DQ signals of the byte. When the 8 DQ lines toggle simultaneously, the resulting current draw through the package wire inductance may cause the power rail to collapse, resulting in a delay of the output of the DQ signals. This “push out” will subtract from the available set up time budget. If nothing is done, when the bit rate doubles, the percentage of the budget occupied by SSO push out will also double. To reduce the contribution of SSO, the package wire inductance must be reduced. This can be done by switching from bond wire to flip chip or increasing the number of power/ground pairs in the interface. Other less effective measures would be using double bonds on power and ground, or adding decoupling. Decoupling is most effective when on-die. Decoupling can also be placed close to the die on the pad ring, within the package or surface-mounted to the package. Adding capacitance on the PCB will likely have too much effective inductance to efficiently decouple high-frequency power/ground noise. Any of these solutions increase the cost; therefore, the architect should plan for 800 Mbps operation from the beginning. Planning is the key. When designing for 400 Mbps, anticipate what will be required to get to 800 Mbps and include it on the front end. To summarize there are two options: Customer assembled kit – requires assembly
In today’s deep sub-micron technologies the following effects impact yield and must be eliminated.
6.0 IP Testing & Verification Thorough validation of IP involves two phases: pre-silicon and post-silicon verification. For pre-silicon verification, the following steps may be necessary:
Figure 5 The received eye for PCI Express - On board diagnostics For post-silicon verification, the following steps are necessary:
The first time test vectors/code are tested on real silicon is when the SoC comes back. So how does the test engineer incorporate the PCI Express compliance eye-mask into the ATE without changing the signal? He certainly cannot hang a high speed tester on each pin. The test engineer’s concern is that with the simple pass/fail test of the external loop-back approach he will not know how much margin he has – chips that pass the loop back test that are marginal may fail over time when subjected to real world conditions. Referring again to Figure 5, using voltage and phase margining the link can be reliably tested in a production test environment. These limits are set by simple input and compare vectors that are provided with the IP. A comparison of this approach with the conventional method is outlined in Table 1.Table 1 High speed SerDes testing comparison 8.0 Conclusions This paper reviewed in detail the important considerations for selecting an IP vendor. The key considerations can be summarized as:
Navraj Nandra is Director, Marketing Mixed-Signal IP at Synopsys. Acknowledgements Contributions to this paper were made by Gervais Fong, David Wallace, Joe Guiliano, Bob Lefferts (Synopsys) Boris Litinsky (Juniper Networks)
|
Home | Feedback | Register | Site Map |
All material on this site Copyright © 2017 Design And Reuse S.A. All rights reserved. |