Seminar Report

  • November 2019
  • PDF

This document was uploaded by user and they confirmed that they have the permission to share it. If you are author or own the copyright of this book, please report to us by using this DMCA report form. Report DMCA


Overview

Download & View Seminar Report as PDF for free.

More details

  • Words: 6,095
  • Pages: 28
1

Seminar on: Trends in PC system memory

Introduction

PC technologies continue to evolve as performance demands grow, and system memory technology is no exception. Throughout the history of the PC, memory technology has steadily progressed in capacity and performance to meet the increasing requirements of other PC hardware subsystems and software. In the past, there have been relatively clear industry transitions from one memory technology to its successor. However, today there are multiple choices— PC133, RDRAM, and Double Data Rate (DDR)— and more choices may exist in the future as providers of DRAM system memory accommodate a growing variety of platforms and form factors. These range from small handheld devices to high-end servers, each with different power, space, speed, and capacity requirements for system memory. This seminar focuses on PC system memory issues and trends. It begins by reviewing the role of memory in the system and the key memory parameters that affect system performance. This report then presents the basics of memory technology and today's alternatives. Finally says about the key upcoming transitions in the DDR and Rambus memory technologies.

Dept of E&CE PDACE Gulbarga

2

Seminar on: Trends in PC system memory

Role of Memory in the System

The primary role of memory is to store code and data for the processor. Although caching and other processor architecture features have reduced its dependency on memory performance, the processor still requires most of the memory bandwidth. Figure 1 shows the major consumers of memory bandwidth: the processor, graphics subsystem, PCI devices (such as high-speed communications devices), and hard drives. Other lower bandwidth interfaces such as the USB and parallel ports must also be accommodated. The memory hub provides an interface to system memory for all of the high bandwidth devices. The I/O hub schedules requests from other devices into the memory hub. Memory plays a key role in the efficient operation of I/O devices such as graphics adapters and disk drives. In a typical system, most data transfers move through system memory. For example, when transferring a file from the network to a local disk, the PCI host adapter transfers data from the network to memory. This is commonly referred to as direct memory access (DMA), as opposed to programmed I/O (PIO), in which the processor is directly involved in all data transfers. The processor, after performing any required formatting operations, initiates a transfer from memory to local disk storage. Once initiated, the data is transferred directly from memory to disk without any further processor involvement. In summary, the system memory functions as the primary storage component for processor code and data, and as a centralized transfer point for most data movement in today's systems

Dept of E&CE PDACE Gulbarga

3

Seminar on: Trends in PC system memory

Performance Factors

Memory parameters that impact system performance are capacity, bandwidth, and latency.

Capacity How does capacity impact system performance? The first step in answering this question is to describe the

memory hierarchy. Table 1 shows the capacities and speeds

of various storage mechanisms found in a typical mainstream desktop computer in service today.

These storage mechanisms range from the very fast, but low-capacity, Level 1 (L1) cache memory to the much slower, but higher-capacity, disk drive. Ideally, a computer would use the fastest available storage mechanisms—in this case L1 cache—for all data. However, the laws of physics (which dictate that higher capacity storage mechanisms are slower) and cost considerations prevent this. Instead, PCs use a mechanism called “virtual memory,” which makes use of the L1 and L2 cache, main system memory, and the hard drive. The virtual memory mechanism allows a programmer to use more memory than is physically available in the system, and to keep the most frequently and recently used data in the fastest storage. When more memory is needed than is available in system memory, some data or code must be stored on disk. When the processor accesses data not available in memory, information that has not been accessed recently is saved to the hard drive. The system then uses the vacated memory space to complete the processor's request. However, disk access is comparatively slow and system performance

Dept of E&CE PDACE Gulbarga

4

Seminar on: Trends in PC system memory

is significantly impacted if the processor must frequently wait for disk access. Adding system memory reduces this probability. The amount of capacity required to reduce disk activity to an acceptable level depends on the operating system and the type and number of active applications, including background tasks. Semiconductor technology has provided capacity improvements consistent with Moore's law—greater than a 1.4x compound annual growth rate—and the outlook is for similar increases over the next few years. These increases have exceeded the requirements of mainstream desktop PCs and, as a result, the number of memory slots is being reduced from three to two in many of today's client platforms. However, servers and high-end workstations continue to take advantage of the capacity increases.

Typical Capacity

Current Speed

Level 1 cache

32KB

4ns

Level 2 cache

256KB

20ns

System memory

128MB

100ns

Hard drive 4GB* 8ms *Although many disk drives contain much more storage capacity, 4 GB is the memory-addressing limitation in most PC systems.

Table 1. Memory Storage Mechanisms

Dept of E&CE PDACE Gulbarga

5

Seminar on: Trends in PC system memory

Bandwidth Memory bandwidth is a measure of the rate at which data can be transferred to and from memory, typically expressed in megabytes per second (MB/sec). Peak bandwidth is the theoretical maximum transfer rate between any device and memory. In practice, peak bandwidth is reduced by interference from other devices and by the “lead-off” time required for a device to receive the first bit of data after initiating a memory request. There should be adequate memory bandwidth to support the actual data rates of the highest-speed devices and to provide enough headroom to prevent significant interference between devices. In many systems, memory and I/O hubs are designed to accommodate peak requirements by buffering transfers and scheduling conflicting memory requests. Table 2 shows the data rates of various system components over the last 4 years. Although the need for memory bandwidth is not directly proportional to these data rates, the upward trend is obvious. Memory systems have done a fairly good job of keeping up

Figure1.Major Consumers of Memory Bandwidth

Dept of E&CE PDACE Gulbarga

6

Seminar on: Trends in PC system memory

with system requirements over this period of time, moving from 533 MB/sec to 2133 MB/sec. Dual-memory interfaces using Rambus or DDR memory boost bandwidth to 3200 MB/sec.

Latency Latency is a measure of the delay from the data request until the data is returned. It is a function of peak bandwidth, lead-off time, and interference between devices. In general, processors are more sensitive to latency than bandwidth because they work with smaller blocks of data and can waste a significant number of clocks waiting for critical data. In contrast, I/O data transfers are relatively long, and bandwidth is a more important consideration than latency. Data transfers moving to and from system memory must pass through the memory hub and, in many cases, the I/O hub. These components are collectively referred to as the chip set or core logic, and are major contributors to the latency from a device to memory. They can restrict or exploit system memory capabilities and must be designed to provide a balance of bandwidth, buffering, and scheduling of data transfers for optimum memory performance. Over the last 4 years, memory bandwidth has kept up with system needs, but latency improvements have lagged. Current Rambus and DDR technologies double the memory bandwidth over100-MHz SDRAM, but do not reduce latency. Ideally, latency should be reduced in proportion to processor clock rate increases.

CAS Latency The RAM module is an organized collection of integrated chips (ICs). Each module is controlled by the memory controller, which handles the signals going from the CPU to

Dept of E&CE PDACE Gulbarga

7

Seminar on: Trends in PC system memory

the RAM. There are two signals--- the Row Access Strobe (RAS) and the Column Access Strobe (CAS). Each memory chip is divided into rows and columns, making it look like a cell matrix. Each cell ha s a row address and a column address. When the CPU sends a signal to the memory controller, it first accesses the row by putting an address on the memory’s address pins, and activation the RAS signal. Then it waits a few clock cycles--this is the RAS-to-CAS Delay. It then puts the column address on the address pins, and activates the CAS signal. Then there’s a wait for another few clock cycles, which is the CAS delay. Finally, the data appears on the pins of the RAM. The CAS delay is called CAS Latency. Lower CAS latency provides faster data access. CAS-2 is where the CPU waits for 2 clock cycles; CAS-3 is where it waits for 3 clock cycles, and so on. However, CAS-2 affords a performance gain over CAS-3 only while playing games or over clocking the CPU. The CAS latency is usually 2.5ns for DDR RAM.

New processor architecture features—improved caches, more support for out-oforder execution, and prefetch instructions for applications that are sensitive to processor latency—have helped to offset this lag in latency improvements.

MEMORY TECHNOLOGY BASICS

While the complex world of main system Memory technology can be regarded from many angles, it is beneficial to begin with a look at two large-scale perspectives, ROM (Read-Only Memory) and RAM (Random Access Memory). An important distinction Dept of E&CE PDACE Gulbarga

8

Seminar on: Trends in PC system memory

between the two is that RAM is volatile memory, that is, any data existing on the memory is erased when the system shuts down. By comparison, ROM memory is referred to as nonvolatile, meaning its data content remains intact, irrespective of shutdown and boot activity. Note that RAM and ROM are commonly used terms, and are used here, although both types actually allow random access.

ROM As the name implies, ROM memory can only be read in operation, preventing the rewriting of contents as part of its normal function. Basic ROM stores critical information in computers and other digital devices, information whose integrity is vital to system operation and is unlikely to change. Other commonly used forms of ROM are:

PROM (Programmable ROM) Programmable ROM is a blank memory chip open for one-time only recording of information that is then stored permanently.

EPROM (Erasable Programmable ROM) This is PROM equipped with a special sensor to which application of a UV light erases stored information, allowing subsequent rewriting.

EEPROM (Electrically Erasable Programmable ROM) This type of ROM is rewrite able by way of software. It is used in flash BIOS, in which the software allows users to upgrade the stored BIOS information (flashing). Dept of E&CE PDACE Gulbarga

9

Seminar on: Trends in PC system memory

RAM

The Random component of RAM’s name actually refers to this type of memory’s ability to access any single byte of information without contacting or affecting neighboring bytes. The main difference between a ROM and a RAM is that a RAM

is that a RAM is written under normal operation, while a ROM is programmed outside the computer and is only normally read. RAM plays a major role in system operations and specifically performance. Essentially, the more complex a program is, the more its execution will benefit from the present of both ample and efficient RAM access. RAM takes two forms, SRAM (Static RAM), and DRAM (Dynamic RAM). A detailed discussion of both follows.

SRAM (Static RAM) Static RAM memory devices retain data for as long as DC power is applied. Because no special action (except power) is required to retain stored data, these devices are called static memory. They re also called volatile memory because they will not retain data without power. The SRAM stores temporary data and is used when the size of the read/write memory is relatively small. Static RAM provides significant advantages for performance, in that it holds data without needing to be refreshed constantly. This allows access times as low as 10 ns as well as a shorter cycle time. A disadvantage is Static RAM’s high cost to produce, limiting most of its practical applications to memory caching functions. There are three types of SRAM, Async RAM, Sync RAM, and Pipeline Burst RAM.

Dept of E&CE PDACE Gulbarga

10

Seminar on: Trends in PC system memory

Async SRAM The designation Async is short for Asynchronous, meaning here that the SRAM functions with no dependence on the system clock. Async SRAM is an older type of SRAM, normally limited to L2 cache only.

Sync SRAM Obviously, this is a synchronized form of SRAM, matched to the system clock’s operations. The marked speed advantage of this feature also dictates a higher cost.

Pipeline Burst SRAM Most common, this type of SRAM directs larger data packets of data (requests) to the memory (pipelining) all at the same time, allowing a much quicker reaction on the part of the RAM, thus increasing access speed. This type of SRAM accommodates bus speeds in excess of 66MHz, hence its popularity.

DRAM (Dynamic RAM)

DRAM is essentially the same as SRAM, except that it retains data for only 2 or 4 ms on an integrated capacitor. After 2 or 4 ms, the contents of the DRAM must be completely rewritten (refreshed) because the capacitor, which store a logic 1 or logic 0, lose their charges. In order to refresh a DRAM, the contents of a section of the memory must periodically be read or written. Any read or write automatically refreshes an entire Dept of E&CE PDACE Gulbarga

11

Seminar on: Trends in PC system memory

section of the DRAM. The number of bits refreshed depends on the size of the memory component and its internal organization. Refresh cycles are accomplished by doing a read, a write, or a special refresh cycle that doesn’t read or write data. The refresh cycle is totally internal to the DRAM and is accomplished while other memory components in the system operate. This type of refresh is called either hidden refresh, transparent refresh, or sometimes cycle stealing. For example, it takes the 8086/8088, running at a 5 MHz clock rate, 800 ns to do a read or write. Because the DRAM must have a refresh cycle every 15.6 µs, this means that for every 19 memory reads or writes, the memory system must run a refresh cycle or memory data will be lost. This represents a loss of 5% of the computer’s time, a small price to pay for the savings represented by using the dynamic RAM. While the raw speed consideration renders it a secondary choice to SRAM, its superior cost advantages make it a vastly more popular choice for system memory.

In the last few years, a tremendous upsurge has occurred in the evolution of DRAM technology. Where originally, nearly all PC system memory was confined to FPM (Fast Page Mode) DRAM. Rapid growth in both CPU and motherboard bus speeds, however, spurred development of improved DRAM methods to provide performance equal to faster and faster system capabilities. Numerous options are now, and have been, available, some popular, some less so. They fall into two main categories, Asynchronous and Synchronous, with the latter taking the lead in recent years.

Dept of E&CE PDACE Gulbarga

12

Seminar on: Trends in PC system memory

Asynchronous DRAM models FPM (Fast Page Mode) An improvement on its Page Mode Access predecessor, FPM quickly became the most widely utilized access mode for DRAM, almost universally installed for a time, and still widely supported. Its bypassing of power-hungry sense and restore current is a major benefit. With speeds originally limited to around 120 ns, and later improvements to as low as 60, FPM was still unable to keep up with the soon-to-be ubiquitous 66 MHz system bus. Presently, FPM is rarely used, having been usurped by several superior technologies. Ironically, its rarity of deployment has resulted in it often being more costly than other forms.

EDO (Extended Data Output) EDO, or Hyper Page mode, as it is sometimes called, was the last significant improvement to be made available to DRAM customers before the development of synchronous DRAM interface technology. Its advantage over FPM lies in its ability, by not turning off the output buffers, to allow an access operation to commence prior to the completion of the previous operation. This provides a performance improvement over FPM DRAM, with no increase in silicon usage, and hence, package size. Improvements in the 30-40% range are seen in the implementation of EDO DRAM. Additionally, it supports memory bus speeds up to 83 MHz, sacrificing little or no performance capability. Given proper chip support, even 100 MHz bus speeds are accessible, although with much lower results than the newer synchronous forms. Users whose bus speed requirements are no higher than 83 MHz should see no clear advantage in upgrading from

Dept of E&CE PDACE Gulbarga

13

Seminar on: Trends in PC system memory

EDO to synchronous embodiments. Nonetheless, the rise in widely available higher chipset capabilities has, for the most part, left EDO in the past alongside FPM DRAM. BEDO (Burst Extended Data Output) BEDO was a largely unsuccessful attempt to further improve on EDO performance. It utilized a burst mode combined with dual bank architecture to beat EDO capabilities. Simply put, the BEDO advantage lay in its ability to prepare 3 subsequent addresses internally following an initial address input. This technique allowed it to overcome time delays resulting from the input of each new address. The timing of its development was, however, poor, in that by the time it became a possible alternative, most large manufacturers had devoted most of their development energies towards SDRAM and related advancement. Consequently, industry wisdom, led by Intel, negated the viability of BEDO as an acceptable solution, irrespective of its possible merits.

SDRAM (Synchronous DRAM) Models As parallel advances made it clear that the future belonged to bus speeds in excess of 66 MHz, it became incumbent upon developers to overcome latency problems inherent in existing DRAM forms. The development of workable solutions in the area of synchronous operations was the result, accomplishing not only the ability to embrace higher bus speeds, but other advantages as well.

DDR SDRAM (Double Data Rate Synchronous) Presently accepted as the de facto industry standard for system memory, DDR This section provides a more detailed view of DDR SDRAM (and in some cases, simply DRAM), its advantages to users, especially those in data-intensive environments. About Unbuffered and Registered modules Generally, in the present marketplace, different

Dept of E&CE PDACE Gulbarga

14

Seminar on: Trends in PC system memory

implementations of DIMM memory are available in two major types, Unbuffered and Registered.

About Speed Ratings

It is worthwhile to be aware of the conventions used for rating the relative speed of SDRAM modules. While ratings for asynchronous DRAM are listed in nanoseconds, an additional MHz rating is applied to SDRAM, since its synchronous nature requires that it be compatible with system bus speed. Unfortunately, exactly matching exact SDRAM ratings with bus speeds, that is, for example, to say that a 100 MHz/10 ns-rated SDRAM speed will operate properly with a 100 MHz system speed, is not practically correct. This stems from the fact that the SDRAM MHz rating refers to optimum conditions in the system, a scenario that rarely, if ever, occurs in real-world operations. More operable would be to “over-match” the SDRAM rating to compensate. Thus, a much more realistic qualification is undertaken by manufacturers, in which 100MHz SDRAM is intended for operation with a system speed of, say, 83 MHz.

To make the entire process easier, Intel implemented a Speed Rating system for SDRAM qualification, its now universally exercised “PC” rating. The qualification, intended as an aid to both manufacturers and users, takes into account the above noted considerations in addition to other internal timing characteristics. As much as any acrossthe-board rating system, this scale comes very close to assuring the compatibility of SDRAM with its host system speed. According to the Intel system, a PC100-compliant

Dept of E&CE PDACE Gulbarga

15

Seminar on: Trends in PC system memory

SDRAM module will perform well with a 100 MHz system. Likewise its PC66 and PC133 ratings. The ratings also guarantee compatibility with Intel protocols, such as PC100 with Intel BX motherboard.

Current Memory Technologies

Over time, PC system memory has evolved to increase capacity and bandwidth, and reduce latency. There has been progress at the component and interface level. While the basic component core architecture has not changed significantly, the interfaces to the component cores have and continue to evolve. Speed and capacity advances in the core have been driven by advances in semiconductor technology. Interfaces have also benefited from semiconductor advances, but with notable architectural changes. PC memory interfaces have transitioned through conventional DRAM, page-mode DRAM (or fast page-mode), and extended data out (EDO) DRAM. Today, three main memory interfaces are used: • Synchronous DRAM (SDRAM) • DDR SDRAM • Rambus

SDRAM SDRAM is a 64-bit-wide interface (or 72 bits in implementations that include error correction capability). The interface clock rate began at 66 MHz, evolved to 100 MHz, and is now capable of operating at 133 MHz in system memory implementations. (100- and 133-MHz implementations

Dept of E&CE PDACE Gulbarga

16

Seminar on: Trends in PC system memory are often referred to as PC100 and PC133.) The bandwidth in MB/sec is equal to the clock multiplied by 8. (A 64-bit wide interface transfers 8 bytes per clock.) This yields 1066 MB/sec bandwidth for current 133-MHz SDRAM. Both latency and bandwidth have improved in proportion to the clock speed increases. A single system memory semiconductor component may provide 4, 8, or 16 bits on each data transfer. Regardless of the bits per transfer, each component contains the same amount of storage (in bits) for a given DRAM component generation. Multiple components are mounted on a memory module (or DIMM). The system's memory hub interfaces with one or more of these memory modules. An SDRAM module combines the components to provide 64 bits on each data transfer. Desktop and portable systems use 8- and 16-bit component interfaces to provide 64-bit interfaces in lower capacities. For example, a typical desktop or portable SDRAM module might consist of four 16-bit components, each with a 128megabit (Mb) capacity, for a total of 64 MB of memory. In contrast, 4- and 8-bit components are used for the higher capacities needed in servers. An SDRAM module used in a server might contain 16 4-bit components, each with 128-Mb capacity, for a total capacity of 256 MB of memory. Component capacities have increased over the life of SDRAM in proportion to the advances in the semiconductor industry.

Current SDRAM Implementations 133-MHz SDRAM is used in today's value and mainstream computers. Its performance is adequate for the application mix in the target markets, but it is expected to be replaced rapidly as the memory performance demands of new system components, operating systems, and applications increase. Dual SDRAM interfaces are used in servers for added bandwidth and capacity. In system memory implementations, SDRAM frequencies are not expected to increase beyond 133 MHz. Instead, DDR will be used beyond the 133-MHz performance point. (SDRAM frequencies beyond 133 MHz are available, but are used in specialized applications such as graphics controllers.)

Dept of E&CE PDACE Gulbarga

17

Seminar on: Trends in PC system memory

DDR SDRAM

DDR SDRAM (see Table2-DDR Naming Conventions) is similar to standard SDRAM, but adds strobes that operate at the same frequency as the clock. These strobes travel with the data from the memory hub to the memory components on writes and to the memory hub from the memory components on reads. Data is transferred on both the rising and falling edges of the strobe. This architectural change, coupled with lower signal levels and advances in semiconductor processes, doubles the data rates and the bandwidth. However, DDR does not reduce latency. In general, the PC1600 implementations have greater latency than PC133. Despite this, PC1600 provides a performance advantage for most systems and applications in which it is implemented. Like SDRAM, PC system memory implementations of DDR have 64-bit data paths (or 72 bits for error correction capability). Bandwidths for DDR are 1600 MB/sec for PC1600 and 2133 MB/sec for PC2100. Component capacities for volume production DDR began at 128 Mb and will be available at least through the 1-Gb generation.

One of the goals of DDR was a smooth technology transition. Many memory hubs are designed to operate with either SDRAM or DDR SDRAM modules. However, practical space and cost considerations, driven in part by different module connectors, has prevented any practical system-level application of this capability. DDR has minor incremental system- and component level cost penalties over standard SDRAM, but these costs will decline with volume and advances in semiconductor processes.

Dept of E&CE PDACE Gulbarga

18

Seminar on: Trends in PC system memory

Original Name

Name Based On Clock Rate

Name based on Band-width

DDR SDRAM (100 MHz)

DDR200

PC1600

DDR266

PC2100

DDR SDRAM (133 MHz)

Table2- DDR naming conventions Some topological differences between DDR SDRAM and SDRAM DIMM modules are as shown below in the table:

SDRAM

DDR

Profile

Voltage

3.3

2.5

Pin Count

168

184

2

1

Notches

Table 3: topological differences between DDR SDRAM and SDRAM DIMM modules

Dept of E&CE PDACE Gulbarga

19

Seminar on: Trends in PC system memory

DDR modules, like their SDRAM predecessors, arrive in their DIMM modules. Although motherboards designed to implement DDR are similar to those that use SDRAM, they are not backward compatible with motherboards that support SDRAM. You cannot use DDR in earlier SDRAM based motherboards, nor can you use SDRAM on motherboards that are designed for DDR.

Rambus

SDRAM and DDR SDRAM share many architectural and signaling features. Both use a parallel data bus, mainly available in component widths of x8 or x16, both have a single addressing command bus that must be shared to transmit row and column addresses. DDR SDRAM increases data bandwidth over conventional SDRAM by transmitting data on both edges of the synchronous clock signal, thus in theory “doubling” the data rate of the memory. It is important to note that DDR SDRAM does not double the address command bandwidth of the system by using both edges of the clock on the command bus, a factor that ultimately limits the real world performance gain from using DDR signaling on the data bus. RDRAM memory, which pioneered double data rate technology for DRAM in 1990, takes a different approach. It combines a conventional DRAM core with a high-speed serial interface called the RDRAM Channel.

The Rambus architecture is significantly different from previous system memory interfaces. RDRAM memory uses separate buses to transmit row and column address

Dept of E&CE PDACE Gulbarga

20

Seminar on: Trends in PC system memory

information on both edges of the clock. This enables higher efficiency by allowing the issue of a column packet while a row packet is simultaneously being issued to another device, a feature not present in SDRAM or DDR. Both commands and data are communicated on packet buses from the memory hub to the modules. There are two command buses—RAS bus and CAS bus—and one data bus. Eight transfers on both edges of a 400-MHz clock are used for all packets. The separate command and data buses, along with identical packet lengths, allow memory hubs to improve the scheduling of memory transfers to achieve higher memory data bus utilization. The result is higher actual bandwidth. One of the benefits of this architecture is that it allows the banks of the individual devices to be additive, pooling them to decrease the probability of bank conflict and thus minimize bank conflict effects. RDRAM memory is capable of servicing a memory request every 10ns. This allows RDRAM to be implemented in a capacity that better matches the needs of a particular system. However, each memory component must be capable of providing the full bandwidth needed by the system. This requirement, coupled with other internal architecture specifications, have resulted in a cost premium for RDRAM components.

Unlike SDRAM, motherboards that support Rambus memory require that all memory slots in the motherboard must be populated. Not with memory modules necessarily, but with a combination of memory modules and Continuity Modules, more commonly referred to as CRIMMs. As an example, on SDRAM based motherboards, you could have 4 memory slots and fill only one with a memory module, and the system would operate. On Rambus based motherboards, all blank or unfilled memory sockets must be populated with either a memory module (in this case a RIMM™ Rambus Inline Memory Modules) Dept of E&CE PDACE Gulbarga

21

Seminar on: Trends in PC system memory

or a Continuity Module, (CRIMM) to complete the memory path to the bus. Rambus memory is available in either ECC or non-ECC format

Rambus memory is a proprietary technology of Rambus Inc., therefore manufacturers that want to produce it are required to pay a royalty to Rambus Inc. DDR designs, on the other hand, are open architecture. Since Rambus is an entirely new design, there are other cost factors to be added in, such as an entirely new module manufacturing and testing process. The results of the bandwidth study for typical memories of different types, operating frequencies and core latency (tRAC) timings can be referenced in the Table 4 below.

Table 4: Effective Bandwidth of 256Mbyte Memory Modules

Upcoming Memory Interface Transitions

There are key upcoming transitions in the DDR and Rambus memory technologies, as well as an emerging new technology for example - Advanced DRAM Technology (ADT).

Dept of E&CE PDACE Gulbarga

22

Seminar on: Trends in PC system memory

ADT ADT is an emerging memory interface technology targeted for introduction slightly after DDR. Details of this interface have not been publicly disclosed; however, all major DRAM component manufacturers are involved in defining both the DDR and ADT interfaces. It is expected that ADT will initially target desktop applications. There is a strong possibility that only one of these standards will be adopted by the PC industry, and the emerging standard will incorporate features from both technologies.

Rambus Rambus has plans for a new high-speed memory interface called Yellowstone. This next-generation interface is intended to provide 3.2 Gb/sec per differential signal pair between DRAM components and the memory interface. A 200-millivolt differential signal is planned to support this data rate. The Yellowstone is expected to be within mainstream PC memory cost budgets.

Multiple Memory Interfaces Although memory technology has kept pace with increasing capacity requirements, system memory bandwidth is not keeping pace with the demands of the rest of the system. The ratio of bandwidth to capacity for a given memory module implementation is falling over time. A dual interface helps to alleviate the resulting performance issues. Multiple memory interfaces in high-end systems are not new, but their use in mainstream desktop PCs is expected to begin soon.

Dept of E&CE PDACE Gulbarga

23

Seminar on: Trends in PC system memory

Future RAM Technologies

Improvements in the memory interface performance require advances in semiconductor technology and the techniques used to maintain signal integrity and timing across the interface. There are challenges on the component side to providing capacity and speed increases at the pace of Moore's law; it is difficult to make components bigger and faster at the same time. Containing cost, reducing size, controlling emissions, and maintaining the reliability of the system memory interface becomes more difficult as performance increases. Future interface transitions must address all of these issues.

MRAM Manufacturers such as IBM, Motorola and Infineon are specifically involved in the research of a new form of solid-state memory called Magnetoresistive RAM (MRAM). This type of RAM uses the principle of alignment of magnetic particles as the basis of storing zeroes and ones in a digital system. An inherent advantage with MRAM is that it since it’s based on the principle of magnetism: it retains whatever information is contained in it even after power is switched off. Additionally, these elements do not need an electric charge to retain information, hence power consumption of these devices is minimal compared to that of DRAM. Finally with the absence of the delays associated with transferring electricity between the storage elements for retaining the information, MRAM is expected to be up to 30 times faster than DRAM. MRAM is thus poised to give rise to a new breed of instantly-on computing devises. Dept of E&CE PDACE Gulbarga

24

Seminar on: Trends in PC system memory

The technology is already up and running in some high-tech labs around the world. But only after the fabrication process for this memory becomes commercially available, it will make its way into mainstream home PCs and handhelds.

IRAM One approach to reduce the problem of processor-memory communication delay is Intelligent RAM (IRAM)—integrating processor and memory on a single die. Of course, for the solution to be complete, memory latency needs to be reduced, too. Currently, IRAM products are no commercially available: however, several university and company research labs are working on IRAM architectures.

NRAM Nano RAM(NRAM) being developed by a company called Nantero, based in Woburn, Massachusetts, may be out ay any time now. MRAM is non-volatile, like MRAM, so your computer can boot up within a second. The chips would run at 2 GHz, and last virtually forever. The concept behind NRAM is simple: combine carbon nanotubes—a molecule of carbon atoms connected like a tube and sealed at both ends with traditional semiconductor techniques. Ones and zeroes revisited: A nanotube is bent using electrostatic forces so that the top touches the bottom. Bent nanotubes can be unbent too. Bent means a 1, and unbent means a 0. And billions of nanotubes can be squeezed into a four-inch wafer.

Dept of E&CE PDACE Gulbarga

25

Seminar on: Trends in PC system memory

Conclusion In today’s world of ever growing need of faster and high-end computers the need for faster and better memory technology is never ending. Consider, for example, the Intel’s newest Pentium4 processors have front-side bus architectures operating at data speeds of 533MHz and 800MHz. This translates to a peak data bandwidth of 4.2GB/s (533MHz x 8 Bytes) and 6.4GB/s (800MHz x 8 Bytes) respectively! Considering the fact that the computer industry has come so far from the FPT to the RDRAM it can now be said that a day will come when every user will be able to start his/her PC like a bulb with out any delay (no time requirement for booting)! Giving rise to a new breed of instantly on computing devises.

Dept of E&CE PDACE Gulbarga

26

Seminar on: Trends in PC system memory

Reference: 1. www.pcguide.com/ref/ram/tech.htm 2. www.rambus.com 3. www.dell.com/r&d 4. www.dewassoc.com 5. www.kingston.com/newtech/MKF_520DDR 6. digit Your Technology Navigator (www.thinkdigit.com) (Feb 2003, June 2003 July 2003, March 2004) 7. www.jedec.org 8. THE INTEL MICROPROCESSORS 8086/8088, 80186/80188, 80286, 80386, 80486, PENTIUM, AND PENTIUM PRO PROCESSOR Architecture, Programming, and Interface By: Barray B. Brey (Forth Edition, Page No: 353)

Dept of E&CE PDACE Gulbarga

27

Seminar on: Trends in PC system memory

APPENDIX –A - Motherboards at a Glance

Dept of E&CE PDACE Gulbarga

28

Seminar on: Trends in PC system memory

INDEX A

L

Latency, 6

About Speed Ratings, 14 ADT, 22

APPENDIX –A - Motherboards at a Glance, 27 Async SRAM, 10 Asynchronous DRAM models, 12

M

MEMORY TECHNOLOGY BASICS, 7 MRAM, 23 Multiple Memory Interfaces, 22

B N

Bandwidth, 5 NRAM, 24 C

P

Capacity, 3 CAS Latency, 6

Performance Factors, 3

Conclusion, 25

Pipeline Burst SRAM, 10 PROM (Programmable ROM), 8

Current Memory Technologies, 15

Current SDRAM Implementations, 16 R D

RAM, 6, 7, 9, 10, 11, 23, 24

DDR SDRAM, 17

Rambus, 19, 22

DDR SDRAM (Double Data Rate Synchronous), 13

Reference, 26

DRAM (Dynamic RAM), 10

ROM, 8

Role of Memory in the System, 2

E

S

EDO (Extended Data Output), 12 EEPROM (Electrically Erasable Programmable ROM), 8 EPROM (Erasable Programmable ROM), 8

SDRAM, 15 SDRAM (Synchronous DRAM) Models, 13 SRAM, 9 Sync SRAM, 10

F

T

Figure1.Major Consumers of Memory Bandwidth, 5 FPM (Fast Page Mode), 12

Table 1. Memory Storage Mechanisms, 4

Future RAM Technologies, 23 I

Table 3: topological differences between DDR SDRAM and SDRAM DIMM modules, 18 Table 4: Effective Bandwidth of 256Mbyte Memory Modules, 21

Table2- DDR naming conventions, 18 U

Introduction, 1 IRAM, 24

Upcoming Memory Interface Transitions, 21 Dept of E&CE PDACE Gulbarga

Related Documents

Seminar Report
May 2020 6
Seminar Report
November 2019 16
Seminar Report
November 2019 19
Seminar Report
May 2020 11
Seminar Report
June 2020 9
Seminar Report
November 2019 18