Beruflich Dokumente
Kultur Dokumente
1. INTRODUCTION
In the past years, computers are made with single core processors. They
used DDR3, DDR2, DDR1, RDRAM etc. as the primary memory for computing.
For better performance and high speed of computing, the vendors introduced
multi-core processors. Multi-core processors are processors with more than one
CPU. For example, Intels Core 2 Duo is a dual core processor, i3,i5 and i7 are
quad core processors. But the limitation of these processors is, they still uses
DDR3 or DDR2 RAM. The performance of these systems is limited by memory
system bandwidth.
In this situation, Micron inc. introduced a new memory technology called
Hybrid Memory Cube (HMC) which have entirely new architecture in which
DRAM layers are arranged in a three dimensional architecture that improves
latency, band width, power consumption and density.
2.1.1 SRAM
SRAM chip uses bi-stable latches or flip-flop to store each bit. As
computer processes data it changes its states from 1 or 0 and thus refreshing is not
required so SRAM are very fast. SRAM is the nearest of memory or cache in
processor. Also it is very expensive, as it requires much larger density. Four to six
transistors are required to store a bit using SRAM.
2.1.2 DRAM
DRAM (Dynamic Random Access Memory) is the main memory used for
all desktop and larger computers. Each elementary DRAM cell is made up of a
single MOS transistor and a storage capacitor. So, high package density and
cheaper than SRAM . Each storage cell contains one bit of information. This
charge, however, leaks off the capacitor due to the sub-threshold current of the
cell transistor. Therefore, the charge must be refreshed several times each second.
in
the
mid-1990s
as
replacement
developed
for
then-
prevalent DIMM SDRAM memory architecture. It has higher speed than SDRAM
memory architecture since data is transferred on both the rising and falling edges
of the clock signal. Compared to other contemporary standards, Rambus shows a
significant increase in latency, heat output, manufacturing complexity, and cost.
Because of the way Rambus designed RDRAM, RDRAM's die size is inherently
larger than similar SDRAM chips. RDRAM's die size is larger because it is
required to house the added interface and results in a 10-20 percent price premium
at 16-megabit densities and adds about a 5 percent penalty at 64MB.
2.1.7 DDR1
Dept. of ECE, College of Engg., Poonjar
2.1.8 DDR2
DDR2 is the next generation of memory developed after DDR. DDR2
increased the data transfer rate referred to as bandwidth by increasing the
operational frequency to match the high FSB frequencies and by doubling the prefetch buffer data rate. There will be more about the memory pre-fetch buffer data
rate later in this section. DDR2 is a 240 pin DIMM design that operates at 1.8
volts. The lower voltage counters the heat effect of the higher frequency data
transfer. DRR operates at 2.5 volts and is a 188 pin DIMM design. DDR2 uses a
different motherboard socket than DDR, and is not compatible with motherboards
designed for DDR. The DDR2 DIMM key will not align with DDR DIMM key. If
the DDR2 is forced into the DDR socket, it will damage the socket and the
memory will be exposed to a high voltage level. Also be aware the DDR is 188
pin DIMM design and DDR2 is a 240 pin DIMM design. The bandwidth of
DDR2 is 5.34 GBps.
2.1.9 DDR3
Dept. of ECE, College of Engg., Poonjar
2.1.10 DDR4
Dept. of ECE, College of Engg., Poonjar
3.5 Scalability
To increase the bandwidth of memory, more number of data lines are
needed. It does not possible to a large extent in currently using RAM modules.
So, scalability of memory is not possible.
10
5. ARCHITECTURE OF HMC
The HMC DRAM is a 68mm2, 50nm, 1Gb die segmented into multiple
autonomous partitions. Each partition includes two independent memory banks
for a total of 128 banks per HMC. Each partition supports a closed page policy
and full cache line transfers, 32 to 256 bytes. Memory vaults are vertical stacks of
DRAM partitions (Fig.6).
11
12
13
14
15
16
17
generator
generates
commands
for
the
target
The requested location of the selected bank is written by data in the row
buffer (memory write).
18
8. TSV TECHNOLOGY
TSV is an important developing technology that utilizes short, vertical
electrical connections or vias that pass through a silicon wafer in order to
establish an electrical connection from the active side to the backside of the die,
thus providing the shortest interconnect path and creating an avenue for the
ultimate in 3D integration. TSV technology offers greater space efficiencies and
higher interconnects densities than wire bonding and flips chip stacking. When
combined with micro bump bonding and advanced flip chip technology, TSV
technology enables a higher level of functional integration and performance in a
smaller form factor.
While two die could be combined using conventional wire bonding
techniques, the inductive losses would reduce the speed of data exchange, in turn
eroding the performance benefit. TSV addresses the data exchange issues of wire
bonding, and offers several other attractive advantages. For example, TSV allows
for shorter interconnects between the die, reducing power consumption caused by
long horizontal wiring, and eliminating the space and power wasted by buffers
(repeaters that propel a signal through a lengthy circuit). TSV also reduces
electrical parasitic in the circuit (i.e., unintended electrical effects), increasing
device switching speeds. Moreover, TSV accommodates much higher
input/output density than wire bonding, which consumes much more space.
19
8.1.2 Design
As TSV is adopted in increasingly complex applications that combine
different types of chips, design guidelines and software must keep pace to address
a variety of issues. With several thousand interconnections between die, chip
architecture and layout must undergo fundamental changes. Designers for each
chip type used in the integration scheme will have to leverage the same master
layout to line up connection points between the chips. At the same time, designers
will need to consider possible heat generation issues. Stacked chips may overheat
if some thermal management mechanism is not included in the design. Hot spots
and temperature gradients strongly affect reliability. Conventional twodimensional (2D) thermal management techniques will not be adequate for 3D
problems; more sophisticated solutions will be required.
20
8.1.3 Manufacturing
For greatest cost-effectiveness, the full manufacturing sequence requires
seamless integration and optimization between traditional steps in wafer
processing and back-end packaging. The entire flow must be optimized to deliver
the greatest performance (yield, reliability) for the highest productivity (cost). In
TSV applications, the process sequence differs, depending on whether the via first
or via last approach is used (as addressed below). In either approach, the TSV
stack undergoes bonding, thinning, wafer processing on bonded/thinned wafers,
and subsequent de-bonding. Wafers are typically bonded to carriers (glass or
dummy silicon) and thinned down to a thickness ranging from 30 to 125m. This
introduces new manufacturing challenges, including thermal budget control. To
preserve the adhesive integrity of the bonding material once wafers are bonded,
processing temperatures cannot exceed 200C. Another manufacturing challenge
is the need for new automatic test capabilities compatible with 3D integration to
ensure electrical functionality of the finished devices. Chip stacking can be die-todie (chip-to-chip), die-to-wafer or two whole wafers stacked together. For early
mainstream applications, die-to-wafer integration will be most common,
involving the stacking of a single Known Good Die (KGD) onto a KGD on a
wafer. Wafer-to-wafer integration will become more common only as consistently
high yielding wafers are produced.
21
22
23
8.3.1 Vias-Last
Dept. of ECE, College of Engg., Poonjar
24
In via-last technology, vias are formed after the die have been fabricated,
i.e, after both the front end of line (FEOL) and BEOL processes have been
completed. While direct oxide bonding is normally carried out at elevated
temperatures (i.e. > 1000C) to achieve a strong, robust bond interface, the higher
temperatures required are not always compatible with processed wafers. In order
to bond completed ICs, a low temperature direct bond process is essential. Direct
oxide bonding often produces very high interface bond strength at low
temperatures. In one embodiment, the technology (Figure 12) involves a quick
plasma treatment (activation) followed by an aqueous ammonium hydroxide rinse
(termination). Such activation/termination processes can be easily implemented at
CMOS wafer foundries, IDMs or OSATs. Subsequent to direct bonding, vias
would then be formed by etching through the top chip down to the connecting pad
for the lower chip.
25
8.3.2Vias-first
In via-first technology, vias are introduced into the wafers either
before device formation[ (FEOL) or just before BEOL interconnect. In either
case, this would occur in the fab prior to completion of the die (wafer). Metal
metal bonding is favored by the industry for 3D IC integration because it
simultaneously forms both the mechanical and electrical bond. It is also generally
accepted that via-first technologies will be significantly easier to manufacture
since processing is at wafer scale and the vias are shallower/smaller. However, a
significant drawback for one process copper thermal compression (CuTC)
bonding is throughput. This bonding process involves heating the bonded
wafers to 350 400C for 30+ minutes under pressure, requiring that the prebonded wafer pair must spend considerable time at one bonding station. To
address this bottleneck, commercial aligner/bonder tool manufacturers have
developed multiple bonding stations, which add significant costs to these tools.
Without question, the industry is looking for a lower CoO method to bond wafers
for 3D integration.
The direct oxide bond is initiated in a few seconds with a standard
pick-and-place tool (W2W or D2W). Wafers are subsequently batch heated in a
cleanroom oven to ~ 300C to form a low resistance electrical bond at the aligned
Ni-Ni interface. Since the activated/terminated oxide layers are bonded together
with high strength, the Ni-Ni interface is subject to sufficient internal pressure so
that when the nickel expands at elevated temperature due to higher coefficient of
thermal expansion (CTE), a reliable metallic bond results. The combination of a
lower cost pick-and-place tool and high wafer throughput lead to lower CoO by
using this low temperature oxide bonding process.
26
27
10. ADVANTAGES
Reduced latency With vastly more responders built into HMC, we expect
lower queue delays and higher bank availability, which can provide a substantial
system latency reduction, which is especially attractive in network system
architectures.
Increased bandwidth A single HMC can provide more than 15x the
Smaller physical systems HMCs stacked architecture uses nearly 90% less
11. DISADVANTAGES
Cost is high (30%higher than DDR3)
Designing is complex.
Manufacturing issues.
Need new motherboard form factor.
28
12. FUTURE
Now that the baseline technology has been verified, additional variations
of HMC will follow. DRAM process geometry and cell size will continue to
shrink to half the prize it is today. This and improved stacking will allow greater
density for a given cube bandwidth and area. HMC devices will extend beyond
8GB. The number of TSV connections will double to create a cube capable of 320
GB/s and beyond.
New interconnects will be optimized for a given system topology. Shortreach SERDES are being developed that are capable of less than 1 pj/bit.
Medium-reach SERDES will serve 810 inches of FR4. Silicon photonics will
extend the reach to 10 meters and beyond. Atomic memory operation, scatter
/gather, floating point processors, cache coherency and meta data are natural
candidates for inclusion in the HMC. Data manipulation closest to where the data
resides is the highest-bandwidth, lowest-power solution.
13. CONCLUSION
Dept. of ECE, College of Engg., Poonjar
29
14. REFERENCES
Dept. of ECE, College of Engg., Poonjar
30
http://ieeexplore.ieee.org
VLSI Technology (VLSIT), 2012 Symposium on
Date of Conference: 12-14 June 2012
Author(s): Jeddeloh, J. Keeth, B.
Page(s): 87 - 88
Product Type: Conference Publications
www.google.com
http://www.micron.com
http://en.wikipedia.org
www.youtube.com
31