throbber
SEMICONDUCTOR
`MEMORIES
`
`A Handbook of Design, Manufacture, and Application
`Second Edition
`
`Betty Prince
`
`IV 2005
`IPR2014-00317
`
`

`
`Betty Prince is New Product Planning Manager for worldwide memory
`marketing at Texas Instruments. She previously worked for N V Philips
`as strategic memory marketing manager and as business manager for
`the static RAM part of the European Megaproject. She earlier worked at
`Motorola as memory strategic marketing and applications manager, as
`European memory marketing manager and as memory project
`engineering manager.
`
`Dr Prince has participated in the memory experts group of the European
`Electronic Components Association, has been on the Electronic
`Industry Association JEDEC memory committee for nine years, is
`chair'man ofthe JEDEC low voltage technical standards committee, and
`is a member of the IEEE. She holds several patents in the memory area.
`
`She has a BSc and MSc in physics from the University of New Mexico
`and California respectively, an MBA in international marketing and a
`PhD in economics from the University ofTexas.
`
`

`
`Semiconductor
`Memories
`Second Edition
`
`

`
`Semiconductor
`Memories
`A Handbook of Design,
`Manufacture and Application
`
`Second Edition
`
`Betty Prince
`Texas Instruments, USA
`
`JOHN WILEY & SONS
`Chichester
`• New York
`• Brisbane
`
`• Toronto
`
`• Singapore
`
`

`
`Second edition of the book Semiconductor Memories by B. Prince and G. Due-Gundersen
`
`Copyright © 1983, 1991 by John Wiley & Sons Ltd.
`Baffins Lane, Chichester
`West Sussex P019 IUD, England
`
`All rights reserved.
`
`No part of this book may be reproduced by any means,
`or transmitted, or translated into a machine language
`without the written permission of the publisher.
`
`Other Wiley Editorial Offices
`
`John Wiley & Sons, Inc., 605 Third Avenue,
`New York, NY 10158-0012, USA
`
`Jacaranda Wiley Ltd, G.P.O. Box 859, Brisbane,
`Queensland 4001, Australia
`
`John Wiley & Sons (Canada) Ltd, 22 Worcester Road,
`Rexdale, Ontario M9W Ill, Canada
`
`John Wiley & Sons (SEA) Pte Ltd, 37 Jalan Pemimpin #05-04,
`Block B, Union Industrial Building, Singapore 2057
`
`Library of Congress Cata/oging-in-Publication Data
`
`Prince, Betty.
`Semiconductor memories I Betty Prince. -
`p.
`cm.
`Includes bibliographical references and index.
`ISBN 0 471 92465 2
`1. Semiconductor storage devices.
`TK7895.M4P74 1991
`62I.39'732--dc20
`
`I. Title.
`
`2nd ed.
`
`91-6943
`ClP
`
`British Library Cataloguing-in-Publication Data
`
`Prince, Betty
`Semiconductor memories. -
`I. Title
`621.3815
`
`ISBN 0 471 92465 2
`
`2nd ed.
`
`Typeset by Techset Composition Limited, Salisbury, Wiltshire
`Printed in Great Britain by Courier InternationaL East Kilbride
`
`

`
`3 TRENDS IN MEMORY
`APPLICATIONS
`
`3.1 VARIETIES OF DATA STORAGE DEVICES BY MEDIA
`
`Over the years many different media have been tried for the basic data storage
`function as shown in Figure 3.1. Logical bits have been stored in mechanical devices
`such as paper tape and cards, in magnetic moving media such as bubbles, hard disks,
`floppy disks and core, in optical moving media such as magneto-optical disk and
`holographic devices, and in solid state media which include semiconductors and
`ferroelectrics. Semiconductors divide by technology into MaS, bipolar and charge
`coupled devices, and by function into RAM, ROM, and SAM. MaS and bipolar are
`available in all three functions, while charge coupled devices are available only as SAM.
`
`3.2 FUNCTIONAL CHARACTERISTICS OF VARIOUS SEMICONDUCTOR
`MEMORIES
`
`The basic differences in function in semiconductor memories have come about as a
`result of specific requirements of the differing systems applications in which they are
`used. No single set of memory characteristics is optimal for all systems, nor has any
`single memory yet been made which has all of the optimum characteristics.
`Functional characteristics that are significant for the systems environment include
`performance (speed), power dissipation (heat), memory density (number of storage
`bits per chip), chip size (memory cost), size of package (system cost), external
`organization of the memory, reprogrammability (endurance of the memory to repeated
`write-erase cycles), long term reliability characteristics, ability to retain data when
`the power is off (volatility), length of time the data is retained when dc power is
`on without an active refresh of the data (data retention), interface voltage levels into
`the system (TTL, ECL, CMOS), optimal power supply voltage levels, moisture
`resistance (hermeticity) of the package, and the amount of logic integrated on the
`memory rather than used separately in the system.
`There are several major product types of semiconductor memories differing by
`various sets of these characteristics. This was determined historically by the basic
`
`

`
`FUNCTIONAL CHARACTERISTICS OF VARIOUS SEMICONDUCTOR MEMORIES
`
`49
`
`~ala'IO"g'med~
`
`Mechanical
`
`Moving media
`
`Solid state
`
`/\
`/ \
`lape /1 /1 If\ 1\
`
`Paper Punch
`
`Magnetic
`
`Optical
`
`Semiconductor
`
`Ferroelectric
`
`Bubble Hard
`disk
`
`Floppy Compact Holographic MOS Bipolar CCD FE-MOS Core
`disk
`disk
`
`Figure 3.1
`Various catagories of data storage devices.
`
`Characteristic
`
`DRAM
`
`SRAM
`
`EPROM
`
`ROM
`
`EEPROM
`
`NVRAM
`
`Number of devices
`in cell
`Relative cell size
`Density (1990)
`Overhead cost
`
`Volatile
`(power off)
`Data retention
`(d.c. power on)
`In system
`reprogram mabie
`Number of
`reprogram times
`(endurance)
`Typical write
`(Reprogram) speed
`Typical read
`speed (ns)
`Number of
`read cycles
`
`1.5
`
`4-6
`
`"1.5
`
`1.0
`
`2.5
`
`8-9
`
`1.5
`4Mb
`yes
`refresh
`logic
`yes
`
`4-6
`1Mb
`no
`
`yes
`
`"1.5
`4Mb
`
`1.0
`4Mb
`yes
`yes
`mask
`UVerase
`programmer charges
`no
`no
`
`4ms
`
`Cf)
`
`10 years
`
`yes
`
`yes
`
`no
`
`Cf)
`
`Cf)
`
`100
`PROM(1)
`
`Cf)
`
`no
`
`0
`
`~ 9-10
`1Mb
`16k
`no
`no
`
`no
`
`no
`
`10 years
`
`10 years
`
`yes
`
`yes
`
`10000
`-1000000
`
`10000
`XJ (write)
`
`100 ns
`
`25 ns
`
`30 min
`
`2.5 s
`
`100
`
`25
`
`100
`
`100
`
`200
`
`200
`
`Cf)
`
`Cf)
`
`.X)
`
`Cf)
`
`Cf)
`
`Cf)
`
`Figure 3.2
`Characteristics of MOS memory product types.
`
`

`
`50
`
`TRENDS IN MEMORY APPLICATIONS
`
`memory cell. Semiconductor memories are split by major cell type into DRAMs,
`SRAMs, SAMs, EPROMs, EEPROMs, and ROMs. General characteristics of the basic
`MOS memory types are shown in Figure 3.2. Minor variations include the NVRAM
`and the CAM.
`The structure of the memory array and periphery determines the input and output
`organization. The logic in the periphery also determines functional subcategories of
`memories such as video and multi-port DRAMs, synchronous and asynchronous
`SRAMs.
`A further subset is defined by the Input and Output (I/O) levels and configuration
`of the memory. I/O levels can be TTL, ECL, or CMOS. Inputs, outputs, and addresses
`can be multiplexed or non-multiplexed, serial or parallel.
`
`3.2.1 MOS memory selection by system requirement
`
`Different systems require different MOS memory characteristics. For example, a
`memory system requiring an infinitely reprogrammable memory would use an SRAM
`or a DRAM. If it is a large memory system with commercial cost pressure then it
`will use the highest density, lowest cost option-the DRAM. If it is a large very fast
`system requiring a random read access memory, then it will either use SRAM for the
`main memory or have a small SRAM cache which can make a lower cost but slower
`DRAM memory appear faster in the system. If it is important that the main memory
`not lose its data when the power fails,
`there are again several options. Either
`SRAM can be used for the main memory due to its low standby power dissipation
`and backed up by a battery, or the main memory can be backed up by an EEPROM.
`As another example a system which needs a small quantity of fast infinitely writable
`non-volatile memory will use an NVRAM, bearing in mind that its non-volatile
`reprogrammability is the same as the EEPROM-about 10 000 reprogram cycles.
`In general DRAMs and EPROMs tend to be optimized for high density and low
`cost and used in large systems where many parts need to be assembled in a small
`space. The large quantity of parts used mean that the per-unit overhead costs are low.
`DRAMs have the lowest cost of any RAM and the highest density because they
`have the smallest memory cell, consisting of one transistor and a capacitor. The
`capacitor stores the charge when the memory is in the' l' state. The capacitor occupies
`a large percentage of the room in the memory cell. In high density DRAMs it is
`usually either stacked on top of the transistor or lowered into a narrow trench in the
`silicon. A DRAM, therefore, effectively has a one plus transistor cell area.
`The penalty that is paid for this small cell size and consequent high density is the
`need for the memory system to provide a periodic refresh to the cell to restore the
`charge to the capacitor as it leaks away. During the time that the memory cell is being
`'refreshed' no other operation can take place in that area of the array. This' dead
`time' slows down the effective speed of the DRAM in the system. While the
`percentage of the operating time that is occupied with refresh is decreasing as higher
`quality capacitors are constructed, refresh is still a significant factor in the timing of
`the system.
`All of the capacitors in the memory array must be periodically refreshed. This is
`done by accessing each row in the array, one row at a time. When a row is accessed,
`
`

`
`FUNCTIONAL CHARACTERISTICS OF VARIOUS SEMICONDUCTOR MEMORIES
`
`51
`
`it is turned on and voltage is applied to the row. This recharges each capacitor on
`the row to its initial value. For example, on a 1Mb DRAM refresh can be performed
`as either a single burst of 512 consecutive refresh cycles (one cycle per row) every
`8 ms, or it can be distributed over time, that is, one 200 ns refresh cycle every 15.6 fls
`(8 ms per 512 rows = 15.6 fls per row). A combination of these two methods is also
`permitted as long as each row is refreshed at least every 8 ms. Details of refreshing
`the DRAM will be considered in Chapter 6. The requirement for refresh does, however,
`add to the complexity of the system and hence to the overhead cost of a system
`using DRAMs.
`EPROMs also have an overhead cost. They are erasable, but must be removed
`from the system and placed in an ultraviolet light which erases the memory through
`a transparent window in the EPROM package. They must then be reprogrammed in
`a separate programmer before being returned to the system. This makes EPROMs
`cumbersome to use.
`SRAMs and EEPROMs, on the other hand, are easy to use from the system
`sophistication viewpoint and do not have the overhead costs, but are more expensive
`per unit. In a small system where the cost of system overhead is a relatively
`higher part of the system cost, the total cost of the SRAMs and EEPROMs moves
`closer to that of the DRAMs and EPROMs, and the determining factor becomes the
`ease of use.
`Where in-system reprogrammability is required, neither EPROMs, nor ROMs
`can be used: however systems requiring data to be retained when the power goes off
`must use one of the non-volatile memories~ROMs, EPROMs, EEPROMs, non(cid:173)
`volatile RAMs, or CMOS SRAMs with battery back-up. If the high performance
`of the semiconductor memory is not required these applications can also use one of
`the slower non-volatile moving media or mechanical memory types.
`ROMs and EPROMs are used primarily to hold software which will not be changed.
`In lower densities and speeds, they were used mainly for holding setup and
`configuration data for peripheral or bootstrap programs. As higher speeds and densities
`appeared they were used to hold entire operating systems, real-time kernels and canned
`applications programs, and in consumer toys and games.
`Battery backed up SRAMs (BRAMs) are used for applications in peripherals
`requiring high speed random access combined with non-volatility. The parts, devel(cid:173)
`oped first by Mostek, contain lithium batteries in a compartment of the SRAM
`package. They also contain various types of voltage sensing circuits that switch over
`to the batteries when external power is removed.
`With relatively small numbers of setup parameters needing to be sorted and changed
`only occasionally, non-volatile SRAMs, which are also called shadow RAMs, can be
`used. These combine RAM and EEPROM on one chip. BRAMs can also be used.
`Shadow RAMs (developed by Xicor) combine a normal static RAM array with an
`EEPROM array. When the chip is powered up, the contents of the EEPROM are
`written into the RAM, providing default configuration parameters. These parameters
`can be changed while the system is on, and the EEPROM can be reprogrammed to
`change the default setting.
`Other characteristics are common choices to all the memory types but differ in
`various system applications, such as organization of the memory, common or separate
`
`

`
`52
`
`TRENDS IN MEMORY APPLICATIONS
`
`inputs and outputs, and whether the memory is synchronous or asynchronous. Depth
`of memory required in the system affects the required organization of the memory
`and also the width of memory word, or number of inputs and outputs. Considerations
`of bus width and error correction are often important determiners of desired memory
`width. Whether a memory is required to be synchronous or can be asynchronous can
`be determined by the presence or absence of a system clock, the desired speed
`of the system, and the required retention of the last accessed data while new
`data are incoming.
`
`3.3 MEMORY USAGE IN COMPUTER APPLICATIONS
`
`To see how differing systems applications require these differing characteristics, a few
`typical systems using memories from various end system market segments are
`considered now.
`Figure 3.3 divides the total memory usage for DRAMs and SRAMs in major
`applications segments for 1988 and forecast for 1992.
`
`1988
`$6.256 billion
`
`PCsond
`office
`automation
`53%
`
`(a)
`
`Other
`computer
`24%
`
`Telecom
`8.8%
`
`1992
`$8.972 billion
`
`PCs ond
`office
`
`Telecom
`--\!!!-- 10. 3 %
`
`Industrial/
`other
`7.1%
`Consumer
`8.6%
`
`Consumer
`5%
`
`Military
`3.5%
`
`1988
`$1.712 billion
`
`1992
`$2.644 billion
`
`PCs and
`office
`
`PCs and
`office
`Other com~uter automation
`22.8:b
`39%
`
`(b)
`
`Consumer
`5%
`
`Industrial/
`other
`15%
`
`Industriall
`
`Telecom
`7.5%
`
`Consumer
`10%
`
`Figure 3.3
`(a) DRAM and (b) SRAM usage by user market segment for 1988 and 1992. (From HTE
`Research Inc.)
`
`

`
`MAINFRAME STORAGE HIERARCHIES
`
`53
`
`Computer application is both the largest of the electronic systems market and also
`uses the highest average value of memory per system. As a result 73% of all memories
`go into some type of computer system. These include mainframes, supercomputers,
`parallel processing systems and graphics workstations.
`
`3.4 COMPUTER STORAGE HIERARCHIES
`
`Throughout the 1980s there was an increasing mismatch between the speed of the
`DRAMs, which are used for main memory, and the speed of the microprocessors,
`used as computing elements in all but the largest computers. Figure 3.4 illustrates this
`for the RISC (reduced instruction set computers), CISC (complex instruction set
`computer), MPU (microprocessor units), and for four generations of DRAM. Fast
`SRAM speeds increased at the same rate as that of the processors. A hierarchy of
`storage layers divided by speed developed.
`A typical computer storage hierarchy is shown in simplified form in Figure 3.5.
`Indications are given of memory types frequently used at each level. Memory is
`organized in several layers of decreasing speed going from fast cache to slow archival
`memory. Progress in mass storage technology continues to add layers to the traditional
`system storage hierarchy.
`If infinite amounts of low cost, random access, non-volatile memory bits were
`available at the speed required by the processors with no wait states, there would be
`little reason to use differing types of memory. Since this is not the case, memory
`hierarchies are used to provide a reasonable facsimile of the ideal memory in the
`system.
`
`100
`
`50
`
`20
`
`10
`
`5
`
`2
`
`N
`I
`::2
`
`w
`....J
`<.J
`>-
`<.J
`
`o Rise
`• else
`• DRAM
`
`20
`
`50
`
`'" c
`
`100
`
`200
`
`300
`
`82 83 84 85 86 87 88 89 90
`YEAR
`
`Figure 3.4
`Speed trends of MPUs and DRAMs. (From Kushiyama [29], Toshiba 1991, with permission
`of IEEE.)
`
`

`
`54
`
`TRENDS IN MEMORY APPLICATIONS
`
`First
`level
`cache
`
`SRAM
`
`Main memory
`
`DRAM
`
`Archive I
`
`Mognetic disk/
`magnetic tope
`
`Archive II
`
`Optical disk
`
`Figure 3.5
`Typical computer system data storage hierarchy.
`
`3.4.1 Main memory
`
`The storage hierarchy exists to support the main memory. The storage device used
`most frequently as main memory in large computers is the dynamic RAM. This
`is because the DRAM is optimized for low cost and high density, which are the major
`needs when fitting very large quantities of memories into a system. What is traded
`off is a higher cost in system overhead needed to refresh these dynamic RAMs,
`volatility, and a speed penalty.
`When it is necessary to interface a fast processor with a slower memory, the
`processor is slowed down since its cycle time is faster than the memory address access
`time. One of the ways to handle this interface situation is to use a cache to decrease
`the effective time to access the main memory.
`
`3.4.2 Overview of cache
`
`Cache memories are buffers that hold selected information from larger and usually
`slower main memory so that processors can reach this information more rapidly.
`The cache is based on the principle that certain information has a higher probability
`at any given time of being selected next by the processor than other information. If
`these higher probability data are stored in a cache with a faster access time then the
`average speed of access of the data will be increased.
`There is a trade off between memory performance, size, and cost, which encourages
`
`

`
`MAINFRAME STORAGE HIERARCHIES
`
`55
`
`system designers to find ways to use memory more efficiently. System designers have
`historically had two choices available in RAM memory devices: fast and expensive
`SRAMs or slow and cheap DRAMs. While SRAMs with 25-35 ns access times can
`keep pace with a 10 MHz processor, a DRAM with even 80 ns access time will force
`this same processor to wait. Given the large size requirement in current memory
`arrays, the importance of finding an optimum trade-off between size and performance
`is, however, of increased importance.
`Caching schemes are common which mix a large amount of slower less expensive
`DRAM with a small amount of faster more expensive SRAM. A variety of such
`schemes can be constructed to allow the processor to operate at maximum speed, and
`are a way to ensure that the data a processor needs will be available when the
`processor needs it without having to wait for the data to be read from their normal
`location in main DRAM memory.
`In general most caches are designed on the assumption that a processor is likely
`in the near future to need the information it is currently working on, plus other data
`which are stored near it in the main memory. Cache performance depends on two
`factors: access time and hit ratio.
`The hit ratio is defined as the probability of finding the needed information in the
`cache, which is a hit, divided by the probability of having to go to the main memory,
`which is a miss. This hit ratio, which is the ratio of the number of hits divided by
`the number of tries, is dependent on the size of the cache and on the number of bytes
`that are brought in on any given fetch from main memory.
`The access time for the cache is the mean time it takes to access the cache on a
`hit. The access time for the main memory must include the initial attempt to find the
`data in the cache. The overall performance of the memory system will then be a
`function of the cache access time, the cache hit ratio and the access time of the main
`memory for references that are not found in the cache.
`There are two general types of caches: a direct mapped cache and an associative
`or tag cache.
`
`3.4.3 Direct mapped cache
`
`The contents of a direct mapped cache correspond exactly to the data that were
`obtained from main memory and are addressed by the same addressing scheme as
`the main memory. The latest data obtained are normally retained by the cache along
`with adjacent data from the main memory. The least used data in cache at the time
`are replaced by the latest data.
`Direct mapped caches are used primarily with smaller main memory having
`direct address mapping schemes. Most direct mapped caches consist of small amounts
`of SRAMs embedded in microprocessors.
`
`3.4.4 Associative cache
`
`An associative cache, also called a cache tag comparator, is more useful with large
`memory systems which include large main memory and perhaps even lower levels
`of archival storage.
`
`

`
`56
`
`TRENDS IN MEMORY APPLICATIONS
`
`Frequently these large memory systems are organized as demand paged virtual
`memory in which memory is organized in pages of a fixed size that have a given
`virtual address. Pages beyond the range of existing physical RAM can be stored in
`large, slower archival memory, usually disk. and swapped into physical RAM when
`needed and where space is freed by another page being swapped back to disk. (In
`this case the main memory also acts as a cache for the disk.) When pages reside in
`physical memory, they also have a physical address which is different from their
`virtual address.
`There is a trade-off in page size between the frequency with which one must swap
`pages from the disk and the time it takes to make such a swap. A memory management
`unit (MMU) is used to translate between virtual and physical addresses. Addresses
`are organized in terms of a page address, called a 'TAG', and an 'offset' which locates
`the data within that page.
`An associative cache, upon receiving an address from the processor, associates or
`compares the TAG with the page address to see if that page address is in cache. If
`so, the offset address is checked and if there is a 'hit'. that is, if the data are in the
`cache, they are immediately transferred from the cache to the processor.
`An example of a single set associative cache is shown in Figure 3.6. This cache
`includes a TAG indicating the page address of data stored in the cache, the offset
`address which locates the data within the page, and the actual data. When a request
`is received from the processor, the TAG field in the address requested is compared
`to the TAG field of data held in the cache. If the TAG field is valid and the page
`offset is present then the data in the cache are returned to the processor.
`
`Processor
`
`r----------------
`I
`I
`I
`No I Comparator J
`
`Address
`I
`
`I
`
`Yes t
`Offset
`
`Tog
`
`Data
`
`l-
`
`I
`I
`I L __________________
`_.J
`
`Main memory
`
`I
`
`Figure 3.6
`Single set associative cache.
`
`

`
`MAINFRAME STORAGE HIERARCHIES
`
`57
`
`A two-way set associate cache would allow the same offset from two different
`pages to be cached. A fully associative cache would allow any number of pages to
`be cached. A fully associative cache is, however, more complex and hence more
`expensive than a simpler cache.
`Cache hierarchies historically included a processor, main memory, and a cache to
`buffer the speed of the main memory for input to the processor. Now cache hierarchies
`can be multilevel with fast processor, cache 1, virtual main memory, cache 2, and
`archive memory. Speeds range down from a few tens of nanoseconds for the first-level
`cache to a few milliseconds for the archive.
`
`3.4.5 Application-specific cache chips
`
`A cache can be constructed of logic and RAM chips or it can be bought as an SRAM
`with the logic integrated on the chip. A number of applications specific SRAMs are
`available as caches for memory systems, although there is always a trade-off for the
`system designer between using a more expensive cache chip, or buying less expensive
`standard SRAM and implementing the cache logic off-chip.
`Many levels of sophistication are possible in cache design, usually determined by
`the speed requirements and cost constraints of the system.
`
`3.4.6 Dual-port SRAMs as cache buffers
`
`An ideal cache could be accessed simultaneously by both a processor and another
`processor or a peripheral. This could result in bus contention unless the cache had
`two ports and the stem had two independent buses.
`Even with two ports, an ideal cache buffer would need to be able to handle both
`external processors accessing the same address simultaneously without resulting in
`indeterminate data in the cache if one processor was attempting to read when another
`was writing to the same location. In reality, this situation is frequently handled by
`having a wait state for one of the processors or by allowing the possibility of having
`information incorrect. Contention arbitration logic can be provided to eliminate
`overlapping operations to the same memory location. It can either allow contention
`to be ignored and both operations to proceed, or it can arbitrate and delay one port
`until the other port's operation is completed. When this occurs, a busy flag is sent
`to the side delayed. A priority port can be specified. Another problem that arises is
`having different width buses on the processor and the peripherals side of the dual port.
`
`3.4.7
`
`Interleaving memory to gain system speed
`
`An additional option for the fast processor and slow memory combination is
`interleaving the memory to increase the data rate. The most basic type of interleaving
`scheme is the two-way interleave in which odd addresses are stored in one memory
`
`

`
`58
`
`TRENDS IN MEMORY APPLICATIONS
`
`array and even addresses are stored in another. The first access will be at the speed
`of the memory array, but consecutive accesses from separate blocks will be at about
`twice the speed of the memory arrays.
`This works best in the case of consecutive addressing from the processor where
`there is a high level of interleaving. This scheme works since consecutive fetches are
`directed at alternate memory banks so there is a time penalty only on the first access.
`For example, two 100 ns memory arrays could be interleaved to give an average
`access time close to 50 ns, rather than using a more expensive 50 ns memory.
`The simple two-way interleaving scheme totally fails if, for example, complex
`numbers are being used where the real and imaginary parts are accessed consecutively
`but stored separately. Since each address access involves both memory arrays, the
`time advantage of alternating slower arrays disappears, leading to a failure of the
`interleaving scheme. A four-way interleave solves this problem. The interleaving
`scheme helps to speed up both instruction and data fetches from memory.
`Caches and interleaving can also be combined, for example, to speed up instruction
`fetches in time to be executed when needed. Ideally, the fetching would occur on
`machine cycles where no other task is using main memory. The cache memory acts
`as a buffer between the main memory and the control unit, greatly reducing the
`memory bandwidth consumed by instruction fetches.
`
`3.4.8 Archive or mass storage
`
`In addition to main memory and cache, system requirements for more memory and
`for higher capacity peripherals have resulted in the use of a variety of storage devices
`whose main requirements are very high density, very low cost, and non-volatility.
`These include magnetic disk, floppy disk, optical disk, and magnetic tape. Optical
`disks are available in three forms which include optical read-only-memory (ROM),
`optical write-once-read-many (WORM), which is a PROM, and most recently optical
`multiwrite, which is a form of EEPROM.
`Reel-to-reel has mostly been replaced by tape cartridges. Hard disks are not being
`displaced because their capacities continue to increase and their size to shrink. For
`example, the 5 ~ inch hard disks have increased from the 10 to 20M byte capacities
`common in the mid 1980s to 200Mbyte and higher in 1989, while 14 inch hard disks
`now have capacities in the gigabyte range.
`Interface standards for optimizing the speed and transparency of links between
`computers and peripherals have developed, such as SCSI (Small Computer System
`Interface) which is used primarily for small systems and IPI which was developed for
`larger systems. These interface standards allow designers to optimize the storage
`capacities of their devices.
`To accommodate these additional options the system I/O interface must become
`more flexible and supply faster data transfers.
`Solid state memory is still too expensive to be a major contender for most mass
`storage applications. It is, however, used in special applications whose reliability
`requirements preclude the use of mechanical storage systems. Since the cost per bit
`
`

`
`MEMORIES FOR VIDEO PROCESSING
`
`59
`
`of MOS memory continues to drop faster than that of the older media, an increasing
`number of applications could potentially be filled with MOS memory in the future.
`The low cost DRAM and flash memories are the main contenders among the MOS
`memories for a share of the archive memory market.
`
`3.5 SUPERCOMPUTERS AND PARALLEL PROCESSORS
`
`Performance in today's supercomputers is determined in part by the high performance
`communications networks connecting the various processors in the total computer
`system. Supercomputers run in parallel processing type clusters linked by these high
`speed networks.
`Applications where parallel processing makes sense include those where a massive
`amount of data must be processed, for example, image processing. The quantity of
`information represented by a typical digital image means that the total processing
`time is likely to be minutes or hours for a single image when a Single micro- or
`miniprocessor is used. The use of parallel processing can Significantly reduce the time
`for such computations.
`Parallel processors also require parallel input-output capabilities and mass storage
`arrays that bring enough I/O and disk performance to match the required levels of
`computational performance.
`Since speed is critical in these super fast systems, circuit elements need to be as
`close as possible to shorten the length of interconnects. Some supercomputer systems
`use high density cassettes of SRAM memory chips without packages to create the
`high density memory arrays needed. Other systems embed the data storage and data
`acquisition subsystems modules in the array of processors. In either case significantly
`enhanced computational power results so that complex processing of analog data
`becomes possible at very high speeds. Image and signal processing are among the
`possible applications.
`
`3.6 MEMORIES FOR VIDEO PROCESSING AND GRAPHICS WORKSTATIONS
`
`The advent of the microprocessor in the 1980s increased the demand for memories
`in small systems such as personal computers, word processors, computer terminals,
`work stations and CAD-CAM systems. The applications of RAMs in small systems
`can largely be divided into main memory and display memory.
`The ratio of number of RAMs used in display memories for storing character data
`and processing graphic data to the number used in the main memory is much larger
`in small systems than in mainframe systems, and in fact in small systems the number
`of display memories is on a par with the number of memories used in the main
`memory of the small system.
`The size of display memories per system ranges from several tens of kilobytes to
`as many as several megabytes. It is expected that during the 1990s a significant
`percentage of DRAMs will be used for display purposes. Display memories use
`
`

`
`60
`
`TRENDS IN MEMORY APPLICATIONS
`
`standard DRAMs integrated into the system with stand-alone logic chips and also
`include several applications specific DRAMs such as frame buffers and video DRAMs
`for bit-mapped displays.
`The basic function of a frame buffer is to send display data to the video display
`monitor at a designated speed. The performance level of a system is determined by
`the efficiency with which the processor can refresh the display data while continuing
`this basic function. A frame buffer works best with two access ports. One port intakes
`new data from the CPU and the other sends display data to the CRT.
`There are two ways to use a standard single port DRAM in this application. One
`is to permit the CPU access to the DRAM only during the blanking phase of the
`display as illustrated in Figure 3.7(a). CPU access during the data display phase is
`restricted to sending display data to the CRT. This r

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket