{"id":931,"date":"2015-11-05T20:46:02","date_gmt":"2015-11-05T20:46:02","guid":{"rendered":"https:\/\/courses.candelalearning.com\/zelixite115resources\/?post_type=chapter&#038;p=931"},"modified":"2016-04-21T23:04:06","modified_gmt":"2016-04-21T23:04:06","slug":"reading-random-access-memory","status":"publish","type":"chapter","link":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/chapter\/reading-random-access-memory\/","title":{"raw":"Reading: Random Access Memory","rendered":"Reading: Random Access Memory"},"content":{"raw":"<div class=\"hatnote\">\r\n<div class=\"thumbinner\">\r\n\r\n[caption id=\"\" align=\"alignright\" width=\"220\"]<img class=\"thumbimage\" src=\"https:\/\/s3-us-west-2.amazonaws.com\/candimgs\/LXQ9KC\/220px-Memory_module_DDRAM_20-03-2006.jpg\" srcset=\"\/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/c\/ca\/Memory_module_DDRAM_20-03-2006.jpg\/330px-Memory_module_DDRAM_20-03-2006.jpg 1.5x, \/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/c\/ca\/Memory_module_DDRAM_20-03-2006.jpg\/440px-Memory_module_DDRAM_20-03-2006.jpg 2x\" alt=\"Photo of strips of random-access memory.\" width=\"220\" height=\"165\" data-file-width=\"2304\" data-file-height=\"1728\" \/> Example of writable volatile\u00a0random-access memory: Synchronous\u00a0Dynamic RAM modules, primarily used as main memory in personal computers, workstations, and servers.[\/caption]\r\n\r\n<div class=\"thumbcaption\">\r\n<div class=\"magnify\"><b>Random-access memory<\/b> (<b>RAM<\/b> <span class=\"nowrap\"><span class=\"IPA nopopups\">\/<span title=\"'r' in 'rye'\">r<\/span><span title=\"\/\u00e6\/ short 'a' in 'bad'\">\u00e6<\/span><span title=\"'m' in 'my'\">m<\/span>\/<\/span><\/span>) is a form of computer data storage. A random-access memory device allows data items to be accessed (read or written) in almost the same amount of time irrespective of the physical location of data inside the memory. In contrast, with other direct-access data storage media such as hard disks, CD-RWs, DVD-RWs and the older drum memory, the time required to read and write data items varies significantly depending on their physical locations on the recording medium, due to mechanical limitations such as media rotation speeds and arm movement delays.<\/div>\r\n<\/div>\r\n<\/div>\r\n<\/div>\r\nToday, random-access memory takes the form of integrated circuits. RAM is normally associated with volatile types of memory (such as\u00a0DRAM memory modules), where stored information is lost if power is removed, although many efforts have been made to develop non-volatile RAM chips.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>Other types of non-volatile memory exist that allow random access for read operations, but either do not allow write operations or have limitations on them. These include most types of ROM and a type of flash memory called <i>NOR-Flash<\/i>.\r\n\r\nIntegrated-circuit RAM chips came into the market in the late 1960s, with the first commercially available DRAM chip, the Intel 1103, introduced in October 1970.\r\n<h2><span id=\"History\" class=\"mw-headline\">History<\/span><\/h2>\r\n<div class=\"thumb tright\">\r\n<div class=\"thumbinner\">\r\n\r\n[caption id=\"\" align=\"alignleft\" width=\"220\"]<img class=\"thumbimage\" src=\"https:\/\/s3-us-west-2.amazonaws.com\/candimgs\/LXQ9KC\/220px-Early_SSA_accounting_operations.jpg\" srcset=\"\/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/f\/f7\/Early_SSA_accounting_operations.jpg\/330px-Early_SSA_accounting_operations.jpg 1.5x, \/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/f\/f7\/Early_SSA_accounting_operations.jpg\/440px-Early_SSA_accounting_operations.jpg 2x\" alt=\"Old black and white photo of tabulating counters\" width=\"220\" height=\"269\" data-file-width=\"473\" data-file-height=\"578\" \/> These IBM tabulating machines from the 1930s used mechanical counters to store information[\/caption]\r\n\r\n<div class=\"thumbcaption\">\r\n<div class=\"magnify\">\u00a0Early computers used relays, mechanical counters\u00a0or delay lines for main memory functions. Ultrasonic delay lines could only reproduce data in the order it was written. Drum memory could be expanded at relatively low cost but efficient retrieval of memory items required knowledge of the physical layout of the drum to optimize speed. Latches built out of vacuum tube triodes, and later, out of discrete transistors, were used for smaller and faster memories such as registers. Such registers were relatively large and too costly to use for large amounts of data; generally only a few dozen or few hundred bits of such memory could be provided.<\/div>\r\n<\/div>\r\n<\/div>\r\n<\/div>\r\nThe first practical form of random-access memory was the Williams tube starting in 1947. It stored data as electrically charged spots on the face of a cathode ray tube. Since the electron beam of the CRT could read and write the spots on the tube in any order, memory was random access. The capacity of the Williams tube was a few hundred to around a thousand bits, but it was much smaller, faster, and more power-efficient than using individual vacuum tube latches. Developed at the University of Manchester in England, the Williams tube provided the medium on which the first electronically stored-memory program was implemented in the Manchester Small-Scale Experimental Machine (SSEM) computer, which first successfully ran a program on 21 June 1948.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>In fact, rather than the Williams tube memory being designed for the SSEM, the SSEM was a testbed to demonstrate the reliability of the memory.\r\n\r\n[caption id=\"\" align=\"alignleft\" width=\"220\"]<img class=\"thumbimage\" src=\"https:\/\/s3-us-west-2.amazonaws.com\/candimgs\/LXQ9KC\/220px-8_bytes_vs._8Gbytes.jpg\" srcset=\"\/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/c\/c0\/8_bytes_vs._8Gbytes.jpg\/330px-8_bytes_vs._8Gbytes.jpg 1.5x, \/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/c\/c0\/8_bytes_vs._8Gbytes.jpg\/440px-8_bytes_vs._8Gbytes.jpg 2x\" alt=\"SD card\" width=\"220\" height=\"165\" data-file-width=\"2048\" data-file-height=\"1536\" \/> A portion of a core memory with a modern flash RAM SD card on top[\/caption]\r\n\r\nMagnetic-core memory was invented in 1947 and developed up until the mid-1970s. It became a widespread form of random-access memory, relying on an array of magnetized rings. By changing the sense of each ring's magnetization, data could be stored with one bit stored per ring. Since every ring had a combination of address wires to select and read or write it, access to any memory location in any sequence was possible.\r\n\r\n[caption id=\"\" align=\"alignright\" width=\"220\"]<img class=\"thumbimage\" src=\"https:\/\/s3-us-west-2.amazonaws.com\/candimgs\/LXQ9KC\/220px-Bundesarchiv_Bild_183-1989-0406-022_VEB_Carl_Zeiss_Jena_1-Megabit-Chip.jpg\" srcset=\"\/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/e\/ec\/Bundesarchiv_Bild_183-1989-0406-022%2C_VEB_Carl_Zeiss_Jena%2C_1-Megabit-Chip.jpg\/330px-Bundesarchiv_Bild_183-1989-0406-022%2C_VEB_Carl_Zeiss_Jena%2C_1-Megabit-Chip.jpg 1.5x, \/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/e\/ec\/Bundesarchiv_Bild_183-1989-0406-022%2C_VEB_Carl_Zeiss_Jena%2C_1-Megabit-Chip.jpg\/440px-Bundesarchiv_Bild_183-1989-0406-022%2C_VEB_Carl_Zeiss_Jena%2C_1-Megabit-Chip.jpg 2x\" alt=\"Photo of a megabit chip\" width=\"220\" height=\"154\" data-file-width=\"791\" data-file-height=\"553\" \/> 1 Megabit chip \u2013 one of the last models developed by VEB Carl Zeiss Jena in 1989[\/caption]\r\n\r\nMagnetic core memory was the standard form of memory system until displaced by solid-state memory in integrated circuits, starting in the early 1970s. Robert H. Dennard invented dynamic random-access memory (DRAM) in 1968; this allowed replacement of a 4 or 6-transistor latch circuit by a single transistor for each memory bit, greatly increasing memory density at the cost of volatility. Data was stored in the tiny capacitance of each transistor, and had to be periodically refreshed every few milliseconds before the charge could leak away.\r\n\r\nPrior to the development of integrated read-only memory (ROM) circuits, <i>permanent<\/i> (or <i>read-only<\/i>) random-access memory was often constructed using diode matrices driven by address decoders, or specially wound core rope memory planes.\r\n<h2><span id=\"Types_of_RAM\" class=\"mw-headline\">Types of RAM<\/span><\/h2>\r\nThe two main forms of modern RAM are static RAM (SRAM) and dynamic RAM (DRAM). In SRAM, a bit of data is stored using the state of a six transistor memory cell. This form of RAM is more expensive to produce, but is generally faster and requires less power than DRAM and, in modern computers, is often used as cache memory for the CPU. DRAM stores a bit of data using a transistor and capacitor pair, which together comprise a DRAM memory cell. The capacitor holds a high or low charge (1 or 0, respectively), and the transistor acts as a switch that lets the control circuitry on the chip read the capacitor's state of charge or change it. As this form of memory is less expensive to produce than static RAM, it is the predominant form of computer memory used in modern computers.\r\n\r\nBoth static and dynamic RAM are considered <i>volatile<\/i>, as their state is lost or reset when power is removed from the system. By contrast, read-only memory (ROM) stores data by permanently enabling or disabling selected transistors, such that the memory cannot be altered. Writeable variants of ROM (such as EEPROM and flash memory) share properties of both ROM and RAM, enabling data topersist without power and to be updated without requiring special equipment. These persistent forms of semiconductor ROM include USB flash drives, memory cards for cameras and portable devices, etc. ECC memory (which can be either SRAM or DRAM) includes special circuitry to detect and\/or correct random faults (memory errors) in the stored data, using parity bits or error correction code.\r\n\r\nIn general, the term <i>RAM<\/i> refers solely to solid-state memory devices (either DRAM or SRAM), and more specifically the main memory in most computers. In optical storage, the termDVD-RAM is somewhat of a misnomer since, unlike CD-RW or DVD-RW it does not need to be erased before reuse. Nevertheless, a DVD-RAM behaves much like a hard disc drive if somewhat slower.\r\n<h2><span id=\"Memory_hierarchy\" class=\"mw-headline\">Memory hierarchy<\/span><\/h2>\r\nOne can read and over-write data in RAM. Many computer systems have a memory hierarchy consisting of processor registers, on-die SRAM caches, external caches, DRAM, pagingsystems and virtual memory or swap space on a hard drive. This entire pool of memory may be referred to as \"RAM\" by many developers, even though the various subsystems can have very different access times, violating the original concept behind the <i>random access<\/i> term in RAM. Even within a hierarchy level such as DRAM, the specific row, column, bank,rank, channel, or interleave organization of the components make the access time variable, although not to the extent that rotating storage media or a tape is variable. The overall goal of using a memory hierarchy is to obtain the higher possible average access performance while minimizing the total cost of the entire memory system (generally, the memory hierarchy follows the access time with the fast CPU registers at the top and the slow hard drive at the bottom).\r\n\r\nIn many modern personal computers, the RAM comes in an easily upgraded form of modules called memory modules or DRAM modules about the size of a few sticks of chewing gum. These can quickly be replaced should they become damaged or when changing needs demand more storage capacity. As suggested above, smaller amounts of RAM (mostly SRAM) are also integrated in the CPU and other ICs on the motherboard, as well as in hard-drives, CD-ROMs, and several other parts of the computer system.\r\n<h2><span id=\"Other_uses_of_RAM\" class=\"mw-headline\">Other uses of RAM<\/span><\/h2>\r\nIn addition to serving as temporary storage and working space for the operating system and applications, RAM is used in numerous other ways.\r\n<h3><span id=\"Virtual_memory\" class=\"mw-headline\">Virtual memory<\/span><\/h3>\r\nMost modern operating systems employ a method of extending RAM capacity, known as \"virtual memory\". A portion of the computer's hard drive is set aside for a <i>paging file<\/i> or a<i>scratch partition<\/i>, and the combination of physical RAM and the paging file form the system's total memory. (For example, if a computer has 2 GB of RAM and a 1 GB page file, the operating system has 3 GB total memory available to it.) When the system runs low on physical memory, it can \"swap\" portions of RAM to the paging file to make room for new data, as well as to read previously swapped information back into RAM. Excessive use of this mechanism results in thrashing and generally hampers overall system performance, mainly because hard drives are far slower than RAM.\r\n<h3><span id=\"RAM_disk\" class=\"mw-headline\">RAM disk<\/span><\/h3>\r\nSoftware can \"partition\" a portion of a computer's RAM, allowing it to act as a much faster hard drive that is called a RAM disk. A RAM disk loses the stored data when the computer is shut down, unless memory is arranged to have a standby battery source.\r\n<h3><span id=\"Shadow_RAM\" class=\"mw-headline\">Shadow RAM<\/span><\/h3>\r\nSometimes, the contents of a relatively slow ROM chip are copied to read\/write memory to allow for shorter access times. The ROM chip is then disabled while the initialized memory locations are switched in on the same block of addresses (often write-protected). This process, sometimes called <i>shadowing<\/i>, is fairly common in both computers and embedded systems.\r\n\r\nAs a common example, the BIOS in typical personal computers often has an option called \u201cuse shadow BIOS\u201d or similar. When enabled, functions relying on data from the BIOS\u2019s ROM will instead use DRAM locations (most can also toggle shadowing of video card ROM or other ROM sections). Depending on the system, this may not result in increased performance, and may cause incompatibilities. For example, some hardware may be inaccessible to the operating system if shadow RAM is used. On some systems the benefit may be hypothetical because the BIOS is not used after booting in favor of direct hardware access. Free memory is reduced by the size of the shadowed ROMs.\r\n<h2><span id=\"Recent_developments\" class=\"mw-headline\">Recent developments<\/span><\/h2>\r\nSeveral new types of <i>non-volatile<\/i> RAM, which will preserve data while powered down, are under development. The technologies used include carbon nanotubes and approaches utilizing Tunnel magnetoresistance. Amongst the 1st generation MRAM, a 128 KiB (<span class=\"nowrap\">128 \u00d7 2<sup>10<\/sup><\/span> bytes) chip was manufactured with 0.18\u00a0\u00b5m technology in the summer of 2003.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>In June 2004, Infineon Technologies unveiled a 16\u00a0MiB (16\u00a0\u00d7\u00a02<sup>20<\/sup> bytes) prototype again based on 0.18\u00a0\u00b5m technology. There are two 2nd generation techniques currently in development: thermal-assisted switching (TAS)\u00a0which is being developed by Crocus Technology, and spin-transfer torque (STT) on which Crocus, Hynix, IBM, and several other companies are working.\u00a0Nantero built a functioning carbon nanotube memory prototype 10\u00a0GiB (10\u00a0\u00d7\u00a02<sup>30<\/sup> bytes) array in 2004. Whether some of these technologies will be able to eventually take a significant market share from either DRAM, SRAM, or flash-memory technology, however, remains to be seen.\r\n\r\nSince 2006, \"solid-state drives\" (based on flash memory) with capacities exceeding 256 gigabytes and performance far exceeding traditional disks have become available. This development has started to blur the definition between traditional random-access memory and \"disks\", dramatically reducing the difference in performance.\r\n\r\nSome kinds of random-access memory, such as \"EcoRAM\", are specifically designed for server farms, where low power consumption is more important than speed.\r\n<h2><span id=\"Memory_wall\" class=\"mw-headline\">Memory wall<\/span><\/h2>\r\nThe \"memory wall\" is the growing disparity of speed between CPU and memory outside the CPU chip. An important reason for this disparity is the limited communication bandwidth beyond chip boundaries. From 1986 to 2000, CPU speed improved at an annual rate of 55% while memory speed only improved at 10%. Given these trends, it was expected that memory latency would become an overwhelming bottleneck in computer performance.\r\n\r\nCPU speed improvements slowed significantly partly due to major physical barriers and partly because current CPU designs have already hit the memory wall in some sense. Intel\u00a0summarized these causes in a 2005 document.\r\n<blockquote>\u201cFirst of all, as chip geometries shrink and clock frequencies rise, the transistor leakage current increases, leading to excess power consumption and heat... Secondly, the advantages of higher clock speeds are in part negated by memory latency, since memory access times have not been able to keep pace with increasing clock frequencies. Third, for certain applications, traditional serial architectures are becoming less efficient as processors get faster (due to the so-called Von Neumann bottleneck), further undercutting any gains that frequency increases might otherwise buy. In addition, partly due to limitations in the means of producing inductance within solid state devices, resistance-capacitance (RC) delays in signal transmission are growing as feature sizes shrink, imposing an additional bottleneck that frequency increases don't address.\u201d<\/blockquote>\r\nThe RC delays in signal transmission were also noted in Clock Rate versus IPC: The End of the Road for Conventional Microarchitectures which projects a maximum of 12.5% average annual CPU performance improvement between 2000 and 2014.\r\n\r\nA different concept is the processor-memory performance gap, which can be addressed by 3D computer chips that reduce the distance between the logic and memory aspects that are further apart in a 2D chip.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>Memory subsystem design requires a focus on the gap, which is widening over time.\u00a0The main method of bridging the gap is the use of caches; small amounts of high-speed memory that houses recent operations and instructions nearby the processor, speeding up the execution of those operations or instructions in cases where they are called upon frequently. Multiple levels of caching have been developed in order to deal with the widening of the gap, and the performance of high-speed modern computers are reliant on evolving caching techniques.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>These can prevent the loss of performance that the processor has, as it takes less time to perform the computation it has been initiated to complete.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>There can be up to a 53% difference between the growth in speed of processor speeds and the lagging speed of main memory access.","rendered":"<div class=\"hatnote\">\n<div class=\"thumbinner\">\n<div style=\"width: 230px\" class=\"wp-caption alignright\"><img loading=\"lazy\" decoding=\"async\" class=\"thumbimage\" src=\"https:\/\/s3-us-west-2.amazonaws.com\/candimgs\/LXQ9KC\/220px-Memory_module_DDRAM_20-03-2006.jpg\" srcset=\"\/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/c\/ca\/Memory_module_DDRAM_20-03-2006.jpg\/330px-Memory_module_DDRAM_20-03-2006.jpg 1.5x, \/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/c\/ca\/Memory_module_DDRAM_20-03-2006.jpg\/440px-Memory_module_DDRAM_20-03-2006.jpg 2x\" alt=\"Photo of strips of random-access memory.\" width=\"220\" height=\"165\" data-file-width=\"2304\" data-file-height=\"1728\" \/><\/p>\n<p class=\"wp-caption-text\">Example of writable volatile\u00a0random-access memory: Synchronous\u00a0Dynamic RAM modules, primarily used as main memory in personal computers, workstations, and servers.<\/p>\n<\/div>\n<div class=\"thumbcaption\">\n<div class=\"magnify\"><b>Random-access memory<\/b> (<b>RAM<\/b> <span class=\"nowrap\"><span class=\"IPA nopopups\">\/<span title=\"'r' in 'rye'\">r<\/span><span title=\"\/\u00e6\/ short 'a' in 'bad'\">\u00e6<\/span><span title=\"'m' in 'my'\">m<\/span>\/<\/span><\/span>) is a form of computer data storage. A random-access memory device allows data items to be accessed (read or written) in almost the same amount of time irrespective of the physical location of data inside the memory. In contrast, with other direct-access data storage media such as hard disks, CD-RWs, DVD-RWs and the older drum memory, the time required to read and write data items varies significantly depending on their physical locations on the recording medium, due to mechanical limitations such as media rotation speeds and arm movement delays.<\/div>\n<\/div>\n<\/div>\n<\/div>\n<p>Today, random-access memory takes the form of integrated circuits. RAM is normally associated with volatile types of memory (such as\u00a0DRAM memory modules), where stored information is lost if power is removed, although many efforts have been made to develop non-volatile RAM chips.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>Other types of non-volatile memory exist that allow random access for read operations, but either do not allow write operations or have limitations on them. These include most types of ROM and a type of flash memory called <i>NOR-Flash<\/i>.<\/p>\n<p>Integrated-circuit RAM chips came into the market in the late 1960s, with the first commercially available DRAM chip, the Intel 1103, introduced in October 1970.<\/p>\n<h2><span id=\"History\" class=\"mw-headline\">History<\/span><\/h2>\n<div class=\"thumb tright\">\n<div class=\"thumbinner\">\n<div style=\"width: 230px\" class=\"wp-caption alignleft\"><img loading=\"lazy\" decoding=\"async\" class=\"thumbimage\" src=\"https:\/\/s3-us-west-2.amazonaws.com\/candimgs\/LXQ9KC\/220px-Early_SSA_accounting_operations.jpg\" srcset=\"\/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/f\/f7\/Early_SSA_accounting_operations.jpg\/330px-Early_SSA_accounting_operations.jpg 1.5x, \/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/f\/f7\/Early_SSA_accounting_operations.jpg\/440px-Early_SSA_accounting_operations.jpg 2x\" alt=\"Old black and white photo of tabulating counters\" width=\"220\" height=\"269\" data-file-width=\"473\" data-file-height=\"578\" \/><\/p>\n<p class=\"wp-caption-text\">These IBM tabulating machines from the 1930s used mechanical counters to store information<\/p>\n<\/div>\n<div class=\"thumbcaption\">\n<div class=\"magnify\">\u00a0Early computers used relays, mechanical counters\u00a0or delay lines for main memory functions. Ultrasonic delay lines could only reproduce data in the order it was written. Drum memory could be expanded at relatively low cost but efficient retrieval of memory items required knowledge of the physical layout of the drum to optimize speed. Latches built out of vacuum tube triodes, and later, out of discrete transistors, were used for smaller and faster memories such as registers. Such registers were relatively large and too costly to use for large amounts of data; generally only a few dozen or few hundred bits of such memory could be provided.<\/div>\n<\/div>\n<\/div>\n<\/div>\n<p>The first practical form of random-access memory was the Williams tube starting in 1947. It stored data as electrically charged spots on the face of a cathode ray tube. Since the electron beam of the CRT could read and write the spots on the tube in any order, memory was random access. The capacity of the Williams tube was a few hundred to around a thousand bits, but it was much smaller, faster, and more power-efficient than using individual vacuum tube latches. Developed at the University of Manchester in England, the Williams tube provided the medium on which the first electronically stored-memory program was implemented in the Manchester Small-Scale Experimental Machine (SSEM) computer, which first successfully ran a program on 21 June 1948.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>In fact, rather than the Williams tube memory being designed for the SSEM, the SSEM was a testbed to demonstrate the reliability of the memory.<\/p>\n<div style=\"width: 230px\" class=\"wp-caption alignleft\"><img loading=\"lazy\" decoding=\"async\" class=\"thumbimage\" src=\"https:\/\/s3-us-west-2.amazonaws.com\/candimgs\/LXQ9KC\/220px-8_bytes_vs._8Gbytes.jpg\" srcset=\"\/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/c\/c0\/8_bytes_vs._8Gbytes.jpg\/330px-8_bytes_vs._8Gbytes.jpg 1.5x, \/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/c\/c0\/8_bytes_vs._8Gbytes.jpg\/440px-8_bytes_vs._8Gbytes.jpg 2x\" alt=\"SD card\" width=\"220\" height=\"165\" data-file-width=\"2048\" data-file-height=\"1536\" \/><\/p>\n<p class=\"wp-caption-text\">A portion of a core memory with a modern flash RAM SD card on top<\/p>\n<\/div>\n<p>Magnetic-core memory was invented in 1947 and developed up until the mid-1970s. It became a widespread form of random-access memory, relying on an array of magnetized rings. By changing the sense of each ring&#8217;s magnetization, data could be stored with one bit stored per ring. Since every ring had a combination of address wires to select and read or write it, access to any memory location in any sequence was possible.<\/p>\n<div style=\"width: 230px\" class=\"wp-caption alignright\"><img loading=\"lazy\" decoding=\"async\" class=\"thumbimage\" src=\"https:\/\/s3-us-west-2.amazonaws.com\/candimgs\/LXQ9KC\/220px-Bundesarchiv_Bild_183-1989-0406-022_VEB_Carl_Zeiss_Jena_1-Megabit-Chip.jpg\" srcset=\"\/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/e\/ec\/Bundesarchiv_Bild_183-1989-0406-022%2C_VEB_Carl_Zeiss_Jena%2C_1-Megabit-Chip.jpg\/330px-Bundesarchiv_Bild_183-1989-0406-022%2C_VEB_Carl_Zeiss_Jena%2C_1-Megabit-Chip.jpg 1.5x, \/\/upload.wikimedia.org\/wikipedia\/commons\/thumb\/e\/ec\/Bundesarchiv_Bild_183-1989-0406-022%2C_VEB_Carl_Zeiss_Jena%2C_1-Megabit-Chip.jpg\/440px-Bundesarchiv_Bild_183-1989-0406-022%2C_VEB_Carl_Zeiss_Jena%2C_1-Megabit-Chip.jpg 2x\" alt=\"Photo of a megabit chip\" width=\"220\" height=\"154\" data-file-width=\"791\" data-file-height=\"553\" \/><\/p>\n<p class=\"wp-caption-text\">1 Megabit chip \u2013 one of the last models developed by VEB Carl Zeiss Jena in 1989<\/p>\n<\/div>\n<p>Magnetic core memory was the standard form of memory system until displaced by solid-state memory in integrated circuits, starting in the early 1970s. Robert H. Dennard invented dynamic random-access memory (DRAM) in 1968; this allowed replacement of a 4 or 6-transistor latch circuit by a single transistor for each memory bit, greatly increasing memory density at the cost of volatility. Data was stored in the tiny capacitance of each transistor, and had to be periodically refreshed every few milliseconds before the charge could leak away.<\/p>\n<p>Prior to the development of integrated read-only memory (ROM) circuits, <i>permanent<\/i> (or <i>read-only<\/i>) random-access memory was often constructed using diode matrices driven by address decoders, or specially wound core rope memory planes.<\/p>\n<h2><span id=\"Types_of_RAM\" class=\"mw-headline\">Types of RAM<\/span><\/h2>\n<p>The two main forms of modern RAM are static RAM (SRAM) and dynamic RAM (DRAM). In SRAM, a bit of data is stored using the state of a six transistor memory cell. This form of RAM is more expensive to produce, but is generally faster and requires less power than DRAM and, in modern computers, is often used as cache memory for the CPU. DRAM stores a bit of data using a transistor and capacitor pair, which together comprise a DRAM memory cell. The capacitor holds a high or low charge (1 or 0, respectively), and the transistor acts as a switch that lets the control circuitry on the chip read the capacitor&#8217;s state of charge or change it. As this form of memory is less expensive to produce than static RAM, it is the predominant form of computer memory used in modern computers.<\/p>\n<p>Both static and dynamic RAM are considered <i>volatile<\/i>, as their state is lost or reset when power is removed from the system. By contrast, read-only memory (ROM) stores data by permanently enabling or disabling selected transistors, such that the memory cannot be altered. Writeable variants of ROM (such as EEPROM and flash memory) share properties of both ROM and RAM, enabling data topersist without power and to be updated without requiring special equipment. These persistent forms of semiconductor ROM include USB flash drives, memory cards for cameras and portable devices, etc. ECC memory (which can be either SRAM or DRAM) includes special circuitry to detect and\/or correct random faults (memory errors) in the stored data, using parity bits or error correction code.<\/p>\n<p>In general, the term <i>RAM<\/i> refers solely to solid-state memory devices (either DRAM or SRAM), and more specifically the main memory in most computers. In optical storage, the termDVD-RAM is somewhat of a misnomer since, unlike CD-RW or DVD-RW it does not need to be erased before reuse. Nevertheless, a DVD-RAM behaves much like a hard disc drive if somewhat slower.<\/p>\n<h2><span id=\"Memory_hierarchy\" class=\"mw-headline\">Memory hierarchy<\/span><\/h2>\n<p>One can read and over-write data in RAM. Many computer systems have a memory hierarchy consisting of processor registers, on-die SRAM caches, external caches, DRAM, pagingsystems and virtual memory or swap space on a hard drive. This entire pool of memory may be referred to as &#8220;RAM&#8221; by many developers, even though the various subsystems can have very different access times, violating the original concept behind the <i>random access<\/i> term in RAM. Even within a hierarchy level such as DRAM, the specific row, column, bank,rank, channel, or interleave organization of the components make the access time variable, although not to the extent that rotating storage media or a tape is variable. The overall goal of using a memory hierarchy is to obtain the higher possible average access performance while minimizing the total cost of the entire memory system (generally, the memory hierarchy follows the access time with the fast CPU registers at the top and the slow hard drive at the bottom).<\/p>\n<p>In many modern personal computers, the RAM comes in an easily upgraded form of modules called memory modules or DRAM modules about the size of a few sticks of chewing gum. These can quickly be replaced should they become damaged or when changing needs demand more storage capacity. As suggested above, smaller amounts of RAM (mostly SRAM) are also integrated in the CPU and other ICs on the motherboard, as well as in hard-drives, CD-ROMs, and several other parts of the computer system.<\/p>\n<h2><span id=\"Other_uses_of_RAM\" class=\"mw-headline\">Other uses of RAM<\/span><\/h2>\n<p>In addition to serving as temporary storage and working space for the operating system and applications, RAM is used in numerous other ways.<\/p>\n<h3><span id=\"Virtual_memory\" class=\"mw-headline\">Virtual memory<\/span><\/h3>\n<p>Most modern operating systems employ a method of extending RAM capacity, known as &#8220;virtual memory&#8221;. A portion of the computer&#8217;s hard drive is set aside for a <i>paging file<\/i> or a<i>scratch partition<\/i>, and the combination of physical RAM and the paging file form the system&#8217;s total memory. (For example, if a computer has 2 GB of RAM and a 1 GB page file, the operating system has 3 GB total memory available to it.) When the system runs low on physical memory, it can &#8220;swap&#8221; portions of RAM to the paging file to make room for new data, as well as to read previously swapped information back into RAM. Excessive use of this mechanism results in thrashing and generally hampers overall system performance, mainly because hard drives are far slower than RAM.<\/p>\n<h3><span id=\"RAM_disk\" class=\"mw-headline\">RAM disk<\/span><\/h3>\n<p>Software can &#8220;partition&#8221; a portion of a computer&#8217;s RAM, allowing it to act as a much faster hard drive that is called a RAM disk. A RAM disk loses the stored data when the computer is shut down, unless memory is arranged to have a standby battery source.<\/p>\n<h3><span id=\"Shadow_RAM\" class=\"mw-headline\">Shadow RAM<\/span><\/h3>\n<p>Sometimes, the contents of a relatively slow ROM chip are copied to read\/write memory to allow for shorter access times. The ROM chip is then disabled while the initialized memory locations are switched in on the same block of addresses (often write-protected). This process, sometimes called <i>shadowing<\/i>, is fairly common in both computers and embedded systems.<\/p>\n<p>As a common example, the BIOS in typical personal computers often has an option called \u201cuse shadow BIOS\u201d or similar. When enabled, functions relying on data from the BIOS\u2019s ROM will instead use DRAM locations (most can also toggle shadowing of video card ROM or other ROM sections). Depending on the system, this may not result in increased performance, and may cause incompatibilities. For example, some hardware may be inaccessible to the operating system if shadow RAM is used. On some systems the benefit may be hypothetical because the BIOS is not used after booting in favor of direct hardware access. Free memory is reduced by the size of the shadowed ROMs.<\/p>\n<h2><span id=\"Recent_developments\" class=\"mw-headline\">Recent developments<\/span><\/h2>\n<p>Several new types of <i>non-volatile<\/i> RAM, which will preserve data while powered down, are under development. The technologies used include carbon nanotubes and approaches utilizing Tunnel magnetoresistance. Amongst the 1st generation MRAM, a 128 KiB (<span class=\"nowrap\">128 \u00d7 2<sup>10<\/sup><\/span> bytes) chip was manufactured with 0.18\u00a0\u00b5m technology in the summer of 2003.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>In June 2004, Infineon Technologies unveiled a 16\u00a0MiB (16\u00a0\u00d7\u00a02<sup>20<\/sup> bytes) prototype again based on 0.18\u00a0\u00b5m technology. There are two 2nd generation techniques currently in development: thermal-assisted switching (TAS)\u00a0which is being developed by Crocus Technology, and spin-transfer torque (STT) on which Crocus, Hynix, IBM, and several other companies are working.\u00a0Nantero built a functioning carbon nanotube memory prototype 10\u00a0GiB (10\u00a0\u00d7\u00a02<sup>30<\/sup> bytes) array in 2004. Whether some of these technologies will be able to eventually take a significant market share from either DRAM, SRAM, or flash-memory technology, however, remains to be seen.<\/p>\n<p>Since 2006, &#8220;solid-state drives&#8221; (based on flash memory) with capacities exceeding 256 gigabytes and performance far exceeding traditional disks have become available. This development has started to blur the definition between traditional random-access memory and &#8220;disks&#8221;, dramatically reducing the difference in performance.<\/p>\n<p>Some kinds of random-access memory, such as &#8220;EcoRAM&#8221;, are specifically designed for server farms, where low power consumption is more important than speed.<\/p>\n<h2><span id=\"Memory_wall\" class=\"mw-headline\">Memory wall<\/span><\/h2>\n<p>The &#8220;memory wall&#8221; is the growing disparity of speed between CPU and memory outside the CPU chip. An important reason for this disparity is the limited communication bandwidth beyond chip boundaries. From 1986 to 2000, CPU speed improved at an annual rate of 55% while memory speed only improved at 10%. Given these trends, it was expected that memory latency would become an overwhelming bottleneck in computer performance.<\/p>\n<p>CPU speed improvements slowed significantly partly due to major physical barriers and partly because current CPU designs have already hit the memory wall in some sense. Intel\u00a0summarized these causes in a 2005 document.<\/p>\n<blockquote><p>\u201cFirst of all, as chip geometries shrink and clock frequencies rise, the transistor leakage current increases, leading to excess power consumption and heat&#8230; Secondly, the advantages of higher clock speeds are in part negated by memory latency, since memory access times have not been able to keep pace with increasing clock frequencies. Third, for certain applications, traditional serial architectures are becoming less efficient as processors get faster (due to the so-called Von Neumann bottleneck), further undercutting any gains that frequency increases might otherwise buy. In addition, partly due to limitations in the means of producing inductance within solid state devices, resistance-capacitance (RC) delays in signal transmission are growing as feature sizes shrink, imposing an additional bottleneck that frequency increases don&#8217;t address.\u201d<\/p><\/blockquote>\n<p>The RC delays in signal transmission were also noted in Clock Rate versus IPC: The End of the Road for Conventional Microarchitectures which projects a maximum of 12.5% average annual CPU performance improvement between 2000 and 2014.<\/p>\n<p>A different concept is the processor-memory performance gap, which can be addressed by 3D computer chips that reduce the distance between the logic and memory aspects that are further apart in a 2D chip.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>Memory subsystem design requires a focus on the gap, which is widening over time.\u00a0The main method of bridging the gap is the use of caches; small amounts of high-speed memory that houses recent operations and instructions nearby the processor, speeding up the execution of those operations or instructions in cases where they are called upon frequently. Multiple levels of caching have been developed in order to deal with the widening of the gap, and the performance of high-speed modern computers are reliant on evolving caching techniques.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>These can prevent the loss of performance that the processor has, as it takes less time to perform the computation it has been initiated to complete.<span style=\"font-size: 13.3333px; line-height: 20px;\">\u00a0<\/span>There can be up to a 53% difference between the growth in speed of processor speeds and the lagging speed of main memory access.<\/p>\n\n\t\t\t <section class=\"citations-section\" role=\"contentinfo\">\n\t\t\t <h3>Candela Citations<\/h3>\n\t\t\t\t\t <div>\n\t\t\t\t\t\t <div id=\"citation-list-931\">\n\t\t\t\t\t\t\t <div class=\"licensing\"><div class=\"license-attribution-dropdown-subheading\">CC licensed content, Shared previously<\/div><ul class=\"citation-list\"><li>Random access memory. <strong>Provided by<\/strong>: Wikipedia. <strong>Located at<\/strong>: <a target=\"_blank\" href=\"https:\/\/en.wikipedia.org\/wiki\/Random-access_memory\">https:\/\/en.wikipedia.org\/wiki\/Random-access_memory<\/a>. <strong>License<\/strong>: <em><a target=\"_blank\" rel=\"license\" href=\"https:\/\/creativecommons.org\/licenses\/by-sa\/4.0\/\">CC BY-SA: Attribution-ShareAlike<\/a><\/em><\/li><\/ul><\/div>\n\t\t\t\t\t\t <\/div>\n\t\t\t\t\t <\/div>\n\t\t\t <\/section>","protected":false},"author":74,"menu_order":6,"template":"","meta":{"_candela_citation":"[{\"type\":\"cc\",\"description\":\"Random access memory\",\"author\":\"\",\"organization\":\"Wikipedia\",\"url\":\"https:\/\/en.wikipedia.org\/wiki\/Random-access_memory\",\"project\":\"\",\"license\":\"cc-by-sa\",\"license_terms\":\"\"}]","CANDELA_OUTCOMES_GUID":"","pb_show_title":"on","pb_short_title":"","pb_subtitle":"","pb_authors":[],"pb_section_license":""},"chapter-type":[],"contributor":[],"license":[],"class_list":["post-931","chapter","type-chapter","status-publish","hentry"],"part":24,"_links":{"self":[{"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/pressbooks\/v2\/chapters\/931","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/pressbooks\/v2\/chapters"}],"about":[{"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/wp\/v2\/types\/chapter"}],"author":[{"embeddable":true,"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/wp\/v2\/users\/74"}],"version-history":[{"count":4,"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/pressbooks\/v2\/chapters\/931\/revisions"}],"predecessor-version":[{"id":1285,"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/pressbooks\/v2\/chapters\/931\/revisions\/1285"}],"part":[{"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/pressbooks\/v2\/parts\/24"}],"metadata":[{"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/pressbooks\/v2\/chapters\/931\/metadata\/"}],"wp:attachment":[{"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/wp\/v2\/media?parent=931"}],"wp:term":[{"taxonomy":"chapter-type","embeddable":true,"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/pressbooks\/v2\/chapter-type?post=931"},{"taxonomy":"contributor","embeddable":true,"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/wp\/v2\/contributor?post=931"},{"taxonomy":"license","embeddable":true,"href":"https:\/\/courses.lumenlearning.com\/sanjacinto-computerapps\/wp-json\/wp\/v2\/license?post=931"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}