On Wednesday, November 14, 2018, 11:52:43 AM PST, juan <[email protected]> 
wrote:  
 
 On Wed, 14 Nov 2018 19:00:52 +0000 (UTC)
jim bell <[email protected]> wrote:

 
>> My company, SemiDisk Systems, was very close to the first disk emulator for 
>> a number of types of PC, including the S-100, TRS-80 Model II, IBM PC, Epson 
>> Q-10.https://www.pcworld.com/article/246617/storage/evolution-of-the-solid-state-drive.html

>    IIRC you also worked for intel designing memory chips? Excuse my rather 
>naive question but...Did you see/hear at that time any hints that chips  were 
>being tampered with or somehow backdooored  because of 'national security'? 
I didn't design memory chips.  I was a "product engineer" for a specific 
self-refreshing dynamic RAM (otherwise called a "pseudo-static") device called 
a 2186.   
https://www.ebay.com/p/Vintage-Intel-D2186a-30-8k-X-8-Pseudo-Static-RAM-D2186-2186-SRAM/1918155784
    Vintage Intel D2186a-30 8k X 8 Pseudo Static RAM D2186 2186 SRAM | eBay    
It, along with a 32K x 8 "21D1", were Intel's first by-8 dynamic RAMs.
Product engineers design the test programs which check out the performance of a 
chip, using (at that time) an ultra-fast dedicated computer made by Teradyne. 
https://www.teradyne.com/products/test-solutions/semiconductor-test     This 
computer very accurately placed clock edges, to a position and accuracy of a 
small fraction of a nanosecond.   The 2186 was tricky by the standards of the 
day, partly due to the self-refreshing feature, but also because the 2186 (and 
21D1) were the first Intel memory devices (possibly the first from anyone?) 
that employed "redundancy":  Previous memory devices were essentially unusable 
if even a single bit, or row, or column failed.  The 2186 incorporated many 
spare rows, and spare columns, which could be programmed in to substitute for 
bits, rows, and columns that had failed.  

My program tested the chip, then took the map of bad rows, columns, and bits, 
and first checked to see if the part could be made good, at least 
theoretically, if the available rows and columns would solve the visible 
problems.  If that appeared to be possible, my program determined which 
redundant rows and columns needed to be activated, and at which row and column 
they needed to be placed at.  From this, a bit stream was generated that was 
clocked into the chip, one bit at a time, and was used to blow poly-silicon 
links (fuses) in a write-once memory area.  That was the memory area which told 
the chip where to access the redundant rows and columns, instead of the array 
rows and columns.  
In fact, I was the first person at Intel, and perhaps in the world, who saw the 
flash(es) through the microscope of the as-being-blown fuses on these chips.  
Intel was doing this redundancy before anyone else, I believe.  ×
Pseudo-static DRAMs refreshed themselves, with the (possible) aid of RFSH 
signal that might occasionally be applied to the chip.  Myself, I didn't think 
that DRAMs were hard to use, having designed a digital circuit and a DRAM card 
using an old Motorola DRAM called a "6605", that I got cheaply.  
https://computerarchive.org/files/mirror/www.bitsavers.org/pdf/motorola/_dataBooks/1979_Motorola_Memory_Data_Book.pdf
I don't think that the 2186 was successful, mostly because Intel eventually got 
out of the DRAM business, and mostly that because other manufacturers got much 
better and more efficient than Intel was.   
I was never in a position to hear if chips could be "backdoored".  
            Jim Bell




  

Reply via email to