A Caltech Library Service

Neurally inspired silicon learning : from synapse transistors to learning arrays


Diorio, Christopher J. (1997) Neurally inspired silicon learning : from synapse transistors to learning arrays. Dissertation (Ph.D.), California Institute of Technology. doi:10.7907/vbyq-fy15.


A computation is an operation that can be performed by a physical machine. We are familiar with digital computers: Machines based on a simple logic function (the binary NOR) and optimized for manipulating numeric variables with high precision. Other computing machines exist: The neurocomputer, the analog computer, the quantum computer, and the DNA computer all are known. Neurocomputers-defined colloquially as computing machines comprising nervous tissue-exist; that they are computers also is certain. Nervous tissue solves ill-posed problems in real time. The principles underlying neural computation, however, remain for now a mystery. I believe that there are fundamental principles of computation that we can learn by studying neurobiology. If we can understand how biological information-processing systems operate, then we can learn how to build circuits and systems that deal naturally with real-world data. My goal is to investigate the organizational and adaptive principles on which neural systems operate, and to build silicon integrated circuits that compute using these principles. I call my approach silicon neuroscience: the development of neurally inspired silicon-learning systems. I have developed, in a standard CMOS process, a family of single-transistor devices that I call synapse transistors. Like neural synapses, synapse transistors provide nonvolatile analog memory, compute the product of this stored memory and the applied input, allow bidirectional memory updates, and simultaneously perform an analog computation and determine locally their own memory updates. I have fabricated a synaptic array that affords a high synapse-transistor density, mimics the low power consumption of nervous tissue, and performs both fast, parallel computation and slow, local adaptation. Like nervous tissue, my array simultaneously and in parallel performs an analog computation and updates the nonvolatile analog memory. Although I do not believe that a single transistor can model the complex behavior of a neural synapse completely, my synapse transistors do implement a local learning function. I consider their development to be a first step toward achieving my goal of a silicon learning system.

Item Type:Thesis (Dissertation (Ph.D.))
Subject Keywords:(Electrical Engineering)
Degree Grantor:California Institute of Technology
Division:Engineering and Applied Science
Major Option:Electrical Engineering
Thesis Availability:Public (worldwide access)
Research Advisor(s):
  • Mead, Carver
Thesis Committee:
  • Mead, Carver (chair)
  • Koch, Christof
  • Psaltis, Demetri
  • Abu-Mostafa, Yaser S.
Defense Date:13 May 1997
Record Number:CaltechETD:etd-01092008-080326
Persistent URL:
Default Usage Policy:No commercial reproduction, distribution, display or performance rights in this work are provided.
ID Code:88
Deposited By: Imported from ETD-db
Deposited On:25 Jan 2008
Last Modified:31 Aug 2022 00:09

Thesis Files

PDF (Diorio_c_1997.pdf) - Final Version
See Usage Policy.


Repository Staff Only: item control page