+ All Categories
Home > Documents > Design and implementation of a biologically realistic ...surements obtained from a fabricated analog...

Design and implementation of a biologically realistic ...surements obtained from a fabricated analog...

Date post: 20-Oct-2020
Category:
Upload: others
View: 1 times
Download: 0 times
Share this document with a friend
22
Design and Implementation of a Biologically Realistic Olfactory Cortex in Analog VLSI JOSE C. PRINCIPE, FELLOW, IEEE, VITOR G. TAVARES, JOHN G. HARRIS, AND WALTER J. FREEMAN, FELLOW, IEEE Invited Paper This paper reviews the problem of translating signals into sym- bols preserving maximally the information contained in the signal time structure. In this context, we motivate the use of nonconver- gent dynamics for the signal to symbol translator. We then describe a biologically realistic model of the olfactory system proposed by Walter Freeman that has locally stable dynamics but is globally chaotic. We show how we can discretize Freeman”s model using digital signal processing techniques, providing an alternative to the more conventional Runge–Kutta integration. This analysis leads to a direct mixed-signal (analog amplitude/discrete time) implementa- tion of the dynamical building block that simplifies the implementa- tion of the interconnect. We present results of simulations and mea- surements obtained from a fabricated analog very large scale inte- gration (VLSI) chip. Keywords—Analog VLSI implementation, digital simulation models, neural assemblies, nonlinear dynamics. I. DESIGN OF SIGNAL TO SYMBOL TRANSLATORS There are many important differences between biological and man-made information processing systems. Animals have goal-driven behavior and have explored inductive principles throughout the course of evolution to work reliably in a non-Gaussian, nonstationary, nonlinear world. Autonomous man-made systems with sensors and compu- tational algorithms (animats) are still unable to match these capabilities. Engineered systems may be made faster, more accurate, but at the expense of specialization, which brings brittleness. Our present model of computation was inherited Manuscript received January 1, 2000; revised February 1, 2001. This work was supported in part by the Office of Naval Research under Grant N00014-01-1-0405 and by the National Science Foundation under Grant ECS-9900394. The work of V. G. Tavares was supported by Fundacao para a Ciencia e Tecnologia and Fundacao Luso-Americana para o Desenvolvimento scholarships. J. C. Principe, V. G. Tavares, and J. G. Harris are with the Computational NeuroEngineering Laboratory, University of Florida, Gainesville, FL 32611 USA. W. J. Freeman is with the Department of Molecular and Cell Biology, University of California, Berkeley, CA 94720 USA. Publisher Item Identifier S 0018-9219(01)05407-X. from Turing and Von Neumann and is based on the theory of formal systems [26]. The formal model is invaluable and quite general for the purpose of symbolic processing [22]. However, we have to remember that symbols do not exist in the real world. The real world provides time varying signals, usually faint signals corrupted by noise. Hence, the critical issue for accurate and robust interpretation of real-world signals by animats is not only how to process symbols but also how to transform signals into symbols. The processing device that transforms signals into symbols is called here the signal-to-symbols translator ( ). We can specify an optimal as a device that is able to capture the goal-rel- evant information contained in the signal time structure and map it with as little excess irrelevant information as possible to a stable representation in the animat’s computational framework. There is no generally accepted theory to optimally design for the unconstrained signals found in the real world. s fall between two very important areas of research: symbolic and signal processing, which use very diverse tools and techniques. Symbolic manipulation is unable to deal with time signals, while signal processing techniques are ill-pre- pared to deal with symbols. Many hybrid approaches using the minimum description length principle [40] and optimal signal processing principles [42], pattern recognition [14], neural networks [41], or other machine learning approaches [28] have been proposed. A framework where the s is modeled as a dynam- ical system coupled to the external world seems a produc- tive alternative. To bring a biological flavor, we will center the discussion in distributed, adaptive arrangements of non- linear processing elements (PEs) called coupled lattices [28]. The appeal of these coupled lattices is that they have po- tentially very rich autonomous dynamics and the external world signals can directly influence the dynamics as forcing inputs. Von Neumann was the first to propose cellular au- tomata, which are discrete-state and discrete-time coupled 0018–9219/01$10.00 © 2001 IEEE 1030 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001
Transcript
  • Design and Implementation of a BiologicallyRealistic Olfactory Cortex in Analog VLSI

    JOSE C. PRINCIPE, FELLOW, IEEE, VITOR G. TAVARES, JOHN G. HARRIS,ANDWALTER J. FREEMAN, FELLOW, IEEE

    Invited Paper

    This paper reviews the problem of translating signals into sym-bols preserving maximally the information contained in the signaltime structure. In this context, we motivate the use of nonconver-gent dynamics for the signal to symbol translator. We then describea biologically realistic model of the olfactory system proposed byWalter Freeman that has locally stable dynamics but is globallychaotic. We show how we can discretize Freeman”s model usingdigital signal processing techniques, providing an alternative to themore conventional Runge–Kutta integration. This analysis leads toa direct mixed-signal (analog amplitude/discrete time) implementa-tion of the dynamical building block that simplifies the implementa-tion of the interconnect. We present results of simulations and mea-surements obtained from a fabricated analog very large scale inte-gration (VLSI) chip.

    Keywords—Analog VLSI implementation, digital simulationmodels, neural assemblies, nonlinear dynamics.

    I. DESIGN OFSIGNAL TO SYMBOL TRANSLATORS

    There are many important differences between biologicaland man-made information processing systems. Animalshave goal-driven behavior and have explored inductiveprinciples throughout the course of evolution to workreliably in a non-Gaussian, nonstationary, nonlinear world.Autonomous man-made systems with sensors and compu-tational algorithms (animats) are still unable to match thesecapabilities. Engineered systems may be made faster, moreaccurate, but at the expense of specialization, which bringsbrittleness. Our present model of computation was inherited

    Manuscript received January 1, 2000; revised February 1, 2001. Thiswork was supported in part by the Office of Naval Research under GrantN00014-01-1-0405 and by the National Science Foundation under GrantECS-9900394. The work of V. G. Tavares was supported by Fundacaopara a Ciencia e Tecnologia and Fundacao Luso-Americana para oDesenvolvimento scholarships.

    J. C. Principe, V. G. Tavares, and J. G. Harris are with the ComputationalNeuroEngineering Laboratory, University of Florida, Gainesville, FL 32611USA.

    W. J. Freeman is with the Department of Molecular and Cell Biology,University of California, Berkeley, CA 94720 USA.

    Publisher Item Identifier S 0018-9219(01)05407-X.

    from Turing and Von Neumann and is based on the theoryof formal systems [26]. The formal model is invaluable andquite general for the purpose of symbolic processing [22].However, we have to remember that symbols do not exist inthe real world. The real world provides time varying signals,usually faint signals corrupted by noise. Hence, the criticalissue for accurate and robust interpretation of real-worldsignals by animats is not only how to process symbols butalso how to transform signals into symbols. The processingdevice that transforms signals into symbols is called herethe signal-to-symbols translator ( ). We can specify anoptimal as a device that is able to capture the goal-rel-evant information contained in the signal time structure andmap it with as little excess irrelevant information as possibleto a stable representation in the animat’s computationalframework.

    There is no generally accepted theory to optimally designfor the unconstrained signals found in the real world.

    s fall between two very important areas of research:symbolic and signal processing, which use very diverse toolsand techniques. Symbolic manipulation is unable to deal withtime signals, while signal processing techniques are ill-pre-pared to deal with symbols. Many hybrid approaches usingthe minimum description length principle [40] and optimalsignal processing principles [42], pattern recognition [14],neural networks [41], or other machine learning approaches[28] have been proposed.

    A framework where the s is modeled as a dynam-ical system coupled to the external world seems a produc-tive alternative. To bring a biological flavor, we will centerthe discussion in distributed, adaptive arrangements of non-linear processing elements (PEs) called coupled lattices [28].The appeal of these coupled lattices is that they have po-tentially very rich autonomous dynamics and the externalworld signals can directly influence the dynamics as forcinginputs. Von Neumann was the first to propose cellular au-tomata, which are discrete-state and discrete-time coupled

    0018–9219/01$10.00 © 2001 IEEE

    1030 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • lattices, as a paradigm for computation [52]. Grossberg [18]and later Hopfield [25] proposed dynamic content associa-tive memories (DCAMs) as the simplest model for s byanalogy with the Ising model of ferromagnetics. In order tocontrast this method with statistical inference, we will brieflydescribe DCAM operation now. DCAMs are dynamical sys-tems with point attractors in their dynamics that correspondto the memorized information. The input serves as an ini-tial condition to the system state, and the stable dynamicsrelax to the closest fixed point, so the dynamical system’sbasins of attraction provide a natural similarity measure. Al-though mimicking statistical CAMs, the principle of opera-tion of DCAMs are very different. DCAMs have limited ca-pacity and display many spurious fixed points [24], so theyare not directly usable as s. The reason DCAMs havea capacity limited to (a fraction of) the number of inputs isthat their weights parameterize in a one-to-one manner thecorrespondence between the system attractors and the inputpatterns. This is not only a problem with the Hopfield model,but is equally shared by recurrent neural networks used forgrammar inference (which have been proven Turing equiva-lent) [16], and other dynamic neural networks [8].

    Although many other researchers have investigateddynamical principles to design and implement informationprocessing systems (mainly in the biophysics [33], [37]and computational neuroscience communities [20]), thisline of research is still a niche compared with the statisticalapproach. We are slowly realizing that the limited repertoireof dynamical behavior (fixed points) implemented by theseDCAMs constrain their use as information processingdevices for signals that carry information in their timestructure. For instance, the point attractor has no dynamicalmemory (i.e., the system forgets all previous inputs whenit reaches the fixed point) while the dynamic memory ofthe limit cycle is constrained to the period; only chaoticsystems display long-term dynamic memory due to thesensitivity to initial conditions. This sensitivity carries theproblem of susceptibility to noise, but a possible solution isto utilize a chaotic attractor created by a dynamical systemwith singularities of at least second order (third-order ODE).A chaotic attractor is still a stable representation, mightexist in a high-dimensional space (much higher than thedimensionality of our three-dimensional world), and, moreimportantly, its dimensionality can be controlled by param-eters of the s. Forseeably, such systems are capable ofusing the inner structure of trajectories within the attractorfor very high information storage and rapid recall, but westill do not fully understand how to control the stability ofrecall, in particular, in the presence of noise.

    Amazingly, signals with very complex time structure (con-sistent with chaotic generators) are measured from mam-malian cortex during sensory processing [12]. Unlike ex-isting computational systems, brains and the sensory corticesembedded within them are highly unstable. They jump fromone quasi-stable state to the next at frame rates of 5–10/sunder central control. The “code” of each frame is a 2-D spa-tial pattern of amplitude modulation (AM) of an aperiodic(chaotic) carrier wave. The AM contents of each frame are

    Fig. 1. Proposed computer architecture for real-world signalinterpretation.

    selected by input and shaped by synaptic connection weightsembodying past learning. These frames are found in the vi-sual, auditory, somatic, and olfactory cortices. Having thesame “code,” they readily combine to form multisensory im-ages (Gestalts) that provides a landscape of basins and attrac-tors for learned classes of stimuli. Our thesis is that sensorycortices explore cooperatively the nonlinear dynamical be-havior of neurons yielding global dynamics that are capableof codifying into spatial patterns minute changes perceivedin the inputs. By creating “chaotic” dynamics, brains are off-setting their limitation of limited size with the richness of thecooperative behavior away from equilibrium conditions. Thisreasoning not only opens a new research direction to studycomputational models beyond symbolic processing, but alsocharts the construction of high-performance information pro-cessing systems for real-world signals.

    Our aim is to construct a that operates in accordancewith the neurodynamics of the cerebral cortex, and that hasthe sensitivity, selectivity, adaptiveness, speed, and toleranceof noise that characterizes human sensation. Fig. 1 shows ourproposed methodology to attack the problem of interpretingreal-world data. The processing chain is composed of a neu-romorphic processor followed by a Turing machine (digitalcomputer). The neuromorphic will accept signals andproduce symbols as outputs, and will be the focus of thispaper. The role of the neuromorphic is to serve as an in-terface between the infinite complexity of the real world andthe countable infinite capacities of conventional symbolic in-formation processors.

    The problem is how to specify the properties and controlthe organization of high dimensional nonlinear (chaotic) dy-namics to operate as an information processing system com-parable to cortical systems.

    II. FREEMAN’S COMPUTATIONAL MODEL

    In a series of seminal papers spanning more than 25 years[11], [14], Freeman has laid down the basis for a biologi-cally realistic computational model of the olfactory system.The brain is a vast pool of excitable cells with a dense inter-connectivity that can be modeled as a spatio-temporal latticeof nonlinear processing elements (PEs) of the form

    order (1)

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1031

  • Fig. 2. Static nonlinear function sigmoidal characteristic.

    where and are constants. Equation (1) presages Gross-berg’s additive model [19], and each PE is a second-orderordinary differential equation to take into consideration theindependent dynamics of the wave density for the actiondendrites and the pulse density for the parallel action ofaxons. No auto-feedback from the nonlinearity output

    is allowed. The operation at trigger zones isnonlinear (wave to pulse transformation) and is incorporatedin the model through , a static sigmoidalnonlinearity, which is nonsymmetric as opposed to the non-linearities used in neurocomputing. This asymmetry is animportant source of instability in the neural network becauseof asymmetric excitatory and inhibitory gain.is defined by (2) and plotted in Fig. 2 [12]. In the model,is fixed at all levels to

    if

    if

    (2)

    A set of inputs may be prewired for temporal processingexpressed by as in the gamma model [39]. Freeman’smodel is also an additive model because the topological pa-rameters and are independent of the input. The dy-namical parameters and are real-time constants of twodecaying eigenfunctions and are determined experimentallyfrom brain tissue [11]. This is also true for all the other topo-logical parameters. is an external input.

    If we follow the engineering literature, a neural networkwould be an arbitrary arrangement of these PEs. However,in the cortex there is a very tight relation between topolog-ical arrangement of neurons and function [44] that is totallymissed in neurocomputing. Freeman’s model incorporatesneurophysiological knowledge by modeling the functionalconnection among groups of neurons, called neural assem-blies. Although each neural assembly is built from thousandsto millions of neurons, the functional connections have beencatalogued in an hierarchy of topologies by Katchalsky [11]

    among others. As a tribute to his contributions, the topolo-gies are named K sets. The hierarchy levels are designatedby K0, KI, KII, KIII.

    A. The K0 Set

    The simplest structure in the hierarchy of sets is the K0 set.This set has the characteristic of having no functional inter-connections among the elements of the neural assembly. Theneural population in a K0 share common inputs and commonsign for the outputs as shown in Fig. 3. It can accept severalspatial inputs that are weighted and summed, and then con-volved with a linear time invariant system defined by the lefthand side of (1). The output state resulting from the con-volution is magnitude shaped by (2) to form the output. Wewill represent a K0 set by a square, as in Fig. 3. This is thebasic building block necessary for the construction of all theother higher level structures. A K0 network is a parallel ar-rangement of K0 sets.

    B. The KI Network

    The KI set is formed from K0 units interconnected throughlateral feedback of common sign; hence, we can have exci-tatory (denoted by sign in Fig. 4) and inhibitory KI sets(denoted by a sign). There is no auto-feedback. A numberof KI sets connected only by forward lateral feedback con-nections also comprises a K1 network. Each connection rep-resents a weight that can be constant or time varying andis obtained from neurophysiological measurements. Thesemodels are lumped circuit models of the dynamics.

    C. The KII Set and Network

    The KII set is built from at least two KI sets with densefunctional interconnections [11]. The most interesting inter-connectivity occurs when the KII set is formed by 2KI (or4 K0) sets with opposite polarity (i.e., two excitatory K0and two inhibitory) (Fig. 5). KII sets have fixed connectingweights, and they form oscillators with the parameter set-tings found in biology. The response measured at the outputof the excitatory K0 to an impulse applied to its input is ofa high-pass filter type (damped oscillation). However, with asustained input it evolves into undamped oscillation, whichvanishes after the input rests into a zero state. The KII setis then an oscillator controlled by the input with an intrinsicfrequency setup by the connection weights.

    When the KII sets are connected together they form a KIInetwork, which becomes a set of mutually coupled oscilla-tors. Note that the excitatory K0 of each set is connectedthrough weights to all the other excitatory K0 sets of thenetwork. Likewise for the inhibitory K0 sets. The excitatoryweights interconnecting the individual KII sets in the net-work are adapted with Hebbian learning, while the inhibitoryweights are fixed with a value obtained from physiologicalmeasurements [13].

    The input to the KII network is a vector of values, each ap-plied to the excitatory K0 (named M1 in Fig. 5) in the KII set.The output is spatially distributed over the network, as wediscuss next. When an external excitation is applied to one

    1032 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • Fig. 3. The K0 set and its components.

    Fig. 4. The KI network made up of an arrangement of K0 sets (Periglomular cells).

    Fig. 5. KII network made up of tetrads of K0 sets. M: Mitral cells. G: Granular cells.

    or more KII sets in the network, those sets tend to build upenergy on the other sets through excitatory (positive) inter-connections feedback. The oscillation magnitude of the setsthat do not have any excitation will depend on the magnitudeof the coupling gain defined by the interconnecting weights.If these interconnections are adapted to an input pattern, adistributed representation for that input is achieved. Whena known pattern is applied to the system input after training,then a recognizable output oscillation pattern emerges acrossthe KII sets. In this perspective, the KII network functionsas an associative memory [30] that associates an input bit

    pattern with the spatial distributed coupling that is storedin the interconnecting pathways, and that becomes “visible”from the outside as larger amplitude oscillations along theKII sets that have larger weights. The excitatory connec-tions represent cooperative behavior while inhibitory con-nections introduce competitiveness into the system. They arecrucial for contrast enhancement between the channels [13](a channel is an input or output taken from a single set). Foran “ON–OFF” (1/0) type of input pattern, the learning is ac-complished with a modified Hebbian rule as follows [13]. Ifbetween any two input channels the product is equal

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1033

  • Fig. 6. KIII network as a model for the olfactory system.

    to both “on” (1), then the weight between M1and M1 isincreased to a high value, otherwise its value is unchanged.At the beginning with no learned patterns, all these connec-tion weights are at a low value. Observe that the KII networkdoes not conform to traditional information processing sys-tems that possess well-defined inputs and outputs. Here, theoutput is spatially distributed and exists briefly in time.

    D. The KIII Network—A Model for the Olfactory System

    The KIII network is topologically depicted in Fig. 6 andis a functional model of the olfactory system. The input tothe model is a layer of excitactory neurons modeled as a KInetwork [periglomerular layer (PG)]. This layer projects tothe olfactory bulb (OB), which is modeled by a KII network(with zero baseline) of a large size (channels). The OBlayer sends outputs to a KII set (positive baseline) that rep-resents the anterior olfactory nucleus (AON) and to anotherKII set (negative baseline), the prepiriform cortex (PC), thatin turn send its output to the entorhinal cortex (EC) and backto the OB and AON layers. Therefore, the KIII network is avariant of the KII network set with several layers of KII basicelements connected through dispersive delays in Fig. 6.The dispersive delays arise from the fact that the axons thatinterconnect the different layers in the brain have differentthicknesses and lengths. As a consequence the action poten-tials received are dispersively delayed with respect to eachother. The resulting operation is a low-pass filtering of theaxon density pulses [49]. The way the overall network oper-ates is very similar to the KII network set; however, the dy-

    Fig. 7. Nonlinear cell embedded in a higher order system [F (s)is the Laplace transform of the left hand side of (3)].

    namics are qualitatively different. Because the intrinsic fre-quencies of oscillation of the KII sets in the different layersare incommensurable and the system is tightly coupled withdispersive delays, the different layers will never lock to thesame frequency and chaotic behavior arises.

    The best place to analyze the dynamic behavior of theoverall system is the state space of the OB layer. The inputswill effectively “pull” the system to predetermined regionsof the state space associated with learned input patterns. Thischaotic motion can be “read-out” as spatial AM patterns ofactivity over the KII network. This output has been shownto provide a realistic account for the EEG obtained with anarray of electrodes placed on the olfactory bulb [14], and alsomimics the behavior of electrophysiological measurementsmade on the visual cortex of monkeys.

    From the explanation of the KII network, we can see thatit is unlike any studied in information processing. On onehand, it is a set of coupled-oscillators in a time-space lat-tice that is locally stable but globally chaotic. When an inputis applied to the PEs, they oscillate with their characteristicfrequency, and the oscillation is spatially propagated. On theother hand, the coupling in space is dependent upon the con-nection weights that are learned, so the information is coded

    1034 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • Fig. 8. Discrete representation of a KI set feedback paths.

    in spatial amplitude patterns of quasi-sinusoidal waves. Al-though we still cannot fully characterize the properties of theKII network for information processing, the dynamical char-acteristics of each piece have been specified by Freeman andco-workers [50], so we can build a system with similar dy-namics in analog very large scale integration (VLSI). Thenext section will address the method of analysis and the im-plementation details.

    III. D ISCRETIZATION OF THEMODEL COMPONENTS

    Freeman’s model is normally simulated using Runge–Kutta integration [38] for each input, what is time con-suming and it is not amenable to simple modifications in thetopology. Here, we develop an equivalent digital model usingdigital signal processing (DSP) techniques at the systemlevel. This digitalization approach has been successfullyapplied to physical phenomena that accept a linear modelsuch as the vocal tract model in speech synthesis [40]. Fornonlinear systems there is no transformation that can mapphase space to a sampled time domain, like the conformalmapping from the to the domains [46]. However, withthe special structure of the K0 model (linear dynamicsfollowed by a static nonlinearity) the following method-ology to design equivalent discrete time linear systems wassuccessful in preserving the general dynamical behaviorof the continuous-time model. Similar methods have beenapplied to discretize the Hopfield model [3]. Equation (3)and Fig. 7 shows a general formal structure where a K0 setfits. The forcing input in the right-hand side of (3) representsthe static nonlinear connections from other similar K0 sets,which themselves might also receive nonlinear excitationfrom this given th K0 set. The differential equation inthe left side of (3) represents a linear time invariant (LTI)system.

    (3)

    The procedure that is proposed to discretize (3) decom-poses this equation as topology and dynamics. The dynamicscan be mapped with linear DSP techniques to the discretetime domain. The connecting structure between individualcells is preserved except that delays are incorporated in the

    Fig. 9. Impulse response at different taps of a fifth-order gammafilter.

    feedback pathways to avoid instantaneous time calculations.There are two basic problems with this methodology. First,it assumes that the continuous dynamical system is synchro-nized, which is unrealistic, but there are ways to mitigate thisshortcoming [3]. Second, the dynamics between the analogand discretized models may differ. For the Hopfield network,the terminal dynamics of both (analog and discrete) systemshave been shown equivalent [37]. However, the terminal dy-namics for the KII set are a limit cycle, instead of a fixedpoint. The inclusion of a delay in the feedback path couldchange the frequency of oscillation between the analog andthe digital counterpart, but the implementation discussed inSection IV avoids this problem.

    One obvious advantage that arises from the proposed ap-proach is that anetwork of dynamical componentscan im-plement the resulting discrete time model. This means thatadding or deleting features or new components to the systemis simple and does not imply rewriting the global state equa-tions [as long as they follow the rule defined in (3)]. The dig-ital system dynamics can also be probed at different compo-nents to understand the system better, and the resulting digitalsystem computes the output solutions in a sample by samplebasis for real time simulations. These characteristics are in-valuable to perform testing of the analog simulator explainedin the next section. An added advantage is that this method-ology has been easily incorporated into standard commercialneural network packages.

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1035

  • Fig. 10. Gamma filter structure.

    A. Discrete Network Implementation of the OlfactorySystem Model

    Each of the individual elements of the KIII model followsthe form of equation (3). The continuous-time instantaneousfeedback is mapped into a delayed feedback operator. Fig. 8illustrates the procedure for a KI set. Feedforward connec-tions need not to be delayed. The exception is the time dis-persion term , but since this time operation is linearthe procedure proposed in the previous section can still beapplied to the overall system with delays.

    IV. I MPULSE INVARIANCE TRANSFORMATION

    The simplest method to discretize the dynamics of (1) isto take an impulse invariant [36] transformation that samplesthe impulse response of the dynamical system and preservesits shape. The discrete time system that results by sampling(4) with a suitable sampling frequency is shownin (5)

    (4)

    (5)

    Since Freeman’s model dynamics have a low-passbehavior, the potential aliasing error that arises from theimpulse invariance transformation can be effectively con-trolled by decreasing . In the results to be presented, thesampling frequency was chosen 20 times larger thanthe maximum frequency pole of the Laplace characteristicequation of (2). The resulting difference equation forFreeman’s model is (6)

    (6)

    The function is the delayed sampled version of thecomposite forcing input on the right-hand side of (1) or (3).Hence, the impulse invariant transformation naturally incor-porates the delay required to avoid the instantaneous compu-tations when the topological part of the dynamical model (3)is discretized.

    Fig. 11. System identification (ID) procedure.

    A. Gamma Basis Decomposition

    Global optimization methods or backpropagation throughtime (BPTT) [49] could be used to automatically set the pa-rameters of the KIII network provided input and output sig-nals were available. However, this method would becomedifficult to implement due to the issue of stability. In fact,rewriting (6) as

    (7)

    and taking the transform, we obtain an infinite impulseresponse system with a pair of poles

    (8)

    During adaptation of and , the system could becomeunstable. Alternatively, one can decompose the impulse re-sponse (7) over the set of real exponentials, which has beencalled the gamma bases (also known as the gamma kernels)given by (9) [39], [8]

    (9)

    The impulse responses of the gamma kernels (a cascadeof low-pass filters of equal time constant) resembles many ofthe signals collected or modeled in neurophysiology (Fig. 9)[2]. The gamma kernels form a set of complete real basisfunctions, which means that any signal in (finite power)can be decomposed in gamma kernels with an arbitrarilysmall error [8]. The gamma kernels can be used as delayoperators in substitution to the ideal delay operator of DSP.Weighting the values of each gamma delay operator createsa generalized feedforward filter called the gamma filter [39],which is written as (10).

    1036 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • Fig. 12. Gamma basis approximation of the impulse response of (6).

    The vector in (10) is the projectionvector. The gamma filter block diagram is shown in Fig. 10

    (10)

    Decomposing (5) over the gamma basis follows the well-established system identification framework, as representedin Fig. 11 [17]. A white noise source is simultaneously fed tothe sampled differential equation (7) [denoted by ] to bemodeled and to the gamma filter . The outputs of thesetwo systems are subtracted and the resulting value is used asan error for adaptation of the gamma parameters. Adaptationalgorithms have been developed previously for the gammafilter parameters and the projection vector [8].

    The determination of the filter order for the case of (6)is relatively easy because we are decomposing the unknownsystem in the set of real exponentials. To gain a better under-standing, we will address the characteristics of the gammakernel next.

    B. Gamma Filter Stability, Resolution, and Depth ofMemory

    If we consider the transform of (9) for gamma ker-nels, we have a transfer function between the input and thelast gamma tap as

    (11)

    Therefore, the delay mechanism for the gamma filter isa cascade of identical first-order IIR sections (Fig. 10). Forstability, the gamma filter feedback parameter has to be re-stricted to . Yet, the condition is very easy to testsince is common to all kernels in the filter.

    Table 1Gamma Parameters

    The gamma filter has an infinite region of support, whichmeans that a low-order filter can model long time dependen-cies. The region of support of the impulse response is veryimportant for system identification, and for processing oftime information it implements a short-term memory mech-anism. Memory in the present context means how far in thepast are the dependencies to compute the present output. Thememory depth for the gamma filter with taps is [8]

    (12)

    The gamma filter is able to trade memory depth by resolu-tion. In fact, we can write , where is the orderof the filter, the memory depth given by (12), andis thetime resolution. In a FIR filter, the memory depth is the filterorder . From (12), we have the extra feedback parameterto change the memory depth of the gamma filter. However,there is a tradeoff between time resolution and memory depth,which is controlled by the order of the system. The resolutionis problem dependent and most of the time the tradeoff is re-solved heuristically. In our dynamical system modeling, wepicked the lowest order due to implementation constraints.

    C. Differential Equation Digitalization Results

    From biological evidence, the parametersand assumethe values 220/s and 720/s, respectively [11]. Since the polesare real in this case, we can obtain a good fit with a second-

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1037

  • Fig. 13. Delay response for different dispersive delays of the KIII model.

    order gamma filter. The sampling frequency was chosenas 20 times larger than the largest frequency pole. Fig. 12 rep-resents the impulse invariant system response to an impulse.The gamma filter approximation of the sampled differentialequation is also plotted. The resulting match is very good,although the approximation filter has only a double pole at

    . Table 1 summarizes the gamma filter parametersvalues found using the discussed system ID approach.

    D. Dispersive Delays Digitization

    As discussed earlier, the time dispersion was introduced inthe KIII network to model the different axonal interconnec-tion lengths and thicknesses. These functions can be modeledas in (13)

    (13)

    which can be rewritten in a convolution form as

    (14)

    where denotes the convolution and is the step func-tion. Equation (14) represents a FIR low-pass filter in dis-crete time and is recognized as a linear phase filterof Type I or II depending if the number of samplesis an even or odd number, respectively. A linear phase systemis nondispersive. To be more biologically realistic, the filtershould have nonlinear phase. A simple low-pass IIR systemlike a first-order gamma kernel has nonlinear phase and soimplements a more realistic dispersive delay.

    The gamma kernel has the added feature of allowing for ananalytical procedure to calculate the time dispersion. From(14), we may interpret as a memory structure withtime depth of . A gamma kernel has a depth ofmemory given by (12), so we may takeand find a with a given order of the system. However,as mentioned before, time resolution is decreased with thisdesign and may be a concern. In the present implementa-tion, a cascade of two dispersive delays was utilized

    . Following this procedure, we get the disper-sive delay response of Fig. 13.

    E. Digital Simulation Results

    After discretization, any software packages (DSP orneural networks) or even hardware simulators can beutilized in the simulation. The environment used for thesimulation task is a commercial neural network packagecalled “NeuroSolutions” [35]. The software has all the basicbuilding blocks necessary to construct the K sets and thedispersive delays. The user interface enabled the creationof the topologies of each set using an icon-based interface.We performed many tests on each K set, but here we willreport on the more interesting ones (forced responses).First, we show in Fig. 14 a comparison of the KII outputspectrum to a “high” input obtained using Runge–Kutta(ODE45 in Matlab [32]) and our modeling method witha sampling period of s. The KII systemresponse was simulated with an input square wave with anamplitude of six (arbitrary units) when excitation is present,and zero otherwise. The parameter set is summarized inTable 2 where K are the weights between the mitral (E)and granular cell (I) denoted by M and G in Fig. 5.

    We plot in Fig. 14 the magnitude spectrum of the oscil-lating time response using the fast Fourier transform (FFT)

    1038 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • Fig. 14. Comparison of KII output spectrum using Runge–Kutta and our discretization.

    Table 2KII Parameter Set

    Table 3Patterns Stored

    window method, with a 12 000 sample window and 36 av-erages. As we can observe, there is a perfect match betweenthe Runge–Kutta spectrum and our methodology.

    We have further simulated a channel KII networkand stored five patterns in the network according to the mod-ified Hebbian rule discussed earlier. The patterns are definedin Table 3. The results to be shown are for the recall of Pat-tern 5.

    An excitation drives the system into a limit cycle that lastswhile the input is applied. Fig. 15(a) shows the time responseof channel 19 (driven by a “1”) and Fig. 15(b) the limit cycleplotted between the excitactory versus the inhibitory states.When the input is applied, the KII state moves from the origin

    (a)

    (b)

    Fig. 15. Time series of KII with a square input and phase plot.

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1039

  • Fig. 16. KII 20 channel network response to input pattern 5 (Table 3) [channel response (“Input”)].

    to a large limit cycle centered away from the origin. It staysthere for the duration of the input, and when the system inputvanishes the output returns to zero.

    The phase plot resembles a Shilnikov singularity, but wehave not investigated further the local characteristics of theunstable point [45]. This result also indicates that the KII maybe stable in the sense of Lyapunov, i.e., there is a potentialfunction that models the system as energy conservative. Fi-nally, Fig. 16 shows the outputs of all the KII sets in the KIInetwork after the application of pattern 5 at the input (recallof pattern 5). The input pattern produces a large amplitudeoscillation in the channels that have 1 in the input pattern, sowe have the expected association between inputs and the AMoscillation that Freeman describes in the OB of the rabbit fol-lowing a sniff, which was also simulated in [10].

    One interesting property of this output is that it is transientin time, disappearing when the input vanishes. This behavioris very different from Hopfield networks, where the systemdynamics stay in the fixed point after the input is removed.Extra energy would have to be supplied to the system to getaway from the fixed point, while here the KII network is al-ways ready for real-time processing of temporal patterns.

    The KIII network was constructed in NeuroSolutions withonly channels in the OB layer because the simula-tion takes much longer. Table 4 summarizes the parameterset used for the simulation. All the results were generated forthe recall of pattern 2. The input was a smooth rising pulsewithin the values of 2 and 8 as specified in [49].

    In the full KIII system, we have many possible probepoints. We will start by analyzing the resting state at the PG,OB, AON, and PC layers. If we observe the time signals atany of these locations in the absence of input, we will seerandom fluctuations of a quasi-periodic activity that resem-

    Table 4Parameters Set for the KIII Network Simulations (see Fig. 6)

    bles the electroencephalogram’s background activity. Whenwe plot the activity as phase plots between the inhibitory andexcitatory states (Fig. 17), we see phase plots that resemblestrange attractors of different shapes across the four layers.Therefore, the KIII has a high-order nonconvergent dynam-ical behavior compatible with a chaotic attractor. We havenot made any tests for chaoticity yet, but a simple powerspectrum estimation of the time series generated by theAON (E1 state) clearly shows a type of spectrum that is

    1040 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • Fig. 17. Phase plots at different layers of the KIII network (no excitation).

    Fig. 18. FFT at the AON layer (no excitation).

    once again compatible with a self similar attractor (Fig. 18).So we conclude that the expected behavior simulated usingRunge–Kutta integration in [49] is again obtained in ourdiscretized model.

    Let us now turn to the OB layer and analyze the responseto a pulse input. Fig. 19 shows the time series for the recall ofpattern two. Notice that the response of the channels where

    the “high” level of the input was stored do indeed show a dcoffset during the time the pulse is applied. We were also ex-pecting an increase in oscillatory amplitude during the pulse,but the small size of the OB layer (only eight channels) mayexplain the difference.

    We investigated in more detail the structure of the oscilla-tions among the different channels of the OB layer. When we

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1041

  • Fig. 19. Time Series for Pattern 2 recall [channel (“Input”)].

    create the phase plot between the excitactory and inhibitorystate of channel 2, we clearly see [Fig. 20 (a)] that the basalattractor (for the “off” input) changes to a “wing” when theinput switches to a high amplitude. Channels that have astored pattern oscillate in perfect synchronization [Fig. 20 (b)and (c)] and share the same basal state evolution. This is alsoreported by Freeman [13]. However, channels belonging todifferent templates oscillate with different incommensurablephases [Fig. 20 (d)]. We remark that in the beginning of thesimulation they start in phase, but they loose synchrony veryquickly thereafter. Freeman states that this lack of coherenceis not apparent in the signals collected from the rabbit nor inanalog implementations of the model. He states that it corre-sponds to the breakdown of the chaotic attractor due possiblyto quantization errors that create limit cycles in phase space[14].

    Therefore, we conclude that when excited, the KIIIsystem undergoes a transition from the basal to a differentchaotic behavior characteristic of the pattern being recalled.When the excitation vanishes, the system always returns tothe initial attractor (Fig. 20). The “on” channels are distin-guishable by the higher offset and signal dynamics (Fig. 20).All these results are consistent with Freeman’s results usingRunge–Kutta numerical methods [49], while here we useour discretization of the linear part of the dynamics usingthe gamma approximation.

    V. ANALOG CIRCUIT IMPLEMENTATIONS

    The previous sections presented Freeman’s model of theolfactory system and its DSP implementation. The flexibility,immunity to noise, and modularity of the DSP implemen-tation is unsurpassed when compared to analog VLSI im-

    plementations. However, in the general framework of sdiscussed in the introduction, an analog implementation alsopresents advantages of its own, such as:

    • natural coupling to the analog real world;• computation is intrinsically parallel;• the computation time is practically instantaneous (only

    set by the delay imposed by each component);• there are no roundoff problems, i.e., the amplitude res-

    olution is effectively infinite;• it generally renders smaller circuits;• it is one order of magnitude more power efficient.

    These benefits motivated the development of an analogVLSI implementation, but unlike the traditional design ap-proach, we seek to preserve as much as possible the func-tional correspondence to the digital design to help us set inthe digital simulator the parameters of the analog KIII model.Otherwise, finding appropriate parameters in the analog ver-sion of the KIII becomes a daunting task. Our analog de-sign approach is centered on acontinuous amplitude and dis-crete timeimplementation of the dynamic building blocks(mixed-signal implementation) for the following reasons.

    • We wish to preserve a continuous amplitude representa-tion due to the negative effect of the roundoff errors re-ported in implementingchaoticdynamicalsystems[50].

    • The most serious design constraint in the VLSI imple-mentation of the KIII is the size of the interconnect. If afull analog (i.e., analog amplitude and continuous time)implementation is selected, the area required for theinterconnect would be prohibitive. Another notoriousadvantage of the mixed-signal solution is the controlgained over the timing of events. The synchronous re-sponse update at discrete times provide free time slots

    1042 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • Fig. 20. Behavior with input “OFF–ON–OFF” cycles.

    to implement other functions such as reading inputs,parameters, and transparent multiplexing with the con-sequent optimal resource management.

    • Discrete time implementation also brings important ad-vantages for testing since a one-to-one correspondencecan be established between the DSP simulation and thechip components. This is evident from the fact that bothhave the same mathematical representations, differenceequations, and transform. The difference resides ondimensionality that can easily be translated betweenboth domains by substitutingby .

    • Unexpectedly, time discretization also leads to a veryefficient implementation of the dynamics in the formof a novel circuit called Filter and Hold (F&H, patentpending) that decreases the size of the capacitors andresistors for a given time constant. The discretizationof the KIII system provides a clear roadmap for theanalog implementation. As was discussed above, theonly components required are gamma filters for thedynamics and dispersive delays, static nonlinearities,summing nodes, and the vast interconnect.

    As a conclusion, the high-level formal simulation pro-posed in earlier sections becomes the centerpiece for thediscrete-time analog implementation. We have access toa formal high-level representation that can, in a sampleby sample basis, predict how the real system will behave.The VLSI system can be flexibly tested and the parametersrefined with the digital simulator, which can be directlybrought onto the designed chip by external active control.

    Of course, this tight coupling between the mixed signaland the DSP implementations also creates extra difficultiesbecause the exact mathematical functions are seldom imple-mentable in VLSI. Hence, the design cycle should include aredefinition of the formal DSP system to include the smallalterations imposed by implementation limitations. If bothbehave similarly, the analog implementation will be a goodrepresentation of the original continuous model, since theprevious results showed that the digital model reproducesthe Runge–Kutta simulation well. This is the procedurefollowed in the next sections, where we describe each ofthe building blocks and present measurements from circuitsdesigned in MOSIS 1.2-m AMI technology using thesubthreshold CMOS regime for low power.

    A. Nonlinear Asymmetrical Sigmoidal Function

    The function responsible for the nonlinear characteristicof Freeman olfactory model is static and has the peculiarityof being asymmetric around the zero point (Fig. 2). A preciseimplementation of (2) is not an easy task, but we succeededin approximating the sigmoid with a modified operationaltransconductor amplifier (OTA). The important aspects ofthe original function are preserved, such as exponential sig-moidal shape and asymmetry around the input axis. The ex-ponential shape is set by placing the MOS transistor in thesubthreshold operating region, where it displays an exponen-tial relationship between the gate source voltage and the draincurrent. The asymmetry is accomplished by unmatching thecurrent mirror to act as an active charge to a differential input

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1043

  • Fig. 21. Approximated nonlinear function (left withN = 5; I = 1; V T = 1) and measuredresponses in the chip for two different bias currents (the bias current ofG is I = 600 nA).

    Fig. 22. Implemented asymmetric nonlinear function (voltage in/voltage out).

    stage. The output current relation with the input differentialvoltage can be represented by (15), which has the requiredshape as Fig. 21 shows

    (15)

    The offset that is evident from Fig. 21 (left panel) is pro-duced by the current mirror unbalancing; however, it can becancelled out by subtracting a voltage at the cell input equalto the offset value. This can be automatically done by placingfeedback around a similar cell and applying the output to theinverting terminal [48]. The residual offset will be only due tothe natural fabrication process. After the offset cancellation,the sigmoidal curve is shown in Fig. 21 (left panel). The finalschematic is shown in Fig. 22. The current is converted to avoltage with a amplifier and the OTA was implementedusing a wide-range topology to prevent the inversion of cur-rentat the input differential pair when the output voltage dropsbelow a certain value. Chip measurements with two different

    bias currents are shown in Fig. 21 (right panel), and the overallshape matches the circuit simulations reasonably well.

    Notice that the dynamic range of the DSP implementa-tion and the chip are different. However, in the chip we pre-serve the 1/5 ratio between the dc unbalance and the satu-ration level that is implemented by Freeman’s nonlinearitywith . We choose an nA to yield a 60-mVdynamic range, but different bias currents preserve the ratiowhile changing the gain and the saturating levels. We areonly concerned with the slope of the nonlinearity that did notmatch (2). This will impact the parameter setting between thetwo implementations, but for comparison purposes (15) wasalso programmed into the digital simulator.

    B. Dynamics—A Nanopower Analog Discrete Time Circuit

    The olfactory system is complex, highly interconnected,with many similar blocks repeated many times. The powerconsumption increases proportionally to the system order,meaning that effectively power consumption is a strict con-

    1044 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • Fig. 23. F&H implementation for the K0 dynamics.

    Fig. 24. F&H measured AC response together with predicted response by (16) (solid).

    straint for the implementation. In discrete time, there are ba-sically two filtering techniques available: the switched ca-pacitor (SC) [5] and switched current (SI) [26], but they bothutilize large areas and are not compatible with nanowatt con-sumption. A third alternative would use analog filtering witha sample and hold (S/H) at the input and output to imple-ment continuous time processing of a discrete time signal[36]; however, it does not bring any advantage when com-pared with the above two approaches.

    Small area and nanopower consumption usually go hand-in-hand. Simpler circuits take less elements and in generalwill require less power. We present an example of a novel dis-crete-time technique, that in essence is analog in magnitude,but uses time discretization in a very simple way to get verylong time constants with little area and low power. The tech-nique was named “filter and hold (F&H)” and it is a mixed

    analog/discrete time design that combines analog continuoustime conventional filtering with sampling [47], [48].

    The F&H technique is based on the very intuitive idea ofallowing a capacitor to integrate a current duringsecondsand halting it during a time of seconds. This process isrepeated every seconds. The resulting time constant valueis then multiplied by the duty cycle, which is defined as theratio of over the sampling period. The multiplicative ef-fect to the time constant is solely based on the duty-cycle ofthe clock and not on the sampling period. Therefore, largetime constants can be achieved without a decrease in sam-pling frequency or with high capacitor (or transistors in SIsystems) ratios. The switching scheme is simple and boththe area and power requirements can be made small. Furtherdetails about F&H are out of the scope of this paper; how-ever, we have shown that the procedure is general for any

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1045

  • Fig. 25. K0 model with a voltage summing node. (a) Excitactory and (b) inhibitory versions.

    Fig. 26. KII schematic (R andR ; j; i = E or I , correspond toK of Table 5).

    filter order and type, i.e., high-pass, low-pass, bandpass andband-reject filters for sampled input signals.

    The F&H circuit used for the K0 is shown in Fig. 23. Thedifference equation can easily be found taking the step re-sponse [47], and the corresponding frequency response be-comes

    (16)

    The very low frequency poles (35 Hz and 114 Hz) wererealized with a very low power implementation (100 nW)and with a fairly low area consumption of 200m 150

    m ( 1 pF, 4 pF), the duty cycle is 1%, whichmeans that the equivalent capacitors will be virtually scaledby 100 ( 100 pF, 400 pF). Fig. 24 shows thepredicted and measured frequency responses.

    C. K0 Cells

    At this point, the K0 model is almost complete. In orderto allow interconnectivity, an input summing node is needed.The summing node is a simple voltage adder, as representedin Fig. 25. Since the adder inverts the signal, the output non-linearity was changed to take into account the signal inver-sion. To build the higher level models, excitatory and in-hibitory K0 cells are needed. The voltage adder configura-

    Table 5KII–K0 Interconnections Gain in Analog and Digital Simulations

    tion inverts the input signals. In order to preserve the propersign between input and output, the nonlinear circuits weretopologically changed to reverse the signal polarity. Fig. 25shows conceptually the needed nonlinear function change toensure the proper signs. The circuit to perform these non-linear functions is a small variant of that of Fig. 22. Only theunbalancing transistors and input positions change.

    Each designed K0 takes an area of 500m 160 m andthe power consumption is roughly 300W to ensure drivingcapabilities in the summing amplifier of Fig. 25.

    1046 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • Fig. 27. Measured result of KII (E1) output for a pulse input.

    D. A Coupled KII–K0 Measured Results

    A KII set was implemented in analog VLSI with the com-ponents described in the previous sections. On the same chip,we also implemented several K0 sets and for the tests, weconnected one of the K0 sets to the KII as shown in Fig. 26.All the resistors are external, and for this setup we have twomore coupling parameters K and K linking the KII andthe K0. Our first experiment was to find the values of thecoupling coefficients that made sense from an analog circuitpoint of view (all the values in the digital simulations aredimensionless multipliers). For dynamic modeling, the ratioof resistors is what is important for the simulations, so it isthe quantity presented in Table 5. The feedback resistor is

    100 K .Due to the F&H implementation of the dynamics and

    differences in the nonlinearity, the coupling coefficients ofTable 2 did not provide the sought after dynamics for theKII, so a tuning of parameters was necessary. We present inTable 5 a parameter set that produced the expected dynam-ical behavior. This reparametrization was anticipated, butaccording to our methodology, we now need to make surethat we redefine the DSP model to match the new dynamicalbehavior.

    First, we measured the poles of the analog K0 to be110/s and 360/s. The bias current for the nonlinearitywas 40 nA, very close to the designed value. Next, wesubstituted in our digital implementation the gamma filtermodel by the F&H equation (16), and the original nonlin-

    earity by (15). With these pole values and the same couplingparameters of Table 5, our digital simulation produced theexpected dynamical behavior that we report next. Hence, wehave reenforced our conviction that the digital simulation canbe used as a flexible test bed to set analog system parameters;we just need to model all the components more accurately.

    The waveform measured in the VLSI chip with a forcinginput pulse is plotted in Fig. 27. Here, we plot the digitalinput and the measured response at the excitatory state (acmeasurement). This figure should be compared with Fig. 15obtained from the digital simulation.

    In Fig. 28, we show the response between the K0 and theexcitatory state of the KII set, for both the digital simulationand the waveforms collected from the chip.

    In Fig. 29, we select the KII inhibitory state to show theexpected out-of-phase response. We conclude that the newdigital simulation correctly follows the KII VLSI implemen-tation time evolution with forced inputs both for the phase,frequency, and relative magnitudes.

    We conclude that the major sources of error responsiblefor some observable differences are the F&H implementa-tion and the shape of the nonlinearity. The amplifierswere designed as a simple differential pair with emitter de-generating diode connected transistors to increase the linearinput range and diminish the value. It can be shown thatfor each stage in Fig. 23 when the switch is “on,” the differ-ential equation is [33]

    (17)

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1047

  • Fig. 28. Measured and simulated KII (E1 and G) time sequence and phase plot. All K0 cells oscillateat the same frequency and in phase.

    The nonlinear hyperbolic tangent function will have a re-shaping effect on the signals and will effectively limit thesignals’ magnitudes (the amplitude difference we see in thesimulations of Figs. 28 and 29). For a better matching be-tween the simple and flexible digital implementation and thereal chip implementation, this effect needs to be included inthe model or a more linear or other more linear F&Htopology should be designed.

    We are beginning to understand the effect of the amplitudenonlinearity on the overall dynamics, which is the missinglink to go back and forth between the parameterization of thedigital simulator and the analog implementation. ObservingTable 5, the ratio of K /K is smaller than the scale factorfor the other gains of Table 2. A more careful analysis ofFig. 21 reveals that the exact nonlinear function proposedby Freeman has a higher slope and consequently requires asmaller input for the same saturating output levels. This isthe gain factor responsible for such a parameter change. Infact, the normalization factor needed to match the slopesof the two functions is about . Furthermore, theimplemented nonlinearity has 25 mV, and a posi-tive saturating level of 80 mV. To scale up this function tothe 1 and 5 saturating levels used in the digital simula-tions of Freeman’s nonlinearity, we have to make in (15)

    , and the resulting normalizing factor becomes, where the numerator

    comes from the fact that Freeman’s nonlinearity has a ratioof five between the saturating levels. This scaled-up functionhas the same saturating levels as the one plotted in Fig. 21,but with different slope due to the different normalization of

    . If now we correct Freeman’s nonlinearity slopewith a normalizing factor of

    , and perform a digital simulation with the same condi-tions of Figs. 28 and 29, the same dynamics (including theoscillating frequency) are obtained except for the obviousscaled-up amplitudes.

    Other higher order effects, such as clock feedthrough [53]and extra delays due to finite bandwidth of all the neededcomponents, have not been accounted for in the digital repre-sentation. We conclude that our discretization methodologyis a good prediction of how the mixed-signal analog systembehaves dynamically, and will be a tremendous asset to theparametrization of future, more complex analog implemen-tations. But with the present silicon technology, these chipswill still have relatively few components to attack the prob-lems of representing real-world signals. One very promisingarea of research is to seek implementations with nanotech-nology components. They hold the promise of shrinking the

    1048 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • Fig. 29. Measured and simulated KII (I1 and G) time sequence and phase plot. All K0 cells oscillateat the same frequency but out of phase.

    size and bring new properties such as short-term memory,which is very difficult to implement in analog VLSI. Bi-ology can provide the insight for new computer architectures,and to take full advantages of the physics and chemistry ofmacromolecules, the language of dynamics should be uti-lized to describe the components of the next generation ofnanoinformation processing devices.

    VI. CONCLUSION

    The work described in this paper encompasses severaldisciplines, and it is a prototype of the issues that will befaced in creating animats possessing the exquisite informa-tion processing capabilities found in animals. We think thateffectively transferring information from real-world signalsto stable representations is a key link in this process thatrequires new insights, mixed analog/digital solutions and adynamic theory of information processing.

    Translating signals into internal representations is not aneasy task as the size of sensory cortices of mammals demon-strate. Understanding neurophysiology and computationalneuroscience is the starting point, but the description lan-guage and the proper level of abstraction are very importantto achieve a model that can be studied and implemented

    using the tools of engineering. Freeman’s model of theolfactory system shows how different the information pro-cessing solutions found in biology are from the conventionalengineering approaches. Pragmatically, we should extractfrom this model what is essential, and discount the require-ments of the “wetware.” Unfortunately, we still are far frombeing able to understand the model to make this distinction.First and foremost, we lack a dynamic theory of informationprocessing. Great strides and interest followed Hopfield’swork on dynamic associative memories, but his model onlyutilizes the simplest of the singularities: the fixed point.Recently, analog processors have been built for image pro-cessing using dynamical processing elements [7], and chaosis being applied in communication systems [1] and signalprocessing [21]. Freeman [12], Haken [20], and many othersare postulating nonconvergent dynamics for informationprocessing, but we need to understand and quantify what isgained by increasing the order of the singularities, and thedifficulties associated with chaotic attractors. The sensitiveto initial conditions of chaotic attractors bring dynamicmemory, but on the other hand require exactly the sameinputs to reach the same attractor, which is obviously impos-sible with the variability of real-world signals. The role ofnetwork noise here seems crucial to smooth out the attractor

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1049

  • landscape. The intriguing aspect is that chaos appears as acontrol “knob” that can fine-tune the sensitivity-specificitycurve of the system.

    From an engineering point of view analyzing, simulating,and building coupled lattices of nonlinear oscillators is achallenge. Albeit an approximation, our method of digitizingthe system dynamics is effective and yields digital modelsthat describe to a first approximation the complex dynamicbehavior. When the digital simulations are sufficiently accu-rate, we can do all this using digital computers. The issue isnot the Runge–Kutta integration precision, but the discretiza-tion of phase space. Contrary to what is claimed in somepapers [50], we do not see any constraint on using digitalcomputers, provided integer arithmetic is utilized in the sim-ulations. The bottleneck is the time to perform the simula-tions using integer arithmetic. Without a digital simulator, theparametrization of the analog model will be hopeless. Hence,further improvements to fully understand the nonlinear cou-pling are needed, as well as a mechanism to fine tune theparameters from data through adaptation.

    The analysis of the computational properties of the KIIImodel have barely started. We should quantify the capacity,the robustness to noise, and the speed of recall of the asso-ciative memory. For animats, we should also create a prin-cipled readout of information from the KIII model, some-thing that biology did not bother to create since the informa-tion is used internally by the animal. We should also under-stand how to go from a simple associative memory paradigmto higher order cortical functions needed for roaming, goaldriven behavior, and decision making. But the flexible archi-tecture of the KIII model already incorporates coupling fromcentral brain structures that will be implemented by the ani-mats’ symbolic processing stage.

    We showed that each piece of Freeman’s model is easyto implement, and the design has low power and small area.Our F&H circuit enables analog amplitude and discrete timeimplementations to facilitate interconnectivity. We are nowtesting a full KIII model using the components described inthis paper that will yield a first-generation .

    REFERENCES

    [1] H. Abarbanel,Analysis of Observed Chaotic Data. New York:Springer, 1996.

    [2] M. Abeles,Corticonics. Cambridge, U.K.: Cambridge Univ. Press,1991.

    [3] D. Amit, Modeling Brain Function. New York: Cambridge Univ.Press, 1989.

    [4] D. Ballard,An Introduction to Natural Computation. Cambridge,MA: MIT Press, 1997.

    [5] M. Beckerman,Adaptive Cooperative Systems. New York: Wiley,1997.

    [6] R. Brodersen, P. Gray, and D. Hodges, “MOS switched-capacitorfilters,” Proc. IEEE, vol. 67, pp. 61–75, Jan. 1979.

    [7] L. Chua,CNN: A Paradigm for Complexity. Singapore: World Sci-entific, 1998.

    [8] B. de Vries, “Temporal processing with neural networks: The de-velopment of the gamma model,” Ph.D. dissertation, Univ. Florida,1991.

    [9] D. Donoho, “Sparse components of images and optimal atomic de-composition,” Stanford Univ., Tech. Rep., 1998.

    [10] J. Eisenberg, W. Freeman, and B. Burke, “Hardware architecture of aneural network model simulating pattern recognition by the olfactorybulb,” Neural Networks, vol. 2, pp. 315–325, 1989.

    [11] W. Freeman,Mass Action in the Nervous System. New York: Aca-demic, 1975.

    [12] , “Tutorial on neurobiology from single neurons to brainchaos,”Int. J. Biforcation Chaos, vol. 2, no. 3, pp. 451–482, 1992.

    [13] W. Freeman, Y. Yao, and B. Burke, “Central pattern generating andrecognizing on olfactory bulb: A correlation learning rule,”NeuralNetw., vol. 1, pp. 277–288, 1988.

    [14] W. Freeman, “Mesoscopic neurodynamics: From neuron to brain,”J. Physiol., 2000, to be published.

    [15] K. Fukunaga,Introduction to Statistical Pattern Recognition. NewYork: Academic, 1990.

    [16] L. Giles, C. Miller, D. Chen, H. Chen, G. Sun, and Y. Lee, “Learningand extracting finite state automata with second order recurrentneural networks,”Neural Comput., vol. 4, pp. 393–405, 1992.

    [17] G. Goodwin and K. Sin,Adaptive Filtering Prediction and Con-trol. Englewood Cliffs, NJ: Prentice-Hall, 1984.

    [18] S. Grossberg, “Pattern learning by functional-differential neural net-works with arbitrary path weights,” inDelay and Functional Differ-ential Equations and Their Applications, Schmitt, Ed. New York:Academic, 1972, pp. 121–160.

    [19] S. Grossberg,Studies of Mind and Brain. Amsterdam, The Nether-lands: Reidel, 1982.

    [20] H. Haken,Principles of Brain Function New York, 1996.[21] S. Haykin and S. Puthusserypady, “Chaotic dynamics of sea clutter,”

    Chaos, vol. 7, no. 4, pp. 777–802, 1997.[22] J. Haugeland,Artificial Intelligence, the Very Idea. Cambridge,

    MA: MIT Press, 1975.[23] C. Helstrom, Elements of Signal Detection and Estima-

    tion. Englewood Cliffs, NJ: Prentice-Hall, 1995.[24] J. Hertz, A. Krogh, and R. Palmer,Introduction to the Theory of

    Neural Computation. Reading, MA: Addison-Wesley, 1991.[25] J. Hopfield, “Neurons with graded response have collective com-

    putational properties like those of two-state neurons,” inProc. Nat.Acad. Sci., vol. 81, 1984, pp. 3088–3092.

    [26] J. Hopcroft and J. Ullman,Introduction to Automata Theory, Lan-guages and Computation. Reading, MA: Addison-Wesley, 1979.

    [27] J. Hughes, N. Bird, and I. Macbeth, “Switched currents—A newtechnique for analog sampled data signal processing,” inProc. IEEEInt. Symp. Circuits and Systems, 1989, pp. 1584–1587.

    [28] K. Kaneko, Theory and Applications of Coupled Map Lat-tices. New York: Wiley, 1993.

    [29] M. Kearns and V. Vazirani,An Introduction to ComputationalLearning Theory. Cambridge, MA: MIT Press, 1994.

    [30] T. Kohonen, Self-Organization and Associative Memory. NewYork: Springer-Velag, 1988.

    [31] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient basedlearning applied to document recognition,”Proc. IEEE, vol. 86, pp.2278–2324, Nov. 1998.

    [32] Matlab Users Manual. Cambridge, MA: Mathworks, 1998.[33] C. Mead,Analog VLSI and Neural Systems. Reading, MA: Ad-

    dison-Wesley, 1989.[34] J. Mendel and K. Fu,Adaptive Learning and Pattern Recognition

    Systems. New York: Academic, 1970.[35] “Neurosolutions,” in Neurosolutions Manual. Ganesville, FL:

    NeuroDimensions Inc., 1995.[36] A. Oppenheim and R. Schafer,Discrete-Time Signal Pro-

    cessing. Englewood Cliffs, NJ: Prentice-Hall, 1989.[37] P. Peretto,An Introduction to the Modeling of Neural Net-

    works. Cambridge, U.K.: Cambridge Univ. Press, 1994.[38] W. Press, B. Flannery, S. Teukolsky, and W. Vetterling,Numerical

    recipes: The art of scientific computing. Cambridge, U.K.: Cam-bridge Univ. Press, 1986.

    [39] J. Principe, B. de Vries, and P. Oliveira, “The gamma filter a newclass of adaptive IIR filters with restricted feedback,”IEEE Trans.Signal Processing, vol. 41, no. 2, pp. 649–656, 1993.

    [40] L. Rabiner and R. Schafer,Digital Processing of Speech Sig-nals. Englewood Cliffs, NJ: Prentice-Hall, 1978.

    [41] J. Rissanen, Stochastic Complexity in StatisticalInquiry. Singapore: World Scientific, 1989.

    [42] F. Rosenblatt,Principles of Neurodynamics: Perceptrons and theTheory of Brain Mechanisms. Washington, DC: Spartan, 1962.

    [43] L. Scharf, Statistical Signal Processing, MA: Addison-Wesley,1991.

    [44] G. Shepard,The Synaptic Organization of the Brain. London,U.K.: Oxford Univ. Press, 1990.

    [45] L. Shilnikov, “Strange attractors and dynamical models,”J. CircuitsSyst. Comput., vol. 3, pp. 1–10, 1993.

    1050 PROCEEDINGS OF THE IEEE, VOL. 89, NO. 7, JULY 2001

  • [46] K. Steiglitz, “The equivalence of digital and analog signal pro-cessing,”Inform. Contr., vol. 8, pp. 455–467, 1965.

    [47] V. Tavares, J. Principe, and J. Harris, “F&H filter: A novel ultra-lowpower discrete time filter,”Electron. Lett., vol. 35, pp. 1226–1227,July 1999.

    [48] V. Tavares, J. Principe, and J. Harris, “A silicon olfactory bulb oscil-lator,” in IEEE Int. Symp. Circuits and Systems, vol. V, Switzerland,2000, pp. 397–400.

    [49] P. J. Werbos, “Back propagation through time: What it does and howto do it?,”Proc. IEEE, vol. 78, pp. 1550–1560, Oct. 1990.

    [50] Y. Yao and W. Freeman, “Model of biological pattern recognitionwith spatially chaotic dynamics,”Neural Netw., vol. 3, pp. 153–170,1990.

    [51] V. Vapnik, The Nature of Statistical Learning Theory. New York:Springer, 1995.

    [52] J. Von Neumann,Theory of Self-Reproducing Automata, Burks,Ed. Urbana: Univ. Illinois Press, 1966.

    [53] W. B. Wilson, H. Z. Massoud, E. J. Swanson, R. George Jr., andR. B. Fair, “Measurement and modeling of charge feedthrough inn-channel MOS analog switches,”IEEE J. Solid-State Circuits, vol.SC-20, pp. 1206–1213, Dec. 1985.

    Jose C. Principe(Fellow, IEEE) is a Professorof Electrical and Computer Engineering andBiomedical Engineering at the University ofFlorida, Gainesville, FL, where he teachesadvanced signal processing, machine learningand artificial neural networks (ANNs) modeling.He is a BellSouth Professor and the Founder andDirector of the University of Florida Computa-tional NeuroEngineering Laboratory (CNEL).His primary area of interest is processing of timevarying signals with adaptive neural models.

    The CNEL Lab has been studying pattern recognition principles basedon information theoretic criteria (entropy and mutual information), theuse of ANNs for dynamical modeling, and speech and object recognitionapplications. He is collaborating with neuroscientists in a computationalmodel of the olfactory cortex toward its analog VLSI implementation. Hehas more than 70 publications in refereed journals, ten book chapters, and160 conference papers. He directed 35 Ph.D. dissertations and 45 Masterthesis. He recently wrote an interactive electronic book entitledNeural andAdaptive Systems: Fundamentals Through Simulation(New York: Wiley).

    Dr. Principe is Secretary of the Technical Committee on Neural Networksof the IEEE Signal Processing Society, member of the Board of Governorsof the International Neural Network Society, and Editor in Chief of the IEEETRANSACTIONS ONBIOMEDICAL ENGINEERING. He is a member of the Ad-visory Board of the University of Florida Brain Institute.

    Vitor G. Tavares received the Licenciatura andM.S. degrees in electrical engineering from theUniversity of Aveiro, Portugal, in 1991 and 1994.

    He is presently completing the Ph.D. degreein electrical engineering from the ComputationalNeuroEngineering Laboratory at the Universityof Florida, Gainesville. His interests are in thearea of neuromorphic chip design and biomimeticcomputing.

    John G. Harris received the B.S. and M.S.degrees in electrical engineering from theMassachusetts Institute of Technology (MIT),Cambridge, in 1983 and 1986, respectively, andthe Ph.D. degree in computation and neural sys-tems from the California Institute of Technologyin 1991. After a two-year post doc at the MIT Ar-tificial Intelligence Lab, he joined the Electricaland Computer Engineering Department at theUniversity of Florida, Gainesville, in 1993. Hecurrently leads the UF Hybrid Signal ProcessingGroup in researching analog circuits and DSP

    for sensory processing.Dr. Harris is the recipient of an NSF CAREER Award as well as the

    Teaching Improvement Program award at the University of Florida.

    Walter J. Freeman (Fellow, IEEE) studiedphysics and mathematics at the MassachusettsInstitute of Technology, English and philosophyat the University of Chicago, medicine at YaleUniversity (M.D. 1954), internal medicine atJohns Hopkins University, and neurophysiologyat the University of California, Los Angeles.

    He has taught brain science at the University ofCalifornia, Berkeley, CA, since 1959, where he isProfessor of the Graduate School. He is the authorof Mass Action in the Nervous System(1975),So-

    cieties of Brains(1995),How Brains Make Up Their Minds(1999), andNeu-rodynamics: An Exploration of Mesoscopic Brain Dynamics(2000).

    PRINCIPEet al.: DESIGN AND IMPLEMENTATION OF OLFACTORY CORTEX 1051


Recommended