programmable optoelectronic multiprocessors and their comparison with symbolic...

14
Programmable optoelectronic multiprocessors and their comparison with symbolic substitution for digital optical computing F. Kiamilev Sadik C. Esener, MEMBER SPIE R. Paturi Y. Fainman P. Mercier C. C. Guest, MEMBER SPIE Sing H. Lee, MEMBER SPIE University of California, San Diego Electrical and Computer Engineering Department La Jolla, California 92093 Abstract. This paper introduces programmable arrays of optically inter- connected electronic processors and compares them with conventional symbolic substitution (SS) systems. The comparison is made on the basis of computational efficiency, speed, size, energy utilization, programma- bility, and fault tolerance. The small grain size and space -invariant con- nections of SS lead to poor computational efficiency, difficult program- ming, and difficult incorporation of fault tolerance. Reliance on optical gates as its fundamental building elements is shown to give poor energy utilization. Programmable optoelectronic multiprocessor (POEM) systems, on the other hand, provide the architectural flexibility for good compu- tational efficiency, use an energy- efficient combination of technologies, and support traditional programming methodologies and fault tolerance. Although the inherent clock speed of POEM systems is slower than that of SS systems, for most problems they will provide greater computational throughput. This comparison does not take into account the recent ad- dition of crossover interconnect and space- variant masks to the SS ar- chitecture. Subject terms: digital optical computing; symbolic substitution; optical intercon- nections. Optical Engineering 28(4), 396 -409 (April 1989). CONTENTS 1. Introduction 2. Summary descriptions of programmable optoelectronic multiprocessor (POEM) and symbolic substitution (SS) 2.1. POEM architecture 2.1.1. Architecture description 2.1.2. Implementation 2.2. SS -based computing systems 2.2.1. Architecture description 2.2.2. Optical implementations of SS systems 2.2.2.1. Description of a simple recognition - substitution module 2.2.2.2. Implementation of image shifting and combining operations 2.2.2.3. Data encoding schemes 3. SS systems and 2 -D VLSI mesh 3.1. Simulation of SS by a mesh 3.2. Simulation of a mesh by SS Invited paper OC -113 received Aug. 10, 1988; revised manuscript received Jan. 12, 1989; accepted for publication Jan. 16, 1989. Portions of this paper were presented at the OSA Annual Meeting, Oct. 30 -Nov. 4, 1988, Santa Clara, Calif. © 1989 Society of Photo -Optical Instrumentation Engineers. 396 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 4. System and technological considerations of POEM and SS 4.1. Fundamental considerations for optical gate arrays 4.2. Energy dissipation and latency for POEM and SS 4.2.1. POEM 4.2.2. SS 5. Relative merits of POEM and SS 5.1. Computational efficiency 5.1.1. Speed and size 5.1.2. Complex and multiple rules 5.2. Architectural considerations 5.3. Other considerations 5.3.1. Local communication 5.3.2. Random -access memory implementation and programming methodologies 5.3.3. Fault tolerance 6. Conclusions 7. Acknowledgments 8. References 1. INTRODUCTION The planar nature of electronic very large scale integration (VLSI) technology imposes limits on parallel electronic computing in- terconnect latency and area. t Free -space optically interconnected processing elements (PEs) offer an opportunity to remove this Programmable optoelectronic multiprocessors and their comparison with symbolic substitution for digital optical computing F. Kiamilev Sadik C. Esener, MEMBER SPIE R. Paturi Y. Fainman P. Mercier C. C. Guest, MEMBER SPIE Sing H. Lee, MEMBER SPIE University of California, San Diego Electrical and Computer Engineering Department La Jolla, California 92093 Abstract. This paper introduces programmable arrays of optically inter- connected electronic processors and compares them with conventional symbolic substitution (SS) systems. The comparison is made on the basis of computational efficiency, speed, size, energy utilization, programma- bility, and fault tolerance. The small grain size and space-invariant con- nections of SS lead to poor computational efficiency, difficult program- ming, and difficult incorporation of fault tolerance. Reliance on optical gates as its fundamental building elements is shown to give poor energy utilization. Programmable optoelectronic multiprocessor (POEM) systems, on the other hand, provide the architectural flexibility for good compu- tational efficiency, use an energy-efficient combination of technologies, and support traditional programming methodologies and fault tolerance. Although the inherent clock speed of POEM systems is slower than that of SS systems, for most problems they will provide greater computational throughput. This comparison does not take into account the recent ad- dition of crossover interconnect and space-variant masks to the SS ar- chitecture. Subject terms: digital optical computing; symbolic substitution; optical intercon- nections. Optical Engineering 28(4), 396-409 (April 1989). CONTENTS 1. Introduction 2. Summary descriptions of programmable optoelectronic multiprocessor (POEM) and symbolic substitution (SS) 2.1. POEM architecture 2.1.1. Architecture description 2.1.2. Implementation 2.2. SS-based computing systems 2.2.1. Architecture description 2.2.2. Optical implementations of SS systems 2.2.2.1. Description of a simple recognition- substitution module 2.2.2.2. Implementation of image shifting and combining operations 2.2.2.3. Data encoding schemes 3. SS systems and 2-D VLSI mesh 3.1. Simulation of SS by a mesh 3.2. Simulation of a mesh by SS Invited paper OC-113 received Aug. 10, 1988; revised manuscript received Jan. 12, 1989; accepted for publication Jan. 16, 1989. Portions of this paper were presented at the OSA Annual Meeting, Oct. 30-Nov. 4, 1988, Santa Clara, Calif. © 1989 Society of Photo-Optical Instrumentation Engineers. 4. System and technological considerations of POEM and SS 4.1. Fundamental considerations for optical gate arrays 4.2. Energy dissipation and latency for POEM and SS 4.2.1. POEM 4.2.2. SS 5. Relative merits of POEM and SS 5.1. Computational efficiency 5.1.1. Speed and size 5.1.2. Complex and multiple rules 5.2. Architectural considerations 5.3. Other considerations 5.3.1. Local communication 5.3.2. Random-access memory implementation and programming methodologies 5.3.3. Fault tolerance 6. Conclusions 7. Acknowledgments 8. References 1. INTRODUCTION The planar nature of electronic very large scale integration (VLSI) technology imposes limits on parallel electronic computing in- terconnect latency and area. l Free-space optically interconnected processing elements (PEs) offer an opportunity to remove this 396 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Upload: others

Post on 10-Mar-2020

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

Programmable optoelectronic multiprocessors and theircomparison with symbolic substitution for digital opticalcomputing

F. KiamilevSadik C. Esener, MEMBER SPIER. PaturiY. FainmanP. MercierC. C. Guest, MEMBER SPIESing H. Lee, MEMBER SPIE

University of California, San DiegoElectrical and Computer Engineering

DepartmentLa Jolla, California 92093

Abstract. This paper introduces programmable arrays of optically inter-connected electronic processors and compares them with conventionalsymbolic substitution (SS) systems. The comparison is made on the basisof computational efficiency, speed, size, energy utilization, programma-bility, and fault tolerance. The small grain size and space -invariant con-nections of SS lead to poor computational efficiency, difficult program-ming, and difficult incorporation of fault tolerance. Reliance on opticalgates as its fundamental building elements is shown to give poor energyutilization. Programmable optoelectronic multiprocessor (POEM) systems,on the other hand, provide the architectural flexibility for good compu-tational efficiency, use an energy- efficient combination of technologies,and support traditional programming methodologies and fault tolerance.Although the inherent clock speed of POEM systems is slower than thatof SS systems, for most problems they will provide greater computationalthroughput. This comparison does not take into account the recent ad-dition of crossover interconnect and space- variant masks to the SS ar-chitecture.

Subject terms: digital optical computing; symbolic substitution; optical intercon-nections.

Optical Engineering 28(4), 396 -409 (April 1989).

CONTENTS1. Introduction2. Summary descriptions of programmable optoelectronic

multiprocessor (POEM) and symbolic substitution (SS)2.1. POEM architecture

2.1.1. Architecture description2.1.2. Implementation

2.2. SS -based computing systems2.2.1. Architecture description2.2.2. Optical implementations of SS systems

2.2.2.1. Description of a simple recognition -substitution module

2.2.2.2. Implementation of image shifting andcombining operations

2.2.2.3. Data encoding schemes3. SS systems and 2 -D VLSI mesh

3.1. Simulation of SS by a mesh3.2. Simulation of a mesh by SS

Invited paper OC -113 received Aug. 10, 1988; revised manuscript received Jan.12, 1989; accepted for publication Jan. 16, 1989. Portions of this paper werepresented at the OSA Annual Meeting, Oct. 30 -Nov. 4, 1988, Santa Clara,Calif.© 1989 Society of Photo -Optical Instrumentation Engineers.

396 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

4. System and technological considerations of POEM and SS4.1. Fundamental considerations for optical gate arrays4.2. Energy dissipation and latency for POEM and SS

4.2.1. POEM4.2.2. SS

5. Relative merits of POEM and SS5.1. Computational efficiency

5.1.1. Speed and size5.1.2. Complex and multiple rules

5.2. Architectural considerations5.3. Other considerations

5.3.1. Local communication5.3.2. Random -access memory implementation and

programming methodologies5.3.3. Fault tolerance

6. Conclusions7. Acknowledgments8. References

1. INTRODUCTIONThe planar nature of electronic very large scale integration (VLSI)technology imposes limits on parallel electronic computing in-terconnect latency and area. t Free -space optically interconnectedprocessing elements (PEs) offer an opportunity to remove this

Programmable optoelectronic multiprocessors and their comparison with symbolic substitution for digital optical computing

F. KiamilevSadik C. Esener, MEMBER SPIER. PaturiY. FainmanP. MercierC. C. Guest, MEMBER SPIE

Sing H. Lee, MEMBER SPIE University of California, San Diego Electrical and Computer Engineering

Department La Jolla, California 92093

Abstract. This paper introduces programmable arrays of optically inter­ connected electronic processors and compares them with conventional symbolic substitution (SS) systems. The comparison is made on the basis of computational efficiency, speed, size, energy utilization, programma- bility, and fault tolerance. The small grain size and space-invariant con­ nections of SS lead to poor computational efficiency, difficult program­ ming, and difficult incorporation of fault tolerance. Reliance on optical gates as its fundamental building elements is shown to give poor energy utilization. Programmable optoelectronic multiprocessor (POEM) systems, on the other hand, provide the architectural flexibility for good compu­ tational efficiency, use an energy-efficient combination of technologies, and support traditional programming methodologies and fault tolerance. Although the inherent clock speed of POEM systems is slower than that of SS systems, for most problems they will provide greater computational throughput. This comparison does not take into account the recent ad­ dition of crossover interconnect and space-variant masks to the SS ar­ chitecture.

Subject terms: digital optical computing; symbolic substitution; optical intercon­ nections.

Optical Engineering 28(4), 396-409 (April 1989).

CONTENTS1. Introduction2. Summary descriptions of programmable optoelectronic

multiprocessor (POEM) and symbolic substitution (SS)2.1. POEM architecture

2.1.1. Architecture description2.1.2. Implementation

2.2. SS-based computing systems2.2.1. Architecture description2.2.2. Optical implementations of SS systems

2.2.2.1. Description of a simple recognition- substitution module

2.2.2.2. Implementation of image shifting and combining operations

2.2.2.3. Data encoding schemes3. SS systems and 2-D VLSI mesh

3.1. Simulation of SS by a mesh3.2. Simulation of a mesh by SS

Invited paper OC-113 received Aug. 10, 1988; revised manuscript received Jan.12, 1989; accepted for publication Jan. 16, 1989. Portions of this paper werepresented at the OSA Annual Meeting, Oct. 30-Nov. 4, 1988, Santa Clara,Calif.© 1989 Society of Photo-Optical Instrumentation Engineers.

4. System and technological considerations of POEM and SS4.1. Fundamental considerations for optical gate arrays4.2. Energy dissipation and latency for POEM and SS

4.2.1. POEM4.2.2. SS

5. Relative merits of POEM and SS5.1. Computational efficiency

5.1.1. Speed and size5.1.2. Complex and multiple rules

5.2. Architectural considerations5.3. Other considerations

5.3.1. Local communication5.3.2. Random-access memory implementation and

programming methodologies5.3.3. Fault tolerance

6. Conclusions7. Acknowledgments8. References

1. INTRODUCTIONThe planar nature of electronic very large scale integration (VLSI) technology imposes limits on parallel electronic computing in­ terconnect latency and area. l Free-space optically interconnected processing elements (PEs) offer an opportunity to remove this

396 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 2: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORSAND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

limitation by providing interconnections in three dimensions.2We describe here general- purpose computing systems currentlyunder investigation at the University of California, San Diegothat integrate optoelectronic PEs and free -space programmableoptical interconnects. These systems combine the advantages ofefficient processing abilities of silicon technology and program-mable global communication provided by optical interconnects.We call these systems programmable optoelectronic multipro-cessor (POEM) systems.5

To place the characteristics of POEMs in context, we willcompare them with an alternative general- purpose optical com-puting system based on symbolic substitution (SS) that has beenpresented by Huang et aí.6'7 and Kozaitis.8 Both POEM and SSare being proposed for achieving high performance, general-purpose, and parallel computing. In this paper we examine theperformance potentials and technological limits of these twosystems. The evaluation of these systems is based on their abilityto implement various algorithms efficiently, the power and arearequirements of existing and projected technologies to implementthem, fault tolerance, and ease of programming.

Section 2 provides architectural descriptions as well as ex-ample implementations of POEM and SS systems. In Sec. 3 weestablish the computational equivalence of SS systems to a 2 -Dmesh of VLSI processors. Technological considerations are dis-cussed in Sec. 4, including system size, speed, and energy dis-sipation. In Sec. 5 the relative merits of POEM and SS systemsare compared. Section 6 presents our conclusions.

2. SUMMARY DESCRIPTIONS OF POEM AND SSIn this section we describe briefly the architectures and funda-mental features of POEM and SS. Specific characteristics im-portant for the comparison of the systems are emphasized.

2.1. POEM architecture

2.1.1. Architecture descriptionPOEM systems have a highly parallel architecture based on waferscale integration of optoelectronic PEs and reconfigurable free -space optical interconnects. The POEM machine can be realizedwith an integrated optoelectronic technology, such as silicon/PLZT9'1° for the PE arrays, and dichromated gelatin as thevolume holographic storage medium for the interconnects. ThePOEM architecture can be extended to be reprogrammable orreconfigurable using a real -time volume holographic mediumsuch as photorefractive crystals.

The POEM architecture uses electrical interconnects for localcommunication within a PE and holographic optical intercon-nects for global communication among PEs. As shown in Ref 11,for interconnections longer than a certain break -even length,free -space holographic optical interconnects consume less en-ergy and are faster than their electrical counterparts. Also, free -space interconnects are immune to the crossover constraints ofplanar electronic technology, allowing denser interconnectiontopologies. Furthermore, they release space in the processingplanes used for interconnects, allowing more silicon circuitryon the wafer. The POEM machines use light modulators asoptical transmitters. Compared with active light sources such aslasers or light- emitting diodes, light modulators are attractivebecause they may be easier to integrate with silicon and becausethey dissipate less power on -wafer since electrical to opticalconversion power is dissipated off -wafer. This also allows on-

CONTROL BEAM I

3-DEMORY

OPTICAL NONLINEARMATERIAL FOR

PROGRAMMABLEINTERCONNECTION

CLOCK PULSES ANDCONTROL INSTRUCTIONS

READING BEAMOR OUTPUT FROM OPTICAL MEMORY

CAR y

OPTICALPROCESSING ARRAY

SI ( PHOTODETECTORTHRESHOLD, ALU.(

OPTICALMEMORYARRAY

E -0 MODULATORCONTROLBEAM 2

(a)

(b)

SI(MEMORYELEMENTS)

E -0 MODULATOR

Fig. 1. (a) POEM architecture. (b) Optoelectronic processing ele-ment array.

wafer power dissipation to be independent of the fan -out of theprocessor communication network if electro -optic light modu-lators are used.

The POEM architecture can support any variation of the pa-rameters commonly used to classify parallel architectures: gran-ularity (fine, coarse, or large grain), synchrony [single instruc-tion stream - multiple data stream (SIMD) or multiple instructionstream - multiple data stream (IMD)], and topology. The strengthof POEM machines comes from their efficient implementationof interconnections and the large degree of parallelism and con-nectivity that is inherent in free -space programmable global op-tical interconnections.

2.1.2. ImplementationAs an example, we describe a fine -grain POEM machine [Fig. 1(a)]containing a very large number (100,000 or more) of simpleone -bit silicon processors. An optoelectronic controller, con-nected to a sequential host computer, is used to optically broad-cast the instruction stream and master clock through a computer -generated hologram to the PEs for SIMD processing. The globalinterprocessor communication in POEM is implemented by ac-tivating different interconnection holograms in a volume holo-graphic material of large storage capacity, such as dichromatedgelatin. Each interconnection hologram is recorded with a dif-ferent random phase code. These holograms can be activatedindependently at speeds compatible with the system clock rateby displaying the appropriate random phase code on a smallspatial light modulator. Therefore, unlike conventional parallelsystems, there are no limitations from fixed interconnection to-

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 397

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORS AND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

limitation by providing interconnections in three dimensions. 2^ We describe here general-purpose computing systems currently under investigation at the University of California, San Diego that integrate optoelectronic PEs and free-space programmable optical interconnects. These systems combine the advantages of efficient processing abilities of silicon technology and program­ mable global communication provided by optical interconnects. We call these systems programmable optoelectronic multipro­ cessor (POEM) systems. 5

To place the characteristics of POEMs in context, we will compare them with an alternative general-purpose optical com­ puting system based on symbolic substitution (SS) that has been presented by Huang et al. 6 ' 7 and Kozaitis. 8 Both POEM and SS are being proposed for achieving high performance, general- purpose, and parallel computing. In this paper we examine the performance potentials and technological limits of these two systems. The evaluation of these systems is based on their ability to implement various algorithms efficiently, the power and area requirements of existing and projected technologies to implement them, fault tolerance, and ease of programming.

Section 2 provides architectural descriptions as well as ex­ ample implementations of POEM and SS systems. In Sec. 3 we establish the computational equivalence of SS systems to a 2-D mesh of VLSI processors. Technological considerations are dis­ cussed in Sec. 4, including system size, speed, and energy dis­ sipation. In Sec. 5 the relative merits of POEM and SS systems are compared. Section 6 presents our conclusions.

2. SUMMARY DESCRIPTIONS OF POEM AND SS

In this section we describe briefly the architectures and funda­ mental features of POEM and SS. Specific characteristics im­ portant for the comparison of the systems are emphasized.

2.1. POEM architecture

2.1.1. Architecture descriptionPOEM systems have a highly parallel architecture based on wafer scale integration of optoelectronic PEs and reconfigurable free- space optical interconnects. The POEM machine can be realized with an integrated optoelectronic technology, such as silicon/ PLZT9' 10 for the PE arrays, and dichromated gelatin as the volume holographic storage medium for the interconnects. The POEM architecture can be extended to be reprogrammable or reconfigurable using a real-time volume holographic medium such as photorefractive crystals.

The POEM architecture uses electrical interconnects for local communication within a PE and holographic optical intercon­ nects for global communication among PEs. As shown in Ref 11, for interconnections longer than a certain break-even length, free-space holographic optical interconnects consume less en­ ergy and are faster than their electrical counterparts. Also, free- space interconnects are immune to the crossover constraints of planar electronic technology, allowing denser interconnection topologies. Furthermore, they release space in the processing planes used for interconnects, allowing more silicon circuitry on the wafer. The POEM machines use light modulators as optical transmitters. Compared with active light sources such as lasers or light-emitting diodes, light modulators are attractive because they may be easier to integrate with silicon and because they dissipate less power on-wafer since electrical to optical conversion power is dissipated off-wafer. This also allows on-

OPTICAL NONLINEARMATERIAL FOR

PROGRAMMABLEINTERCONNECTION

CLOCK PULSES AND CONTROL INSTRUCTIONS

E-0 MODULATOR E-0 MODULATOR

(a)

(b)

Fig. 1. (a) POEM architecture, (b) Optoelectronic processing ele­ ment array.

wafer power dissipation to be independent of the fan-out of the processor communication network if electro-optic light modu­ lators are used.

The POEM architecture can support any variation of the pa­ rameters commonly used to classify parallel architectures: gran­ ularity (fine, coarse, or large grain), synchrony [single instruc­ tion stream-multiple data stream (SIMD) or multiple instruction stream-multiple data stream (MIMD)], and topology. The strength of POEM machines comes from their efficient implementation of interconnections and the large degree of parallelism and con­ nectivity that is inherent in free-space programmable global op­ tical interconnections.

2.7.2. ImplementationAs an example, we describe a fine-grain POEM machine [Fig. l(a)] containing a very large number (100,000 or more) of simple one-bit silicon processors. An optoelectronic controller, con­ nected to a sequential host computer, is used to optically broad­ cast the instruction stream and master clock through a computer- generated hologram to the PEs for SIMD processing. The global interprocessor communication in POEM is implemented by ac­ tivating different interconnection holograms in a volume holo­ graphic material of large storage capacity, such as dichromated gelatin. Each interconnection hologram is recorded with a dif­ ferent random phase code. These holograms can be activated independently at speeds compatible with the system clock rate by displaying the appropriate random phase code on a small spatial light modulator. Therefore, unlike conventional parallel systems, there are no limitations from fixed interconnection to-

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 397

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 3: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

KIAMILEV, ESENER, PATURI, FAINMAN, MERCIER, GUEST, LEE

pology among the processors. Instead, the programmable opticalinterconnects are determined by the optoelectronic controller.Therefore, the programmer can implement a topology that bestmatches the current algorithm. In addition, the interconnectionstorage capacity requirement on the holographic material can bereduced if real -time reprogrammable material requirements canbe added. For example, one may envision using photorefractivecrystals or other nonlinear optical materials to apply repro -grammable interconnects to the PEs. In this case, the user willbe capable of reconfiguring the POEM in a very short time tomatch his algorithmic requirements.

The internal data paths of the PEs are implemented electricallyas in a common electronic processor. Each PE has the capabilityto perform logic, conditional execution, data movement, andI/O operations [Fig. 1(b)]. Also, each PE has some local ran-dom- access memory (RAM) to support the conventional pro-gramming models. In general, the grain size of the PEs is gov-erned by the break -even interconnection distance found by equatingthe energy required by the local and global interconnects, andby the computational and concurrency requirements imposed bya given application. For some applications, the amount of re-quired memory governs the grain size of the PE, resulting innonscalable systems. In POEM, the physical size of a PE maybe governed by the size of the RAM even for a small numberof storage cells. However, a RAM function is crucial for per-forming context switching, that is, for handling a number ofprocesses larger than the number of PEs in the system. Opticalmemory systems that will support large memory bandwidth andlarge storage capacity will remove these limitations and increasethe range of application of POEM systems.

The fine -grain POEM machine was designed to apply par-allelism to a wide variety of algorithms. However, because ofthe programmability of optical interconnects and the large num-ber of simple PEs, it is particularly effective for the rapid ex-ecution of symbolic information processing tasks and graph al-gorithms. The fine -grain POEM machine is expected to offerflexibility and high performance in the rapid execution of se-mantic networks, production systems, management of largeknowledge bases, transportation and communication optimiza-tion problems, computer -aided design, VLSI circuit simulation,parallel databases, and game playing. For example, consider theimplementation of a parallel knowledge -base system with POEMarchitecture. Theoretical work by Fahlman11 has shown thatstoring knowledge as a pattern of interconnections between manyvery simple PEs allows searches to be performed very quickly.The basic idea is to store the knowledge as a graph in whichindividual concepts are assigned to PEs and the interconnectionsbetween the PEs represent the relations between the concepts.Search operations are performed by marking specific node pro-cessors and then propagating these markers in parallel throughthe network. The set of conventions and processing algorithmsfor representing the knowledge in such a parallel network iscalled NETL. Fahlman has shown that NEIL is capable ofperforming search operations on the knowledge base, simpledeductions, learning, consistency checks, matching, and sym-bolic recognition tasks. The important and unique feature ofNETL is that the time required to perform a search is essentiallya constant, independent of the size of the knowledge base. Thisis to be contrasted to sequential AI systems, whose search timeincreases linearly with the size of the knowledge base. The NETLsystem can be directly mapped onto the POEM hardware12 byprogramming the optical interconnects. The large number of PEs

398 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

ZERO ONE

;! :06(0.0)-> 0 (0,1) -> 1 (1,0) ->1 (1,1) ->0

Fig. 2. Example of a substitution rule using dual -rail logic.6

Left-hand side Right -hand side

Fig. 3. Murdocca's substitution rule.74

allows POEM NEIL to be used in implementing large knowl-edge bases. The programmable interconnects of POEM removethe overhead and latency that would occur if NEIL were mappedonto a machine with a fixed interconnection.

2.2. SS -based computing systemsIn this section we give a brief review of SS -based computingsystems and some of the proposed optical implementations.

2.2.1. Architecture descriptionThe idea of SS is derived from cellular automata considered byVon Neumann,13 in which locally interconnected cells evolveusing certain transition rules. The motivation for consideringsuch computational models is the desire to show that a collectionof locally interconnected devices (cells) governed by simpletransition rules can exhibit interesting computational properties.

SS is an elaboration of the idea of cellular automata, suitedfor optical implementation. It is a pattern rewriting procedurethat operates in a parallel and space -invariant fashion on a 2 -Dplane of binary pixels. Every occurrence of a given pattern isreplaced by another pattern. Each such pair of patterns is calleda substitution or a transition rule. A pattern is a k x k square ofpixels in which certain pixels are required to have specific binaryvalues. An example of a rule is shown in Fig. 2. All occurrencesof the left -hand side (LHS) pattern are simultaneously replacedby the right -hand side (RHS) pattern. Since a pixel can be com-mon to several shifted venons of the replacement pattern, theinformation in that pixel as a result of the replacement is a logicalOR of the corresponding pixels.

Brenner, Huang, and Streibl7 have suggested a set of sub-stitution rules that are adequate to perform logical operations,thus demonstrating that SS is a general- purpose computing sys-tem. Murdocca14 proposed a general- purpose SS system thatconsists of only one substitution rule (Fig. 3). The choice ofsubstitution rules is determined by such criteria as universality,simplicity, ease of implementation, and efficiency. In particular,we show in Sec. 5.1.2 that the "complexity" of a rule influencesthe energy dissipation of the system.

KIAMILEV, ESENER, PATURI, FAINMAN, MERGER, GUEST, LEE

pology among the processors. Instead, the programmable optical interconnects are determined by the optoelectronic controller. Therefore, the programmer can implement a topology that best matches the current algorithm. In addition, the interconnection storage capacity requirement on the holographic material can be reduced if real-time reprogrammable material requirements can be added. For example, one may envision using photorefractive crystals or other nonlinear optical materials to apply repro­ grammable interconnects to the PEs. In this case, the user will be capable of reconfiguring the POEM in a very short time to match his algorithmic requirements.

The internal data paths of the PEs are implemented electrically as in a common electronic processor. Each PE has the capability to perform logic, conditional execution, data movement, and I/O operations [Fig. l(b)]. Also, each PE has some local ran­ dom-access memory (RAM) to support the conventional pro­ gramming models. In general, the grain size of the PEs is gov­ erned by the break-even interconnection distance found by equating the energy required by the local and global interconnects, and by the computational and concurrency requirements imposed by a given application. For some applications, the amount of re­ quired memory governs the grain size of the PE, resulting in nonscalable systems. In POEM, the physical size of a PE may be governed by the size of the RAM even for a small number of storage cells. However, a RAM function is crucial for per­ forming context switching, that is, for handling a number of processes larger than the number of PEs in the system. Optical memory systems that will support large memory bandwidth and large storage capacity will remove these limitations and increase the range of application of POEM systems.

The fine-grain POEM machine was designed to apply par­ allelism to a wide variety of algorithms. However, because of the programmability of optical interconnects and the large num­ ber of simple PEs, it is particularly effective for the rapid ex­ ecution of symbolic information processing tasks and graph al­ gorithms. The fine-grain POEM machine is expected to offer flexibility and high performance in the rapid execution of se­ mantic networks, production systems, management of large knowledge bases, transportation and communication optimiza­ tion problems, computer-aided design, VLSI circuit simulation, parallel databases, and game playing. For example, consider the implementation of a parallel knowledge-base system with POEM architecture. Theoretical work by Fahlman11 has shown that storing knowledge as a pattern of interconnections between many very simple PEs allows searches to be performed very quickly. The basic idea is to store the knowledge as a graph in which individual concepts are assigned to PEs and the interconnections between the PEs represent the relations between the concepts. Search operations are performed by marking specific node pro­ cessors and then propagating these markers in parallel through the network. The set of conventions and processing algorithms for representing the knowledge in such a parallel network is called NETL. Fahlman has shown that NETL is capable of performing search operations on the knowledge base, simple deductions, learning, consistency checks, matching, and sym­ bolic recognition tasks. The important and unique feature of NETL is that the time required to perform a search is essentially a constant, independent of the size of the knowledge base. This is to be contrasted to sequential AI systems, whose search time increases linearly with the size of the knowledge base. The NETL system can be directly mapped onto the POEM hardware 12 by programming the optical interconnects. The large number of PEs

(0,0)-

Fig. 2. Example of a substitution rule using dual-rail logic.6

Left-hand side Right-hand side

14Fig. 3. Murdocca's substitution rule.

allows POEM NETL to be used in implementing large knowl­ edge bases. The programmable interconnects of POEM remove the overhead and latency that would occur if NETL were mapped onto a machine with a fixed interconnection.

2.2. SS-based computing systemsIn this section we give a brief review of SS-based computing systems and some of the proposed optical implementations.

2.2.1. Architecture descriptionThe idea of SS is derived from cellular automata considered by Von Neumann, 13 in which locally interconnected cells evolve using certain transition rules. The motivation for considering such computational models is the desire to show that a collection of locally interconnected devices (cells) governed by simple transition rules can exhibit interesting computational properties.

SS is an elaboration of the idea of cellular automata, suited for optical implementation. 7 It is a pattern rewriting procedure that operates in a parallel and space-invariant fashion on a 2-D plane of binary pixels. Every occurrence of a given pattern is replaced by another pattern. Each such pair of patterns is called a substitution or a transition rule. A pattern is a kxk square of pixels in which certain pixels are required to have specific binary values. An example of a rule is shown in Fig. 2. All occurrences of the left-hand side (LHS) pattern are simultaneously replaced by the right-hand side (RHS) pattern. Since a pixel can be com­ mon to several shifted verions of the replacement pattern, the information in that pixel as a result of the replacement is a logical OR of the corresponding pixels.

Brenner, Huang, and Streibl7 have suggested a set of sub­ stitution rules that are adequate to perform logical operations, thus demonstrating that SS is a general-purpose computing sys­ tem. Murdocca14 proposed a general-purpose SS system that consists of only one substitution rule (Fig. 3). The choice of substitution rules is determined by such criteria as universality, simplicity, ease of implementation, and efficiency. In particular, we show in Sec. 5.1.2thatthe "complexity" of a rule influences the energy dissipation of the system.

398 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 4: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORSAND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

A general- purpose computing system that employs SS hasthe following structure: The binary plane contains an encodingof the input data and control bits. The substitution rule is appliedto this plane repeatedly for a predetermined number of cycles.We can think of the control bits as the program. If we haveseveral different rules, these can be applied serially or in parallel.When they are applied in parallel, the resultant plane would bethe OR of the resultant planes from the individual substitutionrules.

2.2.2. Optical implementations of SS systems

An optical system for performing SS must provide two basicoperations: pattern recognition and pattern replacement. The mostwidely used approaches for both operations apply a thresholdingoperation to a composite of shifted replicas of the input image.Here we briefly review the ways in which optical systems canproduce shifted image replicas and describe how this capabilityis combined with thresholding and logic -level restoration to pro-vide cascadable building blocks for pattern recognition and pat-tern substitution.

One important choice to be made in specifying a patternrecognition module is whether it will recognize patterns of ones,patterns of zeros, or patterns consisting of ones and zeros. Rec-ognition of patterns containing ones and zeros leads to systemcompactness and operational flexibility but also requires a morecomplex optical system.

2.2.2.1. Description of a simple recognition- substitutionmodule

Implementation of SS is simplified if the pattern to be recognizedconsists of only bright pixels (ones) or only dark pixels (zeros).A bright pixel pattern recognizer is described here.

A replica of the input image is made for each bright pixel inthe pattern to be recognized (the LHS pattern). Each replicaimage is shifted horizontally and vertically by an amount thatbrings a corresponding LHS bright pixel to the position of adesignated origin pixel. All of the shifted replicas of the inputimage are superimposed, producing a composite image havingpixels with different brightnesses. The brightest pixels in thecomposite image will occur at each position where the inputimage matches the LHS pattern. The composite image is incidenton an array of thresholding optical gates whose output leavesonly these brightest pixels (pattern matches) in the bright state.

Once bright pixels marking the locations of pattern matcheshave been obtained, the next step is to substitute the new RHSpattern at each location. For each bright pixel in the RHS pattern,a replica of the image at the output of the threshold array ismade. The replica images are shifted by an amount correspond-ing to the position of the bright pixels in the RHS pattern. Theshifted replicas are superimposed (ORed), with the result thatthe RHS pattern now appears in all locations that a recognitionspot existed. For achieving cascadable modules, an array of gainand isolation devices is included.

2.2.2.2. Implementation of image shifting and combiningoperations

Optical implementations of SS are all based on replicating, shift-ing, and recombining data page images. During pattern recog-nition, a shifted replica of the input image must be formed foreach distinguished bit in the pattern to be recognized. For sub-stitution, a shifted replica of the output of the threshold array

TiltedMirrors

Beamsplitter Cube

Input

I Output

Fig. 4. Beamsplitter used for image replication and shifting.

must be produced for each bright pixel in the substituted pattern.Two approaches to replicating, shifting, and combining imagesfor SS have been published in the literature7'15: geometricaloptics using beamsplitters, mirrors, and prisms and diffractiveoptics using holograms. We briefly review the merits and fun-damental limitations of each in the following:

Several systems using geometrical optics components havebeen proposed for providing the image replication, shifting, andcombining operations for SS. All of them are roughly equivalentto the beamsplitter configuration shown in Fig. 4. Although theseimplementations are very straightforward, the process is inher-ently power- inefficient. In principle, two images may be com-bined without power loss with the use of a polarization beam -splitter, but the ouput image, containing both polarizations, isnot suitable for cascaded stages of lossless combinations. Sincemany rules require detection and substitution of patterns con-taining at least four or more shifted images, a spatial light mod-ulator must be used to regenerate an image with one linearpolarization after each pair combination, or nonpolarized imagecombination must be used for each additional image combina-tion. If this second approach is adopted to combine N images,at least [(N /2) - 1] /(N /2) of the input power is lost.

The alternative to geometrical optics for image replication,shifting, and combining is the use of holograms. In contrast togeometrical optics, volume holograms can be used to losslesslycombine many images with very little loss. A more subtle prob-lem arises with the use of holographic optical elements (HOEs),however. Holograms do not delay wavefronts the way refractiveoptical components do. With holograms, all phase delays aremodulo 27r. This means, for instance, that if a hologram per-forms the function of a lens, wavefronts passing through thecenter of the holographic lens will arrive at the image beforethose passing through the edge. Put another way, pulses of lightwill be stretched in time, placing a lower limit on the clockperiod for an optical system. As an example, a 2.5 cm diameter,f/1 holographic lens will lengthen all pulses of light passingthrough its full aperture by about 50 ps.

2.2.2.3. Data encoding schemes

Two approaches have emerged for recognizing patterns con-taining both ones and zeros. The first approach is dual -rail logic,or position encoding. With this method both the true and falsestates of a logic variable are represented by a bright spot in theoptical array; ones are represented by a bright spot in a specifiedposition, zeros by a bright spot in another position (e.g., seeFig. 2). Thus, the problem of detecting ones and zeros in apattern has been translated into a requirement to detect just onesor just zeros. Processing can proceed as previously describedfor those operations.

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 399

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORS AND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

A general-purpose computing system that employs SS has the following structure: The binary plane contains an encoding of the input data and control bits. The substitution rule is applied to this plane repeatedly for a predetermined number of cycles. We can think of the control bits as the program. If we have several different rules, these can be applied serially or in parallel. When they are applied in parallel, the resultant plane would be the OR of the resultant planes from the individual substitution rules.

2.2.2. Optical implementations of SS systemsAn optical system for performing SS must provide two basic operations: pattern recognition and pattern replacement. The most widely used approaches for both operations apply a thresholding operation to a composite of shifted replicas of the input image. Here we briefly review the ways in which optical systems can produce shifted image replicas and describe how this capability is combined with thresholding and logic-level restoration to pro­ vide cascadable building blocks for pattern recognition and pat­ tern substitution.

One important choice to be made in specifying a pattern recognition module is whether it will recognize patterns of ones, patterns of zeros, or patterns consisting of ones and zeros. Rec­ ognition of patterns containing ones and zeros leads to system compactness and operational flexibility but also requires a more complex optical system.

2.2.2.7. Description of a simple recognition-substitution module

Implementation of SS is simplified if the pattern to be recognized consists of only bright pixels (ones) or only dark pixels (zeros). A bright pixel pattern recognizer is described here.

A replica of the input image is made for each bright pixel in the pattern to be recognized (the LHS pattern). Each replica image is shifted horizontally and vertically by an amount that brings a corresponding LHS bright pixel to the position of a designated origin pixel. All of the shifted replicas of the input image are superimposed, producing a composite image having pixels with different brightnesses. The brightest pixels in the composite image will occur at each position where the input image matches the LHS pattern. The composite image is incident on an array of thresholding optical gates whose output leaves only these brightest pixels (pattern matches) in the bright state.

Once bright pixels marking the locations of pattern matches have been obtained, the next step is to substitute the new RHS pattern at each location. For each bright pixel in the RHS pattern, a replica of the image at the output of the threshold array is made. The replica images are shifted by an amount correspond­ ing to the position of the bright pixels in the RHS pattern. The shifted replicas are superimposed (ORed), with the result that the RHS pattern now appears in all locations that a recognition spot existed. For achieving cascadable modules, an array of gain and isolation devices is included.

2.2.2.2. Implementation of image shifting and combining operations

Optical implementations of SS are all based on replicating, shift­ ing, and recombining data page images. During pattern recog­ nition, a shifted replica of the input image must be formed for each distinguished bit in the pattern to be recognized. For sub­ stitution, a shifted replica of the output of the threshold array

Tilted Mirrors

Fig. 4. Beamsplitter used for image replication and shifting.

must be produced for each bright pixel in the substituted pattern. Two approaches to replicating, shifting, and combining images for SS have been published in the literature7 ' 15 : geometrical optics using beamsplitters, mirrors, and prisms and diffractive optics using holograms. We briefly review the merits and fun­ damental limitations of each in the following:

Several systems using geometrical optics components have been proposed for providing the image replication, shifting, and combining operations for SS. All of them are roughly equivalent to the beamsplitter configuration shown in Fig. 4. Although these implementations are very straightforward, the process is inher­ ently power-inefficient. In principle, two images may be com­ bined without power loss with the use of a polarization beam­ splitter, but the ouput image, containing both polarizations, is not suitable for cascaded stages of lossless combinations. Since many rules require detection and substitution of patterns con­ taining at least four or more shifted images, a spatial light mod­ ulator must be used to regenerate an image with one linear polarization after each pair combination, or nonpolarized image combination must be used for each additional image combina­ tion. If this second approach is adopted to combine N images, at least [(N/2) - l]/(N/2) of the input power is lost.

The alternative to geometrical optics for image replication, shifting, and combining is the use of holograms. In contrast to geometrical optics, volume holograms can be used to losslessly combine many images with very little loss. A more subtle prob­ lem arises with the use of holographic optical elements (HOEs), however. Holograms do not delay wavefronts the way refractive optical components do. With holograms, all phase delays are modulo 2ir. This means, for instance, that if a hologram per­ forms the function of a lens, wavefronts passing through the center of the holographic lens will arrive at the image before those passing through the edge. Put another way, pulses of light will be stretched in time, placing a lower limit on the clock period for an optical system. As an example, a 2.5 cm diameter, f/1 holographic lens will lengthen all pulses of light passing through its full aperture by about 50 ps.

2.2.2 J. Data encoding schemes

Two approaches have emerged for recognizing patterns con­ taining both ones and zeros. The first approach is dual-rail logic, or position encoding. With this method both the true and false states of a logic variable are represented by a bright spot in the optical array; ones are represented by a bright spot in a specified position, zeros by a bright spot in another position (e.g., see Fig. 2). Thus, the problem of detecting ones and zeros in a pattern has been translated into a requirement to detect just ones or just zeros. Processing can proceed as previously described for those operations.

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 399

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 5: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

KIAMILEV, ESENER, PATURI, FAINMAN, MERCIER, GUEST, LEE

The other approach is to encode the binary states of a cellnot with intensity but with orthogonal polarizations of light.16As with simple recognition, a replica of the data plane is pro-duced for each distinguished cell in the LHS pattern, but in thiscase both true and false LHS cells may be specified. Replicascorresponding to zeros in the LHS pattern are passed through ahalfwave plate, thereby inverting the logic value of their bits.Shifting now occurs on all replicas to bring the specified LHScells to the origin. The resulting superposition passes through apolarizer aligned with the true state polarization in the data array.Wherever the data array matches the LHS pattern, all cells withthe true state polarization and cells with a false state polarizationthat has been rotated 90° to the true state are superimposed.Thus, matches are noted by the brightest pixels after passingthrough the analyzer. From this point, the rest of the processfollows that for simple recognition.

Both approaches roughly double the power consumed by thesystem. For dual -rail encoding this occurs because the numberof pixels to represent each bit is doubled and the complexity oflogic and data paths is correspondingly increased. For polar-ization -based encoding, a polarization analyzer is used prior tothe optical gates, thereby discarding half of the power.

3. SS SYSTEMS AND 2 -D VLSI MESH

In this section, we compare optical SS systems with a VLSI2 -D mesh of processors that operates in SIMD mode. We showthat an SS system can be efficiently simulated by a very-fine-grain mesh of processors and that an SS rule can be simulatedusing only a small number of cycles that depends on the size ofthe SS rule. In fact, we specify measures to quantify the com-plexity of an SS rule. On the other hand, we show that an SSsystem is inefficient in simulating a mesh of electronic proces-sors, where each processor has the ability to perform basic arith-metic and data movement operations on one -bit words. Thissimulation requires more space and time. We also give quan-titative estimates of the resources needed to simulate an opticalSS system using a mesh of VLSI processors. As in a mesh, inSS each instance of the rule works only on a small amount ofnearby information.

3.1. Simulation of SS by a meshWe first make the following assumptions about the mesh: Eachprocessor is connected to its four nearest neighbors with bi-directional edges. The operation of each processor is synchro-nized by a global clock. Each processor has instructions forcommunicating with its four neighbors and for computing thelogical operations AND, OR, and NOT.

To simulate an N x N optical SS system by an N x N meshof electronic processors, we further assume, without loss ofgenerality, an SS system based on a single rule, then extend ouranalysis later to handle the case of multiple rules. The basic ideais that each mesh processor (x,y) is responsible for the state ofthe pixel (x,y) in the binary plane of the SS system. We thensimulate the transition rule on the mesh and update the states.In the following, we compute the cost of simulating a transitionrule.

Consider a transition rule that replaces a k x k frame withanother k x k replacement frame based on the existence of acertain search pattern in the frame. A search pattern is specifiedby requiring distinguished pixels to have certain states. Let mbe the number of these pixels. The other (k2 - m) pixels in theframe are "don't- care" pixels because their state does not affect

400 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

the recognition of the pattern. Similarly, the replacement frameis specified by giving the set of distinguished pixels that arerequired to have the value 1. Let n be the number of those pixels.Other pixels in the replacement pattern have the value O. Ouraim is to capture the cost of the complexity of simulating thetransition rule as a function of k, m, and n.

Consider how a pixel in the output plane of an SS systemcan possibly change its state after an application of a transitionrule. Each pixel in the output plane depends on exactly n k x kframes. If at least one of these frames has the required searchpattern, a 1 will be written in the pixel. The presence of a searchpattern in a frame is determined by the m distinguished pixels.Hence, the new state of a cell is determined by a Boolean for-mula, which is an OR of n terms each of which is an AND ofm Boolean variables. We next show how this function can becomputed for each of the pixels in parallel in time 0(k2) andwith a small (O(min(n,2k))) amount of hardware per processorin the mesh.

In the first phase, we compute the AND of the distinguishedpixels for each possible k x k frame. For each frame, we des-ignate a unique pixel to collect and AND together the states ofthe distinguished pixels corresponding to the search pattern. Notethat each pixel appears distinguished in the search patterns ofexactly m frames. Hence, each pixel has to send its state to mdifferent recipients. This transmission can be accomplished in(k - 1) + min((k - 1)m, k(k - 1)) communication cycles ofthe mesh. Furthermore, each processor in the mesh need haveonly O(min(m,2k) + logk) switches. At the end of the firstphase, all of the required products are computed. In the secondphase, each distinguished pixel of the replacement pattern re-ceives n of these products and computes their OR. This againcan be accomplished in (k - 1) + min((k - 1)n, k(k - 1))communication cycles of the mesh with at most O(min(n,2k) +logk) switches per processor. In summary, a transition rule canbe simulated on the mesh in time O(min((m + n)k, k2)) withO(logk + min(m + n,k)) switches per processor.

These bounds work in general. In many specific cases, onecould exploit the regularity of the rule to derive more efficientsimulations. For example, Murdocca's transition rule can besimulated in about eight communication cycles.

The simulation procedure described above does not handlethe processors at the edges or the case of a system in whichseveral rules are being applied simultaneously. The edge pro-cessors can be taken care of by deleting the appropriate productterms. We can simulate a system with several rules by consid-ering the logical OR of the output binary planes that would resultfrom applying the individual transition rules. The cost functionsfor this case would be the same as in the one -rule case withk = max(ki), m = 1mi, and n = 2.ni.

Since there is a limitation on the size of electronic mesh thatcan be implemented at present, we should consider the problemof simulating an N x N SS system with a smaller M x M mesh,where M < N. We assume N/M to be some integer multiple ofk, and we compute the time and space requirements to performthis simulation. The basic idea is to make each processor in themesh responsible for a p x p window of pixels in an SS binaryplane, where p is N /M.

The simulation algorithm we use here is composed basicallyof a communication phase followed by a computation phase. Inthe communication phase, each processor sends pixel state in-formation to its four nearest neighbors such that 4p(k - 1) +4(k - 1)2 state bits are received at each processor. The idea is

KIAMILEV, ESENER, PATURI, FAINMAN, MERGER, GUEST, LEE

The other approach is to encode the binary states of a cell not with intensity but with orthogonal polarizations of light. 16 As with simple recognition, a replica of the data plane is pro­ duced for each distinguished cell in the LHS pattern, but in this case both true and false LHS cells may be specified. Replicas corresponding to zeros in the LHS pattern are passed through a half wave plate, thereby inverting the logic value of their bits. Shifting now occurs on all replicas to bring the specified LHS cells to the origin. The resulting superposition passes through a polarizer aligned with the true state polarization in the data array. Wherever the data array matches the LHS pattern, all cells with the true state polarization and cells with a false state polarization that has been rotated 90° to the true state are superimposed. Thus, matches are noted by the brightest pixels after passing through the analyzer. From this point, the rest of the process follows that for simple recognition.

Both approaches roughly double the power consumed by the system. For dual-rail encoding this occurs because the number of pixels to represent each bit is doubled and the complexity of logic and data paths is correspondingly increased. For polar­ ization-based encoding, a polarization analyzer is used prior to the optical gates, thereby discarding half of the power.

3. SS SYSTEMS AND 2-D VLSI MESHIn this section, we compare optical SS systems with a VLSI 2-D mesh of processors that operates in SIMD mode. We show that an SS system can be efficiently simulated by a very-fine- grain mesh of processors and that an SS rule can be simulated using only a small number of cycles that depends on the size of the SS rule. In fact, we specify measures to quantify the com­ plexity of an SS rule. On the other hand, we show that an SS system is inefficient in simulating a mesh of electronic proces­ sors, where each processor has the ability to perform basic arith­ metic and data movement operations on one-bit words. This simulation requires more space and time. We also give quan­ titative estimates of the resources needed to simulate an optical SS system using a mesh of VLSI processors. As in a mesh, in SS each instance of the rule works only on a small amount of nearby information.

3.1. Simulation of SS by a meshWe first make the following assumptions about the mesh: Each processor is connected to its four nearest neighbors with bi­ directional edges. The operation of each processor is synchro­ nized by a global clock. Each processor has instructions for communicating with its four neighbors and for computing the logical operations AND, OR, and NOT.

To simulate an N X N optical SS system by an N x N mesh of electronic processors, we further assume, without loss of generality, an SS system based on a single rule, then extend our analysis later to handle the case of multiple rules. The basic idea is that each mesh processor (x,y) is responsible for the state of the pixel (x,y) in the binary plane of the SS system. We then simulate the transition rule on the mesh and update the states. In the following, we compute the cost of simulating a transition rule.

Consider a transition rule that replaces a kxk frame with another kxk replacement frame based on the existence of a certain search pattern in the frame. A search pattern is specified by requiring distinguished pixels to have certain states. Let m be the number of these pixels. The other (k2 m) pixels in the frame are "don't-care" pixels because their state does not affect

the recognition of the pattern. Similarly, the replacement frame is specified by giving the set of distinguished pixels that are required to have the value 1. Let n be the number of those pixels. Other pixels in the replacement pattern have the value 0. Our aim is to capture the cost of the complexity of simulating the transition rule as a function of k, m, and n.

Consider how a pixel in the output plane of an SS system can possibly change its state after an application of a transition rule. Each pixel in the output plane depends on exactly n kxk frames. If at least one of these frames has the required search pattern, a 1 will be written in the pixel. The presence of a search pattern in a frame is determined by the m distinguished pixels. Hence, the new state of a cell is determined by a Boolean for­ mula, which is an OR of n terms each of which is an AND of m Boolean variables. We next show how this function can be computed for each of the pixels in parallel in time O(k2) and with a small (O(min(n,2k))) amount of hardware per processor in the mesh.

In the first phase, we compute the AND of the distinguished pixels for each possible kxk frame. For each frame, we des­ ignate a unique pixel to collect and AND together the states of the distinguished pixels corresponding to the search pattern. Note that each pixel appears distinguished in the search patterns of exactly m frames. Hence, each pixel has to send its state to m different recipients. This transmission can be accomplished in (k - 1) 4- min((k l)m, k(k 1)) communication cycles of the mesh. Furthermore, each processor in the mesh need have only O(min(m,2k) + logk) switches. At the end of the first phase, all of the required products are computed. In the second phase, each distinguished pixel of the replacement pattern re­ ceives n of these products and computes their OR. This again can be accomplished in (k 1) + min((k l)n, k(k 1)) communication cycles of the mesh with at most O(min(n,2k) + logk) switches per processor. In summary, a transition rule can be simulated on the mesh in time O(min((m 4- n)k, k2)) with O(logk + min(m 4- n,k)) switches per processor.

These bounds work in general. In many specific cases, one could exploit the regularity of the rule to derive more efficient simulations. For example, Murdocca's transition rule can be simulated in about eight communication cycles.

The simulation procedure described above does not handle the processors at the edges or the case of a system in which several rules are being applied simultaneously. The edge pro­ cessors can be taken care of by deleting the appropriate product terms. We can simulate a system with several rules by consid­ ering the logical OR of the output binary planes that would result from applying the individual transition rules. The cost functions for this case would be the same as in the one-rule case with k = max(ki), m = 2mi, and n = 2m.

Since there is a limitation on the size of electronic mesh that can be implemented at present, we should consider the problem of simulating an N x N SS system with a smaller M x M mesh, where M < N. We assume N/M to be some integer multiple of k, and we compute the time and space requirements to perform this simulation. The basic idea is to make each processor in the mesh responsible for a p x p window of pixels in an SS binary plane, where p is N/M.

The simulation algorithm we use here is composed basically of a communication phase followed by a computation phase. In the communication phase, each processor sends pixel state in­ formation to its four nearest neighbors such that 4p(k 1) + 4(k - I)2 state bits are received at each processor. The idea is

400 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 6: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORSAND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

that each processor gathers a k - 1 wide window of statesaround it so that it has all of the necessary information to computethe new states of its pixels. The time for the computation isO(p2log(mn)), and each processor needs O(mn + p2 + 4p(k -1) + 4(k - 1)2) switches. The overall time for simulating oneapplication of a transition rule is O(p2log(mn) + 4p (k -1) + 4(k - 1)2). In particular, when p » k, the time isO(p2log(mn)) and the hardware cost is 0(p2).

3.2. Simulation of a mesh by SSWe now consider the simulation of a VLSI mesh with an SSsystem. We show that such simulation requires more space andprocessing cycles, even for a very simple mesh.

Consider a mesh of one -bit processors, each having threeregisters capable of performing logical and data movement op-erations. We also have instructions to transport the data betweenthe neighboring processors. To simulate such a system, we makethe following two generous assumptions about the capabilitiesof the SS system: (1) the system can have a large number ofsubstitution rules operating in parallel and (2) the control bitsin the input plane can be changed every cycle.

The basic idea of the simulation is to allocate a window ofSS pixels for each processor. This window contains the spacefor the three registers and the control bits to specify the instruc-tion in dual -rail logic. We use multiple SS rules (about 16)operating in parallel to implement the instruction set.

This scheme gives us the minimal area per processor and onecycle time to execute an instruction. Simple calculations showthat the area required per processor would be at least 25 pixels.Thus, if we assume that the binary plane has 1000 x 1000 pixels,we can at best simulate a 200 x 200 mesh of one -bit processorswith each step of mesh taking one clock cycle of the SS system.

If a larger -grain processor is used or if the above -mentionedassumptions are not feasible, in particular if we have to workwith a single rule, then the corresponding simulation would bemuch more inefficient in terms of both time and area. This wouldimply that any realistic SS system can simulate only a smallmesh (less than 100 processors), taking a large number of cyclesto simulate a cycle of the mesh.

To summarize, we have shown that an SS system is no morepowerful than a fine -grain mesh of processors of similar size.This means that any advantage that can be enjoyed by an SSsystem must come from technological considerations. In the nextsection, we look at the technological aspects.

4. SYSTEM AND TECHNOLOGICALCONSIDERATIONS OF POEM AND SS

Here, we discuss the technological characteristics of both POEMand SS systems. In particular, we determine the energy dissi-pation and speed of these systems. To begin with, let us considersome fundamental characteristics associated with the optical gatesof which these systems are composed.

4.1. Fundamental considerations for optical gate arraysIn the following, we analyze optical gate switching speed andarray size in terms of thermal limitations, optical interconnectdensity, and efficiency of optical and electrical interconnects.

In general, a bound on the number (N x N) of gates in anarray of area A can be found by requiring that heat dissipationcannot be larger than the heat removal per switching cycle. Thus,we have

N2 < PdmaxA

Pc&(1)

where Pd max is the maximum allowable power dissipation den-sity, which is dependent on the thermal characteristics of thematerial and the heat removal technique applied to the device.Pc is the power dissipation density of a single optical gate, andAc is its active area. In addition, the required space bandwidthproduct (SBP) of an optically interconnected system is

SBP ? A .Ac

(2)

In general, A is limited by wafer size, and Ac is limited bylithography or by the optical wavelength. Combining Eqs. (1)and (2), we obtain an upper limit on the size of an optical gatearray imposed by thermal dissipation and optical interconnectdensity as

N2 <- ( . (3)

For an optical gate, the power dissipation density is related tothe switching energy density Ec and the switching speed T byPc = Ec /T. Using this relation in Eq. (3), we can show that theminimum switching speed T of the array is determined by

Pa (SBP)T = (4)

Hence, for a given device and optical interconnect technology,the speed of an optical gate is limited by the array size. Animportant figure of merit for optical gate arrays, therefore, isthe array throughput, given by

N2T-1 < SBPI PÉ)\

(5)

This equation puts an upper limit on the capabilities of anyoptical gate array implemented with a given technology. In thecase of the optoelectronic PE arrays used in POEM, Ac is thearea of a single modulator in each PE and is occupying only asmall fraction of the total PE area. With the simplifying as-sumptions made in Sec. 5.1 for the worst -case calculations, Eq. (5)can also be used to estimate the computational throughput ofPOEM systems. Next, we develop models to evaluate the energydissipation and the latency of POEM and SS systems.

4.2. Energy dissipation and latency for POEM and SSIn this section we determine the energy dissipation and the speedof POEM and SS systems.

4.2.1. POEMThe POEM machine is composed of electronic PEs intercon-nected with holographic optical interconnects. Each PE is madeof logic gates interconnected with electrical interconnects. Theenergy is dissipated essentially in the electrical interconnectionsand in silicon inverters. The maximum PE clock rate is funda-mentally determined by the speed of the longest electrical in-terconnect in the PE, while the speed of interprocessor com-

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 401

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORS AND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

that each processor gathers a k 1 wide window of states around it so that it has all of the necessary information to compute the new states of its pixels. The time for the computation is O(p2log(mn)), and each processor needs O(mn + p2 + 4p(k 1) + 4(k I)2) switches. The overall time for simulating one application of a transition rule is O(p2log(mn) + 4p(k 1) + 4(k - I)2). In particular, when p » k, the time is O(p2log(mn)) and the hardware cost is O(p2).

3.2. Simulation of a mesh by SSWe now consider the simulation of a VLSI mesh with an SS system. We show that such simulation requires more space and processing cycles, even for a very simple mesh.

Consider a mesh of one-bit processors, each having three registers capable of performing logical and data movement op­ erations. We also have instructions to transport the data between the neighboring processors. To simulate such a system, we make the following two generous assumptions about the capabilities of the SS system: (1) the system can have a large number of substitution rules operating in parallel and (2) the control bits in the input plane can be changed every cycle.

The basic idea of the simulation is to allocate a window of SS pixels for each processor. This window contains the space for the three registers and the control bits to specify the instruc­ tion in dual-rail logic. We use multiple SS rules (about 16) operating in parallel to implement the instruction set.

This scheme gives us the minimal area per processor and one cycle time to execute an instruction. Simple calculations show that the area required per processor would be at least 25 pixels. Thus, if we assume that the binary plane has 1000 x 1000 pixels, we can at best simulate a 200 x 200 mesh of one-bit processors with each step of mesh taking one clock cycle of the SS system.

If a larger-grain processor is used or if the above-mentioned assumptions are not feasible, in particular if we have to work with a single rule, then the corresponding simulation would be much more inefficient in terms of both time and area. This would imply that any realistic SS system can simulate only a small mesh (less than 100 processors), taking a large number of cycles to simulate a cycle of the mesh.

To summarize, we have shown that an SS system is no more powerful than a fine-grain mesh of processors of similar size. This means that any advantage that can be enjoyed by an SS system must come from technological considerations. In the next section, we look at the technological aspects.

4. SYSTEM AND TECHNOLOGICAL CONSIDERATIONS OF POEM AND SSHere, we discuss the technological characteristics of both POEM and SS systems. In particular, we determine the energy dissi­ pation and speed of these systems. To begin with, let us consider some fundamental characteristics associated with the optical gates of which these systems are composed.

4.1. Fundamental considerations for optical gate arraysIn the following, we analyze optical gate switching speed and array size in terms of thermal limitations, optical interconnect density, and efficiency of optical and electrical interconnects.

In general, a bound on the number (NxN) of gates in an array of area A can be found by requiring that heat dissipation cannot be larger than the heat removal per switching cycle. Thus, we have

PCAC(1)

where Pdmax is the maximum allowable power dissipation den­ sity, which is dependent on the thermal characteristics of the material and the heat removal technique applied to the device. PC is the power dissipation density of a single optical gate, and AC is its active area. In addition, the required space bandwidth product (SBP) of an optically interconnected system is

AC(2)

In general, A is limited by wafer size, and Ac is limited by lithography or by the optical wavelength. Combining Eqs. (1) and (2), we obtain an upper limit on the size of an optical gate array imposed by thermal dissipation and optical interconnect density as

N2 < -JSBP . (3)

For an optical gate, the power dissipation density is related to the switching energy density Ec and the switching speed T by PC = EC/T. Using this relation in Eq. (3), we can show that the minimum switching speed T of the array is determined by

EC / Ndmax

(4)

Hence, for a given device and optical interconnect technology, the speed of an optical gate is limited by the array size. An important figure of merit for optical gate arrays, therefore, is the array throughput, given by

< SBP /Pdmax\

reT/ ' (5)

This equation puts an upper limit on the capabilities of any optical gate array implemented with a given technology. In the case of the optoelectronic PE arrays used in POEM, Ac is the area of a single modulator in each PE and is occupying only a small fraction of the total PE area. With the simplifying as­ sumptions made in Sec. 5.1 for the worst-case calculations, Eq. (5) can also be used to estimate the computational throughput of POEM systems. Next, we develop models to evaluate the energy dissipation and the latency of POEM and SS systems.

4.2. Energy dissipation and latency for POEM and SSIn this section we determine the energy dissipation and the speed of POEM and SS systems.

4.2.1. POEM

The POEM machine is composed of electronic PEs intercon­ nected with holographic optical interconnects. Each PE is made of logic gates interconnected with electrical interconnects. The energy is dissipated essentially in the electrical interconnections and in silicon inverters. The maximum PE clock rate is funda­ mentally determined by the speed of the longest electrical in­ terconnect in the PE, while the speed of interprocessor com-

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 401

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 7: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

KIAMILEV, ESENER, PATURI, FAINMAN, MERCIER, GUEST, LEE

munication is determined by the longest holographic interconnectin the system.

First, we discuss the energy dissipation and the speed of aPE. The total energy dissipated per clock cycle within a PE isthe sum of the energies spent in switching the electronic logicgates and driving the interconnects. The energy consumed inswitching a logic gate with short connections is dominated bythe gate input capacitance C. If V is the required voltage swing,then the switching energy is given by

CV2

E` = . (6)

When the connections are longer, the wire capacitance dominatesthe gate input capacitance and the switching energy becomesproportional to the length of the electrical wire.

The operating speed of the circuit is inversely proportionalto the connection delay, which depends on the length of the

wire. For short wires it is given by 17

Tshort wire = 2.718Tinven(KLwire) (7)

where Lwíre is the connection length and K is a constant, typicallybetween 0.1 and 0.2 µm -1. The inverter switching time TInv isa technological constant representing the logical gate switchingspeed. This logarithmic dependence of wire delay on wire lengthshows that for locally connected gates, the speed is essentiallydetermined by Tiny. On the other hand, when the connectionsare long the wire delay is proportional to the wire length and isgiven by17

LwireVTlong wire

c(8)

where c is the speed of light and Er is a constant, typically about4. Thus, long electrical connections decrease the speed of op-eration and increase the energy consumption.

We now turn our attention to the energy dissipation and speedof holographic interprocessor connections. Figure 5 illustrates afree -space optical interconnect system. A biasing optical fieldis incident on only the modulators associated with each PE. Thelight, transmitted by a modulator that is turned on, is directedwith holographic interconnects onto the desired detector(s). Theenergy required by such interconnects can be evaluated to beta

Ea = 2VF(C + Cinv)I q + VJ + CMVM (9)

where E0 is the required optical link energy, V is the invertervoltage swing, F is the fan-out, Cpd is the photodetector capac-itance, CM is the modulator capacitance, and VM is the halfwavevoltage of the modulator. The photon energy is represented byhv, and the electronic charge by q. The efficiency of the opticallink is modeled by 1, which includes the efficiencies of themodulator, hologram, and detector. Compared with the energyrequirements of electrical interconnects in Eq. (7), it can beshown that E. is less for long communication distances. Thebreak -even communication length establishes the criteria for theappropriate use of electrical and optical interconnections. As anexample, an optical link realized with a PLZT light modulatorwith 10 µm2 area and a fan-out of 1 using the 2.5 µm processwill dissipate an energy of 50 pi, assuming 60% holographic

402 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

ProcessingPlanes

OpticalSource

Detectory

Fig. 5. Free -space holographic interconnects.

diffraction efficiency and 90% modulator and detector efficien-cies. Compared with the energy required for a typical electricaloff -chip connection of about 1 ni, the optical link consumes lessenergy.

The speed of operation of POEM systems can be limited bythe latency of the global optical links, local electrical intercon-nect delay, or the inverter switching speed. The latency of theglobal optical links will be governed technologically by the lightmodulator speed and fundamentally by the skew introduced byholographic interconnects and by the free -space optical propa-gation delays. Typical achievable speeds with light modulatorsare 0.1 to 1µs with Si/PLZT and 1 to 10 ns with multiplequantum well (MQW) technologies.19 For global holographicoptical interconnections, relative time delays will be introducedamong the PEs by the hologram. This skew can be expressedfrom simple geometrical considerations as

2

Th=-e(11 +D -1 , (10)

where e is the distance between the PE array and the hologramand D is the length of a side of the array. For fixed interconnects,this skew can be compensated for by the introduction of appro-priate optical time -delay elements into different communicationpaths. However, in the case of programmable optical intercon-nects, this compensation technique cannot be used because ofthe time dependence of the relative delays. Nevertheless, themagnitude of the skew is presently less than the latency of thestate -of- the -art MQW light modulators and therefore does notlimit the communication speed. For example, for an optoelec-tronic PE array 15 cm on the side, the signal skew ranges from20 ps to 200 ps as f is varied from 150 to 15 cm. Note that themagnitude of the skew is reduced by increasing f. However,the free -space optical propagation delay increases with a ac-cording to 1-ft. = e /c. Thus, for a given array dimension thereexists an optimal distance e, minimizing the propagation delayand signal skew.

4.2.2. SSHere, we compute the energy dissipation and the delay involvedin a single application of an SS transition rule. Figure 6 is a

KIAMILEV, ESENER, PATURI, FAINMAN, MERGER, GUEST, LEE

munication is determined by the longest holographic interconnect in the system.

First, we discuss the energy dissipation and the speed of a PE. The total energy dissipated per clock cycle within a PE is the sum of the energies spent in switching the electronic logic gates and driving the interconnects. The energy consumed in switching a logic gate with short connections is dominated by the gate input capacitance C. If V is the required voltage swing, then the switching energy is given by

CV2~ (6)

When the connections are longer, the wire capacitance dominates the gate input capacitance and the switching energy becomes proportional to the length of the electrical wire.

The operating speed of the circuit is inversely proportional to the connection delay, which depends on the length of the wire. For short wires it is given by17

Tshort wire = 2.718Tinv^n(KLwire) (7)

where Lwire is the connection length and K is a constant, typically between 0.1 and 0.2 jjum" 1 . The inverter switching time Tinv is a technological constant representing the logical gate switching speed. This logarithmic dependence of wire delay on wire length shows that for locally connected gates, the speed is essentially determined by Tinv - On the other hand, when the connections are long the wire delay is proportional to the wire length and is given by 17

Lwire V^(8)

where c is the speed of light and er is a constant, typically about 4. Thus, long electrical connections decrease the speed of op­ eration and increase the energy consumption.

We now turn our attention to the energy dissipation and speed of holographic interprocessor connections. Figure 5 illustrates a free-space optical interconnect system. A biasing optical field is incident on only the modulators associated with each PE. The light, transmitted by a modulator that is turned on, is directed with holographic interconnects onto the desired detector(s). The energy required by such interconnects can be evaluated to be °18

2VF(Cpd + Cinv)( (9)

where E0 is the required optical link energy, V is the inverter voltage swing, F is the fan-out, Cpd is the photodetector capac­ itance, CM is the modulator capacitance, and VM is the halfwave voltage of the modulator. The photon energy is represented by hv, and the electronic charge by q. The efficiency of the optical link is modeled by T], which includes the efficiencies of the modulator, hologram, and detector. Compared with the energy requirements of electrical interconnects in Eq. (7), it can be shown that EO is less for long communication distances. The break-even communication length establishes the criteria for the appropriate use of electrical and optical interconnections. As an example, an optical link realized with a PLZT light modulator with 10 jim area and a fan-out of 1 using the 2.5 jxm process will dissipate an energy of 50 pJ, assuming 60% holographic

Fig. 5. Free-space holographic interconnects.

diffraction efficiency and 90% modulator and detector efficien­ cies. Compared with the energy required for a typical electrical off-chip connection of about 1 nJ, the optical link consumes less energy.

The speed of operation of POEM systems can be limited by the latency of the global optical links, local electrical intercon­ nect delay, or the inverter switching speed. The latency of the global optical links will be governed technologically by the light modulator speed and fundamentally by the skew introduced by holographic interconnects and by the free-space optical propa­ gation delays. Typical achievable speeds with light modulators are 0.1 to 1 JJLS with Si/PLZT and 1 to 10 ns with multiple quantum well (MQW) technologies. 19 For global holographic optical interconnections, relative time delays will be introduced among the PEs by the hologram. This skew can be expressed from simple geometrical considerations as

(10)

where is the distance between the PE array and the hologram and D is the length of a side of the array. For fixed interconnects, this skew can be compensated for by the introduction of appro­ priate optical time-delay elements into different communication paths. However, in the case of programmable optical intercon­ nects, this compensation technique cannot be used because of the time dependence of the relative delays. Nevertheless, the magnitude of the skew is presently less than the latency of the state-of-the-art MQW light modulators and therefore does not limit the communication speed. For example, for an optoelec- tronic PE array 15 cm on the side, the signal skew ranges from 20 ps to 200 ps as ( is varied from 150 to 15 cm. Note that the magnitude of the skew is reduced by increasing . However, the free-space optical propagation delay increases with ac­ cording to Ttr = /c. Thus, for a given array dimension there exists an optimal distance , minimizing the propagation delay and signal skew.

4.2.2. SSHere, we compute the energy dissipation and the delay involved in a single application of an SS transition rule. Figure 6 is a

402 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 8: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORSAND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

Losslesssplit/shift/Combineopticalsystem

Thresholdarray

Losslessspllit/

S(Gs - err) shift/(data) combine

opticalsystem

(b - S)Ge.

n

+ Sect

(loss)

? S'sn

Fig. 6. Model for computing the energy dissipation in an SS module.

ein(input)

Ilyee (bias energy)

ire. (bias energy)

Geln

(output)

(input) i i (output)

(b)

Fig. 7. Model of an amplifying optical gate array.

diagram of the system we use. We make the following assump-tions about the system:(i) The SS system uses a single rule of (k,m,n) complexity

operating on an N x N pixel image.(ii) The input image contains b bright pixels, each having an

energy of ein(iii) The input image contains S occurrences of the search

pattern.The transition rule produces an N x N output image withS' bright pixels, each having an energy of einThe optical operations of splitting, shifting, combining,and imaging are lossless.The system contains two N x N arrays of optical gates:one for logic -level isolation and restoration (amplification)and the other one for thresholding.Figure 7 shows the model of the three -terminal deviceused in the amplifier array. When light of energy ein entersthe device, light of energy Geit, leaves. In this case, con-servation of energy requires that

et.' ein(input)

et,< etn

(input)

ei ea(output)

ed

(energy loss)

(a)

tam.. T own. 0 (output)

ein(energy loss)

(b)

Fig. 8. Model of a thresholding optical gate array.

eb + e¡n = Gem + eca t (11)

where eb is the bias energy to the amplifier, G is the input -output gain of the device, and eca is the energy dissipatedin switching the device. On the other hand, when no inputlight enters the device, the bias energy eb is dissipated bythe device and no output is produced.

(viii) Figure 8 shows the model of the two -terminal device usedin the threshold array. Such a device is characterized byits threshold energy and switching energy ect. When theinput light is below the threshold energy, no output isproduced and all of the input energy is dissipated at thedevice. But when the input light energy exceeds the thresh-old, output light of energy (ein - ect) is produced and anenergy of ect is dissipated in switching the device.

(ix) The devices are memoryless; that is, at the end of eachclock cycle the optical gate arrays are reset.

We now explain the system energy budget shown in Fig. 6.The bias energy of N2eb is used to power the amplifier array.Since the input has b bright pixels, the output of the amplifierarray also has b bright pixels, each having energy of Get.. Thetotal energy dissipated in the amplifier array is the sum of theenergies dissipated by the b switching cells, which had lightincident on them (beca), and the N2 - b cells, which had nolight incident on them [(N2 - b)eb]

The amplifier array is followed by a lossless optical systemthat produces an image with S pixels above the threshold energyof the threshold devices. This image is incident on the thresholdarray. The energy dissipation in the threshold array is the sumof the energy to switch the devices for S pixels above threshold(Sect) and all of the energy that is below threshold [(b - S)Get].The output produced by the threshold array is an image with Sbright pixels each having an energy of ( Geit, - ect). This imageis passed to the optical system for substitution, which generatesa final output image with S' bright pixels.

Conservation of energy requires that the total energy enteringthe system be equal to the energy leaving the system plus theenergy absorbed. Using this constraint we obtain

eb >_ (n - 1)e;a + eca + act (12)

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 403

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORS AND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

be (input)

Fig. 6. Model for computing the energy dissipation in an SS module.

(bias energy)

(input)

e (loss)

(a)

(bias energy)

B (loss)

Fig. 7. Model of an amplifying optical gate array.

diagram of the system we use. We make the following assump­ tions about the system: (i) The SS system uses a single rule of (k,m,n) complexity

operating on an N x N pixel image, (ii) The input image contains b bright pixels, each having an

energy of Cm- (iii) The input image contains S occurrences of the search

pattern, (iv) The transition rule produces an N X N output image with

S' bright pixels, each having an energy of ein . (v) The optical operations of splitting, shifting, combining,

and imaging are lossless, (vi) The system contains two N x N arrays of optical gates:

one for logic-level isolation and restoration (amplification)and the other one for thresholding,

(vii) Figure 7 shows the model of the three-terminal deviceused in the amplifier array. When light of energy ein entersthe device, light of energy Gein leaves. In this case, con­ servation of energy requires that

0 (output)

e (energy loss)

Fig. 8. Model of a thresholding optical gate array.

6b + Gin — GCin + Cca > (11)

where 65 is the bias energy to the amplifier, G is the input- output gain of the device, and eca is the energy dissipated in switching the device. On the other hand, when no input light enters the device, the bias energy eb is dissipated by the device and no output is produced.

(viii) Figure 8 shows the model of the two-terminal device used in the threshold array. Such a device is characterized by its threshold energy and switching energy ect. When the input light is below the threshold energy, no output is produced and all of the input energy is dissipated at the device. But when the input light energy exceeds the thresh­ old, output light of energy (ein ect) is produced and an energy of ect is dissipated in switching the device.

(ix) The devices are memoryless; that is, at the end of each clock cycle the optical gate arrays are reset.

We now explain the system energy budget shown in Fig. 6. The bias energy of N2eb is used to power the amplifier array. Since the input has b bright pixels, the output of the amplifier array also has b bright pixels, each having energy of Getn . The total energy dissipated in the amplifier array is the sum of the energies dissipated by the b switching cells, which had light incident on them (beca), and the N2 b cells, which had no light incident on them [(N2 b)eb].

The amplifier array is followed by a lossless optical system that produces an image with S pixels above the threshold energy of the threshold devices. This image is incident on the threshold array. The energy dissipation in the threshold array is the sum of the energy to switch the devices for S pixels above threshold (Sect) and all of the energy that is below threshold [(b S)Gein]. The output produced by the threshold array is an image with S bright pixels each having an energy of (Getn ect). This image is passed to the optical system for substitution, which generates a final output image with S' bright pixels.

Conservation of energy requires that the total energy entering the system be equal to the energy leaving the system plus the energy absorbed. Using this constraint we obtain

eb s> (n - l)ein ect . (12)

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 403

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 9: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

KIAMILEV, ESENER, PATURI, FAINMAN, MERCIER, GUEST, LEE

This equation reveals that each pixel needs enough energy tocreate a full rule -substitution pattern and to energize one thresh-old device and one amplifier device. Now, we use Eq. (11) toeliminate the dependence on ein in Eq. (12) to obtain

(-Gn

G-11eb ? eca + Ject- (13)

This equation indicates that the gain of each amplifying devicemust exceed n. The overall energy dissipation can now be com-puted by adding the energy dissipations of the amplifier and thethresholding arrays and using Eq. (13) for eb. This gives

N2ebEdiss = N eb + N ectG-n

=

NZ(G-le+NZeca

G+N2(bS'

G Jct lest-n -n

(14)

The first term in Eq. (14) is the energy required to bias an arrayof N x N optical devices such that the recognition- substitutionoperation can be carried out. This amount of energy is dissipatedunder all conditions. According to Eq. (14), the bias energy isquite large because it is N2(G - 1) /(G - n) times the switchingenergy required per thresholding device plus N2 times the switch-ing energy required by the amplifying devices. The second termin Eq. (14) represents the energy losses associated with differentfan-in and fan -out and can be made to vanish for m = n, i.e.,for constant fan -in and fan -out. For example, assuming G of 5and using Murdocca's simplest rule where (k,m,n) = (3,4,4),we require power dissipation of (36ea + 9eca) for each 3 x 3window. Considering that the switching energy of an opticaldevice is presently equal to the energy of an electronic transistor,Murdocca's SS rule requires the energy equivalent of 45 tran-sistors. We discuss the computational value of such a recogni-tion- substitution module in the next section.

The above argument can easily be extended to an SS systemwith R parallel rules. Such a system is basically equivalent toR one -rule SS systems operating in parallel. Assuming the samegain for the amplifying array, the energy dissipation will beessentially R times larger and is given by

= N JZ[R(G- 1) +(b -S')1

eat + N2eca (15)Eaiss (G- 1)- R(n -1)

We now estimate the latency of a recognition- substitutionmodule. The time required to perform one application of an SSrule is the sum of the time required to switch the optical gatesand the transit time through the optical system. The speed ofthe optical gate is limited by the array size and is given in Eq. (4).The transit time is limited by the complexity of the rule and theimaging optics. For example, for a very simple SS rule such asthe one proposed by Murdocca, the transit time ?transit is pro-portional to

4fTtransit =

c(16)

where f is the lens focal length. Using the expression for res-olution of an Airy pattern, we can express the latency in termsof the SBP of the optical system; the f- number of the lenses,F #; and the optical frequency y as

404 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

Li

Fig. 9. A flip -flop implemented with 50 x 56 pixels using Murdocca'srule.

Ttransit ? 10v-1VSBP(F#)2 (17)

or, using Eq. (3),

Ttrans ? 10v-1(F#)2N1/ PcPdmax

(18)

Note that the latency of an SS system grows as the size of thearray increases. For more complex systems, the optical transittime increases with the parameter m of the rule and the numberof rules R.

5. RELATIVE MERITS OF POEM AND SSBased on the previous analysis, we now compare quantitativelythe performance potential of POEM and SS systems.

5.1. Computational efficiencyThe computational power of SS systems lies in their ability toimplement space -invariant transition rules very quickly. Thecommunication involved in effecting these transition rules isdone by replicating, shifting, and combining images. Such op-erations are easy to accomplish in optics.

But this capability of SS systems does not necessarily translateinto computational efficiency. The computation involved is doneby thresholding or clipping (a nonbinary operation) an analogsignal back to binary form, resulting in inefficient energy uti-lization, especially for complex rules. The communication pro-vided by the transition rules is very local and space invariant.However, many computations, including basic operations suchas addition and multiplication, can be implemented more effi-ciently with space -variant communication. Hence, SS requiresa large number of pixels and substitution cycles to implementoperations such as logic functions, addition, multiplication, etc.

To provide a specific example, consider the implementationof a NAND gate using Murdocca's rule.14 The SS NAND gatetakes 255 pixels of area and requires six applications of thetransition rule. In contrast, an electronic NAND gate requiresfour inverters and takes about one inverter switching time whenshort wires are used. That is, a NAND gate fabricated with1µm CMOS lithography that has a fan-out of 2 takes 400 pswhen the connection length is less than 1 mm. If the inverterswitching energy and the optical gate switching energies areassumed to be the same, then Murdocca's NAND gate requiresfour orders of magnitude more energy than the electronic NANDgate. Another example of wasted space is shown in Fig. 9, wherea flip -flop is implemented with 50 x 56 pixels using Murdocca'srule. Additional examples given by Cloonan20 and Goodman21

KIAMILEV, ESENER, PATURI, FAINMAN, MERGER, GUEST, LEE

This equation reveals that each pixel needs enough energy to create a full rule-substitution pattern and to energize one thresh­ old device and one amplifier device. Now, we use Eq. (11) to eliminate the dependence on ein in Eq. (12) to obtain

eb 5> eca (13)

This equation indicates that the gain of each amplifying device must exceed n. The overall energy dissipation can now be com­ puted by adding the energy dissipations of the amplifier and the thresholding arrays and using Eq. (13) for et>. This gives

b-S'

G-n ect

G-n G-n

(14)

The first term in Eq. (14) is the energy required to bias an array of N x N optical devices such that the recognition-substitution operation can be carried out. This amount of energy is dissipated under all conditions. According to Eq. (14), the bias energy is quite large because it is N2(G 1)/(G - n) times the switching energy required per thresholding device plus N2 times the switch­ ing energy required by the amplifying devices. The second term in Eq. (14) represents the energy losses associated with different fan-in and fan-out and can be made to vanish for m = n, i.e., for constant fan-in and fan-out. For example, assuming G of 5 and using Murdocca's simplest rule where (k,m,n) = (3,4,4), we require power dissipation of (36ect + 9eca) for each 3x3 window. Considering that the switching energy of an optical device is presently equal to the energy of an electronic transistor, Murdocca's SS rule requires the energy equivalent of 45 tran­ sistors. We discuss the computational value of such a recogni­ tion-substitution module in the next section.

The above argument can easily be extended to an SS system with R parallel rules. Such a system is basically equivalent to R one-rule SS systems operating in parallel. Assuming the same gain for the amplifying array, the energy dissipation will be essentially R times larger and is given by

(15)

We now estimate the latency of a recognition-substitution module. The time required to perform one application of an SS rule is the sum of the time required to switch the optical gates and the transit time through the optical system. The speed of the optical gate is limited by the array size and is given in Eq. (4). The transit time is limited by the complexity of the rule and the imaging optics. For example, for a very simple SS rule such as the one proposed by Murdocca, the transit time Transit is pro­ portional to

_ 4f^transit

C(16)

where f is the lens focal length. Using the expression for res­ olution of an Airy pattern, we can express the latency in terms of the SBP of the optical system; the f-number of the lenses, F#; and the optical frequency v as

Fig. 9. A flip-flop implemented with 50 x 56 pixels using Murdocca's rule.

Ttransit ^ 10v 1 \/SBP(F#)2

or, using Eq. (3),

^transit

(17)

(18)

Note that the latency of an SS system grows as the size of the array increases. For more complex systems, the optical transit time increases with the parameter m of the rule and the number of rules R.

5. RELATIVE MERITS OF POEM AND SSBased on the previous analysis, we now compare quantitatively the performance potential of POEM and SS systems.

5.1. Computational efficiencyThe computational power of SS systems lies in their ability to implement space-invariant transition rules very quickly. The communication involved in effecting these transition rules is done by replicating, shifting, and combining images. Such op­ erations are easy to accomplish in optics.

But this capability of SS systems does not necessarily translate into computational efficiency. The computation involved is done by thresholding or clipping (a nonbinary operation) an analog signal back to binary form, resulting in inefficient energy uti­ lization, especially for complex rules. The communication pro­ vided by the transition rules is very local and space invariant. However, many computations, including basic operations such as addition and multiplication, can be implemented more effi­ ciently with space-variant communication. Hence, SS requires a large number of pixels and substitution cycles to implement operations such as logic functions, addition, multiplication, etc.

To provide a specific example, consider the implementation of a NAND gate using Murdocca's rule. 14 The SS NAND gate takes 255 pixels of area and requires six applications of the transition rule. In contrast, an electronic NAND gate requires four inverters and takes about one inverter switching time when short wires are used. That is, a NAND gate fabricated with 1 jxm CMOS lithography that has a fan-out of 2 takes 400 ps when the connection length is less than 1 mm. If the inverter switching energy and the optical gate switching energies are assumed to be the same, then Murdocca's NAND gate requires four orders of magnitude more energy than the electronic NAND gate. Another example of wasted space is shown in Fig. 9, where a flip-flop is implemented with 50 x 56 pixels using Murdocca's rule. Additional examples given by Cloonan20 and Goodman21

404 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 10: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORSAND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

show that many other important Boolean logic modules requiremore area and time when implemented with SS.

In Sec. 3 we showed that it requires a large area to implementa basic processor capable of arithmetic and data movement op-erations. In Sec. 4 we derived the limitation in size and speedof optical gate arrays in terms of thermal considerations. In thefollowing, we show how power considerations limit the size andspeed of POEM and SS systems. In particular, we derive someestimates on speed and size based on the value E,, the minimaldevice switching- energy density. We show that even the bestpossible values of E, cannot support a large and fast SS system.

5.1.1. Speed and size

The speed and size of both systems are governed by Eq. (5).With respect to SBP, POEM systems may enjoy three orders ofmagnitude advantages over SS since the POEM machines usediffractive optics for global connections, while in SS all inter-connects are implemented with refractive optics. Using multi-level -phase holograms, the SBP of diffractive interconnects canbe as large as 1011. On the other hand, lens -based refractiveinterconnects have a SBP of at most 108. The large SBP ofholographic interconnects is used in the POEM architecture toachieve a larger ratio A/Ac in Eq. (2) while retaining a highdegree of concurrency and allowing reasonable area to imple-ment electronic signal processing.

We now consider the information handling capacity of POEMmachines for two different optoelectronic technologies: Si/PLZT9and Si or GaAs IC integrated with MQW modulators.19 Weassume that the processing element is a simple one -bit processorcapable of performing logic, data movement, conditional exe-cution, and communication instructions and has 128 bits of localmemory. Such a processing element can be implemented withabout 104 transistors, resulting in a square area of 105 1.Lm2 using0.5 µm CMOS technology. This number is calculated based onthe layout of a prototype PE designed with 2.5 µm minimumfeature in 1 mm2 area. We also assume that the size of theprocessor plane is limited to 6 in. x 6 in. by the wafer size.Then, there will be 250,000 PEs on the processor plane. Tocompute the operating speed, we assume that the maximumpower dissipation density is 10 W /cm2 and perform our calcu-lation for the worst -case condition when all devices on the waferdissipate the same switching energy as is required to drive theoptical devices. Si/PLZT technology requires 1 pJ /µm2 switch-ing energy density for the PLZT modulators and a typical mod-ulator occupies a 10 µm2 area.9 Using Eq. (5) we obtain max-imum N2T -1 of 1016 operations /s. If we use an area of 105 µm2( = 104 x 10 µm2) to host the PEs associated with every mod-ulator, then the throughput is reduced to 1012 (= 1016 /104)operations/s. Thus, assuming 100% yield on a 6 in. wafer, 250,000globally interconnected PEs, each occupying an area of 105 µm2,can be operated at megahertz rates.

Assuming Si /MQW or GaAs /MQW integration technologyto be available, a similar calculation reveals that one can im-plement 250,000 PEs, all communicating with one another at arate of 0.1 to 1 GHz, because E, = 10 fJ /µm2, which is smallerthan E, = 1 pJ /µm2 for Si/PLZT. Note that the fundamentallimits on the speed achievable with POEM systems is limitedto a few gigahertz by the skew associated with the optical transittime in the global holographic interconnects. Also, the PEs canperform local computations at rates higher than the communi-cation speed.

Holographic interconnects cannot be used in the case of SSbecause of the pulse spreading they introduce at very high speedoperation. Therefore, refractive interconnects must be used inSS, limiting SBP to 108. As can be seen from Eq. (4), usingMQW technology with N2T -1 < 1015, only 1 million opticaldevices will be allowed to operate at a maximum switching rateof 1 GHz. Devices under development that require E, = 1 fl/µm2 at a 10 ps switching rate will allow a maximum opticalgate array size of 316 x 316. In addition to the device size lim-itations, systems using such high speed devices will be limitedby the optical transit time as given by Eq. (18).

5.1.2. Complex and multiple rules

One can alleviate some of the speed and size inefficiencies thataccompany a single rule system such as Murdocca's by usingcomplex and /or multiple rules operating in parallel.8'2 How-ever, complex and multiple rules increase the energy consump-tion and decrease the speed of the system, as shown in Eq. (15).There are additional constraints on the complexity and the totalnumber of rules in an SS system. For example, the m parameterof an SS rule cannot exceed the dynamic range of the thresh -olding device. For proper recognition, the thresholding devicesmust be able to distinguish an input light intensity of em from[1 - (1 /m)]ein. The n parameter is limited by the gain of theamplifier device. As seen in Eq. (13), n imposes a lower limiton G. Increasing the number of rules in the system increases theoptical transit time and the size of the system and dissipatesmore energy. For a large number of rules R, the increase inenergy dissipation is by a factor of R as can be seen fromEq. (15). The optical transit time increases as we stack manyimages of the binary plane in free space. These multiple imagesalso increase the total volume of the system.

Based on the above discussions, complex/multiple rules arenot favored in an SS system. As a consequence, we cannot easilyeliminate the size and speed inefficiencies that accompany theimplementation of basic operations by SS rules.

5.2. Architectural considerationsSince technological considerations do not favor highly complexrules, the communication in an SS system is essentially local.In Sec. 3 we showed that architecturally, an SS system is notmore powerful than a mesh. In this section, we argue that themesh architecture is not always an efficient network topologyfor parallel computation, even though it is easy to implement.We can map certain problems efficiently onto a mesh by usinghighly regular algorithms, consequently facilitating very fastcommunication. But these highly local interconnections limit theperformance of many algorithms. Any algorithm whose outputdepends on almost all of the inputs requires at least N time steps,the diameter of the mesh. On the other hand, networks such asthe hypercube have log(N) diameter. Even though the com-munication in these cube -like architectures tends to be slowerthan that of the mesh, diameter consideration indicates that thesenetworks will ultimately be more efficient. Table I shows thatseveral important prototype problems do have more efficientalgorithms on highly interconnected architectures. Hence, thereal questions are whether better communication schemes canbe developed for the cube -like architectures and at what pointit is advantageous to have slower communicating but more glob-ally connected processors.

POEM systems offer a potential solution because their ar-chitecture overcomes many of the limitations faced by highly

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 405

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORS AND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

show that many other important Boolean logic modules require more area and time when implemented with SS.

In Sec. 3 we showed that it requires a large area to implement a basic processor capable of arithmetic and data movement op­ erations. In Sec. 4 we derived the limitation in size and speed of optical gate arrays in terms of thermal considerations. In the following, we show how power considerations limit the size and speed of POEM and SS systems. In particular, we derive some estimates on speed and size based on the value Ec , the minimal device switching-energy density. We show that even the best possible values of Ec cannot support a large and fast SS system.

5.7.7. Speed and sizeThe speed and size of both systems are governed by Eq. (5). With respect to SBP, POEM systems may enjoy three orders of magnitude advantages over SS since the POEM machines use diffractive optics for global connections, while in SS all inter­ connects are implemented with refractive optics. Using multi­ level-phase holograms, the SBP of diffractive interconnects can be as large as 1011 . On the other hand, lens-based refractive interconnects have a SBP of at most 108 . The large SBP of holographic interconnects is used in the POEM architecture to achieve a larger ratio A/AC in Eq. (2) while retaining a high degree of concurrency and allowing reasonable area to imple­ ment electronic signal processing.

We now consider the information handling capacity of POEM machines for two different optoelectronic technologies: Si/PLZT9 and Si or GaAs 1C integrated with MQW modulators. 19 We assume that the processing element is a simple one-bit processor capable of performing logic, data movement, conditional exe­ cution, and communication instructions and has 128 bits of local memory. Such a processing element can be implemented with about 104 transistors, resulting in a square area of 105 jjim2 using 0.5 jxm CMOS technology. This number is calculated based on the layout of a prototype PE designed with 2.5 |jim minimum feature in 1 mm2 area. We also assume that the size of the processor plane is limited to 6 in. x 6 in. by the wafer size. Then, there will be 250,000 PEs on the processor plane. To compute the operating speed, we assume that the maximum power dissipation density is 10 W/cm2 and perform our calcu­ lation for the worst-case condition when all devices on the wafer dissipate the same switching energy as is required to drive the optical devices. Si/PLZT technology requires 1 pJ/jjim2 switch­ ing energy density for the PLZT modulators and a typical mod­ ulator occupies a 10 |xm2 area. 9 Using Eq. (5) we obtain max­ imum N2T~ l of 10 16 operations/s. If we use an area of 105 jmm2 (= 104 x 10 (xm2) to host the PEs associated with every mod­ ulator, then the throughput is reduced to 1012 (= 10 16/104) operations/s. Thus, assuming 100% yield on a 6 in. wafer, 250,000 globally interconnected PEs, each occupying an area of 105 jxm2 , can be operated at megahertz rates.

Assuming Si/MQW or GaAs/MQW integration technology to be available, a similar calculation reveals that one can im­ plement 250,000 PEs, all communicating with one another at a rate of 0.1 to 1 GHz, because Ec = 10 fj/|xm2 , which is smaller than Ec = 1 pJ/jmm2 for Si/PLZT. Note that the fundamental limits on the speed achievable with POEM systems is limited to a few gigahertz by the skew associated with the optical transit time in the global holographic interconnects. Also, the PEs can perform local computations at rates higher than the communi­ cation speed.

Holographic interconnects cannot be used in the case of SS because of the pulse spreading they introduce at very high speed operation. Therefore, refractive interconnects must be used in SS, limiting SBP to 108 . As can be seen from Eq. (4), using MQW technology with N2T -1 < 10 15 , only 1 million optical devices will be allowed to operate at a maximum switching rate of 1 GHz. Devices under development that require Ec = 1 fJ/ jjim2 at a 10 ps switching rate will allow a maximum optical gate array size of 316x316. In addition to the device size lim­ itations, systems using such high speed devices will be limited by the optical transit time as given by Eq. (18).

5.7.2. Complex and multiple rulesOne can alleviate some of the speed and size inefficiencies that accompany a single rule system such as Murdocca's by using complex and/or multiple rules operating in parallel. 8 '2 How­ ever, complex and multiple rules increase the energy consump­ tion and decrease the speed of the system, as shown in Eq. (15). There are additional constraints on the complexity and the total number of rules in an SS system. For example, the m parameter of an SS rule cannot exceed the dynamic range of the thresh­ olding device. For proper recognition, the thresholding devices must be able to distinguish an input light intensity of ein from [1 (l/m)]ein . The n parameter is limited by the gain of the amplifier device. As seen in Eq. (13), n imposes a lower limit on G. Increasing the number of rules in the system increases the optical transit time and the size of the system and dissipates more energy. For a large number of rules R, the increase in energy dissipation is by a factor of R as can be seen from Eq. (15). The optical transit time increases as we stack many images of the binary plane in free space. These multiple images also increase the total volume of the system.

Based on the above discussions, complex/multiple rules are not favored in an SS system. As a consequence, we cannot easily eliminate the size and speed inefficiencies that accompany the implementation of basic operations by SS rules.

5.2. Architectural considerationsSince technological considerations do not favor highly complex rules, the communication in an SS system is essentially local. In Sec. 3 we showed that architecturally, an SS system is not more powerful than a mesh. In this section, we argue that the mesh architecture is not always an efficient network topology for parallel computation, even though it is easy to implement. We can map certain problems efficiently onto a mesh by using highly regular algorithms, consequently facilitating very fast communication. But these highly local interconnections limit the performance of many algorithms. Any algorithm whose output depends on almost all of the inputs requires at least N time steps, the diameter of the mesh. On the other hand, networks such as the hypercube have log(N) diameter. Even though the com­ munication in these cube-like architectures tends to be slower than that of the mesh, diameter consideration indicates that these networks will ultimately be more efficient. Table I shows that several important prototype problems do have more efficient algorithms on highly interconnected architectures. Hence, the real questions are whether better communication schemes can be developed for the cube-like architectures and at what point it is advantageous to have slower communicating but more glob­ ally connected processors.

POEM systems offer a potential solution because their ar­ chitecture overcomes many of the limitations faced by highly

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 405

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 11: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

KIAMILEV, ESENER, PATURI, FAINMAN, MERCIER, GUEST, LEE

TABLE I. Algorithmic performance 23

SIMDProblem Input SIMD mesh hypercube

Matrix multiply 2 N x N matrices 0(N) O(IogN)Sorting N2 elements 0(N) O(Iog2N)Connected

components N vertex graph 0(0T1) 0(Iog2N)Fast Fourier

transform N element vector 0(N) 0(IogN)

interconnected electronic architectures such as the hypercube.POEM architecture provides a flexible, fast, and parallel envi-ronment through its programmable global optical interconnects.

First, POEM systems can handle half a million PEs on twohighly interconnected wafers, as discussed in Sec. 5.1. Thislarge number and very high density of interconnections is a directresult of the 3 -D nature of the POEM architecture. Although theestimated number of PEs in POEM is already quite impressive,one can envision that the number can be further increased usingmore PE planes and interconnection holograms. Additionally,with multiple PE planes the processor grain size can be increasedwithout reducing the overall number of processors in the system.

Second, optical interconnects provide fast means of globalcommunication with low energy requirements. As a result, POEMsystems can fully use the advantage of highly interconnectedarchitectures.

Third, the topology of the interconnects in POEM is notrestricted to being regular. Space- variant interconnectionholograms24'25 allow arbitrary and irregular communication be-tween processors. In fact, the need for such communication issupported by the theory of parallel algorithms, which shows thatfast arallel algorithms require irregular communication amongPEs .3,26

Finally, POEM architecture allows for programmable inter-connections. This reduces the silicon area required by the routerscommonly used in electronic concurrent computers. In addition,such programmable interconnections are desirable since differentalgorithms dictate different interconnections for efficient imple-mentation.

5.3. Other considerationsIn the following subsections we compare local communication,programming methodologies, and the resistance to technologicaldefects in POEM and in SS.

5.3.1. Local communicationIn this subsection we show that "simulated wires" used in SSare much slower than electrical wires used in POEM systems.The idea of "simulated wires" is repeated application of a ruleto move information across the plane. In fact, one applicationof a rule of complexity (k,m,n) can move information by adistance of at most (2k - 1) pixels. If T is the time required toapply the rule and L is the length of the connection in pixels,then

LTwire -

T(2k - 1) (19)

In particular, for Murdocca's rule with k = 3, the movementof data across a distance equal to the length of several gatesrequires hundreds of applications of the rule, because even the

406 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

simplest logic gates require large area when implemented in SS.In contrast, the delay of a short electrical wire is basically

determined by the inverter switching time and is given in Eq. (7).Moving information over a short distance or across many hundredsof logic gates takes essentially the same amount of time. Thisis due to the small size of electronic logic gates and the loga-rithmic dependence of delay on the length of the wire.

To illustrate the above arguments, we compare the time delayin moving information in SS and in POEM. Assuming that thetime to apply one transition rule is 1 ns and the area of a typicallogic gate is 10 x 10 pixels, the time to move information across10 gates is 100 ns. In contrast, a NAND gate, in 1 CMOStechnology, with a 1 mm long output wire, has a delay of about400 ps and can move information across many hundreds of gates.

In summary, the simulated wires of SS are inferior to theirelectronic counterparts in speed. This result indicates that thedata and the control bits that operate on it must be placed closetogether in the plane.

5.3.2. RAM implementation and programmingmethodologies

The programming flexibility of digital electronic computing isclosely associated with its ability to implement RAMs. In thissubsection we show that SS does not provide an efficient meansof RAM implementation, limiting its applications and increasingprogramming complexity.

In electronics, the speed of local interconnects enables theimplementation of small size RAMs with fast access times. Thisenables POEM machines to perform space -time trade -offs andto handle large problems. In particular, it enables POEM ma-chines to perform context switching for solving problems largerthan the size of the machine. Therefore, POEM programmingcan be accomplished using the conventional stored program con-cept. The instructions and the data can be stored in the memoryand executed by a processor.

On the other hand, SS has slow local communication, makingthe implementation of RAM difficult. A RAM has a requirementthat any bit of storage is accessible in one clock cycle. Consideran S2 bit SS RAM. Assuming that the RAM is laid out as a2 -D array of p x p pixel windows and each window stores onebit, the length of the side of the array is S *p. Thus, the longestsimulated wire in this system is about S *p pixels long. Imple-menting even 100 bits of memory with Murdocca's rule or an-other similar rule would require unacceptable access time dueto wire delays. Thus, the programming methodology in SS mustbe different from the processor -memory model used in POEMand at least for now appears to be more difficult and limited inflexibility.

To overcome the lack of efficient communication capability,lack of efficient memory, and complex logic implementations,researchers have proposed to lay out SS programs as circuits,with data and associated control bits placed in close proximity.This approach seems to be harder, at least for now, to use forprogramming because of the difficulty of laying out the com-putation and making sure the timing is properly arranged. Thus,it appears that SS would be more suited for highly structured,local, fine -grain, space -invariant problems. An application ofSS to such a problem has yet to be demonstrated.

5.3.3. Fault toleranceAny fabrication procedure has a certain yield factor. Therefore,the POEM and SS systems must have resistance to technological

KIAMILEV, ESENER, PATURI, FAINMAN, MERGER, GUEST, LEE

TABLE I. Algorithmic performance.2

ProblemSIMD

Input SIMD mesh hypercube

Matrix multiplySortingConnected

componentsFast Fourier

transform

2 N x N matricesN 2 elements

N vertex graph

N element vector

0(N)0(N)

0(VN)

0(N)

O(logN)0(log2 N)

0(log2N)

O(logN)

interconnected electronic architectures such as the hypercube. POEM architecture provides a flexible, fast, and parallel envi­ ronment through its programmable global optical interconnects.

First, POEM systems can handle half a million PEs on two highly interconnected wafers, as discussed in Sec. 5.1. This large number and very high density of interconnections is a direct result of the 3-D nature of the POEM architecture. Although the estimated number of PEs in POEM is already quite impressive, one can envision that the number can be further increased using more PE planes and interconnection holograms. Additionally, with multiple PE planes the processor grain size can be increased without reducing the overall number of processors in the system.

Second, optical interconnects provide fast means of global communication with low energy requirements. As a result, POEM systems can fully use the advantage of highly interconnected architectures.

Third, the topology of the interconnects in POEM is not restricted to being regular. Space- variant interconnection holograms24'25 allow arbitrary and irregular communication be­ tween processors. In fact, the need for such communication is supported by the theory of parallel algorithms, which shows that fast parallel algorithms require irregular communication among

Finally, POEM architecture allows for programmable inter­ connections. This reduces the silicon area required by the routers commonly used in electronic concurrent computers. In addition, such programmable interconnections are desirable since different algorithms dictate different interconnections for efficient imple­ mentation.

5.3. Other considerationsIn the following subsections we compare local communication, programming methodologies, and the resistance to technological defects in POEM and in SS.

5.3.1. Local communicationIn this subsection we show that "simulated wires" used in SS are much slower than electrical wires used in POEM systems. The idea of "simulated wires" is repeated application of a rule to move information across the plane. In fact, one application of a rule of complexity (k,m,n) can move information by a distance of at most (2k 1) pixels. If T is the time required to apply the rule and L is the length of the connection in pixels, then

Twire = (19)

In particular, for Murdocca's rule with k = 3, the movement of data across a distance equal to the length of several gates requires hundreds of applications of the rule, because even the

simplest logic gates require large area when implemented in SS.In contrast, the delay of a short electrical wire is basically

determined by the inverter switching time and is given in Eq. (7). Moving information over a short distance or across many hundreds of logic gates takes essentially the same amount of time. This is due to the small size of electronic logic gates and the loga­ rithmic dependence of delay on the length of the wire.

To illustrate the above arguments, we compare the time delay in moving information in SS and in POEM. Assuming that the time to apply one transition rule is 1 ns and the area of a typical logic gate is 10 x 10 pixels, the time to move information across 10 gates is 100 ns. In contrast, a NAND gate, in 1 jxm CMOS technology, with a 1 mm long output wire, has a delay of about 400 ps and can move information across many hundreds of gates.

In summary, the simulated wires of SS are inferior to their electronic counterparts in speed. This result indicates that the data and the control bits that operate on it must be placed close together in the plane.

5.3.2. RAM implementation and programming methodologiesThe programming flexibility of digital electronic computing is closely associated with its ability to implement RAMs. In this subsection we show that SS does not provide an efficient means of RAM implementation, limiting its applications and increasing programming complexity.

In electronics, the speed of local interconnects enables the implementation of small size RAMs with fast access times. This enables POEM machines to perform space-time trade-offs and to handle large problems. In particular, it enables POEM ma­ chines to perform context switching for solving problems larger than the size of the machine. Therefore, POEM programming can be accomplished using the conventional stored program con­ cept. The instructions and the data can be stored in the memory and executed by a processor.

On the other hand, SS has slow local communication, making the implementation of RAM difficult. A RAM has a requirement that any bit of storage is accessible in one clock cycle. Consider an S2 bit SS RAM. Assuming that the RAM is laid out as a 2-D array of p X p pixel windows and each window stores one bit, the length of the side of the array is S*p. Thus, the longest simulated wire in this system is about S*p pixels long. Imple­ menting even 100 bits of memory with Murdocca's rule or an­ other similar rule would require unacceptable access time due to wire delays. Thus, the programming methodology in SS must be different from the processor-memory model used in POEM and at least for now appears to be more difficult and limited in flexibility.

To overcome the lack of efficient communication capability, lack of efficient memory, and complex logic implementations, researchers have proposed to lay out SS programs as circuits, with data and associated control bits placed in close proximity. This approach seems to be harder, at least for now, to use for programming because of the difficulty of laying out the com­ putation and making sure the timing is properly arranged. Thus, it appears that SS would be more suited for highly structured, local, fine-grain, space-invariant problems. An application of SS to such a problem has yet to be demonstrated.

5.3.3. Fault toleranceAny fabrication procedure has a certain yield factor. Therefore, the POEM and SS systems must have resistance to technological

406 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 12: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORSAND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

TABLE II. Summary of comparison.

Conventional SS

Connection topology /grain sizeMesh /very fine grain

Partial use of advantages of optics:Exploits parallelism and speed but not connectivityNot expandable like VLSI meshSensitive to technological defects

Addressable memory: unknownNo RAM capabilityProgrammability: unknown difficultyNo possibility for space -time trade -offs

HardwareFast devices (ns -ps)Larger power required (thresholding, splitting, shifting,

combining, and masking)Large amount of energy required to implement Boolean

logicMay be suitable for customized systems

POEM

Connection topology /grain sizeAny connection topology, including mesh /fine grain

Full use of major advantages of optics:Exploits connectivity, parallelism, and speed of opticsLimitation in signal skew in global irregular interconnectionsFault tolerant due to global interconnections

Addressable memory: RAMMetal -oxide semiconductor RAM, small storage cells, fast accessProgrammability: conventionalContext switching capability

HardwareSlower devices (100 ns -100 ps)Small power dissipation per device

Smaller energy required per local interconnect and Booleanoperations

Unsolved integration issuesSuitable for general purpose and special purpose

defects. In POEM architecture the global interconnects are space -variant and programmable. Therefore, faulty processors can beeasily bypassed.

Since the interconnections in SS are space invariant and non -programmable, it seems very difficult to implement any faulttolerance. One possible way of handling faults is to arrange thecontrol and data bit placements to avoid defective cells. Althoughthis is possible, it complicates the design of SS systems becauseit introduces additional constraints to the problem of laying outa computation in the SS plane.

6. CONCLUSIONSOur intent in this paper has been to introduce a new optoelec-tronic parallel computing architecture called the programmableoptoelectronic multiprocessor (POEM). The attractive featuresof this architecture have been established by comparing POEMto symbolic substitution (SS), a parallel optical computing sys-tem widely recognized in the research community. The com-parison has included computational efficiency of the architec-tures, power dissipation and speed of the respective supportingtechnologies, ease of programming, and amenability to faulttolerance. A summary of the comparison appears in Table II.

The POEM architecture is motivated by analyses indicatingefficient and effective means of combining optics and electron-ics. Electronics possesses a very mature technology for switchingdevices, and electrical communication is more efficient thanoptical communication for short distances (less than 1 mm).Thus, small to medium grain electronic processors (about 1000gates) form the core of POEM. For the greater distances ofinterprocessor communication, optical link efficiency comparesso favorably with electrical link efficiency that the price paid(in power dissipation and delay) for optoelectronic conversionsis overcome. In contrast, SS is at the extreme of fine -grainedprocessing and pays a high price for having even its shortestlinks implemented optically. All- electronic multiprocessor sys-tems usually represent the other extreme of coarse -grained pro-cessing and squander substantial power and time by driving longwires.

Also, POEM can incorporate complex global patterns of in-terprocessor communication, which is difficult for SS and all-

electronic systems to achieve. The extremely high clock rate ofSS systems prohibits the use of holographic connection elements,which in turn has required space -invariant communication usingrefractive optics. This reduces SS to the equivalent of a 2 -Dmesh -connected architecture, which is well known to be com-putationally inefficient for the solution of many problems. Thoughthe use of multiple and complex substitution rules mediates againstthis limitation, such rules exact a heavy penalty in system powerdissipation and speed. In this respect, our results agree withthose of other researchers20 who show that space -invariant tran-sition rules do not give efficient implementations of basic corn-putational operations. Thus, we observe that some proponentsof SS and similar systems have begun incorporating global in-terconnections into their designs.27' 8 In particular, recently thearchitecture of SS was modified by the addition of global space -invariant crossover interconnection between the pixels and space -variant custom masks.29 While the hardware remains the same,the computational characteristics of the new modified SS archi-tecture are markedly different from conventional SS. The newSS architecture is no longer equivalent to a mesh in topology,because it uses the connectivity of optics. It has been appliedto realize programmable logic arrays (PLAs),30 an optical RAMdecoder circuit 31 and a two -channel sorting node for a widebanddigital switch.39 Although the new SS architecture may over-come some interconnection problems, it does not directly addressthe following problems: (1) The energy budget of the new SSsystem has not been worked out to determine the possible sizeand speed of the system. (2) In the PLA designed with the newSS approach, all of the gates in the PLA are dissipating energy,whereas in electronic PLAs, only the gates that actually imple-ment the function dissipate energy. Thus, making large opticalPLAs is likely to be energy inefficient. (3) Although the newSS has been applied to realize a PLA, it is questionable whethera PLA is the most efficient method to realize a logic functionin terms of gate count and speed. In fact, the number of opticalgates that are unused in computing a Boolean function increaseswith the complexity of the function. (4) It is questionable whetherusing a regular interconnect to implement irregular functions isa computationally efficient method. Based on these considera-tions, another comparison study of the new SS architecture withPOEM is underway.

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 407

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORS AND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

TABLE II. Summary of comparison.

Conventional SS

Connection topology/grain sizeMesh/very fine grain

Partial use of advantages of optics:Exploits parallelism and speed but not connectivityNot expandable like VLSI meshSensitive to technological defects

Addressable memory: unknown No RAM capability Programmability: unknown difficulty No possibility for space-time trade-offs

HardwareFast devices (ns-ps) Larger power required (thresholding, splitting, shifting,

combining, and masking) Large amount of energy required to implement Boolean

logic May be suitable for customized systems

POEM

Connection topology/grain sizeAny connection topology, including mesh/fine grain

Full use of major advantages of optics:Exploits connectivity, parallelism, and speed of opticsLimitation in signal skew in global irregular interconnectionsFault tolerant due to global interconnections

Addressable memory: RAMMetal-oxide semiconductor RAM, small storage cells, fast access Programmability: conventional Context switching capability

HardwareSlower devices (100 ns-100 ps) Small power dissipation per device

Smaller energy required per local interconnect and Booleanoperations

Unsolved integration issues Suitable for general purpose and special purpose

defects. In POEM architecture the global interconnects are space- variant and programmable. Therefore, faulty processors can be easily bypassed.

Since the interconnections in SS are space invariant and non­ programmable, it seems very difficult to implement any fault tolerance. One possible way of handling faults is to arrange the control and data bit placements to avoid defective cells. Although this is possible, it complicates the design of SS systems because it introduces additional constraints to the problem of laying out a computation in the SS plane.

6. CONCLUSIONSOur intent in this paper has been to introduce a new optoelec- tronic parallel computing architecture called the programmable optoelectronic multiprocessor (POEM). The attractive features of this architecture have been established by comparing POEM to symbolic substitution (SS), a parallel optical computing sys­ tem widely recognized in the research community. The com­ parison has included computational efficiency of the architec­ tures, power dissipation and speed of the respective supporting technologies, ease of programming, and amenability to fault tolerance. A summary of the comparison appears in Table II.

The POEM architecture is motivated by analyses indicating efficient and effective means of combining optics and electron­ ics. Electronics possesses a very mature technology for switching devices, and electrical communication is more efficient than optical communication for short distances (less than 1 mm). Thus, small to medium grain electronic processors (about 1000 gates) form the core of POEM. For the greater distances of interprocessor communication, optical link efficiency compares so favorably with electrical link efficiency that the price paid (in power dissipation and delay) for optoelectronic conversions is overcome. In contrast, SS is at the extreme of fine-grained processing and pays a high price for having even its shortest links implemented optically. All-electronic multiprocessor sys­ tems usually represent the other extreme of coarse-grained pro­ cessing and squander substantial power and time by driving long wires.

Also, POEM can incorporate complex global patterns of in­ terprocessor communication, which is difficult for SS and all-

electronic systems to achieve. The extremely high clock rate of SS systems prohibits the use of holographic connection elements, which in turn has required space-invariant communication using refractive optics. This reduces SS to the equivalent of a 2-D mesh-connected architecture, which is well known to be com­ putationally inefficient for the solution of many problems. Though the use of multiple and complex substitution rules mediates against this limitation, such rules exact a heavy penalty in system power dissipation and speed. In this respect, our results agree with those of other researchers20 who show that space-invariant tran­ sition rules do not give efficient implementations of basic com­ putational operations. Thus, we observe that some proponents of SS and similar systems have begun incorporating global in­ terconnections into their designs. 27 ' In particular, recently the architecture of SS was modified by the addition of global space- invariant crossover interconnection between the pixels and space- variant custom masks. 29 While the hardware remains the same, the computational characteristics of the new modified SS archi­ tecture are markedly different from conventional SS. The new SS architecture is no longer equivalent to a mesh in topology, because it uses the connectivity of optics. It has been applied to realize programmable logic arrays (PLAs),30 an optical RAM decoder circuit.31 and a two-channel sorting node for a wideband digital switch. 9 Although the new SS architecture may over­ come some interconnection problems, it does not directly address the following problems: (1) The energy budget of the new SS system has not been worked out to determine the possible size and speed of the system. (2) In the PLA designed with the new SS approach, all of the gates in the PLA are dissipating energy, whereas in electronic PLAs, only the gates that actually imple­ ment the function dissipate energy. Thus, making large optical PLAs is likely to be energy inefficient. (3) Although the new SS has been applied to realize a PLA, it is questionable whether a PLA is the most efficient method to realize a logic function in terms of gate count and speed. In fact, the number of optical gates that are unused in computing a Boolean function increases with the complexity of the function. (4) It is questionable whether using a regular interconnect to implement irregular functions is a computationally efficient method. Based on these considera­ tions, another comparison study of the new SS architecture with POEM is underway.

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 407

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 13: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

KIAMILEV, ESENER, PATURI, FAINMAN, MERCIER, GUEST, LEE

Although the technology supporting POEM is not as fast thatused in SS, its efficient combination of optics and electronicsand its flexible use of global interconnects gives POEM com-putational power greater than that of SS. Efficient local elec-tronic connections used in POEM allow easy implementation ofrandom access memory. This in turn facilitates traditional pro-gramming methodologies. In SS, data and programming infor-mation must be tightly interleaved because communication dis-tances are limited. The space -variant and programmable opticalinterconnects of POEM allow interprocessor connection topol-ogies that are more efficient than mesh connection and easilyaccommodate fault tolerance through bypassing defective pro-cessors. The space -invariant connections of SS make this dif-ficult to do.

Although the computational performance of POEM using ex-isting technology is already competitive with any other system,it can be expected to improve steadily. Current limitations toPOEM performance are technological: the speed of electronicprocessors and optical modulators, both of which are being ac-tively developed. SS, by relying heavily on high speed for com-putational power, already faces fundamental limits in devicepower dissipation and signal skew due to optical propagation.

POEM architecture is well suited for parallel processing witha variety of processor granularity, synchrony, and interconnec-tion topology. It combines the power of parallel space -variantoptical communication with the flexibility and efficiency of elec-tronics. The fast, global, and programmable interconnections ofPOEM will enhance significantly the capabilities and applicationrange of parallel computing.

7. ACKNOWLEDGMENTS

This work has been funded by the Defense Advanced ResearchProjects Agency under the management of the Air Force Officeof Scientific Research under contract No. 88 -0022 and by theOffice of Naval Research under contract No. N00014 -86 -K-0697. Additional funding has been provided through the Officeof Naval Research Graduate Fellowship program.

8. REFERENCES1. J. W. Goodman, F. I. Leonberger, S. Y. Kung, and R. A. Athale, "Optical

interconnections for VLSI systems," Proc. IEEE 72, 850 (1984).2. L. A. Bergman, W. H. Wu, A. R. Johnston, R. Nixon, S. C. Esener, C. C.

Guest, P. K. Yu, T. J. Drabik, M. R. Feldman, and S. H. Lee, "Holo-graphic optical interconnects for VLSI," Opt. Eng. 25(10), 1009 -1118(1986).

3. W. H. Wu, L. A. Bergman, A. R. Johnston, C. C. Guest, S. C. Esener,P. K. Yu, M. R. Feldman, and S. H. Lee, "Implementation of opticalinterconnections for VLSI," IEEE Trans. Electron Devices 34(3), 706 (1987).

4. R. Barakat and J. Reif, "Lower bounds on the computational efficiency ofoptical computing systems," Appl. Opt. 26, 1015 (1987).

5. F. Kiamilev, S. Esener, Y. Fainman, and S. H. Lee, "Programmable opto-electronic multiprocessors," submitted to J. Opt. Soc. Am. A.

6. A. Huang, "Parallel algorithms for optical digital computers," in Proc.IEEE 1983 Tenth International Optical Computing Conference, p. 13 (1983).

7. K. -H. Brenner, A. Huang, and N. Streibl, "Digital optical computing withsymbolic substitution," Appl. Opt. 25, 3054 (1986).

8. S. P. Kozaitis, "Higher- ordered rules for symbolic substitution," Opt.Commun. 65, 339 (1988).

9. S. H. Lee, S. C. Esener, M. A. Title, and T. J. Drabik, "Two-dimensionalsilicon/PLZT spatial light modulators: design considerations and technol-ogy," Opt. Eng. 25(2), 250 -260 (1986).

10. J. H. Wang, T. H. Lin, S. C. Esener, S. Dasgupta, and S. H. Lee, "NMOStransistors fabricated by simultaneous laser- assisted crystallization and dif-fusion on silicon on electro -optic PLZT," in Proc. Materials ResearchSociety 1987 Fall Meeting, Vol. 100, pp. 675 -680 (1988).

11. S. E. Fahlman, NEIL: A System for Representing and Using Real -WorldKnowledge, MIT Press, Cambridge, Mass. (1979).

12. F. Kiamilev and S. Esener, "Implementation of NETL knowledge -basesystem with programmable opto- electronic multiprocessor architecture,"presented at OSA Topical Meeting on Optical Computing (Salt Lake City,Utah, Feb. 27 -March 1, 1989).

408 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

13. J. Von Neumann, Theory of Self - Reproducing Automata, edited and com-pleted by A. W. Burks, Univ. of Illinois Press, Urbana (1966).

14. M. J. Murdocca, "Digital optical computing with one -rule cellular auto-mata," Appl. Opt. 26, 682 (1987).

15. J. N. Mait and K. -H. Brenner, "Optical symbolic substitution: systemdesign using phase -only holograms," Appl. Opt. 27, 1692 (1988).

16. K. -H. Brenner, "New implementation of symbolic substitution logic," Appl.Opt. 25(18), 3061 (1986).

17. W. C. Dally, A VLSI Architecture for Concurrent Data Structures, KluwerAcademic, Boston, Mass. (1987).

18. M. R. Feldman, S. C. Esener, C. C. Guest, and S. H. Lee, "Comparisonbetween optical and electrical interconnects based on power and speedconsiderations," Appl. Opt. 27, 1742 (1988).

19. D. A. B. Miller, D. S. Chemla, T. C. Damen, T. H. Wood, C. A. Burrus,Jr., A. C. Gossard, and W. Wiegmann, "The quantum well self- electro-optic effect device: optoelectronic bistability and oscillation and self -li-nearized modulation," IEEE J. Quantum Electron. QE -21, 1462 (1985).

20. T. J. Cloonan, "Performance analysis of optical symbolic substitution,"Appl. Opt. 27, 1701 (1988).

21. S. D. Goodman and W. T. Rhodes, "Symbolic substitution applications toimage processing," Appl. Opt. 27, 1708 (1988).

22. C. D. Capps, R. A. Falk, and T. L. Houk, "Optical arithmetic /logic unitbased on residue arithmetic and symbolic substitution," Appl. Opt. 27,1682 (1988).

23. M. J. Quinn, Designing Efficient Algorithms for Parallel Computers, McGrawHill, New York (1987).

24. P. Ambs, Y. Fainman, S. Esener, and S. H. Lee, "Holographic opticalelements for SLM defect removal and for optical interconnects," Appl. Opt27(22), 4761 -4765 (1988).

25. B. K. Jenkins, P. Chavel, R. Forcheimer, A. A. Sawchuk, and T. C. Strand,"Architectural implications of a digital optical processor," Appl. Opt. 23,3465 (1984).

26. R. M. Karp and V. Ramachandran, "A survey of parallel algorithms forshared -memory machines," Univ. of California at Berkeley /Computer Sci-ences Dept. 40 (1988).

27. K. S. Huang, K. Jenkins, and A. A. Sawchuk, "Binary image algebrarepresentation of optical cellular logic and symbolic substitution," J. Opt.Soc. Am. A. 4(12), 87 (1987).

28. J. Taboury, J. M. Wang, P. Chavel, F. Devos, and P. Garda, "Opticalcellular processor architecture. 1. Principles," Appl. Opt. 27, 1643 (1988).

29. M. J. Murdocca and A. Huang, "Symbolic substitution methods for opticalcomputing," in Optical Computing 88, J. W. Goodman, P. Chavel, andG. Roblin, eds., Proc. SPIE 963, 585 -590 (1989).

30. M. J. Murdocca, A. Huang, J. Jahns, and N. Streibl, "Optical design ofprogrammable logic arrays," Appl. Opt. 27, 1651 (1988).

31. M. J. Murdocca and B. Sugla, "Design of an optical random access mem-ory," Appl. Opt. 28, 182 (1989).

Fouad E. Kiamilev was born in Moscow,U.S.S.R, in 1965. He graduated magna cumlaude from the University of California at SanDiego in 1988 with BS degrees in computerscience and electrical engineering. Presently,he is enrolled in a Ph.D. program in the De-partment of Electrical and Computer Engi-neering of the University of California at SanDiego. His research interests include parallelcomputing, optical computing, and VLSI sys-tems. Mr. Kiamilev is a fellow of the Office

of Naval Research Graduate Fellowship program.

Sadik C. Esener received the B.S. degree inElectrical Engineering from the TechnicalUniversity of Istanbul, Turkey, 1979; the MSdegree from the University of Michigan; andhis Ph.D. degree in the same discipline fromthe University of California, San Diego, in 1987.He is currently an assistant professor in theDepartment of Electrical and Computer En-gineering at the University of California, SanDiego, where he is engaged in research inoptoelectronic computing architectures, spa-

tial light modulators, optical VLSI interconnects, 3 -D memories, andhigh speed photodetectors. Dr. Esener is a member of SPIE and PhiBeta Delta.

KIAMILEV, ESENER, PATURI. FAINMAN, MERGER, GUEST, LEE

Although the technology supporting POEM is not as fast that used in SS, its efficient combination of optics and electronics and its flexible use of global interconnects gives POEM com­ putational power greater than that of SS. Efficient local elec­ tronic connections used in POEM allow easy implementation of random access memory. This in turn facilitates traditional pro­ gramming methodologies. In SS, data and programming infor­ mation must be tightly interleaved because communication dis­ tances are limited. The space-variant and programmable optical interconnects of POEM allow interprocessor connection topol­ ogies that are more efficient than mesh connection and easily accommodate fault tolerance through bypassing defective pro­ cessors. The space-invariant connections of SS make this dif­ ficult to do.

Although the computational performance of POEM using ex­ isting technology is already competitive with any other system, it can be expected to improve steadily. Current limitations to POEM performance are technological: the speed of electronic processors and optical modulators, both of which are being ac­ tively developed. SS, by relying heavily on high speed for com­ putational power, already faces fundamental limits in device power dissipation and signal skew due to optical propagation.

POEM architecture is well suited for parallel processing with a variety of processor granularity, synchrony, and interconnec­ tion topology. It combines the power of parallel space-variant optical communication with the flexibility and efficiency of elec­ tronics. The fast, global, and programmable interconnections of POEM will enhance significantly the capabilities and application range of parallel computing.

7. ACKNOWLEDGMENTSThis work has been funded by the Defense Advanced Research Projects Agency under the management of the Air Force Office of Scientific Research under contract No. 88-0022 and by the Office of Naval Research under contract No. N00014-86-K- 0697. Additional funding has been provided through the Office of Naval Research Graduate Fellowship program.

8. REFERENCES1. J. W. Goodman, F. I. Leonberger, S. Y. Kung, andR. A. Athale, "Optical

interconnections for VLSI systems," Proc. IEEE 72, 850 (1984).2. L. A. Bergman, W. H. Wu, A. R. Johnston, R. Nixon, S. C. Esener, C. C.

Guest, P. K. Yu, T. J. Drabik, M. R. Feldman, and S. H. Lee, "Holo­ graphic optical interconnects for VLSI," Opt. Eng. 25(10), 1009-1118 (1986).

3. W. H. Wu, L. A. Bergman, A. R. Johnston, C. C. Guest, S. C. Esener, P. K. Yu, M. R. Feldman, and S. H. Lee, "Implementation of optical interconnections for VLSI," IEEE Trans. Electron Devices 34(3), 706 (1987).

4. R. Barakat and J. Reif, *'Lower bounds on the computational efficiency of optical computing systems," Appl. Opt. 26, 1015 (1987).

5. F. Kiamilev, S. Esener, Y. Fainman, and S. H. Lee, "Programmable opto­ electronic multiprocessors," submitted to J. Opt. Soc. Am. A.

6. A. Huang, "Parallel algorithms for optical digital computers," in Proc.IEEE 1983 Tenth International Optical Computing Conference, p. 13 (1983).

1. K.-H. Brenner, A. Huang, and N. Streibl, "Digital optical computing withsymbolic substitution," Appl. Opt. 25, 3054 (1986).

8. S. P. Kozaitis, "Higher-ordered rules for symbolic substitution," Opt. Commun. 65, 339 (1988).

9. S. H. Lee, S. C. Esener, M. A. Title, andT. J. Drabik, "Two-dimensional silicon/PLZT spatial light modulators: design considerations and technol­ ogy," Opt. Eng. 25(2), 250-260 (1986).

10. J. H. Wang, T. H. Lin, S. C. Esener, S. Dasgupta, and S. H. Lee, "NMOS transistors fabricated by simultaneous laser-assisted crystallization and dif­ fusion on silicon on electro-optic PLZT," in Proc. Materials Research Society 1987 Fall Meeting, Vol. 100, pp. 675-680 (1988).

11. S. E. Fahlman, NETL: A System for Representing and Using Real-World Knowledge, MIT Press, Cambridge, Mass. (1979).

12. F. Kiamilev and S. Esener, "Implementation of NETL knowledge-base system with programmable opto-electronic multiprocessor architecture," presented at OS A Topical Meeting on Optical Computing (Salt Lake City, Utah, Feb. 27-March 1, 1989).

13. J. Von Neumann, Theory of Self-Reproducing Automata, edited and com­ pleted by A. W. Burks, Univ. of Illinois Press, Urbana (1966).

14. M. J. Murdocca, "Digital optical computing with one-rule cellular auto­ mata," Appl. Opt. 26, 682 (1987).

15. J. N. Mait and K.-H. Brenner, "Optical symbolic substitution: system design using phase-only holograms," Appl. Opt. 27, 1692 (1988).

16. K.-H. Brenner, "New implementation of symbolic substitution logic," Appl. Opt. 25(18), 3061 (1986).

17. W. C. Dally, A VLSI Architecture for Concurrent Data Structures, Kluwer Academic, Boston, Mass. (1987).

18. M. R. Feldman, S. C. Esener, C. C. Guest, and S. H. Lee, "Comparison between optical and electrical interconnects based on power and speed considerations," Appl. Opt. 27, 1742 (1988).

19. D. A. B. Miller, D. S. Chemla, T. C. Damen, T. H. Wood, C. A. Burrus, Jr., A. C. Gossard, and W. Wiegmann, "The quantum well self-electro- optic effect device: optoelectronic bistability and oscillation and self-li­ nearized modulation," IEEE J. Quantum Electron. QE-21, 1462 (1985).

20. T. J. Cloonan, "Performance analysis of optical symbolic substitution," Appl. Opt. 27, 1701 (1988).

21. S. D. Goodman and W. T. Rhodes, "Symbolic substitution applications to image processing," Appl. Opt. 27, 1708 (1988).

22. C. D. Capps, R. A. Falk, and T. L. Houk, "Optical arithmetic/logic unit based on residue arithmetic and symbolic substitution," Appl. Opt. 27, 1682 (1988).

23. M. J. Quinn, Designing Efficient Algorithms for Parallel Computers, McGraw Hill, New York (1987).

24. P. Ambs, Y. Fainman, S. Esener, and S. H. Lee, "Holographic optical elements for SLM defect removal and for optical interconnects," Appl. Opt 27(22), 4761-4765 (1988).

25. B. K. Jenkins, P. Chavel, R. Forcheimer, A. A. Sawchuk, and T. C. Strand, "Architectural implications of a digital optical processor," Appl. Opt. 23, 3465 (1984).

26. R. M. Karp and V. Ramachandran, "A survey of parallel algorithms for shared-memory machines," Univ. of California at Berkeley/Computer Sci­ ences Dept. 40 (1988).

27. K. S. Huang, K. Jenkins, and A. A. Sawchuk, "Binary image algebra representation of optical cellular logic and symbolic substitution," J. Opt. Soc. Am. A. 4(12), 87 (1987).

28. J. Taboury, J. M. Wang, P. Chavel, F. Devos, and P. Garda, "Optical cellular processor architecture. 1. Principles," Appl. Opt. 27, 1643 (1988).

29. M. J. Murdocca and A. Huang, "Symbolic substitution methods for optical computing," in Optical Computing 88, J. W. Goodman, P. Chavel, and G. Roblin, eds., Proc. SPIE 963, 585-590 (1989).

30. M. J. Murdocca, A. Huang, J. Jahns, and N. Streibl, "Optical design of programmable logic arrays," Appl. Opt. 27, 1651 (1988).

31. M. J. Murdocca and B. Sugla, "Design of an optical random access mem­ ory," Appl. Opt. 28, 182 (1989). 3

Fouad E. Kiamilev was born in Moscow, U.S.S.R, in 1965. He graduated magna cum laude from the University of California at San Diego in 1988 with BS degrees in computer science and electrical engineering. Presently, he is enrolled in a Ph.D. program in the De­ partment of Electrical and Computer Engi­ neering of the University of California at San Diego. His research interests include parallel computing, optical computing, and VLSI sys­ tems. Mr. Kiamilev is a fellow of the Office

of Naval Research Graduate Fellowship program.

Sadik C. Esener received the B.S. degree inElectrical Engineering from the Technical University of Istanbul, Turkey, 1979; the MS degree from the University of Michigan; and his Ph.D. degree in the same discipline from the University of California, San Diego, in 1987. He is currently an assistant professor in the Department of Electrical and Computer En­ gineering at the University of California, San Diego, where he is engaged in research in optoelectronic computing architectures, spa­

tial light modulators, optical VLSI interconnects, 3-D memories, and high speed photodetectors. Dr. Esener is a member of SPIE and Phi Beta Delta.

408 / OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms

Page 14: Programmable optoelectronic multiprocessors and their comparison with symbolic ...paturi/myPapers/pubs/KiamilevEs... · 2013-10-03 · Programmable optoelectronic multiprocessors

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORSAND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

Ramamohan Paturi is an assistant professorof computer science and engineering at theUniversity of California, San Diego. His cur-rent research interests include parallel com-putation, optical interconnects, and learningwith threshold networks. Dr. Paturi receivedhis BE degree in electronics and communi-cation engineering from Andhra University,Madras, India, in 1981 and his Ph.D. degreein computer science from the PennsylvaniaState University in 1985. He is a member of

the ACM and IEEE Computer Society.

Yeshaiahu Fainman is an associate professorin the Department of Mechanical Engineeringand Applied Mechanics, College of Engi-neering, of the University of Michigan, AnnArbor. He earned a doctorate from the Tech -nion- Israel Institute of Technology in 1983and his master's degree in electrical engi-neering from the Technion -Israel in 1979. From1983 to 1988 he was an assistant researchphysicist at the University of California, SanDiego. His research has been in the area of

optical information processing, holography, and photorefractive op-tics, and he has published more than 25 papers and contributed achapter to a book in the field. He is a member of OSA and has receivedhonors and awards.

P. Mercier: Biography and photograph not available.

Clark C. Guest is an assistant professor in theDepartment of Electrical and Computer En-gineering at the University of California, SanDiego, where he has been since 1984. He re-ceived his BS and MSEE degrees from RiceUniversity in 1975 and 1976, respectively. Hereceived his Ph.D. from the Georgia Instituteof Technology in 1983. His current researchinterests include optical interconnects, neu-rocomputing, computer generated hologra-phy, and pattern recognition.

Sing H. Lee: Biography and photograph appear with the guest editorialin this issue.

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 409

PROGRAMMABLE OPTOELECTRONIC MULTIPROCESSORS AND THEIR COMPARISON WITH SYMBOLIC SUBSTITUTION FOR DIGITAL OPTICAL COMPUTING

Ramamohan Paturi is an assistant professor of computer science and engineering at the University of California, San Diego. His cur­ rent research interests include parallel com­ putation, optical interconnects, and learning with threshold networks. Dr. Paturi received his BE degree in electronics and communi­ cation engineering from Andhra University, Madras, India, in 1981 and his Ph.D. degree in computer science from the Pennsylvania State University in 1985. He is a member of

the ACM and IEEE Computer Society.

Yeshaiahu Fainman is an associate professor in the Department of Mechanical Engineering and Applied Mechanics, College of Engi­ neering, of the University of Michigan, Ann Arbor. He earned a doctorate from the Tech- nion-lsrael Institute of Technology in 1983 and his master's degree in electrical engi­ neering from the Technion-lsrael in 1979. From 1983 to 1988 he was an assistant research physicist at the University of California, San Diego. His research has been in the area of

optical information processing, holography, and photo refractive op­ tics, and he has published more than 25 papers and contributed a chapter to a book in the field. He is a member of OSA and has received honors and awards.

P. Mercier: Biography and photograph not available.

Clark C. Guest is an assistant professor in the Department of Electrical and Computer En­ gineering at the University of California, San Diego, where he has been since 1984. He re­ ceived his BS and MSEE degrees from Rice University in 1975 and 1976, respectively. He received his Ph.D. from the Georgia Institute of Technology in 1983. His current research interests include optical interconnects, neu- rocomputing, computer generated hologra­ phy, and pattern recognition.

Sing H. Lee: Biography and photograph appear with the guest editorial in this issue.

OPTICAL ENGINEERING / April 1989 / Vol. 28 No. 4 / 409

Downloaded From: http://opticalengineering.spiedigitallibrary.org/ on 10/02/2013 Terms of Use: http://spiedl.org/terms