SlideShare una empresa de Scribd logo
1 de 15
Optical computing

Optical computing means performing computations, operations, storage and
transmission of data using light. Instead of silicon chips optical computer uses organic
polymers like phthalocyanine and polydiacetylene.Optical technology promises massive
upgrades in the efficiency and speed of computers, as well as significant shrinkage in
their size and cost. An optical desktop computer is capable of processing data up to
1,00,000 times faster than current models.


Advantages of Optical Computers
Many of the advantages stated below cannot be put into practice today but are
theoretical advantages of the optical technology. Quite the contrary some of the
advantages mentioned are not advantages nowadays but rather disadvantages due to
the early stages of the development: For example a prototype of an optical computer is
bigger than a conventional computer. However the theoretical advantages can be
realized sooner or later.

Optical computers will only be produced if they are cheaper or more powerful than
conventional ones (or best: both of it at the same time). Besides hybrid solutions there
will be a competition between the two different kinds of computing. To be able to
measure optical computers and their future possibilities, conventional computers and
their actual stage of development concerning each advantage and disadvantage are
depicted.

Advantages of optical computers compared to conventional computers:

      Higher performance

      The most significant advantage of optical computers is the potential of higher
      performance. A concrete benchmark test between optical and conventional
      computers is not possible yet. But the performance of an optical computer in an
      advanced stage should be several orders of magnitude higher.

      Conventional computer technology is based on electric current and electrons.
      While electric current is very fast, the average drift velocity of electrons is
      rather slow. In reality there is nothing faster than the speed of light. This makes
      light and photons to the perfect information carrier. Even long distances can be
      bridged within split seconds.
A computer based completely on optical components is the optimum
concerning speed: Full optical RAM saving photons directly if possible or
otherwise indirectly, bus systems using concentrated light beams (lasers) to
communicate, optical processors and holographic drives. All those components
have the capability to be much faster than the existing electric ones. For
example a light beam is able to transmit the whole encyclopedia Britannica
within one second.[38] To lose no speed it is reasonable to use electric devices
as rare as possible because every electric communication device is a potential
bottleneck. At the beginning of the development a full optical computer will
not be possible. Many devices will still work electrical. Even after a lot of
development electronic parts most likely will still remain in optical computers,
e.g. for instruction and controlling tasks.

But using light as information carrier is only one opportunity to accelerate
computers. Using a different architecture for optical computers could be a key
factor to success. Conventional computers normally base on the Von Neumann
architecture. That means data and instructions are kept in the same memory, are
variable and are executed sequentially. One serious disadvantage of the Von
Neumann architecture is the bottleneck between the CPU and the RAM, called
the Von Neumann bottleneck. Since the CPU became the fastest device in
modern computers it is forced to wait for data coming from the RAM. Bus
systems and the RAM are not fast enough to provide data for the CPU without
intermission. The effective processing speed is seriously limited. In optical
computers light could be used in bus systems. Optical communication devices
are much faster than electric ones. So the bottleneck could disappear.

Another key to fasten up optical computers is to compute with higher
parallelism. This implies higher performance and higher bandwidth which is
depicted within the next break.

Higher parallelism

There are two options to achieve higher parallelism[39] in computers:

One is to increase the amount of data which is sent through bus systems and
computed in the CPU at each time. Modern conventional computers and
operating systems base on 32 bit or 64 bit architecture. Often application
software only supports 32 bit systems so the software does not take full
advantage of 64 bit systems. Optical computers can be built with higher
bandwidth. Within one data path several data sets can be transmitted parallel at
the same time using different wavelengths or polarizations. But thereby
attention to interference is necessary. Due to the coherent laser light destructive
interference depending on the phase might occur. The higher parallelism and
the superior velocity of light allow extreme processing speeds. But to take full
advantage of this new architecture operating systems and application software
have to be adjusted to it.

Furthermore data paths are able to cross each other without
interference.[40] This advantage of optical technologies can help to build
architectures and layouts with superior parallelism. New layouts can be more
three-dimensional and thus the space in a computer case can be used more
intensively. With adequate miniaturization of optical components the size of
computers could shrink.

The following figure visualizes the possibilities of communication using light
beams:
Figure 15: Superposition, Crossed Data Paths and Three-Dimensionality

            (At a hybrid system using optoelectronic devices)

                             Own illustration



In conventional computers superposition and crossed data paths are not
possible. Only one data set can be sent in one data path at each time and
crossed data paths lead to loss of data. Besides that the layout of a conventional
computer is rather based on two dimensions. Optical computers take advantage
of all three dimensions. No wires are needed to transmit information, quite the
contrary information can be sent wireless or even in vacuum.
Another option to achieve higher parallelism in computers is to change the
 architecture concerning the temporal aspect. As already mentioned modern
 computers established on the Von Neumann architecture work sequential. To
 achieve more performance parallel computing is required. So more than one
 instruction stream or data stream can be executed simultaneously. According to
 Flynn’s taxonomy computers are divided into four groups concerning their
 architecture. Flynn differentiates between data streams and instruction streams.

                             Single Instruction           Multi Instruction
Single Data                  SISD                         MISD
Multiple Data                SIMD                         MIMD

                          Table 3: Flynn’s Taxonomy

          Own illustration based on STUCKE (1989), Pages 29-31



 Personal Computers base on SISD. Vector processors using SIMD architecture are able
 to perform the same instruction on several data sets. Computers designed in MISD
 operate redundantly. Most of the fastest supercomputers of the world base on MIMD
 design. Thus several processors work independently and asynchronously.

 Concerning optical computers multiple data and instruction streams are useful, too. At
 least for mainframe computers an optical MIMD concept should be established.

 Less consumption

 Current computers consume a lot of energy. Modern CPUs often need over 80 watts in
 idle state, around 120 watts in normal use and up to 250 watts in performance
 mode.[41] For complex visualization a high performance graphics card is needed which
 needs up to 150 watts itself. The computer industry realized that issue several years ago.
 Today especially for notebooks energy-saving components are built.

 Calculation example:

 A computer runs every day around eight hours in normal use and one kilowatt hour costs
 0,15 €. That means that the CPU consumes around 350 kilowatt hours a year.
 Consequently 53 € of energy costs are generated only by the CPU. All components of a
loaded high performance PC (without monitor and periphery) consume within the same
time even 1100 kilowatt hours (165 €).

Another aspect is that a lot of the consumed energy is not used accurately but for idle
power in the form of internal friction in the ICs. Still the amount of energy loss due to
internal friction rises from CPU generation to generation. Due to this friction a lot of heat
is released, which will be explained in the next break. Friction between the elementary
particles does not occur when light is used as information carrier. Optical computers have
the potential to be more power-saving than conventional ones.

Less heat is released

Less heat release using light? Sounds contradictory since light sources radiate heat. But in
optical computers lasers are used as light sources. Those concentrated light beams only
consist of a small spectrum of different wavelengths. Depending on the field of
application lasers have different needs of energy and produce heat to a greater or lesser
extent.

Most modern CPUs are not able to work without a proper airing. The reason for that is
the friction of the electrons in the integrated circuits. Those moving electrons hit each
other within wires and ICs and thus heat is generated. Airing a processor (whether it is a
CPU, GPU or the like) or the whole case needs energy, space and produces noise. Optical
computers could be smaller because there is no need for a fan or free spaces for air
circulation. Big computer centers with thousands of computers need powerful air
conditioning systems. These aggregates are expensive and need a lot of energy too.
Besides that the released heat increases the danger of fire at computer centers.

Less noise

Conventional computers often cause a lot of noise due to rotating fans and drives. High
speed processors accelerated to their architectural limits need enormous active and
passive cooling. In the past small fast spinning fans often were used which created plenty
noise. Some years ago the computer industry realized the problem of the annoying noise
and developed fans with bigger size. Thus the same volume flow was established with a
lower rotary speed and the noise development was reduced.

Actually optical computers could be almost noiseless since probably no fan will be
needed. Light sources (e.g. lasers) can be cooled with passive coolers and heat pipes built
out of aluminum or copper. Those passive coolers evacuate heat silent. Low noise is an
interesting aspect for office and home users as well.
More flexibility in layout

Conventional PCs are built as a rectangular box (desktop) or as a laptop. One reason for
that is the speed of electronic connections. It depends on the length of the cables and
pipelines. On the motherboard the CPU, RAM and graphics card have to be close to each
other to be able to move huge amounts of information. Longer distances imply the
decrease of the practical transfer rate.

Using optical components the distance of communication does not matter. Once the
signal is in an optical fiber it does not matter whether the signal runs 1 meter or 1000
meters. Because of the low damping long-range communication is possible. Still the data
rate is very high and there is no crosstalk.

So the optical computer technology has the potential to change the shape and layout of
computers fundamentally.[42] The components of one computer can be spread across a
car, a building or even a city with almost no loss in performance. Consequently the
server/client and the peer-to-peer architectures could be advanced. Many clients,
terminals or even single components can be connected optically and consequently allow
higher ranges.

Less loss in communication

Today communication often is realized with electric wires or wireless by radio frequency.
The ranges of those communication ways are limited. Data sent through wires needs to be
amplified several times to bridge longer distances.

The communication with optical fibers is almost lossless due to the total internal
reflection.[43] So amplifying the signal is not or only rarely needed. Furthermore a
higher bandwidth is possible, optical communication is insensible to electromagnetic
interfering fields and it is more tap-proof. For high performance communication (e.g. for
backbones) already today fiber optics are used.

Less wear

Wear normally occurs at mechanically moving parts. In conventional computers those
parts are especially fans, hard disk drives and conventional optical removable storages
(CD, DVD, HD-DVD, Blu-ray disks). All those components have in common that they
rotate or move very fast and thus friction is caused. Because of this friction the
mechanical parts wear out and break. As already mentioned, this friction causes heat
which is suboptimal, too.
In optical computers fans possibly will not be needed any more. An optical processor
      does not heat up due to internal friction of the electrons like a conventional does.
      Additionally new technologies for mass storages can be established. Saving in the form
      of holograms or on molecular basis is possible. Those forms do not need fast spinning
      parts and do not wear out so heavily.




Disadvantages of Optical Computers
As already mentioned in Chapter 5.2 today’s optical computers are only prototypes
and do not reflect the whole capabilities of optical computing. Consequently many of
the disadvantages of optical computing are not because of the principle of the new
technology itself but because of contemporary production technologies. The machines
and processes that could be used to manufacture optical computers in mass production
are still in its early stages.

Disadvantages of optical computers compared to conventional computers:

      Optical components and their production is still expensive

      Conventional ICs are produced in high tech factories whose sole task is to
      manufacture ICs. Since the specialization and productivity of those factories is
      rather high the price of the product is low. Experts say the price of one
      transistor matches approximately the price of a printed character in a
      newspaper.[44] Earlier the price was a lot higher but due to new abilities in
      manufacturing and the new factories it decreased.

      Concerning optical components that are particularly manufactured for optical
      computers the development of its price could be similar to the development of
      the IC price. Today there are already optical components e.g. for
      communication and measurement. But they are not specially built to be placed
      in an optical computer. Furthermore there is no high tech and specialized
      factory to produce optical computers or especially designed components for it
      comparable to modern IC factories. Thus the price of an optical computer is
      rather high or respectively there is no complete system that can be bought yet.
It is a kind of prototype and is not really manufactured but build by handwork
of experts out of components produced for a different purpose. Using a
standard architecture and standardized components could be the key to realize
mass production of optical computers. This would decrease the production
price and thus the price of an optical computer will drop as well. Unfortunately
only a few companies have the financial power to develop computers with
integrated optics first and then build a matching factory to produce them. IBM,
Infinion respectively Siemens, Samsung, Intel and AMD can be mentioned at
this point. During a scenario analysis in Chapter 6 possible scenarios
concerning the price and performance of optical computers will be depicted
more detailed.

Optical components are not miniaturized enough yet

Existing modern IC processors are built in a very high density called VLSI
(very-large-scale-integration) or ULSI (ultra-large-scale-integration). There are
several millions of transistors in an area of only approximately some square
millimeters. The size of the transistors still decreases and some parts of the
transistors already reach atomic dimensions.

In contrast to that optical components can be build small and compact but not
really miniaturized. The area of optics ranges from macrooptics over miniature
optics to microoptics.[45] Unfortunately there are no microoptic integrated
circuits specifically developed to assemble a CPU or a motherboard yet. A lot
of development and new processes will be needed.[46]

Problems of exact manufacturing

Today conventional processors and computer components are manufactured
with high precision and in huge charges using robust processes. Manufacturing
problems can be eliminated by proper quality assurance. A change from the
existing assembly method concerning the size of the chip structure to another
size or using bigger wavers can trigger problems. For instance AMD recently
had some problems changing their production from 90nm to 65nm
structure.[47]

Miniaturized optical components have to be built very exactly to work
properly. This exactness often is not reached yet. Small deviations can cause
massive problems diverting light beams. Exact production is expensive which
    leads to the next disadvantage.

    New expensive high-tech factories have to be built

    An IC factory costs several millions up to billions of dollars. Chip
    manufactures do not want to close such factories due to the high investments
    unless they have amortized. So the goal of the chip manufactures is to integrate
    optical components into the existing concepts and create hybrid devices. To
    allow integration optical devices have to be adapted and are not able to show
    their whole capabilities using an all optical architecture. So at first the
    integrated optical components will probably be produced in existing IC
    factories. Later own factories can be built to produce sole optical parts.

    Incompatibility

    Modern Personal Computers (PCs) are put together according to the Von
    Neumann architecture. Application software and especially operating systems
    are programmed to match this architecture. Because of these existing standards
    most of the application software can be used on any computer using a
    Microsoft Windows operating system. The portability is rather high.

    Optical computers may use a different architecture particularly regarding the
    parallelism of the system. Conventional existing computer programs working
    sequentially will have to run emulated in a kind of compatibility mode. So
    these programs cannot use the full calculating speed of optical computers or
    even worse some programs might not even work. Here the question arises
    whether the software drives the hardware or the hardware drives the software.
    Concerning optical computers the software will depend on the new architecture
    of the hardware. Software especially programmed for optical computers
    probably might work on conventional computers. The two different
    architectures existing next to each other with incompatible software is the worst
    scenario. Then software can only be used on one architecture so the portability
    and capability is quite low.



5.3 Comparison of the Advantages and Disadvantages
In Chapters 5.1 and 5.2 advantages and disadvantages of optical computing are stated.
Strong advantages like more performance while less consumption and more
ergonomic (less noise, smaller and more flexible cases) stand opposite to
disadvantages of more temporary nature (incompatibility and production problems).
Consequently the advantages of optical computing overtop the disadvantages. Photons
and light seem to be better than electrons and electric current to carry information.

Especially in the long run the teething problems of the new technology will be coped
and most likely the onward development will bring out high performance computers.
The question how long this will take and the factors influencing it will be discussed
within the next chapter.
Optical computing
Optical computing
Optical computing
Optical computing

Más contenido relacionado

La actualidad más candente

Opticalcomputing final
Opticalcomputing finalOpticalcomputing final
Opticalcomputing finaldivyajyothi405
 
Optical Computing Technology
Optical Computing TechnologyOptical Computing Technology
Optical Computing TechnologyKanchan Shinde
 
Optical computing by abhishek mahajan
Optical computing by abhishek mahajanOptical computing by abhishek mahajan
Optical computing by abhishek mahajanAbhishek Mahajan
 
Optical computing hardcopy
Optical computing hardcopyOptical computing hardcopy
Optical computing hardcopydivyajyothi405
 
Nanocomputers or Future computer Nanotechnology
Nanocomputers or Future computer NanotechnologyNanocomputers or Future computer Nanotechnology
Nanocomputers or Future computer NanotechnologyMAGNIFIER
 
TheThingsConference 2019 Slides of Alex Raimondi
TheThingsConference 2019 Slides of Alex RaimondiTheThingsConference 2019 Slides of Alex Raimondi
TheThingsConference 2019 Slides of Alex RaimondiSchekeb Fateh
 
B2 oral exam presenation
B2 oral exam presenationB2 oral exam presenation
B2 oral exam presenationAhmed Kamal
 
Claytronics | Programmable Matter | PPT
Claytronics | Programmable Matter | PPTClaytronics | Programmable Matter | PPT
Claytronics | Programmable Matter | PPTSeminar Links
 
12 nanotechnology
12 nanotechnology12 nanotechnology
12 nanotechnology09dec207
 

La actualidad más candente (20)

Opticalcomputing final
Opticalcomputing finalOpticalcomputing final
Opticalcomputing final
 
Optical Computing Technology
Optical Computing TechnologyOptical Computing Technology
Optical Computing Technology
 
Optical computing by abhishek mahajan
Optical computing by abhishek mahajanOptical computing by abhishek mahajan
Optical computing by abhishek mahajan
 
Optical Computing
Optical ComputingOptical Computing
Optical Computing
 
Optical computers pdf
Optical computers pdfOptical computers pdf
Optical computers pdf
 
Optical computing hardcopy
Optical computing hardcopyOptical computing hardcopy
Optical computing hardcopy
 
Optical computing1
Optical computing1Optical computing1
Optical computing1
 
C017521927
C017521927C017521927
C017521927
 
Kamna
KamnaKamna
Kamna
 
Nano computing
Nano computingNano computing
Nano computing
 
14 576
14 57614 576
14 576
 
Nanocomputers or Future computer Nanotechnology
Nanocomputers or Future computer NanotechnologyNanocomputers or Future computer Nanotechnology
Nanocomputers or Future computer Nanotechnology
 
TheThingsConference 2019 Slides of Alex Raimondi
TheThingsConference 2019 Slides of Alex RaimondiTheThingsConference 2019 Slides of Alex Raimondi
TheThingsConference 2019 Slides of Alex Raimondi
 
B2 oral exam presenation
B2 oral exam presenationB2 oral exam presenation
B2 oral exam presenation
 
Nanocomputers
NanocomputersNanocomputers
Nanocomputers
 
Claytronics | Programmable Matter | PPT
Claytronics | Programmable Matter | PPTClaytronics | Programmable Matter | PPT
Claytronics | Programmable Matter | PPT
 
Programmable Matter
Programmable MatterProgrammable Matter
Programmable Matter
 
Claytronics
ClaytronicsClaytronics
Claytronics
 
Programmable matter and applications
Programmable matter and applicationsProgrammable matter and applications
Programmable matter and applications
 
12 nanotechnology
12 nanotechnology12 nanotechnology
12 nanotechnology
 

Similar a Optical computing

Full report on light peak technology
Full report on light peak technologyFull report on light peak technology
Full report on light peak technologyVIKAS SINGH BHADOURIA
 
Computer's clasification
Computer's clasificationComputer's clasification
Computer's clasificationMayraChF
 
computer Development PPT created by Pooja Khamkar
computer Development PPT created by Pooja Khamkarcomputer Development PPT created by Pooja Khamkar
computer Development PPT created by Pooja Khamkarpoojakhamkar8
 
CC LECTURE NOTES (1).pdf
CC LECTURE NOTES (1).pdfCC LECTURE NOTES (1).pdf
CC LECTURE NOTES (1).pdfHasanAfwaaz1
 
Introduction to parallel_computing
Introduction to parallel_computingIntroduction to parallel_computing
Introduction to parallel_computingMehul Patel
 
IS 139 Lecture 1 - 2015
IS 139 Lecture 1 - 2015IS 139 Lecture 1 - 2015
IS 139 Lecture 1 - 2015Aron Kondoro
 
Pipelining and ILP (Instruction Level Parallelism)
Pipelining and ILP (Instruction Level Parallelism) Pipelining and ILP (Instruction Level Parallelism)
Pipelining and ILP (Instruction Level Parallelism) A B Shinde
 
Michael Gschwind, Cell Broadband Engine: Exploiting multiple levels of parall...
Michael Gschwind, Cell Broadband Engine: Exploiting multiple levels of parall...Michael Gschwind, Cell Broadband Engine: Exploiting multiple levels of parall...
Michael Gschwind, Cell Broadband Engine: Exploiting multiple levels of parall...Michael Gschwind
 
Chapter 1.pptx
Chapter 1.pptxChapter 1.pptx
Chapter 1.pptxTekle12
 
Interconnect technology Final Revision
Interconnect technology Final RevisionInterconnect technology Final Revision
Interconnect technology Final RevisionMatthew Agostinelli
 
Melkamu kenito electronics in knitting
Melkamu kenito electronics in knittingMelkamu kenito electronics in knitting
Melkamu kenito electronics in knittingMelkamuKenito
 
End of Moore's Law?
End of Moore's Law? End of Moore's Law?
End of Moore's Law? Jeffrey Funk
 
MULTI-CORE PROCESSORS: CONCEPTS AND IMPLEMENTATIONS
MULTI-CORE PROCESSORS: CONCEPTS AND IMPLEMENTATIONSMULTI-CORE PROCESSORS: CONCEPTS AND IMPLEMENTATIONS
MULTI-CORE PROCESSORS: CONCEPTS AND IMPLEMENTATIONSAIRCC Publishing Corporation
 

Similar a Optical computing (20)

Full report on light peak technology
Full report on light peak technologyFull report on light peak technology
Full report on light peak technology
 
Seminar
SeminarSeminar
Seminar
 
Computer's clasification
Computer's clasificationComputer's clasification
Computer's clasification
 
Aqeel
AqeelAqeel
Aqeel
 
computer Development PPT created by Pooja Khamkar
computer Development PPT created by Pooja Khamkarcomputer Development PPT created by Pooja Khamkar
computer Development PPT created by Pooja Khamkar
 
CC LECTURE NOTES (1).pdf
CC LECTURE NOTES (1).pdfCC LECTURE NOTES (1).pdf
CC LECTURE NOTES (1).pdf
 
COA.pptx
COA.pptxCOA.pptx
COA.pptx
 
Introduction to parallel_computing
Introduction to parallel_computingIntroduction to parallel_computing
Introduction to parallel_computing
 
IS 139 Lecture 1 - 2015
IS 139 Lecture 1 - 2015IS 139 Lecture 1 - 2015
IS 139 Lecture 1 - 2015
 
Pipelining and ILP (Instruction Level Parallelism)
Pipelining and ILP (Instruction Level Parallelism) Pipelining and ILP (Instruction Level Parallelism)
Pipelining and ILP (Instruction Level Parallelism)
 
Michael Gschwind, Cell Broadband Engine: Exploiting multiple levels of parall...
Michael Gschwind, Cell Broadband Engine: Exploiting multiple levels of parall...Michael Gschwind, Cell Broadband Engine: Exploiting multiple levels of parall...
Michael Gschwind, Cell Broadband Engine: Exploiting multiple levels of parall...
 
Chapter 1.pptx
Chapter 1.pptxChapter 1.pptx
Chapter 1.pptx
 
Interconnect technology Final Revision
Interconnect technology Final RevisionInterconnect technology Final Revision
Interconnect technology Final Revision
 
Melkamu kenito electronics in knitting
Melkamu kenito electronics in knittingMelkamu kenito electronics in knitting
Melkamu kenito electronics in knitting
 
End of Moore's Law?
End of Moore's Law? End of Moore's Law?
End of Moore's Law?
 
Distributed Computing
Distributed ComputingDistributed Computing
Distributed Computing
 
Parallel computing persentation
Parallel computing persentationParallel computing persentation
Parallel computing persentation
 
Chapter_01.pptx
Chapter_01.pptxChapter_01.pptx
Chapter_01.pptx
 
Par com
Par comPar com
Par com
 
MULTI-CORE PROCESSORS: CONCEPTS AND IMPLEMENTATIONS
MULTI-CORE PROCESSORS: CONCEPTS AND IMPLEMENTATIONSMULTI-CORE PROCESSORS: CONCEPTS AND IMPLEMENTATIONS
MULTI-CORE PROCESSORS: CONCEPTS AND IMPLEMENTATIONS
 

Optical computing

  • 1. Optical computing Optical computing means performing computations, operations, storage and transmission of data using light. Instead of silicon chips optical computer uses organic polymers like phthalocyanine and polydiacetylene.Optical technology promises massive upgrades in the efficiency and speed of computers, as well as significant shrinkage in their size and cost. An optical desktop computer is capable of processing data up to 1,00,000 times faster than current models. Advantages of Optical Computers Many of the advantages stated below cannot be put into practice today but are theoretical advantages of the optical technology. Quite the contrary some of the advantages mentioned are not advantages nowadays but rather disadvantages due to the early stages of the development: For example a prototype of an optical computer is bigger than a conventional computer. However the theoretical advantages can be realized sooner or later. Optical computers will only be produced if they are cheaper or more powerful than conventional ones (or best: both of it at the same time). Besides hybrid solutions there will be a competition between the two different kinds of computing. To be able to measure optical computers and their future possibilities, conventional computers and their actual stage of development concerning each advantage and disadvantage are depicted. Advantages of optical computers compared to conventional computers: Higher performance The most significant advantage of optical computers is the potential of higher performance. A concrete benchmark test between optical and conventional computers is not possible yet. But the performance of an optical computer in an advanced stage should be several orders of magnitude higher. Conventional computer technology is based on electric current and electrons. While electric current is very fast, the average drift velocity of electrons is rather slow. In reality there is nothing faster than the speed of light. This makes light and photons to the perfect information carrier. Even long distances can be bridged within split seconds.
  • 2. A computer based completely on optical components is the optimum concerning speed: Full optical RAM saving photons directly if possible or otherwise indirectly, bus systems using concentrated light beams (lasers) to communicate, optical processors and holographic drives. All those components have the capability to be much faster than the existing electric ones. For example a light beam is able to transmit the whole encyclopedia Britannica within one second.[38] To lose no speed it is reasonable to use electric devices as rare as possible because every electric communication device is a potential bottleneck. At the beginning of the development a full optical computer will not be possible. Many devices will still work electrical. Even after a lot of development electronic parts most likely will still remain in optical computers, e.g. for instruction and controlling tasks. But using light as information carrier is only one opportunity to accelerate computers. Using a different architecture for optical computers could be a key factor to success. Conventional computers normally base on the Von Neumann architecture. That means data and instructions are kept in the same memory, are variable and are executed sequentially. One serious disadvantage of the Von Neumann architecture is the bottleneck between the CPU and the RAM, called the Von Neumann bottleneck. Since the CPU became the fastest device in modern computers it is forced to wait for data coming from the RAM. Bus systems and the RAM are not fast enough to provide data for the CPU without intermission. The effective processing speed is seriously limited. In optical computers light could be used in bus systems. Optical communication devices are much faster than electric ones. So the bottleneck could disappear. Another key to fasten up optical computers is to compute with higher parallelism. This implies higher performance and higher bandwidth which is depicted within the next break. Higher parallelism There are two options to achieve higher parallelism[39] in computers: One is to increase the amount of data which is sent through bus systems and computed in the CPU at each time. Modern conventional computers and operating systems base on 32 bit or 64 bit architecture. Often application software only supports 32 bit systems so the software does not take full advantage of 64 bit systems. Optical computers can be built with higher
  • 3. bandwidth. Within one data path several data sets can be transmitted parallel at the same time using different wavelengths or polarizations. But thereby attention to interference is necessary. Due to the coherent laser light destructive interference depending on the phase might occur. The higher parallelism and the superior velocity of light allow extreme processing speeds. But to take full advantage of this new architecture operating systems and application software have to be adjusted to it. Furthermore data paths are able to cross each other without interference.[40] This advantage of optical technologies can help to build architectures and layouts with superior parallelism. New layouts can be more three-dimensional and thus the space in a computer case can be used more intensively. With adequate miniaturization of optical components the size of computers could shrink. The following figure visualizes the possibilities of communication using light beams:
  • 4. Figure 15: Superposition, Crossed Data Paths and Three-Dimensionality (At a hybrid system using optoelectronic devices) Own illustration In conventional computers superposition and crossed data paths are not possible. Only one data set can be sent in one data path at each time and crossed data paths lead to loss of data. Besides that the layout of a conventional computer is rather based on two dimensions. Optical computers take advantage of all three dimensions. No wires are needed to transmit information, quite the contrary information can be sent wireless or even in vacuum.
  • 5. Another option to achieve higher parallelism in computers is to change the architecture concerning the temporal aspect. As already mentioned modern computers established on the Von Neumann architecture work sequential. To achieve more performance parallel computing is required. So more than one instruction stream or data stream can be executed simultaneously. According to Flynn’s taxonomy computers are divided into four groups concerning their architecture. Flynn differentiates between data streams and instruction streams. Single Instruction Multi Instruction Single Data SISD MISD Multiple Data SIMD MIMD Table 3: Flynn’s Taxonomy Own illustration based on STUCKE (1989), Pages 29-31 Personal Computers base on SISD. Vector processors using SIMD architecture are able to perform the same instruction on several data sets. Computers designed in MISD operate redundantly. Most of the fastest supercomputers of the world base on MIMD design. Thus several processors work independently and asynchronously. Concerning optical computers multiple data and instruction streams are useful, too. At least for mainframe computers an optical MIMD concept should be established. Less consumption Current computers consume a lot of energy. Modern CPUs often need over 80 watts in idle state, around 120 watts in normal use and up to 250 watts in performance mode.[41] For complex visualization a high performance graphics card is needed which needs up to 150 watts itself. The computer industry realized that issue several years ago. Today especially for notebooks energy-saving components are built. Calculation example: A computer runs every day around eight hours in normal use and one kilowatt hour costs 0,15 €. That means that the CPU consumes around 350 kilowatt hours a year. Consequently 53 € of energy costs are generated only by the CPU. All components of a
  • 6. loaded high performance PC (without monitor and periphery) consume within the same time even 1100 kilowatt hours (165 €). Another aspect is that a lot of the consumed energy is not used accurately but for idle power in the form of internal friction in the ICs. Still the amount of energy loss due to internal friction rises from CPU generation to generation. Due to this friction a lot of heat is released, which will be explained in the next break. Friction between the elementary particles does not occur when light is used as information carrier. Optical computers have the potential to be more power-saving than conventional ones. Less heat is released Less heat release using light? Sounds contradictory since light sources radiate heat. But in optical computers lasers are used as light sources. Those concentrated light beams only consist of a small spectrum of different wavelengths. Depending on the field of application lasers have different needs of energy and produce heat to a greater or lesser extent. Most modern CPUs are not able to work without a proper airing. The reason for that is the friction of the electrons in the integrated circuits. Those moving electrons hit each other within wires and ICs and thus heat is generated. Airing a processor (whether it is a CPU, GPU or the like) or the whole case needs energy, space and produces noise. Optical computers could be smaller because there is no need for a fan or free spaces for air circulation. Big computer centers with thousands of computers need powerful air conditioning systems. These aggregates are expensive and need a lot of energy too. Besides that the released heat increases the danger of fire at computer centers. Less noise Conventional computers often cause a lot of noise due to rotating fans and drives. High speed processors accelerated to their architectural limits need enormous active and passive cooling. In the past small fast spinning fans often were used which created plenty noise. Some years ago the computer industry realized the problem of the annoying noise and developed fans with bigger size. Thus the same volume flow was established with a lower rotary speed and the noise development was reduced. Actually optical computers could be almost noiseless since probably no fan will be needed. Light sources (e.g. lasers) can be cooled with passive coolers and heat pipes built out of aluminum or copper. Those passive coolers evacuate heat silent. Low noise is an interesting aspect for office and home users as well.
  • 7. More flexibility in layout Conventional PCs are built as a rectangular box (desktop) or as a laptop. One reason for that is the speed of electronic connections. It depends on the length of the cables and pipelines. On the motherboard the CPU, RAM and graphics card have to be close to each other to be able to move huge amounts of information. Longer distances imply the decrease of the practical transfer rate. Using optical components the distance of communication does not matter. Once the signal is in an optical fiber it does not matter whether the signal runs 1 meter or 1000 meters. Because of the low damping long-range communication is possible. Still the data rate is very high and there is no crosstalk. So the optical computer technology has the potential to change the shape and layout of computers fundamentally.[42] The components of one computer can be spread across a car, a building or even a city with almost no loss in performance. Consequently the server/client and the peer-to-peer architectures could be advanced. Many clients, terminals or even single components can be connected optically and consequently allow higher ranges. Less loss in communication Today communication often is realized with electric wires or wireless by radio frequency. The ranges of those communication ways are limited. Data sent through wires needs to be amplified several times to bridge longer distances. The communication with optical fibers is almost lossless due to the total internal reflection.[43] So amplifying the signal is not or only rarely needed. Furthermore a higher bandwidth is possible, optical communication is insensible to electromagnetic interfering fields and it is more tap-proof. For high performance communication (e.g. for backbones) already today fiber optics are used. Less wear Wear normally occurs at mechanically moving parts. In conventional computers those parts are especially fans, hard disk drives and conventional optical removable storages (CD, DVD, HD-DVD, Blu-ray disks). All those components have in common that they rotate or move very fast and thus friction is caused. Because of this friction the mechanical parts wear out and break. As already mentioned, this friction causes heat which is suboptimal, too.
  • 8. In optical computers fans possibly will not be needed any more. An optical processor does not heat up due to internal friction of the electrons like a conventional does. Additionally new technologies for mass storages can be established. Saving in the form of holograms or on molecular basis is possible. Those forms do not need fast spinning parts and do not wear out so heavily. Disadvantages of Optical Computers As already mentioned in Chapter 5.2 today’s optical computers are only prototypes and do not reflect the whole capabilities of optical computing. Consequently many of the disadvantages of optical computing are not because of the principle of the new technology itself but because of contemporary production technologies. The machines and processes that could be used to manufacture optical computers in mass production are still in its early stages. Disadvantages of optical computers compared to conventional computers: Optical components and their production is still expensive Conventional ICs are produced in high tech factories whose sole task is to manufacture ICs. Since the specialization and productivity of those factories is rather high the price of the product is low. Experts say the price of one transistor matches approximately the price of a printed character in a newspaper.[44] Earlier the price was a lot higher but due to new abilities in manufacturing and the new factories it decreased. Concerning optical components that are particularly manufactured for optical computers the development of its price could be similar to the development of the IC price. Today there are already optical components e.g. for communication and measurement. But they are not specially built to be placed in an optical computer. Furthermore there is no high tech and specialized factory to produce optical computers or especially designed components for it comparable to modern IC factories. Thus the price of an optical computer is rather high or respectively there is no complete system that can be bought yet.
  • 9. It is a kind of prototype and is not really manufactured but build by handwork of experts out of components produced for a different purpose. Using a standard architecture and standardized components could be the key to realize mass production of optical computers. This would decrease the production price and thus the price of an optical computer will drop as well. Unfortunately only a few companies have the financial power to develop computers with integrated optics first and then build a matching factory to produce them. IBM, Infinion respectively Siemens, Samsung, Intel and AMD can be mentioned at this point. During a scenario analysis in Chapter 6 possible scenarios concerning the price and performance of optical computers will be depicted more detailed. Optical components are not miniaturized enough yet Existing modern IC processors are built in a very high density called VLSI (very-large-scale-integration) or ULSI (ultra-large-scale-integration). There are several millions of transistors in an area of only approximately some square millimeters. The size of the transistors still decreases and some parts of the transistors already reach atomic dimensions. In contrast to that optical components can be build small and compact but not really miniaturized. The area of optics ranges from macrooptics over miniature optics to microoptics.[45] Unfortunately there are no microoptic integrated circuits specifically developed to assemble a CPU or a motherboard yet. A lot of development and new processes will be needed.[46] Problems of exact manufacturing Today conventional processors and computer components are manufactured with high precision and in huge charges using robust processes. Manufacturing problems can be eliminated by proper quality assurance. A change from the existing assembly method concerning the size of the chip structure to another size or using bigger wavers can trigger problems. For instance AMD recently had some problems changing their production from 90nm to 65nm structure.[47] Miniaturized optical components have to be built very exactly to work properly. This exactness often is not reached yet. Small deviations can cause
  • 10. massive problems diverting light beams. Exact production is expensive which leads to the next disadvantage. New expensive high-tech factories have to be built An IC factory costs several millions up to billions of dollars. Chip manufactures do not want to close such factories due to the high investments unless they have amortized. So the goal of the chip manufactures is to integrate optical components into the existing concepts and create hybrid devices. To allow integration optical devices have to be adapted and are not able to show their whole capabilities using an all optical architecture. So at first the integrated optical components will probably be produced in existing IC factories. Later own factories can be built to produce sole optical parts. Incompatibility Modern Personal Computers (PCs) are put together according to the Von Neumann architecture. Application software and especially operating systems are programmed to match this architecture. Because of these existing standards most of the application software can be used on any computer using a Microsoft Windows operating system. The portability is rather high. Optical computers may use a different architecture particularly regarding the parallelism of the system. Conventional existing computer programs working sequentially will have to run emulated in a kind of compatibility mode. So these programs cannot use the full calculating speed of optical computers or even worse some programs might not even work. Here the question arises whether the software drives the hardware or the hardware drives the software. Concerning optical computers the software will depend on the new architecture of the hardware. Software especially programmed for optical computers probably might work on conventional computers. The two different architectures existing next to each other with incompatible software is the worst scenario. Then software can only be used on one architecture so the portability and capability is quite low. 5.3 Comparison of the Advantages and Disadvantages
  • 11. In Chapters 5.1 and 5.2 advantages and disadvantages of optical computing are stated. Strong advantages like more performance while less consumption and more ergonomic (less noise, smaller and more flexible cases) stand opposite to disadvantages of more temporary nature (incompatibility and production problems). Consequently the advantages of optical computing overtop the disadvantages. Photons and light seem to be better than electrons and electric current to carry information. Especially in the long run the teething problems of the new technology will be coped and most likely the onward development will bring out high performance computers. The question how long this will take and the factors influencing it will be discussed within the next chapter.