Mr. Suyog M. Potdar
Under The Guidance Of
Prof. V. S. Gullhane
COMPUTER SCIENCE AND ENGINEERING
SIPNA SHIKSHAN PRASARAK MANDAL’S
College of Engineering & Technology, Amravati
Sant Gadge Baba Amravati University,
This is to certify that
Mr. Suyog M. Potdar
Of final year B.E.(Comp. Sci & Engg) has Successfully completed his
and submitted the seminar report in partial fulfillment of the Degree in
Bachelor of Engineering(Comp. Science & Engg) during academic year
(Dept. Comp Sci & Engg) (Dept. Comp Sci & Engg)
COMPUTER SCIENCE AND ENGINEERING
SIPNA SHIKSHAN PRASARAK MANDAL’S
College of Engineering & Technology, Amravati
Sant Gadge Baba Amravati University, Amravati
The making of the seminar needed co-operation and
guidance of a number of people. I therefore consider it my prime
duty to thank all those who had helped me through his venture.
It is my immense pleasure to express my
gratitude to Prof. V. S. Gulhane as my guide who provided me
constructive and positive feedback during the preparation of this
I express my sincere thank to the head of
department Prof. Dr. P. R. Deshmukh (Computer Science And
Engg) and all other staff members of Computer department for
their kind co- operation.
Last but not least,I am thankful to my friends and
library staff members whose encouragement and suggestion helped
me to complete my seminar.
I will be also thankful to my parent whose best
wishes are always with me.
Mr. Suyog M. Potdar
Final Year- Comp. Sci & Engg.
SIPNA’S College of Engg. & Tech.
3 Why... Adaptive Computing?
4 A Glance At Adaptive Computing
4.1 What is Adaptive Computing?
4.2 History of Adaptive Computing
4.3 Current Systems
4.4 Programming of reconfigurable computers
4.7 National Instruments
4.8 Research Projects
4.9 Comparison of systems
4.9.2 Rate of reconfiguration
4.9.3 Host coupling
4.9.5 Tool flow
5 Application with Example
Semantic-based Context-aware Dynamic Service Composition
an industrial Application Adaptive Computing
5.1 Dynamic Service Composition
Adaptive Computing is emerging as an important new organizational
structure for implementing computations. It combines the post-fabrication
programmability of processors with the spatial computational style most
commonly employed in hardware designs.
The result changes traditional “hardware” and “software” boundaries,
providing an opportunity for greater computational capacity and density
within a programmable media. Adaptive Computing must leverage
traditional CAD technology for building spatial designs. Beyond that,
however, reprogrammablility introduces new challenges and opportunities
for automation, including binding-time and specialization optimizations,
regularity extraction and exploitation, and temporal partitioning and
Due to its potential to greatly accelerate a wide variety of
applications, reconfigurable computing has become a subject of a great deal
of research. Its key feature is the ability to perform computations in
hardware to increase performance, while retaining much of the flexibility of
a software solution. In this introduction to reconfigurable computing, we
give an overview of the hardware architectures of reconfigurable computing
machines, and the software that targets these machines, such as compilation
tools. Finally, we consider the issues involved in run-time reconfigurable
systems, which re-use the configurable hardware during program execution.
Adaptive computing environments involve a variety of smart devices, which
tend to overcharge humans with complex or irrelevant interaction. Ambient
Intelligence pushes forward a vision where technology is integrated into
everyday objects with the intent of making users’ interaction with their
surrounding environment simpler and more intuitive. In this paper, we
expose how Ambient Intelligence can benefit from the coupling of a service-
oriented approach and multi-agent systems towards more appropriate
interactions with users. Our approach combines multi-agent techniques with
semantic web services to enable dynamic, context-aware service
composition, thus providing users with relevant high level services
depending on their current context and activity.
[ Ref 4,5]
Today many software programs, devices and resources are deployed as
distributed components. For example, various service providers (e.g.,
Google, Yahoo!) deploy their services (e.g. search, email and map services)
as distributed components through web service Technologies. Various
Audio and Visual devices also become accessible as distributed components
through networking protocols such as Bluetooth, Universal Plug and Play
UPnP Devices, Havi, etc. As a Network becomes more ubiquitous, the
number of distributed components available on the network also grows
In the computer and electronics world, we are used to two different
ways of performing computation: hardware and software. Computer
hardware, such as application-specific integrated circuits (ASICs), provides
highly optimized resources for quickly performing critical tasks, but it is
permanently configured to only one application via a multimillion-dollar
design and fabrication effort.
Computer software provides the flexibility to change applications and
perform a huge number of different tasks, but is orders of magnitude worse
than ASIC implementations in terms of performance, silicon area efficiency,
and power usage.
Field-programmable gate arrays (FPGAs) are truly revolutionary devices
that blend the benefits of both hardware and software. They implement
circuits just like hardware, providing huge power, area, and performance
benefits over software, yet can be reprogrammed cheaply and easily to
implement a wide range of tasks. Just like computer hardware, FPGAs
implement computations spatially, simultaneously computing millions of
operations in resources distributed across a silicon chip. Such systems can be
hundreds of times faster than microprocessor-based designs. However,
unlike in ASICs, these computations are programmed into the chip, not
permanently frozen by the manufacturing process. This means that an
FPGA-based system can be programmed and reprogrammed many times.
Sometimes reprogramming is merely a bug fix to correct faulty behavior, or
it is used to add a new feature. Other times, it may be carried out to
reconfigure a generic computation engine for a new task, or even to
reconfigure a device during operation to allow a single piece of silicon to
simultaneously do the work of numerous special-purpose chips.
However, merging the benefits of both hardware and software does come at
a price. FPGAs provide nearly all of the benefits of software flexibility and
development models, and nearly all of the benefits of hardware efficiency—
but not quite. Compared to a microprocessor, these devices are typically
several orders of magnitude faster and more power efficient, but creating
efficient programs for them is more complex. Typically, FPGAs are useful
only for operations that process large streams of data, such as signal
processing, networking, and the like.
Compared to ASICs, they may be 5 to 25 times worse in terms of area,
delay, and performance. However, while an ASIC design may take months
to years to develop and have a multimillion-dollar price tag, an FPGA
design might only take days to create and cost tens to hundreds of dollars.
For systems that do not require the absolute highest achievable performance
or power efficiency, an FPGA’s development simplicity and the ability to
easily fix bugs and upgrade functionality make them a compelling design
alternative. For many tasks, and particularly for beginning electronics
designers, FPGAs are the ideal choice.
[ Ref 2,5,8]
3.WHY.. ADAPTIVE COMPUTING?
There are two primary methods in traditional computing for the
execution of algorithms. The first is to use an Application Specific
Integrated Circuit, or ASIC, to perform the operations in hardware. Because
these ASICs are designed specifically to perform a given computation, they
are very fast and efficient when executing the exact computation for which
they were designed. However, after fabrication the circuit cannot be altered.
Microprocessors are a far more flexible solution. Processors execute a set of
instructions to perform a computation. By changing the software
instructions, the functionality of the system is altered without changing the
hardware. However, the downside of this flexibility is that the performance
suffers, and is far below that of an ASIC. The processor must read each
memory, determine its meaning, and only then execute it. This results in a
high execution overhead for each individual operation. Reconfigurable
computing is intended to fill the gap between hardware and software,
achieving potentially much higher performance than software, while
maintaining a higher level of flexibility than hardware.
It is well known that the computational power of general-purpose computers
is growing exponentially. Nevertheless, demand for computational power is
growing even faster. This deficit has driven research in new computer
architectures which might overcome some limitations of current
microprocessors. To date, most performance improvements have stemmed
from incremental (though by no means trivial) enhancements of the
theoretical von Neumann Architecture.
All of these designs, including the most recent superscalar
CPUs, still execute a sequenced stream of instructions taken from a fixed
instruction set. The instruction set is a list of all of the operations the
processor can perform, and it is fixed at the time of chip design. In contrast,
it is interesting to explore architectures that do not have this fixed instruction
set limitation — architectures that are reconfigurable or Adaptive.
This is the focus of a field known as Adaptive computing (AC).
Following are some more reasons to give stress on Adaptive Computing :-
_Need for Speed
Modern processors intended for the PC market
(e.g. Pentium 4, Core 2 Duo, I3, Athlon, etc.) are very powerful.
_What if we need more speed?
Supercomputers are very expensive.
Modern supercomputers are built from a large number of (fairly) ordinary
processors running in parallel.
_What if we also need low power?
Modern embedded CPUs (e.g. the ARM) are quite fast.
_ We have to design hardware
Implementing an algorithm as hardware is often the best way to achieve best
possible performance with the lowest possible transistor count and power
_Fortunately. . .
FPGA (Field Programmable Gate Array) technology makes custom
hardware far more accessible.
FPGAs are still 15 to 25 times slower than ASICs and use more power, so
aren’t a perfect solution –
conventional CPUs can still be faster for some things.
_Now the Solution is -
Adaptive Computing (Also known as Reconfigurable Computing)
[ Ref no. 1,2,3]
4. A GLANCE AT ADAPTIVE COMPUTING
4.1] What is Adaptive Computing ?
Adaptivecomputing, also known as Reconfigurable Computing, is a
computer architecture combining some of the flexibility of software with the
high performance of hardware by processing with very flexible high speed
computing fabrics like field-programmable gate arrays (FPGAs). The
principal difference when compared to using ordinary microprocessors is the
ability to make substantial changes to the datapath itself in addition to the
control flow. On the other hand, the main difference with custom hardware,
i.e. application-specific integrated circuits (ASICs) is the possibility to adapt
the hardware during runtime by "loading" a new circuit on the
It also refers to a logic chip that can change its physical circuitry on the fly.
Evolved from programmable architectures such as CPLD and FPGA,
adaptive computing is an order of magnitude faster in rate of reuse (ROR)
and can reconfigure itself in nanoseconds.
4.2] History of Adaptive Computing
The concept of reconfigurable computing has existed since the 1960s, when
Gerald Estrin's landmark paper proposed the concept of a computer made of
a standard processor and an array of "reconfigurable" hardware. The main
processor would control the behavior of the reconfigurable hardware. The
latter would then be tailored to perform a specific task, such as image
processing or pattern matching, as quickly as a dedicated piece of hardware.
Once the task was done, the hardware could be adjusted to do some other
task. This resulted in a hybrid computer structure combining the flexibility
of software with the speed of hardware; unfortunately this idea was far
ahead of its time in needed electronic technology. [ref 2]
In the 1980s and 1990s there was a renaissance in this area of research with
many proposed reconfigurable architectures developed in industry and
academia, such as: COPACOBANA, Matrix, Garp, Elixent, PACT XPP,
Silicon Hive, Montium, Pleiades, Morphosys, PiCoGA. Such designs were
feasible due to the constant progress of silicon technology that let complex
designs be implemented on one chip. The world's first commercial
reconfigurable computer, the Algotronix CHS2X4, was completed in 1991.
It was not a commercial success, but was promising enough that Xilinx (the
inventor of the Field-Programmable Gate Array, FPGA) bought the
technology and hired the Algotronix staff.
4.3] Current systems
The Adaptive computers can be categorized in two classes of architectures:
hybrid computer and fully FPGA based computers. Both architectures are
designed to transport the benefits of reconfigurable logic to large scale
computing. They can be used in traditional CPU cluster computers and
network infra structures.
The hybrid computer combine a single or a couple of reconfigurable logic
chip, FPGAs, with a standard microprocessor CPU by exchanging e.g. one
CPU of a multi CPU board with a FPGA, also known as hybrid-core
computing, or adding a PCI or PCI Express based FPGA expansion card to
the computer. Simplified, they are Von-Neumann based architectures with
an integrated FPGA accelerator. This architectural compromise results in an
reduced scalability of hybrid computers and raises their power consumption.
They have the same bottlenecks as all von Neumann based architectures
have nevertheless it enables users to get an acceleration of their algorithm
without losing their standard CPU based environment.
A relatively new class are the fully FPGA based computers. This class
usually contains no CPUs or uses the CPUs only as interface to the network
environment. Their benefit is to transport the energy efficiency and
scalability of FPGAs fully without compromise to their users. Depending on
the Architecture of the interconnection between the FPGAs these machines
are fully scalable even across single machine borders. Their bus system and
overall architecture eliminate the bottlenecks of the von Neumann
Examples for hybrid computers
Known examples for hybrid computers are the XD1. A machine designed by
OctigaBay. Cray supercomputer company (not affiliated with SRC
Computers) acquired OctigaBay and its reconfigurable computing platform,
which Cray marketed as the XD1 until recently. SGI sells the RASC
platform with their Altix series of supercomputers. SRC Computers, Inc.
has developed a family of reconfigurable computers based on their
IMPLICIT+EXPLICIT architecture and MAP processor. A current example
of hybrid-core computers is Convey Computer Corporation's HC-1, which
has both an Intel x86 processor and a Xilinx FPGA coprocessor. Another
example is CompactRIO from National Instruments.
Examples for fully FPGA based computers
There are a several academic projects who tried to create fully FPGA based
computers for several markets. One is the COPACOBANA the Cost
Optimized Codebreaker and Analyzer. A spin-off company SciEngines of
the COPACOBANA-Project of the universities of Bochum and Kiel in
Germany currently sells the fourth generation of fully FPGA based
computers, COPACOBANA RIVYERA (Reconfigure Versatally your raw
4.4] Programming of reconfigurable computers
The FPGA reconfiguration can be accomplished either via the traditional
hardware description languages (HDLs), which can be generated directly or
by using electronic design automation ("EDA") or electronic system level
("ESL") tools, employing high level languages like the graphical tool
Starbridge Viva or C-based languages like for example Impulse C from
Impulse Accelerated Technologies, SystemC, LISA, Handel-C from
Celoxica, DIME-C from Nallatech, C-to-Verilog.com or Mitrion-C from
Mitrionics, or graphical programming languages like LabVIEW.
Mitrionics has developed a SDK that enables software written using a single
assignment language to be compiled and executed on FPGA-based
computers. The Mitrion-C software language and Mitrion processor enable
software developers to write and execute applications on FPGA-based
computers in the same manner as with other computing technologies, such
as graphical processing units (“GPUs”), cell-based processors, parallel
processing units (“PPUs”), multi-core CPUs, and traditional single-core
SRC has developed a "Carte" compiler that takes an existing high-level
languages like C or Fortran, and with a few modifications, compiles them
for execution on both the FPGA and microprocessor. According to SRC
literature, "...application algorithms are written in a high-
level language such as C or Fortran. Carte extracts the maximum parallelism
from the code and generates pipelined hardware logic that is instantiated in
the MAP. It also generates all the required interface code to manage the
movement of data to and from the MAP and to coordinate the
microprocessor with the logic running in the MAP." (note that SRC also
allows a traditional HDL flow to be used). The SRC systems communicate
via the SNAP memory interface, and/or the (optional) Hi-Bar switch.
4.7] National Instruments
National Instruments have developed a hybrid embedded computing system
called CompactRIO. CompactRIO systems consist of reconfigurable chassis
housing the user-programmable FPGA, hot swappable I/O modules, real-
time controller for deterministic communication and processing, and
graphical LabVIEW software for rapid RT and FPGA programming.
4.8] Research projects
The research community is also acting on the subject with projects like
MORPHEUS in Europe which implements on a single 100 mm² 90 nm chip
an ARM9 processor, an eFPGA from Abound Logic (formerly M2000), a
DREAM PiCoGA and a PACT XPP matrix.
Abound Logic contributes to the MORPHEUS project with an embedded
FPGA, and uses the same architecture to make its very large standard
The University of Florida also has its own version of a reconfigurable
computer called Novo-G that is currently the world's fastest and most
4.9] Comparison of systems
As an emerging field, classifications of reconfigurable architectures are still
being developed and refined as new architectures are developed; no unifying
taxonomy has been suggested to date. However, several recurring
parameters can be used to classify these systems.
The granularity of the reconfigurable logic is defined as the size of the
smallest functional unit (configurable logic block, CLB) that is addressed by
the mapping tools. High granularity, which can also be known as fine-
grained, often implies a greater flexibility when implementing algorithms
into the hardware. However, there is a penalty associated with this in terms
of increased power, area and delay due to greater quantity of routing
required per computation. Fine-grained architectures work at the bit-level
manipulation level; whilst coarse grained processing elements
(reconfigurable datapath unit, rDPU) are better optimised for standard data
path applications. One of the drawbacks of coarse grained architectures are
that they tend to lose some of their utilisation and performance if they need
to perform smaller computations than their granularity provides, for example
for a one bit add on a four bit wide functional unit would waste three bits.
This problem can be solved by having a coarse grain array (reconfigurable
datapath array, rDPA) and a FPGA on the same chip.
Coarse-grained architectures (rDPA) are intended for the implementation for
algorithms needing word-width data paths (rDPU). As their functional
blocks are optimized for large computations and typically comprise word
wide arithmetic logic units (ALU), they will perform these computations
more quickly and with more power efficiency than a set of interconnected
smaller functional units; this is due to the connecting wires being shorter,
resulting in less wire capacitance and hence faster and lower power designs.
A potential undesirable consequence of having larger computational blocks
is that when the size of operands may not match the algorithm an inefficient
utilisation of resources can result. Often the type of applications to be run
are known in advance allowing the logic, memory and routing resources to
be tailored (for instance, see KressArray Xplorer) to enhance the
performance of the device whilst still providing a certain level of flexibility
for future adaptation. Examples of this are domain specific arrays aimed at
gaining better performance in terms of power, area, throughput than their
more generic finer grained FPGA cousins by reducing their flexibility.
4.9.2] Rate of reconfiguration
Configuration of these reconfigurable systems can happen at deployment
time, between execution phases or during execution. In a typical
reconfigurable system, a bit stream is used to program the device at
deployment time. Fine grained systems by their own nature require greater
configuration time than more coarse-grained architectures due to more
elements needing to be addressed and programmed. Therefore more coarse-
grained architectures gain from potential lower energy requirements, as less
information is transferred and utilised. Intuitively, the slower the rate of
reconfiguration the smaller the energy consumption as the associated energy
cost of reconfiguration are amortised over a longer period of time. Partial re-
configuration aims to allow part of the device to be reprogrammed while
another part is still performing active computation. Partial re-configuration
allows smaller reconfigurable bit streams thus not wasting energy on
transmitting redundant information in the bit stream. Compression of the bit
stream is possible but careful analysis is to be carried out to ensure that the
energy saved by using smaller bit streams is not outweighed by the
computation needed to decompress the data.
4.9.3] Host coupling
Often the reconfigurable array is used as a processing accelerator attached to
a host processor. The level of coupling determines the type of data transfers,
latency, power, throughput and overheads involved when utilising the
reconfigurable logic. Some of the most intuitive designs use a peripheral bus
to provide a coprocessor like arrangement for the reconfigurable array.
However, there have also been implementations where the reconfigurable
fabric is much closer to the processor, some are even implemented into the
data path, utilising the processor registers. The job of the host processor is to
perform the control functions, configure the logic, schedule data and to
provide external interfacing.
The flexibility in reconfigurable devices mainly comes from their routing
interconnect. One style of interconnect made popular by FPGAs vendors,
Xilinx and Altera are the island style layout, where blocks are arranged in an
array with vertical and horizontal routing. A layout with inadequate routing
may suffer from poor flexibility and resource utilisation, therefore providing
limited performance. If too much interconnect is provided this requires more
transistors than necessary and thus more silicon area, longer wires and more
4.9.5] Tool flow
Generally, tools for configurable computing systems can be split up in two
parts, CAD tools for reconfigurable array and compilation tools for CPU.
The front-end compiler is an integrated tool, and will generate a structural
hardware representation that is input of hardware design flow. Hardware
design flow for reconfigurable architecture can be classified by the approach
adopted by three main stages of design process: technology mapping,
placement algorithm and routing algorithm. The software frameworks differ
in the level of the programming language.
Some types of reconfigurable computers are microcoded processors where
the microcode is stored in RAM or EEPROM, and changeable on reboot or
on the fly. This could be done with the AMD 2900 series bit slice processors
(on reboot) and later with FPGAs (on the fly).
Some[which?] dataflow processors are implemented using reconfigurable
A new method of application development for reconfigurable computing is
being developed by MNB Technologies, Inc, under contract to the United
States Air Force Office of Scientific Research (AFOSR). This approach uses
a national repository of generic algorithms, similar to the BLAS and
LAPACK libraries found at netlib.org. In addition to the repository, the
project is developing a tightly integrated suite of expert system based tools
that largely eliminate the need for an application developer to have any in-
depth knowledge of the underlying hardware or how to use the specialized
Verilog and VHDL hardware description languages. The results of this
research will be available without charge to individuals and organizations
based in the United States.
To compare the effect of various ways to implement an algorithm on the
runtime and energy used, some tools allow compiling the same piece of C
code for a fixed CPU, a soft processor, or compiling directly to FPGA
[Ref no. 9, 10]
4.10] Applications of Adaptive Computing
Some of the Applications are :-
1] Isolation Points.
2] On the use of memory and resources in Minority Games.
3] Semantic Based Context Aware Dynamic Service Composition.
4] Exploiting User Location for Load Balancing WLAN's and improving
5] Self Adaptive Software: Landscape and Research Challenges.[Ref no. 8]
5] Semantic-based Context-aware Dynamic
Service Composition- an Industria Application
Complex services may be dynamically composed through combining
distributed components on demand (i.e., when requested by a user) in order
to provide new services without preinstallation. Several systems have been
proposed to dynamically compose services. However, they require users to
request services in a manner that is not intuitive to the users. In order to
allow a user to request a service in an intuitive form (e.g., using a natural
language), this paper proposes a semantics-based service composition
architecture. The proposed architecture obtains the semantics of the service
requested in an intuitive form, and dynamically composes the requested
service based on the semantics of the service. To compose a service based
on its semantics, the proposed architecture supports semantic representation
of components [through a component model named Component Service
Model with Semantics (CoSMoS)], discovers components required to
compose a service [through a middleware named Component Runtime
Environment (CoRE)], and composes the requested service based on its
semantics and the semantics of the discovered components [through a
service composition mechanism named Semantic Graph-Based Service
This paper presents the design, implementation and empirical evaluation of
the proposed architecture.
5.1] Dynamic Service Composition
The recent development of distributed component technologies such as
CORBA andWeb Service made it possible to componentize various software
programs, devices, and resources and to distribute them over a network. In
such an environment where a large number of various components are
distributed, it is possible to dynamically compose a service on demand,
i.e., composing a service upon receiving a request from a user, through
discovering, combining and executing necessary components. Composing
services on demand (i.e., dynamic service composition) has various
advantages. For instance, by dynamically composing services on demand,
services do not need to be configured or deployed in advance. In addition, by
composing services based on requests from users, it is possible to customize
the services to individual user profiles.
To illustrate the advantage of dynamic service composition, consider the
following example scenario. Suppose Tomwants to take his family to a new
restaurant he recently found on the web, so he wants to print out a map
showing a direction to the restaurant from his house. Assume that: 1) the
restaurant’sWeb server stores the restaurant’s information such as its address
in a structured document (e.g., in XML); 2) Tom’s PC stores his personal
information such as his home address in a database; 3) there is a Web
Service which, given two addresses, generates a map showing a direction
from one address to the other; and 4) Tom has a printer connected to his
home network. In order to print out the map showing the direction from his
home to the restaurant with the technology currently available, Tom has to
manually perform the following steps: a) discover the restaurant’s homepage
and the Web Service that generates a map; b) obtain the addresses of the
restaurant and his home; c) invoke theWeb Service using the addresses
obtained; and d) print out the map generated by theWeb Service. However,
if Tomis not an experienced PC user, it may be difficult for him to perform
these steps. For instance, he may not know how to input the restaurant’s
address to theWeb Service that generates a map. Dynamic service
composition, on the other hand, will automatically compose the direction
printing service, upon Tom’s request, by discovering the four necessary
components, identifying the steps a)–d) and executing them on Tom’s
behalf. Since the direction printing service is composed on demand, it is not
required to be configured or deployed in advance. Also, if Tom’s daughter,
Alice, requests for a direction to the restaurant, and if she carries a PDA with
her, the service may be customized for her such that it shows the map on her
PDA’s display instead of printing it out.
[Ref no. 3,1,8]
Reconfigurable computing is becoming an important part of research in
computer architectures and software systems. By placing the
computationally intense portions of an application onto the reconfigurable
hardware, the overall application can be greatly accelerated. This is because
computing combines the benefits of both software and ASIC
implementations. Like software, the mapped circuit is flexible, and can be
changed over the lifetime of the system or even the execution time of an
application. Similar to an ASIC, reconfigurable systems provide a method to
map circuits into hardware, achieving far greater performance than software
as a result of bypassing the fetch-decode-execute cycle of traditional
microprocessors, and parallel execution of multiple operations.
Reconfigurable hardware systems come in many forms, from a configurable
functional unit integrated directly into a CPU, to a reconfigurable co
processor coupled with a host microprocessor, to a multi-FPGA stand-alone
unit. The level of coupling, granularity of computation structures, and form
of routing resources are all key points in the design of reconfigurable
systems. Compilation tools for reconfigurable systems range from simple
tools that aid in the manual design and placement of circuits, to fully
automatic design suites that use program code written in a high-level
language to generate circuits and the controlling software. The variety of
tools available allows designers to choose between manual and automatic
circuit creation for any or all of the design steps. Although automatic tools
greatly simplify the design process, manual creation is still important for
Finally, run-time reconfiguration provides a method to
accelerate a greater portion of a given application by allowing the
configuration of the hardware to change over time. Apart from the benefits
of added capacity through the use of virtual hardware, run-time
reconfiguration also allows for circuits to be optimized based on run-time
conditions. In this manner, the performance of a reconfigurable system can
approach or even surpass that of an ASIC.
Reconfigurable computing systems have shown the ability to greatly a
ccelerate program execution, providing a high-performance alternative to
software-only implementations. However, no one hardware design has
emerged as the clear pinnacle of reconfigurable design. Although general-
purpose FPGA structures have standardized into LUT-based architectures,
groups designing hardware for reconfigurable computing are currently also
exploring the use of heterogeneous structures and word-width computational
elements. Those designing compiler systems face the task of improving
automatic design tools to the point where they may achieve mappings
comparable to manual design for even high-performance applications.
Within both of these research categories lies the additional topic of run-time
reconfiguration. While some work has been done in this field as well,
research must continue in order to be able to perform faster and more
efficient reconfiguration. Further study into each of these topics is necessary
in order to harness the full potential of reconfigurable computing.
[ Ref no. 4,6,7]
1] D. Mennie and B. Pagurek, “An architecture to support dynamic
of service components,” in Proc. 5th Int. Workshop Component-
Oriented Program., Sophia Antipolis, France, 2000
2.] Estrin, G. 2002. Reconfigurable computer origins: the UCLA fixed-plus-
variable (F+V) structure computer. IEEE Ann. Hist. Comput. 24, 4 (Oct.
2002), 3–9. DOI=http://dx.doi.org/10.1109/MAHC.2002.1114865
3] F. Casati, S. Ilnicki, L.-J. Jin, V. Krishnamoorthy, and M.-C. Shan,
“Adaptive and dynamic service composition in eFlow,” in Proc. Int.
Conf Advanced Inf. Syst. Eng., Stockholm, Sweden, 2000.
4] K. Compton, S. Hauck, “Configurable Computing: A Survey of Systems
Northwestern University, Dept. of ECE Technical Report, 1999.
5] M. Minami, H. Morikawa, and T. Aoyama, “The design and evaluation
of an interface-based naming system for supporting service synthesis
in ubiquitous computing environment,” Trans. Inst. Electron., Inf.
Commun. Eng., vol. J86-B, no. 5, pp. 777–789, May 2003.
6] S. Hauck, “The Roles of FPGAs in Reprogrammable Systems”
Proceedings of the IEEE,
Vol. 86, No. 4, pp. 615-638, April 1998.
7] W. H. Mangione-Smith, B. Hutchings, D. Andrews, A. DeHon
8] ACM Transactions on Autonomous and Adaptive Systems Vol 4 No. 2 ,