Beruflich Dokumente
Kultur Dokumente
the recurring silicon area penalty implied is reduced Our vision is illustrated in Figure 1. The Silicon
(in both cases, unlike FPGAs). Hive reconfigurable cores replace both the ASICs
and the DSP accelerators.
Logistics and inventory Since all accelerators are now reconfigurable,
The costs of managing a large variety of slightly their resources can be time-multiplexed. This allows
different products can translate into expensive for a more optimal utilisation of the available
inefficiencies. That can be solved when product hardware, reducing redundancies. Depending on the
variations are implemented in software. case, it can lead to a smaller silicon foot-print than
can be achieved with ASICs.
Deploying reconfigurable accelerators
We enable the replacement of all ASIC accelerators Our reconfigurable architectures
in SoCs with reconfigurable cores, rendering the The basic component of the architecture of Silicon
SoC fully programmable after fabrication. This way, Hive’s accelerators is the Processing and Storage
flexibility is maintained throughout the product life- Element (PSE). See Figure 2. A PSE is a VLIW-like
cycle. At the same time, our cores will preserve a data-path consisting of several interconnect
computational efficiency (MOPS/W) and a silicon networks (IN), one or more operation-issue slots (IS)
foot-print comparable to that achievable with ASIC with associated function units (FU), distributed
cores. register files (RF) and, optionally, local memory
storage (MEM). PSEs are designed according to a
template that ensures all PSEs are easy and clean
data-paths for a compiler to handle, guaranteeing
high-level of programmability by construction.
A matrix of one or more PSEs, together with a
VLIW-like controller (CTRL) and configuration
memory (CONFIG. MEM), make up a cell. A cell is a
fully-operational processor capable of computing
complete algorithms. PSEs within a cell can
communicate with each other via data
communication lines (CL). Typically, one application
function at a time is mapped onto the matrix of
PSEs. The more PSEs are present, the more the
function can be mapped in space, in a data-flow
manner.
An array of one or more cells, connected together across the streaming array, occupying different sub-
via a data-driven communication mechanism, forms sets of cells, so to implement complete sub-systems.
a streaming array. The communication across cells
takes place through blocking FIFOs accessed from
load/store (LD/ST) units within the cells. Multiple
functions can be concurrently mapped onto the
streaming array, each one occupying a non-
overlapping sub-set of the cells.
Many different trade-offs between the number of
cells, and the number of PSEs per cell, can be
made. In general, the closer the data-rate of a given
function is to the clock rate of the processor, the
more cells should be used. This is because the cycle
budget for what a single cell can do is, in this case,
very small, so the computation must be spread
across potentially smaller cells. For applications with
data-rates far lower than the processor’s clock rate,
the cycle budget is higher, and a small number of
cells with a relatively high number of PSEs may lead Figure 3. Diagram of Silicon Hive’s block accelerators.
to more optimal exploitation of instruction-level
parallelism, higher modularisation, and ease of
programming.
Silicon Hive’s internal design methodology and
tools allow for the very quick design of different PSE
architectures, cells with different numbers of PSEs
interconnected in different ways, and streaming
arrays with different numbers of cells interconnected
in a variety of patterns. This way, Silicon Hive can
timely create a virtually unlimited variety of
application-domain-specific cores, as market
evolution dictates. Each core is reconfigurable after
fabrication, and flexible enough to tackle all
application needs within its target domain.
overheads. When a direct interconnect line is not require the least amount of attention from chip
available between two resources, Silicon Hive’s designers during SoC integration and validation.
spatial compiler spatially schedules the
communication with a minimal number of local hops. Cost-effective, customized solutions
Another distinguishing characteristic of Silicon Hive’s
Silicon efficiency solutions is their domain-specific nature. Many
How can Silicon Hive cores achieve a silicon reconfigurable or reprogrammable processors in the
footprint smaller than the equivalent ASIC solution? market today are catch-all solutions, made for
general-purpose use. As a consequence, the
Higher arithmetic density architectures carry considerable overhead in cost
Unlike traditional processors and DSPs, Silicon Hive and efficiency for any specific application domain.
cores have no hardware dedicated to pipeline Because of its internal design methodology, which
management and operand forwarding logic. Instead, allows for the generation of new cores and compilers
these tasks are statically performed by the spatial within days, Silicon Hive can tune its processors to
compiler. In addition, because all data-path the needs of particular application domains and/or
resources in a Silicon Hive core are distributed, the customers. This enhances the ease and speed with
complexity of the operand networks is much which a customer can integrate and use our solution,
reduced. This way, most silicon in a Silicon Hive and dramatically reduces the silicon overheads
core is dedicated to actually crunching data. This implied. Making cost-effective solutions for very cost-
substantially increases their silicon efficiency when sensitive consumer markets is an inherited part of
compared to traditional processors or DSPs. our culture and history within Philips.
Conclusions
A key enabler to truly programmable and cost-
effective SoCs are reconfigurable accelerators that
can replace currently-used ASIC co-processors with
a comparable computational efficiency (MOPS/W)
and lower silicon overhead. Silicon Hive currently
provides two types of reconfigurable accelerators
that address these market needs. The key
differentiators of our approach are: the robustness
and reliability of our technology, our ability to
customise our solutions to specific customer needs,
the high-level programmability of our cores (ANSI
C), and the fact that they can operate either as
standard DSPs or coarse-grained FPGAs.