Results 1 -
6 of
6
ACCURATE MODELS FOR ESTIMATING AREA AND POWER OF FPGA IMPLEMENTATIONS †
"... This paper presents accurate area and power estimation models for implementations using FPGAs from the Xilinx Virtex-2Pro family. These models are designed to facilitate efficient design space exploration in an automated algorithm-architecture codesign framework. Detailed models for accurately estim ..."
Abstract
-
Cited by 8 (3 self)
- Add to MetaCart
(Show Context)
This paper presents accurate area and power estimation models for implementations using FPGAs from the Xilinx Virtex-2Pro family. These models are designed to facilitate efficient design space exploration in an automated algorithm-architecture codesign framework. Detailed models for accurately estimating the number of slices, block RAMs and 18x18-bit multipliers for fixed point and floating-point IP cores have been developed. These models are also utilized to develop accurate power models that consider the effect of logic power, signal power, clock power and I/O power. In all cases, the model coefficients have been derived by using curve fitting or regression analysis. The modeling error for the IP cores is very small (average 0.95%). The error for fairly large examples such as floating point implementation of 8-point FFTs is also quite small; it is 1.87 % for estimation of number of slices and 3.48 % for estimation of power consumption.
An Automated Framework for Accelerating Numerical Algorithms on Reconfigurable Platforms Using Algorithmic/Architectural Optimization,” in
- IEEE Transactions on Computers
, 2009
"... Abstract—This paper describes TANOR, an automated framework for designing hardware accelerators for numerical computation on reconfigurable platforms. Applications utilizing numerical algorithms on large-size data sets require high-throughput computation platforms. The focus is on N-body interaction ..."
Abstract
-
Cited by 4 (1 self)
- Add to MetaCart
(Show Context)
Abstract—This paper describes TANOR, an automated framework for designing hardware accelerators for numerical computation on reconfigurable platforms. Applications utilizing numerical algorithms on large-size data sets require high-throughput computation platforms. The focus is on N-body interaction problems which have a wide range of applications spanning from astrophysics to molecular dynamics. The TANOR design flow starts with a MATLAB description of a particular interaction function, its parameters, and certain architectural constraints specified through a graphical user interface. Subsequently, TANOR automatically generates a configuration bitstream for a target FPGA along with associated drivers and control software necessary to direct the application from a host PC. Architectural exploration is facilitated through support for fully custom fixed-point and floating point representations in addition to standard number representations such as single precision floating point. Moreover, TANOR enables joint exploration of algorithmic and architectural variations in realizing efficient hardware accelerators. TANOR’s capabilities have been demonstrated for three different N-body interaction applications: the calculation of gravitational potential in astrophysics, the diffusion or convolution with Gaussian kernel common in image processing applications, and the force calculation with vector-valued kernel function in molecular dynamics simulation. Experimental results show that TANOR-generated hardware accelerators achieve lower resource utilization without compromising numerical accuracy, in comparison to other existing custom accelerators. Index Terms—Algorithms implemented in hardware, Reconfigurable hardware, Signal processing systems, Numerical algorithms. F 1
J Sign Process Syst DOI 10.1007/s11265-008-0337-9 Accelerating Machine-Learning Algorithms on FPGAs using Pattern-Based Decomposition
, 2008
"... Abstract Machine-learning algorithms are employed in a wide variety of applications to extract useful information from data sets, and many are known to suffer from superlinear increases in computational time with increasing data size and number of signals being processed (data dimension). Certain pr ..."
Abstract
- Add to MetaCart
Abstract Machine-learning algorithms are employed in a wide variety of applications to extract useful information from data sets, and many are known to suffer from superlinear increases in computational time with increasing data size and number of signals being processed (data dimension). Certain principal machine-learning algorithms are commonly found embedded in larger detection, estimation, or classification operations. Three such principal algorithms are the Parzen window-based, non-parametric estimation of Probability Density Functions (PDFs), K-means clustering and correlation. Because they form an integral part of numerous machine-learning applications, fast and efficient execution of these algorithms is extremely desirable. FPGA-based reconfigurable computing (RC) has been successfully used to accelerate computationally intensive problems in a wide variety of scientific domains to achieve speedup over traditional software implementations. However, this potential benefit is quite often not fully realized because creating efficient FPGA designs is generally carried out in a laborious, case-specific manner requiring a great amount of
Design and Analysis of a Multi-core PDF Estimation Algorithm on FPGAs
"... Abstract — FPGAs (Field Programmable Gate Arrays) and the field of “High Performance Computing ” have been applied to computationally intensive problems in various domains mainly addressing speedup issues. However, there is still a significant need of in-depth research and proof of success with real ..."
Abstract
- Add to MetaCart
(Show Context)
Abstract — FPGAs (Field Programmable Gate Arrays) and the field of “High Performance Computing ” have been applied to computationally intensive problems in various domains mainly addressing speedup issues. However, there is still a significant need of in-depth research and proof of success with real applications for proposing them as solutions for a more general class of problems. Along this line, this work involves the design, development and analysis of a multi-core Probability Density Function (PDF) estimation algorithm using Gaussian kernels on FPGAs. Speedup and performance metrics were obtained and discussions addressing scalability issues suggest that FPGAs are a good choice for a wide class of applications that are primarily based on density function estimations.
ACCELERATION OF MOLECULAR DYNAMICS SIMULATION FOR TERSOFF2 POTENTIAL THROUGH RECONFIGURABLE HARDWARE
, 2012
"... Approval of the thesis: ..."
(Show Context)
signal and image processing, embedded systems, architecture-algorithm codesign, optimization and automation, research translation
, 2013
"... The public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comm ..."
Abstract
- Add to MetaCart
The public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggesstions for reducing this burden, to Washington