VLSI Circuits, Systems, and Architectures for Advanced Image and Video Compression Standards
View this Special IssueResearch Article  Open Access
Christos Ttofis, Theocharis Theocharides, "Hardware Design Considerations for EdgeAccelerated Stereo Correspondence Algorithms", VLSI Design, vol. 2012, Article ID 602737, 17 pages, 2012. https://doi.org/10.1155/2012/602737
Hardware Design Considerations for EdgeAccelerated Stereo Correspondence Algorithms
Abstract
Stereo correspondence is a popular algorithm for the extraction of depth information from a pair of rectified 2D images. Hence, it has been used in many computer vision applications that require knowledge about depth. However, stereo correspondence is a computationally intensive algorithm and requires highend hardware resources in order to achieve realtime processing speed in embedded computer vision systems. This paper presents an overview of the use of edge information as a means to accelerate hardware implementations of stereo correspondence algorithms. The presented approach restricts the stereo correspondence algorithm only to the edges of the input images rather than to all image points, thus resulting in a considerable reduction of the search space. The paper highlights the benefits of the edgedirected approach by applying it to two stereo correspondence algorithms: an SADbased fixedsupport algorithm and a more complex adaptive support weight algorithm. Furthermore, we present design considerations about the implementation of these algorithms on reconfigurable hardware and also discuss issues related to the memory structures needed, the amount of parallelism that can be exploited, the organization of the processing blocks, and so forth. The two architectures (fixedsupport based versus adaptivesupport weight based) are compared in terms of processing speed, disparity map accuracy, and hardware overheads, when both are implemented on a Virtex5 FPGA platform.
1. Introduction
Depth extraction from stereoscopic images is a vital step in several emerging embedded applications, such as robot navigation, obstacle detection for autonomous vehicles, and space and avionics [1]. Depth information is extracted by solving the challenging problem of stereo correspondence, which aims at finding corresponding points (or conjugate pairs) between the input stereo images (usually referred to as reference image and target image ). Given that the input images are rectified, the correspondence of a pixel at coordinate can only be found at the same vertical coordinate y, and within a maximum horizontal bound, called disparity range D (–) [2]. The disparity is computed as the absolute difference between the coordinates of the corresponding pixels in and . The disparities of all corresponding pixels form a disparity map, which once estimated, can be used to extract the depth of the scene by using triangulation [1].
In general, stereo correspondence algorithms mostly follow four steps: matching cost computation, cost (support) aggregation, (3) disparity computation/optimization, and (4) disparity map refinement [2]. Moreover, they are classified into two broad categories: global and local [2]. Global algorithms can produce very accurate results but are slower and computationally more demanding compared to local algorithms, due to their iterative nature and high memory needs [3]. On the other hand, local algorithms are faster and less computationally expensive, hence suitable for the majority of embedded stereo vision applications.
Local algorithms determine the disparity of pixel in by finding the pixel of that yields the lowest score of a similarity measure (see [3] for a review) computed on a support (correlation) window centered on and on each of the – candidates defined by the disparity range. Early local algorithms have followed a simple approach that uses a fixed (typically square) support window during the cost aggregation step. This approach, however, is prone to errors as it blindly aggregates pixels belonging to different disparities, resulting in incorrect disparity estimation at depth discontinuities and regions with low texture [3]. Recent local algorithms improve this basic approach by using multiplewindow and adaptivesupport weight (ADSW) methods [4]. The latter methods represent stateoftheart in local stereo correspondence algorithms, as they can generate disparity maps that approach the accuracy of global algorithms [3]. These methods operate by assigning different weights to the pixels in a support window based on the spatial and color distances to the center pixel [4], or on information extracted from image segmentation [5, 6]. In this way, they aggregate only those neighboring pixels that are at the same disparity. Despite their improvements in accuracy, ADSW algorithms are generally slower than other local algorithms. A survey for different approaches can be found in [2, 3].
The realtime and lowpower constraints of most embedded stereo vision applications complicate the realization of stereo correspondence algorithms. Software implementations usually struggle to meet these constraints, and as a result, some form of hardware acceleration or even a complete custom hardware implementation is preferred [3]. Hardware acceleration of stereo correspondence algorithms has been done extensively using Digital Signal Processors (DSPs) and Graphics Processing Units (GPUs) [3]. These systems, however, involve architectures that are generally not suitable for embedded applications. DSPs do not provide enough computational power to achieve parallel processing, while GPUs consume excessive power. The strong embedded requirements of such applications imply the use of dedicated hardware architectures such as ASICs and FPGAs, which can provide the necessary computational power and the energy efficiency.
The majority of dedicated hardware architectures implement fixedsupport algorithms, providing adequate frame rates. However, numerous embedded vision applications require not only realtime processing speed, but also reliable depth computation. As such, there have been a few attempts recently to implement more complex algorithms in hardware such as ADSW algorithms, but these algorithms are computationally more expensive and thus the frame rate suffers, especially when high resolution images are used. It is imperative that the key to providing a successful realization of an embedded stereo vision system is the careful design of the core architecture so as to provide a good tradeoff between the processing speed and the quality of the resulting disparity map.
This paper presents an approach to the acceleration of hardware implementations of stereo correspondence algorithms for embedded stereo vision systems, by using edge information. The presented approach lies on the extraction of feature points (edges) that are used to restrict the stereo correspondence algorithm only to specific features of the input images (rather than to all image points), thus resulting in a considerable reduction of the search space. The paper presents and analyzes design issues and considerations associated with the edgedirected approach when it is applied to different hardware implementations of stereo correspondence algorithms, including both fixedsupport and even more complex ADSW algorithms.
The paper extends our initial work on a hardware implementation of an edgedirected fixedsupport stereo correspondence algorithm presented in [7]. In particular, the extended work generalizes the idea of using edge information as part of the stereo correspondence process and presents design principles and considerations that arise when integrating edge information on the architecture of an ADSW algorithm as well. The paper also explores the effects of different edge detection algorithms on the ADSWbased architecture, showing the impact of each algorithm on the quality of the disparity maps produced by that architecture. Furthermore, the paper compares the two architectures (fixedsupport based versus ADSWbased) in terms of processing speed, disparity map accuracy and hardware overheads, when both are implemented on a Virtex5 FPGA platform. The experimental results indicate that the presented edgedirected approach for accelerating hardware implementations of stereo correspondence algorithms exhibits high potential for applications with hard realtime constraints, as both architectures achieve remarkable speedups relative to existing hardware architectures that implement equally complex algorithms. The architecture based on the fixedsupport algorithm can meet the realtime requirements of such applications, even for highresolution images. Moreover, the combination of the edgedirected approach with the ADSW algorithm enables the realization of accurate disparity computation systems that can also satisfy the realtime requirements of many embedded vision applications such as pedestrian and obstacle detection.
The rest of this paper is organized as follows. Section 2 discusses related work and Section 3 presents the edgebased stereo correspondence process. Section 4 presents the proposed edgedirected disparity map computation architectures. Section 5 presents the experimental framework and results, and Section 6 concludes the paper, discussing future work.
2. Related Work
There exists a large number of implementations in the literature that solve the stereo correspondence algorithm. Several existing works feature algorithms running on general purpose processors, as well as clusters and multiprocessor systems [8–10]. However, most of these implementations require highend computational equipment in order to compute the disparity map in realtime, especially as the image resolution increases. As a result, the realtime processing of those implementations is limited only to smallsized images (smaller than VGA). Alternatively, specialized hardware, such as Intel MMX [11], Graphics Processing Units (GPUs) [12, 13], and Digital Signal Processors [14–16], has been used to accelerate the disparity map computation. [11] attempts to improve the accuracy of the simple SAD correlation technique by using a multiple window approach that decreases errors at object boundaries. This however requires a lot of computational resources. [12, 13] present GPUbased implementations that achieve high frame rates that address the computational needs of stereo vision algorithms; however, they are not currently suitable for embedded and mobile applications due to their power demands. Approaches implemented on highend fixed point DSPs [14–16] consume lower power; however, they do not provide the parallelism of FPGAs and ASICs, thus are less suitable for stereo vision applications with hard realtime requirements. There exist also attempts to implement stereo vision algorithms on the Cell processor [17, 18], but are subject to restrictions imposed by the Cell platform, mainly due to the limited memory of the Synergistic Processing Elements (SPEs).
The last decade features an emergence in dedicated hardware architectures, as a means to address the aforementioned constraints found in software and specialized hardware approaches. Dedicated hardware architectures, implemented on ASICs and FPGAs, can exploit the parallelism inherent in the stereo correspondence process and optimize memory access patterns to provide fast computation. Most of dedicated hardware implementations have been implemented on FPGA platforms, taking advantage of the reconfiguration offered by these platforms, in order to exploit the intrinsic parallelism of the stereo correspondence algorithm. A stereo depth measurement system on an FPGA is introduced in [19]. It generates disparities on images at 30 fps, implementing a windowbased correlation search technique. Another system presented in [20] yields 20 fps on image sizes. However, the memory access pattern utilized does not provide scalability and performance optimization, limiting the overall performance of the system. In [21], a survey of some FPGA implementations, [22–24], is presented, and the survey highlights the common use of the SAD similarity measure, a relatively simple technique suitable for hardware implementations. Furthermore, [22–24] make extensive use of parallelism and pipelining in order to achieve realtime performance. [24] also raises the impact of the image size on the performance of the disparity map computation; it claims 5063 fps using very small images. The frame rate, however, decreases exponentially as the image size increases.
Other dedicated hardware architectures suitable for realtime disparity map computation are given in [25–31]. These works attempt to implement highperformance stereo correspondence algorithms that fail to achieve realtime performance in software. The FPGAbased architectures presented in [25, 26] implement local fixedsupport algorithms using the SAD similarity measure and compute intermediatesized disparity maps at a rate of 768 and 600 fps, respectively. The system in [27] is based on a phasebased computational model, as an alternative to feature correspondence and correlation techniques. That work exploits the parallel computing resources of FPGA devices to produce dense disparity maps for highresolution images at 52 fps. The supported disparity range, however, is not practical for embedded stereo vision systems, which traditionally utilize much larger disparity ranges, especially when highresolution images are used. Another implementation of a computationally complex disparity algorithm that is based on locally weighted phase correlation is presented in [28] and utilizes 4 FPGAs to produce dense disparity maps of size at the rate of 30 fps. A more recent FPGA implementation of a realtime stereo vision system is presented in [29]. That system generates dense disparity maps based on the Census transform. Lastly, the hardware implementation presented in [31] performs a modified version of the Census transform in both the intensity and the gradient images, in combination with the SAD correlation metric (SADIGMCT algorithm), achieving 60 fps on images.
The majority of the aforementioned implementations adopt local fixedsupport algorithms to achieve realtime performance, as these algorithms can be greatly benefited by the use of parallel and straightforward structures, which are key factors available in dedicated hardware implementations. ADSW algorithms, which traditionally achieve better disparity map accuracy, have been rarely implemented on dedicated hardware architectures. As mentioned above, these algorithms are computationally more expensive compared to fixedsupport algorithms and hence require complex and hardwareunfriendly operations. To the best of our knowledge, only the work in [30] implements a local ADSW stereo correspondence algorithm. That work proposes an accurate, hardwarefriendly disparity estimation algorithm called minicensus ADSW, and its corresponding realtime VLSI architecture that achieves 42 fps on image sizes. There has been a good effort in [30] to reduce the computational complexity of the ADSW and achieve realtime performance. However, that work achieves realtime performance for relatively smallsized images.
This work investigates the integration of edgedirected information into hardware architectures of stereo correspondence algorithms, as a means to restrict the stereo correspondence process only to the specific features (edges) of the input images (rather than to all image points), thus reducing the search space considerably. The proposed approach targets realtime embedded vision applications and can benefit both fixedsupport and ADSW algorithms. When applied to fixedsupport algorithms, it leads to very efficient implementations that can effectively address the hard realtime constraints of such applications even when using high resolution images and large disparity ranges. When applied to ADSW algorithms, the proposed approach provides a good tradeoff between accuracy of the disparity maps and processing speed; the reduction of the search space leads to realtime implementations for image sizes which are larger than the ones used in [30]. A comparison between the proposed work and other existing disparity map computation systems is given in Table 7.
3. EdgeBased Stereo Correspondence Process Overview
3.1. Summary of the Algorithm
In this section we present the overall flow of the edgebased stereo correspondence process. The description is generic and applies to both fixedsupport and adaptivesupport weight (ADSW) algorithms. In general, there are three major tasks associated with an edgebased stereo correspondence algorithm: edge detection, stereo correspondence, and interpolation. Figure 1 illustrates the algorithm’s steps towards computing a disparity value for each pixel in . The algorithm starts by determining whether a pixel corresponds to an edge or not. For each pixel corresponding to an edge, the algorithm extracts an correlation (or support) window centered on in , and an support window centered on in (the coordinate of is , where lies in the range to ). The algorithm then computes a pointwise score for any pixel corresponding to and aggregates the scores spatially over the support windows.
Fixedsupport and ADSW stereo correspondence algorithms differ in the way they perform the pointwise score computation and aggregation steps. The fixedsupport algorithm computes a pointwise score for any pixel corresponding to as the absolute different (AD) of and . After the computation of the AD values, the final aggregated cost is computed by summing all pointwise scores as
Equation (1) holds only for the case of a fixedsupport algorithm. In the case of an ADSW algorithm, each pointwise score is computed by multiplying the absolute difference of and by a weight coefficient and a weight coefficient . These weight coefficients are assigned to each pixel in a support window, based on the pixel’s spatial distance, as well as on its distance in the CIELAB color space with regard to the central pixel ( or) in the window. The final aggregated cost is computed by summing up all the weighted pointwise scores, and normalizing by the weights sum as in (2), where, and are the Euclidean distance between two coordinate pairs and two triplets in the CIELAB color space, respectively, and and are two parameters of the algorithm
Both the fixedsupport and ADSW algorithms compute an aggregated cost for all disparity levels in the range , and the best disparity for the pixel is found by locating the disparity with the minimum aggregated cost through a winnertakesall (WTA) approach. In the case where pixel does not correspond to an edge, the disparity is obtained by an interpolation method.
3.2. Discussion
The edgebased stereo correspondence algorithm described above applies an edge detection process over the input image prior to performing the correlation step. Edge detection returns locations in the image that indicate the presence of an edge [1]. These locations, described by the edge points, determine the outline (i.e., perimeter) of an object and distinguish it from the background and other objects [1, 32]. The correlation window in is moved only to the edges (and not to each possible pixel) along the working scanline, resulting in a considerable reduction in the search space. As a result, the correlation step computes only a disparity value for the pixels that correspond to an edge, while the disparity values of the remaining pixels are computed using interpolation, which however is considered computationally less expensive compared to stereo correspondence.
Another consideration of the edgebased stereo correspondence process is whether it uses the edges only as directive points or as matching primitives as well (instead of using pixel intensities). Since edges are encoded using only one bit per pixel, rather than 8 bits (as used in grayscale images), matching edges instead of pixel intensities reduce the computational space and complexity of the search and match operations, as well as the data path requirements. However, this can be applied only to the fixedsupport algorithm. Computing pointwise scores using binary data returns only zeros or ones, and this would invalidate the weights in the ADSW algorithm.
4. Hardware Realization of EdgeAccelerated Disparity Map Computation Architectures
This section presents the hardware architectures of two different edgebased stereo correspondence algorithms: an SADbased fixedsupport algorithm and an ADSWbased algorithm. We present design principles and considerations about the implementation of these algorithms on hardware, and discuss issues related to the memory structures needed, the organization of the processing blocks, the parallelism exploited, and so forth.
4.1. Disparity Map Computation Hardware Architecture Based on the FixedSupport Algorithm
4.1.1. Design Issues and Requirements
This architecture implements an SADbased, fixedsupport algorithm and uses edges both as directive points and as matching primitives. This requires that the architecture should be able to perform edge detection on both input images. Moreover, the use of binary data during the matching process reduces the computation of the SAD values (cost computation step) to a hamming distance operation that can be directly implemented in hardware using only addition and 1bit subtraction operations. The simplicity of the logic required during the cost computation step enables the parallel design of the architecture, so that to target a large number of search and match operations performed in a single cycle. However, the amount of parallelism that can be extracted depends on the ability of the edge detector to generate edge points (as the matching process is performed using edges). In this work we integrate the Sobel detector mainly due to its simple hardware structure, which can be easily parallelized to provide more than one edge points per single cycle. While the use of edge information reduces the search space and simplifies the logic required, the hardware implementation of the algorithm is not straightforward, but the overall design poses significant challenges in terms of the memory structures used in order to account for the irregular rates of data due to the edge only computation. The architecture requires a clever arrangement of the onchip memory in order to be able to process multiple support windows centered at the edge points, while skipping the nonedge points found between successive edges (every clock cycle).
4.1.2. General Structure
The block diagram of the architecture is shown in Figure 2 and consists of two major hardware units: the Edge Detection Unit (EDU) and the Disparity Computation Unit (DCU). It also consists of a Memory Controller and Control Unit (MCCU) that optimizes memory access based on the algorithm requirements and coordinates data transfers and handshakes between the EDU and the DCU. The EDU converts the rectified input image pair (grayscale images) into a pair of binary images (black and white) that characterize only the edges of the initial images. The black and white images are then fed into the DCU, which performs correlation with the objective to compute the disparity map of the input image pair. While the disparity maps generated by the DCU are sparse (disparity estimates are provided only for points corresponding to edges), the system can generate dense disparity maps as well, by integrating interpolation methods as part of the MCCU. In this work, we use the simple and fast nearest neighbor interpolation method, as our emphasis has been on performance in embedded scenarios.
The EDU and the DCU, which communicate through the use of internal memory (FIFO queues), are pipelined, and thus operate concurrently. They are also provided with scanline buffers, which temporarily store the pixels needed to perform convolution (in the case of the EDU), or correlation (in the case of the DCU). This reduces the clock cycles required to load image data from the input port, by exploiting the fact that working windows moved over the image use overlapping pixels. The scanline buffers are organized into FIFO structures and their size depends on the size of the working window ( for the EDU, for the DCU) and the width of the image, N. The delay to fill the scanline buffers is proportional to the I/O bandwidth.
4.1.3. EDU Architecture Overview
The Edge Detection Unit (EDU) integrated to the system implements the Sobel edge detector, which performs a 2D spatial gradient measurement on the input grayscale images using a pair of convolution masks. The masks hold data values between −2 and 2; thus the overall convolution can be implemented in hardware using shifters instead of multipliers. We integrated this detector mainly due to its simple hardware structure, which can be easily parallelized to provide more than one edge points per single cycle. This is important, as the ability of the DCU to perform parallel computations depends on the ability of the ECU to provide multiple edge points per cycle. The architecture of the EDU is shown in Figure 2(a). The EDU employs hardware features, such as parallelism and pipelining, in an effort to parallelize the repetitive calculations involved in the Sobel operation, and uses optimized memory structures in order to reduce the memory reading redundancy. The detector architecture consists of an I/O controller that reads/writes pixel data from/to the I/O port, onchip memory for storing pixel data (scanline buffers) and the convolution mask values, a series of convolution units (MUL_ADD units), as well as comparators. The architecture is pipelined into 2 stages: INPUT/OUTPUT and PROCESSING. In the INPUT/OUTPUT stage the I/O controller fetches 4 pixels from the I/O port (2 pixels per image), in a rowwise fashion, and forwards them to the input ports of the scanline buffers (16 bits to each buffer), which have a FIFO structure. During the PROCESSING stage, the scanline buffers produce four successive windows per cycle, which are convoluted with the Sobel masks by the MUL_ADD units. This pipeline stage returns 1bit pixel intensity values, which are concatenated into two 16bit registers. The values from the 16bit registers are forwarded to the output port (FIFO queues of the DCU) once every 8 clock cycles.
4.1.4. DCU Architecture Overview
The DCU has been designed in such a way as to calculate sparse disparity maps covering disparity ranges up to 120 pixels and correlation window sizes from to . The architecture of the DCU is shown in Figure 2(b) and involves three major steps: INPUT/OUTPUT, SADs_COMPUTATION, and MINSAD_COMPUTATION. The architecture consists of an I/O controller that reads/writes data from/to the I/O port, onchip memory (FIFO queues, scanline buffers and window buffers) for temporarily storing the edge data, and a unit that searches and finds the position of the edge to be processed (edge tracking unit (ETU)). It also consists of a collection of adders and subtractors to compute the SAD values (1) for all disparity levels, a unit to compute the minimum SAD value, multiplexers, and intermediate pipeline registers. Apart from reading/writing data from/to the I/O port, the I/O controller also acts as a control unit, coordinating memory accesses from the onchip memory, as well as data transfers and handshakes between the components of the DCU.
The DCU unit was designed with emphasis on parallelism, targeting a large number of search and match operations performed in a single clock cycle. This is facilitated by the simplicity of the adders and subtractors used (due to the use of binary data), and by the organization of the adders and comparators in tree structures. Due to its pipelined and parallel structure, the DCU presents good scalability in terms of correlation window size and disparity range. Particularly, it can compute the SAD values for all possible positions of the shifting window with a maximum size of in two clock cycles, and the minimum SAD value for a maximum of 120 disparity levels in three cycles. However, once the pipeline fills up and the FIFO queues are not empty, the DCU can provide a disparity value at the output every clock cycle.
4.1.5. DCU Memory Architecture and Edge Tracking Process
To keep a constant flow of data in the pipeline, the DCU must be able to locate one edge in the reference image every clock cycle, while discarding the nonedge points found between successive edges. At the same time, the DCU must have parallel access to the window surrounding the edge found in the reference image, as well as to the corresponding windows from the target image. For these reasons, each scanline buffer used in the DCU consists of a series of 16bit registers and can store scanlines from an input image. We avoid using 1bit registers in order to facilitate more parallelism and to make the process of discarding the nonedge points fast. The 16bit registers are organized into FIFO structures and allow parallel access to their elements. Specifically, the scanline buffers for the reference image (SBR) output 16 successive windows (stored in the candidate window buffers), while the scanline buffers for the target image (SBT) output 16 successive ) windows (stored in the candidate disparity range buffers). The SBR also outputs a 16bit vector (search vector) from positions to of the scanline, where . The search vector is being searched for potential edge points by the edge tracking unit (ETU), which is the connection point between the INPUT/OUTPUT stage and the remaining stages. The ETU works by locating an edge and its corresponding position in the 16bit search vector every clock cycle. The positions of the edges found during the searching process are used to select the window and disparity range buffers (among the 16 candidates) corresponding to the edge points found; the selected buffers become the input of the next pipeline stage. It must be noted that the ETU requires from 1 cycle (in the best case) to 16 cycles (in the worst) to locate all edges in the search vector. During this period, the edge_found signal is set to 1 and the scanline buffers are disabled, so that the content of the candidate window and candidate disparity range buffers remains constant. When all edges in the search vector are located, the ETU sets the edge_found signal to 0. This informs the I/O controller to fetch new edges from the input FIFO queues and to shift the scanline buffers to the right.
4.1.6. DCU Pipeline Stages Description
The major pipeline stages of the DCU are described below.
(i) INPUT/OUTPUT
This pipeline stage fetches pixel data from the I/O port, executes the edge tracking process, and selects the windows corresponding to the edges found. The data fetched from the input port (the two 16bit vectors produced by the EDU) is stored into FIFO queues. The I/O controller reads data from the input FIFO queues (if they are not empty) and forwards data to the scanline buffers (16bits to each scanline buffer), until the first scanlines from both images are stored into the scanline buffers. After the scanline buffers are filled, the I/O controller reads new pixel data from the queues only if the edge_found signal is set to 0 by the ETU. While the edge_found signal is set to 1, the scanline buffers are disabled, and during this period the edge tracking process described above is performed. If there is new data available at the input during this period, this data is written to the input queues. Furthermore, during this pipeline stage, the I/O controller writes the disparity value computed in the previous cycle to the output port.
(ii) SADs_COMPUTATION
The SAD values for all disparity levels are computed during this pipeline stage. The stage consists of absolute difference (ABDIF) units, which compute the absolute difference between the correlation window (stored in the Window Buffer) and the windows (stored in the disparity range buffer). Each ABDIF unit receives as input two bit vectors, whose elements are the edge points of the correlation windows, and consists of 1bit subtractors that compute the absolute difference of the edge points. The output of each ABDIF unit is an bit vector, which is next added bitwise using binary tree adders (BTA). Given the maximum supported correlation window size, and 1bit pixel intensities, the maximum value of the addition operation cannot be greater than 121. As such, the outputs of the BTA units are 7bit values.
(iii) MINSAD_COMPUTATION
The SAD values for all disparity levels in the range [1 : ] are compared with each other in order to compute the minimum value and its disparity. The comparison is carried out by a collection of 7bit comparators and registers, arranged in tree structure to reduce the delay of the longest path. As stated previously, this stage was further divided into 3 pipeline stages in order to meet the targeted operating frequency (100 MHz). Figure 3 shows the circuit that computes the minimum SAD value and its disparity. Each minSAD unit receives as input two 14bit vectors, each of which is a concatenation of an SAD value (7 bits) and its corresponding disparity (7 bits—up to 120 disparity levels). The minSAD unit compares the two SAD values and outputs the minimum of them along with its disparity. The entire circuit for computing the minimum SAD value and its disparity consists of multiple minSAD units, arranged in a structure of a binary tree of log_{2}() levels.
4.2. Disparity Map Computation Hardware Architecture Based on the ADSW Algorithm
4.2.1. Design Issues and Requirements
This architecture implements an ADSW algorithm, which is computationally more expensive compared to the fixedsupport algorithm both in terms of calculations and memory requirements. This is attributed in a significant way to the complex equation involved in the algorithm (2). In this work, we adopted some hardware optimization techniques that aim to simplify the algorithm and make it hardwarefriendly and suitable for embedded constraints. We first eliminate the use of the spatial distance during the weight computation step based on our observation that it affects the accuracy of the disparity maps slightly. We also use YUV instead of CIELAB color representation during the computation of the weight coefficients. This allows the use of unsigned integers instead of signed floatingpoint integers, which are complex and hardwareunfriendly. Furthermore, the computation of the color distance between two YUV triplets is performed using Manhattan rather than Euclidean distance. In this way, the square and square root operations are replaced by simple absolute difference and addition operations. In addition, the exp function is approximated by the function, which assigns a maximum weight of 256 if the color distance is zero and a weight of 0 if the color distance is greater than 8. This function simplifies the circuits that implement the multiplication of the weight coefficients with the pointwise scores, as multiplications are reduced to left shift operations. The cost function is further simplified by setting to a power of 2 (32 in our case). This converts the division to a right shift operation. Lastly, the denominator of (2) is approximated by the nearest power of 2 during the cost aggregation step, allowing the division to be replaced by a right shift operation.
4.2.2. General Structure
The architecture implementing the ADSW algorithm consists of two major pipeline stages: the Input Stage (IS) and the Calculation Stage (CS). The IS consists of a memory controller, 2 RGB to YUV color space converters (rgb2yuv) and 2 RGB to grayscale converters (rgb2gray). The memory controller fetches the RGB color values corresponding to the support windows and in a columnwise fashion (1 pixel value per input image every clock cycle) from the external memory. Those values are then converted to grayscale by the two rgb2gray units, and to their corresponding 8bit YUV representation by the rgb2yuv units. The image values computed by the IS are temporarily stored into onchip buffers (memory arrangements (MAs)), ensuring that there is always sufficient data for the CS, which is responsible for the calculation of the disparities. The overall system architecture also consists of a control unit that coordinates data transfers and handshakes between the different system units. Figure 4 shows a block diagram of the architecture and the data flow between units.
4.2.3. Memory Architecture
The IS and CS stages communicate through the use of onchip buffers (memory arrangements (MAs)), which temporarily store the pixels required to perform correlation between in and the candidate support windows in . The system is provided with 4 MAs per input image, which store the Y, U, V color values and the grayscale values. Figures 5(a) and 5(b) show the architectures of the onchip MAs for the reference and target image, respectively. Both MAs consist of a column buffer, a series of FIFO queues, and window buffers. The column buffer consists of 8bit registers that store the pixels of an entire column of a support window. It receives one pixel per clock cycle and outputs a column every cycles. The output column is stored in a series of FIFO queues (1 pixel per queue), which are used to allow the memory controller to continuously fetch data from the external memory to the onchip MAs (given that there is free space in the queues) irrespective of the data consumption rate of the CS. The CS consumes data at irregular rates; it consumes a column in cycles if is an edge and in a single cycle if is not an edge.
(a)
(b)
The data from the queues is forwarded to the window buffers, which form the inputs of the CS. The window buffer of the reference MA consists of 8bit registers, while the window buffer of the target MA consists of registers ( is set to 0). The use of registers allows parallel access to the window buffers; after an initial delay of cycles per scanline (dominated by the cycles needed to fill in the window buffer of the target MA), both onchip MAs can provide an window per cycle. The window buffer of the target MA is organized in a cyclic structure and is provided with a series of multiplexers at its input, which determine whether the input comes from the FIFO queues or from the rightmost registers of the window buffer. This structure is adopted to enable data reuse.
4.2.4. Calculation Stage (CS)
The CS consists of an Edge Detection Unit, two units for the generation of the weight coefficients (weight generators), a unit that computes the aggregated costs and selects the disparity with the minimum cost, and a unit responsible for the nearest neighbor (NN) interpolation step. The Edge Detection Unit implements the Canny operator, which performs image smoothing using Gaussian convolution, vertical and horizontal gradient and angle calculation, nonmaximum suppression and thresholding. The detector employs hardware features such as parallelism and pipelining in order to provide an edge point every single clock cycle.
The weight generator computes the weight coefficients for a support window in parallel, based on the YUV color values fetched from the [Y,U,V] MAs, and by utilizing instances of the circuit shown in Figure 6. That circuit consists of a Manhattan distance core and a weight table (LUT). Since the multiplication of the pointwise scores by the weight coefficients (cost aggregation step) is performed using shifters instead of multipliers, each location of the LUT stores the shift amount corresponding to the weight coefficient . This shift amount is equal to the binary logarithm of , except from values of greater than 8, for which a binary logarithm does not exist. In that special case, the corresponding entries in the LUT are set to a number, which is large enough so that the result of a shift operation by that number is equal to zero. The weight coefficient is looked up in the LUT using the color distance generated by the Manhattan distance core as index.
The architecture of the cost aggregator and WTA unit is shown in Figure 7. The unit has been design in a parallel manner and utilizes absolute different circuits that compute the pointwise scores between corresponding pixels in and . Those scores are then shifted by the shift amounts corresponding to the weight coefficients and using a series of left shifters (equivalent to multiplying the scores by and ). The final aggregated cost is computed by summing the outputs of the left shifters using a tree adder, and then normalizing (dividing) it by the weights sum, which, before being used for division, is rounded to the nearest power of 2 by using tree comparators. This enables a costeffective implementation of the division using a right shifter. Finally, the WTA unit is responsible for selecting the disparity with the minimum cost.
5. Experimental Results
5.1. Experimental Platform and Methodology
The architectures presented in this paper have been implemented on the ML505 Evaluation Platform, which features a Virtex5 LX110T FPGA. The basic features of the edgedirected disparity map computation architectures are listed in Table 1, while more details are presented in the following subsections. We evaluated the architectures using rectified synthetic and realworld data, initially stored in the compact flash memory card. The synthetic data includes stereo images from the Middlebury database [2] and the pedestrian dataset in [33], and the realworld data includes stereo images taken in the lab. The images were loaded into the onboard DRAM using the Microblaze softprocessor and were used as input to the systems shown in Figures 2 and 4, respectively. The resulting disparity maps were directed to a TFT monitor. The evaluation results of the synthetic images are shown in Figure 8, while the evaluation results of the realworld images are shown in Figure 9. The architectures were compared in terms of processing speed, disparity map accuracy, and FPGA resource utilization. They were also compared against their equivalent hardware systems that do not integrate the edge detectors in order to emphasize on the benefits of the edgebased approach adopted by the architectures.

5.2. Disparity Map QualityImpact of Edge Detection
To evaluate the quality of the disparity maps generated by the proposed edgedirected hardware architectures, and to examine the impact of the edge detection algorithm in the overall system quality, we use Middlebury stereo pairs, as well as stereo pairs from [33], for which the ground truth disparity maps are known and measure the incorrect disparity estimates using the percentage of bad pixels, a commonly accepted metric [2].
In our previous work in [7], we have investigated the impact of three different edge detectors (Sobel, Canny, and Evolvable [34]) on the accuracy of the disparity maps generated by an SADbased, fixedsupport algorithm, in order to select the best detector to be integrated to the system described in [7]. As we discussed in our previous paper, we have selected the Sobel detector, both for its simplicity on an FPGA and also for its good accuracy. In this paper, we also investigate the impact of the same detectors on the ADSW algorithm. The results about the percentage of bad pixels for the Tsukuba and Venus stereo image pairs are given in Table 2. The average percentages of data reduction associated to the results presented in the table are 82.95%, 70.23%, and 57.67% for the Canny, Sobel, and Evolvable detectors, respectively.

As can be seen, the Canny detector achieves on average better accuracy (lower percentage of bad pixels) compared to the other two edge detectors when is being applied to the ADSW algorithm, but the quality of the disparity maps is lower when using the Canny detector in the fixedsupport algorithm. This difference in the behavior of the Canny detector in the two algorithms lies in the fact that the fixedsupport algorithm utilizes edges both as directive points and also as matching primitives. As a result, the fixed support algorithm works better if the edge detector preserves not only strong edges, but also “busy” texture regions (edge regions with numerous small edge elements), which aid the matching process. The Canny detector does not work well with the fixed support algorithm as it is less likely to be fooled by noise, and more likely to detect true weak edges. As such, we have selected the Canny detector to be integrated into the ADSWbased disparity map computation architecture.
For a detailed quality analysis, we compare the quality of the disparity maps generated by the edgedirected architectures to the disparity maps generated by their equivalent hardware architectures that do not integrate the edge detectors (standalone fixedsupport and ADSW architectures). The standalone fixedsupport architecture has a similar structure with the edgedirected architecture shown in Figure 2 (without the EDU, the ETU and the candidate window and disparity range buffers), which is able to process 8bit pixels instead of edges. The standalone ADSW architecture has a similar structure with the edgedirected architecture shown in Figure 4 (without the edge detection and interpolation units). For simplicity purposes, we will refer to each hardware system configuration as “System 1–4,” as defined in Table 3.

The results extracted by comparing the disparity maps of Systems 1–4 to the ground truth disparity maps are presented in Table 4 for two sample image pairs. The resulting disparity maps are also given in Figure 10 for a qualitative comparison. Obviously, the disparity maps generated by the systems that implement the ADSW algorithm are more accurate, and particularly at depth discontinuity regions and at regions with repetitive patterns, where ADSW algorithms traditionally work better. The results also indicate that the use of the edge detectors does not impact negatively the accuracy of the disparity maps, but, in some cases, they improve the accuracy and especially at depth borders. This is because the standalone architectures suffer from changes in pixels intensities, while the edgedirected architectures potentially struggle this problem by limiting the correspondence search to specific reliable features in the images (edges in our case). It must be noted that the aforementioned results are based on a nearest neighbor interpolation method. It is anticipated that the proposed edgedirected hardware systems can produce disparity maps with better quality by integrating more complex interpolation methods such as bilinear and bicubic interpolation; this however is left as part of future work.
 
map quality results for System 1 and System 3 were extracted using Matlab simulation. 
5.3. Processing Speed
The processing speed of the edgedirected disparity computation architectures described in Section 4 is measured in frames per seconds (fps) and is affected by several algorithmic and implementationspecific parameters. The algorithmic parameters include the amount of data reduction (the percentage of nonedge points over the total image points) obtained from the edge detectors, the image size and the representation of the input images (grayscale or color), as well as the disparity range. The implementationspecific parameters include the operating frequency, the I/O bandwidth from the external memory, and the available FPGA resources.
The architecture based on the fixedsupport approach (System 2) has lower requirements in terms of input bandwidth, as it processes grayscale stereo images. Therefore, the external memory bandwidth can be exploited to fetch multiple pixels per clock cycle compared to the architecture that adopts the ADSW algorithm (System 4), which fetches one pixel per input image. The frame rate of System 2 is also independent from the support window size and the disparity range, as that system utilizes the edge information not only as directive points, but also as matching primitives, thus exploiting more parallelism. System 2 can compute the aggregated costs for all support windows in a single clock cycle. System 4 requires clock cycles to compute the costs for all disparity levels as that system processes color stereo images and uses grayscale pixels as matching primitives. Thus, it can exploit much less parallelism when assuming a constant amount of FPGA resources; particularly, it processes a support window every clock cycle, but requires clock cycles to compute a disparity value in the case where the processing pixel represents an edge. System 4, however, performs the matching process on a smaller number of edge points, since it is directed by the Canny detector, which achieves larger percentage of data reduction compared to the Sobel detector. With respect to the I/O bandwidth and the image size, the performance of both system architectures decreases as the image size increases and the I/O bandwidth decreases, since in the former case there is more data to be processed, while, in the latter case, the amount of data flowing into the system limits the system throughput.
To evaluate the processing speed of the edgebased architectures, we compare them with their equivalent system architectures that do not integrate the edge detectors (standalone architectures). We identify the speedup of the proposed architectures compared to the standalone architectures and provide results when increasing the input image size and the disparity range, in order to illustrate the impact of the edge detector in all cases. We used synthetic stereo images as benchmarks from [2, 33] in order to extract the processing speed of the systems mentioned above. The results are given in Tables 5 and 6.
 
frame rates for the systems without the edge detectors are projected frame rates. 
 
frame rates for the systems without the edge detectors are projected frame rates. 

As it can be seen, the processing speed is inversely proportional to the image size in all cases. Another particular observation is that the architectures based on the ADSW approach are slower even if they operate on higher operating frequencies and have larger memory bandwidth; this is due to the high computational complexity of the ADSW algorithm and due to the limited parallelism that can be exploited by such an algorithm in a resource constraint implementation platform such as an FPGA. The most important observation is that the use of the edge detection offers significant speedups, when compared to the standalone architectures for all image sizes. The impact of the edge detector is even more emphasized in the architecture based on the ADSW algorithm, as the data reduction of the Canny detector is larger compared to the Sobel. While the use of the edge detector in System 4 yields a speedup of ~5, the speedup obtained in System 2 is ~2. This is because System 2 was designed with more parallelism and provides a disparity value every clock cycle, irrespective of whether or not the pixel being processed is an edge. As a result, the edge detector in System 2 needs to provide multiple edges per clock cycle, so the bottleneck in this case is the Edge Detection Unit. Of course, the EDU in System 2 can further be parallelized, so the real limitation in terms of parallelism is the external memory I/O. On the other hand, System 4 exploits parallelism only on the level of a support window. If the pixel being processed is an edge, it requires cycles to compute the disparity, so the edge detector used in System 4 needs to provide only one edge every clock cycle. Table 6 indicates that the disparity range affects only the ADSWbased system architectures, due to the limited parallelism that can be exploited on the targeted FPGA. This however does not necessarily imply that the architecture is not scalable. The parallelism is limited by the amount of hardware resources. Of course, the ADSWbased architecture could be implemented to process multiple support windows in parallel on a larger FPGA. This would increase the frame rate as well.
Table 7 presents a comparison between existing implementations ([12, 13, 15, 16, 19, 20, 22–31]) and the proposed FPGA prototypes (System 2 and System 4). Performance is provided in frames per second (fps), as well as in points × disparity estimates per second (PDS). The proposed edgedirected architecture that implements the fixedsupport algorithm (System 2) achieves a PDS of and is the fastest implementation listed in the table, when considering the input image size and the maximum disparity range supported. System 4 achieves a PDS of for an image size of and a disparity range of 64. Such PDS seems sufficient considering that the algorithm implemented by System 4 is much more complex compared to the algorithms implemented by the other implementations in the table; only [30] implements an ADSW algorithm and achieves lower performance rates. Conclusively, performance results indicate that the proposed edgedirected approach for accelerating hardware implementations of stereo correspondence algorithms exhibits high potential for applications with hard realtime constraints. System 2 can meet the realtime requirements of such applications, even for highresolution images. Moreover, the combination of the edgedirected approach with the ADSW algorithm enables the realization of accurate disparity computations systems that can also satisfy the realtime requirements of many embedded vision applications such as pedestrian and obstacle detection.
5.4. Hardware Overheads
The proposed FPGA systems were evaluated for relevant metrics such as area and operating frequency. Table 8 gives the overall hardware demands of the FPGA prototype that is based on the fixedsupport algorithm (System 2). The table also lists the hardware overheads associated with each of the implemented components. The entire system utilizes ~83.8% of the FPGA slice LUTs and ~60% of the FPGA slice registers. The more hardware demanding component is the DCU, as obviously anticipated, due to the large amount of computations performed by that unit. The EDU requires only a small amount of resources, despite the fact that it processes two pixels per input image. This is attributed to the simplicity of the circuits that compute the convolution operation (using shifters instead of multipliers). Table 9 gives the overall hardware demands of the FPGA prototype that is based on the ADSW algorithm (System 4). That prototype utilizes ~73.1% of the FPGA LUTs and ~64.8% of the FPGA slice registers and can operate at 155 MHz. The slice LUTs are dominated by the cost aggregator and the weight generators, which consume ~44% of the available LUTs. The slice registers are dominated by the onchip MAs, which consume ~28% of the available slice registers.


A closer look in Tables 8 and 9 leads us to the following useful observations.(i)Even though System 2 exploits parallelism aggressively in order to compute the cost values for all support windows in the range [1 ] in a clock cycle, it requires only ~10% more FPGA LUTs than those required by System 4. This indicates that the edgebased approach for stereo correspondence is much more efficient (in terms of hardware usage) when applied to a fixedsupport algorithm, mainly due to the fact that it performs correlation using binary data (edges), thus requiring very simple circuits during the cost computation and aggregation steps.(ii)While the Sobel detector is computationally less demanding than Canny, it requires ~1% more FPGA LUTs. This is due to the fact that the EDU is able to process 4 pixels per clock cycle (2 per input image); thus, the hardware resources have been replicated by a factor of 4. The Canny detector, on the other hand, outputs only a pixel per cycle.(iii)The EDU operates at a slower clock frequency compared to the Canny detector. It could be possible to divide the EDU into more pipeline stages so that to increase its frequency, at the expense of further hardware usage. However, this would be meaningless considering that the DCU works at a slower frequency. Moreover, the frame rates achieved by System 2 are extremely high, and therefore any further increase of the operating frequency to achieve higher frame rates would not justify the extra hardware cost.
6. Conclusion
Stereo correspondence is an important algorithm in several embedded vision applications that require realtime computation of depth information. This paper presented an overview of the use of edge information, as a means to accelerate hardware implementations of stereo correspondence algorithms. The paper analyzed design issues and considerations associated with the edgedirected approach when it was applied to different hardware implementations of both fixedsupport and even more complex ADSW algorithms.
Our immediate plans involve the integration of other interpolation methods into the proposed architectures in order to achieve better disparity map accuracy. We also plan to investigate the applicability of the presented edgedirected approach to global stereo correspondence algorithms, which return better quality on the disparity map. Furthermore, the FPGA prototypes will be extended to fullcustom ASIC designs in order to evaluate largescale performance and power.
References
 B. Cyganek and J. P. Siebert, Introduction to 3D Computer Vision Techniques and Algorithms, Wiley, John & Sons, 2009.
 D. Scharstein and R. Szeliski, “A taxonomy and evaluation of dense twoframe stereo correspondence algorithms,” International Journal of Computer Vision, vol. 47, no. 1–3, pp. 7–42, 2002. View at: Publisher Site  Google Scholar
 M. Z. Brown, D. Burschka, and G. D. Hager, “Advances in computational stereo,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, no. 8, pp. 993–1008, 2003. View at: Publisher Site  Google Scholar
 K.J. Yoon and I.S. Kweon, “Adaptive supportweight approach for correspondence search,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, no. 4, pp. 650–656, 2006. View at: Publisher Site  Google Scholar
 F. Tombari, S. Mattoccia, and L. Di Stefano, “Segmentationbased adaptive support for accurate stereo correspondence,” in Proceedings of the 2nd Pacific Rim Conference on Advances in Image and Video Technology, vol. 4872 of Lecture Notes in Computer Science, pp. 427–438, Springer, December 2007. View at: Google Scholar
 M. Gerrits and P. Bekaert, “Local stereo matching with segmentationbased outlier rejection,” in Proceedings of the 3rd Canadian Conference on Computer and Robot Vision (CRV '06), p. 66, June 2006. View at: Publisher Site  Google Scholar
 C. Ttofis, S. Hadjitheophanous, A. S. Georghiades, and T. Theocharides, “Edgedirected hardware architecture for realtime disparity map computation,” in Proceedings of the IEEE Transactions on Computers, 2012. View at: Google Scholar
 K. Mühlmann, D. Maier, J. Hesser, and R. Männer, “Calculating dense disparity maps from color stereo images, an efficient implementation,” International Journal of Computer Vision, vol. 47, no. 1–3, pp. 79–88, 2002. View at: Publisher Site  Google Scholar
 S. Forstmann, Y. Kanou, J. Ohya, S. Thuering, and A. Schmitt, “Realtime stereo by using dynamic programming,” in Proceedings of the Computer Vision and Pattern Recognition Workshop (CVPRW '04), p. 29, June 2004. View at: Google Scholar
 L. D. Stefano, M. Marchionni, and S. Mattoccia, “A fast areabased stereo matching algorithm,” Image and Vision Computing, vol. 22, no. 12, pp. 983–1005, 2004. View at: Publisher Site  Google Scholar
 H. Hirschmüller, P. R. Innocent, and J. Garibaldi, “Realtime correlationbased stereo vision with reduced border errors,” International Journal of Computer Vision, vol. 47, no. 1–3, pp. 229–246, 2002. View at: Publisher Site  Google Scholar
 R. Yang and M. Pollefeys, “A versatile stereo implementation on commodity graphics hardware,” RealTime Imaging, vol. 11, no. 1, pp. 7–18, 2005. View at: Publisher Site  Google Scholar
 Q. Yang, L. Wang, R. Yang, S. Wang, M. Liao, and D. Nister, “Realtime global stereo matching using hierarchical belief propagation,” in Proceedings of the The British Machine Vision Conference, 2006. View at: Google Scholar
 R.P. M. Berretty, A. K. Riemens, and P. F. Machado, “Realtime embedded system for stereo video processing for multiview displays,” in Proceedings of the Stereoscopic Displays and Virtual Reality Systems XIV, Proceeding of SPIE, San Jose, Calif, USA, January 2007. View at: Publisher Site  Google Scholar
 B. Khaleghi, S. Ahuja, and Q. M. J. Wu, “An improved realtime miniaturized embedded stereo vision system (MESVSII),” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops, pp. 1–8, June 2008. View at: Publisher Site  Google Scholar
 C. Zinner and M. Humenberger, “Distributed realtime stereo matching on smart cameras,” in Proceedings of the 4th ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC '10), pp. 182–189, New York, NY, USA, September 2010. View at: Publisher Site  Google Scholar
 S. Cavanag and M. Manzke, “Real time disparity map estimation on the cell processor,” in Proceedings of the Eurographics Ireland Workshop 2009, pp. 67–74, Trinity College Dublin, December 2009. View at: Google Scholar
 J. Liu, Y. Xu, R. Klette, H. Chen, W. Rong, and T. Vaudrey, “Disparity map computation on a cell processor,” in Proceedings of the IASTED International Conference on Modelling, Simulation, and Identification (MSI '09), Beijing, China, October 2009. View at: Google Scholar
 H. Hile and C. Zheng, “Stereo video processing for depth map,” Tech. Rep., University of Washington, 2004. View at: Google Scholar
 Y. Miyajima and T. Maruyama, “A realtime stereo vision system with FPGA,” in Proceedings of the 13th International Conference on FieldProgrammable Logic and Applications (FPL '03), pp. 448–457, Lisbon, Portugal, 2003. View at: Google Scholar
 L. Nalpantidis, G. Sirakoulis, and A. Gasteratos, “Review of stereo matching algorithms for 3D vision,” in Proceedings of the 16th International Symposium on Measurement and Control in Robotics (ISMCR '07), pp. 116–124, Warsaw, Poland, June 2007. View at: Google Scholar
 M. AriasEstrada and J. M. Xicotencatl, “Multiple stereo matching using an extended architecture,” in Proceedings of the FieldProgramable Logic and Applications, vol. 2778, pp. 203–212, Springer, 2003. View at: Google Scholar
 S. H. Lee, J. Yi, and J. Kim, “Realtime stereo vision on a reconfigurable system,” in Proceedings of the Embedded Computer Systems: Architectures, Modelling and Simulation, vol. 3553, pp. 299–307, Springer, 2005. View at: Google Scholar
 M. Hariyama, Y. Kobayashi, M. Kameyama, and N. Yokoyama, “FPGA implementation of a stereo matching processor based on windowparallelandpixelparallel architecture,” in Proceedings of the IEEE International 48th Midwest Symposium on Circuits and Systems, (MWSCAS '05), vol. 2, pp. 1219–1222, Covington, Ky, USA, August 2005. View at: Publisher Site  Google Scholar
 C. Georgoulas and I. Andreadis, “A realtime occlusion aware hardware structure for disparity map computation,” in Proceedings of the Image Analysis and Process (ICIAP '09), vol. 5716, pp. 721–730, 2009. View at: Google Scholar
 K. Ambrosch, M. Humenberger, W. Kubinger, and A. Steininger, “A SADbased stereo matching using FPGAs,” in Proceedings of the Embedded Computer Vision part II, pp. 121–138, Spinger, 2009. View at: Google Scholar
 J. Díaz, E. Ros, R. Carrillo, and A. Prieto, “Realtime system for highimage resolution disparity estimation,” IEEE Transactions on Image Processing, vol. 16, no. 1, pp. 280–285, 2007. View at: Publisher Site  Google Scholar
 A. Darabiha, J. MacLean, and J. Rose, “Reconfigurable hardware implementation of a phasecorrelation stereoalgorithm,” Machine Vision and Applications, vol. 17, no. 2, pp. 116–132, 2006. View at: Publisher Site  Google Scholar
 S. Jin, J. Cho, X. D. Pham et al., “FPGA design and implementation of a realtime stereo vision system,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 20, no. 1, pp. 15–26, 2010. View at: Publisher Site  Google Scholar
 N. Y. C. Chang, T. H. Tsai, B. H. Hsu, Y. C. Chen, and T. S. Chang, “Algorithm and architecture of disparity estimation with minicensus adaptive support weight,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 20, no. 6, pp. 792–805, 2010. View at: Publisher Site  Google Scholar
 K. Ambrosch and W. Kubinger, “Accurate hardwarebased stereo vision,” Computer Vision and Image Understanding, vol. 114, no. 11, pp. 1303–1316, 2010. View at: Publisher Site  Google Scholar
 M. Raman and H. Aggarwal, “Study and comparison of various image edge detection techniques,” The International Journal of Image Processing, vol. 3, no. 1, pp. 1–12, 2009. View at: Google Scholar
 “A Framework for Evaluating StereoBased Pedestrian Detection Techniques,” http://www.cdvp.dcu.ie/datasets/pedestrian_detection/. View at: Google Scholar
 Z. Vasicek and L. Sekanina, “An evolvable hardware system in Xilinx Virtex II Pro FPGA,” International Journal of Innovative Computing and Applications, vol. 1, no. 1, pp. 63–73, 2007. View at: Google Scholar
Copyright
Copyright © 2012 Christos Ttofis and Theocharis Theocharides. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.