InTech uses cookies to offer you the best online experience. By continuing to use our site, you agree to our Privacy Policy.

Computer and Information Science » Computer Graphics » "Image Processing", book edited by Yung-Sheng Chen, ISBN 978-953-307-026-1, Published: December 1, 2009 under CC BY-NC-SA 3.0 license. © The Author(s).

Chapter 24

Acquisition and Digital Images Processing, Comparative Analysis of FPGA, DSP, PC for the Subtraction and Thresholding.

By Carlos Lujan Ramirez, Ramón Atoche Enseñat and Francisco José Mora Mas
DOI: 10.5772/7065

Article top

Overview

Operation of channel Link
Figure 1. Operation of channel Link
Input and output receiver signals
Figure 2. Input and output receiver signals
Camera Link receiver implementation
Figure 3. Camera Link receiver implementation
Clocks wave form
Figure 4. Clocks wave form
Receive Data Formatting
Figure 5. Receive Data Formatting
Camera recovered image.
Figure 6. Camera recovered image.
Contrast amplification.
Figure 7. Contrast amplification.
Original Image and its thresholding by 150.
Figure 8. Original Image and its thresholding by 150.
Bayer Format.
Figure 9. Bayer Format.
Positioning of the sensors and their respective image
Figure 10. Positioning of the sensors and their respective image
Four possible cases for the interpolation of components R and B
Figure 11. Four possible cases for the interpolation of components R and B
two possible cases for the interpolation of the G component.
Figure 12. two possible cases for the interpolation of the G component.
Results of the converter: RGB image
Figure 13. Results of the converter: RGB image
Background with RGB format
Figure 14. Background with RGB format
Animals in the pond with RGB format
Figure 15. Animals in the pond with RGB format
Subtraction of images in RGB format
Figure 16. Subtraction of images in RGB format
Blurring of images in RGB format
Figure 17. Blurring of images in RGB format
Background of the pond
Figure 18. Background of the pond
Animals in the pond
Figure 19. Animals in the pond
Subtraction of the images in Bayer format
Figure 20. Subtraction of the images in Bayer format
Blurring of the images in Bayer format
Figure 21. Blurring of the images in Bayer format
In (a) we have the background image, in (b) the image with objects to be detected.
Figure 22. In (a) we have the background image, in (b) the image with objects to be detected.
Result of the subtraction.
Figure 23. Result of the subtraction.
Result of the thresholding
Figure 24. Result of the thresholding
Processing time of the process previously described.
Figure 25. Processing time of the process previously described.
Resources used by the FPGA.
Table 2. Resources used by the FPGA.
Hardware processing time.
Figure 26. Hardware processing time.
Illustration of the processing time.
Figure 27. Illustration of the processing time.
Resources used by the FPGA
Table 3. Resources used by the FPGA

Acquisition and Digital Images Processing, Comparative Analysis of FPGA, DSP, PC for the Subtraction and Thresholding.

Carlos Lujan Ramirez1, Ramón Atoche Enseñat1 and Francisco José Mora Mas2

1. Introduction

Given the need to reduce the time involved in image processing, we found that is necessary to use new methods to improve the time response. In our application of real time tracking of different species in the study of marine animal conduct, a reduction in the image processing time provides us with more information that allows us to better predict the animal’s escape trajectory from a predator.

Nowadays, photographic cameras can deliver photographs in various formats, hence the need to provide them with pre-processing in order to deliver the required format. Although Photographic cameras are available that can deliver photos without pre-processing, the format provided is the Bayer format. A characteristic of this format is that we obtain the images almost directly from the CCD sensors and from the analogical to digital converter with no pre-processing. The only requirement is a deserialization step and registers in parallel whose role is to place the Bayer format image in a memory for the investigation (Lujan et al., 2007). In the initial method in our project, the camera is connected to a Frame Grabber which retrieves the image, converts it to RGB, changing it to a scale of greys and subsequently carries out the necessary processing, the first being the subtraction of the background from the image of the animals under study, followed by the application of blurring.

Subtraction of images is widely used when the aim is to obtain the animal’s trajectory (Gonzales et al., 2002), however, we use it to leave only the animals under study, eliminating everything else. A low pass filter (Gonzales et al., 2002) (blurred) is used which unifies any division in some of the animals caused by the subtraction. Occasionally a shrimp would be divided in two due to the kind of background used, recognizing two shrimps instead of one, this error was resolved by the application of the blurring.

The new process would involve recovering the image and then to immediately carry out the image processing of subtraction and blurring, saving time by avoiding the use of a Bayer to RGB encoder and the frame grabber. We included a stage of image recovery by a number of deserielizors and registers in order to deliver the image in Bayer format.

For this investigation, an analysis will be performed on the third part of image processing which begins after the image has been stored in an FPGA. The two processing stages analyzed will be the subtraction between the background and the image where the animals under study are found, and the subsequent thresholding from the subtraction.

This will be implemented in four different systems, and the time and resources consumed will be analyzed; first in a PC computer with a Borland C++ program, second with an embedded microprocessor programmed in C++, third with a DSP also programmed in C++, and finally with a hardware designed in VHDL.

2. Camera link digital images acquisition.

The need to acquire images at high transfer rates, it takes us to use a camera with camera link interface, and then, to save time when recovering the image; we connect the camera with FPGA instead of use a frame grabber.

2.1. The standard camera link.

The protocol camera Link (Camera Link, 2000) is a communication interface for vision applications. The interface extends the basic technology of the channel link to provide a more useful specification for vision applications.

For many years, there has been a need for a standard method for communication in the scientific and industrial digital video market. The manufacturers of frame grabbers and cameras developed products with different connections, making cable production difficult for the manufacturers and very confusing for the consumers. A standard connection between digital cameras and frame grabbers is becoming more necessary as the data transfer rate continues to rise.

Increasingly diverse cameras and advanced data transmission signals have made a standard connection such as camera Link an absolute necessity. The camera Link interface reduces both the support time and the cost of Support. The normal cable will be able to manage the increased speed of the signal and the cable assembly will allow clients to reduce their expenses.

2.2. The standard ANSI/TIA/EIA-644

The LVDS is a pattern of data transmission that uses a balanced interface and a low voltage to solve many of the problems relating to existing technologies. The lower amplitude of the signal reduces the voltage used for the line circuits, and a balance in the signals reduces the coupling noise, allowing greater transfer rates.

The LVDS, standardized in TIA/EIA-644, specifies a maximum transfer rate of 1.923 Gbps. In practice however, the maximum transfer rate is determined by the quality of the communication media between the transmitter and the line receiver. Similarly, the length and characteristics of a transmission line condition the usable transfer rate to the maximum (Cole, E., 2002)(Texas Instruments, 2002).

National Semiconductor (National Semiconductor, 2006) developed the technology of channel link as a solution for flat screens, based on LVDS for the physical layer. The technology was later extended into a method for data transmission. The channel link comprises a pair of drivers in the transmitter and in the receiver. The driver accepts 28 data signals of simple terminal and one clock signal of simple terminal. The data are transmitted in series at a rate of 7: 1 in four lines, and together with the clock they form five differential pairs. Therefore, the receiver accepts the four pairs of data lines in LVDS plus one pair for the clock signal which will recover the 28 bits of data and the clock signal, as shown in Fig. 1.

media/image1.png

Figure 1.

Operation of channel Link

One of the benefits of the channel link transmission method is that it requires fewer conductors for transferring data. Five pairs of cables, therefore, can transmit up to 28 data bits. These cables reduce the size of the connector, allowing the manufacture of smaller cameras. Furthermore, the data transmission rate in the channel link chipset can reach 2.38 Gbits/s.

2.3. Camera connection for digital image acquisition

In the first time the camera was connected a Frame Grabber and this to the computer, we propose not to use the Frame Grabber on the other hand to use a FPGA for image recovery and storage in FPGA memory for later processing. Our advantages is a savings of Frame Grabber, because its cost is high, and on the other hand we gain the total control of the image storage.

media/image2.png

Figure 2.

Input and output receiver signals

Taking into account Sawyer implementation (2008) acquisition stage was redesigned, the FPGA was exchanged for a Virtex 4 FPGA, and a memory was added so that the image could be accessed by other devices.

The Figure 2 shows us the camera signals, which enter the receiver, and this provides us with the 28 bit data in parallel. These 28 bits are composed of 24 data bits and 4 control bits used to determine a valid pixel, an end of line and an end table. Fig. 2 shows how the camera signals enter to the receiver in serial form and the receiver gives us 28 data bits in parallel. These 28 bits are composed of 24 data bits and 4 control bits used to determine a valid pixel, an end of line and an end frame.

Logic Description

The logic for the receiver is simplified by using the cascadable IOB DDR flip-flops. The Fig. 3, shows the 4-bit interface.

Two bits of data are clocked in per line, per high-speed rxclk35 clock period, thus giving 8 bits (4-bit data width) of data to register each high-speed clock. Because this high-speed clock is a 3.5x multiple of the incoming slow clock, the bit deserialization and deframing are slightly different under the two alignment events described in “Clock Considerations.” Basically, 28 bits (4-bit data width) are output each slow speed clock cycle. However, for 4-bit data width, the receiver captures 32 bits per one slow clock cycle and 24 bits the next. The decision as to which bits are output and when is made by sampling the incoming clock line and treating it as another data line. In this way, the multiplexer provides correctly deframed data to the output of the macro synchronous to the low speed clock.

There is no data buffering or flow control included in the macro, because this interface is a continuous high-speed stream and requires neither.

media/image3.png

Figure 3.

Camera Link receiver implementation

It’s important to be careful with the generation of the two clock signals rxclk35 and rxclk35not because the recovery of the correct images depends on them. The time diagram for these signals is shown in Fig. 4.

media/image4.png

Figure 4.

Clocks wave form

The Fig. 5 shows us how the image data are transmitted through the camera link protocol.

media/image5.png

Figure 5.

Receive Data Formatting

Is very important to be careful with the placement of flip flops from deserializer, because the transfer rate of data depends of them. (Lujan et al., 2007). No use of FPGA resources (for example Digital Clock Managers in Xilixs architectures) for skew correction allows us to use in other applications that really need them.

Finally, the pixels of the digital image are stored in memory and this can be used by other devices for future processings.

In Fig. 6 shows an image of 720 x 480 pixels, recovered and stored in memory.

3. Digital images subtraction and thresholding

Two concepts that we use are defined below.

3.1. Image subtraction

The difference between two images f(x,y) and g(x,y) is expressed by:

o(i,j) = f(i,j)  g(i,j)    i,j
(1)

This technique has several applications in segmentation and in image enhancement.

3.2. Thresholding

In order to understand this concept more clearly, we must first understand the concept of contrast amplification.

media/image7.png

Figure 6.

Camera recovered image.

Images with poor contrast are often the result of insufficient or variable illumination, or are due to the non-linearity or small dynamic range of the image sensors. A typical example of transformation is shown in figure 7, which can be expressed by:

v={αu0uaβ(u-a)+vaaubγ(u-b)+vbbuL}
(2)
media/image10.png

Figure 7.

Contrast amplification.

The slopes are taken greater than one in the regions in which contrast amplification is desired.

The parameters a and b can be estimated by examining the histogram of the image. For example, the intervals of grey level where the pixels occur with greater frequency must be amplified in order to enhance the visibility of the image.

A particular example of the above is when a = γ = 0, this case takes de name of cutoff. This cutoff can be useful for instance, when we know that the signal is within the range [a,b].

Thresholding is a special case of cut model in which a binary image is obtained.

media/image11.png

Figure 8.

Original Image and its thresholding by 150.

This can be expressed by:

v={00uaLaub0buL}
(4)

This model can be useful when we know that the image must be binary and when this has not been obtained by the process of digitalization. It can also be used in a segmentation process, as shown in Fig. 8.

4. Bayer format utilization for subtraction and blurring of images

4.1. Bayer format

At present, a large number of formats for storing digital images are available, however, most cameras capture the image in Bayer format (Bayer, 1976) which then passes through a converter and the image is delivered in the RGB format. The use of the Bayer color filter is a popular format for the digital acquisition of color images (National Semiconductor, 2006). A drawing of the color filters is shown in Fig. 9. Half the total number of pixels are green (G), while a quarter of the total are attributed to both red (R) and blue (B).

media/image13.png

Figure 9.

Bayer Format.

In order to obtain the color information, the color image sensor is covered with a red, green, or blue filter, in a repetitive pattern. This pattern or sequence of the filters can vary, but the “Bayer” pattern is widely used, it is an arrangement of 2x2 repetitive and was invented by Kodak. Fig. 10 shows the positioning of the sensors and the representation of the color generated in the image.

media/image14.png

Figure 10.

Positioning of the sensors and their respective image

To obtain an image in bit map, a Bayer to RGB converter is required, there are several methods. The bilinear interpolation method together with that of linear interpolation was chosen; this correlation was taken into consideration as it presents better characteristics in comparison with others (Sakamoto et al, 1998). A brief description of the method is as follows.

media/image15.png

Figure 11.

Four possible cases for the interpolation of components R and B

The values of R and B are interpolated in a straight line from the closest neighbors of the same color. There are four possible cases, as shown in Fig. 11. The missing values of R and B are interpolated on a green pixel, see Fig. 11 (a) and (b), taking the mean values of the two closest neighbors with the same color. For example, in Fig. 11 (a), the value of the blue component on a G pixel will be the average of the blue pixels above and below the G pixel, while the value of the red component will be the average of the two red pixels to the left and right of the G pixel. In Fig. 11 (c) we can observe that, when the value of the blue component is interpolated for an R pixel, we take the average of the four closest blue pixels surrounding the R pixel. Similarly, to determine the value of the red component on a B pixel (Fig.11 (d)), we take the average of the four closest red pixels surrounding the B pixel.

media/image16.png

Figure 12.

two possible cases for the interpolation of the G component.

The part of the linear interpolation method dealing with the correlation is as follows:

In Fig. 12(a) the value of the green component is interpolated on a R pixel. The value used for the G component is:

G(R)={(G1+G3)/2if|R1-R3||R2-R4|(G2+G4)/2if|R1-R3||R2-R4|(G1+G2+G3+G4)/4if|R1-R3|=|R2-R4|}
(5)

For Fig. 12 (b), the value of the green component is interpolated on a B pixel. The value used for the G component is as follows:

G(B)={(G1+G3)/2if|B1-B3||B2-B4|(G2+G4)/2if|B1-B3||B2-B4|(G1+G2+G3+G4)/4if|B1-B3|=|B2-B4|}
(6)

4.2. Implementation of the Bayer to RGB converter in a DSP

In accordance with the method previously described, we implemented the converter in the DSP 6416 of Texas Instruments, operating at a frequency of 1GHz (Texas Instruments, 2008). A summary of the implementation is presented in Table 1.

Size in Bytestotal of instructionsCPU cyclestime
2448283733523727310837.27 ms

Table 1.

results of the implementation in DSP 6416

In the input of the converter, we have a Bayer image with the characteristics shown in Fig. 9. After applying the converter, the output would be three matrixes with the characteristics shown in Fig. 13, in other words we would obtain an image three times the size of the input.

media/image19.png

Figure 13.

Results of the converter: RGB image

4.3. Results obtained in the tracking of the shrimp and crab.

For the following of the shrimp and crab, the first step is to subtract the background from the image of the two animals under study. The results obtained from this process using two different methods are detailed below.

4.3.1. Subtraction of images with RGB format.

First, the background of the pond must be captured (without animals). Fig. 14 shows the background and Fig. 15 shows the animals (mat), both images in RGB format.

p_image2=mat  background
(7)

Once we have the two images, the subtraction of images is carried out (Fig. 16) using the equation (1), after which a blurring is applied (Fig. 17).

4.3.2. Subtraction of images with Bayer format.

The procedure is the same as for the previous images, but in this case using the images in Bayer format directly before passing them through the converter. Fig. 18 shows the background image, and Fig. 19 the image with the animals.

In Fig. 20 we can observe the result of the subtraction and in Fig. 21 the result after blurring.

media/image21.png

Figure 14.

Background with RGB format

media/image22.png

Figure 15.

Animals in the pond with RGB format

media/image23.png

Figure 16.

Subtraction of images in RGB format

media/image24.png

Figure 17.

Blurring of images in RGB format

media/image25.png

Figure 18.

Background of the pond

media/image26.png

Figure 19.

Animals in the pond

media/image27.png

Figure 20.

Subtraction of the images in Bayer format

media/image28.png

Figure 21.

Blurring of the images in Bayer format

5. Use and justification of the tools for digital images blurring and subtraction.

Once determined that is not necessary Bayer's conversion to RGB the next step is to determine the tool that it optimizes the time, to perform subtraction and blurring of the digital images.

The first objective is to separate the background from the objects to be detected. The method used to eliminate the background is that of image subtraction, in other words, the image of the scenery or background is subtracted from the images where the objects to be detected are found, obviously without the presence of the objects. The resulting image will conserve only those agents not found in the background image, which in this case will be the objects. The pixels not pertaining to the objects will remain in a range of values very close to zero and will be perceived in the image as dark or almost black colors.

The difference between two images f(x,y) and h(x,y), expressed as:

g(x,y) = f(x,y)  h(x,y)
(8)

The effect being that only the areas in which f(x,y) and h(x,y) are different will appear in the output image with enhanced details.

The Fig. 22, shows the images that must be subtracted in order to detect the objects, in this case the crab and the shrimp.

media/image30.png

Figure 22.

In (a) we have the background image, in (b) the image with objects to be detected.

The result of the subtraction can be observed in Fig. 23.

media/image31.png

Figure 23.

Result of the subtraction.

The next step is to binarize the image by a thresholding. This is achieved by fixing a constant value of brightness, which will serve as a threshold to be able to distinguish between the sections of the image pertaining to the background and those pertaining to the objects of interest. If the brightness level of the pixel under analysis is greater than the established threshold, its value will be modified and established with the maximum brightness value in an image with an 8 bit resolution, and this value is “255”. If the brightness value is not greater than the threshold, its value will be modified to the minimum brightness value corresponding to deep black, i.e. “0”. The result of the thresholding can be observed in Fig. 24.

media/image32.png

Figure 24.

Result of the thresholding

In every system in which the subtraction and thresholding were implemented, the results shown in the figures above were obtained.

Results obtained in the 4 systems implemented. For all cases the systems was implemented using an image of 102 x 150 pixels in Bayer format (Bayer, 1976)

5.1. With the Borland C++ Builder compiler in a PC.

After carrying out the implementation and executing it in the builder (2002), the next step was to measure the processing times for each process, resulting in 140 ms for the subtraction and 160 ms for the thresholding. An advantage of developing the segmentation algorithms in the Builder compiler is that they can be modified very easily and, as we are dealing with a programming language in C++, this makes the code more portable. Furthermore, a person with little knowledge of hardware development could develop the implementations, and even improve them since this tool is software orientated and an elemental knowledge of programming is sufficient. It is important to note that greater efficiency can be attained in execution times, although the main objective of using this code was its functionality rather than its efficiency. For example, predictive filtering and other techniques of algorithm development could be added. Similarly, the capacity to emigrate this code to other more powerful programming tools, perhaps under other operating systems, could produce better results. This application was run in a computer with a 1.6 Ghz processor with which it could be possible to improve the performance by using computers with better processors.

5.2. With the Microblaze.

After downloading the project in the ML 401 board containing the “Microblaze microprocessor (Xilinx, 2002), the processing times were measured, both for the execution of the subtraction and thresholding only and for the execution of the same processes plus the blurring. In order to obtain the processing times, a flag was added to the main program which indicates when processing begins and which changes state when the execution of the segmentation algorithm ends.

The observation of the times required an oscilloscope where a testing point was placed on an output of the board and the measurement was taken. The results can be seen in Fig. 25.

media/image33.png

Figure 25.

Processing time of the process previously described.

The duration of the width of positive pulse observed in the above figure is the total processing time, and with a tool of the oscilloscope it was possible to measure the period of the positive pulse and the exact time was obtained. The legend +Width(1)=190.0ms can be observed in the lower left-hand corner of the screen, this is the time for the subtraction and thresholding.

An advantage of developing the algorithms in the EDK is that a C++ programming language is used to program the embedded microprocessor, again making the code more portable, however, it has some instructions reserved from the Xilinx. A greater knowledge of hardware design is also required in comparison with the Builder since a hardware platform must be elaborated; however, for the user who does not have this advanced knowledge, an assistant is provided to facilitate the development of the platform

For this implementation a 100MHz master clock is used, which is the oscillating crystal of the board used for this project. The results of the algorithms presented can be improved with the use of more powerful boards, such as faster clocks and boards that can handle these speeds.

Another way to improve these times is by using solid nucleus processors such as the Power PC, which has the advantage of being able to handle higher speeds and which has greater efficiency in clock cycles by instruction.

Resources used:

media/image34.png

Table 2.

Resources used by the FPGA.

5.3 With the processor designed in the Ise Foundation.

Once the complete electronic system(Xilinx,2006), had been elaborated and its functionality verified, the processing time of a 102 x 150 image was measured.

In order to obtain the processing times, a flag was added to the developed hardware which indicates when the image processing begins and when it ends. In this way the total time can be determined by measuring the time between each change in voltage level. The results are shown in Fig. 26.

media/image35.jpeg

Figure 26.

Hardware processing time.

The time measured is 153us for each frame processed and is indicated in the above figure with arrows.

The results were also measured with the use of a tool within the Ise Foundation that generates a test bench with which it is possible to simulate the complete electronic system. To facilitate visualization, the necessary pines were added and, as can be observed, the processing time given is 153 us.

media/image36.png

Figure 27.

Illustration of the processing time.

In the following table the resources used for the implementation of segmentation are specified.

One of the main benefits obtained from developing the segmentation algorithms in the hardware description language (HDL) is the parallelization of the processing, with which greater frame rates would be attained since it would be possible to increase the efficiency of instructions by clock cycles.

For this implementation a 100MHz master clock is used, which is the oscillating crystal in the board used for this project. The results of the algorithms presented could be improved by using more powerful boards, such as faster clocks and boards that can handle these speeds.

media/image37.png

Table 3.

Resources used by the FPGA

5.4. With the TMS320C6416” processor of Texas Instruments.

The subtractor and thresholder were subsequently used in the DSP 6416 of Texas Instruments, operating at a frequency of 1GHz (Texas Instruments, 2008). A summary of the implementation is described in Table 4.

Size in BytesT otal of instructionsCPU cyclesT ime
1564597545062585.06 ms

Table 4.

Resultados de la implementación en DSP 6416.

An advantage of developing the algorithms in the DSP is that a C++ programming language is used, which again gives more portability to the code, however, better results are obtained in comparison with the others implemented in C++, and in this case, an advanced knowledge of hardware design is not required, in contrast with the implementation in the Microblaze.

6. Conclusion

The image recovery is a fundamental part of the research, since all subsequent processes depend on it. To achieve this, departing from signals of Camera Link protocol, is very important to have a lot of care with the location of the receiver's flip-flops, to achieve the lowest possible skew, to reach the highest transfer rate, achieving even not to use the FPGA's resources of clock managing (DCM) and to have them available for future applications.

From the obtained results, we can appreciate that converting the images to RGB format is not necessary, in other words the images obtained from the cameras in Bayer format can be used directly. Furthermore, in images 16 and 20 we can appreciate that the subtraction is better in Bayer format since the result conserves the form of the shrimp more faithfully which is of great importance as this shape will be used to obtain the orientation of the shrimp.

With this we can avoid the entire code in the implementation of the Bayer to RGB converter, and more significantly, we can save time in the conversion, which is a critical parameter in real time following of animal trajectories.

All the tests conducted on the implementations demonstrated the functionality of each one of them as well as their technological expectations.

From the results, we can see that the best results are obtained with the implementation in an FPGA, however, the complexity of programming could be a limiting factor and any change in the hardware might represent another difficulty. The recommendation therefore is to work with the mixture of technologies; in processes requiring greater speed, the FPGA can be used and in processes where this not a critical factor the DSP can be used, thereby making use of the good results obtained in this investigation. With this combination it would be possible to optimise the whole system.

References

1 - B. E. Bayer, 1976 Color imaging array, U.S. Patent, 3971065.
2 - . Builder, 2002 Borland C++ Builder 6, 2002.
3 - Camera Link; 2000 Specifications of the camera link, interface standard for digital camera and frame grabbers, Camera Link, October 2000.
4 - E. Cole, 2002 “Performance of LVDS with Different Cables”, SLLA053B, Application report, Texas Instruments, February 2002
5 - R. Gonzales, R. Woods, 2002 “Digital Image Processing”, Second Edition, Prentice Hall,
6 - C. Lujan, F. Mora, J. Martinez, 2007 “Comparative analysis between the Stratix II (Altera) and Virtex 4 (Xilinx) for implementing a LVDS bus receiver”, Electrical and Electronics Engineering, 2007. ICEEE 2007. 4th International Conference on 5 7 Sept. 2007, Page(s):373- 376.
7 - National Semiconductor, 2006 Channel Link Design Guide, National semiconductor, June 2006
8 - T. Sakamoto, C. Nakanishi, T. Hase, 1998 “Software Pixel Interpolation for Digital Still Camaras Suitable for a 32-bit MCU”, IEEE Transactions on Consumer Electronics, 44 4 November.
9 - N. Sawyer, 2008 “1:7 Deserialization in Spartan-3E/3ª FPGAs at Speeds Up to 666 Mbps”, XAPP485 (1 2), Application notes, Xilinx, 27 May 2008.
10 - Texas Instruments, 2002 “Interface Circuit for TIA/EIA-644 (LVDS)”, SLLA038B, Application notes, Texas Instruments, September 2002.
11 - Texas Instruments, 2008 “Fixed-Point Digital Signal Processor TMS320C6416”, SPRS226K, Data Sheet, Texas Instruments, January 2008.
12 - Xilinx, 2002 MicroBlaze Product Brief, USA, 2002.
13 - Xilinx, 2006 ML 401/ ML 402/ ML 403 Evaluation Platform, User Guide, Xilinx, May 2006.