Open access peer-reviewed chapter

Machine Vision Systems – A Tool for Automatic Color Analysis in Agriculture

By Ernesto Martínez Sandoval, Miguel Enrique Martínez Rosas, Jesús Raúl Martínez Sandoval, Manuel Moises Miranda Velasco and Humberto Cervantes De Ávila

Submitted: June 12th 2017Reviewed: October 25th 2017Published: December 20th 2017

DOI: 10.5772/intechopen.71935

Downloaded: 1356


It was in the early 1960s when machine vision systems initiated researchers and developers have worked on building machines that perform tasks of acquisition, processing, and analysis of images in a wide range of applications for different areas. Currently, along with the new technological advances in electronics, computer systems, image processing, pattern recognition, and mechatronics, it has arose the opportunity to improve machine vision systems development with affordable implementations at lower cost. A machine vision system is the combination of several high-tech techniques, including both hardware and software, used to acquire, process, and analyze images on a machine, which contributes with a set of tools for the extraction of features, such as color and dimension parameters, texture, chemical components, disease detection, freshness, assessment, modeling, and control, among others. Based on former paragraphs, we could say that machine vision systems are appropriate to improve the actual agricultural systems making them more useful, efficient, practical, and reliable.


  • image processing
  • computer vision
  • color analysis
  • habanero chili
  • vineyards

1. Introduction

The evaluation of what surrounds us is done through light, colors, shapes, textures, and intensities, among other characteristics, which originate from different natural phenomena that give rise to spectacular scenes that are visually striking to observe. The way to perceive such an enormous amount of information is achieved through the senses. The perception of the human being is an incredible skill composed of five senses: sight, ear, smell, taste, and touch. The human senses are complex systems within the human body, which in turn are formed by an immense number of well-adapted and calibrated sensors with a very wide range of operation to carry out specific tasks, in addition to adapting and interacting with each other. Some of the senses manage to partially or totally regenerate from time to time. The human being has another essential organ, the brain, which is the main processing unit, responsible for receiving, processing all information, making decisions, and coordinating actions in a synchronized, a fast, and an efficient manner. That is why trying to emulate these capabilities is a huge challenge.

The visual perception of the human being is basically composed by the interaction between the eyes and brain. In general, the human vision system can be described as follows: the eyes are composed of the eyeball and the muscles that control its position. The cornea and lens focus light rays at the back of the eye. The lens regulates the focus for near and far objects as they become more or less globular. On the other hand, the brain is much more complex. Even to this day, there are many unknowns about its operation and even its biological purpose. Computer vision systems (CVS) intend to emulate this sophisticated and dynamic vision system whose operation is very natural and transparent [1].

Computer vision can be understood as the science that develops theory and algorithms to extract useful information about an object or scene within an image, for further analysis. Computer vision systems follow a basic structure that can be divided into three stages illustrated in Figure 1: low, intermediate, and high level of processing. In the first stage, the appropriate image acquisition sensors must be chosen for the type of scene to be captured, as well as to establish the acquisition conditions that best fit the experiment in order to obtain clear and good quality images. For the second stage, we must work on segmentation, which is considered a fundamental process in the image processing. The segmentation process separates useful information from the rest of the scene; for this, there are different segmenting and filtering techniques whose objective is to define and extract desired features within the image. The third stage, of high level performs an analysis with the information resulting from the two previous phases, so that its performance is closely linked. This analysis is focused on the recognition of objects within the scene and interprets their attributes to give results or make decisions. This is based on the information extracted and a knowledge database that serves to store the information of the whole process, as well as to train the vision system. Image processing and image analysis are the core of computer vision systems.

Figure 1.

Stages of the basic structure in which image processing is divided.

Both human beings and systems of vision rely on light to appreciate their surroundings. Therefore, the sources of illumination as well as its quality are crucial parameters for the proper functioning of both biological and artificial systems. Roughly speaking, light is the portion of the electromagnetic spectrum that can be perceived by the human eye (400–800 nm). The reflections of different wavelengths in materials are essentially what we call the color of an object. Depending on the physical characteristics of the materials, light incident on an object can give rise to three phenomena: absorption, when light on the object converts that energy into heat; refraction, when light passes through the material by changing its direction as a function of the refractive index of the material; and reflection, when the material rejects certain wavelengths. The reflected waves are those that can be observed with the naked eye and give rise to the familiar range of colors.

Color is the result of the interaction of the light reflected by the objects upon contact with the cells of the cones of the human vision system. There are different kinds of cones in the human eye, so the level of absorption and the interpretation of the nervous system of these signals give rise to the perception of color. So, it can be said that color is a perception of light reflected by the surface of an object [2].

In computer vision systems, the color attribute is very valuable and is very often used to identify useful information and perform its extraction. It should be mentioned that one of the main factors in defining the quality of products in agro-industry, particularly fruit and vegetables, is their appearance and, therefore, the value of these products in the market. Appearance is a combination of color, shape, size, and texture, among others. As a result, color makes the product very attractive to consumers and helps them buy the product.


2. Machine vision systems

From a general perspective, machine vision systems (MVS) or computer vision systems (CVS) try to emulate human vision in order to gather information from an object without requiring a physical interaction with it [3]. Machine vision is a complex high-tech system which includes an image sensor (usually a charge-coupled device, known as CCD), a frame grabber and a computer with the appropriate software and algorithms [4]. In general, the procedure is as follows: a scene is captured by the image sensor, the analog electrical signal obtained from the sensor is converted to a digital format and sent to a computer, and then it is subsequently processed using algorithms in order to analyze the corresponding image. This procedure is frequently implemented using different setups to accomplish specific application requirements. The importance of this technology for the analysis of agricultural products lies in its nondestructive evaluation characteristic.

Among the several definitions for CVS offered by different authors, Timmermans says that CVS are states that include capture, processing, and analysis of 2D images [5]. Sonka mentions that the objective is to replicate human vision by perceiving and understanding an image electronically [6]. Jha holds that the perception of an object and their optical characteristics in order to perform an interpretation of the results is called vision system [7]. Also, it is defined as a system for automatic acquisition and analysis of images to obtain desired data for interpreting or controlling an activity. The system consists of image acquisition, image processing, and interpretation [8].

CVS are suitable for agricultural applications because when used to obtain the characteristics of fruits and vegetables the task is done quickly, economically, hygienically, consistently, and objectively. This is the reason why the use of CVS has been expanded in many sectors of the industry, such as medical, automation, surveillance, remote sensing, autonomous vehicles, and robot vision, among others [912].

There is a broad range of applications for CVS, ranging from routine inspection through complex vision systems for robots, which shows the flexibility of this technology. On the other hand, its implementation requires a relatively low cost, which makes this technology even more attractive for other applications.

In the case of the food industry, CVS have proved to be an alternative method for inspection of visual attributes in pastries [13, 14], meat [1517], and fish [18, 19]. Because this inspection method is nondestructive, it is widely used in agricultural applications, including inspection and selection of fruits and vegetables [20]. For the agro-industry, the visual aspect of its products is particularly important, since this parameter is a determining factor of the value of the product in the market.

Traditionally, personnel trained for this task carry out the inspection and selection of agricultural products manually. This implies several disadvantages, such as inconsistencies in selection, time consumption, intensive tasks, variability, and subjectivity. In addition, the manual process is tedious, laborious, costly, and strongly dependent on external factors.

On the other hand, the appearance of agricultural products, i.e., their size, shape, and color, and the presence of stains or shocks, have a negative influence in the consumer perception and therefore determine the degree of acceptance prior to a purchase. The consumer also associates a certain internal quality with external characteristics (the appearance), which affects future decisions in the purchases [21].

Quality is a key factor that Kramer defines as:

“Quality of foods may be defined as the composite of those characteristics that differentiate individual units of a product, and have significance in determining the degree of acceptability of that unit to the user” [22].

In reference to fruits and vegetables, these attributes can be classified as follows:

  • Color and appearance are the factors that normally determine whether the product is accepted or not.

  • Flavor (taste and aroma): once the product’s appearance convinces the consumer and the product is tested, the flavors and aromas become more important. Freshness, itching, and sweetness are some of the attributes that can be detected when consuming certain products.

  • Texture can be perceived externally not only when taking the product in our hands but also when you taste it, you have a clear impression of the softness, firmness, or crisp of the fruit or vegetable.

  • Nutritional value is a factor that is usually hidden, but it affects our organism in ways we cannot perceive. However, it is an increasingly important parameter for consumers, scientists, and medical personnel.

These factors are closely related to the process and maturity stage of both the plant and the fruit at the time of harvest, as well as postharvest management conditions.

Shewfelt suggests “A primary disadvantage of instrumental testing is that many instrumental measurements have little relevance to consumer acceptability and thus should never be used to define quality attributes for a specific product. In other words it is better to measure what is really important than to believe something is important because you measure it really well” [23].

Machine vision systems present a viable alternative to extract certain important attributes automatically and, moreover, analyze them for making decisions. Besides, they have been diversified in distinct areas that concern agriculture. Due to its wide range of use, a summarized categorization is presented in Table 1, considering the products that are evaluated, the type of application, and the technology that was employed.

RicePanicle length measuringDual-camera devices[24]
SoybeanDetection of insect-damaged soybeanHyperspectral transmittance image[25]
CauliflowerExtract structural parameters to assess the growthRGB-D[27]
Cherry tomatoesDetection of cuticle defectsHyperspectral fluorescence imagery[28]
Date fruitsDetermining viscoelastic characteristics of date fruitsCCD[29]
Dried figsGradingCCD[30]
FoliarDisease spotsCCD[31]
OrangeOcclusion recoveryCCD[32]
Rapeseed varietiesClassificationCCD scanner[33]

Table 1.

Machine vision systems in agriculture.

CCD: charge-coupled device; RGB-D: color depth camera.

The extraction of attributes of agricultural products presents important challenges due in part to the variety of products that exist. The extraction of useful data is highly dependent on a good segmentation. There are several techniques for segmenting images, three of which are illustrated in Figure 2. The first technique is thresholding, which is a fast technique that converts gray levels into a binary representation. The second technique is edge-based technique, which looks for change points that give way to borders or contours, and, finally, region-based technique, which performs a search around a group of pixels within the image.

Figure 2.

Typical segmentation techniques: (a) thresholding, (b) edge-based, and (c) region-based.

To illustrate the performance of the different segmentation techniques described above, a desktop setup was implemented as shown in Figure 4. The first step is to capture an image; clothing buttons were used in this experiment, which also include a printed circular reference. The second step was to implement routines in MATLAB that process the image with the different segmentation techniques resulting binary images. The performance of the three techniques was satisfactory, but the processing time differed. Thresholding was the fastest, then edge-based and region-based. The resulting binary images are illustrated in Figure 3.

Figure 3.

Applications of segmentation techniques: (a) original image, (b) thresholding, (c) edge-based, and (d) region-based.


3. Hardware and software requirements

In order to acquire all the useful characteristics of the image, it is required to use a suitable light source to obtain the type and quality of illumination required over the scene. Taking into account that the image sensor will receive the light reflected from the scene, the upper level of illumination will be limited by the sensitivity of the sensor. The image sensor resolution must be taken into account, since it generates the amount of pixels from the image, where light information is contained, as it was perceived. On the other hand, the speed of frame grabber is selected considering the shortest period time of the task to be captured; for example, to acquire images on real-time or online applications, a faster frame grabber is required. Once the image is acquired, it is ready to be processed and analyzed; the algorithm used to perform the task is designed to extract specific parameters from the image and then to show them as results in the proper form.

The automatic operation of the setup involves to take into account timing considerations, to assembly high-tech components, to write proper and efficient program codes, and, finally, to synchronize and integrate every component in a single functional system. First of all, the scene capturing place, the proper illumination, and the required insulation conditions to acquire useful images must be selected. The final application of the machine vision system must be defined in order to get a suitable image sensor and a frame grabber that accomplishes the resolution and speed required. Finally, it is necessary to write program codes for three different stages on a machine vision system: to acquire, to process, and to analyze the images. Acquisition refers to the task where the image is taken. During the processing stage, the features are extracted, and the image is enhanced. Finally, once that parameters and features of the image are found, the algorithms of the analyzing stage use them to get a result in context to the goal of the application [34].


4. Color analysis

Our eyes are able to detect what we call visible light, which in fact is a range of electromagnetic spectrum wavelengths between 400 and 800 nanometers. What we observe a scene is actually the light reflected from the surfaces or the radiation emitted by light sources that make us to experience the sensation of color.

In reference to fruits and vegetables, color is derived from the natural pigments in fruits and vegetables, many of which change as the plant proceeds through maturation and ripening. The primary pigments imparting color quality are the fat-soluble chlorophylls (green) and carotenoids (yellow, orange, and red) and the water-soluble anthocyanins (red, blue), flavonoids (yellow), and betalains (red). In addition, enzymatic and nonenzymatic browning reactions may result in the formation of water-soluble brown-, gray-, and black-colored pigments. Color is one of the most important object measurements for image understanding and object description [35]. Using color attributes is more discriminant than simply using grayscale since two image points with the same grayscale value can be differentiated from their color attributes [36].

The level and quality of illumination from a light source affect the performance of the human eye, just as the performance of computer vision systems is affected by the illumination sources used. Sarkar found that by adjusting the illumination, the appearance of an object can be radically altered [37]. Therefore, the lighting system can greatly influence the quality of the images, so it plays an important role in the efficiency and accuracy of the CVS. Gunasekaran noted that a well-designed lighting system can help improve the performance of image analysis by improving contrast [38].

Some of the technologies used to acquire images in the food production sector are charge-coupled device (CCD) camera, magnetic resonance imaging (MRI), ultrasound, X-ray, near-infrared spectroscopy, and computed tomography (CT), among others. From the technologies mentioned, the CCD camera is widely used in computer vision systems for quality assessment, product selection, and product classification. CCD technology has the ability to convert captured light into electrical signals to create images. Depending on the characteristics of both the sensor and the optics used to acquire the images, these can be obtained with high resolution, low noise, and good light-sensitive (ISO) level. These characteristics can be adjusted up to certain limits to adapt the images to different capture scenarios. Both color and monochromatic cameras have been used in the food industry for a wide variety of applications [3942]. There are several attributes used to evaluate the quality of products in the food industry. There are external attributes, such as shape, size, color, texture, and defects, which can be captured by CCD-based vision systems. On the other hand, among the internal attributes of food products that may be measured, we can mention internal structures, quantity of water, and gaps. The extraction of these attributes requires the use of technologies such as ultrasound, MRI, and CT in order to obtain images with sufficient information for processing. From the point of view of image processing, both external and internal attributes present important challenges to adequately identify the information of interest. Moreover, each technology has different qualities, and its usefulness will depend on both the type of information to be extracted and the type of application. In addition, it is advisable to keep the economic factor in mind before deciding the type of technology to use.

The appropriate selection of the color space contributes to enhance color attributes from processed images. Table 2 presents a simple classification of agro-products. The table consists of six categories: the first column helps to classify the products in fruits or vegetables; the second column presents the study applications; the third column has the technology or technique used; the fourth column has the percentage of efficiency reported; and the last column contains the references of each of the submitted works. In fruits, it is clearly noted that apples are a quite popular product and have been evaluated using different color spaces with an accuracy greater than 90%, as well as in banana and peach fruits. The applications range from sorting fruits by color and size, maturity discrimination, and color classification, to vegetables focusing on blemish detection and color measurements. The RGB, HSI, HSV, and L*a*b* color spaces are used depending on applications.

CategoryProductsApplicationsTechnologyAccuracy (%)Reference
FruitAppleSorting by color and sizeHSI90%[43]
AppleMature discriminationRGB95.83%[44]
StarfruitsColor classificationHSI100%[45]
BananaColor measurementRGB, HSV, L*a*b97%[46]
PeachSorting by color and sizeHSI90%[47]
VegetablePotatoColor classificationRGB90%[48]
PotatoBlemish detectionRGB89.6%[49]
TomatoColor classificationRGBNo reported[50]

Table 2.

Analysis by color attributes.

HSI: color space (hue, saturation, intensity); HSV: color space (hue, saturation,value); RGB: color space (red, green, blue); CIELAB: L*a*b* color space.


5. Applications of machine vision systems in agriculture

5.1. Machine vision system implementation

To acquire and analyze the information from an image, it is necessary to implement a machine vision system. The simplest case involves the implementation in a desktop setup. The primary goal of this configuration is to provide the insulation conditions to acquire a clean, high-contrast image. Additionally, in this implementation it is easy to add a number of sensors to monitor inside ambient conditions, in order to provide complementary data for the analysis. These characteristics are quite useful for the agricultural sector to characterize several species as well as a variety of crops under different conditions [51, 52].

A summary of technologies used in a wide range of implementations in the agriculture area is presented in Table 3. Products, applications, technologies (used as a machine vision system), the accuracy of this system, and the corresponding references are included.

ProductsApplicationsTechnologyAccuracy (%)Reference
OilseedsMeasuring thermal propertiesMRNo reported[53]
Passion fruit juiceTracking thermal degradationMRNo reported[54]
PearGrading by external shapeFeatures88.2%[55]
StrawberryBruise detectionH-CVS100%[56]
AppleQuality gradingM-CVS93.5%[57]
AppleChilling injuryH-CVS98.4%[58]
CitrusRottenness detectionH-CVS98%[59]
MangoMango gradingFD89.83%[60]
MushroomEnzymatic browningH-CVS>89%[61]
CitrusDefect detectionT-CVS98.9%[62]
Agricultural productsInternal characterizationX-rayNo reported[63]
Agricultural productsQuarantine scannerX-ray/CCDNo reported[64]

Table 3.

Technologies applied to agriculture.

T-CVS: traditional computer vision system; H-CVS: hyperspectral computer vision system; M-CVS: multispectral computer vision system; R: Magnetic resonance.

Based on an extensive bibliographical research, the advantages and disadvantages of the artificial vision systems applied in agriculture are indicated in Table 4. It is important to mention that the nondestructive characteristic becomes one of its main advantages, since the industry wants to sell the final product, with the best quality to the consumer and in that way increase their profits.

ConsistentAdaptable under specific conditions
DatabaseArtificial lighting
EfficientEnvironment control
FastObject identification
Non expensive

Table 4.

Advantages and disadvantages of computer vision system.

5.2. Case study

5.2.1. Habanero chili color assessment

Setups for 2D vision systems are suitable for capturing images of fruits and vegetables. For this case of study, the setup required a camera hold steady at the center top for capturing the images. It was used a commercial digital single-lens reflex (DSLR) camera as capture device.

As shown in Figure 4, the capture device includes a housing that allows to isolate the experiment from external contaminants such as light and dust. In addition, a control unit is installed to trigger the camera remotely in order to avoid handling vibrations and to ensure that the captures are taken as uniform as possible between samples. In the same way, the illumination lamp is turned on few seconds before to capture the image, which allows to stabilize the lamp and in this way to achieve a more uniform illumination.

Figure 4.

Typical machine vision system component.

To verify the proper performance of the capturing device, it is advisable to include a wireless communication to the control unit in order to perform a remote monitoring. On the other hand, the illumination lamp spreads the light using the white surface of the interior of the case, allowing a better distribution of light, reducing both shadows and reflections in the samples. Immediately, after each image capture, the lamp is turned off to avoid possible damages due to over exposure to the fruits. In addition, a high-contrast background has to be used and thus facilitate a correct segmentation.

To ensure that the conditions and positions of the samples do not change during the capture of the images that are same, each individual sample can be separated using specialized routines, so that they can be processed for further analysis. In order to extract the color attribute automatically from the images, a routine can be implemented to carry out a threshold segmentation and thus to obtain only the information of the region of interest (ROI). Once the information of the camera in RGB format is obtained, a color space conversion to CIELAB was made to show the color changes as a function of time.

Images were taken with a sampling frequency of one image per hour, during 8 days for 35 habanero chili specimens. This generates a total of 6720 sub-images from the 192 images taken. The type of images captured from this system can be seen in Figure 5.

Figure 5.

Type of images captured: (a) samples at the first day (hour 1) and (b) samples at the eighth day (hour 192).

The foundation for development of image processing algorithms was to keep them simple and with a proprietary processing instead of using toolbox functions of MATLAB® for image processing, except for the fundamental ones to loading and displaying images. The first step was to test whether the algorithms were able to capture, process, and analyze the color images of habanero chili. Since these algorithms are part of a larger project that considers migrating these codes to an embedded system, to provide greater versatility and portability to the vision system, expanding its range of applications. Algorithm starts loading an image, from color image database of habanero chili specimens from the previous acquisition stage. Then, a predetermined area segmentation delimits a specific region for each specimen, called sub-image. Next, a color segmentation is performed, using a threshold technique, where the background color parameters are set as the threshold value, so everything else is the color information from the specimens. Then, a color space conversion from RGB to CIELAB is carried out. After that calculations of color average from each specimen are performed and color information in CIELAB components L*a*b*, Cab*, and hab° is generated, this stage results in a matrix with these color parameters. Finally, with this color data, a color analysis is executed, where these parameters can be plotted and observed, as shown in Figure 6.

Figure 6.

Flowchart of image processing algorithm.

In Figure 7, a color diagram is presented with the parameters of a* vs. b* from CIELAB and represents the mean values from each specimen during 8 days of experiment. Each point signifies the average from 24 images taken per day, corresponding to sampling rate. It can be seen how the system is able to detect color changes.

Figure 7.

Color diagram of a* vs. b* of all samples from their mean values per day.

Color information obtained from image processing and expressed in terms CIELAB coordinates was statistically analyzed applying a one-way analysis of variance (ANOVA). The analysis of variance shows highly significant differences between specimens. The obtained results are reported in Table 5 (sum of square [SS], degrees of freedom [DF], mean square [MS], and test statistic [F]). Moreover, Figure 8 represents the analysis of variance between specimens.

SourceSSDFMSF > F0.05
Days22643.273234.7410.97 ≫ 2.010

Table 5.

One-way analysis of variance (ANOVA) between samples of α = 0.05 significance.

Figure 8.

Analysis of variance results between samples, box, and whisker plot.

Another statistical analysis was performed using one-way analysis of variance (ANOVA). The analysis of variance shows highly significant differences between days. The obtained results are reported in Table 6. Moreover, Figure 9 represents the analysis of variance between days.

SourceSSDFMSF > F0.05
Specimens56449.8341660.298.77 ≫ 1.424

Table 6.

One-way analysis of variance (ANOVA) between samples of α = 0.05 significance.

Figure 9.

Analysis of variance results between days, box, and whisker plot.

Highlights: a color image database of habanero chilies at controlled conditions was generated for future developments and analysis. At the moment, developments in this specific kind of application explore noninvasive techniques for color assessment in habanero chili specimens; the information related was null, which increase the scientific value of these results. Algorithms developed are capable to acquire process and analyze habanero chili specimens in postharvest under controlled conditions. The guideline for algorithm development was to write code with basic functions and proprietary code, instead of toolbox functions of MATLAB®, considering code migration for future research and developments. The statistical analysis confirms that image processing algorithms and machine vision system are capable to detect, quantify, and analyze color changes of digital images from habanero chili in an automatic way.

5.2.2. Vineyards

The wine industry is one of the most interesting in the use of vision systems to increase the quality of its crops [65]. An increasing number of wine producers recognize the advantages of understanding the biophysical characteristics and the performance of their vineyards, leading to better management of their resources and making decisions. Wine producers commonly have a goal for the state of maturity they want to achieve for the wine they produce. Such a goal can vary, even within the same grape variety, depending on the type or style of wine that will be made. As previously mentioned, for optimum determination of fruit ripeness, color is one of the most important parameters. In the case of vineyards, a computer vision system can be used to individually locate grapes in clusters as well as grape berries that are used in evaluation samplings. By using segmentation algorithms, sub-images of each grape are used to extract the information of the color parameters in a proper color space.

As mentioned by Martinez-Sandoval et al. [66] as well as Rabatel and Guizard [67], digital images of individual grapes could easily be acquired in the field as long as adequate control of the distance and light conditions over the samples is maintained. The use of computer vision systems would allow estimating the volume of grapes from their visible area, and using suitable image processing algorithms, it would be possible to obtain detailed information such as the color and size of the berries.

The development of methods based on image processing has advanced to the point of being able to automatically detect and count the grapes to accurately predict the yield of a harvest. As indicated by Nuske et al. [68], in order to capture the images of vines (in the visible light spectrum), conventional cameras are used throughout the vineyard. A typical algorithm to perform the image processing of vines can be divided into the following stages [66]:

  1. Detecting potential berry locations with a radial symmetry transform.

  2. Identifying the potential locations that have similar appearance to grape berries.

  3. Group neighboring berries into clusters

A vine image consists of a set of berries with nearly identical color properties, so that the information available for berry separation mainly relies on their contours (as seen in Figure 10) where a bunch of grapes are segmented from the background. Due to the smooth 3D shape of the berries, their contours are accessible as luminance discontinuities in the image. However, the algorithm only works with grayscale images. The main goal is to make a nondestructive estimation of the bunch weight of grapes at an early stage by computer vision using an elliptical model suitable to estimate the volume of the grape from its visible area.

Figure 10.

Image segmentation of bunch of grapes.

Winemakers commonly have a goal for grape maturity that they would like to achieve for the wine they will produce. This objective can vary, even within the same grape variety, depending on the type or style of wine desired [+]. For the characterization of maturation, physicochemical analyses such as sugar content, acid, and pH [69] are performed, sugar content being one of the most important factors [70]. Figure 11 shows a generalized graphical representation of changes in grape composition during development and ripening.

Figure 11.

Grape berry development and ripening.

The sugar content of the grapes is commonly controlled by periodically measuring the content of soluble solids in the ripening berries with a refract meter. It works by measuring refracted light through grape juice in a prism. The sugar level is generally expressed in Brix grade, which represents grams of sugar per 100 grams of juice; the thicker juice shows a higher Brix degree on the scale.

The procedure followed by Murillo-Bracamontes et al. [3] in order to extract color parameters from berry image and to compare them against Brix degree (measured with a refract meter) as well as to their pH value was as follows: the acquired images were converted to grayscale and equalized using a contrast-limited adaptive histogram equation (CLAHE) to improve the contrast of grayscale images (see Figure 12) [71, 72]. A Hough transformation was then applied to locate the grapes (as shown in Figure 13). After completion of the transformation, three vectors corresponding to (x, y) and radius parameters of the berries were obtained; then, the individual berries detected were segmented from the background, stored in sub-pictures, and then converted to the CIELAB color space for analysis. Figure 14 shows a flowchart of the steps performed for this analysis.

Figure 12.

Preprocessing of the image before applying the Hough transform. (a) Original image in grayscale. (b) Results of applying CLAHE equalization.

Figure 13.

Hough transform. (a) Original image in RGB color space. (b) Results of the detected circles (remarked in yellow).

Figure 14.

Algorithm applied to grape images.

Analyzed data yield the same evolution described in Figure 11, which is a good indicator that the procedure provides promising values.


6. Conclusions

Computer vision systems are a viable option for handling agricultural products. The image processing and image analysis tools are well suited for application developments with automatic extraction of external and internal attributes of agro-products. One of the trends is the development of vision applications for in situ use in fields or greenhouses. The properties of being noninvasive and nondestructive make these systems an attractive option for application developments in different areas. However, agro-industry is particularly interesting due to the nature of their products and the standards of quality requirements.



The authors would like to thank the National Council of Science and Technology (CONACYT) for providing the fundings to carry out this research, as well as to the people involved on this research at the Autonomous University of Baja California, Program for Strengthening Educational Quality (PFCE) and the Center for Scientific Research and Higher Education at Ensenada and the State University of Sonora.

© 2017 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution 3.0 License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

How to cite and reference

Link to this chapter Copy to clipboard

Cite this chapter Copy to clipboard

Ernesto Martínez Sandoval, Miguel Enrique Martínez Rosas, Jesús Raúl Martínez Sandoval, Manuel Moises Miranda Velasco and Humberto Cervantes De Ávila (December 20th 2017). Machine Vision Systems – A Tool for Automatic Color Analysis in Agriculture, Automation in Agriculture - Securing Food Supplies for Future Generations, Stephan Hussmann, IntechOpen, DOI: 10.5772/intechopen.71935. Available from:

chapter statistics

1356total chapter downloads

More statistics for editors and authors

Login to your personal dashboard for more detailed statistics on your publications.

Access personal reporting

Related Content

This Book

Next chapter

Automation of Integrated System for Grain Beverages Processing

By Gana Ibrahim Mohammed, Ephraim Micheal and Agidi Gbabo

Related Book

First chapter

Real-Time Sensory Information for Remote Supervision of Autonomous Agricultural Machines

By Daniel D. Mann, Uduak Edet, Mitchell Green, Olayinka Folorunsho, Avery Simundsson and Franklin Ogidi

We are IntechOpen, the world's leading publisher of Open Access books. Built by scientists, for scientists. Our readership spans scientists, professors, researchers, librarians, and students, as well as business professionals. We share our knowledge and peer-reveiwed research papers with libraries, scientific and engineering societies, and also work with corporate R&D departments and government entities.

More About Us