© 2024 The authors. This article is published by IIETA and is licensed under the CC BY 4.0 license (http://creativecommons.org/licenses/by/4.0/).
OPEN ACCESS
For learning about things like strength, level, contrast, range, limits, lost and unseen parts or borders, etc. in the digital world, picture enhancement and edge detection are essential. Identical few computations and approaches exist to show a goal's superior specifics. This work focuses on an edge-finding technique called the Side Searching Method (SSM), which is a simple way of identifying hidden boundaries in images or faces in an unconstrained environment. The other objective of this paper is improvement or enhancement of the object. To achieve this, the darkest portion of the image has been enhanced using the Object Improving Method (OIM) in an unconstrained face recognition environment. The proposed methodologies focus on interactive image exploration. In order to investigate the issue and evaluate recognition ability, the photographs of faces are divided into several characteristics using state-of-the-art deep networks. Numerous promising outcomes were seen, and their studies have the potential to advance Deep Learning techniques toward high accuracy and practical application to tackle the difficult challenge of unconstrained face recognition. The OIM approach enhances the instruction of the Gamma, Second Gamma, Gain, and Cutoff parameters, which helps to improve the quality of pictures. Variations in gamma values result in distinct visual effects. While the SSM approach is useful for more accurately studying the image side boundaries, a lower gamma value accentuates details in certain intensity ranges, while a higher gamma value emphasizes contrast.
face recognition, Side Searching Method (SSM), Object Improving Method (OIM), edge detection (searching), enhancement (improvement)
In proposed methodology work focuses on an edge-finding technique called the SSM, which is a simple way for identifying hidden boundaries in images. The other upcoming improvement method locates its route to depiction and improves the darkest observation of an aim. Another aim is, the darkest portion of the image has been enhanced using the OIM in Unconstrained face environment.
The stochastic and deterministic representations of images during image processing improve the quality of images by detaching presentation. The author presents a categorized and compared analysis of edge detection techniques [1]. In the restoration process, we genuinely try to reconstruct an image as closely as we can to the original. Lack of differentiation, random sound, disruption, geometric disturbances, etc. is examples of destruction.
The technique of running a two-dimensional image through an electronic computer is known as electronic image processing. It suggests running two-dimensional information electronically. A small number of intricate or real-world objects are represented by a range of electronic images [2, 3]. In this introduction section, some crucial ideas for image processing that were utilized in this manuscript are discussed.
Due to various application types in a dynamic environment, the unconstrained recognition problem has grown in importance. Due to the diverse nature of applications in a dynamic environment, the unconstrained reputation problem has become more prevalent. The studies have been driven in an unconstrained environment by forensic technology, video surveillance, and computer vision. Face reputation in controlled environments is a mature technological approach, according to various algorithms that have nearly error-free recognition [4]. Within the managed reputation techniques, managed face reputation has reached a benchmark. In addition, it finds its most common application in biometric systems, which rely on people willingly striking an identifying position. Facial recognition is difficult to implement in video surveillance because subjects may not always appear willingly, increasing the likelihood of false positives. Uncontrolled face reputation refers to applications below out of control situations (such as varying illumination, expression, and pose, etc.). It is a more realistic approach because it makes it easier to identify faces in actual, diverse global environments.
1.1 Edge searching
Edge searching is a crucial and prominent component of face recognition in image processing. A picture's edge explains the separation between a subject and its own history. An advantage can be identified by a significant change in the value of the picture depth goal. Consequently, a benefit divides two halves into several strengths [5, 6]. Edge judgment is sometimes a quick way to acquire information from photos before moving on to purposeful elimination and topic separation in image-processing software. This method captures the borders between foreground and background elements in a photograph if there are sharp discontinuities in the illumination. The visual illustration of these boundaries is the benefit plan [7-10]. In order to express useful pixels, it's critical to know how far apart two pixels are. Some of the often-used benefit detection workers include Slope, Sobel, Prewitt, Log, Canny Operators, etc.
1.2 Image optimization
Image enhancement techniques continue to be preferred in many uses where the excellence of the image is significant [11, 12]. Change may serve as the governing mechanism for displaying and storing discovered data, which makes it an undeniably significant feature in practically any separate research on image excellence.
The primary argument for image enhancement may be that the displayed image has obvious qualities that the original image lacked. It has been suggested to use a variety of methods, such as power transformation, to improve photos that have undergone histogram equalization filtering and irregular illumination transformation. Lowering the dynamic range and/or increasing the difference can commonly improve an understanding image [13-16].
1.3 Unconstrained face recognition
According to study [17], DL has lately become the industry standard in face recognition and a number of other AI-related fields. They question, "Can DL really resolve the issue of face recognition?" and, if not, "What is the problem for DL approaches in facial recognition?" They contend that one of the challenges for DL, particularly in unrestricted face recognition, is the issue of face image quality. They separate photos of faces into discrete traits to research the problem and gauge recognition performance using cutting-edge deep networks. There were a number of exciting results, and their researches can suggest advancing DL approaches toward high accuracy & useful application to handle the challenging problem of unconstrained face recognition.
The Id Quality of the unconstrained face in an unrestrained situation is indicated in this study [18]. The result is a matching picture face that has been recognized from an image series. Additionally, it provides the same approaches to expression alteration, lightness, and inequality [19]. It has elements like discrepancy, lighting, and expression-difference techniques. There are potential uses for face recognition and image retrieval. Currently, a variety of techniques are available for top-view facial recognition. In the last few years, a variety of methods for computer vision facial recognition have emerged. Face identification is difficult to do in real life, though. The growing interest in unrestricted advantageous face recognition is being fueled by the extensive use of internet media, which includes community-based strategies and video surveillance monitoring. Experience assessment is now more crucial than ever. According to this study, status is managed within the context of chart assumption. They are able to identify a mystical encounter using a variety of techniques. The study focuses on the alternatives put forth for the unrestricted face recognition quality region and makes reference to a method put forward by Reference for recognizing individuals based on their faces (from Gallery Image). RFG recognition and DCT position-sensitive hashing are combined for scalability and quick recoveries. This study aims to improve presentation effectiveness using an open-ended face recognition system Reference Face Based Approach. When recommending approaches, MATLAB will be utilized for simulation results [20-23].
2.1 Technique based on convolution
Since the outcome of convolution at each pixel is simply the multiplication of local pixels from the image's cause and pixels in a quantity [24, 25], convolution is a regional operation. Although, in actuality, the dumbbells will represent the filtration's opinions, the large number of pixels on the display will be the final pixel value. The kernel is the name for the screen [1].
$[u, v]=\sum_{l, m} b[l, m] h[u-l, v-m]$ (1)
where, $u$ and $v$ are presents variables.
2.2 Gradient operator
The idea for the gradient operator is based on the first or second derivative of the gray level. The first kind may indicate benefit points, with greater advantage points (large magnitudes) being provided by higher gray-level alterations [26]. Two wishes, one for each aspect of the benefit, are returned by the next type. The advantage of this is that, if a mark is strained between your two wishes, the point where they intersect the zero axis will be the advantage's midpoint, ideally allowing us to determine side position with sub-pixel accuracy. The possibility of zero-crossing occurring at a fractional pixel time is explained by sub-pixel precision. The classic benefit indicator uses first-changes to determine the picture's slope [27]. When the incline is higher than the limit, a product is present inside the image. The gradient of stage ($w$, $z)$ in relation to image $j(w, z)$ denotes as follows:
$\nabla j(w, z)=\left[\begin{array}{ll}G_w & G_z\end{array}\right]=\left|\begin{array}{ll}\frac{\partial j}{\partial w} & \frac{\partial j}{\partial z}\end{array}\right|$ (2)
The vector's weight is:
$\nabla j=\operatorname{mag}(\nabla j)=\left[\begin{array}{ll}G_w^2 & G_z^2\end{array}\right]^{1 / 2}$ (3)
And its course as:
$\varphi(w, z)=\arctan \left(G_z / G_w\right)$ (4)
where, Gw and Gz are the slopes in the $w$ and $z$ paths, respectively. The three equations mentioned above are used to determine the slope of each pixel in the image. In fact, the image can be approached by using a tiny region program convolution. Prewitt, Robert, and Sobel owner are examples of companies that are trending.
Laplacian operator uses the second type, and the owner is k known to be:
$\nabla^2 j(w, z)=\frac{\partial^2 j(w, z)}{\partial w^2}+\frac{\partial^2 j(w, z)}{\partial z^2}$ (5)
The Laplacian agent locates the favorable spots, screening bigger areas surrounding the pixel, on the other hand, deteriorating at sides and shapes. Moreover, when the gray phase asset goal varies, the Laplacian filtering fails to discover the alignment of benefit.
2.3 Filter, LOG
Gaussian Mixed Gaussian Selection Laplacian Using the Laplacian, it may be said that:
$G_\delta(w, z)=\frac{1}{2 \pi \delta^2} \exp \left(-\frac{w^2+z^2}{2 \delta^2}\right)$ (6)
When s and t are used in place of w and z in Eq. (7), after a Gaussian operator and j(s, t) are convolution to smooth the image, the following formula is used to determine your advantage [28]:
$\nabla^2\left[G_\delta(s, t) * j(s, t)\right]=\left[\nabla^2 G_\delta(s, t) * j(s, t)\right]$ (7)
Gaussian edge sensors reduce the sound by removing the image and are symmetrical throughout the benefit. The main owner, Canny, used Gaussian's type for Canny to convolve the image [1].
2.4 Canny operator
John Canny developed a method for detecting edges called Canny edge detection when I was a student at MIT in 1983 [29]. It's the gold standard, the most effective, and the most popular approach to edge detection. It filters out the background noise before attempting to recover the image's edges. When compared to other approaches, Canny is superior for edge extraction and yields desirable results. Several edge picture details are under the Canny operator's command, and noise is effectively suppressed.
The first-purchase type of Gaussian function can be used as the aspect indicator. The edge alert was developed to become a benefit indication that satisfies the following three criteria [27, 30, 31]:
A multi-step technique called canny edge detection can simultaneously locate edges while suppressing noise.
$f(\gamma, \delta)=F_\sigma(\gamma, \delta) * h(\gamma, \delta)$
where,
$F_\sigma(\gamma, \delta)=\frac{1}{\sqrt{2 \pi \sigma^2}} \exp \left(-\frac{\gamma^2+\delta^2}{2 \sigma^2}\right)$
$A(\gamma, \delta)=\sqrt{f_\gamma^2(\gamma, \delta)+f_\delta^2(\gamma, \delta)}$
And
$\theta(\gamma, \delta)=\tan ^{-1}\left[\frac{f_\delta(\gamma, \delta)}{f_\gamma(\gamma, \delta)}\right]$
$A_J(\gamma, \delta)=\left\{\begin{array}{c}A(\gamma, \delta) ; i f A(\gamma, \delta)>J \\ 0 ; \text { Otherwise }\end{array}\right.$
Here J is selected in such a way that all edge elements are preserved, while the majority of the noise is muted.
2.5 Operator of Prewitt
The side indication is unquestionably the right method for figuring out where and how an edge is oriented. The advantage acknowledgement immediately within the kernel obtains the placement by employing the ideal response. An edge incline vector is calculated at each level by the adjacent indication of the basis image. Enhancing a side image starts at the slop vector level.
$\left[\begin{array}{ccc}1 & 1 & 1 \\ 0 & 0 & 0 \\ -1 & -1 & -1\end{array}\right]$
2.6 Operator Sobel
Using the Sobel estimate toward the type, the Sobel method locates sides. This moves down the sides in these conditions, where the slope of J is at its best. The Sobel operator was the most popular benefit recognition operator prior to the introduction of approaches with an academic base.
$\left[\begin{array}{ccc}1 & 2 & 1 \\ 0 & 0 & 0 \\ -1 & -2 & -1\end{array}\right]$
2.7 Sensor for zero crossing
To detect zero crossing at each pixel ᴩ of the exhausted photo G(s, t), one approach would use 3×3 regions centered at ᴩ. At least two other pixels nearby should likewise change when g reaches zero. Four exam scenarios-two diagonal, one up/down, one left/right-will be included. The complete assessment of the arithmetic distinction must exceed the threshold if the assessment of g(w, z) has been associated with a boundary, in addition to the hallmark of other neighbors differing, before we may come in touch with ᴩ at a zero-crossing pixel [32].
It will discover standard comparison improvement methods. They fall into two categories: frequency domain method and particular locations [33, 34]. In a novel site technique, a picture's pixels are directly changed, and their values are adjusted based on the ethics of the surrounding pixels [11, 35]. A picture's Fourier transform is modified to offer frequency-domain methods.
3.1 Histogram equivalence
A graph known as a histogram only displays the percentage of a picture that has a certain power. By modifying the histogram of the image and redistributing all pixel values to as closely resemble the intended histogram as feasible, equalization is a method that yields a gray diagram. This method is useful for generating low-contrast images with an equal number of pixels in each result to produce a grayscale [11].
3.2 Changes to energy regulations
The end graphic of the energy regulatory adjustment is tied to its input picture [36].
$G(m, n)=b[E(m, n)]^¥$ (8)
They are continuous, b and ¥. The intensity value's level of rise is determined by the value of ¥ [37]. During an electric law transformation, every pixel in the face is elevated to a specific proponent cost. By choosing the suitable price proponent, one can increase brightness at either a high or low cost [20, 38].
3.3 Process of sigmoid
This is known as a non-linear function. The "G" shape of the function is where the term "sigmoid" originates. Arithmetical experts refer to this objective as the logistic function when S(m) is used for an input with 1, such as a gain [11]. So that, the function represented by:
$S(m)=\frac{1}{1+e^{-g(m)}}$ (9)
3.4 Gamma adjustment
Mark is a mathematical parameter used in image processing that represents the nonlinearity of power reproduction. Gamma correction is a pre-process for nonlinearity that runs an indicator from a level cost [39, 40].
Let Jo represent the output image, Ji represent the input image and $\gamma$ represents the numerical parameter which describes the non-linearity of intensity reproduction.
$J_0(m, n)=\left[J_i(m, n)\right]^{\frac{1}{\gamma}}$ (10)
It makes use of a modified operator and bases its acquisition of the subsequent matrix on an approximation.
Operator $\mathrm{A}=\left[\begin{array}{ccc}0 & 0.11111 & 0 \\ 0.11111 & -0.44444 & 0.11111 \\ 0 & 0.11111 & 0\end{array}\right]$
Both methods are novel and quick for object improvement and side-finding from faces. Aimed at side improvement and detection, these are quite helpful. In order to complete our assignment, we've developed the following calculations based on the above-mentioned projected providers:
4.1 Algorithm for side searching
In the SS method, images from a self-collected dataset are initially loaded with results [41]. A variety of well-established image processing techniques were utilized to ensure optimal feature extraction and noise reduction. Initially, the RGB images were converted to grayscale, effectively reducing color information while preserving structural details. This transition helped to simplify the following steps of analysis. The vertical Sobel operator, implemented as a 3×3 matrix, is then used to identify vertical axes in grayscale images. This operator effectively highlighted the variations in intensity along the vertical axis, thereby emphasizing the vertical edge characteristics of the images. To reduce noise interference and further improve the image quality, a median filter was consistently used. This method effectively refined the images by replacing the value of every pixel with the median value of its neighboring pixels, thereby reducing noise while preserving boundaries. The Canny edge detection method is then used, which is an essential step for identifying prominent edges within images.
These preprocessing methods produced a modified and enhanced vertical Sobel edge image. This image effectively emphasized the vertical edge features, successfully laying the foundation for later analysis and interpretation.
Convert the RGB image into a grayscale image
The technique of converting a color image to a grayscale one while preserving the essential details is intricate. It's possible that the grayscale version of the image will lack some of the color image's contrasts, clarity, shadow, and structure. The new technique uses reduction, RGB approximation, and the addition of chrominance and brightness to transform the color image into a grayscale image. The study's grayscale images created by the technique verify that the method maintains the color image's essential qualities, including its colors, sharpness, shadows, and overall design [42].
Since it first searches the edges from picture data and then executes various operations, it is also possible to refer to the SSM as an edge searching algorithm. This algorithm adheres to the many steps discussed below:
Q(Ѓ, £))=Convolution (A, J).
SSM mathematical representation
Suppose J=[Image]Ѓ×£×3
In this case, Ѓ and £ stand for rows and columns, respectively. And J represents the three-layer image.
Than the formula,
$A(\dot{\Gamma}, £)=J(\dot{\Gamma}, £, 1)$ (11)
$B(\dot{\Gamma}, £)=J(\dot{\Gamma}, £, 2)$ (12)
$C(\dot{\Gamma}, £)=J(\dot{\Gamma}, £, 3)$ (13)
D(Ѓ, £)=(A(Ѓ, £)+B(Ѓ, £)+C(Ѓ, £))/3
Then, implement histogram equalization,
$I(\dot{\Gamma}, £)=H_e(J(\dot{\Gamma}, £)$ (14)
He represents the equalization of the histogram.
$K(\dot{\Gamma}, £)=Z_c(I(\dot{\Gamma}, £))$ (15)
The zero crossing edge detection is represented here by Zc.
$L(\dot{\Gamma}, £)=C_e(J)$ (16)
Ce here is the canny edge detector.
G1(Ѓ, £)=P(Ѓ, £) OR T(Ѓ, £)
P= $\left[\begin{array}{ccc}0 & 0.11111 & 0 \\ 0.11111 & -0.44444 & 0.11111 \\ 0 & 0.11111 & 0\end{array}\right]$
Q(Ѓ, £)=Conv.(P, J)
U(Ѓ, £)=Ce(Q(Ѓ, £))
Then
G2(Ѓ, £)=U(Ѓ, £)+G1(Ѓ, £).
4.2 Algorithm for object improving
In OIM, the acquisition of images from a self-compiled dataset with results that depict the objects or scenes of interest is the first step in the object enhancement method [41]. After obtaining these images, the next stage is to prepare them for enhancement. This preparation begins by converting the original images, which are typically in RGB format with complete color, to grayscale images. Grayscale images have a single channel that represents various hues of gray based on the original image's light intensity. After converting each image to grayscale, the object enhancement method adjusts the contrast and luminance of each image using a power-law transformation. This transformation is crucial because it permits the fine-tuning of image characteristics. The following gamma values are used for this purpose: 10, 1, 1, and 0.5. These gamma values serve as parameters for adjusting the amount of contrast and luminance applied to each image. To implement the power-law transformation with a particular gamma value, say 10, the method first selects this value as the transformation's parameter. It then iterates through each pixel in the grayscale image, multiplying the intensity value of each pixel by the selected gamma value (in this case, 10). Adjusted Pixel Value=Original Pixel Value 10 is the formula for this adjustment. The outcome is a grayscale image with increased contrast and luminosity. Depending on the implementation, a normalization step may be performed to ensure that the adjusted pixel values lie within the valid intensity range, which for 8-bit images typically ranges from 0 to 255. This procedure is then repeated for each of the selected gamma values: 5, 1, and 0.5, resulting in a series of transformed images with distinct contrast and luminance levels. These enhanced images can be further analyzed or contrasted to determine the most effective enhancement for the specific objects in the dataset, enabling improved object recognition or analysis according to the goals of the current task. The following steps are described below:
Preprocessing in OIM method
The OIM is a series of image processing steps applied to an input image to enhance its visual quality and features.
Convert into RGB to GRAY
When an image is processed, it is first digitised so that various manipulations may be applied to it. This method is used to improve an image or get meaningful data from it. This study converts RGB into gray.
Grayscale Image
Grayscale refers to a colorless scale of gray tones. Black is the darkest color conceivable, while white is the lightest. White indicates complete transmission of light, whereas black indicates no light at all is being transmitted or reflected. Grays in between are shown by varying the intensity of the RGB primaries [43].
Apply power-law transformation to adjust the image's contrast and brightness.
Contrast improvement is a crucial factor to consider when performing image enhancement, whether in the spatial or Fourier domains. The techniques can be broken down even further in the spatial realm into subcategories like grey level transformations and histogram processing. As was previously indicated, histogram equalisation has the potential downside of reducing contrast. The pictures' histograms can be specified or matched manually, depending on the user's preferences. Similar amounts of user input are needed for power law transformations or piece-wise linear transformation functions. In the former, the exponent in the transformation function must be selected, while in the latter, the straight lines that make up the transformation function must be specified with respect to their slopes and ranges.
Common explanations for the power-law change include:
$s=c r^\gamma$
In this equation, c is a constant and s and r are the grayscale values of the pixels in the output and input pictures, respectively [44].
Different gamma values are used in this research which includes 10, 1, 1, and 0.5. The gamma value acts as a parameter that controls the degree of enhancement applied to the image. Different gamma values yield different visual effects. A higher gamma value emphasizes contrast, while a lower gamma value enhances details in specific intensity ranges. The equation used for gamma values are described below:
Let Jo represent the output image and Ji represent the input image.
$\mathrm{J}_{\mathrm{o}}(\mathrm{m}, \mathrm{n})=\left[\mathrm{J}_i(\mathrm{~m}, \mathrm{n})\right]^{\frac{1}{\gamma}}$ (17)
By applying these transformations with varying gamma values, the OIM attempts to find the right balance between contrast enhancement and detail preservation to enhance the overall appearance of the image or highlight specific features of interest. The specific choice of gamma values depends on the desired outcome and the features of the input image.
This algorithm is known as the OIM. The steps of this approach are described below:
Sign=1/1+(1+exp (gain*(cutoff-Z3)).
Tenancy's use the photo below to verify the suggested strategy and the SSM for edge identification for face recognition that is described in part 4.1.
It is feasible to demonstrate the impact of the Canny edge sensor Figure 1 shows the original photo and Figure 2 concludes the original colour to grayscale converted original photo. Figure 3 represents the original photograph, a canny edge detector.
In prior techniques, the operator A's convolved picture was used with a skillful edge detector after the original image had been convolved with the operator A [45, 46]. It came across the picture in Figure 4.
The benefits of recognition have been discovered to be achieved by following our recommended course of action. This formula is helpful for obtaining all hidden images in addition to outlining the borders of objects. Figure 5 for the bordered image of the intended SSM illustrates this outcome.
Figure 1. Original colour photo
Figure 2. Colour to grayscale converted original photo
Figure 3. On the original photograph, a canny edge
Figure 4. Image that has been condensed using operator A and then put through the Canny filter
Figure 5. Planned SSM edged picture
Now take into consideration the OIM that was suggested for image enhancement and covered in part 4.2. After performing experiments on photos using various parameter values and discovering that those parameters produced satisfactory results, it chose the parameter and its values, which are displayed in Table 1. Figure 6 (a) shows the original photo and the image below, which is depicted in Figure 6 (b), has been taken into consideration for the verification of our findings. Please refer to the updated graphics below for details on the various gamma, gain, cutoff, and gamma 2 settings.
After being put through the proposed OIM, the image is shown in Figure 6 (c), with the cut-off set to 0.8, the gamma set to 2, the gain set to 4, and the second gamma set to 0.8. This clearly shows the finished product, which is evidently not visible in the first picture due to its extreme brightness.
The image is shown in Figure 6 (d) after being put through the proposed OIM, with the following settings: cut-off set to worth 0.8, gamma set to worth 2, gain set to worth 6, and second gamma set to worth 7. The original image's dark side is definitely evident in this version, but not sufficiently so.
Figure 6 (e) shows the image after it has been run through the suggested OIA, with the cut-off set to 0.6, the gamma set to 2, the gain set to 6, and the second gamma set to 12. The original image's dark side is actually discernible in this representation. As a result, several cost guidelines define the resulting image. Gain regulates which difference is different. The cutoff is positioned in relation to the gray value in a way where the difference is either reduced or raised.
Table 1. Experimental parameter and selected values
Type of Image |
Parameters and Value |
|||
Gamma |
Gain |
Cutoff |
2nd Gamma |
|
Contrast Image |
2 |
4 |
0.8 |
0.8 |
Improved Image |
2 |
6 |
0.8 |
7 |
Intensity image |
2 |
6 |
0.6 |
12 |
(a)
(b)
(c)
(d)
(e)
Figure 6. (a). Original colour photo; (b). Colour to grayscale converted original photo; (c). Demonstrates a contrast image with the following values: Gain=4, Cutoff=0.8, Gamma=2, and Second Gamma=0.8; (d). Image improvement at Gain=6, Cutoff=0.8, Gamma=2, and Second Gamma=7; (e). Strength image at Gain=6, Cutoff=0.6, Gamma=2, and Second Gamma=12
Based on the previously modified technique in the field of image processing, it first looked at the previous findings before using our computations to provide better results. We have used the Side Search Method (SSM) and the Object Improvement Method (OIM) in this study. The OIM method contributes to improving the excellence of images by improving the instruction of the Gamma, Second Gamma, Gain, and Cutoff parameters. Different gamma values yield different visual effects. A higher gamma value emphasizes contrast, while a lower gamma value enhances details in specific intensity ranges, while the SSM technique is good for more precisely examining the picture side borders. We used both methods to process a variety of photos. Here, we come to the conclusion that both algorithms are successful in addressing the present demands of image processing in an unconstrained face recognition environment with side searching and object improving. We propose emerging a mix method that combines neural and deep learning approaches in the future to improve the photo detection of moving objects.
We acknowledge the help rendered by colleagues in simulation work.
[1] Roushdy, M. (2006). Comparative study of edge detection algorithms applying on the grayscale noisy image using morphological filter. GVIP Journal, 6(4): 17-23.
[2] Luevano, L.S., Chang, L., Méndez-Vázquez, H., Martínez-Díaz, Y., González-Mendoza, M. (2021). A study on the performance of unconstrained very low resolution face recognition: Analyzing current trends and new research directions. IEEE Access, 9: 75470-75493. https://doi.org/10.1109/ACCESS.2021.3080712
[3] Wan, M., Chen, X., Zhan, T., Yang, G., Tan, H., Zheng, H. (2023). Low-rank 2D local discriminant graph embedding for robust image feature extraction. Pattern Recognition, 133: 109034. https://doi.org/10.1016/j.patcog.2022.109034
[4] Zhang, Y., Huang, Y., Yu, S., Wang, L. (2019). Cross-view gait recognition by discriminative feature learning. IEEE Transactions on Image Processing, 29: 1001-1015. https://doi.org/10.1109/TIP.2019.2926208
[5] Panda, C.S., Patnaik, S. (2009). Filtering corrupted image and edge detection in restored grayscale image using derivative filters. International Journal of Image Processing, 3(3): 105-119.
[6] Gupta, S., Maurya, A., Agrawal, A.K. (2022). Unconstrained face recognition system using deep neural network: A review. In 2022 4th International Conference on Advances in Computing, Communication Control and Networking (ICAC3N), Greater Noida, India, pp. 961-966. https://doi.org/10.1109/ICAC3N56670.2022.10074417
[7] Samopa, F., Asano, A. (2009). Hybrid image thresholding method using edge detection. International Journal of Computer Science and Network Security, 9(4): 292-299.
[8] Yao, Y., Ju, H. (2009). A sub-pixel edge detection method based on canny operator. In 2009 Sixth International Conference on Fuzzy Systems and Knowledge Discovery, Tianjin, China, pp. 97-100. https://doi.org/10.1109/FSKD.2009.573
[9] Cha, J., Kim, G. (2007). Enhanced snake algorithm using the proximal edge search method. In Computational Science and Its Applications-ICCSA 2007: International Conference, Kuala Lumpur, Malaysia. Proceedings, Part I. Springer Berlin Heidelberg, 7: 1083-1095. https://doi.org/10.1007/978-3-540-74472-6_90
[10] Dubey, S.R., Singh, S.K., Singh, R.K. (2016). Multichannel decoded local binary patterns for content-based image retrieval. IEEE Transactions on Image Processing, 25(9): 4018-4032. https://doi.org/10.1109/TIP.2016.2577887
[11] Choi, D.H., Jang, I.H., Kim, M.H., Kim, N.C. (2008). Color image enhancement using single-scale retinex based on an improved image formation model. In 2008 16th European Signal Processing Conference, Lausanne, Switzerland, pp. 1-5. http://eurasip.org/Proceedings/Eusipco/Eusipco2008/papers/1569104826.pdf.
[12] Xu, G., Ren, L., Liu, Y. (2014). Flash‐Aware page replacement algorithm. Mathematical Problems in Engineering, 2014(1): 136246. https://doi.org/10.1155/2014/136246
[13] Manjunathachari, A.K., Prasad, K.S. (2006). Implementation of image processing operations using simultaneous multithreading and buffer processing. GVIP Journal, 6(3): 47-53. https://static.aminer.org/pdf/PDF/000/344/797/analysis_and_design_of_parallel_algorithms_and_implementations_for_some.pdf.
[14] Dagar, N.S., Dahiya, P.K. (2016). Soft computing techniques for edge detection problem: A state-of-the-art review. International Journal of Computer Applications, 136(12). https://doi.org/10.5120/ijca2016908615
[15] Ghimpeţeanu, G., Batard, T., Bertalmío, M., Levine, S. (2015). A decomposition framework for image denoising algorithms. IEEE Transactions on Image Processing, 25(1): 388-399. https://doi.org/10.1109/TIP.2015.2498413
[16] Zhu, Q., Mai, J., Shao, L. (2015). A fast single image haze removal algorithm using color attenuation prior. IEEE Transactions on Image Processing, 24(11): 3522-3533. https://doi.org/10.1109/TIP.2015.2446191
[17] Guo, G., Zhang, N. (2018). What is the challenge for deep learning in unconstrained face recognition?. In 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi'an, China, pp. 436-442. https://doi.org/10.1109/FG.2018.00070
[18] Zeng, J., Qiu, X., Shi, S. (2021). Image processing effects on the deep face recognition system. Mathematical Biosciences and Engineering, 18(2): 1187-1200. https://doi.org/10.3934/mbe.2021064
[19] Tyagi, R., Tomar, G.S., Shrivastava, L. (2023). Unconstrained face recognition from image sequence. In 2023 IEEE World Conference on Applied Intelligence and Computing (AIC), Sonbhadra, India, pp. 772-777. https://doi.org/10.1109/AIC57670.2023.10263884
[20] Vaddi, R.S., Boggavarapu, L.N.P., Vankayalapati, H.D., Anne, K.R. (2011). Comparative analysis of contrast enhancement techniques between histogram equalization and CNN. In 2011 Third International Conference on Advanced Computing, Chennai, India, pp. 106-110. https://doi.org/10.1109/ICoAC.2011.6165157
[21] Oloyede, M., Hancke, G., Myburgh, H., Onumanyi, A. (2019). A new evaluation function for face image enhancement in unconstrained environments using metaheuristic algorithms. EURASIP Journal on Image and Video Processing, 1-18. https://doi.org/10.1186/s13640-019-0418-7
[22] Tyagi, R., Tomar, G.S., Shrivastava, L. (2021). Unconstrained face detection of multiple humans present in the video. Wireless Personal Communications, 118(2): 901-917. https://doi.org/10.1007/s11277-020-08050-2
[23] Tyagi, R., Tomar, G.S., Shrivastava, L. (2019). Reference face based technique for unconstrained face recognition from images gallery. International Journal of Computer Graphics, 10(1): 1-16. http://doi.org/10.21742/ijcg.2019.10.1.01
[24] Young, I.T., Gerbrands, J., Vliet, L.J.V. (1995). Fundamental of image processing. https://web.archive.org/web/20170808042633id_/http://archivemiq.free.fr/archives/archives_S3/Signal/Cours%20divers/FIP2.2.pdf.
[25] Joshi, S.R., Koju, R. (2012). Study and comparison of edge detection algorithms. In 2012 Third Asian Himalayas International Conference on Internet, Kathmundu, Nepal, pp. 1-5. https://doi.org/10.1109/AHICI.2012.6408439
[26] Mohan, S., Kumar, S.S. (Eds.). (2013). Proceedings of the Fourth International Conference on Signal and Image Processing 2012 (ICSIP 2012). Springer Science & Business Media, Vol. 221. https://doi.org/10.1007/978-81-322-1000-9
[27] Gonzalez, R.C., Woods, R.E. (2008). Digital image processing third edition. https://dl.ebooksworld.ir/motoman/Digital.Image.Processing.3rd.Edition.www.EBooksWorld.ir.pdf.
[28] Gaetano, R., Pesquet-Popescu, B., Chaux, C. (2013). A convex optimization approach for image resolution enhancement from compressed representations. In 2013 18th International Conference on Digital Signal Processing (DSP), Fira, Greece, pp. 1-8. https://doi.org/10.1109/ICDSP.2013.6622842
[29] Singh, S., Singh, R. (2015). Comparison of various edge detection techniques. In 2015 2nd International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, pp. 393-396. https://doi.org/10.14257/ijsip.2016.9.2.13
[30] Yang, J., Wright, J., Huang, T.S., Ma, Y. (2010). Image super-resolution via sparse representation. IEEE Transactions on Image Processing, 19(11): 2861-2873. https://doi.org/10.1109/TIP.2010.2050625
[31] Tyagi, R., Tomar, G.S., Shrivastava, L. (2016). Unconstrained face recognition quality: A review. International Journal of Signal Processing, Image Processing and Pattern Recognition, 9(11): 199-210. https://doi.org/10.14257/ijsip.2016.9.11.18
[32] Annadurai, S. (2007). Fundamentals of Digital Image Processing. Pearson Education India.
[33] Tyagi, R., Dahiya, P. (2012). An adaptive image enhancement algorithm. International Journal of Advances in Computing and Information Technology, 1(2): 122-127. https://doi.org/10.1016/S0031-3203(96)00145-8
[34] Hasson, N.N., Aljunid, S.A., Ahmad, R.B. (2008). Extract dominant elements and shapes from raster images. In 2008 International Conference on Electronic Design, Penang, Malaysia, pp. 1-4. https://doi.org/10.1109/ICED.2008.4786650
[35] Scott, D.M., McCann, H. (Eds.). (2018). Process Imaging for Automatic Control. CRC Press.
[36] Kumar, K.S., Prasad, S., Saroj, P.K., Tripathi, R.C. (2010). Multiple cameras using real time object tracking for surveillance and security system. In 2010 3rd International Conference on Emerging Trends in Engineering and Technology, Goa, India, pp. 213-218. https://doi.org/10.1109/ICETET.2010.30
[37] Plataniotis, K., Venetsanopoulos, A.N. (2000). Color image processing and application. https://www.comm.toronto.edu/~kostas/Publications2008/pub/bookchapters/2000-SpringerMonograph.pdf.
[38] Kanwal, S., Yousaf, M.H. (2016). Optimization of real time edge enhanced object tracking algorithm on video processor. Technical Journal (Taxila), 19(4): 79-91.
[39] Malviya, L.D., Yadav, R., Lal, J.D. (2010). New Image processing techniques to disclose hidden boundaries. In in 1st International Conference on Intelligent Information System and Management (IISM’10), Coimbatore.
[40] Wan, W., Chen, J., Zhou, Z., Shi, Z. (2022). Self-supervised simple Siamese framework for fault diagnosis of rotating machinery with unlabeled samples. IEEE Transactions on Neural Networks and Learning Systems. https://doi.org/10.1109/TNNLS.2022.3209332
[41] https://drive.google.com/file/d/1Nl0cmZs3L6ZTaDJSD2KsIWgeZH5Z2AUf/view?usp=drive_link, accessed on Jul. 11, 2024.
[42] Saravanan, C. (2010). Color image to grayscale image conversion. In 2010 Second International Conference on Computer Engineering and Applications, Bali, Indonesia, 2: 196-199. https://doi.org/10.1109/ICCEA.2010.192
[43] Maini, R., Aggarwal, H. (2009). Study and comparison of various image edge detection techniques. International Journal of Image Processing (IJIP), 3(1): 1-11. https://www.cscjournals.org/manuscript/Journals/IJIP/Volume3/Issue1/IJIP-15.pdf.
[44] Sha, Z., Zhong, B. (2024). Prediction-Correction line segment detection. In ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Korea, Republic of, pp. 3515-3519. https://doi.org/10.1109/ICASSP48485.2024.10447671
[45] Cheraghi, A. (2016). Searching for an unknown edge in the graph and its tight complexity bounds. Indian Journal of Science and Technology, 9(13). https://doi.org/10.17485/ijst/2016/v9i13/71360
[46] Zhan, C., Duan, X., Xu, S., Song, Z., Luo, M. (2007). An improved moving object detection algorithm based on frame difference and edge detection. In Fourth International Conference on Image and Graphics, Chengdu, China, pp. 519-523. https://doi.org/10.1109/ICIG.2007.153