Skip to main content

ORIGINAL RESEARCH article

Front. Public Health, 27 January 2022
Sec. Digital Public Health
This article is part of the Research Topic Big Data Analytics for Smart Healthcare applications View all 109 articles

An Artificial Intelligence Approach for Expurgating Edible and Non-Edible Items

\nDilip KumarDilip Kumar1Urvashi Bansal
Urvashi Bansal1*Roobaea S. AlroobaeaRoobaea S. Alroobaea2Abdullah M. BaqasahAbdullah M. Baqasah3Mustapha HedabouMustapha Hedabou4
  • 1Department of Computer Science and Engineering, Dr. B. R. Ambedkar National Institute of Technology, Jalandhar, India
  • 2Department Computer Science, College of Computers and Information Technology, Taif University, Taif, Saudi Arabia
  • 3Department of Information Technology, College of Computers and Information Technology, Taif University, Taif, Saudi Arabia
  • 4School of Computer Science, Mohammed VI Polytechnic University, Ben Guerir, Morocco

In the pandemic of COVID-19, it is crucial to consider the hygiene of the edible and nonedible things as it could be dangerous for our health to consume infected things. Furthermore, everything cannot be boiled before eating as it can destroy fruits and essential minerals and proteins. So, there is a dire need for a smart device that could sanitize edible items. The Germicidal Ultraviolet C (UVC) has proved the capabilities of destroying viruses and pathogens found on the surface of any objects. Although, a few minutes exposure to the UVC can destroy or inactivate the viruses and the pathogens, few doses of UVC light may damage the proteins of edible items and can affect the fruits and vegetables. To this end, we have proposed a novel design of a device that is employed with Artificial Intelligence along with UVC to auto detect the edible items and act accordingly. This causes limited UVC doses to be applied on different items as detected by proposed model according to their permissible limit. Additionally, the device is employed with a smart architecture which leads to consistent distribution of UVC light on the complete surface of the edible items. This results in saving the health as well as nutrition of edible items.

1. Introduction

Health is the most critical factor in our life and majorly depends upon what we eat and what is the quality of our daily intake which includes fruits and vegetables having essential minerals, vitamins, and proteins (1). The raw vegetables and fruits may have viruses or pathogens on their surfaces even if some of them are removed by a wash. Due to the COVID-19 pandemic, this becomes essential that the things we are intaking are pathogens free and are virus-free whether it is edible items or our daily usable items. There are many products available in the market for sanitizing edible and non-edible items e.g., UVC rays (24). It is already proven that germicidal UVC has properties that deactivate the pathogens by exposing the surface of fruits and vegetables for a few minutes (58).

In recent times, it is seen that researchers are widely using UVC to perform complex operations such as scientific applications for deactivating viruses; water purification to remove viruses and bacteria from the water; air purification which deactivates the viruses and germs present in the air of the atmosphere (9); sanitizing a room. Additionally, UVC has many applications for deactivating pathogens over different surfaces and mediums. It is also being used with fluorescent which makes the viruses glow when it is seen by microscopes in presence of the UVC rays. UVC is cheap and easily available in the market which further increases its usage.

In the existing solution, there is a limitation that it damages approximately 30% of the edible items especially fruits and vegetables as its being used by industries for decades for removing the pathogens present on the surface of these items (1012). So, the existing solution does not have an automated system that can identify the object and make optimal configuration accordingly and an existing solution does not have any filter which prevents the damage done by the exposure of the direct UVC to this item.

We have designed an Artificial Intelligence employed device to provide an automated self-learning process in which the object is identified by the device itself by using machine learning and make optimal configurations as per the class of the object classified by the object detection model. So, AI enables automation (13) with the optimal setting configuration and makes the device smarter than the existing solutions.

2. Our Contribution

In the view of limitations of existing solutions, we have incorporated certain changes in the proposed device so that edible items do not loose its nutrient values.

• We have used an AI-Enabled UVC sanitization device powered by raspberry pi 4, UVC light 254 nm with polarizing filters for the limited dose to the fruits and vegetables, and a camera sensor for object detection.

• We have used the rotating base to apply UVC light uniformly throughout the object.

• The vacuum pump is used to remove dust particles from inside the box which further maintains the uniform intensity of UVC.

• Inside the box, the mirror is presented on each side so total light reflection occurred when the rays from the UVC lamp strikes the mirrors at different angles.

• The aluminum sheet is also sandwiched between the outer shell and the mirror for maximum heat dissipation when the heavy dose is required by the objects.

• Bipolar fans are also present to cool down the aluminum sheets and can dissipate the heat properly out of the box.

We have used machine learning models for automated UVC sanitization. There are many different objects like vegetables and fruits which vary with proteins structure and have different molecules. It is necessary to detect the object placed inside the box so a camera sensor is used to capture the image of the object placed. This object image is passed to the machine learning model which detects the class of the object and then this class is passed to another classifier which classifies the rotation angle and the safe exposure time for that object based on the similarity.

3. Literature Review

UVC Sanitation is being used for many decades in the industry (14). It is not a new thing that is used for sanitizing the different edible and non-edible items. The industry is using germicidal UVC for raw fruits and vegetables to deactivate the pathogens present on the surface of vegetables and fruits (8). Germicidal UVC of 254 nm has the capability of deactivating pathogens as well as air-born virus-like h1n1 flu and also COVID-19 virus (2, 3, 6, 15). Recently, many devices have come into the market which claims that their devices can sanitize any object. There are many products available in the market like air purifiers using the germicidal UVC.

UVC Sanitization box is available in the market in different sizes which can sanitize the items using the germicidal UVC of 254 nm. There is also a UVC device available that can be used for sanitizing small electronic devices and non-electronics devices (5). These devices are good enough for sanitizing the electronic device and the non-edible items but they are not much effective for edible items like raw vegetables and fruits (3).

Some of the findings were provided in Table 1. As discussed in existing literature, there is a limitation that a few minutes of doses can lead to damage to edible items like fruits and vegetables. So, we proposed a novel idea to overcome this limitation.

TABLE 1
www.frontiersin.org

Table 1. Existing findings.

4. Methodology: System Architecture Overview

4.1. Architecture

The newly built proposed design of the AI-enabled UV sanitization box, as shown in Figures 1, 2, depicts the proposed architecture with the following units:

Processing unit: Raspberry pie 4 would be used as a processing unit.

Object detection unit: This is the unit where all the images captured by the image sensor are pre-processed and then the image of the object is passed to the AI trained model which detects the category of object and results in the category of the object. Here, we have created a machine learning model using the CNN and we trained the model with training images of the different fruits and vegetables.

Sensors: We have used a camera module as an image sensor and temperature sensor along with a cooling module.

Alarm unit: We have used speakers and voice enable system to alert the user.

Cooling unit: The same purpose is solved with heat sink and cooling fans.

UVC rays filter unit: UVC 254 nm is very harmful to the human skin as well as it can damage fruits and vegetables. So, to avoid that damage to the vegetables and fruits, the intensity of the light is controlled using the two polarized sheets (2326). One sheet is fixed and one sheet is made to rotate on both sides directly with the help of a stepper motor. The range of the rotation varies from 0 to 1800. Intensity can be controlled between 0 and 1800 from total light pass to total blockage of the light through the polarizing filter (27).

- Parallel axis: When two sheets are placed parallelly, they allowed the total light to passed through it.

- In between axis: If one sheet is fixed and rotation of the sheet is started in either direction, it will start blocking the light as the sheets are getting crossed to each other. If further rotated, it starts allowing the light through it.

- Crossed axis: If two sheets are crossed then it blocks the light passing through it.

Rotating base unit: A rotating base is used for the even distribution of the UVC rays to the object which is placed on the rotating base. It also prevents the direct penetration of the UVC. The base starts rotating with the start of the sanitization process.

Display unit: The display unit displays all events that occurred at the device. It starts from the device start message and mode selected and then displaying the timer set for the UVC sanitization up to the end of the sanitization process.

FIGURE 1
www.frontiersin.org

Figure 1. Top view of the UVC device.

FIGURE 2
www.frontiersin.org

Figure 2. Front view of the UVC device.

4.2. Mechanism of Sanitization

Smart UVC Sterilization has a special mechanism of sanitization. It is operated from 230-volt AC. There are two shields: one is the outer shell and one is the inner shell. The opening door consists of an auto cut UVC power supply mechanism which is operated using a switch. It will auto cut the power source to UVC when we open the door and a visible lamp will glow to put the objects inside the UVC box. It has a rotating plate that carries the objects. We have a few pre-set items on the operational menu. We can use those pre-sets if we know the edible item otherwise the auto mode is selected.

When the pre-set mode is selected, the UVC Box is set to the predefined fixed mode where the UVC box works on the predefined values for the UVC exposure times and the UVC box runs for the predefined time and set its polarized angle as defined for that mode (ref Figure 3). In the Auto mode, the UVC box uses the AI model to detect the object inside the box. The process is shown in Figure 4.

FIGURE 3
www.frontiersin.org

Figure 3. Pre-set mode operation of UVC sanitization.

FIGURE 4
www.frontiersin.org

Figure 4. Auto mode operation of UVC sanitization.

5. Control Design

Control design process of the AI-enabled UVC device consists of different process flows and control structures used in building the device and process description which defines the different sequences of processes held in the device.

5.1. Main Process Building Block

This section defines the process of the overall workflow of the AI-Enabled device (28). It defines the overall process of the device from the start to the end of the sanitization process. The main process block is defined in Figure 5. The process starts from inputting the edible or non-edible object inside the device and then needs to select the mode of operation of sanitization.

FIGURE 5
www.frontiersin.org

Figure 5. Main process flow diagram.

There are two modes: one is manual which can be used directly for the non-edible items; second mode is auto mode which once selected, it starts capturing the images of the object inserted inside the object. After successfully capturing the image, these images are passed to the object detection model which detects the class of the object and the required UVC exposure time is stored using the UVC dosage calculation formula given in equation 1. According to equation 1, we can easily determine that the total UVC dose time is calculated on the basis of the intensity of light dropped for the particular amount of time.

UVC_Dose=UVC_rays_Intensity(μW/cm2)                        ×UVC_Exposure_Time(seconds).    (1)

On the basis of Table 2, we have calculated the expected doses for the predicted item class. If the class of the object is detected to edible items, then the safe exposure parameters are set according to the detected object, and the rotation angle of the polarizing filter and time of UVC rays exposure is measured and stored using another classifier that classifies the edible item based on the few parameters. Finally, the sanitization process gets started. The procedure is provided in Algorithm 1.

TABLE 2
www.frontiersin.org

Table 2. UV-C doses for SARS-CoV-2.

ALGORITHM 1
www.frontiersin.org

Algorithm 1. UVC_Process.

5.2. Control Process Flow of Camera Sensor

The Camera Sensor is used for capturing the images of the inserted object whenever the auto mode is selected. AI comes into the picture when auto mode is selected and there is a need to classify the class of the object inserted inside the device. Image captured by the camera sensor is saved to the raspberry pi and then it is transferred to the machine learning model which is already trained over the 131 classes of fruits and vegetables. The camera sensor helps to detect the class of the object inserted inside the device.

5.3. Object Detection Process Flow

The object detection process flow is the process that defines the classification of the input Image via a pre-trained model loaded to the raspberry pi. The object detection model is trained over 65,000 images with 131 classes. Image is passed to the model and the model predicts the class and returns the class of the object if it is predicted otherwise return nonedible.

5.4. Polarizer Filter Process Flow

The Polarizer filter process depicts the process of principle used for controlling the intensity of the UVC rays via changing the angle between two polarized sheets. One sheet is mounted with the stepper motor which rotates the motor in both directions. When two sheets are crossed to each other it is 180 degrees angle and zero for full light pass through the polarized sheet. The process starts with the configuration settings from the detected object class and starts rotating to the set angle for the detected class.

5.5. Cooling System Process Flow

The cooling system is an essential part of the device as it takes care of the heat generated inside the device and prevents the excessive heat generated by the UVC rays. To avoid damage from the heat generated by the UVC rays, the cooling system helps the aluminum covering to act as the heat dissipation heat sink (26, 29).

The process of the cooling system consists of a heat sensor module where the auto cut relay is programmed with a temperature setting. The cooling fans control the temperature of the device. The temperature is set to the desired value. The threshold value for the temperature is set to 300C. Whenever the temperature reaches above 300C, it turns on all fans connected to the temperature module. If the temperature goes below 300C, it stops all the fans. This process continues whenever there is a temperature rise and down.

6. Circuit Design

The Circuit design includes the interfacing of the different hardwares. This section depicts the interfacing of the different hardwares used in this device.

6.1. 2004a LCD Display Interfacing

LCD 2004 is a parallel LCD display which is simple and a cost effective solution for adding display characters on a 20 ×4 white and high contrast white text upon a blue background/backlight. The display has 20 characters and 4 lines. It requires 6 data pins to be interfaced with the raspberry pi GPIO pins and other 6 pins with the ground and 5 volts.

The 2004a LCD is interfaced using the i2c adapter which reduces the connection wires from 16 to only a two-wire connection to the I2C adapter. I2C adapter supports many connections to be handled at a time as defined in Figure 6. So, it is used here to reduce the complexity of the raspberry pi. It reduces the number of GPIO pins used by the LCD.

FIGURE 6
www.frontiersin.org

Figure 6. Interfacing of 2004a display LCD with Raspberry Pi.

6.2. NEMA 17 Stepper Motor Interfacing

The Nema17 stepper motor is used for rotating the object holding base. It has high torque which can rotate weight up to 4 kgs. The L298N stepper driver is interfaced with the Nema17 stepper motor and with a raspberry pi to drive the NEMA as shown in Figure 7. Stepper motor drivers need to be powered with 12 volts to drive the NEMA motor.

FIGURE 7
www.frontiersin.org

Figure 7. Interfacing of NEMA 17 stepper motor with Raspberry Pi.

6.3. RJ2003 Stepper Motor Interfacing

It is a stepper motor that consists of 4 coils which allow it to take very precise steps and very accurately it can rotate up to specific angles rotations. It is used for the rotation of the polarizer filter. It makes the filter from a 0-degree angle to 180-degree angle, which means from a total block of light to the total pass of the light through the polarizer filter. It is interfaced with the raspberry pi. In Figure 8, a stepper driver is used to drive the stepper motor that is ULN2003 which is interfaced with both the raspberry pi and stepper motor 28BY-J-48y.

FIGURE 8
www.frontiersin.org

Figure 8. Interfacing of RJ2003 stepper motor with Raspberry Pi.

7. Proposed Technique

7.1. Proposed Object Detection Model

We have used the deep learning method to train the model against the fruits and vegetable dataset of images. We obtained the dataset which is freely available named Fruits-360 which can be downloaded from GitHub or available at Kaggle. It includes about 130,000 images of different categories. This dataset includes high-definition images which are essential to obtain a good classifier (3033).

The structure of the model we used is defined in Figure 9. There are a total of 11 layers used in our model. There are four convolutional layers, 4max pooling layers and 2 fully connected layers, and the output layer is the SoftMax layer. The input layer is a convolutional layer where the filter of 16, 5 x 5 x 4 filter is applied and the max-pooling of 2 x 2 is used with stride=2. Convolutional layer 4 is with parameter (5 x 5 x 64) with dimension 128. The model is trained with the 131 classes of fruits and vegetables and it can be capable of detecting the multiclass of the image also.

FIGURE 9
www.frontiersin.org

Figure 9. CNN model layers.

7.2. Optimal Configuration Prediction Model

This is the model proposed to predict the optimal configuration for the detected object. This model makes a prediction based on the data collected through the experimentations. This collected dataset includes the entity name, its features and configuration such as time of UVC exposure and the angle of rotation of the polarizer.

The model used is a decision tree classifier and it works fine as per the requirement. We can classify the UVC exposure time and the angle of rotation of the polarizer and be able to detect the required configuration of the UVC sanitization process. This classifier gives an accuracy score of the 96.33%.

7.3. Smart Device Control

The device is controllable through a mobile app (android). There are two interface screens of the app. The use of the mobile app is to control the device using the Bluetooth communication channel and can be operatable using the mobile app interfaces (3440). The wireless channel enables the message transfer from mobile to operate the device using the host program running on the raspberry pi and client program on the android app.

7.4. Control Screen

The device is configured for the following operations:

7.4.1. Mode Selection

There are two select modes

• Auto mode

• Default mode

The two modes can be selected from app as shown in the app screen in Figure 10. One of the mode is the auto mode where the AI comes into the picture and takes care of the rest of the process of the sanitization. The other mode is the default mode operating normally for non-edible items. Its configuration is fixed and loaded on the selection.

FIGURE 10
www.frontiersin.org

Figure 10. Mobile app main screen with controls.

8. Results and Discussion

The code is implemented two times, one experiment is done with 20 items with 20 iterations, another one is done with the 25 items with 20 iterations. The goal is to create an optimal configuration for the discovered items.

8.1. Data Set

In this research realm, the two data sets taken are listed below:

• Fruits 360: This dataset includes the 131 classes of images of both fruits and vegetables (41).

• Configuration mapping: This dataset is made from the experiments performed by us for optimal configuration for the different objects.

8.2. System Requirements

The code of the proposed object detection and training model is written in Python and TensorFlow libraries. The code has been implemented to detect the object class and make optimal configuration using the classification model on experiments results to assess the performance of the UVC process algorithm. The following specifications are used to implement the algorithm:

• Processor - Intel(R) i7-8750H CPU @ 2.40GHz 2.40 GHz

• GPU - NVIDIA-GeForce GTX-1050

• RAM - 8 GB DDR4

• Hard disk drive - 256 SSD

• Operating System - Windows 10 Pro x64-bit

9. Results

The machine learning model used is CNN for object detection and we have used the fruits360 dataset (41) and 50 epochs for training the model which gives an accuracy of 96.2% without overfitting and underfitting as we have provided enough training data for training the model. We have trained the model on an i7 intel processor with Nvidia graphics as it is not possible to train a model over the raspberry pi. Due to limited hardware configuration, we trained it over the system and used the trained model in raspberry pi. So, the trained model gives a good result on the raspberry pi as it is trained over 131 classes of the fruits and vegetables of 65,000 images. We have run the device for the different configurations and the different filter parameters. We have run the machine for the following angle of filters as follows:

• 00 angle filter rotation: Here, the light passes through the filters as the normal light passes as it permits the whole rays to pass through the filters.

• 450 angle filter rotation: This angle of rotation allows lights to pass through the filter but with limitations as it does not allow the whole UVC light to get pass through it. It allows 75% of UVC light through.

• 900 angle filter rotation: At this angle of rotation of the filter, 50% of rays are blocked and 50% of rays get passed through the filters.

• 1350 angle filter rotation: At this angle of rotation of the filter, 70% of rays are blocked and 25% of rays get passed through the filters.

• 1800 angle filter rotation: At this angle of rotation of the filter, 100% rays are blocked and 0% rays get passed through the filters.

So, we have used the 45-degree, 90-degree, and 135-degree rotations for our experiments and get the following results for a few of the fruits as shown in Table 3. We have calculated the UVC dose time using equation 1. As our UVC lamp has a UVC intensity of 300 μW/cm2 and the length we have to use is 15 cm. As represented in Table 3, the edible items (especially raw fruits and vegetables) can be sanitized without losing their quality and without any damage to the cell body of the edible item using the proposed system architecture.

TABLE 3
www.frontiersin.org

Table 3. Experimental results.

9.1. Performance Plot

We have plotted the performance graph for our model for the training data and the validation data. We plot the graph between training data loss vs. validation data loss and training data accuracy and validation data accuracy (Figure 11) of the CNN model for object detection. The performance graph plot tells the increase in accuracy with the increase in the epochs and it also describes whether the model is trained properly or not. Figure 11 is the graph plotted between the loss at training and loss at Validation time. As portrayed in the graph, it is nearly converging so there is no overfitting and underfitting in our model. Training accuracy vs. validation accuracy performance graph shows that the model is trained nearly perfect as there is very little difference between the validation and the training accuracy. So, our model gives an accuracy of the 96.2% which is good enough to detect the type of vegetable and fruits from an image and it works fine for the single item or single fruit or vegetable image. It is also capable of predicting the class of multiple objects too. The model works fine for 131 types of categories. It gives very accurate results for the images input to the model. Finally, Figures 1216 provide some additional images of the device under testing. These images are for illustration purposes only.

FIGURE 11
www.frontiersin.org

Figure 11. Training accuracy vs. validation accuracy (Left) training loss vs. validation loss (Right) performance graph.

FIGURE 12
www.frontiersin.org

Figure 12. Device front view.

FIGURE 13
www.frontiersin.org

Figure 13. Device internal view.

FIGURE 14
www.frontiersin.org

Figure 14. Device internal view - rotating base.

FIGURE 15
www.frontiersin.org

Figure 15. Device internal view - processing unit and cooling fans.

FIGURE 16
www.frontiersin.org

Figure 16. Device connectivity.

10. Conclusion and Future Scope

With the proposed system architecture, the edible items (especially raw fruits and vegetables) can be sanitized without losing their quality and without any damage to the cell body of the edible item. The AI-enabled device sanitizes the edible items with different optimal configurations made using object detection and automated configuration detection based on the detected class. The optimal configurations utilize the phenomenon of polarization filters which limits the expose of rays due to which the edible items get sanitized without losing their nutrition. Limited filter UVC doses are exposed uniformly throughout the surface of the item with different configurations with the optimal time.

In future, some points need to be considered to enhance the research work. The developed prototype is smart enough to detect the object inserted inside the box but the UVC 254 nm is not safe for humans as well as for edible items. So, 222 nm Germicidal UVC light can be used instead of 254 nm if it is available. Smart Home integration and API could be done to control the device using IoT, e.g., google assistant. Simultaneous sanitization could be performed for multiple items using optimal parameters.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://www.kaggle.com/moltean/fruits.

Author Contributions

UB and DK carried out the experiment. DK wrote the manuscript with support from UB. DK fabricated the device with the help from RA, AB, and MH. UB supervised the project. All authors contributed to the article and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

The authors are grateful to the Taif University Researchers Supporting Project (Number TURSP-2020/36), Taif University, Taif, Saudi Arabia and Dr. B. R. Ambedkar National Institute of Technology Jalandhar for their support to carry out this work.

References

1. Xiong H, Jin C, Alazab M, Yeh KH, Wang H, Gadekallu TRR, et al. On the design of blockchain-based ECDSA with Fault-tolerant batch verication protocol for blockchain-enabled IoMT. IEEE J Biomed Health Inform. (2021) 1–6. doi: 10.1109/JBHI.2021.3112693

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Santhosh R, Yadav S. Low cost multipurpose UV-C sterilizer box for protection against COVID-19. In: 2021 International Conference on Artificial Intelligence and Smart Systems (ICAIS). Coimbatore: IEEE (2021). p. 1495–8.

Google Scholar

3. Zaffina S, Camisa V, Lembo M, Vinci MR, Tucci MG, Borra M, et al. Accidental exposure to UV radiation produced by germicidal lamp: case report and risk assessment. Photochem Photobiol. (2012) 88:1001–4. doi: 10.1111/j.1751-1097.2012.01151.x

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Guerrero-Beltr·n J, Barbosa-C·novas G. Advantages and limitations on processing foods by UV light. Food Sci Technol Int. (2004) 10:137–47. doi: 10.1177/1082013204044359

CrossRef Full Text | Google Scholar

5. Anderson M. The ultraviolet offense: Germicidal UV lamps destroy vicious viruses. New tech might put them many more places without harming humans. IEEE Spectrum. (2020) 57:50–5. doi: 10.1109/MSPEC.2020.9205549

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Nenova V, Nenova MV, Georgieva DV, Gueorguiev VE. ISO standard implementation impact in COVID-19 erra on UV-lighting devices. In: 2020 Fifth Junior Conference on Lighting (Lighting). Ruse: IEEE (2020). p. 1–4.

Google Scholar

7. Cao Y, Chen W, Li M, Xu B, Fan J, Zhang G. Simulation based design of deep ultraviolet LED array module used in virus disinfection. In: 2020 21st International Conference on Electronic Packaging Technology (ICEPT). Guangzhou: IEEE (2020). p. 1–4.

Google Scholar

8. Yamano N, Kunisada M, Kaidzu S, Sugihara K, Nishiaki-Sawada A, Ohashi H, et al. Long-term effects of 222-nm ultraviolet radiation C sterilizing lamps on mice susceptible to ultraviolet radiation. Photochem Photobiol. (2020) 96:853–62. doi: 10.1111/php.13269

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Alyami H, Alosaimi W, Krichen M, Alroobaea R. monitoring social distancing using artificial intelligence for fighting COVID-19 virus spread. Int J Open Source Softw Process. (2021) 12:48–63. doi: 10.4018/IJOSSP.2021070104

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Dhanamjayulu C, Nizhal U, Maddikunta PKR, Gadekallu TR, Iwendi C, Wei C, et al. Identification of malnutrition and prediction of BMI from facial images using real-time image processing and machine learning. IET Image Process. (2021) 1–12. Available online at: https://doi.org/10.1049/ipr2.12222

Google Scholar

11. Mukhtar H, Rubaiee S, Krichen M, Alroobaea R. An IoT framework for screening of COVID-19 using real-time data from wearable sensors. Int J Environ Res Public Health. (2021) 18:4022. doi: 10.3390/ijerph18084022

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Arul R, Alroobaea R, Tariq U, Almulihi AH, Alharithi FS, Shoaib U. IoT-enabled healthcare systems using block chain-dependent adaptable services. Pers Ubiquitous Comput. (2021) p. 1–15. doi: 10.1007/s00779-021-01584-7

CrossRef Full Text | Google Scholar

13. Wang W, Qiu C, Yin Z, Srivastava G, Gadekallu TR, Alsolami F, et al. Blockchain and PUF-based lightweight authentication protocol for wireless medical sensor networks. IEEE Intern Things J. (2021) 14:1–9. doi: 10.1109/JIOT.2021.3117762

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Bourouis S, Channoufi I, Alroobaea R, Rubaiee S, Andejany M, Bouguila N. Color object segmentation and tracking using flexible statistical model and level-set. Multimed Tools Appl. (2021) 80:5809–31. doi: 10.1007/s11042-020-09809-2

CrossRef Full Text | Google Scholar

15. Juarez-Leon FA, Soriano-Sánchez AG, Rodríguez-Licea MA, Perez-Pinal FJ. Design and implementation of a germicidal UVC-LED lamp. IEEE Access. (2020) 8:196951–62. doi: 10.1109/ACCESS.2020.3034436

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Weber DJ, Rutala WA, Anderson DJ, Chen LF, Sickbert-Bennett EE, Boyce JM. Effectiveness of ultraviolet devices and hydrogen peroxide systems for terminal room decontamination: focus on clinical trials. Am J Infect Control. (2016) 44:e77–e84. doi: 10.1016/j.ajic.2015.11.015

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Scotland HP. Literature review and practice recommendations: existing and emerging technologies used for decontamination of the healthcare environment: hydrogen peroxide. Natl Serv Scotland. (2016). Published on December 5 2016, Available online at: https://www.nss.nhs.scot/publications/literature-review-and-practice-recommendations-existing-and-emerging-technologies-for-decontamination-of-the-health-and-care-environment-airborne-hydrogen-peroxide-v1-1/

Google Scholar

18. Barbut F, Menuet D, Verachten M, Girou E. Comparison of the efficacy of a hydrogen peroxide dry-mist disinfection system and sodium hypochlorite solution for eradication of Clostridium difficile spores. Infection Control Hosp Epidemiol. (2009) 30:507–14. doi: 10.1086/597232

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Holmdahl T, Lanbeck P, Wullt M, Walder M. A head-to-head comparison of hydrogen peroxide vapor and aerosol room decontamination systems. Infection Control Hosp Epidemiol. (2011) 32:831–6. doi: 10.1086/661104

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Fu T, Gent P, Kumar V. Efficacy, efficiency and safety aspects of hydrogen peroxide vapour and aerosolized hydrogen peroxide room disinfection systems. J Hosp Infect. (2012) 80:199–205. doi: 10.1016/j.jhin.2011.11.019

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Haas JP, Menz J, Dusza S, Montecalvo MA. Implementation and impact of ultraviolet environmental disinfection in an acute care setting. Am J Infect Control. (2014) 42:586–90. doi: 10.1016/j.ajic.2013.12.013

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Memarzadeh F, Olmsted RN, Bartley JM. Applications of ultraviolet germicidal irradiation disinfection in health care facilities: effective adjunct, but not stand-alone technology. Am J Infect Control. (2010) 38:S13–S24. doi: 10.1016/j.ajic.2010.04.208

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Lu H. Polarization separation by an adaptive filter. IEEE Trans Aerosp Electron Syst. (1973) 6:954–6. doi: 10.1109/TAES.1973.309675

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Urciuoli A, Buono A, Nunziata F, Migliaccio M. Analysis of local-and non-local filters for multi-polarization SAR coastline extraction applications. In: 2019 IEEE 5th International forum on Research and Technology for Society and Industry (RTSI). Florence: IEEE (2019). p. 28–33.

Google Scholar

25. Lee JS, Grunes MR, De Grandi G. Polarimetric SAR speckle filtering and its implication for classification. IEEE Trans Geosci Remote Sens. (1999) 37:2363–73. doi: 10.1109/36.789635

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Qu T, Wang X, Zhao F. Effect of many air cooling fans in operation on dynamic response of bridge truss. In: 2011 International Conference on Consumer Electronics, Communications and Networks (CECNet). Xianning: IEEE (2011). p. 1194–7.

Google Scholar

27. Calisti M, Carbonara G, Laschi C. A rotating polarizing filter approach for image enhancement. In: OCEANS 2017-Aberdeen. Aberdeen: IEEE (2017). p. 1–4.

Google Scholar

28. Deepa N, Prabadevi B, Maddikunta PK, Gadekallu TR, Baker T, Khan MA, et al. An AI-based intelligent system for healthcare analysis using Ridge-Adaline Stochastic Gradient Descent Classifier. J Supercomput. (2021) 77:1998–2017. doi: 10.1007/s11227-020-03347-2

CrossRef Full Text | Google Scholar

29. Carl R, Lawless P, Decker J. Characterization and mitigation of cooling fan installation penalties in 1U enterprise class servers. In: 2014 Semiconductor Thermal Measurement and Management Symposium (SEMI-THERM). San Jose, CA: IEEE (2014). p. 213–8.

Google Scholar

30. Huang Z, Cao Y, Wang T. Transfer learning with efficient convolutional neural networks for fruit recognition. In: 2019 IEEE 3rd Information Technology, Networking, Electronic and Automation Control Conference (ITNEC). Chengdu: IEEE (2019). p. 358–62.

Google Scholar

31. Pande A, Munot M, Sreeemathy R, Bakare R. An efficient approach to fruit classification and grading using deep convolutional neural network. In: 2019 IEEE 5th International Conference for Convergence in Technology (I2CT). Bombay: IEEE (2019). p. 1–7.

Google Scholar

32. Zhu D, Wang M, Zou Q, Shen D, Luo J. Research on fruit category classification based on convolution neural network and data augmentation. In: 2019 IEEE 13th International Conference on Anti-counterfeiting, Security, and Identification (ASID). Xiamen: IEEE (2019). p. 46–50.

Google Scholar

33. Kausar A, Sharif M, Park J, Shin DR. Pure-cnn: a framework for fruit images classification. In: 2018 International Conference on Computational Science and Computational Intelligence (CSCI). Las Vegas, NV: IEEE (2018). p. 404–8.

Google Scholar

34. Das S, Ganguly S, Ghosh S, Sarker R, Sengupta D. A bluetooth based sophisticated home automation system using smartphone. In: 2016 International Conference on Intelligent Control Power and Instrumentation (ICICPI). Kolkata: IEEE (2016). p. 236–40.

Google Scholar

35. Kumar S, Lee SR. Android based smart home system with control via Bluetooth and internet connectivity. In: The 18th IEEE International Symposium on Consumer Electronics (ISCE 2014). Jeju: IEEE (2014). p. 1–2.

Google Scholar

36. Singh A, Gupta T, Korde M. Bluetooth controlled spy robot. In: 2017 International Conference on Information, Communication, Instrumentation and Control (ICICIC). Indore: IEEE (2017). p. 1–4.

Google Scholar

37. Paala NMA, Garcia NMM, Supetran RA, Fontamillas MELB. Android controlled lawn mower using Bluetooth and WIFI connection. In: 2019 IEEE 4th International Conference on Computer and Communication Systems (ICCCS). Singapore: IEEE (2019). p. 702–6.

Google Scholar

38. Shinde A, Kanade S, Jugale N, Gurav A, Vatti RA, Patwardhan M. Smart home automation system using IR, bluetooth, GSM and android. In: 2017 Fourth International Conference on Image Information Processing (ICIIP). Shimla: IEEE (2017). p. 1–6.

Google Scholar

39. Tripti NF, Farhad A, Iqbal W, Zaman HU. SaveMe: a crime deterrent personal safety android app with a bluetooth connected hardware switch. In: 2018 9th IEEE Control and System Graduate Research Colloquium (ICSGRC). Shah Alam: IEEE (2018). p. 23–6.

Google Scholar

40. Agarwal S, Agarwal N. Interfacing of robot with android app for to and fro communication. In: 2016 Second International Innovative Applications of Computational Intelligence on Power, Energy and Controls With Their Impact on Humanity (CIPECH). Ghaziabad: IEEE (2016). p. 222–6.

Google Scholar

41. Fruit-360 dataset. (2021). Available online at: https://www.kaggle.com/moltean/fruits.

Google Scholar

Keywords: food sanitization, smart sanitization, UVC, germicidal, food safety, machine learning for health, smart system

Citation: Kumar D, Bansal U, Alroobaea RS, Baqasah AM and Hedabou M (2022) An Artificial Intelligence Approach for Expurgating Edible and Non-Edible Items. Front. Public Health 9:825468. doi: 10.3389/fpubh.2021.825468

Received: 30 November 2021; Accepted: 20 December 2021;
Published: 27 January 2022.

Edited by:

Thippa Reddy Gadekallu, VIT University, India

Reviewed by:

Zhaoyang Han, University of Aizu, Japan
Mohamed Mounir, El gazeera High Institute for Engineering and Technology, Egypt

Copyright © 2022 Kumar, Bansal, Alroobaea, Baqasah and Hedabou. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Urvashi Bansal, dXJ2YXNoaSYjeDAwMDQwO25pdGouYWMuaW4=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.