Despite fast technological progress, pharmaceutical industries still face several problems in drug manufacturing. The conventional manufacturing process of drugs involves the possibility of human-made errors and it is not cost-effective in terms of time spent on the verification. Computer vision for quality control in automated manufacturing systems can help significantly in this task by checking for broken or imprint errors on the pills. In this project, I built a proof-of-concept of an anomaly detection system for pharmaceutical pills (capsules) using the Edge Impulse FOMO (Faster Objects, More Objects) algorithm. As the medication makes its way through the production line, images are taken using a web camera and transferred to a Seeed Studio reComputer (Jetson Xavier NX module) which processes the image using the algorithm designed to check if the pills have any anomalies. Real-time inferencing can be monitored remotely using a web browser over a WiFi network.
Hardware SetupThis project requires to do inferencing at greater speed to detect defects in real-time to assist the manufacturing process. The reComputer Jetson Xavier NX (8GB RAM) is used for development and deployment as well. A 5-megapixels USB webcam is used to capture images. The headless reComputer is connected to the home router using a USB WiFi dongle and is accessed using an SSH connection.
The reComputer comes with Ubuntu 18.04 and JetPack 4.6.2 preinstalled. To set this device up in Edge Impulse, ssh to the device and run the following command in the terminal.
$ wget -q -O - https://cdn.edgeimpulse.com/firmware/linux/jetson.sh | bash
Data CollectionWe are using a set of different color empty capsule shells for data generation. The pharmaceutical capsule shell consists of two pre-fabricated, cylindrical sections (a cap and a body) each of which has one rounded, closed-end and one open end. The body has a slightly lower diameter than the cap and fits inside the cap.
To imitate real types of defects, a dark spot on the body and crack or cuts on the cap brim are fabricated on a few samples. The normal capsule samples and plain base are treated as background and only defective samples are labeled with bounding boxes. Although in the real industrial setup there is a controlled illumination condition, we have captured the images with slightly different lighting conditions to generate resilient datasets. Also, it is assumed that in the production line (under the camera frame) there are no objects other than the background and the capsules.
We need to create a new project to upload data to Edge Impulse Studio.
With the camera connected to the reComputer and run the following command.
$ edge-impulse-linux --disable-microphone
This will start a wizard which will ask you to log in. It will show a list of projects and we can select the newly created project from the list. Now go to the Data Acquisition page at the Edge Impulse Studio where we would be able to see the live camera feed in the Training data tab.
By clicking on the Start sampling button, we can capture the images for training. We have collected 206 images (167 for training and 39 for testing). Now we can label the data with defects in the Labeling queue tab as shown in the GIF below.
Go to the Impulse Design > Create Impulse page, click Add a processing block, and then choose Image, which preprocesses and normalizes image data, and optionally allows you to choose the color depth. Also, on the same page, click Add a learning block, and choose Object Detection (Images) which fine-tunes a pre-trained object detection model on your data. Since the defects are usually tiny so we are using a 320x320 image size for better resolution for detection. Now click on the Save Impulse button.
On the Image page, choose RGB as color depth and click on the Save parameters button. The page will be redirected to the Generate Features page.
Now we can start feature generation by clicking on the Generate features button,
After feature generation, go to the Object Detection page and click on Choose a different model and select FOMO. Then click on the Start training button. It will take a few minutes to complete the training.
This project aims to use Edge Impulse Studio to collect data and train a model for the anomaly detection task. In the preliminary trials, the supported off-the-shelf image classification or object detection models perform poorly to detect tiny defects on the small capsules. Usually, Generative Adversarial Networks (GAN) or Autoencoder based models are used for anomaly detection for similar datasets. The Edge Impulse FOMO algorithm works quite well for the given task. FOMO uses an architecture similar to a standard image classification model which splits the input image into a grid and runs the equivalent of image classification across all cells in the grid independently in parallel. By default the grid size is 8x8 pixels, which means for a 320x320 image, the output will be 40x40 as shown in the image below.
For localization, it cuts off the last layers of the classification model and replaces this layer with a per-region class probability map, and subsequently applies a custom loss function that forces the network to fully preserve the locality in the final layer which essentially gives us a heat map of defects locations. FOMO works on the constraining assumption that all the bounding boxes are square and have a fixed size and the objects are spread over the output grid. In our case these assumptions are inherent and constraints become advantages.
Confusion MatrixOnce the training is completed we can see the confusion matrix as shown below. The F1 score is 98.3% which is very good.
In the Model testing page, click on the Classify all button which will initiate model testing with the trained float32 model. The testing accuracy is 89.74%.
To achieve the highest possible inferencing rate we will be running the model using the GPU on the Jetson Xavier NX. On the Deployment page, we will choose NVIDIA TensorRT Library from the Deploy your impulse section.
In the Select Optimizations section, we will choose the float32 model which has better latency which a bit bigger but we have plenty of memory available at Jetson Xavier NX. The model build process starts after clicking on the Build button and will be downloaded automatically in a while.
We will be deploying the model in the executable .eim format that contains signal processing and machine learning code, compiled with optimizations for GPU plus a very simple IPC layer (over a Unix socket). Execute the command below to set up Linux C++ SDK and install OpenCV to build the .eim file.
$ git clone https://github.com/edgeimpulse/example-standalone-inferencing-linux
$ cd example-standalone-inferencing-linux
$ git submodule update --init --recursive
$ sh build-opencv-linux.sh
Now unzip the downloaded model bundle into this repository. We can compile the .eim model with TensorRT using the command below.
$ APP_EIM=1 TARGET_JETSON_NANO=1 make -j
After the successful build process, the model.eim file will be generated in the build directory.
Run InferencingWe will be using the Edge Impulse command line tools to run inferencing. Also, the Jetson Xavier NX power mode is changed to 20W (2 cores) for better utilization of the device.
$ sudo nvpmodel -m 6 && sudo jetson_clocks
$ edge-impulse-linux-runner --model-file ./build/model.eim
[RUN] Starting the image classifier for Edge Impulse Experts / PPAD2 (v16)
[RUN] Parameters image size 320x320 px (3 channels) classes [ 'defect' ]
[RUN] Using camera ELECOM 5MP Webcam starting...
[RUN] Connected to camera
Want to see a feed of the camera and live classification in your browser? Go to http://192.168.3.12:4912
We can see the live inferencing in a web browser by using the URL in the output above.
The inferencing rate is around 165 fps which is blazingly fast. Also, the detection seems to be working great while objects are moving.
This project showcases an industrial use case for pharmaceutical pill defect detection using Edge Impulse FOMO, which can be reproduced with other similar kinds of datasets for anomaly detection. Although the project was created using a Seeed reComputer using a Jetson Xavier NX module with an Nvidia GPU to achieve a high inferencing rate, it can be ported easily to less powerful single-board computers with a reduced inferencing rate.
Comments
Please log in or sign up to comment.