Open Access Paper
11 September 2023 Design of security image contraband detection system based on PP-YOLOE+_DCS
Fan Chen, Hua Jin, Qinghan Li
Author Affiliations +
Proceedings Volume 12779, Seventh International Conference on Mechatronics and Intelligent Robotics (ICMIR 2023); 127791S (2023) https://doi.org/10.1117/12.2688728
Event: Seventh International Conference on Mechatronics and Intelligent Robotics (ICMIR 2023), 2023, Kunming, China
Abstract
Traditional security screening methods mainly use manual identification of security images, there is a low identification of inefficiency, high error of judgement rate, which has become a bottleneck limiting public safety and security. Therefore, to deal with this problem, this paper proposes the security image contraband detection model PP-YOLOE+_DCS, which makes three main improvements on the basis of the PP-YOLOE+ model. To begin with, we introduced deformable convolution within the backbone network that strengthen the models' feature extraction capability. Secondly, we introduced a coordinated attention mechanism among the backbone network and the detection neck for better focusing the model on the object region. Finally, we replaced the original GIOU loss function with the SIOU loss function to improve the detection accuracy and training speed. The improved PP_YOLOE+_DCS model obtained achieves 91.4% detection accuracy, 2.8% improvement compared with the baseline model mAP, only 0.24 M additional parameters, and 420.2 ms inference delay on embedded devices, which provides a new solution for the intelligence of contraband detection.

1.

INTRODUCTION

As the demand for people’s travel increases, the safety of public transportation hubs such as airports, train stations and subway stations is of great concern.X-ray baggage screening is one of the necessary measures to ensure people’s travel security, which can detect and intercept people or items carrying contraband in a timely manner1. However, the traditional security screening method mainly relies on manual identification of contraband in X-ray images, which has low efficiency, time consuming, high false positive rate, and an accuracy rate of only 80-90% 2, 3. Therefore, how to enhance the efficiency and precision of security checks has become a pressing issue in the field of security screening.

With the significant advances in image recognition and natural language processing made by deep learning techniques during recent years, many researchers and academics have begun to explore the application of deep learning to automatically detect prohibited items in X-ray security images.However, deep learning models usually require high-performance computers or cloud servers for training and inference, while embedded devices have unique advantages in deploying deep learning models by virtue of their small size, low power consumption, high performance and real-time response. Therefore, studying an X-ray security image contraband-assisted detection system that can be deployed in embedded devices to better assist security inspectors in their inspection work can effectively reduce the impact of human factors.

Considering the indicators of accuracy and model complexity, this paper selects the S-model model of PP-YOLOE+ as the baseline, and improves and optimizes it to obtain a deep learning model suitable for contraband detection. And using the Paddle Inference inference framework, the contraband detection model is deployed into the embedded device, aiming to realize the whole process of contraband detection in X-ray security images by intelligent means, reducing the work intensity of security inspectors and speeding up the efficiency of security inspection.

2.

RELATED WORK

Recently, the advancement of deep learning techniques has enabled neural networks that can extract features directly from the initial data, which has significant advantages in terms of detection efficiency and accuracy. Object detection algorithms based on deep learning primarily fall into two-stage methods that are based on candidate regions and one-stage methods that are based on regression problems. The advantage of the two-stage method is high accuracy, and the disadvantage is high complexity and computational effort. The advantage of the one-stage method is its simplicity, speed and wide applicability, and the disadvantage is the relatively low detection accuracy.

Currently, X-ray image contraband detection methods are mainly based on targeted improvements and optimizations of the first-stage algorithms. While the improved model of contraband detection method brings some contribution to the contraband detection task, it also reveals some shortcomings. Zhou et al4 proposed an improved X-ray safety image detection algorithm that is based on the YOLOv4 algorithm, which addresses the problems of complex backgrounds, target scale variations and mutual occlusion by introducing techniques such as deformable convolution, GHM loss and non-maximum suppression methods, with the disadvantage that GHM loss requires the computation of the gradient of the average value and the variance of the samples, thus increasing the computational sophistication and lengthening the training time.Ren et al5 put forward the LightRay, a light-weight object detection framework for YOLOv4, that uses MobileNetV3 as its backbone feature extraction network, a light-weight feature pyramid network LFPN and CBAM attention mechanism, which is effective in achieving feature fusion at different scales and enhancing the feature information of small-sized contraband in complex backgrounds. However, when the detection results are unsatisfactory, more computational resources are still required.Liu et al6 put forward a light-weight contraband detection method based on YOLOV4, denoted as LPD-YOLOv4. The method using MobilenetV3 as its backbone feature extraction network and using depth-separated convolution minimizes the parameter count and its computational consumption through optimizing the neck and head.In addition, it is designed with an self-adaptive space and channel attention blocks for improving the feature extraction capability.However, there is a reduced loss of detection accuracy compared to the original YOLOv4 model.Song et al7 have proposed an enhanced YOLOv5 model that combines the CGhost module, Stem module,and Mixup data augmentation method to build contraband recognition capability. However, since Mixup requires the generation of new data samples,it increases the calculation sophistication for the model, which may contribute to the increase of training time and resources.

3.

METHODOLOGIES

Because of the limited calculation performance of the deployed devices, the S-model of the PP-YOLOE+ series algorithm8 is chosen as the baseline for this paper, on which optimization and improvement are carried out. The new model is denoted as PP-YOLOE+_DCS, and its algorithm structure is shown in Figure 1. The optimized PP-YOLOE+_DCS algorithm has three main improvement points, as shown in the solid box part in Figure 1, as follows:

Figure 1.

Structure of the improved PP-YOLOE+_DCS algorithm

00104_PSISDG12779_127791S_page_2_1.jpg

(1) Introducing deformable convolution v29 that is to strengthen the backbone network’s feature extraction capabilities and allow it to better capture the deformation and pose information of the target; (2)Introducing a CA (Coordination attention)10 mechanism that enhances the attraction and perception of the model to key parts of the object among the backbone and feature fusion networks;(3)Substitute the former GIOU loss function of the model with the SIOU loss function11, which is used to redefine the penalty index, and consider the fast convergence of the distance that the predicted box is from the real box to strengthen model’s localization accuracy and matching ability of the object.

3.1

Design of D-Res module

PP-YOLOE+ adopts an multi-scale feature map with three different scales to perform multi-scale feature integration which can adapt to targets of different sizes and have better detection effect on traditional natural images.However, due to the randomness of passenger luggage placement, after X-ray scanning, the contraband in the image will appear as multi-scale and multi-pose features, while ordinary items may be mistaken as contraband due to mutual occlusion and overlap between items.

To handle the above problem, we introduce the deformable convolutional DCNv2 that has modulation mechanism into the backbone network in this paper.Compared with the traditional convolution operation, DCNv2 can handle the spatial variation of target objects flexibly for improving the detection and recognition competence of the model. Since there is no fixed theoretical description for inserting deformable convolution in the network, the DBS module and the D-Res Block module are designed in this paper, and the D-Res module is obtained after replacing the backbone network CSPRes module, as shown in Figure 2. Each DBS module consists of DCNv2, BN and SiLU activation functions, for which n refers to the convolutional kernel size of DCNv2 and m refers to the sampling steps of the convolutional kernels.

Figure 2.

Design of D-Res module

00104_PSISDG12779_127791S_page_3_1.jpg

3.2

Coordinate attention

Since security screening images are different from those in natural scenes, there is a large amount of noise and complex background information.This redundant information can hinder the overall network’s representational capability and affect the model’s overall detection of contraband.The coordinated attention mechanism is a method that can capture both channel relevance and spatial relevance in the feature map, and can adaptively adjust the weight relationship between different channels and different locations, so that the channels and locations with important information can receive more attention.It is shown that the coordination attention mechanism embedded into the backbone network and feature fusion network, shown on the CA module marked by the red solid line box in Figure 1.

The CA mechanism decomposes channel attention that captures features along two spatial directions from two one-dimensional feature encoding processes, one of which is responsible for capturing remote-dependent features and the other retains precise location information.The architecture is shown in Figure 3.

Figure 3.

Architecture of the coordination of attention mechanism

00104_PSISDG12779_127791S_page_4_1.jpg

The CA attention mechanism can be made up of two steps: Coordinated information embedding and Coordinated attention generation. The specific implementation is as follows:

①Coordinate message embedding

As for the input feature map FRC×H×W, the average pooling operation using the pooling kernels of size (H,1) and (1, W) along the width and height directions respectively is performed so that the feature maps 00104_PSISDG12779_127791S_page_4_2.jpg and 00104_PSISDG12779_127791S_page_4_3.jpg in the width and height directions are obtained, which are calculated as shown in Equation (1).

00104_PSISDG12779_127791S_page_4_4.jpg

②Coordinate Attention generation

The feature maps which are encoded in the two directions of width and height are concatenated by the transformation of the information embedding step.Subsequently, the dimensionality is reduced to the original C / r by performing a 1×1 convolution operation in the channel dimension.After that, the batch normalized feature map will be fed into the Sigmoid activation function which results in a feature map of dimension 1 × (W + H) × C / r, as shown in Equation (2).

00104_PSISDG12779_127791S_page_4_5.jpg

After that, the feature map f is subjected to a 1×1 convolution operation following the initial width and height so that the feature maps fh and fw with the same dimension as the initial ones are obtained. Then, they are processed by the Sigmoid activation function to obtain the attention weights gh and gw on the width and height directions, respectively, as shown in Equation (3).

00104_PSISDG12779_127791S_page_4_6.jpg

Finally, the obtained attention weights are multiplied and weighted in accordance with the primal feature map so that a feature map yc with attention weights is obtained, illustrated in Equation (4).

00104_PSISDG12779_127791S_page_4_7.jpg

3.3

SIOU loss function

The PP-YOLOE+ model is calculated by adopting the GIOU loss function as the boundary box regression loss, although GIOU adds the minimum outer rectangle as the penalty term that is between the prediction box and the real box, and solves the problem that the gradient cannot be calculated when IOU is used as the loss function.However, there are still the following drawbacks: The first one is that when the prediction box and the real box overlap, the GIOU loss becomes degraded to IOU loss, which cannot properly reflect the position of the prediction box within the real box and the height-to-width ratio of the prediction box. The second is that the gradient of the GIOU loss function also becomes very large when the intersection area of prediction box and real box is close to 0, which leads to unstable training.

Therefore, it is introduced that the SIOU loss function replaced the pre-existing GIOU loss function.The parameters associated with the SIOU loss function are shown in Figure 4. Where, Bpred denotes that the real box with the location of the center point (bcx, bcy), Bgt denotes that the real box with the location of the center point 00104_PSISDG12779_127791S_page_5_1.jpg, σ is the distance between the center of the real box and the predicted box, w and h denoted that the width and height of the predicted box, wgt and hgt denoted that the width and height of the real box, respectively. The figure contains two rectangular boxes, one is the rectangle that is formed at the center of the real box and the prediction box, which is represented by the dashed box in Figure 4, and whose height and width differences are denoted as ch and cw, respectively; In the other one is the outer solid wireframe, which represents a minimum outer rectangle enclosing the real and predicted boxes, its height and width differences are denoted as ch’ and cw’, respectively.

Figure 4.

Schematic diagram showing the parameters of the prediction box and the real box

00104_PSISDG12779_127791S_page_5_2.jpg

The SIOU loss function that is made up of four cost functions, and they are: angle cost, distance cost, shape cost and IoU cost.

(1) Angle cost. Its calculation formula are shown in (5).

00104_PSISDG12779_127791S_page_5_3.jpg

Among them

00104_PSISDG12779_127791S_page_5_4.jpg
00104_PSISDG12779_127791S_page_5_5.jpg
00104_PSISDG12779_127791S_page_5_6.jpg

Where x is the sine function of angle α. The angle α can be obtained by taking the inverse function of the sine function. In the training process, if 00104_PSISDG12779_127791S_page_5_7.jpg, then choose to minimize α, otherwise minimize β. Note that the angle loss will become 0 when the angle α is equal to 00104_PSISDG12779_127791S_page_5_8.jpg or 0.

(2) Distance cost. In redefining the distance cost, the angle cost is to be considered, which is calculated by the formula shown in (9).

00104_PSISDG12779_127791S_page_6_1.jpg

Among them

00104_PSISDG12779_127791S_page_6_2.jpg

From this equation, it can be seen that the effect of distance cost is greatly reduced when the angle α tends to 0 degrees, and becomes more pronounced when the angle α tends to 90 degrees.

(3) Shape Cost. Its calculation formula is shown in (11).

00104_PSISDG12779_127791S_page_6_3.jpg

Among them

00104_PSISDG12779_127791S_page_6_4.jpg

where the role of the θ parameter as a factor that controls the concern for shape cost, avoiding paying excessive attention to shape loss, and reducing the movement of the prediction box, its value range is [2, 6].

(4) IoU cost. Its calculation formula is shown in (13).

00104_PSISDG12779_127791S_page_6_5.jpg

In conclusion, ultimately the SIOU loss function can be defined in Equation (14).

00104_PSISDG12779_127791S_page_6_6.jpg

4.

EXPERIMENTS

4.1

Dataset source and production

For the training model, the dataset adopted is the OPIXray dataset12 released by the Software Development Key Laboratory for Environment of Beijing University of Aeronautics and Astronautics in 2019.This dataset has a total of 8,885 X-ray images and 5 different knife categories. However, since the only contraband category labeled in this dataset is knives, the detection categories are single and limited, which is not suitable for the application.

Therefore, the following processing is taken to expand the dataset: in the first step, other contraband categories in this dataset are labeled using the labeling tool labelImg, including categories such as laptops, liquids, and cell phones; in the second step, considering the diversity of contraband categories in life, from the HiXray dataset13 and the PIDray dataset14, water, rechargeable batteries, cell phones, laptops, guns, lighters and aerosols. The new dataset, named OPI_MIXray, contains 12 categories of contraband and 30,676 images (training set:validation set:test set=8:1:1).

4.2

The experimental environment and hyperparameter settings

  • (1) Experimental environment: This paper uses AI Studio of Baidu Flying Pulp as the experimental platform, the hardware device has 32GB of video memory and memory capacity, 100GB of hard disk storage space, and the framework uses PaddlePaddle 2.4.0.

  • (2) Hyperparameter settings: The weight is updated via the adoption of a momentum optimizer and the initial momentum value that is 0.9. Learning rate that is set to 0.005 and dynamically adjusted using a cosine annealing learning rate decay and a linear preheat learning rate strategy with the first 5 rounds as preheat rounds.Batchsize is to be set to 32 and Epoch is to be set to 300. After each 10 epochs in training, it is evaluated on the validation dataset and the optimal weights are saved. and save the optimal weights.

  • (3) Deployment test environment: Jeston Nano B01 suite with quad-core ARM A57 processor, 128-core MAXWELL GPU and 4GB LPD-DR memory, with sufficient AI computing power and model deployment framework of Paddle-Inference 2.4.1.

4.3

Model deployment

The model deployment framework uses Paddle-Inference inference framework. The framework is Baidu’s open source high-performance inference engine based on the PaddlePaddle training model, which supports embedded devices such as x86 and ARM. The deployment process is shown in Figure 5. First, Paddle-Inference 2.4.1 is installed on the embedded device Jeston Nano, the trained and improved model is converted into a model deployment file and downloaded to Jeston Nano, and finally the initialization of the deployment file is completed through the API functions provided by the Paddle-Infernece framework, the model is invoked and inference is performed The inference result is obtained, and the location detection and classification of contraband is completed.

Figure 5.

Model deployment inference flow chart

00104_PSISDG12779_127791S_page_7_1.jpg

4.4

D-Res module introduces a comparison of location and attention mechanisms

In this paper, the CSPRes modules of C3, C4 and C5 stages of the backbone network output feature maps are replaced by D-Res modules, and the results under mAP@0.5 are all improved to different degrees than the baseline model, among which the C4 stage has the best results, it is shown in Table 1.

Table 1.

D-Res module and attention mechanism experiments

D-Res moduleC3C4C5Baseline
mAP@0.5/%90.290.489.888.6
Attentional MechanismsCAECACBAMBaseline
mAP@0.5/%90.989.790.488.6

For the attention mechanism, we introduce ECA, CBAM and CA attention mechanisms respectively in this paper, which all have significantly improved the detection accuracy, among them the CA attention mechanism has the best effect, as shown in Table 1. Therefore, we choose to embed CA attention mechanism in the network in this paper, and replace the C4 stage CSPRes module with D-Res module at the output feature map of the backbone network, and adopt SIOU loss function to derive PP-YOLOE+_DCS model.

4.5

Ablation experiments

The ablation of experiments in this paper as shown in Table 2. Where D represents the D-Res module, C represents the CA attention mechanism, and S represents the SIOU loss function. The experimental results demonstrate that all three improvement methods proposed in this paper can improve the detection precision to different degrees compared with the baseline model PP-YOLOE+_S. By introducing the D-Res module, it makes the convolutional kernel better adaptable to contraband of different sizes, shapes and poses. By introducing the CA attention mechanism, it enables the network to acquire the context i southwest west of the region around the contraband in the security check image and reduce the influence of factors such as background and noise. By the replacement of the SIOU loss function, it reduces the localization error between the prediction box and the real box, which improves the detection accuracy and training speed of the model. Finally, with the combination of the three modules, mAP improves by 2.8% compared to the baseline model with only 0.24M additional number of parameters, and FLOPs and Latency both decrease to some extent.

Table 2.

The ablation experiment of the improved model

ModelmAP@0.5/%Params(M)FLOPs(G)Latency(ms)
Baseline88.67.6716.4430.8
+D90.47.8015.6410.7
+C90.97.7816.6450.3
+S90.37.6716.4432.2
+DC90.77.9115.8419.3
+DCS91.4(+2.8)7.91(+0.24)15.8(-0.6)420.2(-10.6)

4.6

Comparison experiment

To further verify the improvement model PP-YOLOE+_DCS performance presented in this paper, the above model was selected for comparison experiments and also compared with other models of PP-YOLOE+ to better understand the relationship between model performance and scale. It can be seen from Table 3 that the improved method PP_YOLOE+_DCS introduced in this paper has reduced its inference delay by 10.6ms compared with that before the improvement, and good detection accuracy has been achieved in mAP@0.5:0.95 and mAP@0.5. This indicates that the improved method presented in this paper helps to improve the feature extraction ability of the model and makes it more capable of handling multi-scale and multi-attitude contraband. The improved model satisfies detection accuracy comparable to that of the complex model, making it more friendly for deployment on resource-limited devices and of practical application value.

Table 3.

Results of experiments with different models

ModelOPI_MIXray
mAP@0.5/%Params(M)FLOPs(G)Latency(ms)
Faster R-CNN75.3138.6560.41586.6
SSD62.32.552.976.2
YOLOv5_S87.27.0516.0420.3
YOLOX_S88.58.9426.7701.4
YOLOE+_S88.67.6716.4430.8
YOLOE+_M91.623.5249.421298.1
YOLOE+_L92.453.26113.42978.8
YOLOE+_X91.1101.30211.55556.7
Ours91.4(+2.8)7.91(+0.24)15.8(-0.6)420.2(-10.6)

4.7

Visualization of test results

Figure 6 shows the detection results for a partial test set (IoU=0.5). Among them, (a) and (c) show the detections results that are from the original PP-YOLOE+_S model, and (b) and (d) show the detections results that are from the improved PP-YOLOE+_DCS model. From the figures, it is observed that some targets with unclear shapes and contours are missed by the original model. And using the improved model, the missed targets can be detected (marked with red boxes). Therefore, the improved model PP-YOLOE+_DCS submitted in it can better capture the feature information of contraband and improve the recognition and localization of contraband.

Figure 6.

Comparison chart of test results

00104_PSISDG12779_127791S_page_9_1.jpg

5.

CONCLUSION

In order to better help security personnel detect contraband, we designed a method that is based on PP-YOLOE+_S for detecting prohibited items in X-ray security images in this paper. Targeted improvements are made to address the characteristics of prohibited items in the X-ray security screening images. DCNv2 is introduced in the backbone network and designed to obtain the D-Res module, which is used in C4 stage, so that it can enhance the accuracy of the model by better capturing the detailed information of prohibited items in X-ray security images with no increased computational effort. Moreover, we also introduce the CA attention mechanism of the backbone network and feature fusion network to further Improve the model’s ability to locate prohibited items and extract critical characteristics. Finally, we replace the original of GIOU loss function with SIOU loss function to reduce the localization error and improve the detection accuracy. Through a series of tests, this paper designs a model that can improve the efficiency and accuracy of contraband recognition, which has engineering significance for the automation and intelligence of security image contraband detection.

ACKNOWLEDGMENTS

This research project was supported by the Education Department of Jilin Province (“Thirteen Five” Scientific Planning Project, Grant No.JJKH20180898KJ), Jilin Education Science Planning Project (“ThirteenFive” Plan, Grant No. GH170043), School-Enterprise Cooperation Program of Yanbian University. (No.YDXQ202301)

REFERENCES

[1] 

Akcay, S., Breckon, T., “Towards automatic threat detection: A survey of advances of deep learning within X-ray security imaging,” Pattern Recognition, 122 108245 (2022). https://doi.org/10.1016/j.patcog.2021.108245 Google Scholar

[2] 

Wang, R., Shi, Y., Cai, M., “Optimization and Research of Suspicious Object Detection Algorithm in X-ray Image,” in 2023 IEEE 6th Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), 1357 –1361 (2023). Google Scholar

[3] 

Ma, C., Zhuo, L., Li, J. et al., “Occluded prohibited object detection in X-ray images with global Context-aware Multi-Scale feature Aggregation,” Neurocomputing, 519 1 –16 (2023). https://doi.org/10.1016/j.neucom.2022.11.034 Google Scholar

[4] 

Zhou, C., Xu, H., Yi, B., Yu, W., Zhao, C., “X-ray security inspection image detection algorithm based on improved YOLOv4,” in 2021 IEEE 3rd Eurasia Conference on IOT, Communication and Engineering (ECICE), 546 –550 (2021). Google Scholar

[5] 

Ren, Y., Zhang, H., Sun, H., et al., “LightRay: Lightweight network for prohibited items detection in X-ray images during security inspection,” Computers and Electrical Engineering, 103 108283 (2022). https://doi.org/10.1016/j.compeleceng.2022.108283 Google Scholar

[6] 

Liu, D., Liu, J., Yuan, P., et al., “Lightweight prohibited item detection method based on YOLOV4 for x-ray security inspection,” Applied Optics, 61 (28), 8454 –8461 (2022). https://doi.org/10.1364/AO.467717 Google Scholar

[7] 

Song, B., Li, R., Pan, X., Liu, X., Xu, Y., “Improved YOLOv5 Detection Algorithm of Contraband in X-ray Security Inspection Image,” in 2022 5th International Conference on Pattern Recognition and Artificial Intelligence (PRAI), 169 –174 (2022). Google Scholar

[8] 

Xu, S., Wang, X., Lv, W., et al., “PP-YOLOE: An evolved version of YOLO,” (2022). Google Scholar

[9] 

Zhu, X., Hu, H., Lin, S., Dai, J., “Deformable convnets v2: More deformable, better results,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 9308 –9316 (2019). Google Scholar

[10] 

Hou, Q., Zhou, D., Feng, J., “Coordinate attention for efficient mobile network design,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 13713 –13722 (2021). Google Scholar

[11] 

Gevorgyan, Z., “SIoU Loss: More Powerful Learning for Bounding Box Regression,” (2022). Google Scholar

[12] 

Wei, Y., Tao, R., Wu, Z., Ma, Y., Zhang, L., Liu, X., “Occluded prohibited items detection: An x-ray security inspection benchmark and de-occlusion attention module,” in Proceedings of the 28th ACM International Conference on Multimedia, 138 –146 (2020). Google Scholar

[13] 

Tao, R., Wei, Y., Jiang, X., Li, H., Qin, H., Wang, J., et al., “Towards real-world X-ray security inspection: A high-quality benchmark and lateral inhibition module for prohibited items detection,” in Proceedings of the IEEE/CVF international conference on computer vision, 10923 –10932 (2021). Google Scholar

[14] 

Wang, B., Zhang, L., Wen, L., Liu, X., Wu, Y., “Towards real-world prohibited item detection: A large-scale x-ray benchmark,” in Proceedings of the IEEE/CVF international conference on computer vision, 5412 –5421 (2021). Google Scholar
© (2023) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Fan Chen, Hua Jin, and Qinghan Li "Design of security image contraband detection system based on PP-YOLOE+_DCS", Proc. SPIE 12779, Seventh International Conference on Mechatronics and Intelligent Robotics (ICMIR 2023), 127791S (11 September 2023); https://doi.org/10.1117/12.2688728
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Object detection

Performance modeling

Education and training

X-ray imaging

X-rays

Instrument modeling

Network security

RELATED CONTENT


Back to Top