XI INTERNATIONAL CONFERENCE
“INFORMATION TECHNOLOGY AND IMPLEMENTATION” (IT&I-2024)
Enhancing Object Detection and Classification in High-Resolution Images Using SAHI Algorithm and Modern Neural Networks
Oleksii Bychkov, Kateryna Merkulova, Yelyzaveta Zhabska and Andrii Yaroshenko
Fundamental Theoretical Aspects
The Slicing Aided Hyper Inference (SAHI) algorithm is a key component of the proposed
approach for object detection and classification in high-resolution images. The main idea behind SAHI
is to divide the large input image into smaller, overlapping patches, which can then be processed
independently by the object detection models. This approach has several advantages over traditional
methods that rely on resizing or cropping the image to fit the input size of the neural network.
The SAHI algorithm consists of the following steps:
1. Slicing:
2. Inference:
Fundamental Theoretical Aspects (cont.)
3. Merging:
4. Post-processing:
Non-Maximum Suppression (NMS)
It removes redundant bounding boxes, retaining only the most confident detections for each object. This is especially important in sliding window approaches, like SAHI, where overlapping patches can generate multiple detections for the same object.
Fundamental Theoretical Aspects (cont.)
How NMS Works:
B1, B2 — bounding boxes of detections
High IoU indicates significant overlap, suggesting the same object.
We conducted experiments to assess the performance of the SAHI algorithm combined with five object detection models. The evaluation considered several factors:
Key Variables:
Experimental Evaluation
Metrics Evaluated:
Experimental Evaluation
A large beach panorama was selected to evaluate the detection of small objects such as people, cars, and boats. The image details are as follows:
Initial Observations:
SAHI Implementation:
Tile Size: Multiple sizes tested (256×256, 512×512, and 1024×1024 pixels).
Overlap Ratio: Assessed at 25% and 50%.
Case Study: Beach Panorama Analysis
Studied Image Preview
Case Study: Beach Panorama Analysis
Experiment Results: Beach Panorama Detection
Figure 1 - results for patch size 512x512 and overlap ratio of 50%.
Figure 2 - results of various combinations of patch sizes,
overlap ratios and detection models with true value
NMS Impact on Object Detection
Figure 3 - detections with NMS on the left and without on the right
Left (With NMS):
Right (Without NMS):
This highlights the effectiveness of NMS in reducing redundant detections and improving the clarity of object detection outcomes.
To validate detection accuracy, the approximate number of people in the photo was manually counted at ~1,100. The experiment evaluated combinations of neural networks and post-processing methods, with key observations as follows:
General Observations:
Experiment Results: Beach Panorama Detection
Figure 4 — results for GREEDYNMM
Experiment Results: Beach Panorama Detection
This table summarizes the results of all previous observations, evaluating:
The table highlights the trade-offs between accuracy and computational performance across different configurations.
Neural Network Comparison:
These results (Figure 2) highlight the strengths and limitations of each network and post-processing combination in detecting small objects.
Experiment Results: Beach Panorama Detection
For detailed results and additional experiments, including:
Please refer to the full paper for more in-depth insights.
Further Experiments
This study proposed a novel framework for object detection in high-resolution images, combining the SAHI algorithm with state-of-the-art neural networks (YOLOv5, YOLOv8, YOLOX, Torchvision, and RetinaNet). Key contributions and findings include:
Key Contributions:
Conclusion
Real-World Applications:
Future Directions:
This research provides a robust, efficient framework for object detection in high-resolution images, paving the way for accurate and reliable systems across diverse applications.
Conclusion
Department of Software Systems and Technologies�Taras Shevchenko National University of Kyiv�Kyiv, UKRAINE
Oleksii Bychkov�bos.knu@gmail.com
Kateryna Merkulova �kate.don11@gmail.com
Yelyzaveta Zhabska�y.zhabska@gmail.com�
Andrii Yaroshenko
andrii.yaroshenko@knu.ua
Thank you for attention!