Soccer player and ball detection and tracking have emerged as an area of intense interest among many analysts and researchers. This is because it aids coaches in team performance evaluation and decision-making to achieve optimal results. However, existing methodologies have failed to effectively detect and track the ball when it moves at high velocity and also to track players under occlusion conditions. You only look once (YOLOv3) and simple online real-time (SORT)-based soccer ball and player tracking approach is proposed, for accurately classifying the detected objects in soccer video and track them in various challenging situations. The proposed methodology consists of two parts: (i) YOLOv3 can detect and classify the objects (i.e., player, soccer ball, and background) and eliminate the detected objects outside the playfield as background; (ii) tracking is achieved using SORT algorithm which employs a Kalman filtering and bounding box overlap. The proposed model achieves tracking accuracy of 93.7% on multiple object tracking accuracy metrics with a detection speed of 23.7 frames per second (FPS) and a tracking speed of 11.3 FPS.
Availability of humungous visual data and increasing in generation of visual data in Security and Surveillance domain made a pathway to Computer Vision algorithms. The existing algorithms are not precise enough for predictive analytics. Sensitive use cases such as action recognition and identifying missing people in huge crowds has thrown a challenging research of drawing accurate and precise results. The existing 2-D plots for action recognition have failed due to unstructured visual data available where the accuracy is around <50%. Due to unstructured visual data, the existing 3-D plots often get overlapped with each other. Although the accuracy is noted >90% which maps it to False Positives. The existing solutions deals with object detection through Boolean logic then Pose Plots are mapped. Our research focus in on reverse engineer the existing solutions by applying smart segmentation to isolate background and then map the pose formula to detect the action. Our proposed solution obliterates the over-lap complications and unravels the False Positives. Our proposed solution achieved accuracy and precision of mAP>0.8 for both images and video feeds.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.