Optimized deep learning-based cricket activity focused network and medium scale benchmark

The recognition of different activities in sports has gained attention in recent years for its applications in various athletic events, including soccer and cricket. Cricket, in particular, presents a challenging task for automatic activity recognition methods due to its closely overlapped activitie...

Full description

Bibliographic Details
Main Authors: Waqas Ahmad, Muhammad Munsif, Habib Ullah, Mohib Ullah, Alhanouf Abdulrahman Alsuwailem, Abdul Khader Jilani Saudagar, Khan Muhammad, Muhammad Sajjad
Format: Article
Language:English
Published: Elsevier 2023-07-01
Series:Alexandria Engineering Journal
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S1110016823003368
Description
Summary:The recognition of different activities in sports has gained attention in recent years for its applications in various athletic events, including soccer and cricket. Cricket, in particular, presents a challenging task for automatic activity recognition methods due to its closely overlapped activities such as cover drive, and pull short, to name a few. Existing methods often rely on hand-crafted features as the limited availability of public data has restricted the scope of research to only the significant categories of cricket activities. To this end, we proposed a cricket activities dataset and an intuitive end-to-end deep learning model for cricket activity recognition. The data is collected from online sources and pre-processed through cleaning, resizing, and organizing. Similarly, an intuitive deep model is designed with a combination of time-distributed 2D CNN layers and LSTM cells for extracting and learning the spatiotemporal information from the input sequences. For benchmarking, we evaluated the model on our cricket datasets and four standard datasets namely UCF101, HMDB51, YouTube action, and Kinetics. The quantitative results show that the proposed model outperforms different variants of recurrent neural networks and achieved an accuracy of 92%, recall of 91%, and F1 score of 91%. Our code and dataset is publicly available for further research on https://drive.google.com/file/d/1c9qcAz4q00qvx4yFA3pSudWFczm1cWUL/view?usp=sharing.
ISSN:1110-0168