Welcome to the SoccerNet Development Kit for the Team Ball Action Spotting challenge. This challenge is an evolution of the original Action Spotting and Ball Action Spotting tasks, which are described in more detail here.
⚽ What is Action Spotting?
Action spotting involves the identification and precise localization of actions within an untrimmed video. Given a video input, the objective is to detect and accurately locate all the actions occurring throughout the video.
- The original SoccerNet Action Spotting task focused on identifying 17 sparse actions across soccer broadcasts, spanning a total of 550 football games. Key features included sparse annotations and a looser evaluation metric.
- The SoccerNet Ball Action Spotting task introduced a new set of soccer videos, with denser annotations for 12 distinct actions. A stricter metric was also employed, with tolerances of just up to 1 second.
- The SoccerNet Team Ball Action Spotting task is a direct extension of the Ball Action Spotting task. It uses the same games and actions but adds the challenge of identifying which team (left or right in the video) performed the action.
In this repository, we focus on the latest specification of the task, using an adaptation of the SoccerNet 2024 Ball Action Spotting challenge winner model, T-DEED, as our baseline. The evaluation metric is also adapted to assess whether methods correctly predict the team performing the action. Further details on this adaptation are provided below.
As mentioned earlier, we use the T-DEED model as the baseline for our task. T-DEED is a Temporal-Discriminability Enhancer Encoder-Decoder designed for end-to-end training with the goal of improving token discriminability. Further details about the model can be found in the original paper and in its dedicated repository. In this repository, we provide the adaptation of T-DEED for the Team Ball Action Spotting task.
To adapt T-DEED for predicting the team performing the action, we add an additional prediction head with a sigmoid activation function, and incorporate a team loss during training based on Binary Cross-Entropy loss. T-DEED’s evaluation in the SoccerNet 2024 Ball Action Spotting challenge benefited from joint training on both the SoccerNet Action Spotting and SoccerNet Ball Action Spotting datasets. We follow the same approach here to leverage the larger original dataset, which helps in learning the actions for Ball Action Spotting.
-
Download the datasets and labels: Follow the instructions provided in this link to download the SoccerNet Action Spotting and SoccerNet Ball Action Spotting datasets, including videos and labels. While the labels and SoccerNet Ball Action Spotting videos can be downloaded directly from the link, the instructions for downloading SoccerNet Action Spotting videos can be found on this webpage. If you have labels from earlier versions of the challenge, ensure that you update them, as the new versions include team annotations for each action.
-
Update label paths:
Modify thelabels_path.txt
files in/data/soccernet
and/data/soccernetball
to point to the folders containing the labels for each dataset. -
Extract frames:
Extract frames for both SoccerNet Action Spotting and SoccerNet Ball Action Spotting using theextract_frames_sn.py
andextract_frames_snb.py
scripts. -
Update the config file:
Update the config file to set the paths for frame directories (for both datasets), and checkpoint and predictions saving locations. For details on the configuration parameters, refer to the corresponding README. -
Run initial training setup:
Execute thetrain_tdeed_bas.py
script, specifying the model name corresponding to the configuration file (SoccerNetBall_baseline
in this case). For the first run, set thestore_mode
parameter in the configuration to "store". This will partition the untrimmed videos into clips and store information such as the starting frame and processed labels, enabling faster data loading during training. -
Train the model:
After the initial setup, runtrain_tdeed_bas.py
again withstore_mode
set to "load" to train the model. The script will train on thetrain
splits, use thevalidation
splits for early stopping, evaluate on thetest
split, and generate predictions for thechallenge
split.
To evaluate this task, we use a modified version of the mAP (mean Average Precision) metric that has been employed in previous Action Spotting challenges. In this adaptation, we compute the AP (Average Precision) for each combination of action (12 different actions) and team (left or right). The final AP for each action is calculated as the weighted average between both team sides, with the weights determined by the number of ground-truth observations. Lastly, the overall Team mAP metric is obtained by averaging the AP values across all 12 actions. As in the previous Ball Action Spotting challenge, a tolerance of 1 second is applied.
You can check /util/eval.py and /util/score.py for the implementation of the evaluation metric. You can evaluate your models on the test split using this repository, or submit your predictions for both the test and challenge splits to the relevant competitions (test and challenge) on Codabench. Additional details regarding the structure of the submission file can be found on the respective Codabench competition pages.
Test split | Challenge split | |||
---|---|---|---|---|
Method | Team mAP@1 | mAP@1 | Team mAP@1 | mAP@1 |
T-DEED (baseline) | 47.18 | 53.59 | 51.72 | 58.38 |
Here, we present the results of the provided T-DEED baseline. Challenge yourself by pushing the limits of your methods and participating in our upcoming 2025 Challenges. The official rules and submission guidelines can be found in ChallengeRules.md. You can submit your predictions to the Codabench evaluation server for both the test and challenge competitions.
A checkpoint of the baseline is available at the following link, allowing you to perform inference using the provided T-DEED model. Please note that results obtained with this checkpoint may differ slightly from the reported results due to variations in different training runs.
This repository is built upon the foundation of the E2E-Spot codebase, and we would like to extend our gratitude for their work.
If you use this repository for your research or wish to refer to our contributions, please use the following BibTeX entries.
@InProceedings{Deliège2020SoccerNetv2,
title={SoccerNet-v2 : A Dataset and Benchmarks for Holistic Understanding of Broadcast Soccer Videos},
author={Adrien Deliège and Anthony Cioppa and Silvio Giancola and Meisam J. Seikavandi and Jacob V. Dueholm and Kamal Nasrollahi and Bernard Ghanem and Thomas B. Moeslund and Marc Van Droogenbroeck},
year={2021},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
}
@article{cioppa2024soccernet,
title={SoccerNet 2023 challenges results},
author={Cioppa, Anthony and Giancola, Silvio and Somers, Vladimir and Magera, Floriane and Zhou, Xin and Mkhallati, Hassan and Deli{\`e}ge, Adrien and Held, Jan and Hinojosa, Carlos and Mansourian, Amir M and others},
journal={Sports Engineering},
volume={27},
number={2},
pages={24},
year={2024},
publisher={Springer}
}
@inproceedings{xarles2024t,
title={T-DEED: Temporal-Discriminability Enhancer Encoder-Decoder for Precise Event Spotting in Sports Videos},
author={Xarles, Artur and Escalera, Sergio and Moeslund, Thomas B and Clap{\'e}s, Albert},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
pages={3410--3419},
year={2024}
}