Predicting the winner of a match is a well-established research topic in sports (see reviews Horvat & Job, 2020; Wunderlich & Memmert, 2021), i.e., in tennis (Kovalchik, 2016), American football (Purucker, 1996), horse racing (Davoodi & Khanteymoori, 2010), basketball (Cao, 2012) or golf (Chae, Park, & So, 2021). Predicting match outcomes is crucial for setting betting odds, researching the nature of a particular sport and for conducting performance analysis (Wunderlich & Memmert, 2021). Particularly in association football (soccer), extensive research exists seeking to predict the exact score (Angelini & Angelis, 2017), the spread (Karlis & Ntzoufras, 2009) or the outcome in terms of a win, draw, and lose (Baboota & Kaur, 2019; Beal, Middleton, Norman, & Ramchurn, 2021; Goddard & Asimakopoulos, 2004). Methodologies employed range from Poisson distribution probabilistic classifiers (Goddard, 2005), Bayesian networks (Owramipur, Eskandarian, & Mozneb, 2013) and Random Forests (Schauberger & Groll, 2018) to Artificial Neural Networks (Igiri & Nwachukwu, 2014).
While existing studies predict match outcomes, researchers could also address the question of what will happen next during a match – whether a team will make the next goal, shot attempt, or corner kick. These types of predictions have not yet been fully explored in research but hold substantial relevance for the betting market, where a wide range of in-play betting options generates significant financial turnover (Jackson, 2015; Killick & Griffiths, 2019; Parke & Parke, 2019). In-play predictions can complement traditional outcome predictions, offering finer granularity and thereby improving forecasting accuracy and scope. Existing studies in inplay forecasting focus on various prediction goals. e.g., some aim to predict the outcome of the match while it is ongoing (Arntzen & Hvattum, 2021; Dobson & Goddard, 2017; Robberechts, van Haaren, & Davis, 2019; Yao, Wang, Zhu, Cao, & Zeng, 2022) or forecast whether a specific shot will result in a goal (Lucey, Bialkowski, Monfort, Carr, & Matthews, 2015). Other approaches utilize imitation learning techniques to predict player movement trajectories (Lindström, Jacobsson, Carlsson, & Lambrix, 2020), using sentiment analysis techniques for goal prediction (Wunderlich & Memmert, 2022) or estimate the likelihood of the next event being a goal (Decroos & Davis, 2019). However, to the best of our knowledge, there are no studies that predict in-play match events beyond the immediate next event.
This article investigates to what extent goals and other success- or scoring-related match events (SREs) in the near future, e.g., shots taken, corner kicks, or box entries, can be predicted using prior performance data from the near past. There is good reason to believe that these predictions can be made utilizing the concept of match momentum (Veo, 2021; Whitmore, 2021), also known as game flow (Ho, McKinley, & Moore, 2018) or dominance (Link, Lang, & Seidenschwarz, 2016), representing the actual performance difference between the two teams. Hence, match momentum can reveal a performance streak for one team if it is on its side, which might result in achieving more SREs and therefore a higher probability of this team scoring. Meanwhile, psychological momentum (Iso-Ahola & Dotson, 2016; Iso-Ahola & Mobily, 1980), or often just momentum, is difficult to measure in soccer (Redwood-Brown et al., 2018), although it is somewhat related to match momentum and vice versa. According to Higham et al. (2005, p. 3), momentum in football is described as a “hidden force [sic]” in a match, flowing between the two teams and representing the energy between the competitors. Momentum is also elaborated in other game sports such as American football (Roebber, Burlingame, & deWinter, 2022), basketball (Arkes & Martinez, 2011) and ice hockey (Leard & Doyle, 2011). We argue that match momentum represents the evolving performance difference between teams over a short period, which can be captured through in-game performance indicators (PIs). This concept aims to link past in-game performance trends to near-future outcomes.
In light of the above, the question arises of which in-play PIs or PI-combinations best predict SREs in general and thus represent the actual team performance. Answering this question also contributes to the discussion of the validity of PIs and to the theory of football (Carling, Wright, Nelson, & Bradley, 2014; Mackenzie & Cushion, 2013). A valid PI should be stable in terms of a small random fluctuation (Davis et al., 2024; Lames & McGarry, 2007) and be associated with a component of success (Hughes & Bartlett, 2002). When a PI demonstrates its validity for (short-term) success, then it is feasible for (in-play) analytical questions (Lames & McGarry, 2007), e.g., to measure the effectiveness of a tactical intervention such as a substitution, change of pressing strategy, or style of play. Beside other techniques (Davis et al., 2024), research suggests using the PIs’ predictive power for success or its components to compare and evaluate them (Herold, Kempe, Bauer, & Meyer, 2021; Hughes & Bartlett, 2002). Recent years have seen some studies on this, but usually on seasonal or match-level (Russomanno, Linke, Geromiller, & Lames, 2020) or with a different methodology than ours (Davis et al., 2024; Decroos & Davis, 2019; Herold et al., 2021), but not for the short-term predictive power as we suggest. Regarding PIs and their value, spatiotemporal player tracking data-based PIs are relatively good predictors, e.g., for goal-scoring opportunities (Wagenaar, Okafor, Frencken, & Wiering, 2017), match outcome (Goes, Kempe, & Lemmink, 2019) and the outcome of the second half (Klemp, Wunderlich, & Memmert, 2021). Others used eventrelated PIs to predict match outcomes (Cintia, Giannotti, Pappalardo, Pedreschi, & Malvaldi, 2015), to identify PIs that discriminate between winning and losing teams (Harrop & Nevill, 2014), rate single actions performed by players (Liu, Luo, Schulte, & Kharrat, 2020) or predict the next event with information from the last actions (Simpson, Beal, Locke, & Norman, 2022; Wang et al., 2024). While some of these studies share similar goals to ours, we focus on shortterm event prediction and exclusively use in-play PI data to predict the occurrence of a special event in the next few minutes.
Given this context, this study aims to explore the applicability of common PIs in soccer to predict SREs. To achieve this, we test to what extent our prediction goal (PG), i.e. the occurrence of an SRE, in a defined future time span (prediction window) can be predicted for one team by using the values of in-play PIs (e.g., passes, time of ball possession, opponents outplayed) within a defined time span from the past (input window). For prediction, we employ five different machine learning models (MLM) for multiple window sizes and combinations. The PIs are based on event and spatiotemporal tracking data from 102 German Bundesliga matches. Consequently, our study is organized in four parts: First, (Part I) we evaluate the inter-correlation between PIs and PGs in matches, quantifying their relationship. Second, (Part II) we explore the performance of different MLMs and rank the PIs based on their mean predictive power for our PGs to quantify their general usability. Subsequently, (Part III) we test the prediction performance of different PI-combinations for one selected scenario (predicting a goal in the next three minutes) in order to evaluate the interaction of the PIs. Finally, (Part IV) we propose a real-world application utilizing all findings for a match momentum metric.
Our research holds significance for both the betting industry and performance analysis. It addresses questions regarding the relationship between input window size and prediction performance, as well as identifying which PI or PI-combinations can effectively utilized as in-game team performance representation. The results can be employed to enhance in-play betting odds and contribute valuable insights to practitioners regarding the validity and usability of PIs for match analysis and real-time decision-making.
Data for our study comprises 102 matches of German professional football leagues, encompassing 82 Bundesliga and 20 Bundesliga 2 matches played across 31 rounds in the 2017/2018 season. Each match record contains three types of data: positional data of players (x/y coordinates), and ball (x/y/z) captured with 25 hertz; event data; and basic match information, e.g., player/team names, and playing positions. Positional data was collected semi-automatically using the TRACAB optical tracking system (ChyronHego, NY), while in-game events were manually recorded by human observers (DFL, 2020). The eight-centimeter accuracy of the tracking system was validated by Linke et al. (2020). All 36 teams are represented with a minimum and maximum of two to twenty matches. Each team had home and away matches, except one team which had only away matches available. For each match and each team, we calculated PIs (see Parametrization). We aggregated this data in intervals to use it as input data (see Motivation and Design). To ensure sufficient SREs in both the validation and test sets, we randomly split the dataset on match-level into training, validation, and test sets with a 60:20:20 ratio. This split also helps ensure robust evaluation across a diverse range of matches and scenarios. For reproducibility and testing purposes, 102 synthetic datasets are provided in the associated GitHub repository available at https://github.com/SteffenLa/Shortterm_Soccer_Event_Prediction.git.
Our study aims to predict SREs based on information from the running game (e.g., whether a goal will be scored by a team in the next few minutes using the information given from the previous minutes). Identifying PIs capable of predicting such events can help to validate their applicability as team performance representation in football; however, finding suitable PIs for this type of prediction is challenging. Some PIs have different timescales and characteristics, which makes it difficult for MLMs to utilize them effectively. For instance, the well-known PI Shots Taken can be useful to predict goals in a match, but occurs infrequently and is only a good predictor for the immediate future (Mead, O’Hare, & McMenemy, 2023). These characteristics seem to make this PI unsuitable for quantifying team performance for more than a specific moment. Similarly, numerous shots in a shorter timespan could be included in this metric. Therefore, we propose a windowing approach to address the issue of overvaluing a PI’s immediate prediction quality by MLMs (see In-Play Prediction Masking). Our approach masks data immediately before the prediction window, giving less weight to PIs that have a direct short-term relation to sparse PGs, such as shots leading to goals. This approach reduces the effect of such a PI on the MLM but still enables its utilization on a longer time scale.
Another challenge is that some individual PIs may not have a good prediction quality or are designed for different prediction time spans (e.g., seconds vs. minutes). However, combining those PIs with others may result in a higher prediction accuracy than their individual contributions. Unfortunately, exploring all of them in-depth is computationally infeasible due to the number of PIs, PGs, MLMs and their specific hyperparameters. Even with a large dataset, models can be over-fitted and may not provide a general answer to which PIs are suitable for representing team performance. To tackle this problem, we structured our experiments to focus on gaining insights into the PIs. By evaluating the average performance of an individual PI on predicting various SREs (Part II) employing default-configured MLMs, similar to existing research (Decroos & Davis, 2019; Sipper, 2022), we can better understand the PIs utility and which MLM can most effectively leverage the PI. We do not advocate for or against a particular MLM, as the performance of most MLMs can be improved through experimental tuning and investing time. Instead, we propose PIs that can be easily leveraged to increase performance to benefit a variety of PGs and MLMs.
After this in Part III, we perform an extensive experimental evaluation of PI-combinations for one selected PG, filter them to a sufficiently small set to obtain a complete combinatorial picture, and run cross-validation to verify their stability with respect to the specific dataset distribution. Note, PI-combinations refer to the use of two or more PIs as input features, without creating new PIs.
To address the confounding issues between PIs and PGs, we propose a simple way to mask the data, called In-Play Prediction Masking. Figure 1 illustrates the design of our approach, which defines a single sample based on its input window (IW), hidden window, and prediction window (PW). Given that a game interruption can last several minutes and most PIs are either unreliable or unavailable during this period, we only consider the effective playing time for the input window, usually about 55 minutes in a match (Siegle & Lames, 2012). The IW consists of the PI values sampled every 5 seconds, with varying window lengths, followed by the hidden window, neither belonging to the IW nor to the PW. The trained model predicts whether an SRE has occurred in the PW or not (binary classification). We use a rolling window approach, where all the windows move through the entire match with a defined step size. Applying In-Play Prediction Masking, we can assign less importance to very short-term predictors that are not useful in measuring a team’s performance over a longer time span. Instead, we concentrate on PIs that represent this performance best and can potentially anticipate future events. Our experiments have shown that this approach is crucial, allowing us to focus on high-quality predictive PIs and to develop MLMs capable of anticipating such events effectively.

Visualization of the In-Play Prediction Masking approach, the rolling window approach, and the utilized PIs; PGs; IW and PW window sizes. The lower band visualizes the rolling window, whereas all windows moved one step further. In brackets the number of configuration feature levels.
This approach incorporates various parameters, including MLMs; IW, PW, and hidden window sizes; rolling window step sizes; input data resolutions; PIs; and PGs. We selected five commonly used MLMs: Logistic Regression (LR), K-Nearest-Neighbors (KNN), Gaussian Naive Bayes (NB), Support Vector Machine (SVM), and Neural Networks (NN). An overview of the default configurations of our utilized MLMs and minor individual changes to fit the dataset can be found in the appendix (Tab. 2). Further, to provide a solid baseline, we implemented a random guessing algorithm based on average historical data. The probability used for the Random Guesser is derived from the ratio of all samples in the training set with the event versus those without the event. Statistically, we conducted a Bernoulli experiment for each prediction sample, establishing a minimum threshold that any sophisticated model should surpass to be deemed useful. The IW and PW sizes were each set to 1, 3, 5, 10, and 15 minutes of effective playing time, which covers a variety of possible time spans for measuring the actual team performance during a match (Fig. 1). We opted for a one-minute hidden window size to ensure that no confounding between the PIs and the PGs was present. The step size of the rolling window was set to the input data resolution of five seconds to create as many observations as possible. To measure the in-game performance of a team, various indicators are publicly available and discussed in research. We selected 14 feasible PIs, which can be calculated while the match is running, e.g., time of ball possession, zone entries, opponents outplayed, or Dangerousity (Link et al., 2016), with only Dangerousity being based on spatiotemporal tracking data. As a game of interactions, one team’s performance is intricately linked to the opposing team’s performance. Consequently, we expanded these 14 PIs for one team by calculating the difference to the opponent in the same interval, resulting in a total of 28 PIs (Tab. 1a). We selected five PGs which are all representative of team performance (Herold et al., 2021; Mackenzie & Cushion, 2013) (Tab. 1b). Features were normalized to a [0,1] range. The scaler was fitted to the training data and subsequently applied to the validation and test sets. To ensure full transparency and reproducibility, all scripts for data processing, model training, and evaluation used in this study are available in the associated GitHub repository.
a) Performance indicators (PIs) and b) Prediction goals (PGs) of a team and their definitions utilized in our study. PIs 1–14 as individual PIs and 15–28 as the difference to the opponent team of the individual ones. The definition of a PI is always based on the performance of the respective team in an interval and is either an event performed, or a metric based on actions of the team. The definition of a PG is that the event happens at minimum once in the respective prediction window for the team.
| No | Abbreviation | Definition |
|---|---|---|
| a) Performance Indicators (PI) | ||
| 1 | PICorner | Number of corner kicks |
| 2 | PIEntrBox | Number of entries of a player with ball possession into the opponent box |
| 3 | PIEntr3rd | Number of entries of a player with ball possession into the attacking third |
| 4 | PIGoal | Number of goals scored |
| 5 | PIShot | Number of shot attempts |
| 6 | PICross | Number of crosses |
| 7 | PITackWon | Number of tacklings won |
| 8 | PIPassBox | Number of successful passes in or into the opponent box |
| 9 | PIPass3rd | Number of successful passes in or into the attacking third |
| 10 | PIBP | Time of ball possession |
| 11 | PIBPBox | Time of ball possession in the box |
| 12 | PIBP3rd | Time of ball possession in the attacking third |
| 13 | PIOutpOpp | Number of outplayed opponent players by successful passes |
| 14 | PIDanger | Goal scoring probability at each moment (Link et al., 2016) |
| 15-28 | PIPI_diff | Difference of both PI values, (Team – Opponent) |
| b) Prediction Goals (PG) | ||
| 1 | PGisGoal | A goal event for the team occurs |
| 2 | PGisShot | A shot event for the team occurs |
| 3 | PGisCorner | A corner kick event for the team occurs |
| 4 | PGisEntrBox | An entry into the opponent box performed by the team occurs |
| 5 | PGisEntr3rd | An entry into the attacking third performed by the team occurs |
For Part I, which focuses on gaining deeper insights into the PIs, we assess the inter-correlation between PIs and our PGs during matches. To compute these correlations, we establish pairs of observations as follows: we aggregate the values of one PI across all intervals in a defined window and correspondingly aggregate the PG values in the subsequent window, with each window of the same length. We then employ a rolling window approach across the entire dataset, utilizing a predefined step size of one minute. This process is repeated for each team across all matches. For correlation we use the Pearson correlation coefficient. The results reveal any correlations among all PIs and the PGs, highlighting which PIs may best represent a PG.
For Parts II and III, we employ the Matthews Correlation Coefficient (MCC) also known as Phi coefficient metric (Guilford, 1954) (Eq. 1), recommended for evaluating models with highly unbalanced datasets (Chicco & Jurman, 2020). Our selected PGs represent sparse events in football matches, e.g., 2.43–3.06 goals (Farias, Bergmann, Vaz, & Pinheiro, 2018) or 10 corner kicks per match (Siegle & Lames, 2012), in our dataset 2.99 and 9.54 per match, respectively. With 139,694 five-second intervals as observation units in our study, the dedicated PGs are present in only 0.22% (goals), 0.7% (corner kicks), 1.28% (shots taken), 2.54% (entries into the opponent box) and 6.97% (entries into the attacking third) intervals. The MCC ranges from −1 to 1, where 0 indicates random prediction. Consequently, while a random guessing model would generate MCC results around 0, indicating chance prediction, trained MLMs are expected to yield MCC values higher than 0, signifying learned and predictive capabilities.
In the final phase of our analysis, we conduct an in-depth exploration of one practical scenario to identify the most effective PI-combinations (Part III). Here, we carry out an extensive model training iteration, covering most of the conceivable configurations specifically for PGisGoal within the defined PW3min. Given the large number of possible combinations, we capped the maximum combined PIs at five. To do so, we combined, without repetition, all 28 PIs up to five
Analyzing the relationship between the selected PIs and our PGs in Part I, Figure 2 shows the resulting Pearson correlation coefficient (r) calculated with a window size of 15 minutes for both, PIs and PGs. For PGisGoal, results show that there is only a low correlation between all PIs and the PG ranging from r = −0.22 to r= 0.13, with the highest results achieved by PIPass3rd, PIDanger, and their team-differences. However, systematic trends emerge across all PGs. Figure 2 illustrates that higher correlations come in two cases. First, with an increasing frequency of an SRE, i.e. the sparse PGisGoal achieves an rmax= 0.13, whereas the most often occurring PGEntr3rd in our dataset has an rmax= 0.57. Second, when utilizing the respective PI differences most of the time (e.g., PIBP and PIBP_diff).

Furthermore, PIGoal, PITackWon, and their differences produce only low correlations with all PGs. Conversely, full pitch time of ball possession (PIBP), time or actions occurring in the final third (PIPass3rd, PIBP3rd, PIEntr3rd), and their team differences exhibit high correlations. Note as an exception, while PIBP correlates low with PGisGoal, time of ball possession in the final third (PIBP3rd) demonstrates a strong correlation. Besides, there is a noticeable trend indicating that an increase in correlation comes with an increase in total time due to differing window lengths. This is presumably because some PIs and PGs are sparse, and by using a larger time window, they are included more often. The complete correlation results for all examined window sizes are in the supplementary material (Tab. S1).
To classify suitable MLMs and PIs (Part II), we conducted the analysis regarding their predictive performance in all configurations. Given the diverse parametrization generated by each configuration (MLM × IW × PW × PI × PG), wherein certain MLMs or PIs perform differently, there is no singular approach to compare and rank the results of all prediction experiments uniformly. Due to our goal of finding suitable PIs on average, we rank our results by the MCCmean for decision-making. Figure 3 shows the MCC results for each PG and MLM, each boxplot contains 700 results of all examined configurations (IW (5) × PW (5) × PI (28)), except for PGisEntr3rd where results of particular configurations (IW10min, IW15min, PW10min, PW15min) were omitted due to the event occurring in every observational unit. Here, LR outperforms all other MLMs in all PGs by their mean results (31% better MCCmean to 2nd rank). Interestingly, the MCCmean ranking of the examined MLMs is the same for all chosen PGs (except for PGisGoal, where NB and SVM switched places two and three). Our baseline model (random guesser) performs as expected with an MCCmean close to 0 (MCCmean_RandomGuesser = −2e−4 to −5e−5). The minimum and maximum MCC results fluctuate among each configuration, as it is highly dependent on the data split. Facing this, we perform 10-fold-cross validation in Part III to obtain more robust performance estimates. The complete experimental results are available in the supplementary material (Tab. S2).

MCC results of all experiments split by MLMs and PGs. Each boxplot contains results of 700 experiments, except for the PGisEntr3rd with 250 experiments. The red dotted line indicates MCC=0.
For our second task in Part II, Figure 4 shows the same experimental results from the task before but now split by the individual PIs. Each boxplot contains 545 experiments each (MLM (5) × IW (5) × PW (5) × PG (5)). The results show the overall Top 3 PIs are PIDanger (MCCmean=0.051 ± 0.053), PIEntr3rd (0.045 ± 0.055) and PIDanger_diff (0.042 ± 0.049). Lowest MCCmean were achieved by PIGoal (−0.006 ± 0.033), PIPassBox (−0.005 ± 0.033), and PITackWon_diff (−0.001 ± 0.024) on average, indicating random results with an MCCmean around 0.

MCC results of all experiments split by PIs. Each boxplot contains the results of 545 experiments. Boxplots are sorted by their MCCmean in descending order. The red dotted line indicates MCC=0.
Going deeper into these results with an additional split by the PGs, Figure 5 shows the Top 3 and Bottom 3 PIs per PG in descending order by their MCCmean. PIDanger is ranked best for PGisGoal (MCCmean= 0.037 ± 0.036) and PGisShot (0.073 ± 0.068), whereas PIEntr3rd is ranked at the top for PGisEntr3rd (0.058 ± 0.041), PGisCorner (0.051 ± 0.047), and PGisEntrBox (0.069 ± 0.056). Additionally, exceptional good results were achieved by PIDanger, PIDanger_diff, and PIPass3rd having all a Top 9 rank for each PG. Similar to our results in Part I, PIGoal, PITackWon, and their team-differences seem not to be useful as individual PIs to predict our PGs and are ranked in the bottom third. Complete PI rankings for each PG are available in the appendix (Fig. 8–12).

a) Top 3 and b) Bottom 3 PIs for each PG. PIs are sorted by their mean MCC results in descending order. Each boxplot contains the results of 125 experiments, except for the PGisEntr3rd with 45 experiments. The red dotted line indicates MCC=0.
For our third part, we conducted extensive experiments where we first searched for the best ten individual PIs and later combined these in all possible combinations for the specific scenario of predicting a goal in the next three minutes (Part III). Figure 6a shows the Top 10 on average best performing PIs selected. The resulting PI-combinations were 10-fold cross-validated on all five IW sizes with LR, as it showed the best results in Part II. Regarding the overall CI95-results, we can select three interesting insights from the Top 10% configurations. First, the resulting ranking of individual PIs shows that some have a significant effect over others. Here, in the Top 10% PI-combinations, PITackWon, PIDanger_diff, and PIEntr3rd_diff occur the most often, 67%, 52%, and 49%, respectively (Fig. 6a). Second, regarding the number of individual PIs in a combination, four and five show the highest appearances with 25% and 23%, respectively (Fig. 6b). Third, the most occurring IW size is 5 minutes with 65% of all combinations, followed by 15 minutes with 32% (Fig. 6c). Interestingly, in 11 out of 512 cases (2%), only one PI was necessary to achieve a good prediction result. Moreover, PIOutpOpp_diff (rank 5, IW5min, CI95 = 0.037, MCCmean = 0.055 ± 0.025), and PIDanger (rank 8, IW15min, CI95 = 0.035, MCCmean = 0.055 ± 0.016) yield an overall Top 10 rank, which shows the power of some individual PIs in their applicability for predicting an SRE. By combining PIOutpOpp_diff and PITackWon, the best result was achieved with CI95 = 0.047 (rank 1, IW5min, MCCmean = 0.061 ± 0.02). Respective to the large fluctuation and dependencies on the data split, the best result with the lowest standard deviation was achieved by combining PIDanger and PIEntr3rd (rank 3, IW15min, CI95 = 0.039, MCCmean = 0.047 ± 0.012). An analysis of the two PIs that most often occur together revealed that PITackWon and PIDanger_diff are present in 37% of the cases where at minimum two PIs are combined. Indeed, this combination has also the second-best result overall (rank 2, IW5min, CI95 = 0.04, MCCmean = 0.061 ± 0.029). The complete results of our selected scenario (predicting a goal in the next three minutes) are presented in the supplementary material (Tab. S3).

The percentual appearance in the Top 10% PI-combinations in the application scenario (Part III) of a) the selected Top 10 individual Pls, b) the number of individual Pls combined, and c) the input window length.
In the last part of our study (Part IV), we show the applicability of our results, whereas we train a new model with our complete dataset and apply it to an unseen match. For this, we use the configuration of the third-ranked model in Part III (PI = PIDanger & PIEntr3rd, IW15min, PW3min, MLMLR, PGisGoal), as it showed only slightly fewer CI95-result (rank 1: CI95=0.047; rank 3: CI95=0.039) but the lowest standard deviation overall (rank 1: MCCmean=0.061 ± 0.02 vs. rank 3: MCCmean=0.047 ± 0.012), indicating the most stable configuration. Then, we predict over the course of the match whether a goal will be scored in the next three minutes for both teams taking part in the match FC Bayern Munich (FCB) vs. SC Paderborn (SCP) of the season 19/20. This match was not part of our study so far and can be seen as an application test for our trained model. Subsequently, in Figure 7, we plot for this match the predicted probabilities, the relevant PIs (PIDanger and PIEntr3rd), and important match events. Additionally, eight important phases are highlighted (P1-8) for explanation (see below). In the appendix, we offer figures for the first and second ranked configurations as well (Fig. 13 and 14). The top three trained models from our analysis are publicly available in the accompanying GitHub repository, facilitating reproducibility and further research.
The psychological momentum, frequently discussed in sports, is difficult to measure directly in football due to the low number of scoring events (Redwood-Brown, Sunderland, Minniti, & O’Donoghue, 2018). As a result, match momentum was introduced to represent the actual performance difference between the competing teams and streaks of one team. If one team has match momentum on its side, it could lead to a psychological advantage and create a “momentum”. Currently, there is no gold standard described for measuring match momentum. However, higher team performance typically leads to more SREs, suggesting that team performance should, in some manner, predict the occurrence of these events.
Therefore, we propose a possible application by utilizing all findings of our study for defining the match momentum metric. We define match momentum as the difference between the predicted goal probabilities of the two teams (Fig. 7). This empirically grounded metric accounts for both teams’ performances and is stable with a high update frequency. For our application scenario, we selected one of the best configurations with the lowest variation, while other configurations also showed strong results and could be used similarly. Our match momentum definition provides a valuable metric for performance analysis since it might help to identify optimal moments for tactical interventions, e.g., when the opponent’s likelihood of scoring is increasing, and it can be applied in setting live betting odds as it enables real-time prediction.

Application of the trained model (rank 3) for an unseen match between FC Bayern Munich (FCB) vs SC Paderborn (SCP) in Season 19/20 which resulted in a 3:2. In the upper half (FCB) and the lower half (SCP) for each team important events (corner kicks, given cards, goals scored, shots taken), PIDanger and PIEntr3rd, and prediction values are illustrated over the course of the match. Also, Dominance by Link et al. (2016) and the goal prediction difference between both teams, as our proposed match momentum metric, are shown. Additionally, eight important sequences (P1-8) are highlighted.
P1 indicates that prediction for the initial minutes of the game is impossible due to insufficient information. To address this, we add zeros for each PI and as soon as information is available, we include it in the prediction. P2 shows that no prediction is made during interruptions in the match, as predictions are only calculated when the match is running. P3 marks a dominant phase with an increasing predicted probability for FCB, where a goal was scored, and an additional shot was taken. P4 involves a goal scored by SCP, with media reports stating, “Out of nowhere, a blunder by [FCB goalkeeper] Neuer brings SCP back into the game!” (Sport.de, 2020). SCP performed poorly in the period leading up to the goal, as indicated by the low number of entries and the missing shots. Predictions based on these PIs cannot foresee such goals. P5 and P6 demonstrate that both teams can simultaneously increase the probability of scoring in the near future by performing well. The prediction for one team is independent of the other team’s performance. At the end of P6, FCB scored their second goal. Similarly to P4, P7 features another unforeseeable goal, with reports noting that FCB was “sniffing around for the decisive third goal” (Sport.de, 2020) just before SCP scored their second goal. In P8, FCB increased the pressure and their performance and so increased their probability of scoring. At the end of this phase, FCB scored the decisive goal and won the match.
Our results reveal several interesting points. The correlation analysis indicates that in 11 out of 14 PIs, the PI difference between the two teams leads to significantly higher correlations than the individual PIs themselves. This is reasonable, since a greater PI difference should correspond to more SREs for the better-performing team. Similarly, increasing the correlation window length (from 1 minute up to 15 minutes in our experiment) increases the correlation. This is expected, as most of the examined PIs and PGs are sparse (i.e., goals, shots taken, corner kicks, entries). In summary, high correlations can be achieved by having more PI information available, the more frequently a PG occurs, and by utilizing the interaction with the opponent (PI differences).
Ranking the PIs by their mean prediction performance across all PGs provides valuable insights into their general usability. The PI ranking varies across the utilized PGs, indicating that there is no universally optimal PI. However, certain PIs consistently appear in the Top 10 results across all PGs, demonstrating their general usability. Conversely, some PIs consistently perform poorly (e.g., PIGoal, PITackWon, PIPassBox, PICorner, and their differences). Additionally, the analysis shows that some PIs ranked in the middle (e.g. PICross or PIBP3rd_diff) achieve high MCC results for particular configurations, which indicates either an overfitting for a specific data split or that in some cases the PI can be used to predict the dedicated PG quite well. As mentioned, our goal was not to find the best configuration for a particular PI, instead, we wanted to find those PIs overall performing on average good with default configurations across different PGs and MLMs. In the future, it is recommended to have a deeper analysis of these PIs, yet they did not perform best on average here. Summarizing, PIs based on spare events, e.g., goals, corner kicks and box actions, seem not to be useful for SRE prediction, whereas PIs based on a range of actions and with high frequency are beneficial. As far as the MCC results of all experiments are concerned, all configurations provide only low predictive performance, which was to be expected since soccer is strongly influenced by chance (Lames, 2018) and our PGs are very sparse. Nevertheless, MCC results indicate that prediction is better than random guessing and can be increased by further hyperparameter tuning. Specifically, for our application scenario, the best configuration achieved an MCCmean approximately 6% better than random guessing. Considering that goals occur only about three times per match on average, this level of improvement represents a meaningful predictive gain in the context of such a sparse and chance influenced event.
Further insights are gained from our application scenario analysis. PITackWon, as a standalone indicator, has a low correlation with our selected PGs and demonstrates poor prediction quality. Conversely, combining PITackWon with other PIs results in the highest performance and most frequent occurrence of a PI in our application scenario, underscoring the potential of combining PIs to enhance predictive power. Similarly, other PIs, such as PIShot and PICross, could be valuable in combination. However, they were not included in our application analysis as they did not rank in the Top 10 during pre-selection, despite achieving Top 3 results in specific PGs. Surprisingly, results in Figure 6b indicate that combining more PIs does not necessarily lead to better prediction results, contrary to the expectation that more information would enhance prediction performance. This discrepancy might be due to the limitations of LR utilizing the full scope of available information, whereas a more sophisticated MLM could potentially make better use of it.
Additionally, PIDanger, PIDanger_diff, and PIEntr3rd consistently show strong results across all selected PGs. This consistency highlights their great value for predicting SREs and representing team performance. They demonstrate higher correlation and prediction quality than others and are more frequently selected among the Top 10% configurations in our application scenario. Moreover, including PIDanger_diff in a PI-combination yields more stable results in the 10-fold cross-validation, with significantly lower standard deviations than without (SDGroup_Danger_diff = 0.0216 and SDGroup_Not_Danger_diff = 0.0239; U-statistic = 42766.5, p < 0.001). For future work, it is recommended to use these well-performing PIs and additionally utilize hyperparameter optimization to achieve even better prediction performance.
Regarding our employed methods, we encountered several issues. Utilizing MLMs with default configurations, without hyperparameter engineering, tuning, and searching (Decroos & Davis, 2019), penalizes deep learning models and those with numerous hyperparameters. However, we propose a method that enables a fair comparison across thousands of model configurations simultaneously. As a result, it highlights which PIs perform reliably across different model architectures. This is particularly useful for researchers and practitioners seeking initial insights without significant resource investment in hyperparameter tuning. Hence, LR was selected as the best-performing MLM presumably due to its fewer hyperparameters, whereas NN, for instance, were not chosen. Nevertheless, all utilized MLMs exhibited high MCC values in particular configurations, indicating their potential usability with hyperparameter optimization in the future. Using the MCCmean for ranking the PIs can lead to a misunderstanding of their predictive power. This is because the mean includes both the best-performing MLMs (e.g., LR) and the worst-performing ones (e.g., KNN). Consequently, the maximum achieved MCCmean for a PI per PG is quite low, e.g., between 0.037 to 0.073 meaning around 4% to 7% better than guessing, which does not accurately reflect the possibly predictive power of each PI. Instead, it highlights their general usability, which was our primary goal.
Furthermore, our examined PIs are restricted to those publicly available and can be calculated in-game. It would be from large interest and value to include and test other metrics, i.e., expected goals (Rudd, 2011), expected thread (StatsBomb, 2024), or VAEP (Decroos, Bransen, van Haaren, & Davis, 2020). These PIs are generally more sophisticated as they integrate multiple dimensions of play and therefore represent a richer description of the current game state, similar to Dangerousity. Although integrating match context variables, such as current league table position, results of recent matches, market value of the teams, or other inter-match metrics, would increase complexity, it could further enhance the accuracy of the predictions, as it has been shown valuable in other studies (Horvat & Job, 2020). While PIDanger requires fine-grained positional tracking data, the majority of performance indicators analyzed in our study are based on event data alone. This ensures that our framework can be applied broadly, including in leagues and competitions where only event-level data is available. Notably, opensource event data of professional matches is accessible to researchers through resources such as StatsBomb’s Open Data (StatsBomb, 2025), offering a valuable foundation for further replication and extension of our work. While the SREs investigated in this study are inherently sparse, this sparsity reflects their relevance in the context of soccer performance, as such events typically occur infrequently within or across possessions. Importantly, the proposed prediction framework is adaptable and can be applied to alternative event types (e.g., turnovers, freekicks) depending on the analytical focus and use case.
Our proposed In-Play Prediction Masking method addresses a known issue where specific PIs predict specific events very accurately within a short time window. For example, entries into the box can predict events, such as shots taken or goals scored, within a short time window of around one minute very good, as most shots and goals occur from inside the box. This confounding effect is problematic for MLMs and their analysis, as our models are intended to predict SREs over a longer time window in the future. To tackle this issue and examine the predictive ability of all PIs equally, we implemented a one-minute hidden window before the PW. This method is not yet described in the literature and could penalize PIs that are better suited for very short-term prediction tasks.
Our study examined different PIs by their predictive ability for SREs. The achieved MCC results are low for events like shots taken and goals due to their sparsity and the difficulty of prediction. Nonetheless, the results indicate that some PIs are in general more useful than others and warrant more intensive examination in the future. Furthermore, the combination of PIs appears to achieve even better results, particularly for PIs that have not been emphasized as important in literature anymore (e.g., Tacklings Won). Additionally, our analyses demonstrated that a 5-minute input window length achieves better results than a 15-minute window. This could be attributed to a psychological advantage based on a performance streak that may not last longer than 5 minutes.
As a possible application we use the difference in both teams’ in-game goal predictions as an empirical based and stable representation of match momentum. This proposed match momentum metric updates more frequently and smoothly than other sparse PIs, and without the high fluctuation seen, for instance, in the Dangerousity metric (Link et al., 2016), making it a valuable tool for performance analysis and live betting markets. However, our proposed methodology for measuring match momentum has limitations, such as imprecise predictions during the initial minutes when the input data is mostly zero-filled, and the absence of predictions during match interruptions. To address these issues, we suggest adapting our methodology by including pre-game information as mentioned and by examining the interruption length influence on different PIs, as a performance advantage can be stopped by a longer interruption. For future research, we suggest utilizing different SRE predictions in combination, which would even better reflect both teams’ actual performance. Overall, our results enhance the understanding of valid PIs by identifying those strongly associated with SREs and demonstrating their predictive stability.
In this paper, we examined the applicability of common in-game PIs for predicting success- or scoring-related match events (SREs) in soccer. To achieve this, we investigated which PIs or PI-combinations (e.g., time of ball possession, opponents outplayed, Dangerousity, shots taken) offer the highest predictive performance, as these should best reflect a team’s current performance. For prediction, we employed data from 102 matches in German professional football leagues, we generated 28 well-known in-game PIs at every moment of effective playing time and tested thousands of machine learning model configurations. Our results demonstrate that while predicting in-game SREs is inherently challenging due to soccer’s chaotic nature and the high influence of chance, in-game performance data contains valuable information that can be leveraged for short-term event prediction. However, PIs vary in their predictive power and general usability across different target events, indicating that some PIs are more applicable than others. Furthermore, combining PIs or using the difference between the competing teams’ PI values generally leads to better predictive performance. In summary, PIs based on rare events, such as goals, corner kicks, and box actions, are less effective for SRE prediction. In contrast, PIs derived from a wider range of actions and collected at high frequency (e.g., ball possession in the last third, Dangerousity, outplayed opponents) are more beneficial. To the best of our knowledge, we are the first to predict in-play events beyond the immediate next event. Building on these findings, we propose a match momentum metric based on the difference in both teams’ goal predictions. This empirically grounded tool can help identify optimal moments for tactical decisions, such as reacting to an opponent’s increasing likelihood of scoring. These insights lay the foundation for developing improved PIs and enhancing performance analysis in both coaching and live betting.