Estimating the Probability of Winning an NFL Game Using Random Forests

Similar documents
Building an NFL performance metric

Statistical methods in sports with a focus on win probability and performance evaluation

Projecting Three-Point Percentages for the NBA Draft

= = 1 2. (6) What is the probability that Jack will get a wooden bat and Peter will not = 25

Predicting the Draft and Career Success of Tight Ends in the National Football League

This file is part of the following reference:

PREDICTING THE FUTURE OF FREE AGENT RECEIVERS AND TIGHT ENDS IN THE NFL

BASKETBALL PREDICTION ANALYSIS OF MARCH MADNESS GAMES CHRIS TSENG YIBO WANG

How Do Injuries in the NFL Affect the Outcome of the Game

Predicting NBA Shots

PREDICTING THE NCAA BASKETBALL TOURNAMENT WITH MACHINE LEARNING. The Ringer/Getty Images

Decision Trees. an Introduction

Bayesian Optimized Random Forest for Movement Classification with Smartphones

NFL Player Evaluation Using Expected Points Added with nflscrapr

Using Spatio-Temporal Data To Create A Shot Probability Model

Predicting the Total Number of Points Scored in NFL Games

PREDICTING the outcomes of sporting events

Navigate to the golf data folder and make it your working directory. Load the data by typing

Rules of Play REDZONE! I. DESCRIPTION OF THE GAME: II. EQUIPMENT INCLUDED: III. PLAY OF THE GAME: IV. HOW TO USE THE GAME EQUIPMENT: 1.

Special Topics: Data Science

Rules of Play REDZONE! I. DESCRIPTION OF THE GAME: III. PLAY OF THE GAME: IV. HOW TO USE THE GAME EQUIPMENT: 1. Game Dice:

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag

Pairwise Comparison Models: A Two-Tiered Approach to Predicting Wins and Losses for NBA Games

Coach s Clipboard

Predicting the use of the Sacrifice Bunt in Major League Baseball. Charlie Gallagher Brian Gilbert Neelay Mehta Chao Rao

Chapter 12 Practice Test

Biostatistics & SAS programming

INTER-AMERICAN TROPICAL TUNA COMMISSION SCIENTIFIC ADVISORY COMMITTEE FOURTH MEETING. La Jolla, California (USA) 29 April - 3 May 2013

Basketball data science

4th Down Expected Points

Lab 11: Introduction to Linear Regression

Football Player s Performance and Market Value

GREY BOARD. Coach s Clipboard

Using Multi-Class Classification Methods to Predict Baseball Pitch Types

Two Machine Learning Approaches to Understand the NBA Data

CS 7641 A (Machine Learning) Sethuraman K, Parameswaran Raman, Vijay Ramakrishnan

COMPLETING THE RESULTS OF THE 2013 BOSTON MARATHON

Deconstructing Data Science

The Reliability of Intrinsic Batted Ball Statistics Appendix

Super Bowl vs. Super Tuesday Spell Plenty of Thrills for All

Lecture 5. Optimisation. Regularisation

A Novel Approach to Predicting the Results of NBA Matches

Predictors for Winning in Men s Professional Tennis

Multilevel Models for Other Non-Normal Outcomes in Mplus v. 7.11

Predictive Model for the NCAA Men's Basketball Tournament. An Honors Thesis (HONR 499) Thesis Advisor. Ball State University.

A school carnival committee features a different version of the Making Purple game, as shown below.

The probability of winning a high school football game.

intended velocity ( u k arm movements

Exercise 11: Solution - Decision tree

Analysis of Variance. Copyright 2014 Pearson Education, Inc.

y ) s x x )(y i (x i r = 1 n 1 s y Statistics Lecture 7 Exploring Data , y 2 ,y n (x 1 ),,(x n ),(x 2 ,y 1 How two variables vary together

Predicting Momentum Shifts in NBA Games

DISMAS Evaluation: Dr. Elizabeth C. McMullan. Grambling State University

Section I: Multiple Choice Select the best answer for each problem.

Has the NFL s Rooney Rule Efforts Leveled the Field for African American Head Coach Candidates?

Mathematics of Pari-Mutuel Wagering

Evaluating and Classifying NBA Free Agents

Football Play Type Prediction and Tendency Analysis

SCTB16 Working Paper SWG 6

Examining NBA Crunch Time: The Four Point Problem. Abstract. 1. Introduction

B. AA228/CS238 Component

An Examination of Decision-Making Biases on Fourth Down in The National Football League. Weller Ross, B.S. Sport Management

JPEG-Compatibility Steganalysis Using Block-Histogram of Recompression Artifacts

Predicting the NCAA Men s Basketball Tournament with Machine Learning

Journal of Quantitative Analysis in Sports. Rush versus Pass: Modeling the NFL

Bayesian Methods: Naïve Bayes

INTRODUCTION Page 2 SETTINGS

Machine Learning Application in Aviation Safety

SPORTSCIENCE sportsci.org

Announcements. Lecture 19: Inference for SLR & Transformations. Online quiz 7 - commonly missed questions

Using Computer Vision and Machine Learning to Automatically Classify NFL Game Film and Develop a Player Tracking System

Predicting the use of the sacrifice bunt in Major League Baseball BUDT 714 May 10, 2007

Decision Making in American Football: Evidence from 7 Years of NFL Data

Deconstructing Data Science

Data Analytics based Deep Mayo Predictor for IPL-9

A One-Parameter Markov Chain Model for Baseball Run Production

NFL Direction-Oriented Rushing O -Def Plus-Minus

Scoring dynamics across professional team sports: tempo, balance and predictability

1. OVERVIEW OF METHOD

Predicting Horse Racing Results with TensorFlow

Taking Your Class for a Walk, Randomly

Southeast Idaho Youth Football League General Game Rules

PLEASE MARK YOUR ANSWERS WITH AN X, not a circle! 2. (a) (b) (c) (d) (e) (a) (b) (c) (d) (e) (a) (b) (c) (d) (e)...

Roland C. Deutsch. April 14, 2010

Standardized catch rates of yellowtail snapper ( Ocyurus chrysurus

Fundamentals of Machine Learning for Predictive Data Analytics

An Analysis of Factors Contributing to Wins in the National Hockey League

3.0. GSIS 3.1 Release Notes. NFL GSIS Support: (877) (212)

2 When Some or All Labels are Missing: The EM Algorithm

A point-based Bayesian hierarchical model to predict the outcome of tennis matches

Factors Affecting the Probability of Arrests at an NFL Game

Modeling Fantasy Football Quarterbacks

A computer program that improves its performance at some task through experience.

EEC 686/785 Modeling & Performance Evaluation of Computer Systems. Lecture 6. Wenbing Zhao. Department of Electrical and Computer Engineering

SAP Predictive Analysis and the MLB Post Season

Should bonus points be included in the Six Nations Championship?

Outline. Terminology. EEC 686/785 Modeling & Performance Evaluation of Computer Systems. Lecture 6. Steps in Capacity Planning and Management

A Game-Theoretic Intelligent Agent for the Board Game Football Strategy

Data Set 7: Bioerosion by Parrotfish Background volume of bites The question:

Predicting Tennis Match Outcomes Through Classification Shuyang Fang CS074 - Dartmouth College

Transcription:

Estimating the Probability of Winning an NFL Game Using Random Forests Dale Zimmerman February 17, 2017

2 Brian Burke s NFL win probability metric May be found at www.advancednflstats.com, but the site has been inactive since Burke joined the ESPN Analytics Department in 2015 Bins every play in a training dataset according to current score, time remaining, and field position Estimates win probability (WP) for a new play by the proportion of training observations in the corresponding bin for which the team on offense won the game, with some adjustment for down and yards to go for a first down Some extrapolation and smoothing used to borrow strength from bins similar to the one corresponding to the new play

3 Lock and Nettleton (LN) random forest approach Similar in some ways to Burke s, but differs as follows: Replaces subjective binning of training data with data-driven partitioning that minimizes prediction error (random forest) Additional variables are included: pre-game point spread, a variable that combines difference in score with time remaining, number of remaining time-outs each team has, and total points scored Allows for assessment of relative importance of each variable

4 Data Training data: Data from all plays from 2001-2011 NFL seasons n = 430,168 plays p = 10 predictor variables Response y i is an indicator variable for whether the team on offense before play i won the game Test data: Data from all plays from 2012 season

Regression trees A classification or regression tree is grown by performing a series of binary splits, each of which splits the p-dimensional space of predictor variables into two parts according to the value of just one of the variables At each split, the single variable is selected and the split-point is chosen to minimize the misclassification error (a greedy algorithm) Stops when when a pre-set minimum node size is reached The end result is a set of (hyper)rectangles corresponding to the terminal nodes; some pruning may then be performed For a regression tree, the predicted response is the average of the observations in each terminal node 5

6

Two-predictor illustration (from Elements of Statistical Learning by Hastie, Tibshirani, and Friedman): 306 9. Additive Models, Trees, and Related Methods R5 R2 t4 X2 X2 R3 t2 R4 R1 t1 t3 X1 X1 X1 t1 X2 t2 X1 t3 X2 t4 R1 R2 R3 X2 X1 R4 R5 FIGURE 9.2. Partitions and CART. Top right panel shows a partition of a two-dimensional feature space by recursive binary splitting, as used in CART, applied to some fake data. Top left panel shows a general partition that cannot be obtained from recursive binary splitting. Bottom left panel shows the tree corresponding to the partition in the top right panel, and a perspective plot of the prediction surface appears in the bottom right panel. 7

8 Bagging and random forests Predictions corresponding to an input x obtained by trees generally have low bias but high variance. To reduce this variance, bagging (bootstrap aggregation) is a useful technique: we fit the same regression tree many times to bootstrap samples of the training data, and average the predictions. This variance may be reduced even more if we can reduce the correlation between the trees. Random forests achieve this by, before each split, randomly selecting a subset of the predictor variables as candidates for splitting.

9 Steps of LN random forest approach 1. Draw a bootstrap sample (of size 430,168) from the training dataset and regard these data as belonging to a single node N 0. 2. Randomly select m (taken to be 2) predictor variables from the 10 predictors. 3. For each selected predictor variable x and all possible cutpoints c, compute the sum of squared errors. 2 k=1 i N k (y i ȳ k ) 2, where N 1 and N 2 are the subsets of training observations with x c and x > c, respectively, and ȳ k is the average of the y i s in N k.

Steps of LN random forest approach, continued 4. Choose the variable and the cutpoint c to minimize the sum of squared errors in Step 3, and split the training data into two subnodes accordingly. 5. Repeat Steps 2 through 4 recursively at each resulting subnode until either: (a) the number of observations in any subnode is less than nodesize (a user-supplied tuning parameter taken to be 200), or (b) the y i s corresponding to all the observations in a subnode are identical. These 5 steps produce a tree. Repeat them B times to obtain a random forest. 10

11 Prediction using the random forest To predict WP using the random forest: 1. For each tree, trace the path of the new play situation down the branches of the tree to a terminal node, and obtain the predicted response for that tree as the proportion of wins in that terminal node. 2. Average the predicted responses across all B trees in the forest.

12 Assessing relative importance of predictor variables Methods: 1. Graphical assessment Plot how WP changes when one variable is changed while holding all the others constant See Figure 3 of paper

13 Assessing relative importance of predictor variables 2. Numerical assessment: Randomly permute the values of predictor variable k with the test set and re-predict WP; For each play i compute the squared error after permutation to the original square error; Repeat the above two steps 100 times and compute the average increase in squared error for play i; Average over all plays to obtain a measure of overall importance of the kth predictor variable. Overall, difference in score and spread were the two most important variables.

14 Addressing other interesting questions Can chart the WP before each play of a game to see how it evolves over the course of the game (i.e. a time series plot) Can determine which play over the course of the game, a season, etc. resulted in the greatest change in WP Can use change in WP after each of several successful play options to assess coaching decisions. For example, in the 2016 AFC Championship Game, Broncos vs. Patriots, the Patriots were favored by 3. With 6:03 left in the game and 3 timeouts remaining, they faced 4th and 1 from the Broncos 16 trailing 20 to 12. Should they go for it or kick a field goal?

15 Punt, kick a field goal, or go for it? If we know, for each of these 3 options, the occurrence probability of each potential outcome (punt returned for touchdown, or punt and net yardage; field goal made or not made; touchdown scored, first down made or not made plus yardage, or touchdown scored against); and the offense s WP for each new game situation that results from each outcome, then we may compute the expected win probability (EWP) corresponding to that option as EWP = (Probability of outcome) (Corresponding WP).

16 Punt, kick a field goal, or go for it? For example, EWP for a field goal at a given distance is EWP FG = θ WP make + (1 θ)wp miss where θ = probability of a made field goal (at the given distance). We may use outcomes of all previous field goal attempts (possibly smoothed by a logistic regression of those outcomes on distance) to estimate θ at the given distance. This approach is due to Zimmerman and Nettleton (2016, Midwest Sports Analytics Meeting).

17 Back to the 2016 AFC Championship game... Coach Bill Belichick elected to go for it, and the Patriots were stopped short. However, it was the right call at the time since EWP(field goal attempt) = 0.199, while EWP(go for it) = 0.281. With 2:25 left in the same game and still 3 timeouts remaining, the Patriots faced 4th and 6 from the Broncos 14 trailing 20 to 12. Now what should they do? Belichick elected to go for it again, and they were again stopped short. This too was the call that maximized EWP, since EWP(field goal attempt) = 0.147 and EWP(go for it) = 0.161.

18 Interestingly, the Patriots lost the game by 2 points, 20 to 18. If they had kicked a field goal in either one of the two fourth down situations, they might have won. DZ cannot help but wonder if this experience played a role in Belichick s decision in the 2017 Super Bowl to kick a field goal rather than go for it on 4th and goal from the 15 with 9:44 left in the 4th quarter.

19 Real-time computing of EWP by coaches? NFL has computers on the sidelines and in the pressbox (http://www.sfgate.com/technology/article/nfl-players-to-usetablet-computers-during-games-5665371.php), but it seems they are allowed to use these only to view static pictures of previous plays. College football doesn t yet allow computers for any in-game purpose (http://www.timesunion.com/sports/college/article/lowtech-college-football-yet-to-embrace-10914002.php).