Can machine learning make you rich from sports betting?

Two years ago, I asked myself if it would be possible to use machine learning to better predict the outcome of soccer games.

I decided to give it a serious try and today, two years and contextual data from 30,000 soccer games later, I’ve gained lots of interesting insights.

The big data challenge: Let the data mining begin

Step 1: To begin with, I harvested as many data points as possible. I mined old game data from every different source and API I could find. Some of the more important ones were Football-data, Everysport, and Betfair.

Step 2: I then merged these data points with their corresponding results, quantified it, and put everything into one database. 

Step 3. Finally, I used the data to train a machine learning model, to be used as my software for predicting upcoming soccer games.

Big data mining.
Percentage of correctly predicted games over time.

How to measure predictions of the unpredictable

The nature of a soccer match is, of course, that it is unpredictable. I guess that’s why we love the game, right?

Still, I was somewhat obsessed with the naive notion that I, armed with a data-driven machine learning model, would be able to predict games better than I normally would. At that point, I based most of my sports bets on emotions (“gut feelings”) rather than actual data.

The first challenge was to find out how to measure whether or not my model was succeeding. I quickly realized that measuring the actual percentage of correctly guessed games didn’t add much value — not without some form of context.

I decided to compare the model’s output with the best guesses of the actual market. The easiest way to assess such data was to harvest market regulated odds. Therefore, I started comparing how my model would perform if betting on Betfair, only because their odds are adjusted based on real people betting real money against each other.

The results: Did my model make me rich?

Fast-forward to today: Now — two years have passed. Has the model made me a rich man?

Well, no.

I soon realized that my predictions, for the most part, were aligned with the market’s best performance.

Since I used a regression-based model, I was able to predict the strength of the probability of a certain outcome of a game. And at the strongest grades of probability, my model predicts roughly 70% of the games correctly. Since the market performs just as well, it makes it difficult for me actually to make any serious money from my bets.

But, to be honest, I never actually thought that I would create a “money machine”, either. Instead, I came to several rather interesting insights about the possibilities (and limitations!) of big data and machine learning:

Accumulated winnings.
Accumulated winnings over time.

Learning 1: Machine learning and diminishing gains

In theory, machine learning should be able to improve over time. The amount of data the model has to learn from grows, enhancing the outcome of the predictions.

Well, this wasn’t my experience at all.

Two years ago I started with about 2,000 games in my database — and with quite limited data sets attached to them. Today, I have almost 30,000 games in the database, complete with lots of metadata covering everything from weather and distances between the teams home grounds to shots and corners.

All this added data — and the fact that the model has been able to “learn” over time! — still didn’t improve its predictions. It seems big data and machine learning only take you so far in trying to predict the unpredictable.

Learning 2: The power of unbiased generalizations

The power of machine learning seems closely tied to its ability to make unbiased generalizations.

For example: Over the past two years, I was curious to see if my model could predict when winning or losing streaks were to be broken. If, for instance, it could predict when Barcelona would finally lose after winning ten games straight. Could my model prove certain anomalies to be significant?

Well, it has shown to be not that good at that.

What I found instead was that the model was surprisingly good at betting against overvalued teams over time.

Last season, I saw how my soccer prediction machine quite often predicted against Borussia Dortmund while the market made another prediction. Dortmund ended up having a bad season making my model advantageous compared to market predictions. This season I have seen the same when it comes to teams like Liverpool and Chelsea.

So the lesson learned is that some people tend to make sports betting decisions based on emotions. Liverpool and Dortmund are teams liked by lots of people and at times, you make predictions with your heart instead of your brain. My machine learning model, well, it does not.

Learning 3: Machine learning and easy gains

If nothing else, I learned that making predictions that outperforms the market is hard. Still, when I started looking at what I had achieved (instead of just obsessing over at what I hadn’t), I found one quite surprising fact:

From a simple Python program and less than 10,000 lines of code, I still had made something that performed just as well as the market. How many man-hours aren’t behind bookies odds models and predictions? The model can pick out attractive bets on a weekly basis, just as any newspaper or expert would. By making generalizations you might not be able to find that one bet that will make you rich — but it may save you lots of time in the right context.

Plotting.
Plotting soccer results.

Implementing machine learning to Wide Ideas

With these insights in mind, I started to look at another project I’ve been involved in for the last five years; the idea platform Wide Ideas, a platform for companies to crowdsource ideas and creativity.

What I wanted to do was to look at the ideas companies gather from their employees and try to predict whether the idea would be implemented or not.

The team and I quantified the data, but instead of shots on goal and weather forecasts, we looked at how many had interacted with an idea — and in what way. And lo and behold; the outcome was on par with the soccer predictions:

We’re now able to make decent predictions on whether or not a creative idea will be implemented or not. We can visualize this in a way to encourage more great ideas through gamification.

Can we find a good idea that doesn’t follow the general patterns of a good idea? No, not really – not yet at least.

Still, for the product, and given that you look at an organization that can harvest 10,000 ideas per year, finding ways to highlight and encourage particular ideas can save time and resources. So just by going from 10,000 ideas to 100 probably good ideas and visualizing the result saves lots of time.

The great gap between making machines just as good as humans and actually making them better than we are.

Big data and machine learning might be able to predict anything from early-stage cancer to making self-driving cars anticipate potential dangers. Models like this will probably prove most useful where generalizations save time.

Take medical implementations, for example. Sifting through thousands of birthmarks pictures, a model could help pick the most likely ones to be cancer, thus saving doctors valuable time and resources.

However, human behavior may prove to be tricky. In what way is human behavior predictable? We’re rationally irrational. We will be able to generalize, placing people into different categories based on what you like to eat, watch or do, but there might just be too many factors that set us apart as individuals.

Will big data and machine learning be able to detect the anomalies — or will it just be superb at making generalizations?

I hope that we’ll experience a future where companies are focusing on actual data analysis, instead of thinking that “big data” by default equals “better data.”

So, until someone proves me wrong (or Arnold Schwarzenegger returns from the future, whichever comes first!), I believe that machine learning should be put to use where generalizations best can save time from real humans. 

The risk otherwise is that we’ll end up with so many metrics that the sheer amount would suffocate any possibility to make sense out of it.


About the Author

This article was written by Ola Lidmark Eriksson, CTO at Wide Ideas.

Recently Published

Key Takeaway: Honey bees, originally tropical insects, evolved complex nest-choosing patterns 600,000 years ago to survive cold climates. However, research into honey bee pressures and behavior rarely takes into account these nest preferences. Researchers have found that tree nests lose less heat than conventional hives and that features of man-made hives inserted for convenience increase […]

Top Picks

Key Takeaway: A study has found that our memory helps us learn from experiences and develop new knowledge by integrating and updating information. Memory can forge inferred connections beyond direct experiences, which can sometimes lead to false inferences. The study found that people may prioritize information from liked sources more than those from disliked ones, […]
Key Takeaway: OpenAI CEO Sam Altman sparked controversy by referencing the 2013 movie “Her” to highlight the novelty of ChatGPT’s latest iteration. Actor Scarlett Johansson accused the company of improperly using her voice after she spurned their offer to make her the voice of ChatGPT’s new virtual assistant. This highlights the “sci-fi feedback loop,” which […]

Trending

I highly recommend reading the McKinsey Global Institute’s new report, “Reskilling China: Transforming The World’s Largest Workforce Into Lifelong Learners”, which focuses on the country’s biggest employment challenge, re-training its workforce and the adoption of practices such as lifelong learning to address the growing digital transformation of its productive fabric. How to transform the country […]

Join our Newsletter

Get our monthly recap with the latest news, articles and resources.

Login

Welcome to Empirics

We are glad you have decided to join our mission of gathering the collective knowledge of Asia!
Join Empirics