The 2020 World Series of Poker (WSOP) Main Event final table at the Rio All-Suite Hotel & Casino in Las Vegas is just a couple of days away. One man traveling across the country from Somerset, New Jersey to compete for $1.5 million and a shot at the bracelet is Ryan “Hagzzz021” Hagerty, who will start the final table third in chips with 5,075,000.
With nearly $2 million in online tournament earnings and $341,128 in the live realm, the 28-year-old Hagerty is among the most experienced players at the table, though it’s only his second WSOP Main Event (he busted Day 2 in 2018 and had to miss last summer due to his sister’s wedding). No matter where he finishes, Hagerty will notch the biggest live score of his career, and what may be surprising to some is the fact that he almost didn’t play.
“It’s pretty sick, especially being the fact that the day of the tournament I wasn’t initially planning on playing it because I didn’t like if you got COVID you get ninth place,” he admitted. “I was back and forth on whether I should play it or not.”
He ultimately decided to play, but not before selling 50 percent of his action.
Wasn’t gonna play, but got some Fomo. Selling to main at 1.2 have 50% and would need to sell out, if anyone wants to gamble lmk
Tableau Des Mains Poker, opening poker hands, clams casino - i'm god (instrumental), monte casino theatre shows Thank you for subscribing! We are so glad to be with you!
— Ryan Hagerty (@Hags_Ryan)Obviously, the investors are quite excited for their respective pieces, but Hagerty is equally so.
“It took a while for it to really sink in that I made the final table. It’s a weird year, playing two days online. It was a grind, a really good tournament. It was a much tougher field than a normal main event being online. It was a battle, but it feels amazing … The last two years I’ve done really well live … I feel like once I get to that final table, it’s been a while since I’ve touched the cards and been to a live table, but I think after a couple of hands I’ll get right back into it. I think I’ll be fine.”
'It’s been a while since I’ve touched the cards and been to a live table, but I think after a couple of hands I’ll get right back into it. I think I’ll be fine.”
Not only does Hagerty get to share the experience with his investors, he actually made the final table with a close friend.
“The final table definitely has some tough players. Actually, one of my best friends in poker for the past three years, Mike Cannon, made the final table so it’s pretty cool to be going to the final table with him.”
Hagerty continued: “When we would play at the Borgata series we would rent out a beach house in Brigantine. He was a part of it and we became really close friends, always pulling for each other. We both had a lot of chips throughout Day 2. We were kind of like, ‘Is this going to happen? Is this really going to happen?’ I couldn’t believe it. It’s pretty amazing to share this experience with somebody I’m pretty close with.”
Hagerty’s poker journey to the WSOP Main Event final table has been a decade in the making. About 10 years ago, he was playing $20 buy-in home games as a high school senior. Before long, he started to absorb strategy material, and in 2013 dove into online poker when it was introduced in the Garden State.
In 2015, he graduated from Rowan University with a degree in journalism and hoped to use that to pursue opportunities in poker media. He even covered the 2015 World Poker Tour $15,000 World Championship – won by Asher Conniff for $973,683 – where he interviewed players such as Jason Mercier, Greg Merson, David Williams, and Ryan Riess.
After moving to Baltimore to be with his girlfriend and a brief stint working at a Target, Hagerty traveled home to New Jersey and binked a $21,000 score in an online Borgata event.
“That was my big bink. Poker was my passion and I just wanted to go for it, so I went for it [after my girlfriend and I broke up],” Hagerty explained. “It’s an up-and-down rollercoaster. It wasn’t always amazing, but these past three years I’ve really honed my skills and saw myself improve a lot. The results started flowing in. That’s basically my journey.”
He added: “There was a point in my career where I was pretty broke so I was Ubering, but I’ve never really had a real job. Basically, for five years I’ve been a pro, and the past three years have been great. It’s nice when you’re playing well and living comfortably as a pro. That’s kind of how it’s been, it’s all come full circle to this.”
Among Hagerty’s biggest scores are $70,865 for winning the 2019 WPTDeepStacks Schenectady, taking down the 2019 Borgata Spring Poker Open Event #12: $400 NLH for $37,472, and winning the 2018 Delaware Poker Championship $550 Main Event for $25,500.
As for the upcoming final table, Hagerty is confident.
“I think my preparation is going to be good, I’m going to ready,” he said. “Whatever happens, happens. As long as I play my A-game I’m not going to try and think about how big the pay jumps are. I’m just going to go in there and play poker … that’s what I’ve been doing all tournament. Even on Day 2, I wasn’t worried about busting the tournament. I can’t be thinking about that; about how big the moment is. I just have to play my game, focus, and I kept that approach until I got to the final table.”
Like all the other players at the final table, Hagerty has taken quarantining very seriously so as to decrease his risk of contracting COVID.
“Throughout this whole year I’ve been quarantining pretty hard, I haven’t really gone out much, so it’s not really going to be too much different for me, not really going to be too much of a lifestyle change for me, I’m used to it,” he said. “I plan on flying out the day before and just plan to quarantine until then. It’s a little bit nerve-wracking regardless. Anybody who makes this final table is going to feel a little bit nervous about it, but hopefully just don’t get COVID.”
For more on Hagerty, listen to his interview on last week’s episode of the PokerNews Podcast:
*Lead photo Borgata.Tim Kelliher.
I recently started a new newsletter focus on AI education. TheSequence is a no-BS( meaning no hype, no news etc) AI-focused newsletter that takes 5 minutes to read. The goal is to keep you up to date with machine learning projects, research papers and concepts. Please give it a try by subscribing below:
I had a long conversation with one of my colleagues about imperfect information games and deep learning this weekend and reminded me of an article I wrote last year so I decided to republish it.
Poker has remained as one of the most challenging games to master in the fields of artificial intelligence(AI) and game theory. From the game theory-creator John Von Neumann writing about poker in his 1928 essay “Theory of Parlor Games, to Edward Thorp masterful book “Beat the Dealer” to the MIT Blackjack Team, poker strategies has been an obsession to mathematicians for decades. In recent years, AI has made some progress in poker environments with systems such as Libratus, defeating human pros in two-player no-limit Hold’em in 2017. Last year, a team of AI researchers from Facebook in collaboration with Carnegie Mellon University achieved a major milestone in the conquest of Poker by creating Pluribus, an AI agent that beat elite human professional players in the most popular and widely played poker format in the world: six-player no-limit Texas Hold’em poker.
The reasons why Pluribus represents a major breakthrough in AI systems might result confusing to many readers. After all, in recent years AI researchers have made tremendous progress across different complex games such as checkers, chess, Go, two-player poker, StarCraft 2, and Dota 2. All those games are constrained to only two players and are zero-sum games (meaning that whatever one player wins, the other player loses). Other AI strategies based on reinforcement learning have been able to master multi-player games Dota 2 Five and Quake III. However, six-player, no-limit Texas Hold’em still remains one of the most elusive challenges for AI systems.
The challenge with six-player, no-limit Texas Hold’em poker can be summarized in three main aspects:
In AI theory, poker is classified as an imperfect-information environment which means that players never have a complete picture of the game. No other game embodies the challenge of hidden information quite like poker, where each player has information (his or her cards) that the others lack. Additionally, an action in poker in highly dependent of the chosen strategy. In perfect-information games like chess, it is possible to solve a state of the game (ex: end game) without knowing about the previous strategy (ex: opening). In poker, it is impossible to disentangle the optimal strategy of a specific situation from the overall strategy of poker.
The second challenge of poker relies on the difficulty of achieving a Nash equilibrium. Named after legendary mathematician John Nash, the Nash equilibrium describes a strategy in a zero-sum game in which a player in guarantee to win regardless of the moves chosen by its opponent. In the classic rock-paper-scissors game, the Nash equilibrium strategy is to randomly pick rock, paper, or scissors with equal probability. The challenge with the Nash equilibrium is that its complexity increases with the number of players in the game to a level in which is not feasible to pursue that strategy. In the case of six-player poker, achieving a Nash equilibrium is computationally impossible many times.
The third challenge of six-player, no-limit Texas Hold’em is related to its dependence on human psychology. The success in poker relies on effectively reasoning about hidden information, picking good action and ensuring that a strategy remains unpredictable. A successful poker player should know how to bluff, but bluffing too often reveals a strategy that can be beaten. This type of skills has remained challenging to master by AI systems throughout history.
Like many other recent AI-game breakthroughs, Pluribus relied on reinforcement learning models to master the game of poker. The core of Pluribus’s strategy was computed via self-play, in which the AI plays against copies of itself, without any data of human or prior AI play used as input. The AI starts from scratch by playing randomly, and gradually improves as it determines which actions, and which probability distribution over those actions, lead to better outcomes against earlier versions of its strategy.
Differently from other multi-player games, any given position in six-player, no-limit Texas Hold’em can have too many decision points to reason about individually. Pluribus uses a technique called abstraction to group similar actions together and eliminate others reducing the scope of the decision. The current version of Pluribus uses two types of abstractions:
To automate self-play training, the Pluribus team used a version of the of the iterative Monte Carlo CFR (MCCFR) algorithm. On each iteration of the algorithm, MCCFR designates one player as the “traverser” whose current strategy is updated on the iteration. At the start of the iteration, MCCFR simulates a hand of poker based on the current strategy of all players (which is initially completely random). Once the simulated hand is completed, the algorithm reviews each decision the traverser made and investigates how much better or worse it would have done by choosing the other available actions instead. Next, the AI assesses the merits of each hypothetical decision that would have been made following those other available actions, and so on.The difference between what the traverser would have received for choosing an action versus what the traverser actually achieved (in expectation) on the iteration is added to the counterfactual regret for the action. At the end of the iteration, the traverser’s strategy is updated so that actions with higher counterfactual regret are chosen with higher probability.
The outputs of the MCCFR training are known as the blueprint strategy. Using that strategy, Pluribus was able to master poker in eight days on a 64-core server and required less than 512 GB of RAM. No GPUs were used.
The blueprint strategy is too expensive to use real time in a poker game. During actual play, Pluribus improves upon the blueprint strategy by conducting real-time search to determine a better, finer-grained strategy for its particular situation. Traditional search strategies are very challenging to implement in imperfect information games in which the players can change strategies at any time. Pluribus instead uses an approach in which the searcher explicitly considers that any or all players may shift to different strategies beyond the leaf nodes of a subgame. Specifically, rather than assuming all players play according to a single fixed strategy beyond the leaf nodes, Pluribus assumes that each player may choose among four different strategies to play for the remainder of the game when a leaf node is reached. This technique results in the searcher finding a more balanced strategy that produces stronger overall performance.
Facebook evaluated Pluribus by playing against an elite group of players that included several World Series of Poker and World Poker Tour champions. In one experiment, Pluribus played 10,000 hands of poker against five human players selected randomly from the pool. Pluribus’s win rate was estimated to be about 5 big blinds per 100 hands (5 bb/100), which is considered a very strong victory over its elite human opponents (profitable with a p-value of 0.021). If each chip was worth a dollar, Pluribus would have won an average of about $5 per hand and would have made about $1,000/hour.
The following figure illustrates Pluribus’ performance. On the top chart, the solid lines show the win rate plus or minus the standard error. The bottom chart shows the number of chips won over the course of the games.
Pluribus represents one of the major breakthroughs in modern AI systems. Even though Pluribus was initially implemented for poker, the general techniques can be applied to many other multi-agent systems that require both AI and human skills. Just like AlphaZero is helping to improve professional chess, its interesting to see how poker players can improve their strategies based on the lessons learned from Pluribus.
Original. Reposted with permission.
Related: