- The paper introduces a framework and novel reward calculation technique for analyzing selfish mining profitability under general stochastic reward functions, capturing complex incentives like MEV.
- It analyzes β-cutoff selfish mining strategies using a Markov Chain and derives attacker profit equations, instantiated with a reward function combining block rewards, transaction fees, and random MEV.
- Results from instantiating a Bitcoin-like reward model demonstrate that considering transaction fees and MEV can significantly decrease the profitability threshold for selfish mining by over 50% compared to block rewards alone and pure selfish mining analysis respectively.
This paper introduces a framework for analyzing strategic behavior in Proof-of-Work consensus mechanisms under general miner reward functions, which can be stochastic, variable in time, and/or ephemeral. These reward functions aim to capture various existing reward sources, including Miner/Maximal Extractable Value (MEV), in modern blockchains. The paper focuses on analyzing the profitability of cutoff selfish mining strategies for reward functions identically distributed across forks, employing a novel reward calculation technique to capture non-linearity in general rewards.
The authors instantiate these results in a combined reward function that represents miner incentives in Bitcoin today. This reward function includes block rewards, linear-in-time transaction fees, and a third random reward based on observed transaction fee spikes. The instantiation enables qualitative observations, quantitative claims, and confirmation of the theoretical analysis using Monte Carlo simulations.
Key contributions of the paper include:
- A general reward function model to capture the aggregate incentives for following a specific strategy under distinct revenue streams.
- A set of properties characterizing subtleties of blockchain rewards.
- A technique to calculate expected attacker profit given an aggregate reward function under mild assumptions about the distribution of the constituent reward sources.
- An instantiated reward function combining block reward, transaction fee, and MEV rewards.
The paper begins by defining a stylized model of Proof-of-Work mining with general stochastic rewards, referred to as the Nakamoto Consensus Game (NCG). The NCG models the set of miners M, where each miner m∈M has hashrate αm. At any time t, there is a public view Vt which consists of the state of the blockchain known to all miners, and a private view Vtm for each miner m which includes Vt and any additional blocks m knows about. Miners are rewarded for creating blocks on the eventual longest chain, which is modeled as a reward function Rm(t,V,B,r,B′)→R, where t is the time, V is the view, B is a block in V, r is randomness, and B′ is a block created by miner m. Each miner m has a strategy that takes as input a time t, a view Vtm, and the reward Rm(t,Vtm,B,r,B′) for extending each block B∈Vtm by a valid block B′∈Bm(t,Vtm,B,r), and outputs (i) a block B∈Vtm to mine on, (ii) contents of the next block B′∈Bm(t,Vtm,B,r), and (iii) a (potentially empty) subset of blocks in Vtm∖Vt to broadcast.
The paper also defines properties of the reward functions, including:
- Miner-Independent Rewards: A reward function R is miner-independent if the set of valid views, the set of valid blocks extending each block in those views, and equal rewards from any such valid block are the same for all miners.
- View-Independent Rewards: A reward function R is view-independent if the probability of reward x extending block B1 with timestamp t′ in view V1 is the same as the probability of reward x extending block B2 with timestamp t′ in view V2.
- Static Rewards: A reward function R is static if the probability of reward x extending block B1 at time t1 with timestamp t1−Δ in view V1 is the same as the probability of reward x extending block B2 at time t2 with timestamp t2−Δ in view V2.
- Persistent Rewards: A reward function R is persistent if for all blocks B′ mined at time t extending block B and resulting in view V, R(t,V,B,r,B′) is bound above by the total rewards minus the sum of claimed rewards on the ancestral chain of B.
The paper then presents examples of reward functions, including transaction fees and Loss-Versus-Rebalancing (LVR), and uses them to illustrate the properties.
The paper analyzes β-cutoff selfish mining strategies, where the attacker withholds blocks if the rewards they earn are below a threshold β. The attacker follows the longest chain and claims all available rewards but withholds any block found where the time since parent is less than β. The paper introduces the β-cutoff Markov Chain, which captures the states of the attacker's hidden chain length.
Using this Markov Chain, the stationary distribution pi is calculated as the probability of being in State i. The stationary distribution measures the probability that a block produced in the Markov Chain is orphaned, \begin{align*}
\lambda = p_1 (1-\alpha) \left(1+\frac{\alpha}{1-2\alpha}\right).
\end{align*}
The per-state attacker rewards, fi, are then calculated as the expected reward of a canonicalized attacker block mined in State i.
For all states i≥2, the expected attacker rewards collected in State i is given by:
\begin{align*}
f_i &= \sum_{j=0}{i-1} \left[\alpha (1-\alpha)j \int_{0}\infty \frac{tj e{-t/(1-\lambda)}}{j!(1-\lambda){j+1}} \mathbb{E}_{r}[R(t)]dt\right]
\end{align*}
The attacker's reward is then calculated as:
\begin{align*}
\text{ATTACKER REWARD} =f_0p_0 + f_1 p_1 + \alpha \sum_{i=2}\infty f_i p_{i-1}.
\end{align*}
The paper instantiates an aggregate reward function, R^, composed of (1) a fixed block reward of size C, (2) a linear-in-time transaction fee reward, and (3) a random "extra" reward of size E awarded to a block based on the outcome of a Bernoulli trial with probability p.
R^ is then defined as:
\begin{align*}
\hat{R}(t) &= C + t + E \cdot \mathds{1}[X=1], \; X\sim \text{Bernoulli}(p).
\end{align*}
With this combined reward function, the paper analyzes the transition probabilities, stationary distribution, and expected attacker rewards.
For example, it was demonstrated that the profitability threshold of β−cutoff selfish mining decreases by about 22\% and another 31\% compared to pure selfish mining when considering other rewards at γ=0.
The paper concludes by discussing potential future research directions, including applying the methodology more broadly to other consensus protocols and expanding the strategy space to capture more realistic mining strategies.
The paper also includes several appendices which provide further details on the derivations and calculations used in the analysis.
- Appendix A: Proof of Lemmas
- Appendix B: Derivation of f3
- Appendix C: Evaluation of f0,(i),f0,(ii),f0,(iii) Integrals
- Appendix D: Derivation of Full Attacker Reward Equation
- Appendix E: Block Rewards Only
- Appendix F: Linear-in-Time Transaction Fees