AI

Tokenomics for AI Arena: Rewarding Collaboration and Performance

Jan 10, 2025

Tokenomics for AI Arena: Rewarding Collaboration and Performance

AI Arena leverages the FLock framework to create a dynamic, fair, and incentivized ecosystem for participants in decentralized machine learning. By aligning rewards with performance, contribution, and community consensus, FLock ensures sustained engagement and system integrity.

The total supply of FLOCK, our native token, is capped at 1 billion tokens. Only the day-one unlocked tokens are minted initially, with additional tokens minted daily through our management smart contract. To know more about the airdrop and token allocation mechanism, visit here.

Key Features of AI Arena Tokenomics

1. Rewarding Honesty and Contribution

FLock incentivizes participantsโ€”training nodes, validators, and delegatorsโ€”to contribute meaningfully. Rewards are determined by:

  • Accumulative staking amounts: Stake size influences the reward pool.
  • Relative performance: Ranked quality of submissions.
  • Community governance: DAO voting sets key parameters like stake weight.

2. Dynamic Token Emission

Daily token emissions fuel rewards, with a flexible ratio split between AI Arena and FL Alliance. Once tasks are created, the distribution of rewards between DAO-verified AI Arena and FL Alliance tasks is dependent on their relative stake amount of active tasks. This blog post focuses only on AI Arena for now, so we will save FL Alliance for another time!

Participant Roles and Rewards

Training Nodes


Training nodes fine-tune AI tasks and ensure ecosystem health through staking and high-quality submissions.

  • Reward Breakdown:
  1. Reward A: Daily emitted rewards.
  2. Reward B: Vested rewards released after task completion.
  • Reward Drivers:
  1. Relative stake in a task.
  2. Delegated stake contributions.
  3. Submission quality, determined by the ranking based on the consensus results of validators.

Formula: 

Step 1: Reward distribution within a single AI Arena task

Within a single AI Arena task, the reward distribution between training nodes and validators is determined based on their relative stake amounts

We assume there are ๏ปฟ๐‘›๐‘›๏ปฟ submissions ๏ปฟ(๐‘‚1,...,๐‘‚๐‘›)(๐‘‚1,...,๐‘‚๐‘›)๏ปฟ from ๏ปฟnn๏ปฟ training nodes with stakes ๏ปฟ(๐‘ก1,...,๐‘ก๐‘›)(๐‘ก1,...,๐‘ก๐‘›)๏ปฟ, and ๏ปฟmm๏ปฟ validators ๏ปฟ(๐‘‰1,...,๐‘‰๐‘š)(๐‘‰1,...,๐‘‰๐‘š)๏ปฟ with stakes ๏ปฟ(๐‘ 1,...,๐‘ ๐‘š)(๐‘ 1,...,๐‘ ๐‘š)๏ปฟ. Each validator ๏ปฟ๐‘‰๐‘—(1โ‰ค๐‘—โ‰ค๐‘š)๐‘‰๐‘—(1\le๐‘—\le๐‘š)๏ปฟ evaluates the ๐‘› models submitted by the training nodes.

Let the total daily reward allocated to a task be denoted as R0โ€‹ and the parameter ฮณ controls the split rewards, defining the balance between fixed and stake-dependent reward components.

The total rewards for training nodes are:

๏ปฟR0โ‹…(ฮณ+(1โˆ’2ฮณ)โ‹…โˆ‘j=1msjโˆ‘i=1nti+โˆ‘j=1msj)R_0\cdot\left(\gamma+(1-2\gamma)\cdot\frac{\sum_{j=1}^ms_j}{\sum_{i=1}^nt_i+\sum_{j=1}^ms_j}\right)๏ปฟ

Step 2: Rewards for training nodes and their delegators

We can now compute the total rewards allocated for the training nodes as well as their delegators, which is based on the quality of their submission and their total amount of stake: 

๏ปฟfi(gi,ti)=giโ‹…tiฮฑtโˆ‘k=1ngkโ‹…tkฮฑtf_i(g_i,t_i)=\frac{g_i\cdot t_i^{\alpha_t}}{\sum_{k=1}^ng_k\cdot t_k^{\alpha_t}}๏ปฟ

In which ๐‘ก๐‘– the total stake amount from the training node ๐‘– as well as its respective delegators, gi is the scores of the submitted models from training node, whereas ๐‘˜ denotes a given training nodeโ€™s rank amongst its peers in the same task. On theother hand, ๐›ผ๐‘ก is a system parameter that determines the influence of the stake on the reward distribution.

Step 3: Rewards for training nodes

If a training node ๐‘–โ€™s stake in the task is ๐‘ก๐‘› and stakes delegated to training node ๐‘– is ๐‘ก๐‘‘ , i.e., ๐‘ก๐‘– = ๐‘ก๐‘› + ๐‘ก๐‘‘, then the actual reward for training node ๐‘– is

๏ปฟfiโ‹…(ฯƒ+(1โˆ’ฯƒ)โ‹…tntn+td)f_i\cdot\left(\sigma+(1-\sigma)\cdot\frac{t_n}{t_n+t_d}\right)๏ปฟ

Note that in the front-end, you will see a โ€œreward-sharing ratioโ€, which refers to (1 - ๐œŽ), which means when reward-sharing ratio is 60%, ๐œŽ is 0.4. This ratio is set by training nodes and validators permissionlessly.

๏ปฟExample: 

Letโ€™s assume daily total rewards for all AI Arena tasks for a given day is 309,157.68. We have 1 task with 2 nodes and 3 validators.

Nodes A and B stake 3,000 and 3,500 FLOCK respectively, while validators A, B and C stake 3,000, 6,000 and 3,000 respectively. Node A also receives an additional 1,000 FLOCK from its delegators, which brings the ๐‘ก๐‘– (total stake including delegated stake) to be 4,000 for Node A. For simplicity, we assume ฮณ to be 0 in this example.

First, for this given task, total rewards for *all* training nodes are:

๏ปฟR0ร—โˆ‘i=1ntiโˆ‘i=1nti+โˆ‘j=1msj=309,157.68ร—65006500+12000โ‰ˆ108,623.7R_0\times\frac{\sum_{i=1}^nt_i}{\sum_{i=1}^nt_i+\sum_{j=1}^ms_j}=309,157.68\times\frac{6500}{6500+12000}\approx108,623.7๏ปฟ

We can then compute the rewards for *Node A and its delegators*. We are assuming that the scores for Node A and B are 0.501435 and 0.498565 respectively. Consider ฮฑtโ€‹=1, rewards for Node A (together with delegators) are: 

๏ปฟfi(gi,ti)=giโ‹…tiโˆ‘k=1ngkโ‹…tk=0.501435ร—4000(0.501435ร—4000)+(0.498565ร—3500)ร—108,623.7=58,084f_i(g_i,t_i)=\frac{g_i\cdot t_i}{\sum_{k=1}^ng_k\cdot t_k}=\frac{0.501435\times4000}{(0.501435\times4000)+(0.498565\times3500)}\times108,623.7=58,084๏ปฟ

Finally, given ๐œŽ=0.4, the actual rewards for *Node A alone* is:

๏ปฟfiโ‹…(ฯƒ+(1โˆ’ฯƒ)tntn+td)=58,084ร—(0.4+0.6ร—30004000)=49,371.40f_i\cdot\left(\sigma+(1-\sigma)\frac{t_n}{t_n+t_d}\right)=58,084\times\left(0.4+0.6\times\frac{3000}{4000}\right)=49,371.40๏ปฟ


2. Validators

Validators evaluate submissions to maintain quality and reliability. Their rewards depend on:

  1. Validation quality: Alignment with majority consensus.
  2. Stake size: Validatorโ€™s direct and delegated stakes.
  3. Validation volume: Number of successfully validated submissions.

Formula: 

Step 1: Reward distribution within a single AI Arena task

Same as how reward distribution is calculated for training nodes as explained above, the rewards for validators in the same given AI Arena task is:

๏ปฟR0โ‹…(ฮณ+(1โˆ’2ฮณ)โ‹…โˆ‘i=1ntiโˆ‘i=1nti+โˆ‘j=1msj)R_0\cdot\left(\gamma+(1-2\gamma)\cdot\frac{\sum_{i=1}^nt_i}{\sum_{i=1}^nt_i+\sum_{j=1}^ms_j}\right)๏ปฟ

Step 2: Reward distribution for validators

If the validator finishes multiple (i.e., ๐‘ ) validation tasks, then its reward is:

๏ปฟfiโ‹…Fโ‹…(ฯƒ+(1โˆ’ฯƒ)โ‹…SvSv+Sd)f_i\cdot F\cdot\left(\sigma+\left(1-\sigma\right)\cdot\frac{S_v}{S_v+S_d}\right)๏ปฟ

Specifically,

  • F refers to the performance of the validation, which is calculated through off-chain consensus
  • fi is the rewards for all validators
  • ฯƒ is the guaranteed validator return against delegators
  • Svโ€‹ is the stake amount of this validator
  • Sd is the stake amount delegated to this validator

Example: 

Continuing with our example above, total rewards for *all validators* will be:

๏ปฟR0ร—โˆ‘j=1msjโˆ‘i=1nti+โˆ‘j=1msj=309,157.68ร—120006500+12000โ‰ˆ200,534.0R_0\times\frac{\sum_{j=1}^ms_j}{\sum_{i=1}^nt_i+\sum_{j=1}^ms_j}=309,157.68\times\frac{12000}{6500+12000}\approx200,534.0๏ปฟ

If we assume validator A has a score of 0.472768, the the reward for * validator A only* is:

๏ปฟfiโ‹…Fโ‹…(ฯƒ+(1โˆ’ฯƒ)SvSv+Sd)=200,534ร—0.369ร—(0.4+0.6ร—30003000+0)โ‰ˆ73,997f_i\cdot F\cdot\left(\sigma+(1-\sigma)\frac{S_v}{S_v+S_d}\right)=200,534\times0.369\times\left(0.4+0.6\times\frac{3000}{3000+0}\right)\approx73,997๏ปฟ

3. Delegators

Delegators support training nodes and validators by staking. Rewards are based on:

  1. Stake quality: Contributions of the selected node/validator.
  2. Stake size: Amount staked by the delegator.

Formula: 

If a delegator has delegated Sdโ€‹ to the above-mentioned Node A, then the daily return for the delegator in this task is:

๏ปฟfiโ‹…(1โˆ’ฯƒ)โ‹…tdtn+tdf_i\cdot(1-\sigma)\cdot\frac{t_d}{t_n+t_d}๏ปฟ

Similarly, if a delegator delegates to a validator, the reward will be:

๏ปฟfiโ‹…(1โˆ’ฯƒ)โ‹…sdsv+Sdf_i\cdot(1-\sigma)\cdot\frac{s_d}{s_v+S_d}๏ปฟ

Example: 

Continuing the example for Node A, for a given delegator delegated 1,000 FLOCK to Node A, the reward for *this delegator alone* is:


๏ปฟfiโ‹…(1โˆ’ฯƒ)โ‹…tdtn+td=58,084ร—(0.6ร—10003000+1000)โ‰ˆ8,712.6f_i\cdot(1-\sigma)\cdot\frac{t_d}{t_n+t_d}=58,084\times\left(0.6\times\frac{1000}{3000+1000}\right)\approx8,712.6๏ปฟ

Note that rewards in delegation pools are time-weighted to balance fairness for long-term participants and incentivize new delegations. As pools grow, rewards stabilize, promoting sustained engagement. Also, delegators must maintain their stake for at least 24 hours before un-delegating. This ensures meaningful contributions and prevents exploitative behaviors.

Parameters like reward splits (ฮณ) are fine-tuned through DAO voting. This democratized control keeps the ecosystem adaptive and equitable.

To participate in and know more about AI Arena

To participate in AI Arena, navigate to our website. If you are interested in learning more about FLOCKโ€™s tokenomics, visit our docs or whitepaper

Our smart contracts underpinning this system are open-source and auditable. Explore the details here.

Weekly newsletter

No spam. Just the latest releases and tips, interesting articles, and exclusive interviews in your inbox every week.

Read about our privacy policy