We incentivize the prediction of future events. The prediction space is based on binary future events such as the ones listed on Polymarket. We are always actively expanding to new data providers. Most of this data is then processed by an LLM pipeline which handles the event generation. We are focused on judgemental forecasting rather than statistical forecasting. We hence expect the models used by miners to be LLMs.
Miners submit their predictions to validators. Each prediction has to be done early enough before the event underlying the prediction settles. Once the event settles, the validators that received the prediction score the miner.
Making predictions is a hard task that requires cross-domain knowledge and intuition. It is often limited in explanatory reasoning and domain-specific (the expert in predicting election results will differ from the one predicting the progress in rocket-engine technology) ([1]). At the same time it is fundamental to human society, from geopolitics to economics.
LLMs approach or surpass human forecasting abilities. They near on average the crowd prediction on prediction market events ([1]), and surpass humans in predicting neuroscience results ([2]). They are also shown to be calibrated with their predictions i.e confident when right. Through their generalization capabilities and unbounded information processing, LLMs have the potential to automate the prediction process or complement humans.
The value of the subnet first relies in the improvement of the efficiency of prediction markets. This value can be extracted by validators through arbitrage. The validators may obtain a better knowledge of the probability of an event settling and communicate this information to a prediction market by opening a position.
The first applications built on top of our subnet could be related to prediction markets. A trader could query our market to obtain the most up to date and relevant predictions to their portfolio based on the current news landscape (LLMs would be constantly aggregating the most up to date and relevant news articles). They could then readjust their positions accordingly or trade directly on this information.
In the long term, a validator could provide paid economic forecasts or more generally the output of any forward-looking task addressed to an LLM ([2]). A customer might then provide a series of paid sub-queries related to the information they aim at retrieving.
Miners compete by sending to the validators for each binary event
The prediction
A reference providing a baseline miner strategy is the article "Approaching Human Level Forecasting with Langage Models" ([1]). The authors fine-tune an LLM to generate predictions on binary events (including the ones listed on Polymarket) which nears the performance of human forecasters when submitting a forecast for each prediction, and which beats human forecasters in a setting where the LLM can choose to give a prediction or not based on its confidence.
According to the article, the performance of forecasting LLMs depends significantly on the amount of data one can retrieve for a given prediction or event (for example prediction market data). If our subnet is able to continually produce new synthetic data, miners should be able to beat the SoA.
Validators record the miners' predictions and score them once the events settle. At each event settlement, a score is added to the moving average of the miner's score. We implement a cutoff for the submission time of a prediction. The cutoff is set at 24 hours before the resolution date for most events.
For a binary event
-
$o_q = 1$ if the event is realized, -
$o_q = 0$ otherwise.
The Brier score
-
If
$o_q = 1$ :$$S(p_i, 1) = 1 - (1 - p_i)^2$$ -
If
$o_q = 0$ :
$$S(p_i, 0) = 1 - p_i^2.$$
This strictly proper scoring rule incentivizes miners to report their true beliefs.
For each event, the forecast period between the issue date and the resolution date is divided into
Once the event resolves, the validator computes the Brier score for each prediction in the time series, producing a corresponding series of scores. These scores are then aggregated into a weighted average, where later (more recent) predictions are weighted more heavily. Specifically, for submission interval
A detailed explanation of this process is available here.
In the updated system, the peer scoring mechanism replaces the legacy Brier score method. For each binary event, a miner
For miner (i), the peer score on a given event that resolves positively (
This calculation is performed for every submission interval, and the resulting peer scores are stored as a time series. Exacty as above, in case of no prediction being submitted a miner gets the worst possible score under the peer scoring rule. The same exponential weighting scheme is applied to compute a weighted average peer score for each miner, with the weight for interval
For each event, the weighted average peer score is then added to a moving average
Finally, the validator applies an extremising function to
-
Scoring with exponentially decreasing weights until settlement date and linear differentiation mechanism
-
Synthetic event generation with central resolution using ACLED data
-
Scoring with exponential differentiation mechanism
-
Comprehensive and granular analytics
-
Synthetic event generation with UMA resolution - human verifiers resolve our events through the OOv2
-
Synthetic event generation from news data using an LLM
-
Validator v2 - modular and much higher throughput
-
Scoring v2 (batches, peer score)
-
Exposing the silicon crowd predictions
-
Decentralisation of event generation and validator dynamic desirability (inspired from SN13)
-
Decentralisation of event generation and validator dynamic desirability (inspired from SN13)
-
Trustless event resolution
-
Commit-reveal on the miners' predictions
-
Scoring a reasoning component
-
Data generation for iterative fine-tuning of prediction focused LLMs
-
Reasoning component
-
Data generation for iterative fine-tuning of forecasting LLMs
Regarding instructions and requirements, see here for validators and here for miners.
A detailed explanation of how to set up a wallet can be found here. We also provide some indications here.
Reference ID | Author(s) | Year | Title |
---|---|---|---|
1 | Halawi and al. | 2024 | Approaching Human Level Forecasting with Langage Models |
2 | Luo and al. | 2024 | LLM surpass human experts in predicting neuroscience results |
This repository is licensed under the MIT License.
# The MIT License (MIT)
# Copyright © 2023 Yuma Rao
# Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated
# documentation files (the “Software”), to deal in the Software without restriction, including without limitation
# the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software,
# and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
# The above copyright notice and this permission notice shall be included in all copies or substantial portions of
# the Software.
# THE SOFTWARE IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO
# THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
# THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
# OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
# DEALINGS IN THE SOFTWARE.