More to come: learn.klimchitsky.com
This is great analytics, man!
We have the momentum and the resources, so no worries.
People will start to see that DFX is building some amazing things and honestly, I am excited to see it flourish.
Thank you for doing that.
It’s gonna be a tool for protocol analysis in the vein of what Gauntlet is doing, but based on a set of DeFi-native premises:
- Gauntlet is running their simulations with agent-based models representing users interacting with a protocol. A model of a user is based on a set of theoretical assumptions about user behaviour patterns. This approach was developed for TradFi, where the bulk of real life data is either not digitised at all (much of the b2c interaction happens offline) or isn’t available (much of market data is private). Hence, agent-based modelling with theoretical assumptions about incomplete data is justifiable. However, for DeFi agent-based modelling is a suboptimal legacy framework. Since all data about transactions and user interactions with the protocol is open and available for modelling, we can learn from real life data a model of a living protocol, or parts of it, and models of user interactions with it. Moreover, this model will be continuously fine-tuned with new data emerging.
- Transactions model is only half of the story. The other half is community sentiment manifested on twitter, discord and discourse. In offline economy inflation expectations and consumer sentiment influence consumer behaviour and central banks of the world when modelling national economies gauge it with polls. In DeFi we have the luxury to model community sentiment not with approximating polls, but again with real life data, while constantly fine-tuning the model.
We can learn from real life data models of onchain activity & of community sentiment and then merge them to get a true-to-life DeFi-native model of a protocol, which can be constantly fine-tune. Then it can be used to build tools for explorable + explainable DeFi serving purposes of both DAOs and DeFi investors: running stress tests and alternative scenarios, classifying protocols and tokens, detecting user behaviour patterns, making forecasts about certain protocol KPIs like TVL and that of partnering protocols.