Outlier Ventures Data-Driven Token Design and Optimization

Outlier Ventures Token Design and Optimization

Dr. Achim Struve | Speaker

Sissi | Translator

Editor’s Note:

In the token ecosystem, achieving sustainable development is crucial. Outlier Ventures’ latest video provides a comprehensive perspective on the key issues facing the token ecosystem and offers practical solutions and tools to address these challenges.

The video emphasizes the principles and methods of token engineering, which provide a new perspective for planning and constructing token systems. Additionally, a series of practical tools such as agent-based simulation tools, QTM, etc., are introduced. These tools can provide valuable information at different stages and help projects make wise decisions. With the help of these auxiliary tools, Web3 startups have the opportunity to achieve sustainable growth.

Outlier Ventures’ video brings us a new understanding, highlighting the crucial role of token engineering and related tools in project teams’ response to changes. These tools have proven to be powerful weapons for adapting to the ever-changing token ecosystem. This understanding is formed through in-depth research and practice of the token ecosystem, enabling participants to better understand the dynamics of the ecosystem and make wiser and more visionary decisions. The following is a translation and summary of the content of the video. For more content on token engineering, please refer to previous articles in this public account.

Three Stages of Token Design and Optimization

Discovery Stage

In building a successful token ecosystem, key steps need to be taken at the macro level of the ecosystem. First, the problem must be clearly defined, and the challenges faced must be clearly stated. Second, the flow of value between stakeholders needs to be clearly defined to ensure the robustness and balance of the ecosystem. At the same time, a thorough discussion and consideration of the entire ecosystem and the rationality of its tokens, including the reasonable use of tokens, are required. In addition, high-level planning is also indispensable, covering how to effectively use tokens and design solutions for various aspects. These key steps are essential elements in building a successful token ecosystem.

Design Stage

In the construction of the token ecosystem, parameterization is another key step that involves the use of quantitative tools such as spreadsheets and simulation tools like cadCAD, Token Spice, Machinations, etc. These tools can help people obtain optimized and validated models for risk analysis and prediction, thus gaining in-depth understanding of the trends in token supply and valuation. Through these quantitative tools, a better understanding of the operation of the ecosystem can be achieved, providing strong support for its design and optimization.

Deployment Stage

The deployment stage is crucial, as it translates the previous theoretical analysis and design into practice by actually deploying the ecosystem onto the blockchain. In this stage, various tools are used, including different programming languages such as Solidity, Rust, etc., and deployment environments such as Hardhat. Through this process, the final result is the actual ecosystem token or product, enabling it to be realized and operated on the blockchain.

Token Design Tools

In the three different stages (discovery, design, and deployment), we need to use a series of tools, and the focus and types of these tools can also vary in different fields. They are not only applicable to the DeFi field, but also to various application projects, infrastructure, games, and other fields.

When considering the details, there are two perspectives: one perspective believes that we can view the ecosystem from a qualitative perspective, and using market standards is enough without any simulation. The other perspective believes that a digital twin is needed to create a 1:1 simulation of the entire ecosystem because it involves a large amount of financial risk. As we move towards a more precise direction and increase resource intensity, the required programming knowledge will also increase. This also increases the requirements for users – they need to have programming skills to handle more complex models, which may affect user friendliness. Therefore, to build a more detailed ecosystem model, more programming knowledge is required, and a sufficient understanding of mathematics is also needed.

In the token ecosystem, there are various tools available to help us understand and design the system. On the left side of the “Token Design Tools” diagram, there are spreadsheet models and some qualitative tools, such as problem statements, stakeholder problem statements, stakeholder mapping, and specific value flows. We can even use AI-driven reasoning, such as using machine learning models to draft the first token design. In the middle part of the diagram, there is QTM (Quantitative Token Model), which is also a spreadsheet model, but it covers multiple different fields, not just DeFi. This wide coverage may result in a loss of accuracy, but it can indeed help startups gain first-hand insights and a preliminary understanding of their token ecosystem.

On the left side of the diagram, there are simulation tools such as cadCAD that can model the ecosystem 1:1 in complex environments. Overall, choosing the right tools and methods in the token ecosystem is crucial for the success of startups. Different types of tools can provide valuable information at different stages, helping companies make wise decisions and promoting the continuous development of the ecosystem.

Overview of QTM

QTM is a quantitative token model that adopts a fixed simulation time of 10 years, with each time step being one month, so it is more like a macro simulation model rather than a highly accurate model. At the beginning of each time step, tokens are emitted into the ecosystem, so the model includes incentive modules, token ownership modules, airdrop modules, etc. These tokens will then be distributed into several meta buckets for further refinement of generalized utility. Rewards payments are defined from these utility tools. In addition, aspects of off-chain business are also considered, such as general financial conditions of the business, which can include burning or repurchasing, as well as measuring user adoption or defining user adoption scenarios.

However, it should be emphasized that the quality of the model’s output depends on the quality of the input. Therefore, before using QTM, sufficient market research must be conducted to obtain more accurate input information and gain a deeper understanding of what is happening. This can result in output results that are closer to real-world situations. QTM is considered an educational tool for early-stage startups to help them understand their own ecosystem, but no financial advice should be derived from it, nor should it rely solely on its results.

Data Analysis

Next, let’s understand the different types of data that can be extracted from a data analysis perspective. First, we can observe the overall market development from a macro market perspective, including the DeFi market and the cryptocurrency market. Then, we can focus on indicators of fundraising rounds to understand the project’s financing situation, such as the amount of funds raised, valuation, and supply sales in different rounds. Second, we can also study the behavior patterns of participants to gain a deeper understanding of other people’s investment habits.

Compared to traditional finance, on-chain data has significant differences because on-chain data is publicly visible to everyone, and almost every transaction in the ecosystem can be viewed. As a result, various indicators can be obtained, such as user growth, total locked value (TVL), trading volume, etc. More importantly, we can observe how different incentive mechanisms affect the operation of the ecosystem. In addition, social media platforms such as Twitter, Reddit, Discord, and Telegram play important roles in token economics and project performance.

These pieces of information are publicly available and very valuable data. We should make full use of this data to better understand ecosystem parameters and validate our model.

Here is an example where we can look at data similar to vesting creations. Although this example is just a small sample, in general, vesting periods of different stakeholders can be observed. In the chart above, you can see the minimum, average, median, and maximum vesting periods, which are analyses of vesting periods in different domains. It can also be further segmented to differentiate different industry sectors. This way, you can see that the data distribution in different sectors may vary greatly. Although these numbers may not always be optimal, they provide us with a starting point.

Let’s take another example regarding the historical balance of a token bucket. Taking Maple Finance as an example, you can see the situation of its native token and track all transactions in the entire ecosystem, categorizing them into specific “token buckets,” such as addresses related to Maple, addresses related to centralized exchanges, and addresses related to decentralized exchanges, etc. In this way, we can view the balance of each stakeholder and observe what is happening in the entire ecosystem.

In this example, it may be found that the balance of all Maple addresses has been decreasing until mid-July 22nd, due to the introduction of the staking contract, which resulted in a large supply of tokens being allocated to the staking contract. We can also observe the participation of venture capitalists in this staking program, and this information can be directly inferred from the chart. In addition, we can also observe how the balance of exchanges changes over time, which is very helpful for understanding what is happening in the ecosystem. Finally, we can study the behavior of individual staking or specific addresses to gain a deeper understanding of the relevant situations.

In the token ecosystem, observing the behavior of specific addresses can provide important information about token liquidity. For example, when tokens are sent from the staking contract to specific addresses, we can understand how the recipients handle these tokens. Do they choose to reinvest these tokens, send them back to the staking contract, or sell them off or deploy them elsewhere? These are key pieces of information for understanding the behavior of each stakeholder through analysis, and we can feed this data back into our model to help adjust the model.

This model can not only analyze the behavior of token recipients at individual addresses, but also analyze representative aggregated stakeholder groups. For example, we can analyze multiple token projects such as Maple, Goldfinch, and TrueFi, and find that approximately 38% of tokens are sent back to the staking contract as the first transaction after being received through the staking contract. In contrast, the proportion for centralized exchanges is about 8%, and for decentralized exchanges, it is about 14%. By reviewing the token bucket allocation at a certain point in time on QTM, we can understand the circulation supply of tokens. These values can be applied to our parameters to gain a preliminary understanding of the behavior of the ecosystem.

With this data, we can make predictions, such as predicting the balance supply of different buckets in the ecosystem in the next decade or so, including foundations, teams, staking allocations, overall circulation supply, and liquidity pools, etc. Simulations or predictions of prices can also be conducted. It should be emphasized that these predictions are not for speculation or financial advice, but rather help us understand the relationship between supply attribution and token demand, in order to understand the balance between these two factors.

In addition, other aspects can also be analyzed, such as the allocation of different utility parts. For example, we can understand how many tokens are staked, how many are used for liquidity mining incentive programs, or if there is a burning mechanism, how many tokens are burned. If tokens can be used in stores or elsewhere, we can also observe monthly utility rewards to understand the value of these incentives in USD. Understanding the overall usage of tokens is very important, especially when considering cost factors in incentivizing the ecosystem.

Data-Driven Model

Another theme is a new way of thinking about token distribution plans. Sometimes people think that a very long token distribution plan is all that’s needed, but this is not always good because it means that the supply is very low in the initial circulation, leading to speculation and potential manipulation in the market. Therefore, we propose introducing a token distribution mechanism that adjusts the allocation, unaffected by market demand. In other words, predicting the demand of the ecosystem is not necessary because the distribution will be controlled by the controller based on certain predefined key performance indicators. These key performance indicators can include TVL, trading volume, user adoption rate, business profitability, and so on. In this example, the token price is simply used.

In the token ecosystem, the relationship between allocation and price can be understood by analyzing real token examples. For example, in the first year of the ecosystem, a large supply enters the market through allocation, but due to the product possibly not being mature enough and the market demand being insufficient, adoption may be low, resulting in a decrease in token price. This situation can be simulated using models like QTM, from which similar behavior can be observed: in the initial stage, the token price may decrease due to a large supply entering the market. Then, over time, as adoption changes and users start joining and bringing in increased revenue, buybacks can be conducted and the price will eventually recover.

In the model, three different demand scenarios can be simulated: logical function, linear function, and exponential growth. The controller essentially manages different emission levels at different time points, and it can be seen that for each different growth and demand scenario, the controller will manage different release amounts at different time points.

When the token price rises, more tokens will be released into the ecosystem, which may cause early investors to sell tokens and lead to a price decrease. Conversely, when the price is below the predetermined price, the token issuance will decrease. However, the token issuance will not drop to zero because we need to ensure that all initial investors eventually receive their fair share. Through this control mechanism, the token price will rise again, ultimately reducing volatility and stabilizing the ecosystem.

Price is a very important indicator in the ecosystem. A 90% price drop in a year would be very detrimental to the ecosystem. Although we know that we cannot accurately predict the future, we should at least consider the demand aspect and try to model and predict it. This does not mean pursuing a specific result or specific value, but exploring the entire available solution space through Monte Carlo simulations and parameter scans. Doing so can help us understand the possibilities in different scenarios and formulate more comprehensive and flexible strategies.

In addition, we can also allocate these ownerships in different weighted distributions. For example, in the initial stage, the ecosystem incentives may receive a larger allocation of tokens, while the team may receive a smaller share. However, over time, the situation may change because we do not want to rely solely on token ownership to drive the development of the ecosystem. We aim to establish a sustainable growth model.

Original video: https://www.youtube.com/watch?v=Yv7VvLVELCQ

We will continue to update Blocking; if you have any questions or suggestions, please contact us!

Share:

Was this article helpful?

93 out of 132 found this helpful

Discover more

Blockchain

How far are we from the real DAO?

Foreword: What is DAO? DAO is the abbreviation of Distributed Autonomous Organization, which is called "Decentra...

Blockchain

Do the bull engine start? After a year, the Bitcoin Bulls and Bears indicator "Super Gubby" first turned cattle

The Bitcoin daily chart recently issued a bullish signal, and the highly-accepted “super guppy” bull-be...

Blockchain

Cryptocurrency Essence Series: The Rise and Development of the Mining Industry

Original author: Chris Dannen, Leo Zhang and Martín Beauchamp, have worked iteration capital (Iterative Capital)...

Blockchain

After visiting Libra’s base (Switzerland), US legislators have not let go of their guard

According to foreign media, after visiting Switzerland to meet with the country's financial regulators, US legis...

Blockchain

Babbitt Column | Dusk of the Ethereum Empire

This article does not constitute any investment advice, investment is risky, and you need to be cautious when enterin...

Blockchain

In Devcon 5, Osaka, I saw Ethereum's diverse DeFi ecology.

Devcon5 is over, it's time to write a review! This time Tmono took me and Tina, Feng Liu, Cao and Suji to a grou...