Navigating Borrow Rates Trends in DeFi: Insights from the AAVE & Compound Ecosystem
The evolving landscape of decentralized finance (DeFi) continues to draw both intrigue and investment from the broader cryptocurrency community. Among the myriad developments, one trend that stands out is the increasing borrowing rates of DeFi assets across lending protocols. This shift has implications for investors, borrowers, and the DeFi ecosystem at large. In this blog, we'll delve into the reasons behind these rising rates, compare various assets on the Ethereum network, and analyz...
State of Personalization [Web2 v/s Web3]
Personalization has been a driving force behind the success of Web2 platforms, shaping user experiences and setting new standards for digital interaction. Whether it's the curated playlists on Spotify, product recommendations on Amazon, or the tailored content feeds on social media, personalization drives engagement to the extent that we almost take it for granted. However, such personalization is fundamentally absent in the web3 ecosystem, despite the wealth of available on-chain data. ...
State of DeFi Leverage: Jan'24
2024 started strong for DeFi, with a mix of events through the first month. Restaking was one of the strongest narratives within the sector witnessing over $1.7B worth of inflows.Source: DeFiLlamaThere were multiple developments within the leverage landscape throughout the month.Starting with MakerDAO passing two executive votes, with the key highlight being an increase of Spark’s DAI debt ceiling from 800M to 1.2B.Morpho labs continued to ship, with Blockanalytica together with B.protocol co...
Leverage liquidity layer of the DeFi stack.
Navigating Borrow Rates Trends in DeFi: Insights from the AAVE & Compound Ecosystem
The evolving landscape of decentralized finance (DeFi) continues to draw both intrigue and investment from the broader cryptocurrency community. Among the myriad developments, one trend that stands out is the increasing borrowing rates of DeFi assets across lending protocols. This shift has implications for investors, borrowers, and the DeFi ecosystem at large. In this blog, we'll delve into the reasons behind these rising rates, compare various assets on the Ethereum network, and analyz...
State of Personalization [Web2 v/s Web3]
Personalization has been a driving force behind the success of Web2 platforms, shaping user experiences and setting new standards for digital interaction. Whether it's the curated playlists on Spotify, product recommendations on Amazon, or the tailored content feeds on social media, personalization drives engagement to the extent that we almost take it for granted. However, such personalization is fundamentally absent in the web3 ecosystem, despite the wealth of available on-chain data. ...
State of DeFi Leverage: Jan'24
2024 started strong for DeFi, with a mix of events through the first month. Restaking was one of the strongest narratives within the sector witnessing over $1.7B worth of inflows.Source: DeFiLlamaThere were multiple developments within the leverage landscape throughout the month.Starting with MakerDAO passing two executive votes, with the key highlight being an increase of Spark’s DAI debt ceiling from 800M to 1.2B.Morpho labs continued to ship, with Blockanalytica together with B.protocol co...
Leverage liquidity layer of the DeFi stack.
Share Dialog
Share Dialog

Subscribe to Lucidity Finance

Subscribe to Lucidity Finance
As we building towards powering on-chain recommendations, some of the most frequent and important questions we hear are around the quantity and quality of on-chain data available to power meaningful recommendations. While model architectures play an important role towards addressing those concerns, that only takes us half way.
Data is one of the three fundamental neural scaling constraints and this blog aims to deconstruct the key challenges, processes and considerations in building a rich and robust data pipeline from onchain model builder’s standpoint. The other two constraints are compute and parameters (model size), but more on them later.
First things first, building a data pipeline for AI models involves multiple steps, each with its own set of challenges, more so when you’re targeting onchain usecases.
This includes gathering data from a variety of onchain / offchain sources relevant to the use case (in our case any datasets that can contribute to characterisation of a wallet/asset/protocol smart contracts).
Onchain data is simply not being rich enough given the limited history and engagement as compared to Internet data. For instance, Ethereum has processed ~2.5 billion transactions till date; only ~8 billion even if we consider top 10 chains. Compare this with ~15 trillion data tokens from publicly available sources used to pre-train Llama 3 on.
DataDAOs: Decentralized networks enabling community-owned data ecosystems. Over the past year, several DataDAOs have emerged, aggregating diverse datasets across domains like finance, healthcare, and social media. Projects such as Masa, Vana, Grass, etc have gained significant traction, facilitating user-owned data collection and management across varied sectors.
Synthetic data: Artificially generating datasets that mirror real-world data, in a privacy-preserving manner. While its scalability still needs to be seen, synthetic data can be a very cost-effective to the AI model’s rapidly increasing data needs. Teams like Firstbatch, Openledger, Mizu have been building in this direction.
Once the data is collected, it needs to be cleaned to make it usable to training AI models, which involves multiple steps. The goal here is to ensure data integrity by removing noise and inconsistencies, adding context and structure, and refining outputs for usability in downstream applications.
Blockchain data can be noisy and incomplete. The data collected across different chains is also often not standardized and not interoperable. Furthermore, incentivised crowdsourcing of data can lead to corrupt user behaviours and sybil attacks. The labelling process can also be time consuming, costly, and lead to inherent biases in resulting datasets.
Labelling: There have been interesting developments around building incentivised networks of massive-scale, community-owned, labelled datasets and trustless verifications to eliminate biases. Teams including FractionAI, Kiva, Synesis, etc. have been leading the charge on this front.
Data aggregators and marketplaces: Enable permissionless exchange and monetisation of tokenised and processed datasets across ecosystems. Teams including Ocean Protocol, 0xScope have made notable strides, offering frameworks for data processing, sharing and pricing in a decentralized manner.
This involves secure and scalable storage and handling of the collected data. Key considerations include maximising throughputs and minimising latency while managing growing data volumes queried across distributed systems. This is essential for large-scale model training and serving.
Onchain AI applications and agentic networks will demand high throughputs and data availability. Siloed data clusters and unavailability of high-performant data storage solutions can massively impede the development of onchain intelligence networks.
DA layers: 0G’s DA layer offers infinitely-scalable infrastructure specifically focused on handling vast datasets required by AI models, queriable across its own storage and also third party storage solutions.
Storage solutions: Solutions like Hyperline’s unified data lakehouse, and Arweave’s permanent, tamper-proof storage, can enable seamless preservation and querying of datasets used in model training and auditing.
AI models building need increasing amounts of offchain user data (in addition to onchain data) to enable meaningfully personalised onchain interactions. That involves ensuring that the data collectd is authentic, tamper-proof, and doesn’t contain users’ PII (personally identifiable information).
The privacy and the security of the user-owned data are serious concerns for any type of data collection task, more so with high-volume of meaningful data required for model training. Verification of data authenticity is also an equally daunting task especially when dealing with crowdsourced/offchain data.
ZK-proofs: Space and Time’s zk-proven SQL-based data warehouse combines a decentralized data warehouse with “Proof of SQL” protocol to allow querying onchain/offchain data and verify the result in a trustless manner. This enables enforcing verifiability of inputs at scale.
Privacy: Provably focuses on privacy-preserving, verifiable analytics to enable AI and data teams to collaborate on live private data collections. This Developments in FHE (Fully Homomorphic Encryption) for privacy-preserving machine learning are also worth mentioning here.
The foundations of a robust data stack are being steadily established to meet the growing demand for rich, diverse, scalable, and verifiable data essential for onchain intelligence networks. With advancements in data collection, augmentation, storage, querying, and verification, the ecosystem is evolving to address the complexities of powering onchain AI. The challenges, though significant, seem surmountable.
Lucidity, focused on enabling personalized recommendations—one of the most impactful onchain AI applications—is dedicated to working with innovative teams across the data pipeline. Together, we aim to drive meaningful progress toward a more inclusive, transparent, and efficient Web3 ecosystem.
As we building towards powering on-chain recommendations, some of the most frequent and important questions we hear are around the quantity and quality of on-chain data available to power meaningful recommendations. While model architectures play an important role towards addressing those concerns, that only takes us half way.
Data is one of the three fundamental neural scaling constraints and this blog aims to deconstruct the key challenges, processes and considerations in building a rich and robust data pipeline from onchain model builder’s standpoint. The other two constraints are compute and parameters (model size), but more on them later.
First things first, building a data pipeline for AI models involves multiple steps, each with its own set of challenges, more so when you’re targeting onchain usecases.
This includes gathering data from a variety of onchain / offchain sources relevant to the use case (in our case any datasets that can contribute to characterisation of a wallet/asset/protocol smart contracts).
Onchain data is simply not being rich enough given the limited history and engagement as compared to Internet data. For instance, Ethereum has processed ~2.5 billion transactions till date; only ~8 billion even if we consider top 10 chains. Compare this with ~15 trillion data tokens from publicly available sources used to pre-train Llama 3 on.
DataDAOs: Decentralized networks enabling community-owned data ecosystems. Over the past year, several DataDAOs have emerged, aggregating diverse datasets across domains like finance, healthcare, and social media. Projects such as Masa, Vana, Grass, etc have gained significant traction, facilitating user-owned data collection and management across varied sectors.
Synthetic data: Artificially generating datasets that mirror real-world data, in a privacy-preserving manner. While its scalability still needs to be seen, synthetic data can be a very cost-effective to the AI model’s rapidly increasing data needs. Teams like Firstbatch, Openledger, Mizu have been building in this direction.
Once the data is collected, it needs to be cleaned to make it usable to training AI models, which involves multiple steps. The goal here is to ensure data integrity by removing noise and inconsistencies, adding context and structure, and refining outputs for usability in downstream applications.
Blockchain data can be noisy and incomplete. The data collected across different chains is also often not standardized and not interoperable. Furthermore, incentivised crowdsourcing of data can lead to corrupt user behaviours and sybil attacks. The labelling process can also be time consuming, costly, and lead to inherent biases in resulting datasets.
Labelling: There have been interesting developments around building incentivised networks of massive-scale, community-owned, labelled datasets and trustless verifications to eliminate biases. Teams including FractionAI, Kiva, Synesis, etc. have been leading the charge on this front.
Data aggregators and marketplaces: Enable permissionless exchange and monetisation of tokenised and processed datasets across ecosystems. Teams including Ocean Protocol, 0xScope have made notable strides, offering frameworks for data processing, sharing and pricing in a decentralized manner.
This involves secure and scalable storage and handling of the collected data. Key considerations include maximising throughputs and minimising latency while managing growing data volumes queried across distributed systems. This is essential for large-scale model training and serving.
Onchain AI applications and agentic networks will demand high throughputs and data availability. Siloed data clusters and unavailability of high-performant data storage solutions can massively impede the development of onchain intelligence networks.
DA layers: 0G’s DA layer offers infinitely-scalable infrastructure specifically focused on handling vast datasets required by AI models, queriable across its own storage and also third party storage solutions.
Storage solutions: Solutions like Hyperline’s unified data lakehouse, and Arweave’s permanent, tamper-proof storage, can enable seamless preservation and querying of datasets used in model training and auditing.
AI models building need increasing amounts of offchain user data (in addition to onchain data) to enable meaningfully personalised onchain interactions. That involves ensuring that the data collectd is authentic, tamper-proof, and doesn’t contain users’ PII (personally identifiable information).
The privacy and the security of the user-owned data are serious concerns for any type of data collection task, more so with high-volume of meaningful data required for model training. Verification of data authenticity is also an equally daunting task especially when dealing with crowdsourced/offchain data.
ZK-proofs: Space and Time’s zk-proven SQL-based data warehouse combines a decentralized data warehouse with “Proof of SQL” protocol to allow querying onchain/offchain data and verify the result in a trustless manner. This enables enforcing verifiability of inputs at scale.
Privacy: Provably focuses on privacy-preserving, verifiable analytics to enable AI and data teams to collaborate on live private data collections. This Developments in FHE (Fully Homomorphic Encryption) for privacy-preserving machine learning are also worth mentioning here.
The foundations of a robust data stack are being steadily established to meet the growing demand for rich, diverse, scalable, and verifiable data essential for onchain intelligence networks. With advancements in data collection, augmentation, storage, querying, and verification, the ecosystem is evolving to address the complexities of powering onchain AI. The challenges, though significant, seem surmountable.
Lucidity, focused on enabling personalized recommendations—one of the most impactful onchain AI applications—is dedicated to working with innovative teams across the data pipeline. Together, we aim to drive meaningful progress toward a more inclusive, transparent, and efficient Web3 ecosystem.
<100 subscribers
<100 subscribers
No activity yet