# Roadmap

## 2025: Establishing the AI Data Foundation

In 2025, LazAI will focus on developing the foundational components of its ecosystem: launching an AI agent for data alignment, deploying a high-performance blockchain for AI execution, and implementing data-driven PoS + Quorum-Based BFT consensus. This year will be testnet-focused, allowing developers to experiment with AI data anchoring and verification mechanisms.

### Phase 1: AI Agent Alith (Q1-Q2 2025) – AI-Powered Data Coordination

LazAI will launch Alith, a simple, composable, high-performance, and Web3-friendly AI agent framework designed to facilitate decentralized AI data processing, governance, and alignment. Unlike traditional AI agents that run in silos, Alith will interact with LazAI on-chain and off-chain AI data sources, ensuring data provenance and verification.

* [x] **Multimodal Model and Data Interpretation:** Supports text-, image-, and voice-based data to enable seamless AI model interaction with decentralized sources.
* [x] **Deployment and Workflow:** SDKs for Rust, Python, and Node.js will enable custom AI data processing and application workflows.
* [x] **High Performance on-Chain Inference:** Leverages techniques including graph optimization, model compression, use of GPU coprocessors, and JIT/AOT compilation for high-performance inference.
* [x] **Data Management:** Alith will help users coordinate, verify, and build decentralized AI datasets for model training and inference based on LazAI.

**Successfully Delivered:** Establish AI-native data workflows, allowing developers to access verified, decentralized AI datasets for training and inference, and build high-performance AI Agent applications.

{% hint style="success" %}
For more info, visit Alith [website](https://lazai.network/alith) and [docs](https://alith.vercel.app/docs).&#x20;
{% endhint %}

### Phase 2: Testnet – AI Data Settlement & Blockchain Infrastructure (Q2-Q3 2025)

LazAI will launch its testnet, providing a blockchain environment optimized for AI data integrity, validation, and settlement. This testnet will serve as a sandbox for developers experimenting with AI data tokenization, provenance tracking, and alignment verification.

* [ ] **Data Anchoring Token (DAT):** A new asset standard that tokenizes AI datasets and training pipelines, ensuring verifiability.
* [ ] **AI Data Processing and Inference/Training Workflow:** Supports high-throughput, parallelized AI data transactions, reducing latency for model updates and training workflows.
* [ ] **LazAI Verified Computing Framework:** Ensuring the authenticity, integrity, and verifiability of AI data based on DAT protocol is critical to building a trustworthy AI ecosystem.&#x20;

**Expected Outcome:** Developers gain access to a secure, scalable blockchain infrastructure where AI data can be registered, exchanged, and verified with on-chain provenance guarantees.

{% hint style="success" %}
DAT is live on Pre-Testnet, visit [here](https://predat.lazai.network/).&#x20;
{% endhint %}

### Phase 3: Mainnet – AI Data-Driven PoS + Quorum-Based BFT Consensus (Q3-Q4 2025)

LazAI will officially launch its mainnet, introducing a hybrid consensus protocol tailored for AI data verification and governance.

* [ ] **Quorum-Based BFT Consensus:** AI data alignment and model verification will be validated through a decentralized quorum-based mechanism.
* [ ] **PoS Economic Model for AI Data Validation:** Validators will be required to stake tokens to secure AI data pipelines, ensuring accountability and economic incentives for honest AI verification.
* [ ] **Decentralized AI Data Arbitration:** iDAO-led quorums will resolve disputes over AI dataset ownership, alignment, and correctness.

Expected Outcome: A robust, AI-data-driven blockchain with verifiable training datasets, model alignment guarantees, and tamper-proof AI execution workflows.

## 2026: Scaling AI Data Integrity & Web3 Interoperability

With a strong blockchain foundation in place, LazAI will shift its focus toward ensuring AI data security, cross-chain AI dataset interoperability, and decentralized AI workflow automation.

### Phase 4: Mainnet Upgrade – The Fastest AI Data-Optimized Blockchain (Q1–Q2 2026)

LazAI will implement its first major mainnet upgrade, reinforcing its position as the fastest blockchain for AI data transactions and provenance tracking.

* [ ] **Real-Time AI Data Anchoring:** AI models and datasets will automatically anchor data hashes to ensure transparency and immutability.
* [ ] **Privacy-Preserving AI Data Processing:** Zero-Knowledge Proofs (ZKPs) will be used to verify AI data integrity without exposing sensitive datasets.
* [ ] **Federated AI Data Verification:** A multi-party AI dataset verification system will enable collaborative, decentralized training across multiple chains.

**Expected Outcome:** LazAI will emerge as the leading blockchain for AI data integrity, verification, and decentralized training workflows.

### Phase 5: Strengthening Arbitration & AI Data Security (Q2–Q3 2026)

To reinforce trust in AI data sources and computation, LazAI will implement enhanced dispute resolution and decentralized AI security guarantees.

* [ ] **Optimistic Proofs for AI Data Disputes:** Enables low-cost, fast arbitration over model ownership and dataset integrity.
* [ ] **ZK Proofs for AI Model Transparency:** Cryptographic proofs will ensure AI models are ethically trained and aligned with predefined standards.
* [ ] **LAV (Logical Assertion Verification) Integration:** AI training pipelines will be verified against predefined rules using ZK/OP proofs.
* [ ] **Decentralized AI Agent Coordination:** AI agents will be able to autonomously resolve AI data-related disputes through iDAO-governed mechanisms.

**Expected Outcome:** LazAI will set a new standard for AI data validation, ensuring models remain ethical, accountable, and bias-resistant.

### Phase 6: AI Data Interoperability & Web3 Infrastructure Integration (Q3–Q4 2026)

LazAI will expand beyond its native blockchain, integrating with decentralized AI data platforms, oracles, and federated learning networks.

* [ ] **Cross-Chain AI Dataset Provenance:** AI training datasets will be securely registered across multiple blockchains, enabling trust-minimized AI workflows.
* [ ] **Decentralized AI Compute Networks:** LazAI will interact with decentralized compute resources, ensuring scalable AI model training without centralized dependencies.
* [ ] **Web3 AI Data Bridges:** Establishing multi-chain AI data validation protocols, allowing models to be trained on one chain and verified on another.
* [ ] **Interoperable AI Data Staking & Lending:** Enabling decentralized AI dataset monetization through liquidity pools for AI training data.

**Expected Outcome:** LazAI will become the global hub for AI data exchange, enabling seamless AI asset movement across blockchain ecosystems.

## Long-Term Vision: The AI Data Infrastructure for Web3

LazAI is building the first AI-native blockchain ecosystem centered around AI data integrity, accessibility, and provenance tracking. Our roadmap ensures fair, scalable, and verifiable AI model governance.

🔹 **2025:** Building AI Data Infrastructure – Launching AI agents, enabling AI dataset tokenization, and establishing decentralized AI arbitration mechanisms.

🔹 **2026:** Scaling AI Data Governance – Enhancing AI dataset privacy, security, and interoperability.

🔹 **Beyond 2026:** Expanding AI data monetization, federated learning, and autonomous AI data governance.
