<aside> <img src="notion://custom_emoji/1033d2c0-f40d-4c2f-9583-313aa8d2b337/1a080ee7-e5ee-8009-889f-007a7a44cbfb" alt="notion://custom_emoji/1033d2c0-f40d-4c2f-9583-313aa8d2b337/1a080ee7-e5ee-8009-889f-007a7a44cbfb" width="40px" /> TABLE OF CONTENTS
</aside>
<aside> <img src="notion://custom_emoji/1033d2c0-f40d-4c2f-9583-313aa8d2b337/1a080ee7-e5ee-8009-889f-007a7a44cbfb" alt="notion://custom_emoji/1033d2c0-f40d-4c2f-9583-313aa8d2b337/1a080ee7-e5ee-8009-889f-007a7a44cbfb" width="40px" /> MIZU is the first Edge AI Data Network, turning personal devices into a self-hosted AI agent ecosystem. Powered by local models, MIZU’s AI agents autonomously manage, process, and share personalized data across apps like Telegram and Twitter—securely and seamlessly
</aside>
Edge AI is 5–10x larger than traditional data centers, with a compute capacity of 20–40 ZFLOPS compared to just 4 ZFLOPS in centralized infrastructure. With the rise of self-hosted AI models via platforms like Ollama and LMStudio, individuals are increasingly running AI locally.
However, turning edge devices into a functional AI inference network remains a challenge due to unreliable nodes, execution verification issues, and fragmented compute power. Addressing these barriers is critical to unlocking the full potential of decentralized AI.
Unlike cloud servers, which operate under strict uptime guarantees, edge devices are inherently unreliable. Home computers and personal GPUs can be turned off, go to sleep, or disconnect at any moment, making it difficult to ensure a consistent inference service.
Various approaches have been proposed to mitigate this, most commonly through request redundancy, where multiple devices handle the same query, and only the fastest valid response is accepted. While this method can improve reliability, it introduces significant inefficiencies by duplicating compute work.
Another approach focuses on smart request routing, where tasks are dynamically assigned to devices with a history of stable uptime. Additionally, economic incentives are being explored, rewarding nodes that maintain high availability while penalizing those that frequently drop offline.
One of the biggest risks in a decentralized AI network is the ability for bad actors to cheat the system. If a worker can claim to have performed inference but return incorrect or empty results, the network loses integrity.
Current approaches to execution verification fall into three categories:
Unlike cloud environments with standardized hardware, Edge AI networks must deal with a wide range of devices—from high-performance desktops to low-power mobile phones. Managing this diversity is a logistical challenge, especially when some devices can barely run small models, while others can handle large-scale inference.
One approach is to move edge devices into controlled environments, such as micro data centers, and limit the number of supported hardware SKUs (e.g., Mac Mini or Intel NUCs). While this simplifies hardware management, it fundamentally contradicts the purpose of Edge AI, centralizing compute power rather than distributing it.
Alternatively, some research explores federated inference, where workloads are distributed across multiple weaker devices to enable larger model execution on edge hardware. While this approach makes it possible to run bigger models, individual device latency remains high, and throughput is significantly limited, reducing overall efficiency.
Lessons from MIZU: Phones vs. Laptops
In Phase 1, MIZU focused on integrating smartphones into the network. While some phones can run 1–3B parameter models, we shifted our focus to laptops and desktops in Phase 2 due to two critical challenges:
Although we continue to explore ways to integrate smartphones, our immediate focus is on devices that ensure network stability and reliability. To achieve this, MIZU prioritizes devices that meet two key criteria:
Sufficient performance to run AI models effectively.
Sustained uptime, ensuring that inference nodes remain available for long periods
MIZU is building an edge AI inference network optimized for asynchronous workloads, ensuring scalability, efficiency, and cost-effectiveness.
Unlike cloud-based models that prioritize low-latency responses, MIZU is designed for large-scale batch inference and data processing, where cost and scalability matter more than real-time latency.
MIZU functions as both a routing and caching layer, reducing redundant processing while ensuring reliability across distributed edge devices.
MIZU employs a hybrid verification approach that combines optimistic challenges and per-worker sampling, ensuring worker execution is trustworthy while maintaining efficiency. Each MIZU Pool owner can enable per-worker sampling to enhance result reliability. Over time, workers with consistent performance will require fewer checks, reducing network overhead.
Reputation-Based Sampling Strategy
The default sampling strategy defines a reputation score (r) for each worker:
The sampling rate (s) is calculated as s = Max(1, r) / 10,000
With this approach, a well-behaving worker will see its sample rate drop to 0.01% after 200 correct inferences, ensuring minimal verification overhead while maintaining execution integrity.
MIZU is continuously refining verification strategies, with plans to introduce more dynamic reputation models in future updates.
MIZU allows anyone to own and operate a MIZU Pool, where pool owners:
This market-driven model encourages efficiency and network stability while distributing decision-making across multiple stakeholders.
Most decentralized AI networks struggle due to a lack of real-world demand. MIZU addresses this by launching its first major application: The Open Content Aggregator—a tool that collects, processes, and redistributes content from Telegram, Twitter, and other decentralized data sources.
This aggregator runs on periodic AI pipelines, making it a perfect fit for MIZU’s async inference model, where:
By focusing on real demand, MIZU ensures steady AI workload distribution, fueling network adoption.
Instead of relying solely on edge devices from day one, MIZU follows a hybrid model to ensure reliability while transitioning to full decentralization:
This gradual transition ensures businesses can lower costs while maintaining high performance.
While software solutions drive current progress, specialized AI hardware could accelerate edge inference adoption. MIZU is exploring custom AI workstations, designed to function as decentralized inference nodes, offering
These explorations could redefine how AI inference is performed at the edge, further reducing reliance on centralized infrastructure.