Developing truly open and decentralized AI is one of society’s most critical challenges. This requires a complete rethinking of the networks, platforms, and client applications that make it easy to perform decentralized AI model training. Here is one such case study that breaks away from the shackles of centralized data and model control, enabling greater access that puts AI development directly in the hands of people around the world.
Use Case: Decentralized AI Model Training

Problem Statement
Centralized AI systems are plagued by the same problems as the centralized web, wherein large corporations leverage and control the data to fulfill their own objectives against the public interest. Consequently, there is low public participation, less data and inaccuracies in the outcome of AI models.

Realization Approach
Decentralized AI, or dAI, aims to counter this problem by embracing a DePIN-based decentralized compute infrastructure that empowers businesses and developers to harness the power of federated learning by providing a secure and scalable infrastructure.

Solution Space
This approach helps address data privacy concerns by enabling the training of models on distributed data without the need for centralized storage while still leveraging collective insights from multiple sources. It also allows AI companies to unlock the potential of collaborative learning while ensuring data security and compliance with privacy regulations.
Featured DePIN Solution

Akash Network offers a decentralized compute marketplace with an open network that lets users buy and sell computing resources securely and efficiently. It also supports high-performance GPUs for AI applications and is secured by the AKT token.
This case study outlines the integration between Akash and FLock.io that enables users to easily train AI models on decentralized compute. It illustrates how Akash, the first open-source Supercloud, gives AI developers access to the high-performance compute resources needed to train AI models with FLock.io, a platform for decentralized machine learning.
An Introduction to FLock.io
FLock’s mission is to democratize AI through decentralized, blockchain-based systems. FLock facilitates an open and collaborative environment where participants can contribute models, data, and computing resources, in exchange for on-chain rewards based on their traceable contributions.
FLock is a platform focused on federated machine learning. Federated learning is a distributed approach to training AI models where the data remains decentralized, allowing for privacy-preserving and collaborative learning. FLock aims to empower businesses and developers to harness the power of federated learning by providing a secure and scalable infrastructure. By enabling the training of models on distributed data without the need for centralized storage, FLock helps address data privacy concerns while still leveraging collective insights from multiple sources. The platform offers a range of tools and APIs to facilitate the development, deployment, and monitoring of federated learning models, making it easier for organizations to adopt this innovative technology. With FLock, companies can unlock the potential of collaborative learning while ensuring data security and compliance with privacy regulations.
The Decentralized AI Training Pipeline
An AI model is a mathematical representation of a real-world process, designed to make predictions or decisions based on input data. It is the output of training a machine learning algorithm on a dataset, resulting in a set of parameters and rules that can be used to map new input data to the desired output.
AI model training is built around three foundational components: data, algorithms, and compute resources. Data provides the examples and patterns that the AI model learns from to make accurate predictions or decisions. Algorithms define the mathematical models and learning processes that enable the AI system to extract insights and knowledge from the data. Compute resources provide the computational power and infrastructure required to process large amounts of data, execute complex algorithms, and train sophisticated AI models efficiently.
Once a model is initially trained, there are two approaches for updating or improving it: bring the data to the model, or bring the model to the data. The former is often referred to as traditional machine learning, the latter as federated machine learning (or simply ‘federated learning’). In traditional machine learning, data is collected and centralized in a single location, where models are trained using the available compute resources, typically owned or controlled by a single entity. In contrast, federated learning enables models to be trained on distributed data across multiple devices or nodes, without the need to centralize the data. Federated learning allows for collaborative model training while preserving data privacy, whereas traditional machine learning requires data to be shared and processed in a central location.
Although these two approaches are not mutually exclusive, the heuristic is helpful when discussing AI training pipelines and how they differ in centralized and decentralized contexts.
FLock’s core user base (Task Creators, Training Nodes, and Validators) – reflect, at a high level, the AI training pipeline. Some training tasks are defined based on a base model and dataset (task creation). The model is improved based on new data and/or parameter tuning (training), and then its performance is measured by some score (e.g., loss), and this score is validated on a subset of data that the model hasn’t seen before (validation) in order to update the model. The implementation details vary widely depending on the approach (traditional machine learning vs federated learning) and whether pipeline tasks take place in a centralized or decentralized context.
Given the overlap, separating what we mean by centralized and decentralized AI model training is important. The key questions to ask are:
- Who controls task definition?
- Who owns the data used for task definition? What are the processes by which the data is collected? To what extent is the data public or private?
- Who owns the compute resources used for training and validation? Who controls access to these compute resources?
- How transparent are the inner workings of the model (e.g. parameters and algorithms)?
- Which aspects of the training pipeline are open-source and accessible to the public?
- To what extent can users update the model for their own domain-specific use cases?
- Who governs the model, and who governs the AI training process?
Indeed, the reason big tech firms like OpenAI, Meta, Google, and NVIDIA dominate the AI space is because they’ve vertically integrated the AI training stack. Such a centralized AI training process has clear advantages, including more granular control over the training pipeline. However, this approach is fraught with risks and ethical concerns. Decentralized AI training, especially based on blockchain tech, offers a better path forward but is not without its own set of challenges.
The UX for decentralized AI model training has room for improvement. The experience is disjointed and requires significant technical expertise that spans machine learning, blockchain interoperability knowledge, and network engineering. Even top-skilled machine learning engineers with Web2 backgrounds struggle to take advantage of capabilities unlocked with a decentralized AI training approach due to high barriers to entry.
These are the precise barriers to entry that Akash and FLock.io aim to break down.
Birds of a Feather: Akash & FLock.io
As the adage goes, “birds of a feather flock together.” As the decentralized AI movement takes flight, Akash and FLock.io are flying in unison. Akash is an open network that lets users buy and sell computing resources securely and efficiently. FLock.io is an open network that lets users train, validate, and govern AI models democratically and transparently.
Akash is purpose-built for public utility. Compute is available to anyone on the network through Akash’s peer-to-peer marketplace, and the network does not limit the types of tasks or workloads that can be deployed. Similarly, FLock.io is built for and by the community, and the platform does not restrict where models are trained, validated, hosted, and/or deployed.
With a clear alignment on visions for the future and shared values of openness, permissionlessness, and composability, it’s clear to see why Akash and FLock.io are closely aligned.
As the first open-source Supercloud, Akash enables users to permissionlessly access cloud resources — including high-performance NVIDIA H100s, A100s, A6000s, 4090s, and many more. This is a massive unlock as it gives AI developers access to state-of-the-art GPUs to perform training and validation at affordable prices. Not only that, the Akash Console provides a user-friendly interface that abstracts away the complexity of compute resource management – a key friction point in the AI training pipeline – that allows AI developers to spend more time and energy focusing on what they are best at.
This is why FLock.io has integrated with Akash to create 1-click templates that make it easy to run FLock.io nodes on the Akash Supercloud.
For a complete step by step guide to deploy FLock training node on decentralized compute infrastructure, refer to the original post on Akash Network.


