Research

Paper

AI LLM February 20, 2026

SeedFlood: A Step Toward Scalable Decentralized Training of LLMs

Authors

Jihun Kim, Namhoon Lee

Abstract

This work presents a new approach to decentralized training-SeedFlood-designed to scale for large models across complex network topologies and achieve global consensus with minimal communication overhead. Traditional gossip-based methods suffer from message communication costs that grow with model size, while information decay over network hops renders global consensus inefficient. SeedFlood departs from these practices by exploiting the seed-reconstructible structure of zeroth-order updates and effectively making the messages near-zero in size, allowing them to be flooded to every client in the network. This mechanism makes communication overhead negligible and independent of model size, removing the primary scalability bottleneck in decentralized training. Consequently, SeedFlood enables training in regimes previously considered impractical, such as billion-parameter models distributed across hundreds of clients. Our experiments on decentralized LLM fine-tuning demonstrate thatSeedFlood consistently outperforms gossip-based baselines in both generalization performance and communication efficiency, and even achieves results comparable to first-order methods in large scale settings.

Metadata

arXiv ID: 2602.18181
Provider: ARXIV
Primary Category: cs.LG
Published: 2026-02-20
Fetched: 2026-02-23 05:33

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2602.18181v1</id>\n    <title>SeedFlood: A Step Toward Scalable Decentralized Training of LLMs</title>\n    <updated>2026-02-20T12:38:42Z</updated>\n    <link href='https://arxiv.org/abs/2602.18181v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2602.18181v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>This work presents a new approach to decentralized training-SeedFlood-designed to scale for large models across complex network topologies and achieve global consensus with minimal communication overhead. Traditional gossip-based methods suffer from message communication costs that grow with model size, while information decay over network hops renders global consensus inefficient. SeedFlood departs from these practices by exploiting the seed-reconstructible structure of zeroth-order updates and effectively making the messages near-zero in size, allowing them to be flooded to every client in the network. This mechanism makes communication overhead negligible and independent of model size, removing the primary scalability bottleneck in decentralized training. Consequently, SeedFlood enables training in regimes previously considered impractical, such as billion-parameter models distributed across hundreds of clients. Our experiments on decentralized LLM fine-tuning demonstrate thatSeedFlood consistently outperforms gossip-based baselines in both generalization performance and communication efficiency, and even achieves results comparable to first-order methods in large scale settings.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.LG'/>\n    <published>2026-02-20T12:38:42Z</published>\n    <arxiv:primary_category term='cs.LG'/>\n    <author>\n      <name>Jihun Kim</name>\n    </author>\n    <author>\n      <name>Namhoon Lee</name>\n    </author>\n  </entry>"
}