Skip to main content

6 posts tagged with "Research"

Technical research posts covering MoE architectures, distributed training, and AI alignment.

View All Tags

MoE and Catastrophic Forgetting: How Expert Isolation Gives You Domain Specialization Without Destroying General Capability

· 6 min read
Research & Engineering

Catastrophic forgetting is the central unsolved problem of continual learning. You train a model to be better at task B, and it forgets how to do task A. The more you specialize, the more you destroy. BlockZero's MoE architecture makes this tradeoff avoidable — by construction.

The Compounding Expert Library — Why Every Training Job Makes the Next One Cheaper

· 5 min read
Research & Engineering

Most AI customization is one-off work. A consulting firm spends six months fine-tuning a model for a client, and when the engagement ends, all that accumulated knowledge walks out the door. The next client starts from scratch. There is no compounding.

BlockZero is built around a fundamentally different model: every training job produces a reusable expert module that compounds in value with every subsequent use.

TEFT: Targeted Expert Fine-Tuning — How We Reduce Communication Overhead by Orders of Magnitude

· 7 min read
Research & Engineering

This post introduces TEFT (Targeted Expert Fine-Tuning) — the protocol at the core of BlockZero — and explains how it achieves communication-efficient, quality-gated distributed MoE fine-tuning over a permissionless network. This is the research paper translated into blog form.