Skip to content
@Mesh-LLM

Mesh LLM

Democratize and Decentralized Models and Agents.

Mesh LLM logo

Mesh LLM

mesh-llm turns spare compute into a peer-to-peer inference cloud for open models.

mesh-llm pools GPUs across macOS and Linux machines so teams, researchers, and agents can run local or open-weight models through one OpenAI-compatible endpoint. It can serve a model on one node, distribute large models across nearby peers, route requests to specialized models, and let agents coordinate through mesh gossip.

Work in progress — use with caution.

What it is for

  • Share spare GPU capacity across trusted machines.
  • Run open models locally without a centralized inference provider.
  • Serve an OpenAI-compatible API at http://localhost:9337/v1.
  • Route requests across multiple nodes, models, and capabilities.
  • Experiment with distributed inference, MoE expert sharding, and agent collaboration.

Quick start

curl -fsSL https://github.com/Mesh-LLM/mesh-llm/releases/latest/download/mesh-bundle.tar.gz | tar xz \
  && mkdir -p ~/.local/bin \
  && mv mesh-bundle/* ~/.local/bin/

Join the public mesh:

mesh-llm --auto

Or start your own mesh with a model:

mesh-llm --model GLM-4.7-Flash

Learn more

Built with Rust, iroh, and llama.cpp.

Pinned Loading

  1. mesh-llm mesh-llm Public

    Distributed AI/LLM for the people. Share compute privately or publicly to power your agents and chat.

    Rust 958 120

Repositories

Showing 7 of 7 repositories

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Top languages

Loading…

Most used topics

Loading…