Read the Architecture
Open the interactive architecture map to understand query flow from orchestrator to agents.
IEEE-Style Research Project
Orchestrated local inference across multiple agent models with modular aggregation, reproducible benchmarks, statistical significance testing, and publication-ready outputs.
Use this public portal in three simple steps.
Open the interactive architecture map to understand query flow from orchestrator to agents.
Go to Playground, set your public orchestrator URL, and run one strategy with fixed seed.
Review benchmark charts, CSV/JSON exports, and the full IEEE manuscript from this same site.
This project presents a complete local distributed AI network in which one FastAPI orchestrator routes prompts to four local Ollama-based LLM agents, aggregates answers with multiple ensemble strategies, and logs benchmark-ready metrics. The system implements majority voting, dynamic weighted voting, inverse surprising popularity (ISP), topic-based routing, and two-round debate. Evaluation is automated over MMLU, GSM8K, and TruthfulQA with deterministic controls, repeated trials, confidence intervals, paired t-tests, and Wilcoxon signed-rank tests. All outputs are exported as CSV, JSON, PNG figures, and IEEE LaTeX tables, with an accompanying manuscript package for conference submission.
Watch the full visual explanation of this distributed AI ensemble project, including architecture flow, live multi-agent querying, optimization, and benchmark outcomes.
Click any node in the architecture map to view role details.
This playground calls your orchestrator /query API directly.
For public use, set a public HTTPS endpoint (for example, your tunnel URL) with CORS enabled on the orchestrator.
https://ai.marcbd.site as the Orchestrator API URL./health.majority, Seed 42, Temperature 0.0, Deterministic ON, Max tokens 24-64.Support Endpoint
Recommended public endpoint for this project:
https://ai.marcbd.site
Waiting for query...
No response yet.
Agreement distribution will appear after query.
| Agent | Model | Answer | Latency (ms) | Tokens | Status |
|---|---|---|---|---|---|
| Run a query to populate per-model results. | |||||
Only the public abstract is shown by default. Full manuscript reading and downloads are restricted behind your follow, subscribe, permission-request, and password gate.
This project presents a complete local distributed AI network in which one FastAPI orchestrator routes prompts to four local Ollama-based LLM agents, aggregates answers with multiple ensemble strategies, and logs benchmark-ready metrics. The system implements majority voting, dynamic weighted voting, inverse surprising popularity, topic-based routing, and two-round debate. Evaluation is automated over MMLU, GSM8K, and TruthfulQA with deterministic controls, repeated trials, confidence intervals, paired t-tests, and Wilcoxon signed-rank tests. All outputs are exported as CSV, JSON, PNG figures, and IEEE LaTeX tables, with an accompanying manuscript package for conference submission.
Visitors must follow my GitHub, subscribe to my YouTube channel, send a permission request, and then enter the approved access password to download the protected paper package.
Complete all three steps, then enter the approved password to unlock the protected paper package.
Access is stored only for this browser session.
This public preview shows only the first page of the manuscript. Full access remains protected.
The full manuscript is no longer published as plain PDF, Word, LaTeX, or text files in this public branch.
It is distributed only as an encrypted .tar.gpg package. After download, use the approved
password to decrypt and extract the files locally.