A single AI model is a single point of failure. It reflects the training data, cultural assumptions, and capability limitations of one organization, in one country, at one point in time.
HexaCore solves this.
HexaCore is the multi-model AI infrastructure I designed for the RCT Ecosystem. It routes tasks across 7 AI models — selected for complementary strengths, cost profiles, and geopolitical diversity — and verifies results through SignedAI consensus before any output reaches users.
The result: 0.3% hallucination rate (industry: 12–15%), 30–40% cost reduction vs always using the most capable model, and guaranteed geopolitical balance in every production decision.
The Seven Models
| Role | Model | Region | Cost/1M Tokens | Context | Best For | |---|---|---|---|---|---| | Supreme Architect | Claude Opus 4.6 | 🇺🇸 US | $5/$25 | 1M | Critical decisions | | Lead Builder | Kimi K2.5 | 🇨🇳 CN | $0.45/$2.25 | 200K | Complex programming | | Junior Builder | MiniMax M2.1 | 🇨🇳 CN | $0.27/$0.95 | 128K | Routine code | | Specialist | Gemini 3 Flash | 🇺🇸 US | $0.50/$3.00 | 1M | Finance & health | | Librarian | Grok 4.1 Fast | 🇺🇸 US | $0.20/$0.50 | 2M | Long-context reading | | Humanizer | DeepSeek V3.2 | 🇨🇳 CN | $0.25/$0.38 | 128K | Chat & creative | | Regional Thai | Typhoon v2 70B | 🇹🇭 TH | $0.40/$1.20 | 128K | Thai NLP |
Geopolitical Balance: Why It Matters
The Western/Eastern distribution (3W + 3E + 1 Regional) is intentional. LLMs trained predominantly in one region inherit the cultural assumptions, news biases, and regulatory viewpoints of that region.
For enterprise AI serving ASEAN markets:
- Western models (Claude, Gemini, Grok) excel at English-language reasoning trained on Western corpora
- Eastern models (Kimi, MiniMax, DeepSeek) have stronger East Asian cultural context and often stronger multilingual reasoning
- Regional Thai (Typhoon v2) is specifically trained on Thai language and cultural context — critical for Thai NLP tasks
When SignedAI consensus requires agreement from both Western and Eastern models, cultural bias is automatically detected and surfaced — not hidden.
Intelligent Task Routing
HexaCore's Task Router selects the optimal model based on task type and complexity:
| Task Type | Primary Model | Fallback | |---|---|---| | Complex coding | Kimi K2.5 | Claude Opus | | Simple/routine code | MiniMax M2.1 | Kimi K2.5 | | Long document reading | Grok 4.1 Fast | Gemini 3 Flash | | Finance or health queries | Gemini 3 Flash | Claude Opus | | Thai language tasks | Typhoon v2 70B | DeepSeek V3.2 | | Creative or chat | DeepSeek V3.2 | MiniMax M2.1 | | Critical decisions | Claude Opus 4.6 | (requires Tier 8 consensus) |
This routing saves 30–40% in API costs vs always routing to Claude Opus, while maintaining quality through task-appropriate model selection and circuit breaker fallback chains.
SignedAI Consensus Tiers
For production-critical queries, HexaCore routes through SignedAI — a multi-model consensus system with 4 tier levels:
| Tier | Models | Threshold | Cost/1M | Use Case | |---|---|---|---|---| | S | 1 model | N/A | $0.32 | Fast, low-stakes | | 4 | 4 (2W:2E) | 50% (2/4) | $0.45 | Standard operations | | 6 | 6 (3W:3E) | 67% (4/6) | $1.00 | Production consensus | | 8 | 8 (4W:4E) | 75% (6/8) | $2.60 | Critical decisions |
Voting Methods
SignedAI uses four voting algorithms:
- Majority vote — Simple >50% agreement
- Weighted vote — Models weighted by confidence score for the specific task type
- Ranked choice — Models rank each other's outputs; highest accumulated rank wins
- Jaccard similarity — Measures semantic similarity across all outputs (not just agreement/disagreement)
The Jaccard similarity method is particularly powerful for detecting when models superficially agree but produce subtly different reasoning chains — a common source of consensus hallucination in naive implementations.
Cost Optimization: The Math
| Scenario | Model | Cost/1M Tokens | |---|---|---| | Simple chat | DeepSeek V3.2 | $630 | | Routine code | MiniMax M2.1 | $1,220 | | Complex code | Kimi K2.5 | $2,700 | | Finance/health | Gemini 3 Flash | $3,500 | | Critical decision | Claude Opus 4.6 | $30,000 |
A system that always uses Claude Opus for everything spends ~47x more than one that routes correctly with HexaCore. Even accounting for SignedAI consensus overhead, HexaCore delivers 3.74x cost reduction through intelligent routing.
Fault Tolerance: Circuit Breaker Pattern
HexaCore uses the Circuit Breaker pattern to protect against model failures:
- Closed state: Normal operation. Requests flow to the primary model.
- Open state: After consecutive failures, the circuit opens. All requests route to fallback models.
- Half-open state: After 30s recovery window, one test request is sent. If it succeeds, the circuit closes.
The circuit breaker covers all 7 models independently. If Kimi K2.5 is experiencing issues, complex coding tasks automatically route to the fallback chain (Claude Opus or MiniMax) with no manual intervention.
Key metrics:
- Recovery time: 30 seconds automatic
- Fallback chain depth: 3 levels
- Throughput: 18 req/s aggregate (6 models × 3 req/s each)
- Queue capacity: 1,000 concurrent requests
Frequently Asked Questions
Why 7 models instead of 1 or 2?
Seven models provides the optimal balance between geopolitical diversity, cost optimization, and consensus reliability. Fewer models limit diversity; more models increase cost and latency without proportional quality gains.
Can HexaCore work with models not listed here?
Yes. HexaCore uses a registry-based architecture. New models can be added by registering them with their capability profile, cost data, and geopolitical classification. The Task Router automatically incorporates new registrations.
How does Typhoon v2 differ for Thai language tasks?
Typhoon v2 70B is specifically trained on Thai language corpora. For Thai NLP tasks, it achieves significantly better Thai linguistic accuracy than Western or general Eastern models, which are trained primarily on English and Mandarin. For ASEAN enterprise customers, Thai language quality is a non-negotiable requirement.
Does SignedAI slow down every request?
Only for Tier 4+ consensus. Tier S (single model) has no consensus overhead. Most casual queries use Tier S. Production queries use Tier 4 or 6. Only critical decisions (creating legal documents, medical outputs, high-value financial decisions) use Tier 8.
Summary
HexaCore is the infrastructure layer that makes the RCT Ecosystem's performance guarantees possible:
- 7 models with complementary strengths and geopolitical balance (3W + 3E + 1 Regional Thai)
- Intelligent routing — right model for each task type (30-40% cost savings)
- SignedAI consensus — Tier S/4/6/8 with majority, weighted, ranked, and Jaccard voting
- Circuit breaker — automatic fault tolerance with 30s recovery
- 18 req/s aggregate throughput — scalable across 6 concurrent models
The result: 0.3% hallucination rate, 99.98% uptime, and 3.74x cost reduction — all built on a balanced, verifiable, multi-model foundation.
This article was written by Ittirit Saengow, founder and sole developer of RCT Labs.
สิ่งที่องค์กรควรสรุปจากบทความนี้
HexaCore is the multi-model AI routing infrastructure at the heart of the RCT Ecosystem. This article explains how 7 AI models (3 Western + 3 Eastern + 1 Regional Thai) are selected, balanced, and verified to achieve 0.3% hallucination and 30-40% cost savings vs single-model deployments.
เชื่อมจากความรู้ไปสู่การประเมินระบบจริง
ทุกบทความเชิงวิจัยควรเชื่อมต่อไปยัง solution page, authority page, และ conversion path เพื่อให้การอ่านไม่จบแค่ traffic
บทความก่อนหน้า
สมการ FDIA อธิบาย: F = (D^I) × A ขับเคลื่อน Constitutional AI อย่างไร
FDIA คือรากฐานทางคณิตศาสตร์ของ RCT Labs ซึ่งเป็นสมการสี่ตัวแปรที่ควบคุมวิธีที่ระบบ AI ผลิตผลลัพธ์ที่น่าเชื่อถือ บทความนี้อธิบายทุกส่วนประกอบ ทำไม Intent ทำหน้าที่เป็นตัวยก และ FDIA บรรลุความแม่นยำ 0.92 เทียบกับ baseline อุตสาหกรรม ~0.65 ได้อย่างไร
บทความถัดไป
The Intent Operating System: Why Enterprise AI Needs an Orchestration Layer
An LLM is not an operating system. It is an application. Enterprise AI needs what every enterprise software system needs: an orchestration layer that manages resources, enforces policies, routes tasks, and maintains state. This is what an Intent OS provides — and why the RCT Ecosystem is built as one.
Ittirit Saengow
Primary authorอิทธิฤทธิ์ แซ่โง้ว คือผู้ก่อตั้ง นักพัฒนาเพียงคนเดียว และผู้เขียนหลักของ RCT Labs — แพลตฟอร์มระบบปฏิบัติการ AI แบบ constitutional ที่สร้างขึ้นอย่างอิสระตั้งแต่สถาปัตยกรรมจนถึงการเผยแพร่ เขาคิดค้นสมการ FDIA (F = (D^I) × A) ข้อกำหนดโปรโตคอล JITNA (RFC-001) สถาปัตยกรรม 10 ชั้น ระบบ 7-Genome และกระบวนการ RCT-7 แพลตฟอร์มทั้งหมด ทั้งโครงสร้างสองภาษา ระบบ SEO ระดับองค์กร ไมโครเซอร์วิส 62 ตัว อัลกอริทึม 41 ชุด และงานวิจัยทั้งหมดที่เผยแพร่ สร้างโดยคนเพียงคนเดียวในกรุงเทพฯ ประเทศไทย