Skip to content

Stabilarity Hub

Menu
  • Home
  • Research
    • Healthcare & Life Sciences
      • Medical ML Diagnosis
    • Enterprise & Economics
      • AI Economics
      • Cost-Effective AI
      • Spec-Driven AI
    • Geopolitics & Strategy
      • Anticipatory Intelligence
      • Future of AI
      • Geopolitical Risk Intelligence
    • AI & Future Signals
      • Capability–Adoption Gap
      • AI Observability
      • AI Intelligence Architecture
      • AI Memory
      • Trusted Open Source
    • Data Science & Methods
      • HPF-P Framework
      • Intellectual Data Analysis
      • Reference Evaluation
    • Publications
      • External Publications
    • Robotics & Engineering
      • Open Humanoid
      • Open Starship
    • Benchmarks & Measurement
      • Universal Intelligence Benchmark
      • Shadow Economy Dynamics
      • Article Quality Science
  • Tools
    • Healthcare & Life Sciences
      • ScanLab
      • AI Data Readiness Assessment
    • Enterprise Strategy
      • AI Use Case Classifier
      • ROI Calculator
      • Risk Calculator
      • Reference Trust Analyzer
    • Portfolio & Analytics
      • HPF Portfolio Optimizer
      • Adoption Gap Monitor
      • Data Mining Method Selector
    • Geopolitics & Prediction
      • War Prediction Model
      • Ukraine Crisis Prediction
      • Gap Analyzer
      • Geopolitical Stability Dashboard
    • Technical & Observability
      • OTel AI Inspector
    • Robotics & Engineering
      • Humanoid Simulation
    • Benchmarks
      • UIB Benchmark Tool
    • Article Evaluator
    • Open Starship Simulation
  • API Gateway
  • About
    • Contributors
  • Contact
  • Join Community
  • Terms of Service
  • Login
  • Register
Menu

UIB Benchmark API Documentation

UIB Benchmark API

Universal Intelligence Benchmark — 8 dimensions, any model, via API

Base URL https://hub.stabilarity.com/api/v1/uib/
Authentication: X-API-Key header. Get your free key →

Endpoints

GET/v1/uib/status
Service status — total benchmark runs and number of distinct models evaluated.
POST/v1/uib/run
Run the full UIB benchmark against any model. You provide your own model API key. Returns composite score + per-dimension breakdown.
FieldTypeDescription
modelstringModel identifier (e.g. “gpt-4”, “claude-sonnet-4-5”, “llama-3-70b”)
api_keystringYour model provider API key (OpenAI, Anthropic, etc.)
dimensionsarrayOptional — subset of dimensions to run. Default: all 8.
api_basestringOptional — custom API base URL for local/self-hosted models (e.g. http://localhost:11434/v1)
GET/v1/uib/leaderboard?limit=20&benchmark_type=uib
Current leaderboard ranked by average composite score. Filter by benchmark_type.
GET/v1/uib/results?limit=50&offset=0
Paginated list of all benchmark run results. Optional: benchmark_type filter.
GET/v1/uib/reports
Paginated list of all individual benchmark reports. Same params as /results.
GET/v1/uib/reports/{run_id}
Full detail for a single benchmark run — all dimension scores, task-level results, and metadata.
GET/v1/uib/dimensions
List all 8 UIB dimensions with task counts and descriptions.

The 8 Dimensions

Causal Reasoning

Pearl’s causal hierarchy, intervention vs observation, confound detection. 5 tasks.

Embodied Intelligence

Physical reasoning, robot control, manipulation, navigation. 5 tasks.

Multimodal Synthesis

Cross-modal reasoning, sensor fusion, modality transfer. 5 tasks.

Temporal & Planning

Long-horizon planning, scheduling, trend analysis, temporal reasoning. 5 tasks.

Social Cognition

Theory of mind, negotiation, sarcasm, team dynamics. 5 tasks.

Tool Creation

Algorithm design, DSL creation, self-improvement, optimization. 5 tasks.

Domain Transfer

Cross-domain analogy, concept mapping, abstraction. 5 tasks.

Resource Efficiency

Compression theory, cost-normalized intelligence, speed prior. 3 tasks.

Examples

Run a full benchmark

# Run UIB on GPT-4 curl -X POST -H “X-API-Key: YOUR_STABILARITY_KEY” \ -H “Content-Type: application/json” \ -d ‘{“model”:”gpt-4″,”api_key”:”YOUR_OPENAI_KEY”}’ \ https://hub.stabilarity.com/api/v1/uib/run

Run specific dimensions only

# Test only causal and social dimensions curl -X POST -H “X-API-Key: YOUR_STABILARITY_KEY” \ -H “Content-Type: application/json” \ -d ‘{“model”:”claude-sonnet-4-5″,”api_key”:”YOUR_KEY”,”dimensions”:[“causal”,”social”]}’ \ https://hub.stabilarity.com/api/v1/uib/run

Benchmark a local model (Ollama, vLLM, etc.)

# Point to your local OpenAI-compatible endpoint curl -X POST -H “X-API-Key: YOUR_STABILARITY_KEY” \ -H “Content-Type: application/json” \ -d ‘{“model”:”llama-3-70b”,”api_key”:”none”,”api_base”:”http://localhost:11434/v1″}’ \ https://hub.stabilarity.com/api/v1/uib/run

Get leaderboard

curl -H “X-API-Key: YOUR_KEY” \ “https://hub.stabilarity.com/api/v1/uib/leaderboard?limit=10”

Example response (leaderboard)

[ { “model”: “deepseek-v4”, “avg_score”: 72.8, “runs”: 3, “dimensions”: { “causal”: 78.2, “embodied”: 61.4, “multimodal”: 69.1, “temporal”: 74.5, “social”: 71.3, “tool_creation”: 82.6, “transfer”: 68.9, “efficiency”: 76.4 } } ]

Links

UIB Benchmark Tool (interactive) · Full API Gateway docs · Source code on GitHub · Research article: UIB Composite Score

Open Source & Contributions Welcome The UIB benchmark is open source. Found a bug or want to contribute? File an issue or PR on GitHub, or email contact@stabilarity.com.

Recent Posts

  • Interpretable Models vs Post-Hoc Explanations: True Cost Comparison for Enterprise AI
  • XAI Tool Economics: The Cost Structure of Explanation Generation
  • Transparent AI Sourcing: Build vs Buy Economics When Explanations Matter
  • XAI Observability: Monitoring Explainability Drift in Production Models
  • Manufacturing AI Observability: Monitoring Explanation Quality in Predictive Maintenance Systems

Research Index

Browse all articles — filter by score, badges, views, series →

Categories

  • ai
  • AI Economics
  • AI Memory
  • AI Observability & Monitoring
  • AI Portfolio Optimisation
  • Ancient IT History
  • Anticipatory Intelligence
  • Article Quality Science
  • Capability-Adoption Gap
  • Cost-Effective Enterprise AI
  • Future of AI
  • Geopolitical Risk Intelligence
  • hackathon
  • healthcare
  • HPF-P Framework
  • innovation
  • Intellectual Data Analysis
  • medai
  • Medical ML Diagnosis
  • Open Humanoid
  • Research
  • ScanLab
  • Shadow Economy Dynamics
  • Spec-Driven AI Development
  • Technology
  • Trusted Open Source
  • Uncategorized
  • Universal Intelligence Benchmark
  • War Prediction

About

Stabilarity Research Hub is dedicated to advancing the frontiers of AI, from Medical ML to Anticipatory Intelligence. Our mission is to build robust and efficient AI systems for a safer future.

Language

  • Medical ML Diagnosis
  • AI Economics
  • Cost-Effective AI
  • Anticipatory Intelligence
  • Data Mining
  • 🔑 API for Researchers

Connect

Facebook Group: Join

Telegram: @Y0man

Email: contact@stabilarity.com

© 2026 Stabilarity Research Hub

© 2026 Stabilarity Hub | Powered by Superbs Personal Blog theme
Stabilarity Research Hub

Open research platform for AI, machine learning, and enterprise technology. All articles are preprints with DOI registration via Zenodo.

185+
Articles
8
Series
DOI
Archived

Research Series

  • Medical ML Diagnosis
  • Anticipatory Intelligence
  • Intellectual Data Analysis
  • AI Economics
  • Cost-Effective AI
  • Spec-Driven AI

Community

  • Join Community
  • MedAI Hack
  • Zenodo Archive
  • Contact Us

Legal

  • Terms of Service
  • About Us
  • Contact
Operated by
Stabilarity OÜ
Registry: 17150040
Estonian Business Register →
© 2026 Stabilarity OÜ. Content licensed under CC BY 4.0
Terms About Contact
Language: 🇬🇧 EN 🇺🇦 UK 🇩🇪 DE 🇵🇱 PL 🇫🇷 FR
Display Settings
Theme
Light
Dark
Auto
Width
Default
Column
Wide
Text 100%

We use cookies to enhance your experience and analyze site traffic. By clicking "Accept All", you consent to our use of cookies. Read our Terms of Service for more information.