UNIVERSITAS AI

Research at Universitas AI

Advancing the science of reliable intelligence.

From memory-first inference to autonomous agent coordination.

The Paradigm Shift

AI research has undergone three architectural inversions between 2023-2025. We're building the infrastructure for this new reality.

01

Memory Bottleneck Dominance

Long-context inference (≥16K tokens) exhibits arithmetic intensity collapse, making memory bandwidth—not compute throughput—the performance ceiling.

AMD MI300X (192GB, 5.3 TB/s) achieves 112% of NVIDIA H100 performance on 32K-context workloads despite 34% lower FLOPs.

02

Post-Training Supremacy

Reinforcement learning in verifiable domains now delivers 37-40× higher capability returns per dollar than pre-training scaling.

DeepSeek R1 achieves 79.8% on AIME mathematics (vs GPT-4's 13.4%) through $8-15M RL on a $5.6M pre-trained base.

03

Inference Economics Dominance

For production models, 90%+ of total cost is serving, not training. Reasoning models generating 10K+ token outputs cost 185× more per query.

Meta allocates 97.3% of 600K GPUs to inference rather than training, fundamentally reshaping infrastructure requirements.

Research Vectors

Vector 01

Frontier AI Research

From Compression to Abundance

Memory-first paradigm, post-training scaling, and reasoning emergence. Hardware inversions that change optimal AI infrastructure.

Key Papers:

  • Memory-First Paradigm in Frontier AI Research
  • ROCmOpt: Memory-First Inference for AMD GPUs

Measured Impact:

AMD MI300X: 12% faster at 32K context • 108-112% H100 performance • 34% TCO reduction

Technical Papers →
Vector 02

Open Model Infrastructure

Distributed Supercomputing

Multi-cloud orchestration, hardware-aware deployment, and cross-cloud state coordination protocols.

Key Protocols:

  • Constellation (hardware-aware deployment)
  • DEO (inference-aware autoscaling)
  • KVBridge (cross-cloud state coordination)

Measured Impact:

80% faster cold starts • 40-60% cost savings • 88% faster scaling response

View Protocols →
Vector 03

Agent Coordination

Standards for Autonomous AI

Metadata standards enabling trust, discovery, and autonomous operation across organizational boundaries.

Key Standards:

  • AgentFacts (universal KYA standard)
  • Private Facts (operational boundaries)
  • mcpFacts (tool discovery at scale)

Open Standard:

Apache 2.0 licensed • 10-category metadata schema • Multi-authority verification

Explore Standards →

Measured Impact

75%

TTFT Reduction

Through KV cache coordination

40-60%

Cost Savings

Via hardware-aware deployment

88%

Faster Scaling

8min → 60s automated response

79.8%

AIME Score

Via pure RL post-training

80%

Faster Cold Starts

180s → 12s regional caching

85%

Routing Accuracy

Cross-cloud coordination

Publications & Code

Academic Papers

Open Model Infrastructure for Multi-Cloud AI

Constellation, DEO, KVBridge, and ROCmOpt protocols enabling distributed supercomputing.

November 2025 Download PDF →

From Compression to Abundance: Memory-First Paradigm

Hardware inversions and post-training supremacy in frontier AI research.

November 2025 Download PDF →

Coordination Infrastructure for Distributed AI Systems

From single models to ecosystem scale: agent coordination and multi-cloud execution.

November 2025 Download PDF →

AgentFacts: Universal KYA Standard

Verified AI agent metadata and deployment standards for enterprise coordination.

May 2025 Download PDF →

Open Source

open-model-infra/constellation

Hardware-aware model deployment protocol

★ 234
Apache 2.0 Python

jaredgrogan/agentfacts_standard

Universal agent metadata standard

★ 189
Apache 2.0 Spec

open-model-infra/deo

Inference-aware autoscaling for multi-cloud GPU clusters

★ 156
Apache 2.0 Python

open-model-infra/kvbridge

Cross-cloud KV cache coordination

★ 142
Apache 2.0 Python

Research Team

Jared James Grogan

Affiliate, MBZUAI Institute of Foundation Models

"We're solving the coordination problems that frontier research actually faces today. Not theoretical futures—the fragmented multi-cloud reality researchers navigate right now."

Institutional Partners

MIT Media Lab Lambda Labs Hugging Face Coinbase

Get Involved

🚀

Implement

Deploy our protocols in your infrastructure. Open source, Apache 2.0 licensed.

Get Started
💡

Contribute

Submit PRs, add papers, join working groups. Build the future of AI infrastructure.

GitHub →
🤝

Partner

Research collaborations, institutional partnerships, and funding opportunities.

Contact →