Rebellions.ai Logo
Development

Rebellions.ai

Energy-efficient AI inference accelerators and software for hyperscale data centers.
8.4
Rating
Custom
Price
6
Key Features

Overview

Rebellions.ai (Rebellions Inc.) develops purpose-built AI inference accelerators (chiplets, SoCs, servers) and a GPU-class software stack to enable high-throughput, energy-efficient LLM and multimodal inference at hyperscale. Their product family includes REBEL chiplets (REBEL-Quad), the ATOM SoC family, and ATOM-Max server/pod systems, paired with the RBLN (Rebellions) SDK, Model Zoo, and developer tools to support production deployments, mixed-precision execution, and distributed rack-to-rack scaling. The company emphasizes UCIe-Advanced chiplet interconnects, HBM3E/GDDR6 memory options, and software compatibility with PyTorch, vLLM, Triton, and Hugging Face for fast adoption in existing pipelines. Rebellions targets energy- and cost-sensitive data-center inference workloads by combining hardware efficiency with an integrated software stack. Public-facing site materials do not list per-unit pricing or public subscription tiers; pricing and procurement are handled via enterprise/direct sales contact.

Details

Developer
rebellions.ai
Launch Year
2020
Free Trial
No
Updated
2025-12-07

Features

Chiplet SoC Architecture

REBEL-Quad and REBEL chiplets use UCIe-Advanced to present multiple chiplets as a single virtual die, enabling high compute density and low-latency chiplet-to-chiplet communication.

High-Bandwidth Memory & Memory Subsystem

Supports HBM3E (144 GB, multi-TB/s effective bandwidth) and GDDR6-based ATOM-Max variants to deliver the memory throughput required for long-context LLM inference.

Mixed-Precision, High-Throughput Execution

Native mixed-precision pipeline (FP16/FP8 and narrower formats) enabling high TFLOPS/TOPS performance while maintaining single-pipeline execution and kernel compatibility.

Rebellions (RBLN) SDK and Model Zoo

GPU-class, PyTorch-native SDK with compiler/runtime, profiling, Triton backend, vLLM/Hugging Face integration, and >300 supported models for rapid onboarding and optimization.

Rack-to-Rack & Scalable Pod Designs

ATOM-Max Server/Pod and RDMA-friendly designs allow horizontal scaling from a single server to large clusters with orchestration and pod-level management.

Developer Tooling & Observability

Profiler, driver/firmware stack, runtime modules, system-management tools, and example integrations (Kubernetes, Docker, Ray) to support deployment and performance tuning.

Screenshots

Rebellions.ai Screenshot
Rebellions.ai Screenshot
Rebellions.ai Screenshot

Pros & Cons

Pros

  • Very high throughput-per-watt focus (energy-efficient inference hardware).
  • Modern chiplet-based designs (UCIe-Advanced) and HBM3E support for high memory bandwidth.
  • Full-stack approach (hardware + SDK + Model Zoo) and strong developer tooling (PyTorch-native, Triton, vLLM integration).
  • Product options for single-server and rack/pod scale (ATOM-Max Server/Pod, REBEL-Quad).
  • Global presence and strong industry partnerships (Arm, Samsung Foundry, SK Telecom, Pegatron, Marvell).

Cons

  • No public pricing or self-serve purchase flows—enterprise sales/contact required.
  • Detailed benchmarks and independent performance/price comparisons are limited on the public site.
  • Focused on data-center deployments; not targeted at hobbyists or consumer use.

Compare with Alternatives

FeatureRebellions.aiEnCharge AIHailo
PricingN/AN/AN/A
Rating8.4/108.1/108.2/10
Compute ArchitectureChiplet SoC architectureCharge-domain analog IMC architectureDataflow accelerator architecture
Memory BandwidthHBM3E high-bandwidth memoryAnalog IMC with limited external bandwidthOn-chip buffers modest memory bandwidth
Precision & ThroughputMixed-precision high-throughput executionHigh-efficiency analog IMC throughputINT8-optimized low-power throughput
Scalability & PodsYesPartialPartial
Developer SDKYesYesYes
Observability ToolingYesPartialPartial
Form Factor SupportYesYesYes
Power EfficiencyHyperscale energy-efficient inferenceHigh energy efficiency sustainability focusedLow-power edge optimized

Audience

DevelopersBuild, optimize, and deploy LLM and multimodal models using the RBLN SDK and Model Zoo.
EnterprisesDeploy energy-efficient inference at scale in data centers with ATOM-Max servers and REBEL-Quad appliances.
Cloud & OEMsIntegrate chiplet-based accelerators into rack-scale and sovereign deployments for hyperscale inference.

Tags

aiinferencenpuchipletHBM3EUCIeLLMsdkdevopsenergy-efficiency

Related Articles (7)