Turn your GPUs into
callable AI services.
AionRay helps teams connect owned GPUs and idle compute, then run training, deployment, OpenAI-compatible APIs, app monetization, compute rental, billing records, revenue records, and troubleshooting from one console.
Everything you need to ship an AI product, in one chain.
Training, API, billing, and revenue records should not live in separate tools. Put them together and you ship faster.
- Upload data or point to a local dataset
- Launch fine-tune jobs on attached machines
- Publish as an inference service in one click
- Works with common OpenAI client calls
- Per-request or per-token billing pipeline built in
- End-user credentials, quotas, and bills centralized
- Wrap the endpoint as a published app
- Configure access policy, price, quota
- Subscription and usage revenue are tracked in the ledger
OpenAI-style clients
connect with simple config.
AionRay implements the Chat / Completions / Embeddings / Models routes. Swap the base_url and key, then keep your existing SDK flow for core requests.
The public contract covers the core `/v1/*` routes below. Advanced request fields are forwarded to the underlying model when the selected service supports them.
Canonical public routes: `/v1/chat/completions`, `/v1/completions`, `/v1/embeddings`, `/v1/models`.Tool calls, structured output, and multimodal content depend on the model and runtime you pick.from openai import OpenAI
client = OpenAI(
base_url="https://your-gateway/v1",
api_key="sk-xxxxxxxx",
)
resp = client.chat.completions.create(
model="your-model",
messages=[
{"role": "user", "content": "Hello"},
],
)One console for training, deployment, billing, and distribution.
Keep the core AI product workflow in one place, instead of switching between training, serving, metering, and revenue accounting tools.
Who uses AionRay
Turn an idea into a subscription AI tool — from training to launch, no tool switching.
Ship a domain model as a metered API. Access policy, quota, audit, and billing are ready out of the box.
Package internal models into services for internal or external use, with quota, billing, and audit in one place.
5 steps to get going
- 1Sign up and open the console
- 2Connect your own GPU, or rent one from the marketplace
- 3Upload data, start a fine-tune
- 4Deploy the inference service, or publish as an app
- 5Set pricing and budgets, then start billing
Start from your own GPU or model.
Connect compute, train or import a model, deploy it as an API, package it into a subscribable app, and keep operations visible from one console.