Skip to main content

Getting Started

Genesis Cloud Inference service allows users to interact with a variety of LLMs through a unified API interface.

The service is powered by our own GPU infrastructure ensuring high availability and performance.

OpenAI Compatible Endpoints

https://inference-api.genesiscloud.com/openai/v1

Available Models and Pricing

During the beta phase of the Inference Service, the available models and their pricing may constantly change. Use the /models endpoint to get the currently available models and pricing (no authentication required):

https://inference-api.genesiscloud.com/openai/v1/models

Authentication

Requests to the completions endpoints must be authenticated. This is done using the Authorization HTTP header that must be provided with each request.

To generate an API token visit the Keys and Tokens section of the console dashboard. Make sure to select Inference Service when creating a new token.

Please Note: It is absolutely essential that you keep your secret token private as it provides access to everything in your Genesis Cloud account. If you accidentally publish your token, please immediately revoke it.

OpenAI Compatability

Our inference service is OpenAI compatible. This means you cannot only access the service using REST, but also OpenAI libraries (JavaScript and Python). We currently support the /models, /chat/completions and /completions endpoints with more to come.

External Tools (UIs, code editors, etc.)

To configure external tools like open-webui, code editors, etc. simply supply our URL and your token. No further configuration is required.

Open WebUI config example

JavaScript Example

import openai from "openai";

const openai = new openai({
apikey: "INFERENCE_API_TOKEN",
baseurl: "https://inference-api.genesiscloud.com/openai/v1",
});

const response = await openai.chat.completions.create({
model: "deepseek-ai/DeepSeek-V3-0324",
messages: [
{
role: "user",
content: "Explain to me how AI works",
},
],
});

console.log(response.choices[0].message);

Python Example

from openai import OpenAI

client = OpenAI(
api_key="INFERENCE_API_TOKEN",
base_url="https://inference-api.genesiscloud.com/openai/v1"
)

response = client.chat.completions.create(
model="deepseek-ai/DeepSeek-V3-0324",
messages=[
{
"role": "user",
"content": "Explain to me how AI works"
}
]
)

print(response.choices[0].message)

REST Example

curl "https://inference-api.genesiscloud.com/openai/v1/chat/completions" \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $INFERENCE_API_TOKEN" \
-d '{
"model": "deepseek-ai/DeepSeek-V3-0324",
"messages": [
{
"role": "user",
"content": "Explain to me how AI works"
}
],
"stream": true,
"stream_options": {
"include_usage": true
}
}'