Category:AI on demand: Difference between revisions

From MediaWiki
Jump to navigation Jump to search
No edit summary
 
(5 intermediate revisions by 2 users not shown)
Line 6: Line 6:
All of our LLM services are reachable via the OpenAI-compatible gateway https://llm.stoney-cloud.com/.
All of our LLM services are reachable via the OpenAI-compatible gateway https://llm.stoney-cloud.com/.


= Usage =
Check out the [[AI on demand: Usage]] page. You can list the current models, show your own usage and more.
We assume you've received your API key from us in our usual manner.
 
== Usage - List all models ==
=== Usage - List all models - Generic ===
<syntaxhighlight lang="bash">
# Set your personal key:
STONEY_KEY=sk-...
 
# List all generic models:
curl -s https://llm.stoney-cloud.com/v1/models \
  -H "Authorization: Bearer $STONEY_KEY" \
  | jq .
</syntaxhighlight>
 
Example output:
<syntaxhighlight lang='json'>
{
  "data": [
    {
      "id": "MinerU2.5-2509-1.2B",
      "object": "model"
    },
    {
      "id": "MiniMaxAI/MiniMax-M2.5",
      "object": "model"
    },
    {
      "id": "swiss-ai/Apertus-70B-Instruct-2509",
      "object": "model"
    }
  ],
  "object": "list"
}
</syntaxhighlight>
 
=== Usage - List all models - Audio ===
<syntaxhighlight lang="bash">
# Set your personal key
STONEY_KEY=sk-...
 
# List all audio models
curl -s https://llm.stoney-cloud.com/v1/audio/models \
  -H "Authorization: Bearer $STONEY_KEY" \
  | jq .
</syntaxhighlight>
 
Example output:
<syntaxhighlight lang='json'>
{
  "data": [
    {
      "id": "mistralai/Voxtral-Mini-3B-2507",
      "object": "model"
    }
  ],
  "object": "list"
}
</syntaxhighlight>
 
== Usage - Inspecting your usage ==
We issue one key per model for now. The usage is thus per-model-per-key.
 
<syntaxhighlight lang="bash">
# Set your personal key
STONEY_KEY=sk-...
 
curl -s https://llm.stoney-cloud.com/v1/usage \
  -H "Authorization: Bearer $STONEY_KEY" \
  | jq .
</syntaxhighlight>
 
Example output:
<syntaxhighlight lang="json">
{
  "models": [
    {
      "cost_chf": 0.000002,
      "model": "MinerU2.5-2509-1.2B",
      "output": 16,
      "input": 70
    },
    {
      "cost_chf": 0.054537,
      "model": "MiniMaxAI/MiniMax-M2.5",
      "output": 3603,
      "input": 10097
    },
    {
      "cost_chf": 0.0016,
      "model": "mistralai/Voxtral-Mini-3B-2507",
      "seconds": 8
    },
    {
      "cost_chf": 0.003774,
      "model": "swiss-ai/Apertus-70B-Instruct-2509",
      "output": 551,
      "input": 443
    }
  ],
  "month": "2026-03",
  "total_cost_chf": 0.059913
}
</syntaxhighlight>


[[Category: Artificial Intelligence]]
[[Category: Artificial Intelligence]]

Latest revision as of 08:35, 8 April 2026

Overview

This category collects all AI on demand related pages.

stepping stone AG is proud to serve customers with current Language Models and OCR solutions.

All of our LLM services are reachable via the OpenAI-compatible gateway https://llm.stoney-cloud.com/.

Check out the AI on demand: Usage page. You can list the current models, show your own usage and more.

Pages in category "AI on demand"

The following 5 pages are in this category, out of 5 total.