Category:AI on demand: Difference between revisions

From MediaWiki
Jump to navigation Jump to search
No edit summary
 
(2 intermediate revisions by the same user not shown)
Line 6: Line 6:
All of our LLM services are reachable via the OpenAI-compatible gateway https://llm.stoney-cloud.com/.
All of our LLM services are reachable via the OpenAI-compatible gateway https://llm.stoney-cloud.com/.


= Usage =
Check out the [[AI on demand: Usage]] page. You can list the current models, show your own usage and more.
We assume you've received your API key from us in our usual manner.
 
== Usage - List all models ==
=== Usage - List all models - Generic ===
<syntaxhighlight lang="bash">
# Set your personal key:
STONEY_KEY=sk-...
 
# List all generic models:
curl -s https://llm.stoney-cloud.com/v1/models \
  -H "Authorization: Bearer $STONEY_KEY" \
  | jq .
</syntaxhighlight>
 
Example output:
<syntaxhighlight lang='json'>
{
  "data": [
    {
      "id": "MinerU2.5-2509-1.2B",
      "object": "model"
    },
    {
      "id": "MiniMaxAI/MiniMax-M2.5",
      "object": "model"
    },
    {
      "id": "swiss-ai/Apertus-70B-Instruct-2509",
      "object": "model"
    }
  ],
  "object": "list"
}
</syntaxhighlight>
 
=== Usage - List all models - Audio ===
<syntaxhighlight lang="bash">
# Set your personal key:
STONEY_KEY=sk-...
 
# List all audio models:
curl -s https://llm.stoney-cloud.com/v1/audio/models \
  -H "Authorization: Bearer $STONEY_KEY" \
  | jq .
</syntaxhighlight>
 
Example output:
<syntaxhighlight lang='json'>
{
  "data": [
    {
      "id": "mistralai/Voxtral-Mini-3B-2507",
      "object": "model"
    }
  ],
  "object": "list"
}
</syntaxhighlight>
 
== Usage - Inspecting your usage ==
We issue one key per model for now. The usage is thus per-model-per-key.
 
<syntaxhighlight lang="bash">
# Set your personal key:
STONEY_KEY=sk-...
 
# Set desired month in the form YYYY-MM, for example: '2026-03'.
MONTH=$(date +%Y-%m)
 
curl -s https://llm.stoney-cloud.com/v1/usage?month="$MONTH" \
  -H "Authorization: Bearer $STONEY_KEY" \
  | jq .
</syntaxhighlight>
 
Example output:
<syntaxhighlight lang="json">
{
  "models": [
    {
      "cost_chf": 0.000002,
      "model": "MinerU2.5-2509-1.2B",
      "output": 16,
      "input": 70
    },
    {
      "cost_chf": 0.054537,
      "model": "MiniMaxAI/MiniMax-M2.5",
      "output": 3603,
      "input": 10097
    },
    {
      "cost_chf": 0.0016,
      "model": "mistralai/Voxtral-Mini-3B-2507",
      "seconds": 8
    },
    {
      "cost_chf": 0.003774,
      "model": "swiss-ai/Apertus-70B-Instruct-2509",
      "output": 551,
      "input": 443
    }
  ],
  "month": "2026-03",
  "total_cost_chf": 0.059913
}
</syntaxhighlight>


[[Category: Artificial Intelligence]]
[[Category: Artificial Intelligence]]

Latest revision as of 08:35, 8 April 2026

Overview

This category collects all AI on demand related pages.

stepping stone AG is proud to serve customers with current Language Models and OCR solutions.

All of our LLM services are reachable via the OpenAI-compatible gateway https://llm.stoney-cloud.com/.

Check out the AI on demand: Usage page. You can list the current models, show your own usage and more.

Pages in category "AI on demand"

The following 5 pages are in this category, out of 5 total.