phaya

Use the Phaya SaaS backend to generate images, videos, audio, music, and run LLM chat completions via simple REST API calls. Use when the user wants to generate media, call AI models, or use the Phaya API for image/video/audio/text generation.

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "phaya" with this command: npx skills add boombignose/phaya

Phaya Media API

Phaya is a FastAPI backend that brokers AI media generation across KIE.ai (Sora 2, Veo 3.1, Seedance, Kling, Seedream, Suno), Google Gemini TTS, and OpenRouter LLMs.

Auth

All endpoints require a Bearer token or API key:

Authorization: Bearer <your_api_key>

Set these environment variables before using this skill:

export PHAYA_API_KEY="your_api_key_here"   # required — all endpoints
export PHAYA_BASE="https://your-api-host/api/v1"  # required — your Phaya instance URL

Get your profile and credit balance:

  • GET /api/v1/user/profile — full profile
  • GET /api/v1/user/credits{ "credits_balance": 84.90, ... }

Rate limit: 60 requests/minute per API key.

Cost Warning

This skill calls a paid credit system. Every generation deducts real credits from your account. Video generation can cost 8–50 credits per job. Start with cheap endpoints (text-to-image at 1 credit) to verify connectivity before running expensive jobs. Credits are auto-refunded on failure, but not on successful jobs you don't use.

Recommendation: create a scoped API key with a small credit balance for initial testing.

Credit System

Every generation costs credits deducted on job creation; auto-refunded on failure.

CreditsService
0.5image-to-video (FFmpeg local), Sora 2 character creation
1.0text-to-image (Z-Image)
1.5Seedream 5.0
2–4Nano Banana 2 (1K/2K/4K resolution)
3.0Text-to-music (Suno)
2–35Seedance 1.5 Pro (resolution × duration × audio)
8.0Sora 2 video
1.21–1.82/secKling 2.6 motion control (720p/1080p)
15.0Veo 3.1 fast (veo3_fast)
50.0Veo 3.1 quality (veo3)

Job / Polling Pattern

Every generation is async. Create endpoints return job_id immediately; poll the status endpoint.

POST /api/v1/<service>/create   →  { "job_id": "uuid" }
GET  /api/v1/<service>/status/{job_id}  →  { "status": "...", "<media>_url": "..." }

Status values:

  • Image/music endpoints: PENDING, QUEUED, PROCESSING, COMPLETED, FAILED
  • Speech/subtitle endpoints: PENDING, PROCESSING, COMPLETED, FAILED
  • Video/download endpoints: processing, completed, failed, cancelled

Response URL field by media type:

Media typeResponse field
Imagesimage_url
Videosvideo_url
Audio / musicaudio_url (music also returns audio_urls[])
Sora 2 charactercharacter_id (a string ID, not a URL)

Poll every 3–5 seconds until the terminal status is reached.

Quick Start

1. Generate an image (text-to-image)

import httpx, time

BASE = "https://your-api-host/api/v1"
HEADERS = {"Authorization": "Bearer YOUR_API_KEY"}

r = httpx.post(f"{BASE}/text-to-image/generate", headers=HEADERS, json={
    "prompt": "A futuristic city at sunset, ultra-detailed",
    "aspect_ratio": "16:9"
})
job_id = r.json()["job_id"]

while True:
    s = httpx.get(f"{BASE}/text-to-image/status/{job_id}", headers=HEADERS).json()
    if s["status"] == "COMPLETED":
        print("Image URL:", s["image_url"])
        break
    if s["status"] == "FAILED":
        raise RuntimeError("Job failed")
    time.sleep(4)

2. Generate a video (Sora 2 text-to-video)

r = httpx.post(f"{BASE}/sora2-text-to-video/create", headers=HEADERS, json={
    "prompt": "A dragon flying over mountains at dawn",
    "aspect_ratio": "landscape",
    "n_frames": "10"          # "10" or "15" as a string
})
job_id = r.json()["job_id"]
# Poll /sora2-text-to-video/status/{job_id} → s["video_url"]

3. Chat with Phaya-GPT

r = httpx.post(f"{BASE}/phaya-gpt/chat/completions", headers=HEADERS, json={
    "messages": [{"role": "user", "content": "Hello, what can you do?"}],
    "stream": False
})
print(r.json()["message"]["content"])   # flat dict — NOT choices[0].message.content

Additional Resources

Local Binary Requirements

Most features only need python3 (or curl) and your API key — all AI processing is remote.

Two optional features invoke local binaries on your machine:

  • POST /image-to-video/create (FFmpeg local) — requires ffmpeg installed locally
  • POST /video-download/create (yt-dlp) — the download runs server-side via yt-dlp on the Phaya host, not locally

No local AI models, GPU, or disk-intensive operations are required by this skill.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

Expedy

Expedy integration. Manage Organizations, Pipelines, Users, Filters. Use when the user wants to interact with Expedy data.

Registry SourceRecently Updated
General

Evenium

Evenium integration. Manage Events, Users, Roles. Use when the user wants to interact with Evenium data.

Registry SourceRecently Updated
General

Exhibitday

ExhibitDay integration. Manage Organizations. Use when the user wants to interact with ExhibitDay data.

Registry SourceRecently Updated
General

Enigma

Enigma integration. Manage Deals, Persons, Organizations, Leads, Projects, Activities and more. Use when the user wants to interact with Enigma data.

Registry SourceRecently Updated