Skip to content
AI Primer
πŸ€– ML/AINew

gemini-interactions-api

google-geminiby google-gemini4 days ago3.5k

Use this skill when writing code that calls the Gemini API for text generation, multi-turn chat, multimodal understanding, image generation, streaming responses, background research tasks, function calling, structured output, or migrating from the old generateContent API. This skill covers the Interactions API, the recommended way to use Gemini models and agents in Python and TypeScript.

Install

npx skills add https://github.com/google-gemini/gemini-skills --skill gemini-interactions-api
Show step-by-step
  1. 1

    Open your terminal

    • Mac: Press ⌘ Space, type "Terminal", press Enter
    • Windows: Press Win R, type "cmd", press Enter
  2. 2

    Paste the command above and press Enter

    Use the Copy command button, then paste in your terminal (Mac: ⌘V, Windows: Ctrl V).

  3. 3

    Restart Claude Code

    Close and reopen Claude Code, or start a new session, so it picks up the new skill.

Where it lives
~/.claude/skills/google-gemini--gemini-skills--skills--gemini-interactions-api/
β”œβ”€β”€ SKILL.md
└── ... (skill resource files)
View on GitHub

Comments

Always review skill code before installing. Third-party skills may contain scripts that run on your machine.

Related skills

πŸ€– ML/AI

comfyui

Generate images, video, and audio with ComfyUI β€” install, launch, manage nodes/models, run workflows with parameter injection. Uses the official comfy-cli for lifecycle and direct REST/WebSocket API for execution.

by NousResearch Β· 12 days ago146.2k
πŸ€– ML/AI

hyperframes

Create HTML-based video compositions, animated title cards, social overlays, captioned talking-head videos, audio-reactive visuals, and shader transitions using HyperFrames. HTML is the source of truth for video. Use when the user wants a rendered MP4/WebM from an HTML composition, wants to animate text/logos/charts over media, needs captions synced to audio, wants TTS narration, or wants to convert a website into a video.

by NousResearch Β· 7 days ago146.2k
πŸ’» Developer Tools

claude-api

Build, debug, and optimize Claude API / Anthropic SDK apps. Apps built with this skill should include prompt caching. Also handles migrating existing Claude API code between Claude model versions (4.5 β†’ 4.6, 4.6 β†’ 4.7, retired-model replacements). TRIGGER when: code imports `anthropic`/`@anthropic-ai/sdk`; user asks for the Claude API, Anthropic SDK, or Managed Agents; user adds/modifies/tunes a Claude feature (caching, thinking, compaction, tool use, batch, files, citations, memory) or model (Opus/Sonnet/Haiku) in a file; questions about prompt caching / cache hit rate in an Anthropic SDK project. SKIP: file imports `openai`/other-provider SDK, filename like `*-openai.py`/`*-generic.py`, provider-neutral code, general programming/ML.

by anthropics Β· 20 days ago132.8k
πŸ€– ML/AI

paddleocr-text-recognition

Use this skill whenever the user wants text extracted from images, photos, scans, screenshots, or scanned PDFs. Returns exact machine-readable strings with line-level text and optional bbox coordinates. Strong accuracy for CJK, small print, and handwritten text. Trigger terms: OCR, ζ–‡ε­—θ―†εˆ«, 图片转文字, ζˆͺε›Ύθ―†ε­—, 提取图中文字, 扫描识字, θ―†ε­—, ηΊ―ζ–‡ε­—, plain text extraction, 坐标, 检桋摆, bbox, bounding box, image to text, screenshot, photo scan, recognize text.

by PaddlePaddle Β· 18 days ago77.7k