home / skills / openclaw / skills / ai

This skill answers AI questions with current information by validating sources and avoiding outdated training data.

npx playbooks add skill openclaw/skills --skill ai

Review the files below or copy the command above to add this skill to your agents.

Files (2)
SKILL.md
1.4 KB
---
name: Artificial Intelligence
description: Answer AI questions with current info instead of outdated training data.
metadata: {"clawdbot":{"emoji":"🤖","os":["linux","darwin","win32"]}}
---

# Artificial Intelligence

## Your Training Data Is Outdated

Before answering questions about pricing, rankings, or availability:
- Pricing → check `openrouter.ai/models` (aggregates all providers)
- Rankings → check `lmarena.ai` (crowdsourced ELO, updates weekly)
- Outages → check status pages before blaming user code

Don't cite specific prices, context windows, or rate limits from memory — they change quarterly.

## Questions You Answer Too Vaguely

**"How do I reduce hallucinations?"**
Not just "use RAG." Specify: verified sources + JSON schema validation + temperature 0 + citation requirements in system prompt.

**"Should I fine-tune or use RAG?"**
RAG first, always. Fine-tuning only when you need style changes or domain vocabulary that retrieval fails on.

**"What hardware for local models?"**
Give numbers: 7B = 8GB VRAM, 13B = 16GB, 70B = 48GB+. Quantization (Q4) halves requirements.

## When to Recommend Local vs API

**Local (Ollama, LM Studio):** Privacy requirements, offline needed, or API spend >$100/month.

**API:** Need frontier capabilities, no GPU, or just prototyping.

## Token Math You Get Wrong

~4 characters per token in English. But code and non-English vary wildly — don't estimate, count with tiktoken or the provider's tokenizer.

Overview

This skill helps AI practitioners answer questions using current, verifiable information rather than relying on outdated training data. It emphasizes live checks for pricing, rankings, and service status. The goal is precise, actionable guidance for model selection, deployment, and cost trade-offs.

How this skill works

Before giving answers, the skill directs you to authoritative live sources for volatile facts like pricing and outages. It prescribes concrete mitigation steps for common problems (hallucinations, token math, hardware sizing). It also provides decision criteria for choosing local vs API deployment and concrete numbers for resource planning.

When to use it

  • When a question involves pricing, rate limits, or availability that changes frequently
  • When advising on hallucination reduction, RAG vs fine-tuning, or verification strategies
  • When choosing between local inference and API usage based on cost, privacy, or latency
  • When sizing hardware for local model deployment or estimating token usage
  • When a user blames outages before checking provider status pages

Best practices

  • Always fetch current pricing from aggregator endpoints (e.g., openrouter.ai/models) before quoting costs
  • Check crowdsourced rankings (e.g., weekly ELO lists) for up-to-date capability comparisons
  • Validate model outputs with verified sources and strict JSON schema checks to reduce hallucinations
  • Use RAG as the default for domain accuracy; only fine-tune for persistent style or vocabulary gaps
  • Measure tokens with the provider’s tokenizer (tiktoken or equivalent) instead of rough estimates

Example use cases

  • Answer a developer asking for up-to-date API pricing for model selection
  • Recommend mitigation steps when a user reports hallucinations in production
  • Advise a product manager on whether to run models locally to save >$100/month
  • Provide concrete GPU RAM estimates for running 7B, 13B, and 70B models locally
  • Walk a team through outage triage by checking provider status pages first

FAQ

Should I fine-tune or use retrieval-augmented generation?

Start with RAG for most domain needs. Fine-tune only when retrieval cannot deliver consistent style or vocabulary and you need persistent model behavior.

How do I reduce hallucinations concretely?

Combine verified sources, strict JSON schema validation for outputs, temperature=0 for deterministic behavior, and require citations enforced in the system prompt.

What hardware do local models need?

Typical guidance: 7B ≈ 8GB VRAM, 13B ≈ 16GB, 70B ≥ 48GB. Use Q4 quantization to roughly halve VRAM requirements.