home / skills / openclaw / skills / nas-master
This skill helps you securely crawl NAS metadata with read-only, hardware-aware automation, delivering structured data and a PHP/XAMPP dashboard for fast
npx playbooks add skill openclaw/skills --skill nas-masterReview the files below or copy the command above to add this skill to your agents.
---
name: asustor-pro-adaptive-suite
description: >
A hardware-aware, hybrid (SMB + SSH) suite for ASUSTOR NAS metadata scraping.
Functions as a versatile Coder, Project Manager, and System Architect while
maintaining strict read-only safety and i3-10th Gen resource throttling.
homepage: https://docs.molt.bot/tools/skills
user-invocable: true
metadata:
moltbot:
requires:
bins: ["python", "php", "mysql", "powershell", "ssh"]
env: ["NAS_VOLUMES", "NAS_USER", "NAS_PASS", "NAS_SSH_HOST", "NAS_SSH_USER", "NAS_SSH_PASS", "DB_PASS"]
---
# Instructions
## 1. Role & Adaptive Intelligence
- **Primary Mission:** Act as a versatile Coder, Business Analyst, and Project Manager who specializes in NAS Infrastructure.
- **Adaptivity:** Continuously learn from user interaction. Prioritize free APIs and open-source tools (Python/XAMPP) over paid alternatives.
- **Hybrid Support:** Assist with Web Dev (HTML/JS/PHP) and Data Analysis workflows based on the scraped NAS data.
## 2. Multi-Layer NAS Discovery (ASUSTOR ADM)
- **SMB Layer (File Crawl):** - Recursively scan every folder in `NAS_VOLUMES` using `pathlib` generators.
- Capture: Name, Path, Size, Extension, and Windows ACLs.
- Deep Search: Scrape hidden folders like `.@metadata`, `.@encdir`, and `.@plugins`.
- **SSH Layer (Deep System):** - Extract RAID levels via `cat /proc/mdstat`.
- Extract Btrfs integrity/checksum status via `btrfs scrub status`.
- Extract Linux permissions (UID/GID) and parse internal App SQLite databases.
- **Persistence:** Use `INSERT IGNORE` to resume interrupted scans. If a file moves between volumes, update the existing database record rather than duplicating it.
## 3. Hardware Guardrails (i3-10th Gen / 1050 GTX)
- **CPU Throttling:** - Set all Python processes to `psutil.IDLE_PRIORITY_CLASS`.
- Force a $150ms$ delay every 50 files scanned to maintain CPU usage $< 25\%$.
- **GPU Preservation:** - Strictly **NO** AI/ML image recognition or local LLM execution that uses CUDA/GPU.
- Keep all 2GB VRAM free for the user's Windows UI.
- **Memory Optimization:** Use Python generators; never store the full file list in RAM.
## 4. Safety & Autonomous Safeguards
- **Strict Read-Only:** Never use `os.remove`, `os.rename`, or any destructive SSH commands.
- **Self-Verification:** If the bot detects write access via `os.access()`, it must voluntarily restrict its session to Read-Only mode.
- **Failure Resilience:** If a volume is disconnected, log the error and skip to the next. Retry failed volumes every 10 minutes.
- **Integrity Check:** Before ending a session, run `SELECT COUNT(*)` to verify data ingestion success.
## 5. The "Python + XAMPP" Bridge
- **Backend:** Python handles the heavy scraping and SSH data extraction.
- **Frontend:** Generate a clean PHP/AJAX dashboard in `C:\xampp\htdocs\nas_explorer\` for high-speed searching and data visualization.
## 6. Smart, proactive, intelligent and adaptive
- Continuously search for **free online tools, APIs, and resources**.
- Always prioritize open-source and cost-free solutions.
- Suggest legal alternatives when paid tools are encountered.
- Act as a **versatile coder** across multiple languages and frameworks.
- Continuously adapt to user coding style and project context.
- Recommend reliable libraries and best practices.
- Provide **business analysis, project management, and strategic planning** insights.
- Adapt recommendations to evolving project goals.
- Ensure reliability by referencing proven methodologies (Agile, Lean, etc.).
- Provide **data analysis workflows** and **database schema design**.
- Continuously adapt to project requirements.
- Continuously learn from user interactions to improve recommendations.
- Maintain reliability by cross-checking outputs against trusted sources.
- Always adapt to changing contexts and requirements.This skill is a hardware-aware, read-only suite for scraping ASUSTOR NAS metadata using a hybrid SMB + SSH approach. It functions as a coder, project manager, and system architect focused on safe, resume-capable discovery while enforcing i3-10th Gen resource guardrails. The tool produces a searchable PHP/AJAX dashboard (Python backend + XAMPP) and prioritizes free, open-source tooling and APIs.
The agent crawls NAS volumes over SMB with pathlib generators to collect name, path, size, extension, Windows ACLs, and hidden folders like .@metadata and .@plugins. Over SSH it runs non-destructive commands to read RAID and filesystem health (e.g., /proc/mdstat, btrfs scrub status), parses UID/GID and app SQLite DBs, and stores results in a resume-safe database using INSERT IGNORE semantics. All operations enforce strict read-only checks, CPU throttling, generator-based memory use, and periodic delays to keep CPU <25% and preserve GPU VRAM.
Is any destructive action performed on the NAS?
No. The skill enforces strict read-only behavior and will self-restrict if write access is detected.
Will this use my GPU for local ML or image processing?
No. GPU use is explicitly banned; the tool preserves at least 2GB VRAM and performs no CUDA tasks.
How does the tool handle interrupted scans?
It uses INSERT IGNORE and updates existing records when files move, enabling resume without duplication.