Automate web browsers, scrape pages, search the web, and run AI prompts on live websites via HARPA AI Grid REST API
Skills(SKILL.md)は、AIエージェント(Claude Code、Cursor、Codexなど)に特定の能力を追加するための設定ファイルです。
詳しく見る →Automate web browsers, scrape pages, search the web, and run AI prompts on live websites via HARPA AI Grid REST API
Build and run Gemini 2.5 Computer Use browser-control agents with Playwright. Use when a user wants to automate web browser tasks via the Gemini Computer Use model, needs an agent loop (screenshot → function_call → action → function_response), or asks to integrate safety confirmation for risky UI actions.
Prompt injection detection and security scanning for OpenClaw agents. Installs the ai-sentinel plugin via OpenClaw CLI, configures plugin settings, and offers local (Community) or remote (Pro) classification with dashboard reporting. All configuration changes require explicit user confirmation.
>
>
工作进度检查技能 - 定期检查待办事项 + 子代理超时/消失检测与自动恢复 + 全量会话监控
>
Search through past OpenClaw session transcripts (JSONL files in `~/.openclaw/agents/*/sessions/`).
Read supported mailbox providers such as 126, 163, and Gmail, identify invoice attachments or invoice download links, archive invoices by month, deduplicate by invoice number and amount, and prepare monthly reports plus delivery bundles. Use when OpenClaw needs to sync invoice mail, investigate failed downloads, package a month's archive, or summarize totals and high-value invoices.
Answers questions based on the content of uploaded documents (PDF, DOCX, TXT), supporting individual files or entire folders.
saved-markdown
*"At every crossroads lies a message. Chance is the messenger. You are the reader."*
Safety guardrails for autonomous agents. Intercepts destructive actions, requires structured approval before irreversible operations, with CLI-backed cryptographic audit trails and policy enforcement. Your lobster has claws. Keel is the rubber band.
Extract architectural decisions from conversations. Identifies problem-solution pairs, trade-off discussions, and explicit choices. Use when analyzing session transcripts for ADR generation.
Turn a set of local paths (or a beagle project's conventional knowledge locations) into a cited, structured extraction of insights, context, decisions, and raw detail.
Automatic graph layout using dagre with React Flow (@xyflow/react). Use when implementing auto-layout, hierarchical layouts, tree structures, or arranging nodes programmatically. Triggers on dagre, auto-layout, automatic layout, getLayoutedElements, rankdir, hierarchical graph.
Implements agents using Deep Agents. Use when building agents with create_deep_agent, configuring backends, defining subagents, adding middleware, or setting up human-in-the-loop workflows.
Implements stateful agent graphs using LangGraph. Use when building graphs, adding nodes/edges, defining state schemas, implementing checkpointing, handling interrupts, or creating multi-agent systems with LangGraph.
Architectural guidance for building node-based UIs with React Flow. Use when designing flow-based applications, making decisions about state management, integration patterns, or evaluating whether React Flow fits a use case.
Implements React Flow node-based UIs correctly using @xyflow/react. Use when building flow charts, diagrams, visual editors, or node-based applications with React. Covers nodes, edges, handles, custom components, state management, and viewport control.
React Flow (@xyflow/react) for workflow visualization with custom nodes and edges. Use when building graph visualizations, creating custom workflow nodes, implementing edge labels, or controlling viewport. Triggers on ReactFlow, @xyflow/react, Handle, NodeProps, EdgeProps, useReactFlow, fitView.
React Router v7 best practices for data-driven routing. Use when implementing routes, loaders, actions, Form components, fetchers, navigation guards, protected routes, or URL search params. Triggers on createBrowserRouter, RouterProvider, useLoaderData, useActionData, useFetcher, NavLink, Outlet.
本技能用于将用户的查询请求转发到 DeepSeek 网页版,利用 DeepSeek 的联网搜索能力获取最新信息,同时减轻主模型的负担。
抖音视频下载工具 - 解析抖音链接,下载视频并发送
Analyze bank transactions, categorize spending, track monthly budgets, detect overspending and anomalies. Outputs interactive HTML report.
Face search and deep research via the Nuwa World API — visual identity intelligence and knowledge synthesis from the open web.
Analyze job descriptions, extract real hiring signals, assess candidate fit, and provide resume tailoring advice.
Search, download, and upload files in a single Seafile library using a repo token. Use when an agent needs library-scoped filename or path search, exact-path downloads, or uploads into a Seafile library via the repo-token API.
Start a Selenium‑controlled Chrome browser, open a URL, take a screenshot, and report progress. Supports headless mode and optional proxy.
Verified molt swarms - cryptographically prove your identity with Ed25519 signatures and WorldID proof-of-personhood. Register with services and verify unique human operators.
Cryptographic identity and coordination for AI agents. Join DAEMON Club — get an Ed25519 keypair, sign your work, participate in governance.
>
>
>
台股月營收抓取與分析。從公開資訊觀測站(MOPS)取得上市/上櫃公司當月營收,計算YoY/MoM年增率與月增率,輸出BEAT/MISS/NEUTRAL信號。用於:(1)月營收公布後自動分析 (2)觸發投資Pipeline決策 (3)更新ANALYSIS_INDEX監控條件
Interact with PikaBoard task management API. Use when creating, updating, listing, or managing tasks. Agent-first kanban for AI teams. Triggers on: tasks, kanban, board, todo, backlog, sprint.
xhs-publish
Import Personal Documents. Use when the user wants to import a local directory or explicit files into 材序 at the ingest stage, including “导入一批材料”“先把这个目录吃进去”“先建立可继续建库的解析上下文”. Prefer caixu-skill when the user asks for the full end-to-end mainline or is unsure which stage to run. This skill creates or loads a library, starts an ingest pipeline run, lets the agent choose per-file routes over low-level OCR and parser tools, persists ParsedFile records, and hands off library_id plus run_id to build-asset-library. Do not use it for asset extraction, lifecycle judgment, package building, or submission.
Web scraping using local Crawl4AI instance. Use for fetching full page content with JavaScript rendering. Better than Tavily for complex pages. Unlimited usage.
Fetch articles from PubMed Central using NCBI APIs. Search journals, retrieve full text via OAI-PMH, batch harvest for RAG pipelines. No API key required.
Universal Ragflow API client for RAG operations. Create datasets, upload documents, run chat queries against knowledge bases. Self-hosted RAG platform integration.
Financial analysis and research workflows for market research, equity research, comparable companies, precedent transactions, and DCF valuation. Use when asked to build or critique DCF models, comps/precedents tables, competitor or strategy analysis, market sizing, or to deliver analyst-ready outputs in Excel, PowerPoint, or Markdown from CSV/Excel/SQL/API/web data.
Target company evaluation and deal qualification for PE and strategic buyers
SEC EDGAR filing analysis for M&A due diligence — extract financials, detect risks, and track corporate events from 10-K, 10-Q, and 8-K filings
Scans OpenClaw skills for security vulnerabilities and suspicious patterns before installation
Boost your productivity with automated task management
Expert Cinema Director skill for Seedance 2.0 (ByteDance) — high-fidelity video generation using technical camera grammar and multimodal references. Supports text-to-video, image-to-video, and video extension.
Fetches current top trending topics on X (Twitter) for any country using public aggregators.
>
>