Discover Agent Skills for web scraping & data collection. Browse 21skills for Claude, ChatGPT & Codex.
Extract data, manipulate documents, and programmatically generate PDF files using specialized libraries and tools.
Extracts and analyzes competitor advertising strategies across platforms to provide actionable messaging and creative insights.
Downloads high-quality videos and audio from YouTube and other platforms for offline viewing, editing, or archival.
Orchestrates a structured, multi-agent workflow to conduct deep-dive research, synthesize information from multiple sources, and generate comprehensive reports.
Accesses USPTO APIs to perform comprehensive patent and trademark searches, retrieve examination histories, and analyze intellectual property data.
Accesses the ClinicalTrials.gov API v2 to search, filter, and export clinical study data for medical research and patient matching.
Automates biomedical literature searches and programmatic data extraction from the PubMed database using E-utilities and advanced MeSH queries.
Performs AI-powered web searches with real-time information and source citations to access data beyond the model's knowledge cutoff.
Extracts and saves YouTube video subtitles or transcripts to local text files using command-line tools or automated browser interaction.
Performs real-time AI web searches with citations using Perplexity models to provide up-to-date information and scientific literature.
Equips Claude with high-performance web search capabilities and deep content extraction tools powered by the Tavily API.
Converts diverse file formats including PDFs, Office documents, and media into structured, token-efficient Markdown for LLM processing.
Transforms unstructured files like PDFs, Word documents, and presentations into structured Pydantic models using LlamaExtract services.
Parses and extracts structured content from complex PDF documents using LlamaParse and agentic OCR capabilities.
Replicates existing websites into production-ready Next.js 16 and Tailwind CSS v4 codebases using Firecrawl MCP.
Downloads high-quality video and audio content from YouTube and HLS-based streaming platforms while resolving common authentication and formatting issues.
Automates multi-step information gathering and synthesis using structured planning and parallel subagents.
Conducts deep web investigations by delegating tasks to specialized subagents and synthesizing findings into organized reports.
Searches and retrieves life sciences preprints from the bioRxiv database with advanced filtering and PDF download capabilities.
Automates web content retrieval using a progressive four-tier fallback strategy to bypass bot detection and access restrictions.
Implements a four-tier progressive scraping strategy to bypass bot detection and reliably extract web content.
Conducts systematic web research through autonomous subagent delegation and multi-source synthesis.
Retrieves web content through a four-tier progressive fallback strategy to bypass bot detection and access restrictions.
Implements a four-tier progressive escalation strategy to reliably scrape web content and bypass advanced bot detection.
Converts complex file formats including PDF, Office documents, and media into clean Markdown optimized for LLM processing.
Downloads high-quality videos and audio from YouTube and other platforms for offline access and archival.
Automates web content extraction using a four-tier fallback strategy to bypass bot detection and CAPTCHAs.
Conducts deep, multi-faceted web research by orchestrating parallel subagents to plan, gather, and synthesize complex information.
Downloads high-quality video and audio content from YouTube and other platforms directly through your terminal workspace.
Extracts subtitles and transcripts from YouTube videos directly into local text files using command-line tools or browser automation.
Scroll for more results...