web scraping & data collection Claude 스킬을 발견하세요. 15개의 스킬을 탐색하고 AI 워크플로우에 완벽한 기능을 찾아보세요.
Extract data, manipulate documents, and programmatically generate PDF files using specialized libraries and tools.
Downloads high-quality videos and audio from YouTube and other platforms for offline viewing, editing, or archival.
Orchestrates a structured, multi-agent workflow to conduct deep-dive research, synthesize information from multiple sources, and generate comprehensive reports.
Accesses USPTO APIs to perform comprehensive patent and trademark searches, retrieve examination histories, and analyze intellectual property data.
Performs AI-powered web searches with real-time information and source citations to access data beyond the model's knowledge cutoff.
Extracts and saves YouTube video subtitles or transcripts to local text files using command-line tools or automated browser interaction.
Performs real-time AI web searches with citations using Perplexity models to provide up-to-date information and scientific literature.
Equips Claude with high-performance web search capabilities and deep content extraction tools powered by the Tavily API.
Transforms unstructured files like PDFs, Word documents, and presentations into structured Pydantic models using LlamaExtract services.
Parses and extracts structured content from complex PDF documents using LlamaParse and agentic OCR capabilities.
Downloads high-quality video and audio content from YouTube and HLS-based streaming platforms while resolving common authentication and formatting issues.
Automates multi-step information gathering and synthesis using structured planning and parallel subagents.
Automates web content extraction using a four-tier fallback strategy to bypass bot detection and CAPTCHAs.
Downloads high-quality videos and audio from YouTube and other platforms for offline access and archival.
Searches and retrieves life sciences preprints from the bioRxiv database with advanced filtering and PDF download capabilities.
Conducts deep web investigations by delegating tasks to specialized subagents and synthesizing findings into organized reports.
Converts complex file formats including PDF, Office documents, and media into clean Markdown optimized for LLM processing.
Automates web content retrieval using a progressive four-tier fallback strategy to bypass bot detection and access restrictions.
Conducts deep, multi-faceted web research by orchestrating parallel subagents to plan, gather, and synthesize complex information.
Retrieves web content through a four-tier progressive fallback strategy to bypass bot detection and access restrictions.
Implements a four-tier progressive scraping strategy to bypass bot detection and reliably extract web content.
Implements a four-tier progressive escalation strategy to reliably scrape web content and bypass advanced bot detection.
Downloads high-quality video and audio content from YouTube and other platforms directly through your terminal workspace.
Conducts systematic web research through autonomous subagent delegation and multi-source synthesis.
Extracts subtitles and transcripts from YouTube videos directly into local text files using command-line tools or browser automation.
Manages YouTube channel tracking by automating video content collection, transcript retrieval, and structured summary generation.
Extracts text and structural data from complex Microsoft Word documents containing nested tables, checkboxes, and multi-layered cell layouts.
Extracts clean source code from GitHub file URLs using the GitHub CLI to bypass web scraping noise and HTML clutter.
Empowers Claude with real-time internet research capabilities by integrating Gemini's Google Search tool directly into the terminal workflow.
Searches and retrieves life sciences preprints from the bioRxiv server using keywords, authors, date ranges, and categories.
Scroll for more results...