web scraping & data collection Claude 스킬을 발견하세요. 17개의 스킬을 탐색하고 AI 워크플로우에 완벽한 기능을 찾아보세요.
Performs concurrent, recursive web scraping to extract clean text content while maintaining URL-based directory structures and respecting domain boundaries.
Automates complex web data extraction, page crawling, and document parsing using the Firecrawl API.
Extracts clean web content, captures screenshots, and parses PDFs using the powerful Firecrawl API.
Empowers Claude with real-time web search, content extraction, and advanced crawling capabilities using the Tavily API.
Fetches web documentation, extracts specific topics using AI subagents, and generates structured markdown summaries.
Converts entire PDF documents into clean, structured Markdown while preserving formatting, tables, and images for seamless context loading.
Empowers Claude with real-time web search, automated content extraction, and deep research capabilities using the Tavily API.
Provides real-time web search, content extraction, and automated site crawling capabilities powered by the Tavily API.
Generates structured JSON templates for the Obsidian Web Clipper by analyzing web page metadata and verifying CSS selectors.
Integrates Gemini AI models with real-time web content through Google Search grounding to provide verified information and automatic citations.
Extracts and correlates high-quality food images from delivery platforms and restaurant websites to populate digital catalogs and menus.
Scrapes web pages and WeChat articles to produce clean, noise-free Markdown content for processing, translation, or archival.
Downloads videos, audio, and clean paragraph-style transcripts from YouTube and other yt-dlp supported platforms.
Downloads high-quality video and audio from YouTube and HLS streaming platforms like Vimeo and Mux using yt-dlp and ffmpeg.
Orchestrates multi-agent parallel workflows to perform comprehensive web research, data collection, and structured report generation.
Conducts deep, multi-source web research and synthesis without cluttering the primary conversation context.
Enforces a rigorous, test-driven development workflow for building and maintaining web scrapers and data extraction agents.
Reads and processes RSS/Atom feeds from any blog URL with intelligent automatic feed discovery.
Simplifies the creation of standardized data collection forms in Excel for mobile survey platforms like ODK and KoBoToolbox.
Conducts comprehensive market analysis and trend forecasting across the consumer, technology, healthcare, and finance sectors.
Installs and configures the Jina AI MCP server to enable high-quality web scraping and URL content extraction within Claude Code.
Automates the installation and configuration of the VeyraX MCP server for enhanced web search and content extraction.
End of results