Python crawler, the full name is Python web crawler, is a program or script that automatically captures World Wide Web information according to certain rules. It is mainly used to capture securities trading data and weather data. , website user data, image data, etc., Python has a large number of built-in libraries to support the normal functions of web crawlers, mainly of several types. The following article will introduce it to you.
1. Python crawler network library
Python crawler network library mainly includes: urllib, requests, grab, pycurl, urllib3, httplib2, RoboBrowser, MechanicalSoup, mechanism, socket, Unirest for Python, hyper , PySocks, treq and aiohttp, etc.
2. Python web crawler framework
The Python web crawler framework mainly includes: grab, scrapy, pyspider, cola, portia, restkit and demiurge, etc.
3. HTML/XML parser
● lxml: an efficient HTML/XML processing library written in C language. Supports XPath.
● cssselect: Parse DOM tree and CSS selector.
● Pyquery: Parse DOM tree and jQuery selector.
●BeautifulSoup: Inefficient HTML/XML processing library, implemented in pure Python.
● html5lib: Generates the DOM of HTML/XML documents according to the WHATWG specification. This specification is used in all current browsers.
● Feedparser: Parse RSS/ATOM feeds.
●MarkupSafe: Provides safe escaped strings for XML/HTML/XHTML.
● xmltodict: A Python module that makes processing XML feel like JSON.
● xhtml2pdf: Convert HTML/CSS to PDF.
● untangle: Easily convert XML files into Python objects.
4. Text processing
A library for parsing and manipulating simple text.
● difflib: (Python standard library) helps perform differential comparisons.
● Levenshtein: Quickly calculate Levenshtein distance and string similarity.
● fuzzywuzzy: fuzzy string matching.
● esmre: Regular expression accelerator.
● ftfy: Automatically organize Unicode text to reduce fragmentation.
5. Specific format file processing
Library for parsing and processing specific text formats.
● tablib: A module that exports data to XLS, CSV, JSON, YAML and other formats.
●textract: Extract text from various files, such as Word, PowerPoint, PDF, etc.
●messytables: A tool for parsing messy tabular data.
● Rows: A common data interface that supports many formats (currently supports CSV, HTML, XLS, TXT: more will be provided in the future!).
The above is the detailed content of What libraries are used to write crawlers in Python?. For more information, please follow other related articles on the PHP Chinese website!