A Python package & command-line tool to gather text on the Web ============================================================== .. meta:: :description lang=en: Trafilatura is a Python package and command-line tool designed to gather text on the Web. Its main applications are web crawling, downloads, scraping, and extraction of main texts, comments and metadata. .. image:: https://img.shields.io/pypi/v/trafilatura.svg :target: https://pypi.python.org/pypi/trafilatura :alt: Python package .. image:: https://img.shields.io/pypi/pyversions/trafilatura.svg :target: https://pypi.python.org/pypi/trafilatura :alt: Python versions .. image:: https://img.shields.io/codecov/c/github/adbar/trafilatura.svg :target: https://codecov.io/gh/adbar/trafilatura :alt: Code Coverage .. image:: https://static.pepy.tech/badge/trafilatura/month :target: https://pepy.tech/project/trafilatura :alt: Downloads .. image:: https://img.shields.io/badge/DOI-10.18653%2Fv1%2F2021.acl--demo.15-blue :target: https://aclanthology.org/2021.acl-demo.15/ :alt: Reference DOI: 10.18653/v1/2021.acl-demo.15 | .. image:: trafilatura-demo.gif :alt: Demo as GIF image :align: center :width: 85% :target: https://trafilatura.readthedocs.org/ Description ----------- Trafilatura is a **Python package and command-line tool** designed to gather text on the Web. It includes discovery, extraction and text processing components. Its main applications are **web crawling, downloads, scraping, and extraction** of main texts, metadata and comments. It aims at staying **handy and modular**: no database is required, the output can be converted to various commonly used formats. Going from raw HTML to essential parts can alleviate many problems related to text quality, first by avoiding the **noise caused by recurring elements** (headers, footers, links/blogroll etc.) and second by including information such as author and date in order to **make sense of the data**. The extractor tries to strike a balance between limiting noise (precision) and including all valid parts (recall). It also has to be **robust and reasonably fast**, it runs in production on millions of documents. This tool can be **useful for quantitative research** in corpus linguistics, natural language processing, computational social science and beyond: it is relevant to anyone interested in data science, information extraction, text mining, and scraping-intensive use cases like search engine optimization, business analytics or information security. Features ~~~~~~~~ - Web crawling and text discovery: - Focused crawling and politeness rules - Support for sitemaps (TXT, XML) and feeds (ATOM, JSON, RSS) - URL management (blacklists, filtering and de-duplication) - Seamless and parallel processing, online and offline: - URLs, HTML files or parsed HTML trees usable as input - Efficient and polite processing of download queues - Conversion of previously downloaded files - Robust and efficient extraction: - Main text (with LXML, common patterns and generic algorithms: jusText, fork of readability-lxml) - Metadata (title, author, date, site name, categories and tags) - Formatting and structural elements: paragraphs, titles, lists, quotes, code, line breaks, in-line text formatting - Comments (if applicable) - Output formats: - Text (minimal formatting or Markdown) - CSV (with metadata, `tab-separated values `_) - JSON (with metadata) - XML (with metadata, text formatting and page structure) and `TEI-XML `_ - Optional add-ons: - Language detection on extracted content - Graphical user interface (GUI) - Speed optimizations Evaluation and alternatives ~~~~~~~~~~~~~~~~~~~~~~~~~~~ For detailed results see the `benchmark `_ and `evaluation script `_. To reproduce the tests just clone the repository, install all necessary packages and run the evaluation script with the data provided in the *tests* directory. Other evaluations: ^^^^^^^^^^^^^^^^^^ - Most efficient open-source library in *ScrapingHub*'s `article extraction benchmark `_ - Best overall tool according to Gaël Lejeune & Adrien Barbaresi, `Bien choisir son outil d'extraction de contenu à partir du Web `_ (2020, PDF, French) In a nutshell ------------- Primary installation method is with a Python package manager: ``pip install trafilatura``. See `installation documentation `_. With Python: .. code-block:: python >>> import trafilatura >>> downloaded = trafilatura.fetch_url('https://github.blog/2019-03-29-leader-spotlight-erin-spiceland/') >>> trafilatura.extract(downloaded) # outputs main content and comments as plain text ... On the command-line: .. code-block:: bash $ trafilatura -u "https://github.blog/2019-03-29-leader-spotlight-erin-spiceland/" # outputs main content and comments as plain text ... For more information please refer to `usage documentation `_ and `tutorials `_. .. raw:: html License ------- *Trafilatura* is distributed under the `GNU General Public License v3.0 `_. If you wish to redistribute this library but feel bounded by the license conditions please try interacting `at arms length `_, `multi-licensing `_ with `compatible licenses `_, or `contacting me `_. See also `GPL and free software licensing: What's in it for business? `_ Context ------- These documentation pages also provide information on `concepts behind data collection `_ as well as practical tips on how to gather web texts (see `tutorials `_). Contributing ~~~~~~~~~~~~ Contributions are welcome! See `CONTRIBUTING.md `_ for more information. Bug reports can be filed on the `dedicated page `_. Many thanks to the `contributors `_ who submitted features and bugfixes! Roadmap ~~~~~~~ For planned enhancements and relevant milestones see `issues page `_. Author ~~~~~~ This effort is part of methods to derive information from web documents in order to build `text databases for research `_ (chiefly linguistic analysis and natural language processing). Extracting and pre-processing web texts to the exacting standards of scientific research presents a substantial challenge for those who conduct such research. Web corpus construction involves numerous design decisions, and this software package can help facilitate text data collection and enhance corpus quality. - Barbaresi, A. `Trafilatura: A Web Scraping Library and Command-Line Tool for Text Discovery and Extraction `_, Proceedings of ACL/IJCNLP 2021: System Demonstrations, 2021, p. 122-131. - Barbaresi, A. "`Generic Web Content Extraction with Open-Source Software `_", Proceedings of KONVENS 2019, Kaleidoscope Abstracts, 2019. - Barbaresi, A. "`Efficient construction of metadata-enhanced web corpora `_", Proceedings of the `10th Web as Corpus Workshop (WAC-X) `_, 2016. .. image:: https://img.shields.io/badge/DOI-10.18653%2Fv1%2F2021.acl--demo.15-blue :target: https://aclanthology.org/2021.acl-demo.15/ :alt: Reference DOI: 10.18653/v1/2021.acl-demo.15 .. image:: https://zenodo.org/badge/DOI/10.5281/zenodo.3460969.svg :target: https://doi.org/10.5281/zenodo.3460969 :alt: Zenodo archive DOI: 10.5281/zenodo.3460969 .. code-block:: shell @inproceedings{barbaresi-2021-trafilatura, title = {{Trafilatura: A Web Scraping Library and Command-Line Tool for Text Discovery and Extraction}}, author = "Barbaresi, Adrien", booktitle = "Proceedings of the Joint Conference of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: System Demonstrations", pages = "122--131", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2021.acl-demo.15", year = 2021, } You can contact me via my `contact page `_ or on `GitHub `_. Software ecosystem ~~~~~~~~~~~~~~~~~~ .. image:: software-ecosystem.png :alt: Software ecosystem :align: center :width: 65% *Trafilatura*: `Italian word `_ for `wire drawing `_. `Known uses of the software `_. Corresponding posts on `Bits of Language `_ (blog). Further documentation ===================== .. toctree:: :maxdepth: 2 installation usage tutorials evaluation corefunctions used-by background Indices and tables ================== * :ref:`genindex` * :ref:`modindex` * :ref:`search`