miniflux-legacy/docs/full-article-download.markdown
2015-04-10 20:34:48 -04:00

1.8 KiB

Full article download

For feeds that accept only a summary, it's possible to download the full content directly from the original website.

How the content grabber works?

  1. Try with rules first (Xpath patterns) for the domain name
  2. Try to find the text content by using common attributes for class and id
  3. Finally, if nothing is found, the feed content is displayed

The content downloader use a fake user agent, actually Google Chrome under Mac Os X.

However the content grabber doesn't work very well with all websites.

The best results are obtained with Xpath rules file.

How to write a grabber rules file?

Add a PHP file to the directory rules, the filename must be the domain name with the suffix .php:

Example with the BBC website, www.bbc.co.uk.php:

<?php

return array(
    'test_url' => 'http://www.bbc.co.uk/news/world-middle-east-23911833',
    'body' => array(
        '//div[@class="story-body"]',
    ),
    'strip' => array(
        '//script',
        '//form',
        '//style',
        '//*[@class="story-date"]',
        '//*[@class="story-header"]',
        '//*[@class="story-related"]',
        '//*[contains(@class, "byline")]',
        '//*[contains(@class, "story-feature")]',
        '//*[@id="video-carousel-container"]',
        '//*[@id="also-related-links"]',
        '//*[contains(@class, "share") or contains(@class, "hidden") or contains(@class, "hyper")]',
    )
);

Actually, only body, strip and test_url are supported.

Don't forget to send a pull request or a ticket to share your contribution with everybody.

List of content grabber rules

List of rules included by default.