Extracting text from HTML with Reporter

Published on 31 August 2013

Note: This post originally appeared on November 11, 2012 on Visual Revenue's blog. The text from the Github repository has been appended for completeness.

At Visual Revenue, we use many great open source software packages. Our favorites include: ubuntu, vim, emacs, git, mongodb, selenium, redis, ipython, scikit-learn, pandas, R, and D3. This open source tech helps us to develop a state-of-the-art recommendation platform for news editors. I'm happy to announce that we can now give something back to the open source community: Reporter.

Reporter is being developed at Visual Revenue, Inc. where it is used to extract the main text from news articles. The name Reporter and internal terms are inspired by the news domain.

Reporter is a flexible tool that extracts text from HTML. At Visual Revenue, we actively use it to extract the main text from news articles. It is written in Python, which allows you to embed it in your own Python applications. You can also use Reporter from the command line.

In short, Reporter:

  • Extracts the main text from HTML.
  • Uses a white-box scoring algorithm to determine the main text container.
  • Can easily be extended.
  • Supports Unicode without pain.
  • Has awesome debugging facilities.

Below, I'll explain how the scoring algorithm works.


You can read more about Reporter on Github: https://github.com/visualrevenue/reporter. It's also available as a PyPi package: http://pypi.python.org/pypi/reporter/0.1.2.


Reporter can be invoked from the command line:

$ reporter.py --url URL

The HTML from URL will be parsed by Beautiful Soup and the main text will be printed on stdout. If the --debug flag is added, the text and HTML will be saved to file. The HTML will be styled as follows. Each tag will get a background color based on its score, ranging from red (low score) to green (high score). Moreover, the tag that is selected as news container (see below) will have a blue dashed line.


If the --test flag is given, all files in ./test/input will be processed, and the text and HTML will be saved, as in --debug. This is useful for processing many local files, so that these only have to downloaded once.

Please see ./reporter.py --help for more options.

Reporter can also be used from Python:

my_reporter = Reporter()
print my_reporter.report_news()

Scoring algorithm

To extract the main text from an HTML document, Reporter gives each HTML tag (e.g., DIV, H1, and P) a score. The text contained in the tag with the highest score is returned as the main text of the news article.

The main part of the scoring algorithm is based on traversing the parsed HTML and works as follows. Reporter traverses the HTML in reverse order, i.e., it starts at the leaves of the DOM tree. Each tag is scored either as a paragraph or as a container. A tag is considered to be a paragraph (in the abstract sense, not in the P sense) when it contains more than 10 characters*, otherwise it is considered to be a container. The exact scoring of a tag is defined in the Autocue. An Autocue is a list of scoring rules that get triggered at various stages. For example, when a tag is to be scored as a paragraph, one rule may count the number of words and return 2 points per word. Once a tag (and its siblings) are scored, its parent is scored. If the parent is also considered to be a paragraph, which happens, for instance, with the P tag in:

<DIV><P>Hello World, this is the <B>Reporter package</B></P></DIV>

the scores of the B tag are discarded and the complete text is re-scored. The DIV tag is scored as a container because (in this case) it contains no text by itself. In fact, there is an important scoring rule which penalises containers. If such a rule would not be included, the HTML tag would always receive the highest score, which would not be very effective.

*) Currently, this is the only heuristic that is hard-coded. In Readability, which served as the inspiration for Reporter, all scoring is hard-coded.

As mentioned, a scoring rule is triggered at a certain stage as the Reporter is processing the Autocue. Below, we list and explain the seven triggers with Python code. (The complete default Autocue is in autocues.py, which is easily extensible with additional rules.)

  • HTML, operates on the raw HTML. For example: split a paragraph with two consecutive line breaks into two paragraphs

    default_autocue.append((RegExReplacer(pattern='<br */? *>[\\r\\n]\*<br */? *>', repl='</p><p>'), HTML))
  • PRE_TRAVERSAL, scores or prunes (deletes tags) before the DOM is traversed. This is useful for getting rid of specific tags such as footers, or give positive scores to certain tags For example, delete all comments (specific to a certain news property):

    default_autocue.append((CSSSelector("div#comments", Pruner()),

Now, the HTML will be traversed as explained above.

  • EVAL_PARAGRAPH, scores a tag as a paragraph. For example, by counting words.

      (Scorer(RegExMatcher("(\w)+(['`]\w)?", factor=2, name="word"),
      reset_children=True), EVAL_PARAGRAPH))
  • EVAL_CONTAINER, scores a tag as a container. For example, combining the scores of the children tags with a 70 points penalty, giving a minimal score of 0.

      (ScoreAggregator(start_score=-70, vmin=0),

This concludes the traversing of the HTML.

  • POST_TRAVERSAL, scores or prunes tags after Reporter has traversed the HTML.

The tag with the highest score is selected as news container.

  • NEWS_CONTAINER is like POST_TRAVERSAL but only applies to the tag that is selected as news container.

    Example: penalize DIVs inside the news container:

      (CSSSelector("div", Scorer(FixedValue(-60))),

    Example: Get rid of any tags that have a score below -50:

      mode="upper", actor=Pruner()), NEWS_CONTAINER))
  • NEWS_TEXT, operates on the text inside the news container. For example, put all text on one line:

    default_autocue.append((RegExReplacer(pattern='\s+', repl=' '), NEWS_TEXT))

Now, we can return the final text as the main text of the HTML!

If you like what I had to say then you may want to follow me on Twitter.