Track changes on websites via git
This tool checks all the websites listed in its config. When a change is detected, the new site is added to a git commit. It can then be inspected via normal git tooling.
Basically it's curl, sed++ and then git commit in a neat package.
See it in action (literally in GitHub Actions).
- GitHub Releases
- Arch Linux User Repository (AUR)
- Docker Hub Image
- Via rust and cargo: Clone →
cargo install --path .
Check out website-stalker-example which runs within GitHub actions.
-
First create a new folder / git repository for tracking website changes
mkdir personal-stalker cd personal-stalker git init website-stalker example-config > website-stalker.yaml
-
Add your favorite website to the configuration file
website-stalker.yaml. Also make sure to set the value of from to an email address of yours.website-stalker example-config > website-stalker.yaml nano website-stalker.yaml -
Run your newly added website. If you added
https://apple.com/newsroomuse something like this to test if everything works like you want:website-stalker run apple
-
Set up a cronjob /
systemd.timerexecuting the following command occasionallywebsite-stalker run --all --commit
The config describes a list of sites. Each site has a URL. Additionally, each site can have editors which are used before saving the file. Each editor manipulates the content of the URL.
# This is an example config
# The filename should be `website-stalker.yaml`
# and it should be in the working directory where you run website-stalker.
#
# For example run `website-stalker example-config > website-stalker.yaml`.
# Adapt the config to your needs and set the FROM email address which is used as a request header:
# https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/From
#
# And then do a run via `website-stalker run --all`.
---
from: my-email-address
sites:
- url: "https://edjopato.de/post/"
editors:
- css_select: article
- css_remove: a
- html_prettify
- regex_replace:
pattern: "(Lesezeit): \\d+ \\w+"
replace: $1
- url: "https://edjopato.de/robots.txt"There is a bigger config in my example repository. The example repository is also used by me to detect changes of interesting sites.
Options which are globally configured at the root level of the configuration file website-stalker.yaml.
Used as the From header in the web requests.
It is a required field.
The idea here is to provide a way for a website host to contact whoever is doing something to their web server. As this tool is self-hosted and can be run as often as the user likes this can annoy website hosts. While this tool is named "stalker" and is made to track websites it is not intended to annoy people.
This tool sets the User-Agent header to website-stalker/<version> https://github.com/EdJoPaTo/website-stalker and the From header to the user configured value.
This way both the creator and the user of this tool can be reached in case of problems.
from: my-email-addressAlternatively you can specify FROM via environment variable
export WEBSITE_STALKER_FROM=my-email-addressOptions available per site besides the editors which are explained below.
One or multiple URLs can be specified. The simple form is a single URL:
sites:
- url: "https://edjopato.de/"
- url: "https://edjopato.de/post/"It's also possible to specify multiple URL at the same time. This is helpful when multiple sites are sharing the same options (like editors).
sites:
- url:
- "https://edjopato.de/"
- "https://edjopato.de/post/"Allows HTTPS connections with self-signed or invalid / expired certificates.
From reqwests documentation:
You should think very carefully before using this method. If invalid certificates are trusted, any certificate for any site will be trusted for use. This includes expired certificates. This introduces significant vulnerabilities, and should only be used as a last resort.
Do you have a need for self-signed certificates or the usage of the system certificate store? Please share about it in Issue #39.
sites:
- url: "https://edjopato.de/post/"
accept_invalid_certs: trueOnly use HTTP/1 for the web request.
Back-ends might use HTTP/2 fingerprinting which could result in different or unusable output depending on what the back-end assumes about the client. HTTP/1 is a simpler protocol which does not allow such kinds of back-end optimizations.
sites:
- url: "https://edjopato.de/post/"
http1_only: trueOnly show warning when the site errors.
This is useful for buggy services which are sometimes just gone or for pages which will exist in the future but are not there yet. Personal example: A bad DNS configuration which lets the website appear nonexistent for some time.
This setting also skips errors from editors.
sites:
- url: "https://edjopato.de/might-appear-in-the-future"
ignore_error: trueOverrides the URL based default filename of the site.
Normally the filename is automatically derived from the URL.
For the following example it would be something like de-edjopato-api-token-0123456789-action-enjoy-20weather.html.
With the filename options it is saved as de-edjopato-api-weather.html instead.
sites:
- url: "https://edjopato.de/api?token=0123456789&action=enjoy%20weather"
filename: de-edjopato-api-weatherAdd additional HTTP headers to the request to the given site.
This is useful for sites that respond differently based on different headers.
Each header Key/Value pair is supplied as YAML String separated with a : followed by a space.
This is the same syntax as HTTP uses which sadly collides with YAML.
YAML assumes something with a : is an object.
Therefor you have to make sure to quote the headers.
Using a YAML object / key/value pair is also not possible as some header keys are allowed multiple times.
sites:
- url: "https://edjopato.de/"
headers:
- "Cache-Control: no-cache"
- "User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:106.0) Gecko/20100101 Firefox/106.0"Editors are manipulating the content of a webpage to simplify comparing them later on.
For example: If you are interested in the content of a webpage the <head> with changing style-sheets isn't interesting to you.
When keeping it, it will still create diffs which end up being commits.
This will create noise you're probably just going to ignore.
That's why editors exist.
Think of editors like a pipeline, the next one gets the input of the one before.
As some editors are assuming HTML input, they won't work (well) with non HTML input.
For example its kinda useless to use html_prettify after html_textify as text won't end up being pretty HTML.
For this reason editors like css_select are still producing valid HTML output.
There are probably more tasks out there that might be useful as editors. Feel free to provide an issue for an editor idea or create a Pull Request with a new editor.
Replaces every matching HTML element with its child nodes and returns the HTML.
Instead of css_remove this does not remove all the child nodes below.
Examples:
editors:
- css_flatten: div
- css_flatten: a[href^="#"] # flatten all local links away (starting with a #)Tries to remove every instance of matching HTML elements and returns the remaining HTML.
Opposite of css_select.
When the child nodes should be kept, use css_flatten.
Examples:
editors:
- css_remove: article
- css_remove: h1 a
- css_remove: h1 > aUse CSS Selectors to grab every instance of matching HTML elements and returns all of them.
If no matching HTML elements are found, this editor errors.
Examples:
editors:
- css_select: article
- css_select: h1 a
- css_select: h1 > aSort elements matching to the given CSS Selector. Other elements not matching are kept. Elements below different parents are sorted independently.
Basic example:
<div><p>C</p><p>B</p></div>
<div><p>D</p><p>A</p></div>with p as the selector will sort into this:
<div><p>B</p><p>C</p></div>
<div><p>A</p><p>D</p></div>Examples:
editors:
# Sort all articles
- css_sort:
selector: articleThe above example sorts by the whole element (outerHTML).
In order to sort by something specific for a given HTML element, editors can be used.
editors:
# Sort articles by their heading
- css_sort:
selector: article
sort_by: # the specified editors are applied to every selected HTML element independently
- css_select: h2This might still sort in surprising ways as things like attributes are still included (<h2 class="a">Z</h2> is sorted before <h2 class="z">A</h2>).
Therefore, editors like html_textify or html_sanitize are likely a good idea to be used in sort_by.
Tip: debug_files can help you understand what is happening. But don't forget to remove it after you are done testing:
editors:
- css_sort:
selector: article
sort_by:
- css_select: h2
- html_sanitize
- debug_files: /tmp/website-stalker/You can also reverse the sorting:
editors:
- css_sort:
selector: article
reverse: trueReplace HTML tags matching a given CSS Selector.
For example, the following config will replace all h3 tags with h2 tags.
editors:
- css_tag_replace:
selector: h3
replace: h2 <html>
<head></head>
<body>
- <h3 class="green">
+ <h2 class="green">
Hello
- </h3>
+ </h2>
World
</body>
</html>This can be helpful to ensure some kind of structure especially when editors like html_markdownify are used.
Think about a website where only some subsections are of interest and selected via the css_select.
While the header contains some h1 the selected part skips the h2 headings and continues with h3 headings.
Also, <strong> are (incorrectly) used as subheadings.
Parsing this to Markdown results in less optimal structure as h2 are skipped and <strong> don't result in headings.
Ideal would be a single h1 and then continuous depending on the depth h2, h3 and so on.
The following can help with that:
editors:
# Select the header and some interesting sections
- css_select: header, main section.interesting
# First migrate the h3 tags to h2 so there is no gap
- css_tag_replace:
selector: h3
replace: h2
# Then migrate all strong tags to proper headings
- css_tag_replace:
selector: strong
replace: h3
# In the end parse to Markdown
- html_markdownifyThis editor passes its input through without modifying it. The content is written to a file in the given directory. The filename is created from the current UNIX Timestamp.
This is neat when looking at steps in between editors is of interest. Especially for editors like RSS which use editors per item this can be handy to look at the steps in between.
Warning: It's not recommended committing these files.
debug_files should be removed before when committing the config.
It might have unintended side effects or might spam your repository with many potentially large files.
Examples:
editors:
- debug_files: /tmp/website-stalker/Formats the input HTML as Markdown.
Example:
editors:
- html_markdownifyFormats the input HTML as pretty HTML.
Example:
editors:
- html_prettifyStrip down HTML to its minimal form.
Example:
editors:
- html_sanitizeOnly returns text content of HTML elements within the input.
Example:
editors:
- html_textifyParses the input HTML for URLs. URLs are parsed into their canonical, absolute form.
Example:
editors:
- html_url_canonicalizeFormats the input JSON as pretty JSON.
Example:
editors:
- json_prettifySelects from an input JSON to narrow it down.
Inspired by jq but with only very basic selection support.
Due to the current, naive selector implementation it might support syntax jq doesnt support.
Syntax that works with this editor and jq is intended behaviour.
Syntax that only works with this editor and not jq is considered a bug and might change its behaviour on any release.
This editor is intentionally kept simple and does not intend to support the whole jq featureset.
But it does not mean it should stay in its current state.
If you think it should support other simple syntax of jq, feel free to create an issue or pull request and then lets discuss about it.
Example:
editors:
- json_simple_select: .foo[2].barSearches the input with a Regex pattern and replaces all occurrences with the given replace phrase.
Grouping and replacing with $1 also works.
Examples:
editors:
# Remove all occurrences of that word
- regex_replace:
pattern: "tree"
replace: ""
# Remove all numbers
- regex_replace:
pattern: "\\d+"
replace: ""
# Find all css files and remove the extension
- regex_replace:
pattern: "(\\w+)\\.css"
replace: $1Creates an RSS 2.0 Feed from the input.
An RSS item is generated for every item_selector result.
The other selectors can be used to find relevant information of the items.
The content is the full result of the item_selector.
It can be further edited with every available editor.
Defaults:
title: When a<title>exists, it will be used. Otherwise, it's empty.item_selector:articletitle_selector:h2link_selector:acontent_editorscan be omitted when empty
Examples:
# Fully specified example
- url: "https://edjopato.de/post/"
editors:
- rss:
title: EdJoPaTos Blog
item_selector: article
title_selector: h2
link_selector: a
content_editors:
- css_remove: "h2, article > a, div"
- html_textify
# Minimal working example
- url: "https://edjopato.de/post/"
editors:
- rss: {}- Website Changed Bot is a Telegram Bot which might potentially use this tool later on
- bernaferrari/ChangeDetection is an Android app for this
- dgtlmoon/changedetection.io can be self-hosted and configured via web interface
- Feed me up, Scotty! creates RSS feeds from websites
- htmlq command line tool to format / select HTML (like
jqfor HTML) - urlwatch