I have a small project at home, where I need to scrape a website for links every once in a while and save the links in a txt file.
The script need to run on my Synology NAS, therefore the script needs to be written in bash script or python without using any plugins or external libraries as I can't install it on the NAS. (to my knowledge anyhow)
A link looks like this:
<a href="http://www.example.com">Example text</a>
I want to save the following to my text file:
Example text - http://www.example.com
I was thinking I could isolate the text with curl and some grep (or perhaps regex). First I looked into using Scrapy or Beutifulsoup, but couldn't find a way to install it on the NAS.
Could one of you help me put a script together?