How to use wget to download all links from the site and save to a text file

wget

I am trying to download all links from aligajani.com. There are 7 of them, excluding the domain facebook.com–which I want to ignore. I don't want to download from links that start with facebook.com domain.

Also, I want them saved in a .txt file, line by line. So there would be 7 lines.

Here's what I've tried so far. This just downloads everything. Don't want that.

wget -r -l 1 http://aligajani.com

Best Answer

wget does not offer such an option. Please read its man page.

You could use lynx for this:

lynx -dump -listonly http://aligajani.com | grep -v facebook.com > file.txt

From its man page:

   -listonly
          for -dump, show only the list of links.
Related Question