Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- #!/bin/sh
- # Download URLs from file (called queue) with history.
- # License: LGPLv2
- # I schedule this to run every night.
- # The file called queue must be in the same dir
- # as this script and lines starting with # are ignored.
- cp queue queue.tmp
- sed "/ *#/d" queue |
- while read url; do
- echo "#$url" >> queue.tmp
- wget -nv "$url"
- done
- sed -n "/ *#/p" queue.tmp > queue
- rm -f queue.tmp
Add Comment
Please, Sign In to add comment