Skip to content

extracts PDF links from websites and dumps them to the stdout and as a textfile. only works for links pointing to files with the ".pdf" extension

License

Notifications You must be signed in to change notification settings

moiSentineL/pdflinkextractor

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 

Repository files navigation

pdflinkextractor

Originally made by Glutanimate.

This script finds any pdf files hyperlinked in a webpage and provides output of the links or download them.

Usage

Bash One Liner

curl -fsSL https://nibirsan.org/pdflinkextractor/script.sh | sh -s - [-d] <website> 

Using -d will enable you to download the files instead of just saving the links.

To save the links to a file, just do > file at the end of the command.

Local

git clone https://github.com/moiSentineL/pdflinkextractor.git
cd pdflinkextractor && chmod +x script.sh
./script.sh [-d] <website>

Tip

Alias it for better access.

Dependencies

You have to have wget and lynx installed:

Ubuntu

sudo apt-get install wget lynx

Arch Linux

sudo pacman -S wget lynx

About

extracts PDF links from websites and dumps them to the stdout and as a textfile. only works for links pointing to files with the ".pdf" extension

Topics

Resources

License

Stars

Watchers

Forks

Languages