forked from floodsung/Deep-Learning-Papers-Reading-Roadmap
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge pull request floodsung#17 from mjanv/scripting
Add script for automatic papers download
- Loading branch information
Showing
2 changed files
with
108 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,106 @@ | ||
import os | ||
import re | ||
import urllib2 | ||
import shutil | ||
import argparse | ||
import mistune | ||
import bs4 as BeautifulSoup | ||
|
||
def download_pdf(link, location, name): | ||
try: | ||
response = urllib2.urlopen(link) | ||
file = open(os.path.join(location, name), 'w') | ||
file.write(response.read()) | ||
file.close() | ||
except urllib2.HTTPError: | ||
print('>>> Error 404: cannot be downloaded!\n') | ||
raise | ||
|
||
def clean_pdf_link(link): | ||
if 'arxiv' in link: | ||
link = link.replace('abs', 'pdf') | ||
if not(link.endswith('.pdf')): | ||
link = '.'.join((link, 'pdf')) | ||
if 'github' in link: | ||
link = '.'.join((link, 'html')) | ||
return link | ||
|
||
def clean_text(text, replacements = {' ': '_', '/': '_', '.': '', '"': ''}): | ||
for key, rep in replacements.items(): | ||
text = text.replace(key, rep) | ||
return text | ||
|
||
def print_title(title, pattern = "-"): | ||
print('\n'.join(("", title, pattern * len(title)))) | ||
|
||
def get_extension(link): | ||
extension = os.path.splitext(link)[1][1:] | ||
if extension in ['pdf', 'html']: | ||
return extension | ||
if 'pdf' in extension: | ||
return 'pdf' | ||
return 'pdf' | ||
|
||
def shorten_title(title): | ||
m1 = re.search('[[0-9]*]', title) | ||
m2 = re.search('".*"', title) | ||
if m1: | ||
title = m1.group(0) | ||
if m2: | ||
title = ' '.join((title, m2.group(0))) | ||
return title[:50] + ' [...]' | ||
|
||
|
||
if __name__ == '__main__': | ||
|
||
parser = argparse.ArgumentParser(description = 'Download all the PDF/HTML links into README.md') | ||
parser.add_argument('-d', action="store", dest="directory") | ||
parser.add_argument('--no-html', action="store_true", dest="nohtml", default = False) | ||
parser.add_argument('--overwrite', action="store_true", default = False) | ||
results = parser.parse_args() | ||
|
||
output_directory = 'pdfs' if results.directory is None else results.directory | ||
|
||
forbidden_extensions = ['html', 'htm'] if results.nohtml else [] | ||
|
||
if results.overwrite and os.path.exists(output_directory): | ||
shutil.rmtree(output_directory) | ||
|
||
with open('README.md') as readme: | ||
readme_html = mistune.markdown(readme.read()) | ||
readme_soup = BeautifulSoup.BeautifulSoup(readme_html, "html.parser") | ||
|
||
point = readme_soup.find_all('h1')[1] | ||
|
||
failures = [] | ||
while point is not None: | ||
if point.name: | ||
if re.search('h[1-2]', point.name): | ||
if point.name == 'h1': | ||
h1_directory = os.path.join(output_directory, clean_text(point.text)) | ||
current_directory = h1_directory | ||
elif point.name == 'h2': | ||
current_directory = os.path.join(h1_directory, clean_text(point.text)) | ||
os.makedirs(current_directory) | ||
print_title(point.text) | ||
|
||
if point.name == 'p': | ||
link = point.find('a') | ||
if link is not None: | ||
link = clean_pdf_link(link.attrs['href']) | ||
ext = get_extension(link) | ||
if not ext in forbidden_extensions: | ||
print(shorten_title(point.text) + ' (' + link + ')') | ||
try: | ||
name = clean_text(point.text.split('[' + ext + ']')[0]) | ||
download_pdf(link, current_directory, '.'.join((name, ext))) | ||
except: | ||
failures.append(point.text) | ||
|
||
point = point.next_sibling | ||
|
||
print('Done!') | ||
if failures: | ||
print('Some downloads have failed:') | ||
for fail in failures: | ||
print('> ' + fail) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,2 @@ | ||
mistune>=0.7.2 | ||
beautifulsoup4>=4.4.1 |