You cannot select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
dot-files/dot_local/bin/executable_get-amazon-art.tmpl

175 lines
5.1 KiB
Python

# chezmoi:template:left-delimiter=#{{
#{{- /* vim: set filetype=python: */ -}}
#{{ if .include_legacy -}}
#!/usr/bin/env python
"""
This is a program that attempts to get all the album art from amazon
as is possible.
"""
import os
import sys
import urllib
import urllib2
from BeautifulSoup import BeautifulSoup
def walklevel(some_dir, level=1):
some_dir = some_dir.rstrip(os.path.sep)
assert os.path.isdir(some_dir)
num_sep = some_dir.count(os.path.sep)
for root, dirs, files in os.walk(some_dir):
yield root, dirs, files
num_sep_this = root.count(os.path.sep)
if num_sep + level <= num_sep_this:
del dirs[:]
def get_search_dirs(base = '.'):
ret = []
for root, dirs, files in walklevel(base, 0):
for the_dir in dirs:
new_path = "%s/%s" %(root, the_dir)
for root2, dirs2, files2 in walklevel(new_path, 0):
for found_dir in dirs2:
ret.append("%s/%s" % (new_path, found_dir))
return ret
def get_term_from_path(the_dir):
"""Gets a search term from the directory"""
parts = the_dir.split('/')
return {
'artist': parts[-2],
'album': parts[-1]
}
def get_url_amazon(term):
"""Returns the amazon search url get_url"""
params = {
'url': 'search-alias=digital-music',
'x': '0',
'y': '0',
'field-keywords': term
}
return "http://www.amazon.com/s/ref=nb_sb_noss?%s" % urllib.urlencode(params)
def get_url(term):
"""Returns the amazon search url get_url"""
params = {
'mode': 'b',
'QT': term
}
return "http://www.emusic.com/search.html?%s" % urllib.urlencode(params)
def get_album_img_src_amazon(url):
"""parses the url and find the link for the album"""
html = urllib2.urlopen(url)
soup = BeautifulSoup(html)
imgSrc = soup.find('a', {'id': 'mp3StoreShovelerShvlLink0'}).img['src']
imgParts = imgSrc.split('_');
imgParts.pop(-2)
return '_'.join(imgParts)
def get_album_img_srcs_emusic(url):
"""parses the url and find the link for the album"""
html = urllib2.urlopen(url)
soup = BeautifulSoup(html)
ul = soup.find('ul', {'class': 'resultList'})
li = ul.findAll('li')[0]
img = li.find('img')
img_src = img['src']
img_parts = img_src.split('/')
img_parts[-1] = '600x600.jpg'
large_img = '/'.join(img_parts)
img_parts[-1] = '1400x1400.jpg'
huge_img = '/'.join(img_parts)
return {
'large': large_img,
'huge': huge_img
}
def get_file_name(info, size):
"""docstring for get_file_name"""
return 'imgs/%s.%s.%d.jpg' % (info['artist'], info['album'], size)
def save_file(in_stream, size, info):
"""docstring for save_file"""
file_name = get_file_name(info, size)
output = open(file_name, 'wb')
output.write(in_stream.read())
output.close()
def main():
"""The main script."""
# Get the path for the directory
try:
base_dir = sys.argv[1]
except IndexError:
print "Must enter path of directory to search. Example: %s ./foo" % \
sys.argv[0]
os.abort()
if os.path.exists(base_dir) == False:
print "Could not find path: '%s'" % base_dir
os.abort()
# Remove trailing /
if base_dir[-1] == "/":
base_dir = base_dir[:-1]
search_dirs = get_search_dirs(base_dir)
for the_dir in search_dirs:
search_terms = get_term_from_path(the_dir)
found_file = False
#for tmp_size in (1400, 600, 500):
for tmp_size in (500, ):
tmp_file_name = get_file_name(search_terms, tmp_size)
if os.path.exists(tmp_file_name):
print "Skipping check, found '%s'" % tmp_file_name
found_file = True
break
if found_file:
continue
#img_src = None
#url = get_url(search_terms['album'])
#try:
# img_src = get_album_img_srcs_emusic(url)
#except:
# img_src = None
#if img_src != None:
# img_file = urllib2.urlopen(img_src['huge'])
# if img_file.info().getmaintype() == 'image':
# save_file(img_file, 1400, search_terms)
# print "Saved huge emusic cover for %s" % search_terms['album']
# continue
# img_file = urllib2.urlopen(img_src['large'])
# if img_file.info().getmaintype() == 'image':
# save_file(img_file, 600, search_terms)
# print "Saved large emusic cover for %s" % search_terms['album']
# continue
# If emusic doesn't have it, try amazon
url = get_url_amazon("%s %s" % (search_terms['artist'], search_terms['album']))
try:
img_src = get_album_img_src_amazon(url)
img_file = urllib2.urlopen(img_src)
except:
print "Skipped, could not find cover art for %s" % search_terms['album']
continue
save_file(img_file, 500, search_terms)
print "Saved large amazon cover for %s" % search_terms['album']
if __name__ == "__main__":
main()
#{{- end }}