mirror of
https://github.com/cooperhammond/irs.git
synced 2024-12-22 17:35:28 +00:00
297 lines
10 KiB
Python
297 lines
10 KiB
Python
import os, sys, time, re, select, requests, subprocess
|
|
import urllib.request, urllib.parse
|
|
from termcolor import colored
|
|
from urllib.request import Request, urlopen
|
|
|
|
from mutagen.id3 import APIC
|
|
from mutagen.easyid3 import EasyID3
|
|
from mutagen.mp3 import MP3
|
|
from bs4 import BeautifulSoup
|
|
import youtube_dl, mutagen
|
|
|
|
"""def test_system():
|
|
try:
|
|
# Put import stuff here
|
|
except Exception as e:
|
|
e = str(e).split(" ")[-1]
|
|
print ("Please install the module %s with:\npip3 install %s")
|
|
exit(0)
|
|
if subprocess.check_output("mpv --version", shell=True) != 0:
|
|
print ("mpv is not installed, please install it with:\nsudo apt-get install mpv")
|
|
if subprocess.check_output("peerflix --version", shell=True) != 0:
|
|
print ("peerflix is not installed, install it with:\nsudo apt-get install peerflix")
|
|
if "Rakshasa" in subprocess.check_output("rtorrent -h", shell=True):
|
|
print ("rtorrent is not installed, install it with:\nsudo apt-get install rtorrent")
|
|
|
|
def quiet_output(string):
|
|
output = subprocess.check_output(string, shell=True)
|
|
"""
|
|
|
|
def download_album_art(album, band):
|
|
search = "%s %s" % (album, band)
|
|
url = "http://www.seekacover.com/cd/" + urllib.parse.quote_plus(search)
|
|
page = requests.get(url).text
|
|
soup = BeautifulSoup(page)
|
|
done = False
|
|
for img in soup.findAll('img'):
|
|
if done == False:
|
|
try:
|
|
if search.lower() in img['title'].lower():
|
|
url = img['src']
|
|
urllib.request.urlretrieve(url, "cover.jpg")
|
|
done = True
|
|
except Exception:
|
|
pass
|
|
|
|
def embed_mp3(art_location, song_path):
|
|
music = mutagen.id3.ID3(song_path)
|
|
|
|
music.delall('APIC')
|
|
|
|
music.add(APIC(
|
|
encoding=0,
|
|
mime="image/jpg",
|
|
type=3,
|
|
desc='',
|
|
data=open(art_location, "rb").read()
|
|
)
|
|
)
|
|
music.save()
|
|
|
|
def find_mp3(song, author):
|
|
print ("'%s' by '%s'\n" % (song, author))
|
|
query_string = urllib.parse.quote_plus({"search_query" : ("%s %s lyrics" % (song, author))})
|
|
html_content = urllib.request.urlopen("http://www.youtube.com/results?" + query_string)
|
|
search_results = re.findall(r'href=\"\/watch\?v=(.{11})', html_content.read().decode())
|
|
audio_url = ("http://www.youtube.com/watch?v=" + search_results[0])
|
|
return audio_url
|
|
|
|
def rip_mp3(song, author, album, tracknum):
|
|
audio_url = find_mp3(song, author)
|
|
command = 'youtube-dl --metadata-from-title "%(title)s" --extract-audio --audio-format mp3 --add-metadata ' + audio_url
|
|
os.system(command)
|
|
for filename in os.listdir("."):
|
|
if str(audio_url).strip("http://www.youtube.com/watch?v=") in filename:
|
|
os.rename(filename, song + ".mp3")
|
|
|
|
try:
|
|
googled = search_google(song, author)
|
|
mp3file = MP3(song + ".mp3", ID3=EasyID3)
|
|
print ("\n%s Metadata parsing:" % output("s"))
|
|
try:
|
|
mp3file['title'] = song
|
|
except Exception:
|
|
mp3file['title'] = ""
|
|
mp3file.save()
|
|
print ('\t%s Title parsed: %s' % (output("g"), mp3file['title']))
|
|
|
|
mp3file['artist'] = author
|
|
mp3file.save()
|
|
print ('\t%s Author parsed: %s' % (output("g"), mp3file['artist']))
|
|
if album == "":
|
|
for i, j in enumerate(googled):
|
|
if "Album:" in j:
|
|
album = (googled[i + 1])
|
|
try:
|
|
mp3file['album'] = album
|
|
mp3file.save()
|
|
except Exception:
|
|
mp3file['album'] = ""
|
|
mp3file.save()
|
|
print ('\t%s Album Auto-parsed: %s' % (output("g"), mp3file['album']))
|
|
else:
|
|
try:
|
|
mp3file['album'] = album
|
|
mp3file.save()
|
|
except Exception:
|
|
mp3file['album'] = ""
|
|
mp3file.save()
|
|
print ('\t%s Album parsed: %s' % (output("g"), mp3file['album']))
|
|
|
|
if mp3file['album'][0] != "":
|
|
try:
|
|
download_album_art(mp3file['album'][0], author)
|
|
embed_mp3("cover.jpg", song + ".mp3")
|
|
print ("\t%s Album art Auto-parsed!" % output("g"))
|
|
except Exception as e:
|
|
print ("%s There was a problem with Auto-parsing the album art of: '%s'" % (output("e"), song))
|
|
print (e)
|
|
|
|
for i, j in enumerate(googled):
|
|
if "Released:" in j:
|
|
date = (googled[i + 1])
|
|
try:
|
|
mp3file['date'] = date
|
|
except Exception:
|
|
mp3file['date'] = ""
|
|
mp3file.save()
|
|
print ('\t%s Release date Auto-parsed: "%s"' % (output("g"), mp3file['date'][0]))
|
|
|
|
if tracknum != "":
|
|
mp3file['tracknumber'] = str(tracknum)
|
|
mp3file.save()
|
|
|
|
print ('\n%s "' % output("g") + song + '" by ' + author + ' downloaded successfully\n')
|
|
except Exception as e:
|
|
print (e)
|
|
|
|
def output(string):
|
|
if string == "q":
|
|
return colored("[?]", "magenta")
|
|
elif string == "e":
|
|
return colored("[-]", "red")
|
|
elif string == "g":
|
|
return colored("[+]", "green")
|
|
elif string == "s":
|
|
return colored("[*]", "yellow")
|
|
|
|
def visible(element):
|
|
if element.parent.name in ['style', 'script', '[document]', 'head', 'title']:
|
|
return False
|
|
elif re.match('<!--.*-->', str(element)):
|
|
return False
|
|
return True
|
|
|
|
def search_google(song_name, band):
|
|
# takes 2 strings, one as the song_name and the other as the band's name, with the keyword as what prefaces the info you want.
|
|
try:
|
|
string = "%s %s" % (song_name, band)
|
|
filename = 'http://www.google.com/search?q=' + urllib.parse.quote_plus(string)
|
|
hdr = {
|
|
'User-Agent':'Mozilla/5.0', # Honestly, this is just so that google doesn't prevent your from looking at them.
|
|
'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8', # creds are here.
|
|
}
|
|
texts = BeautifulSoup(urlopen(Request(filename, headers=hdr)).read(), 'html.parser').findAll(text=True)
|
|
|
|
visible_texts = list(filter(visible, texts))
|
|
return visible_texts
|
|
except Exception as e:
|
|
print ("%s There was an error with Auto-parsing." % output("e"))
|
|
return ""
|
|
|
|
def get_album(album_name, artist, what_to_do):
|
|
visible_texts = search_google(album_name, artist)
|
|
songs = []
|
|
num = True
|
|
for i, j in enumerate(visible_texts):
|
|
if 'Songs' in j:
|
|
if visible_texts[i + 1] == "1":
|
|
indexed = i
|
|
while num == True:
|
|
try:
|
|
if type(int(visible_texts[indexed])) is int:
|
|
a = visible_texts[indexed + 1]
|
|
songs.append(a)
|
|
indexed += 1
|
|
except:
|
|
indexed += 1
|
|
if indexed >= 1000:
|
|
num = False
|
|
else:
|
|
pass
|
|
if what_to_do == "download":
|
|
for i, j in enumerate(songs):
|
|
rip_mp3(j, artist, album_name, i + 1)
|
|
elif what_to_do == "stream":
|
|
for i in songs:
|
|
a = find_mp3(i, artist)
|
|
command = 'mpv "%s" --no-video' % a
|
|
os.system(command)
|
|
|
|
|
|
def get_torrent_url(search_url):
|
|
search_request_response = requests.get(search_url, verify=False)
|
|
soup = BeautifulSoup(search_request_response.text, 'html.parser')
|
|
movie_page = 'https://kat.cr' + (soup.find_all("a", class_="cellMainLink")[0].get('href'))
|
|
|
|
search_url = requests.get(movie_page, verify=False)
|
|
soup = BeautifulSoup(search_url.text, 'html.parser')
|
|
torrent_url = 'https:' + soup.find_all('a', class_='siteButton')[0].get('href')
|
|
return torrent_url
|
|
|
|
def main():
|
|
#test_system()
|
|
#try:
|
|
i = 0
|
|
args = sys.argv
|
|
del args[i]
|
|
what_to_do = args[i]
|
|
del args[i]
|
|
|
|
if what_to_do not in ("download", "stream"): raise Exception
|
|
|
|
media = args[i]
|
|
del args[i]
|
|
|
|
if media == "song":
|
|
song = (" ".join(args)).split(" by ")
|
|
if what_to_do == "stream":
|
|
command = 'mpv "%s" --no-video' % find_mp3(song[0], song[1])
|
|
os.system(command)
|
|
elif what_to_do == "download":
|
|
rip_mp3(song[0], song[1], "", "")
|
|
|
|
elif media == "album":
|
|
album_name = (" ".join(args)).split(" by ")
|
|
get_album(album_name[0], album_name[1], what_to_do)
|
|
|
|
elif media == "movie":
|
|
query = 'https://kat.cr/usearch/' + urllib.parse.quote_plus((" ".join(args) + " category:movies"))
|
|
if what_to_do == "stream":
|
|
torrent_url = get_torrent_url(query)
|
|
os.system('peerflix "%s" -a --vlc' % torrent_url)
|
|
exit(0)
|
|
elif what_to_do == "download":
|
|
os.system("rtorrent '%s'" % get_torrent_url(query))
|
|
exit(0)
|
|
|
|
elif media == "tv":
|
|
query = 'https://kat.cr/usearch/' + urllib.parse.quote_plus((" ".join(args) + " category:tv"))
|
|
if what_to_do == "stream":
|
|
torrent_url = get_torrent_url(query)
|
|
os.system('peerflix "%s" -a --vlc' % torrent_url)
|
|
exit(0)
|
|
elif what_to_do == "download":
|
|
os.system("rtorrent '%s'" % get_torrent_url(query))
|
|
exit(0)
|
|
else:
|
|
raise Exception
|
|
|
|
"""except Exception as e:
|
|
print (e)
|
|
invalid_format()"""
|
|
|
|
def columns(columns):
|
|
for row in columns:
|
|
print("{: >15} {: >15}".format(*row))
|
|
def invalid_format():
|
|
# I feel like there should be an easier way to write out help for command-line interfaces ...
|
|
print ("Usage:")
|
|
print (" UPS [what-to-do] [type-of-media] <movie-name>")
|
|
print (" UPS [what-to-do] [type-of-media] <tv-show> [episode]")
|
|
print (" UPS [what-to-do] [type-of-media] <album> by <artist>")
|
|
print (" UPS [what-to-do] [type-of-media] <song> by <artist>\n")
|
|
print ("what-to-do:")
|
|
d_or_s = [["download", "will download the specified media"],
|
|
["stream", "will stream the specified media"]]
|
|
columns(d_or_s)
|
|
print ("type-of-media:")
|
|
media = [["song", "finds a song, needs the artist as well"],
|
|
["album", "finds all songs on the album, needs the artist as well"],
|
|
["movie", "will find a movie"],
|
|
["tv", "will find a tv show, needs to have the episode specified"]]
|
|
columns(media)
|
|
print ("tv show episode should be formatted as:")
|
|
episode = [["s01e01", "s12e01"], ["s03e05", "s05e17"], ["... ", "... "]]
|
|
columns(episode)
|
|
print ("examples:")
|
|
examples = [["$ UPS stream movie Fight Club"],
|
|
["$ UPS download album Jazz by Queen"],
|
|
["$ UPS stream song Where Is My Mind by The Pixies"],
|
|
["$ UPS download tv mr robot s01e01"]]
|
|
for row in examples:
|
|
print("\t {: >15}".format(*row))
|
|
|
|
if __name__ == '__main__':
|
|
main()
|