diff options
Diffstat (limited to 'tirante/tirante.py')
-rw-r--r-- | tirante/tirante.py | 105 |
1 files changed, 16 insertions, 89 deletions
diff --git a/tirante/tirante.py b/tirante/tirante.py index b29ef5b..c179475 100644 --- a/tirante/tirante.py +++ b/tirante/tirante.py @@ -25,85 +25,12 @@ import urllib3 from bs4 import BeautifulSoup import requests -# Main manga source. -MAIN_URL = 'https://manganelo.com/manga/' -# Manga name. -MANGA_NAME = 'Kimetsu no Yaiba' -# Manga name in the form of how appears in the url. -MANGA_NAME_URL = 'kimetsu_no_yaiba/' - -# PC main file location. -MANGA_DIR = 'E:\\Mangas\\' -# PC main manga data location. -MANGA_DATA_DIR = ''.join(['C:\\Users\\Lorentzeus\\Google Drive\\', - 'Personal\\Python\\mangas\\test_data']) - - -def get_chapters_list(main_url=MAIN_URL, - manga_name_url=MANGA_NAME_URL, - manga_name=MANGA_NAME, - reverse_sorted=True): - """ - Retrieves chapter urls and names. Returns a list of lists - containing the url and the title of the chapter. - main_url: Main webpage name (source). - manga_name_url: Name of the manga in the url format - that's used by the webpage. - manga_name: Actual name of the manga, as it appears in the webpage. - reverse_sorted: Sorting of the final array. - """ - - manga_url = ''.join([MAIN_URL, MANGA_NAME_URL]) - - # Not actually a file, but the content of the html. - html = urllib3.PoolManager().request('GET', manga_url) - - # Get the data from the html and parse it. - soup = BeautifulSoup(html.data, 'html.parser') - - # Get the "rows" class, this contains the url - # and title data for each chapter. - # Deletes the first tag, since it's not useful. - soup_rows = soup.find_all('div', {'class': 'row'}) - del soup_rows[0] - - # Creates a list to store date for each url and chapter name. - chapter_list = [] - - for row in soup_rows: - - # Gets the url name from the a tag. - href = row.a['href'] - # Same, for the title. Deletes every ocurrance of the manga name, - # unwanted characters and then gets everyword. - title_words = row.a['title'].replace(manga_name, '').replace('?', '') - title_words = title_words.replace(':', '').replace('-', '') - title_words = title_words.replace('...', '').replace(',', '').split() - - # Doing all the work in oneliner doesn't work for some chapters, - # for some reason. - # title = '_'.join(row.a['title'].replace(manga_name, '') - # .replace(':', '').replace('-', '').lower().split()) - - # Lowers every word and appends it to a new list, - # then it gets joined with '_' as a sep. - title_words_lower = [] - for word in title_words: - title_words_lower.append(word.lower()) - - title = '_'.join(title_words_lower) - - # print(href, title) - chapter_list.append([href, title]) - - if reverse_sorted: - return chapter_list[::-1] - else: - return chapter_list +# Project specific imports. +from gcl import get_chapters_list def chapters_list_to_csv(chapters_list, - manga_name=MANGA_NAME): + manga_name): """ Creates a csv file from the input chapter_list. chapters_list: List of data of the chapters. @@ -200,11 +127,11 @@ def chapter_image_csv_to_list(chapter_image_csv): return out_chapter_image_list -def create_database(main_url=MAIN_URL, - manga_name_url=MANGA_NAME_URL, - manga_name=MANGA_NAME, - manga_dir=MANGA_DIR, - manga_data_dir=MANGA_DATA_DIR): +def create_database(main_url, + manga_name_url, + manga_name, + manga_dir, + manga_data_dir): """ Creates a database from zero, made of csv files. main_url: Main webpage name (source). @@ -258,11 +185,11 @@ def create_database(main_url=MAIN_URL, print(''.join([chapter_name_ext, ' already exists.'])) -def update_database(main_url=MAIN_URL, - manga_name_url=MANGA_NAME_URL, - manga_name=MANGA_NAME, - manga_dir=MANGA_DIR, - manga_data_dir=MANGA_DATA_DIR): +def update_database(main_url, + manga_name_url, + manga_name, + manga_dir, + manga_data_dir): """ Updates the database already created, adding missing ones. main_url: Main webpage name (source). @@ -357,9 +284,9 @@ def download_chapter(image_list): download_image(image) -def download_manga(manga_name=MANGA_NAME, - manga_dir=MANGA_DIR, - manga_data_dir=MANGA_DATA_DIR): +def download_manga(manga_name, + manga_dir, + manga_data_dir): """ Downloads a whole manga, saving it to subfolders. Uses the database already created. |