mirror of
https://github.com/KevinMidboe/bulk-downloader-for-reddit.git
synced 2026-01-11 03:35:36 +00:00
Compare commits
89 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
280147282b | ||
|
|
b7baf07fb5 | ||
|
|
aece2273fb | ||
|
|
f807efe4d5 | ||
|
|
743d887927 | ||
|
|
da5492858c | ||
|
|
cebfc713d2 | ||
|
|
f522154214 | ||
|
|
27cd3ee991 | ||
|
|
29873331e6 | ||
|
|
8a3dcd68a3 | ||
|
|
ac323f2abe | ||
|
|
32d26fa956 | ||
|
|
137481cf3e | ||
|
|
9b63c55d3e | ||
|
|
3a6954c7d3 | ||
|
|
9a59da0c5f | ||
|
|
d56efed1c6 | ||
|
|
8f64e62293 | ||
|
|
bdc43eb0d8 | ||
|
|
adccd8f3ba | ||
|
|
47a07be1c8 | ||
|
|
1a41dc6061 | ||
|
|
50cb7c15b9 | ||
|
|
a1f1915d57 | ||
|
|
3448ba15a9 | ||
|
|
ff68b5f70f | ||
|
|
588a3c3ea6 | ||
|
|
8f1ff10a5e | ||
|
|
9338961b2b | ||
|
|
94bc1c115f | ||
|
|
c19d8ad71b | ||
|
|
4c8de50880 | ||
|
|
3e6dfccdd2 | ||
|
|
20b9747330 | ||
|
|
be7508540d | ||
|
|
ccd9078b0a | ||
|
|
43cf0a4d42 | ||
|
|
3693cf46f8 | ||
|
|
04152e8554 | ||
|
|
210238d086 | ||
|
|
90e071354f | ||
|
|
426089d0f3 | ||
|
|
7ae6c6385d | ||
|
|
97d15f9974 | ||
|
|
172cd72dc1 | ||
|
|
af29492951 | ||
|
|
5633b301f3 | ||
|
|
5ed855af28 | ||
|
|
7ccf2fb7f9 | ||
|
|
2297e9ed86 | ||
|
|
401e014059 | ||
|
|
eb31d38c44 | ||
|
|
747fefea14 | ||
|
|
80cc4fade3 | ||
|
|
c26843c7fc | ||
|
|
a14edc9f5a | ||
|
|
d685860c22 | ||
|
|
dcf9f35273 | ||
|
|
7fdf03aa24 | ||
|
|
25d61a4c78 | ||
|
|
558eb107f4 | ||
|
|
6e74630050 | ||
|
|
2fd9248715 | ||
|
|
457b8cd21c | ||
|
|
e953456ead | ||
|
|
ed0564fba0 | ||
|
|
5378555f74 | ||
|
|
95ef308915 | ||
|
|
436f867f2e | ||
|
|
91d71565cc | ||
|
|
c7b7361ded | ||
|
|
cd81a6c38b | ||
|
|
1623722138 | ||
|
|
dad5669441 | ||
|
|
35d54d1eb1 | ||
|
|
394b864d86 | ||
|
|
837281c3c6 | ||
|
|
e6b648d8b3 | ||
|
|
cfaf2de7db | ||
|
|
80546d7094 | ||
|
|
139a81a0e7 | ||
|
|
9bb0a5da7f | ||
|
|
6f2273f182 | ||
|
|
b5d6165802 | ||
|
|
b98815376f | ||
|
|
d9586f99b8 | ||
|
|
76711892a2 | ||
|
|
bfea548eab |
3
.gitignore
vendored
3
.gitignore
vendored
@@ -2,4 +2,5 @@ build/
|
|||||||
dist/
|
dist/
|
||||||
MANIFEST
|
MANIFEST
|
||||||
__pycache__/
|
__pycache__/
|
||||||
src/__pycache__/
|
src/__pycache__/
|
||||||
|
config.json
|
||||||
91
README.md
91
README.md
@@ -1,7 +1,7 @@
|
|||||||
# Bulk Downloader for Reddit
|
# Bulk Downloader for Reddit
|
||||||
This program downloads imgur, gfycat and direct image and video links of saved posts from a reddit account. It is written in Python 3.
|
Downloads media from reddit posts.
|
||||||
|
|
||||||
**PLEASE** post any issue you have with the script to [Issues](https://github.com/aliparlakci/bulk-downloader-for-reddit/issues) tab. Since I don't have any testers or contributers I need your feedback.
|
## [Download the latest release](https://github.com/aliparlakci/bulk-downloader-for-reddit/releases/latest)
|
||||||
|
|
||||||
## What it can do
|
## What it can do
|
||||||
- Can get posts from: frontpage, subreddits, multireddits, redditor's submissions, upvoted and saved posts; search results or just plain reddit links
|
- Can get posts from: frontpage, subreddits, multireddits, redditor's submissions, upvoted and saved posts; search results or just plain reddit links
|
||||||
@@ -13,86 +13,17 @@ This program downloads imgur, gfycat and direct image and video links of saved p
|
|||||||
- Saves a reusable copy of posts' details that are found so that they can be re-downloaded again
|
- Saves a reusable copy of posts' details that are found so that they can be re-downloaded again
|
||||||
- Logs failed ones in a file to so that you can try to download them later
|
- Logs failed ones in a file to so that you can try to download them later
|
||||||
|
|
||||||
## [Download the latest release](https://github.com/aliparlakci/bulk-downloader-for-reddit/releases/latest)
|
## **[Compiling it from source code](docs/COMPILE_FROM_SOURCE.md)**
|
||||||
|
*\* MacOS users have to use this option.*
|
||||||
|
|
||||||
## How it works
|
## Additional options
|
||||||
|
Script also accepts additional options via command-line arguments. Get further information from **[`--help`](docs/COMMAND_LINE_ARGUMENTS.md)**
|
||||||
- For **Windows** and **Linux** users, there are executable files to run easily without installing a third party program. But if you are a paranoid like me, you can **[compile it from source code](docs/COMPILE_FROM_SOURCE.md)**.
|
|
||||||
- In Windows, double click on bulk-downloader-for-reddit file
|
|
||||||
- In Linux, extract files to a folder and open terminal inside it. Type **`./bulk-downloader-for-reddit`**
|
|
||||||
|
|
||||||
- **MacOS** users have to **[compile it from source code](docs/COMPILE_FROM_SOURCE.md)**.
|
|
||||||
|
|
||||||
Script also accepts **command-line arguments**, get further information from **[`--help`](docs/COMMAND_LINE_ARGUMENTS.md)**
|
|
||||||
|
|
||||||
## Setting up the script
|
## Setting up the script
|
||||||
Because this is not a commercial app, you need to create an imgur developer app in order API to work.
|
You need to create an imgur developer app in order API to work. Go to https://api.imgur.com/oauth2/addclient and fill the form (It does not really matter how you fill it).
|
||||||
|
|
||||||
### Creating an imgur app
|
|
||||||
* Go to https://api.imgur.com/oauth2/addclient
|
|
||||||
* Enter a name into the **Application Name** field.
|
|
||||||
* Pick **Anonymous usage without user authorization** as an **Authorization type**\*
|
|
||||||
* Enter your email into the Email field.
|
|
||||||
* Correct CHAPTCHA
|
|
||||||
* Click **submit** button
|
|
||||||
|
|
||||||
It should redirect to a page which shows your **imgur_client_id** and **imgur_client_secret**
|
It should redirect you to a page where it shows your **imgur_client_id** and **imgur_client_secret**.
|
||||||
|
|
||||||
\* Select **OAuth 2 authorization without a callback URL** first then select **Anonymous usage without user authorization** if it says *Authorization callback URL: required*
|
## [FAQ](docs/FAQ.md)
|
||||||
|
|
||||||
## FAQ
|
## [Changes on *master*](docs/CHANGELOG.md)
|
||||||
### How do I open self post files?
|
|
||||||
- Self posts are held at reddit as styled with markdown. So, the script downloads them as they are in order not to lose their stylings.
|
|
||||||
However, there is a [great Chrome extension](https://chrome.google.com/webstore/detail/markdown-viewer/ckkdlimhmcjmikdlpkmbgfkaikojcbjk) for viewing Markdown files with its styling. Install it and open the files with [Chrome](https://www.google.com/intl/tr/chrome/).
|
|
||||||
|
|
||||||
However, they are basically text files. You can also view them with any text editor such as Notepad on Windows, gedit on Linux or Text Editor on MacOS
|
|
||||||
|
|
||||||
### How can I change my credentials?
|
|
||||||
- All of the user data is held in **config.json** file which is in a folder named "Bulk Downloader for Reddit" in your **Home** directory. You can edit
|
|
||||||
them, there.
|
|
||||||
|
|
||||||
## Changes on *master*
|
|
||||||
### [24/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/7a68ff3efac9939f9574c2cef6184b92edb135f4)
|
|
||||||
- Added OP's name to file names (backwards compatible)
|
|
||||||
- Deleted # char from file names (backwards compatible)
|
|
||||||
- Improved exception handling
|
|
||||||
|
|
||||||
### [23/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/7314e17125aa78fd4e6b28e26fda7ec7db7e0147)
|
|
||||||
- Split download() function
|
|
||||||
- Added erome support
|
|
||||||
- Remove exclude feature
|
|
||||||
- Bug fix
|
|
||||||
|
|
||||||
### [22/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/a67da461d2fcd70672effcb20c8179e3224091bb)
|
|
||||||
- Put log files in a folder named "LOG_FILES"
|
|
||||||
- Fixed the bug that makes multireddit mode unusable
|
|
||||||
|
|
||||||
### [21/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/4a8c2377f9fb4d60ed7eeb8d50aaf9a26492462a)
|
|
||||||
- Added exclude mode
|
|
||||||
|
|
||||||
### [20/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/commit/7548a010198fb693841ca03654d2c9bdf5742139)
|
|
||||||
- "0" input for no limit
|
|
||||||
- Fixed the bug that recognizes none image direct links as image links
|
|
||||||
|
|
||||||
### [19/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/41cbb58db34f500a8a5ecc3ac4375bf6c3b275bb)
|
|
||||||
- Added v.redd.it support
|
|
||||||
- Added custom exception descriptions to FAILED.json file
|
|
||||||
- Fixed the bug that prevents downloading some gfycat URLs
|
|
||||||
|
|
||||||
### [13/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/9f831e1b784a770c82252e909462871401a05c11)
|
|
||||||
- Change config.json file's path to home directory
|
|
||||||
|
|
||||||
### [12/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/50a77f6ba54c24f5647d5ea4e177400b71ff04a7)
|
|
||||||
- Added binaries for Windows and Linux
|
|
||||||
- Wait on KeyboardInterrupt
|
|
||||||
- Accept multiple subreddit input
|
|
||||||
- Fixed the bug that prevents choosing "[0] exit" with typing "exit"
|
|
||||||
|
|
||||||
### [11/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/a28a7776ab826dea2a8d93873a94cd46db3a339b)
|
|
||||||
- Improvements on UX and UI
|
|
||||||
- Added logging errors to CONSOLE_LOG.txt
|
|
||||||
- Using current directory if directory has not been given yet.
|
|
||||||
|
|
||||||
### [10/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/ffe3839aee6dc1a552d95154d817aefc2b66af81)
|
|
||||||
- Added support for *self* post
|
|
||||||
- Now getting posts is quicker
|
|
||||||
1
_config.yml
Normal file
1
_config.yml
Normal file
@@ -0,0 +1 @@
|
|||||||
|
theme: jekyll-theme-cayman
|
||||||
83
docs/CHANGELOG.md
Normal file
83
docs/CHANGELOG.md
Normal file
@@ -0,0 +1,83 @@
|
|||||||
|
# Changes on *master*
|
||||||
|
## [27/01/2019](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/b7baf07fb5998368d87e3c4c36aed40daf820609)
|
||||||
|
- Clarified the instructions
|
||||||
|
|
||||||
|
## [28/08/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/d56efed1c6833a66322d9158523b89d0ce57f5de)
|
||||||
|
- Adjusted algorith used for extracting gfycat links because of gfycat's design change
|
||||||
|
- Ignore space at the end of the given directory
|
||||||
|
|
||||||
|
## [16/08/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/d56efed1c6833a66322d9158523b89d0ce57f5de)
|
||||||
|
- Fix the bug that prevents downloading imgur videos
|
||||||
|
|
||||||
|
## [15/08/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/adccd8f3ba03ad124d58643d78dab287a4123a6f)
|
||||||
|
- Prints out the title of posts' that are already downloaded
|
||||||
|
|
||||||
|
## [13/08/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/50cb7c15b9cb4befce0cfa2c23ab5de4af9176c6)
|
||||||
|
- Added alternative location of current directory for config file
|
||||||
|
- Fixed console prints on Linux
|
||||||
|
|
||||||
|
## [10/08/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/8f1ff10a5e11464575284210dbba4a0d387bc1c3)
|
||||||
|
- Added reddit username to config file
|
||||||
|
|
||||||
|
## [06/08/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/210238d0865febcb57fbd9f0b0a7d3da9dbff384)
|
||||||
|
- Sending headers when requesting a file in order not to be rejected by server
|
||||||
|
|
||||||
|
## [04/08/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/426089d0f35212148caff0082708a87017757bde)
|
||||||
|
- Disabled printing post types to console
|
||||||
|
|
||||||
|
## [30/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/af294929510f884d92b25eaa855c29fc4fb6dcaa)
|
||||||
|
- Now opens web browser and goes to Imgur when prompts for Imgur credentials
|
||||||
|
|
||||||
|
## [26/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/1623722138bad80ae39ffcd5fb38baf80680deac)
|
||||||
|
- Improved verbose mode
|
||||||
|
- Minimalized the console output
|
||||||
|
- Added quit option for auto quitting the program after process finishes
|
||||||
|
|
||||||
|
## [25/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/1623722138bad80ae39ffcd5fb38baf80680deac)
|
||||||
|
- Added verbose mode
|
||||||
|
- Stylized the console output
|
||||||
|
|
||||||
|
## [24/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/7a68ff3efac9939f9574c2cef6184b92edb135f4)
|
||||||
|
- Added OP's name to file names (backwards compatible)
|
||||||
|
- Deleted # char from file names (backwards compatible)
|
||||||
|
- Improved exception handling
|
||||||
|
|
||||||
|
## [23/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/7314e17125aa78fd4e6b28e26fda7ec7db7e0147)
|
||||||
|
- Splited download() function
|
||||||
|
- Added erome support
|
||||||
|
- Removed exclude feature
|
||||||
|
- Bug fixes
|
||||||
|
|
||||||
|
## [22/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/6e7463005051026ad64006a8580b0b5dc9536b8c)
|
||||||
|
- Put log files in a folder named "LOG_FILES"
|
||||||
|
- Fixed the bug that makes multireddit mode unusable
|
||||||
|
|
||||||
|
## [21/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/4a8c2377f9fb4d60ed7eeb8d50aaf9a26492462a)
|
||||||
|
- Added exclude mode
|
||||||
|
|
||||||
|
## [20/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/7548a010198fb693841ca03654d2c9bdf5742139)
|
||||||
|
- "0" input for no limit
|
||||||
|
- Fixed the bug that recognizes none image direct links as image links
|
||||||
|
|
||||||
|
## [19/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/41cbb58db34f500a8a5ecc3ac4375bf6c3b275bb)
|
||||||
|
- Added v.redd.it support
|
||||||
|
- Added custom exception descriptions to FAILED.json file
|
||||||
|
- Fixed the bug that prevents downloading some gfycat URLs
|
||||||
|
|
||||||
|
## [13/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/9f831e1b784a770c82252e909462871401a05c11)
|
||||||
|
- Changed config.json file's path to home directory
|
||||||
|
|
||||||
|
## [12/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/50a77f6ba54c24f5647d5ea4e177400b71ff04a7)
|
||||||
|
- Added binaries for Windows and Linux
|
||||||
|
- Wait on KeyboardInterrupt
|
||||||
|
- Accept multiple subreddit input
|
||||||
|
- Fixed the bug that prevents choosing "[0] exit" with typing "exit"
|
||||||
|
|
||||||
|
## [11/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/a28a7776ab826dea2a8d93873a94cd46db3a339b)
|
||||||
|
- Improvements on UX and UI
|
||||||
|
- Added logging errors to CONSOLE_LOG.txt
|
||||||
|
- Using current directory if directory has not been given yet.
|
||||||
|
|
||||||
|
## [10/07/2018](https://github.com/aliparlakci/bulk-downloader-for-reddit/tree/ffe3839aee6dc1a552d95154d817aefc2b66af81)
|
||||||
|
- Added support for *self* post
|
||||||
|
- Now getting posts is quicker
|
||||||
@@ -5,20 +5,24 @@ See **[compiling from source](COMPILE_FROM_SOURCE.md)** page first unless you ar
|
|||||||
***Use*** `.\bulk-downloader-for-reddit.exe` ***or*** `./bulk-downloader-for-reddit` ***if you are using the executable***.
|
***Use*** `.\bulk-downloader-for-reddit.exe` ***or*** `./bulk-downloader-for-reddit` ***if you are using the executable***.
|
||||||
```console
|
```console
|
||||||
$ python script.py --help
|
$ python script.py --help
|
||||||
usage: script.py [-h] [--directory DIRECTORY] [--link link] [--saved]
|
usage: script.py [-h] [--directory DIRECTORY] [--NoDownload] [--verbose]
|
||||||
[--submitted] [--upvoted] [--log LOG FILE]
|
[--quit] [--link link] [--saved] [--submitted] [--upvoted]
|
||||||
[--subreddit SUBREDDIT [SUBREDDIT ...]]
|
[--log LOG FILE] [--subreddit SUBREDDIT [SUBREDDIT ...]]
|
||||||
[--multireddit MULTIREDDIT] [--user redditor]
|
[--multireddit MULTIREDDIT] [--user redditor]
|
||||||
[--search query] [--sort SORT TYPE] [--limit Limit]
|
[--search query] [--sort SORT TYPE] [--limit Limit]
|
||||||
[--time TIME_LIMIT] [--NoDownload]
|
[--time TIME_LIMIT]
|
||||||
|
|
||||||
This program downloads media from reddit posts
|
This program downloads media from reddit posts
|
||||||
|
|
||||||
optional arguments:
|
optional arguments:
|
||||||
-h, --help show this help message and exit
|
-h, --help show this help message and exit
|
||||||
--directory DIRECTORY
|
--directory DIRECTORY, -d DIRECTORY
|
||||||
Specifies the directory where posts will be downloaded
|
Specifies the directory where posts will be downloaded
|
||||||
to
|
to
|
||||||
|
--NoDownload Just gets the posts and stores them in a file for
|
||||||
|
downloading later
|
||||||
|
--verbose, -v Verbose Mode
|
||||||
|
--quit, -q Auto quit afer the process finishes
|
||||||
--link link, -l link Get posts from link
|
--link link, -l link Get posts from link
|
||||||
--saved Triggers saved mode
|
--saved Triggers saved mode
|
||||||
--submitted Gets posts of --user
|
--submitted Gets posts of --user
|
||||||
@@ -38,8 +42,6 @@ optional arguments:
|
|||||||
--limit Limit default: unlimited
|
--limit Limit default: unlimited
|
||||||
--time TIME_LIMIT Either hour, day, week, month, year or all. default:
|
--time TIME_LIMIT Either hour, day, week, month, year or all. default:
|
||||||
all
|
all
|
||||||
--NoDownload Just gets the posts and store them in a file for
|
|
||||||
downloading later
|
|
||||||
```
|
```
|
||||||
|
|
||||||
# Examples
|
# Examples
|
||||||
|
|||||||
@@ -1,16 +1,14 @@
|
|||||||
# Compiling from source code
|
# Compiling from source code
|
||||||
## Requirements
|
## Requirements
|
||||||
### Python 3 Interpreter
|
### Python 3 Interpreter
|
||||||
Latest* version of **Python 3** is needed. See if it is already installed [here](#finding-the-correct-keyword-for-python). If not, download the matching release for your platform [here](https://www.python.org/downloads/) and install it. If you are a *Windows* user, selecting **Add Python 3 to PATH** option is mandatory.
|
Latest* version of **Python 3** is needed. See if it is already installed [here](#finding-the-correct-keyword-for-python). If not, download the matching release for your platform [here](https://www.python.org/downloads/) and install it. If you are a *Windows* user, selecting **Add Python 3 to PATH** option when installing the software is mandatory.
|
||||||
|
|
||||||
\* *Use Python 3.6.5 if you encounter an issue*
|
\* *Use Python 3.6.5 if you encounter an issue*
|
||||||
## Using terminal
|
## Using terminal
|
||||||
### To open it...
|
### To open it...
|
||||||
- **On Windows 8/8.1/10**: Press the File tab on **Windows Explorer**, click on **Open Windows PowerShell** or **Open Windows Command Prompt** or look for *Command Prompt* or *PowerShell* in *Start Menu*.
|
- **On Windows**: Press **Shift+Right Click**, select **Open Powershell window here** or **Open Command Prompt window here**
|
||||||
|
|
||||||
- **On Windows 7**: Press **WindowsKey+R**, type **cmd** and hit Enter or look for *Command Prompt* or *PowerShell* in *Start Menu*.
|
- **On Linux**: Right-click in a folder and select **Open Terminal** or press **Ctrl+Alt+T**.
|
||||||
|
|
||||||
- **On Linux**: Right-click in a folder and select **Open Terminal** or press **Ctrl+Alt+T** or look for **Terminal** in the programs.
|
|
||||||
|
|
||||||
- **On MacOS**: Look for an app called **Terminal**.
|
- **On MacOS**: Look for an app called **Terminal**.
|
||||||
|
|
||||||
@@ -39,4 +37,4 @@ python -m pip install -r requirements.txt
|
|||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
Now, you can go to [Using command-line arguments](COMMAND_LINE_ARGUMENTS.md)
|
Now, you can go to [Using command-line arguments](COMMAND_LINE_ARGUMENTS.md)
|
||||||
|
|||||||
23
docs/FAQ.md
Normal file
23
docs/FAQ.md
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
# FAQ
|
||||||
|
## What do the dots resemble when getting posts?
|
||||||
|
- Each dot means that 100 posts are scanned.
|
||||||
|
|
||||||
|
## Getting posts is taking too long.
|
||||||
|
- You can press Ctrl+C to interrupt it and start downloading.
|
||||||
|
|
||||||
|
## How are filenames formatted?
|
||||||
|
- Self posts and images that are not belong to an album are formatted as **`[SUBMITTER NAME]_[POST TITLE]_[REDDIT ID]`**.
|
||||||
|
You can use *reddit id* to go to post's reddit page by going to link **reddit.com/[REDDIT ID]**
|
||||||
|
|
||||||
|
- An image in an imgur album is formatted as **`[ITEM NUMBER]_[IMAGE TITLE]_[IMGUR ID]`**
|
||||||
|
Similarly, you can use *imgur id* to go to image's imgur page by going to link **imgur.com/[IMGUR ID]**.
|
||||||
|
|
||||||
|
## How do I open self post files?
|
||||||
|
- Self posts are held at reddit as styled with markdown. So, the script downloads them as they are in order not to lose their stylings.
|
||||||
|
However, there is a [great Chrome extension](https://chrome.google.com/webstore/detail/markdown-viewer/ckkdlimhmcjmikdlpkmbgfkaikojcbjk) for viewing Markdown files with its styling. Install it and open the files with [Chrome](https://www.google.com/intl/tr/chrome/).
|
||||||
|
|
||||||
|
However, they are basically text files. You can also view them with any text editor such as Notepad on Windows, gedit on Linux or Text Editor on MacOS
|
||||||
|
|
||||||
|
## How can I change my credentials?
|
||||||
|
- All of the user data is held in **config.json** file which is in a folder named "Bulk Downloader for Reddit" in your **Home** directory. You can edit
|
||||||
|
them, there.
|
||||||
@@ -1,3 +1,4 @@
|
|||||||
|
bs4
|
||||||
requests
|
requests
|
||||||
praw
|
praw
|
||||||
imgurpython
|
imgurpython
|
||||||
174
script.py
174
script.py
@@ -10,10 +10,11 @@ import logging
|
|||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
import time
|
import time
|
||||||
|
import webbrowser
|
||||||
from io import StringIO
|
from io import StringIO
|
||||||
from pathlib import Path, PurePath
|
from pathlib import Path, PurePath
|
||||||
|
|
||||||
from src.downloader import Direct, Gfycat, Imgur, Self, Erome
|
from src.downloader import Direct, Erome, Gfycat, Imgur, Self
|
||||||
from src.errors import *
|
from src.errors import *
|
||||||
from src.parser import LinkDesigner
|
from src.parser import LinkDesigner
|
||||||
from src.searcher import getPosts
|
from src.searcher import getPosts
|
||||||
@@ -22,7 +23,7 @@ from src.tools import (GLOBAL, createLogFile, jsonFile, nameCorrector,
|
|||||||
|
|
||||||
__author__ = "Ali Parlakci"
|
__author__ = "Ali Parlakci"
|
||||||
__license__ = "GPL"
|
__license__ = "GPL"
|
||||||
__version__ = "1.5.0"
|
__version__ = "1.6.4.1"
|
||||||
__maintainer__ = "Ali Parlakci"
|
__maintainer__ = "Ali Parlakci"
|
||||||
__email__ = "parlakciali@gmail.com"
|
__email__ = "parlakciali@gmail.com"
|
||||||
|
|
||||||
@@ -38,20 +39,34 @@ def getConfig(configFileName):
|
|||||||
if "reddit_refresh_token" in content:
|
if "reddit_refresh_token" in content:
|
||||||
if content["reddit_refresh_token"] == "":
|
if content["reddit_refresh_token"] == "":
|
||||||
FILE.delete("reddit_refresh_token")
|
FILE.delete("reddit_refresh_token")
|
||||||
|
|
||||||
|
if not all(False if content.get(key,"") == "" else True for key in keys):
|
||||||
|
print(
|
||||||
|
"Go to this URL and fill the form: " \
|
||||||
|
"https://api.imgur.com/oauth2/addclient\n" \
|
||||||
|
"Enter the client id and client secret here:"
|
||||||
|
)
|
||||||
|
webbrowser.open("https://api.imgur.com/oauth2/addclient",new=2)
|
||||||
|
|
||||||
for key in keys:
|
for key in keys:
|
||||||
try:
|
try:
|
||||||
if content[key] == "":
|
if content[key] == "":
|
||||||
raise KeyError
|
raise KeyError
|
||||||
except KeyError:
|
except KeyError:
|
||||||
print(key,": ")
|
FILE.add({key:input(" "+key+": ")})
|
||||||
FILE.add({key:input()})
|
|
||||||
return jsonFile(configFileName).read()
|
return jsonFile(configFileName).read()
|
||||||
|
|
||||||
else:
|
else:
|
||||||
FILE = jsonFile(configFileName)
|
FILE = jsonFile(configFileName)
|
||||||
configDictionary = {}
|
configDictionary = {}
|
||||||
|
print(
|
||||||
|
"Go to this URL and fill the form: " \
|
||||||
|
"https://api.imgur.com/oauth2/addclient\n" \
|
||||||
|
"Enter the client id and client secret here:"
|
||||||
|
)
|
||||||
|
webbrowser.open("https://api.imgur.com/oauth2/addclient",new=2)
|
||||||
for key in keys:
|
for key in keys:
|
||||||
configDictionary[key] = input(key + ": ")
|
configDictionary[key] = input(" "+key+": ")
|
||||||
FILE.add(configDictionary)
|
FILE.add(configDictionary)
|
||||||
return FILE.read()
|
return FILE.read()
|
||||||
|
|
||||||
@@ -62,10 +77,26 @@ def parseArguments(arguments=[]):
|
|||||||
description="This program downloads " \
|
description="This program downloads " \
|
||||||
"media from reddit " \
|
"media from reddit " \
|
||||||
"posts")
|
"posts")
|
||||||
parser.add_argument("--directory",
|
parser.add_argument("--directory","-d",
|
||||||
help="Specifies the directory where posts will be " \
|
help="Specifies the directory where posts will be " \
|
||||||
"downloaded to",
|
"downloaded to",
|
||||||
metavar="DIRECTORY")
|
metavar="DIRECTORY")
|
||||||
|
|
||||||
|
parser.add_argument("--NoDownload",
|
||||||
|
help="Just gets the posts and stores them in a file" \
|
||||||
|
" for downloading later",
|
||||||
|
action="store_true",
|
||||||
|
default=False)
|
||||||
|
|
||||||
|
parser.add_argument("--verbose","-v",
|
||||||
|
help="Verbose Mode",
|
||||||
|
action="store_true",
|
||||||
|
default=False)
|
||||||
|
|
||||||
|
parser.add_argument("--quit","-q",
|
||||||
|
help="Auto quit afer the process finishes",
|
||||||
|
action="store_true",
|
||||||
|
default=False)
|
||||||
|
|
||||||
parser.add_argument("--link","-l",
|
parser.add_argument("--link","-l",
|
||||||
help="Get posts from link",
|
help="Get posts from link",
|
||||||
@@ -137,13 +168,6 @@ def parseArguments(arguments=[]):
|
|||||||
choices=["all","hour","day","week","month","year"],
|
choices=["all","hour","day","week","month","year"],
|
||||||
metavar="TIME_LIMIT",
|
metavar="TIME_LIMIT",
|
||||||
type=str)
|
type=str)
|
||||||
|
|
||||||
parser.add_argument("--NoDownload",
|
|
||||||
help="Just gets the posts and store them in a file" \
|
|
||||||
" for downloading later",
|
|
||||||
action="store_true",
|
|
||||||
default=False)
|
|
||||||
|
|
||||||
|
|
||||||
if arguments == []:
|
if arguments == []:
|
||||||
return parser.parse_args()
|
return parser.parse_args()
|
||||||
@@ -160,9 +184,10 @@ def checkConflicts():
|
|||||||
else:
|
else:
|
||||||
user = 1
|
user = 1
|
||||||
|
|
||||||
|
search = 1 if GLOBAL.arguments.search else 0
|
||||||
|
|
||||||
modes = [
|
modes = [
|
||||||
"saved","subreddit","submitted","search","log","link","upvoted",
|
"saved","subreddit","submitted","log","link","upvoted","multireddit"
|
||||||
"multireddit"
|
|
||||||
]
|
]
|
||||||
|
|
||||||
values = {
|
values = {
|
||||||
@@ -175,15 +200,18 @@ def checkConflicts():
|
|||||||
if not sum(values[x] for x in values) == 1:
|
if not sum(values[x] for x in values) == 1:
|
||||||
raise ProgramModeError("Invalid program mode")
|
raise ProgramModeError("Invalid program mode")
|
||||||
|
|
||||||
if values["search"]+values["saved"] == 2:
|
if search+values["saved"] == 2:
|
||||||
raise SearchModeError("You cannot search in your saved posts")
|
raise SearchModeError("You cannot search in your saved posts")
|
||||||
|
|
||||||
if values["search"]+values["submitted"] == 2:
|
if search+values["submitted"] == 2:
|
||||||
raise SearchModeError("You cannot search in submitted posts")
|
raise SearchModeError("You cannot search in submitted posts")
|
||||||
|
|
||||||
if values["search"]+values["upvoted"] == 2:
|
if search+values["upvoted"] == 2:
|
||||||
raise SearchModeError("You cannot search in upvoted posts")
|
raise SearchModeError("You cannot search in upvoted posts")
|
||||||
|
|
||||||
|
if search+values["log"] == 2:
|
||||||
|
raise SearchModeError("You cannot search in log files")
|
||||||
|
|
||||||
if values["upvoted"]+values["submitted"] == 1 and user == 0:
|
if values["upvoted"]+values["submitted"] == 1 and user == 0:
|
||||||
raise RedditorNameError("No redditor name given")
|
raise RedditorNameError("No redditor name given")
|
||||||
|
|
||||||
@@ -237,12 +265,15 @@ class PromptUser:
|
|||||||
|
|
||||||
if programMode == "subreddit":
|
if programMode == "subreddit":
|
||||||
|
|
||||||
subredditInput = input("subreddit (enter frontpage for frontpage): ")
|
subredditInput = input("(type frontpage for all subscribed subreddits,\n" \
|
||||||
|
" use plus to seperate multi subreddits:" \
|
||||||
|
" pics+funny+me_irl etc.)\n\n" \
|
||||||
|
"subreddit: ")
|
||||||
GLOBAL.arguments.subreddit = subredditInput
|
GLOBAL.arguments.subreddit = subredditInput
|
||||||
|
|
||||||
while not (subredditInput == "" or subredditInput.lower() == "frontpage"):
|
# while not (subredditInput == "" or subredditInput.lower() == "frontpage"):
|
||||||
subredditInput = input("subreddit: ")
|
# subredditInput = input("subreddit: ")
|
||||||
GLOBAL.arguments.subreddit += "+" + subredditInput
|
# GLOBAL.arguments.subreddit += "+" + subredditInput
|
||||||
|
|
||||||
if " " in GLOBAL.arguments.subreddit:
|
if " " in GLOBAL.arguments.subreddit:
|
||||||
GLOBAL.arguments.subreddit = "+".join(GLOBAL.arguments.subreddit.split())
|
GLOBAL.arguments.subreddit = "+".join(GLOBAL.arguments.subreddit.split())
|
||||||
@@ -269,7 +300,7 @@ class PromptUser:
|
|||||||
GLOBAL.arguments.time = "all"
|
GLOBAL.arguments.time = "all"
|
||||||
|
|
||||||
elif programMode == "multireddit":
|
elif programMode == "multireddit":
|
||||||
GLOBAL.arguments.user = input("\nredditor: ")
|
GLOBAL.arguments.user = input("\nmultireddit owner: ")
|
||||||
GLOBAL.arguments.multireddit = input("\nmultireddit: ")
|
GLOBAL.arguments.multireddit = input("\nmultireddit: ")
|
||||||
|
|
||||||
print("\nselect sort type:")
|
print("\nselect sort type:")
|
||||||
@@ -361,10 +392,7 @@ def prepareAttributes():
|
|||||||
|
|
||||||
GLOBAL.arguments.link = GLOBAL.arguments.link.strip("\"")
|
GLOBAL.arguments.link = GLOBAL.arguments.link.strip("\"")
|
||||||
|
|
||||||
try:
|
ATTRIBUTES = LinkDesigner(GLOBAL.arguments.link)
|
||||||
ATTRIBUTES = LinkDesigner(GLOBAL.arguments.link)
|
|
||||||
except InvalidRedditLink:
|
|
||||||
raise InvalidRedditLink
|
|
||||||
|
|
||||||
if GLOBAL.arguments.search is not None:
|
if GLOBAL.arguments.search is not None:
|
||||||
ATTRIBUTES["search"] = GLOBAL.arguments.search
|
ATTRIBUTES["search"] = GLOBAL.arguments.search
|
||||||
@@ -394,7 +422,7 @@ def prepareAttributes():
|
|||||||
ATTRIBUTES["submitted"] = True
|
ATTRIBUTES["submitted"] = True
|
||||||
|
|
||||||
if GLOBAL.arguments.sort == "rising":
|
if GLOBAL.arguments.sort == "rising":
|
||||||
raise InvalidSortingType
|
raise InvalidSortingType("Invalid sorting type has given")
|
||||||
|
|
||||||
ATTRIBUTES["limit"] = GLOBAL.arguments.limit
|
ATTRIBUTES["limit"] = GLOBAL.arguments.limit
|
||||||
|
|
||||||
@@ -431,6 +459,9 @@ def isPostExists(POST):
|
|||||||
|
|
||||||
possibleExtensions = [".jpg",".png",".mp4",".gif",".webm",".md"]
|
possibleExtensions = [".jpg",".png",".mp4",".gif",".webm",".md"]
|
||||||
|
|
||||||
|
"""If you change the filenames, don't forget to add them here.
|
||||||
|
Please don't remove existing ones
|
||||||
|
"""
|
||||||
for extension in possibleExtensions:
|
for extension in possibleExtensions:
|
||||||
|
|
||||||
OLD_FILE_PATH = PATH / (
|
OLD_FILE_PATH = PATH / (
|
||||||
@@ -457,6 +488,8 @@ def isPostExists(POST):
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
def downloadPost(SUBMISSION):
|
def downloadPost(SUBMISSION):
|
||||||
|
|
||||||
|
"""Download directory is declared here for each file"""
|
||||||
directory = GLOBAL.directory / SUBMISSION['postSubreddit']
|
directory = GLOBAL.directory / SUBMISSION['postSubreddit']
|
||||||
|
|
||||||
global lastRequestTime
|
global lastRequestTime
|
||||||
@@ -465,13 +498,12 @@ def downloadPost(SUBMISSION):
|
|||||||
"imgur":Imgur,"gfycat":Gfycat,"erome":Erome,"direct":Direct,"self":Self
|
"imgur":Imgur,"gfycat":Gfycat,"erome":Erome,"direct":Direct,"self":Self
|
||||||
}
|
}
|
||||||
|
|
||||||
|
print()
|
||||||
if SUBMISSION['postType'] in downloaders:
|
if SUBMISSION['postType'] in downloaders:
|
||||||
|
|
||||||
print(SUBMISSION['postType'].upper())
|
|
||||||
|
|
||||||
if SUBMISSION['postType'] == "imgur":
|
if SUBMISSION['postType'] == "imgur":
|
||||||
|
|
||||||
if int(time.time() - lastRequestTime) <= 2:
|
while int(time.time() - lastRequestTime) <= 2:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
credit = Imgur.get_credits()
|
credit = Imgur.get_credits()
|
||||||
@@ -482,13 +514,18 @@ def downloadPost(SUBMISSION):
|
|||||||
+ " Minutes " \
|
+ " Minutes " \
|
||||||
+ str(int(IMGUR_RESET_TIME%60)) \
|
+ str(int(IMGUR_RESET_TIME%60)) \
|
||||||
+ " Seconds")
|
+ " Seconds")
|
||||||
|
|
||||||
|
if credit['ClientRemaining'] < 25 or credit['UserRemaining'] < 25:
|
||||||
|
printCredit = {"noPrint":False}
|
||||||
|
else:
|
||||||
|
printCredit = {"noPrint":True}
|
||||||
|
|
||||||
print(
|
print(
|
||||||
"Client: {} - User: {} - Reset {}".format(
|
"==> Client: {} - User: {} - Reset {}\n".format(
|
||||||
credit['ClientRemaining'],
|
credit['ClientRemaining'],
|
||||||
credit['UserRemaining'],
|
credit['UserRemaining'],
|
||||||
USER_RESET
|
USER_RESET
|
||||||
)
|
),end="",**printCredit
|
||||||
)
|
)
|
||||||
|
|
||||||
if not (credit['UserRemaining'] == 0 or \
|
if not (credit['UserRemaining'] == 0 or \
|
||||||
@@ -496,7 +533,7 @@ def downloadPost(SUBMISSION):
|
|||||||
|
|
||||||
"""This block of code is needed
|
"""This block of code is needed
|
||||||
"""
|
"""
|
||||||
if int(time.time() - lastRequestTime) <= 2:
|
while int(time.time() - lastRequestTime) <= 2:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
lastRequestTime = time.time()
|
lastRequestTime = time.time()
|
||||||
@@ -530,16 +567,14 @@ def download(submissions):
|
|||||||
FAILED_FILE = createLogFile("FAILED")
|
FAILED_FILE = createLogFile("FAILED")
|
||||||
|
|
||||||
for i in range(subsLenght):
|
for i in range(subsLenght):
|
||||||
print("\n({}/{})".format(i+1,subsLenght))
|
print(f"\n({i+1}/{subsLenght}) – r/{submissions[i]['postSubreddit']}",
|
||||||
print(
|
end="")
|
||||||
"https://reddit.com/r/{subreddit}/comments/{id}".format(
|
print(f" – {submissions[i]['postType'].upper()}",end="",noPrint=True)
|
||||||
subreddit=submissions[i]['postSubreddit'],
|
|
||||||
id=submissions[i]['postId']
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
if isPostExists(submissions[i]):
|
if isPostExists(submissions[i]):
|
||||||
print(submissions[i]['postType'].upper())
|
print(f"\n" \
|
||||||
|
f"{submissions[i]['postSubmitter']}_"
|
||||||
|
f"{nameCorrector(submissions[i]['postTitle'])}")
|
||||||
print("It already exists")
|
print("It already exists")
|
||||||
duplicates += 1
|
duplicates += 1
|
||||||
downloadedCount -= 1
|
downloadedCount -= 1
|
||||||
@@ -603,23 +638,33 @@ def download(submissions):
|
|||||||
downloadedCount -= 1
|
downloadedCount -= 1
|
||||||
|
|
||||||
if duplicates:
|
if duplicates:
|
||||||
print("\n There was {} duplicates".format(duplicates))
|
print(f"\nThere {'were' if duplicates > 1 else 'was'} " \
|
||||||
|
f"{duplicates} duplicate{'s' if duplicates > 1 else ''}")
|
||||||
|
|
||||||
if downloadedCount == 0:
|
if downloadedCount == 0:
|
||||||
print(" Nothing downloaded :(")
|
print("Nothing downloaded :(")
|
||||||
|
|
||||||
else:
|
else:
|
||||||
print(" Total of {} links downloaded!".format(downloadedCount))
|
print(f"Total of {downloadedCount} " \
|
||||||
|
f"link{'s' if downloadedCount > 1 else ''} downloaded!")
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
|
VanillaPrint(
|
||||||
|
f"\nBulk Downloader for Reddit v{__version__}\n" \
|
||||||
|
f"Written by Ali PARLAKCI – parlakciali@gmail.com\n\n" \
|
||||||
|
f"https://github.com/aliparlakci/bulk-downloader-for-reddit/"
|
||||||
|
)
|
||||||
GLOBAL.arguments = parseArguments()
|
GLOBAL.arguments = parseArguments()
|
||||||
|
|
||||||
if GLOBAL.arguments.directory is not None:
|
if GLOBAL.arguments.directory is not None:
|
||||||
GLOBAL.directory = Path(GLOBAL.arguments.directory)
|
GLOBAL.directory = Path(GLOBAL.arguments.directory.strip())
|
||||||
else:
|
else:
|
||||||
GLOBAL.directory = Path(input("download directory: "))
|
GLOBAL.directory = Path(input("\ndownload directory: ").strip())
|
||||||
|
|
||||||
print("\n"," ".join(sys.argv),"\n")
|
print("\n"," ".join(sys.argv),"\n",noPrint=True)
|
||||||
|
print(f"Bulk Downloader for Reddit v{__version__}\n",noPrint=True
|
||||||
|
)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
checkConflicts()
|
checkConflicts()
|
||||||
@@ -628,35 +673,21 @@ def main():
|
|||||||
|
|
||||||
if not Path(GLOBAL.configDirectory).is_dir():
|
if not Path(GLOBAL.configDirectory).is_dir():
|
||||||
os.makedirs(GLOBAL.configDirectory)
|
os.makedirs(GLOBAL.configDirectory)
|
||||||
GLOBAL.config = getConfig(GLOBAL.configDirectory / "config.json")
|
GLOBAL.config = getConfig("config.json") if Path("config.json").exists() \
|
||||||
|
else getConfig(GLOBAL.configDirectory / "config.json")
|
||||||
|
|
||||||
if GLOBAL.arguments.log is not None:
|
if GLOBAL.arguments.log is not None:
|
||||||
logDir = Path(GLOBAL.arguments.log)
|
logDir = Path(GLOBAL.arguments.log)
|
||||||
download(postFromLog(logDir))
|
download(postFromLog(logDir))
|
||||||
sys.exit()
|
sys.exit()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
POSTS = getPosts(prepareAttributes())
|
POSTS = getPosts(prepareAttributes())
|
||||||
except InsufficientPermission:
|
except Exception as exc:
|
||||||
print("You do not have permission to do that")
|
logging.error(sys.exc_info()[0].__name__,
|
||||||
sys.exit()
|
exc_info=full_exc_info(sys.exc_info()))
|
||||||
except NoMatchingSubmissionFound:
|
print(log_stream.getvalue(),noPrint=True)
|
||||||
print("No matching submission was found")
|
print(exc)
|
||||||
sys.exit()
|
|
||||||
except NoRedditSupoort:
|
|
||||||
print("Reddit does not support that")
|
|
||||||
sys.exit()
|
|
||||||
except NoPrawSupport:
|
|
||||||
print("PRAW does not support that")
|
|
||||||
sys.exit()
|
|
||||||
except MultiredditNotFound:
|
|
||||||
print("Multireddit not found")
|
|
||||||
sys.exit()
|
|
||||||
except InvalidSortingType:
|
|
||||||
print("Invalid sorting type has given")
|
|
||||||
sys.exit()
|
|
||||||
except InvalidRedditLink:
|
|
||||||
print("Invalid reddit link")
|
|
||||||
sys.exit()
|
sys.exit()
|
||||||
|
|
||||||
if POSTS is None:
|
if POSTS is None:
|
||||||
@@ -683,7 +714,6 @@ if __name__ == "__main__":
|
|||||||
except KeyboardInterrupt:
|
except KeyboardInterrupt:
|
||||||
if GLOBAL.directory is None:
|
if GLOBAL.directory is None:
|
||||||
GLOBAL.directory = Path(".\\")
|
GLOBAL.directory = Path(".\\")
|
||||||
print("\nQUITTING...")
|
|
||||||
|
|
||||||
except Exception as exception:
|
except Exception as exception:
|
||||||
if GLOBAL.directory is None:
|
if GLOBAL.directory is None:
|
||||||
@@ -692,4 +722,4 @@ if __name__ == "__main__":
|
|||||||
exc_info=full_exc_info(sys.exc_info()))
|
exc_info=full_exc_info(sys.exc_info()))
|
||||||
print(log_stream.getvalue())
|
print(log_stream.getvalue())
|
||||||
|
|
||||||
input("\nPress enter to quit\n")
|
if not GLOBAL.arguments.quit: input("\nPress enter to quit\n")
|
||||||
|
|||||||
@@ -1,13 +1,15 @@
|
|||||||
import io
|
import io
|
||||||
|
import json
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
import urllib.request
|
import urllib.request
|
||||||
from html.parser import HTMLParser
|
from html.parser import HTMLParser
|
||||||
|
from multiprocessing import Queue
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from urllib.error import HTTPError
|
from urllib.error import HTTPError
|
||||||
|
|
||||||
import imgurpython
|
import imgurpython
|
||||||
from multiprocessing import Queue
|
from bs4 import BeautifulSoup
|
||||||
|
|
||||||
from src.errors import (AlbumNotDownloadedCompletely, FileAlreadyExistsError,
|
from src.errors import (AlbumNotDownloadedCompletely, FileAlreadyExistsError,
|
||||||
FileNameTooLong, ImgurLoginError,
|
FileNameTooLong, ImgurLoginError,
|
||||||
@@ -23,8 +25,7 @@ def dlProgress(count, blockSize, totalSize):
|
|||||||
|
|
||||||
downloadedMbs = int(count*blockSize*(10**(-6)))
|
downloadedMbs = int(count*blockSize*(10**(-6)))
|
||||||
fileSize = int(totalSize*(10**(-6)))
|
fileSize = int(totalSize*(10**(-6)))
|
||||||
sys.stdout.write("\r{}Mb/{}Mb".format(downloadedMbs,fileSize))
|
sys.stdout.write("{}Mb/{}Mb\r".format(downloadedMbs,fileSize))
|
||||||
sys.stdout.write("\b"*len("\r{}Mb/{}Mb".format(downloadedMbs,fileSize)))
|
|
||||||
sys.stdout.flush()
|
sys.stdout.flush()
|
||||||
|
|
||||||
def getExtension(link):
|
def getExtension(link):
|
||||||
@@ -54,6 +55,23 @@ def getFile(fileDir,tempDir,imageURL,indent=0):
|
|||||||
As too long file names seem not working.
|
As too long file names seem not working.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
headers = [
|
||||||
|
("User-Agent", "Mozilla/5.0 (Windows NT 10.0; Win64; x64) " \
|
||||||
|
"AppleWebKit/537.36 (KHTML, like Gecko) Chrome/67.0.3396.87 "\
|
||||||
|
"Safari/537.36 OPR/54.0.2952.64"),
|
||||||
|
("Accept", "text/html,application/xhtml+xml,application/xml;" \
|
||||||
|
"q=0.9,image/webp,image/apng,*/*;q=0.8"),
|
||||||
|
("Accept-Charset", "ISO-8859-1,utf-8;q=0.7,*;q=0.3"),
|
||||||
|
("Accept-Encoding", "none"),
|
||||||
|
("Accept-Language", "en-US,en;q=0.8"),
|
||||||
|
("Connection", "keep-alive")
|
||||||
|
]
|
||||||
|
|
||||||
|
opener = urllib.request.build_opener()
|
||||||
|
if not "imgur" in imageURL:
|
||||||
|
opener.addheaders = headers
|
||||||
|
urllib.request.install_opener(opener)
|
||||||
|
|
||||||
if not (os.path.isfile(fileDir)):
|
if not (os.path.isfile(fileDir)):
|
||||||
for i in range(3):
|
for i in range(3):
|
||||||
try:
|
try:
|
||||||
@@ -61,13 +79,14 @@ def getFile(fileDir,tempDir,imageURL,indent=0):
|
|||||||
tempDir,
|
tempDir,
|
||||||
reporthook=dlProgress)
|
reporthook=dlProgress)
|
||||||
os.rename(tempDir,fileDir)
|
os.rename(tempDir,fileDir)
|
||||||
print(" "*indent+"Downloaded"+" "*10)
|
|
||||||
break
|
|
||||||
except ConnectionResetError as exception:
|
except ConnectionResetError as exception:
|
||||||
print(" "*indent + str(exception))
|
print(" "*indent + str(exception))
|
||||||
print(" "*indent + "Trying again\n")
|
print(" "*indent + "Trying again\n")
|
||||||
except FileNotFoundError:
|
except FileNotFoundError:
|
||||||
raise FileNameTooLong
|
raise FileNameTooLong
|
||||||
|
else:
|
||||||
|
print(" "*indent+"Downloaded"+" "*10)
|
||||||
|
break
|
||||||
else:
|
else:
|
||||||
raise FileAlreadyExistsError
|
raise FileAlreadyExistsError
|
||||||
|
|
||||||
@@ -86,14 +105,16 @@ class Erome:
|
|||||||
|
|
||||||
extension = getExtension(IMAGES[0])
|
extension = getExtension(IMAGES[0])
|
||||||
|
|
||||||
|
"""Filenames are declared here"""
|
||||||
|
|
||||||
title = nameCorrector(post['postTitle'])
|
title = nameCorrector(post['postTitle'])
|
||||||
print(post["postSubmitter"]+"_"+title+"_"+post['postId']+extension)
|
print(post["postSubmitter"]+"_"+title+"_"+post['postId']+extension)
|
||||||
|
|
||||||
fileDir = directory / (
|
fileDir = directory / (
|
||||||
POST["postSubmitter"]+"_"+title+"_"+POST['postId']+extension
|
post["postSubmitter"]+"_"+title+"_"+post['postId']+extension
|
||||||
)
|
)
|
||||||
tempDir = directory / (
|
tempDir = directory / (
|
||||||
POST["postSubmitter"]+"_"+title+"_"+POST['postId']+".tmp"
|
post["postSubmitter"]+"_"+title+"_"+post['postId']+".tmp"
|
||||||
)
|
)
|
||||||
|
|
||||||
imageURL = "https:" + IMAGES[0]
|
imageURL = "https:" + IMAGES[0]
|
||||||
@@ -221,8 +242,11 @@ class Imgur:
|
|||||||
post['mediaURL'] = content['object'].link
|
post['mediaURL'] = content['object'].link
|
||||||
|
|
||||||
post['postExt'] = getExtension(post['mediaURL'])
|
post['postExt'] = getExtension(post['mediaURL'])
|
||||||
|
|
||||||
title = nameCorrector(post['postTitle'])
|
title = nameCorrector(post['postTitle'])
|
||||||
|
|
||||||
|
"""Filenames are declared here"""
|
||||||
|
|
||||||
print(post["postSubmitter"]+"_"+title+"_"+post['postId']+post['postExt'])
|
print(post["postSubmitter"]+"_"+title+"_"+post['postId']+post['postExt'])
|
||||||
|
|
||||||
fileDir = directory / (
|
fileDir = directory / (
|
||||||
@@ -281,6 +305,8 @@ class Imgur:
|
|||||||
+ "_"
|
+ "_"
|
||||||
+ images[i]['id'])
|
+ images[i]['id'])
|
||||||
|
|
||||||
|
"""Filenames are declared here"""
|
||||||
|
|
||||||
fileDir = folderDir / (fileName + images[i]['Ext'])
|
fileDir = folderDir / (fileName + images[i]['Ext'])
|
||||||
tempDir = folderDir / (fileName + ".tmp")
|
tempDir = folderDir / (fileName + ".tmp")
|
||||||
|
|
||||||
@@ -377,12 +403,17 @@ class Gfycat:
|
|||||||
except IndexError:
|
except IndexError:
|
||||||
raise NotADownloadableLinkError("Could not read the page source")
|
raise NotADownloadableLinkError("Could not read the page source")
|
||||||
except Exception as exception:
|
except Exception as exception:
|
||||||
|
#debug
|
||||||
|
raise exception
|
||||||
raise NotADownloadableLinkError("Could not read the page source")
|
raise NotADownloadableLinkError("Could not read the page source")
|
||||||
|
|
||||||
POST['postExt'] = getExtension(POST['mediaURL'])
|
POST['postExt'] = getExtension(POST['mediaURL'])
|
||||||
|
|
||||||
if not os.path.exists(directory): os.makedirs(directory)
|
if not os.path.exists(directory): os.makedirs(directory)
|
||||||
title = nameCorrector(POST['postTitle'])
|
title = nameCorrector(POST['postTitle'])
|
||||||
|
|
||||||
|
"""Filenames are declared here"""
|
||||||
|
|
||||||
print(POST["postSubmitter"]+"_"+title+"_"+POST['postId']+POST['postExt'])
|
print(POST["postSubmitter"]+"_"+title+"_"+POST['postId']+POST['postExt'])
|
||||||
|
|
||||||
fileDir = directory / (
|
fileDir = directory / (
|
||||||
@@ -413,30 +444,25 @@ class Gfycat:
|
|||||||
|
|
||||||
url = "https://gfycat.com/" + url.split('/')[-1]
|
url = "https://gfycat.com/" + url.split('/')[-1]
|
||||||
|
|
||||||
pageSource = (urllib.request.urlopen(url).read().decode().split('\n'))
|
pageSource = (urllib.request.urlopen(url).read().decode())
|
||||||
|
|
||||||
theLine = pageSource[lineNumber]
|
soup = BeautifulSoup(pageSource, "html.parser")
|
||||||
lenght = len(query)
|
attributes = {"data-react-helmet":"true","type":"application/ld+json"}
|
||||||
link = []
|
content = soup.find("script",attrs=attributes)
|
||||||
|
|
||||||
for i in range(len(theLine)):
|
if content is None:
|
||||||
if theLine[i:i+lenght] == query:
|
|
||||||
cursor = (i+lenght)+1
|
|
||||||
while not theLine[cursor] == '"':
|
|
||||||
link.append(theLine[cursor])
|
|
||||||
cursor += 1
|
|
||||||
break
|
|
||||||
|
|
||||||
if "".join(link) == "":
|
|
||||||
raise NotADownloadableLinkError("Could not read the page source")
|
raise NotADownloadableLinkError("Could not read the page source")
|
||||||
|
|
||||||
return "".join(link)
|
return json.loads(content.text)["video"]["contentUrl"]
|
||||||
|
|
||||||
class Direct:
|
class Direct:
|
||||||
def __init__(self,directory,POST):
|
def __init__(self,directory,POST):
|
||||||
POST['postExt'] = getExtension(POST['postURL'])
|
POST['postExt'] = getExtension(POST['postURL'])
|
||||||
if not os.path.exists(directory): os.makedirs(directory)
|
if not os.path.exists(directory): os.makedirs(directory)
|
||||||
title = nameCorrector(POST['postTitle'])
|
title = nameCorrector(POST['postTitle'])
|
||||||
|
|
||||||
|
"""Filenames are declared here"""
|
||||||
|
|
||||||
print(POST["postSubmitter"]+"_"+title+"_"+POST['postId']+POST['postExt'])
|
print(POST["postSubmitter"]+"_"+title+"_"+POST['postId']+POST['postExt'])
|
||||||
|
|
||||||
fileDir = directory / (
|
fileDir = directory / (
|
||||||
@@ -459,6 +485,9 @@ class Self:
|
|||||||
if not os.path.exists(directory): os.makedirs(directory)
|
if not os.path.exists(directory): os.makedirs(directory)
|
||||||
|
|
||||||
title = nameCorrector(post['postTitle'])
|
title = nameCorrector(post['postTitle'])
|
||||||
|
|
||||||
|
"""Filenames are declared here"""
|
||||||
|
|
||||||
print(post["postSubmitter"]+"_"+title+"_"+post['postId']+".md")
|
print(post["postSubmitter"]+"_"+title+"_"+post['postId']+".md")
|
||||||
|
|
||||||
fileDir = directory / (
|
fileDir = directory / (
|
||||||
@@ -478,7 +507,8 @@ class Self:
|
|||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def writeToFile(directory,post):
|
def writeToFile(directory,post):
|
||||||
|
|
||||||
|
"""Self posts are formatted here"""
|
||||||
content = ("## ["
|
content = ("## ["
|
||||||
+ post["postTitle"]
|
+ post["postTitle"]
|
||||||
+ "]("
|
+ "]("
|
||||||
|
|||||||
@@ -67,7 +67,7 @@ class NoMatchingSubmissionFound(Exception):
|
|||||||
class NoPrawSupport(Exception):
|
class NoPrawSupport(Exception):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
class NoRedditSupoort(Exception):
|
class NoRedditSupport(Exception):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
class MultiredditNotFound(Exception):
|
class MultiredditNotFound(Exception):
|
||||||
|
|||||||
@@ -29,7 +29,7 @@ def LinkParser(LINK):
|
|||||||
ShortLink = False
|
ShortLink = False
|
||||||
|
|
||||||
if not "reddit.com" in LINK:
|
if not "reddit.com" in LINK:
|
||||||
raise InvalidRedditLink
|
raise InvalidRedditLink("Invalid reddit link")
|
||||||
|
|
||||||
SplittedLink = LINK.split("/")
|
SplittedLink = LINK.split("/")
|
||||||
|
|
||||||
|
|||||||
120
src/searcher.py
120
src/searcher.py
@@ -1,4 +1,5 @@
|
|||||||
import os
|
import os
|
||||||
|
import sys
|
||||||
import random
|
import random
|
||||||
import socket
|
import socket
|
||||||
import webbrowser
|
import webbrowser
|
||||||
@@ -8,7 +9,7 @@ from prawcore.exceptions import NotFound, ResponseException, Forbidden
|
|||||||
|
|
||||||
from src.tools import GLOBAL, createLogFile, jsonFile, printToFile
|
from src.tools import GLOBAL, createLogFile, jsonFile, printToFile
|
||||||
from src.errors import (NoMatchingSubmissionFound, NoPrawSupport,
|
from src.errors import (NoMatchingSubmissionFound, NoPrawSupport,
|
||||||
NoRedditSupoort, MultiredditNotFound,
|
NoRedditSupport, MultiredditNotFound,
|
||||||
InvalidSortingType, RedditLoginFailed,
|
InvalidSortingType, RedditLoginFailed,
|
||||||
InsufficientPermission)
|
InsufficientPermission)
|
||||||
|
|
||||||
@@ -47,6 +48,7 @@ def beginPraw(config,user_agent = str(socket.gethostname())):
|
|||||||
|
|
||||||
self.client = self.recieve_connection()
|
self.client = self.recieve_connection()
|
||||||
data = self.client.recv(1024).decode('utf-8')
|
data = self.client.recv(1024).decode('utf-8')
|
||||||
|
str(data)
|
||||||
param_tokens = data.split(' ', 2)[1].split('?', 1)[1].split('&')
|
param_tokens = data.split(' ', 2)[1].split('?', 1)[1].split('&')
|
||||||
params = {
|
params = {
|
||||||
key: value for (key, value) in [token.split('=') \
|
key: value for (key, value) in [token.split('=') \
|
||||||
@@ -92,6 +94,7 @@ def beginPraw(config,user_agent = str(socket.gethostname())):
|
|||||||
reddit = authorizedInstance[0]
|
reddit = authorizedInstance[0]
|
||||||
refresh_token = authorizedInstance[1]
|
refresh_token = authorizedInstance[1]
|
||||||
jsonFile(GLOBAL.configDirectory / "config.json").add({
|
jsonFile(GLOBAL.configDirectory / "config.json").add({
|
||||||
|
"reddit_username":str(reddit.user.me()),
|
||||||
"reddit_refresh_token":refresh_token
|
"reddit_refresh_token":refresh_token
|
||||||
})
|
})
|
||||||
else:
|
else:
|
||||||
@@ -101,6 +104,7 @@ def beginPraw(config,user_agent = str(socket.gethostname())):
|
|||||||
reddit = authorizedInstance[0]
|
reddit = authorizedInstance[0]
|
||||||
refresh_token = authorizedInstance[1]
|
refresh_token = authorizedInstance[1]
|
||||||
jsonFile(GLOBAL.configDirectory / "config.json").add({
|
jsonFile(GLOBAL.configDirectory / "config.json").add({
|
||||||
|
"reddit_username":str(reddit.user.me()),
|
||||||
"reddit_refresh_token":refresh_token
|
"reddit_refresh_token":refresh_token
|
||||||
})
|
})
|
||||||
return reddit
|
return reddit
|
||||||
@@ -114,7 +118,7 @@ def getPosts(args):
|
|||||||
reddit = beginPraw(config)
|
reddit = beginPraw(config)
|
||||||
|
|
||||||
if args["sort"] == "best":
|
if args["sort"] == "best":
|
||||||
raise NoPrawSupport
|
raise NoPrawSupport("PRAW does not support that")
|
||||||
|
|
||||||
if "subreddit" in args:
|
if "subreddit" in args:
|
||||||
if "search" in args:
|
if "search" in args:
|
||||||
@@ -125,8 +129,6 @@ def getPosts(args):
|
|||||||
if args["user"] == "me":
|
if args["user"] == "me":
|
||||||
args["user"] = str(reddit.user.me())
|
args["user"] = str(reddit.user.me())
|
||||||
|
|
||||||
print("\nGETTING POSTS\n.\n.\n.\n")
|
|
||||||
|
|
||||||
if not "search" in args:
|
if not "search" in args:
|
||||||
if args["sort"] == "top" or args["sort"] == "controversial":
|
if args["sort"] == "top" or args["sort"] == "controversial":
|
||||||
keyword_params = {
|
keyword_params = {
|
||||||
@@ -145,8 +147,8 @@ def getPosts(args):
|
|||||||
}
|
}
|
||||||
|
|
||||||
if "search" in args:
|
if "search" in args:
|
||||||
if args["sort"] in ["hot","rising","controversial"]:
|
if GLOBAL.arguments.sort in ["hot","rising","controversial"]:
|
||||||
raise InvalidSortingType
|
raise InvalidSortingType("Invalid sorting type has given")
|
||||||
|
|
||||||
if "subreddit" in args:
|
if "subreddit" in args:
|
||||||
print (
|
print (
|
||||||
@@ -158,7 +160,7 @@ def getPosts(args):
|
|||||||
sort=args["sort"],
|
sort=args["sort"],
|
||||||
subreddit=args["subreddit"],
|
subreddit=args["subreddit"],
|
||||||
time=args["time"]
|
time=args["time"]
|
||||||
).upper()
|
).upper(),noPrint=True
|
||||||
)
|
)
|
||||||
return redditSearcher(
|
return redditSearcher(
|
||||||
reddit.subreddit(args["subreddit"]).search(
|
reddit.subreddit(args["subreddit"]).search(
|
||||||
@@ -170,23 +172,23 @@ def getPosts(args):
|
|||||||
)
|
)
|
||||||
|
|
||||||
elif "multireddit" in args:
|
elif "multireddit" in args:
|
||||||
raise NoPrawSupport
|
raise NoPrawSupport("PRAW does not support that")
|
||||||
|
|
||||||
elif "user" in args:
|
elif "user" in args:
|
||||||
raise NoPrawSupport
|
raise NoPrawSupport("PRAW does not support that")
|
||||||
|
|
||||||
elif "saved" in args:
|
elif "saved" in args:
|
||||||
raise NoRedditSupoort
|
raise ("Reddit does not support that")
|
||||||
|
|
||||||
if args["sort"] == "relevance":
|
if args["sort"] == "relevance":
|
||||||
raise InvalidSortingType
|
raise InvalidSortingType("Invalid sorting type has given")
|
||||||
|
|
||||||
if "saved" in args:
|
if "saved" in args:
|
||||||
print(
|
print(
|
||||||
"saved posts\nuser:{username}\nlimit={limit}\n".format(
|
"saved posts\nuser:{username}\nlimit={limit}\n".format(
|
||||||
username=reddit.user.me(),
|
username=reddit.user.me(),
|
||||||
limit=args["limit"]
|
limit=args["limit"]
|
||||||
).upper()
|
).upper(),noPrint=True
|
||||||
)
|
)
|
||||||
return redditSearcher(reddit.user.me().saved(limit=args["limit"]))
|
return redditSearcher(reddit.user.me().saved(limit=args["limit"]))
|
||||||
|
|
||||||
@@ -201,7 +203,7 @@ def getPosts(args):
|
|||||||
sort=args["sort"],
|
sort=args["sort"],
|
||||||
subreddit=args["subreddit"],
|
subreddit=args["subreddit"],
|
||||||
time=args["time"]
|
time=args["time"]
|
||||||
).upper()
|
).upper(),noPrint=True
|
||||||
)
|
)
|
||||||
return redditSearcher(
|
return redditSearcher(
|
||||||
getattr(reddit.front,args["sort"]) (**keyword_params)
|
getattr(reddit.front,args["sort"]) (**keyword_params)
|
||||||
@@ -215,7 +217,7 @@ def getPosts(args):
|
|||||||
sort=args["sort"],
|
sort=args["sort"],
|
||||||
subreddit=args["subreddit"],
|
subreddit=args["subreddit"],
|
||||||
time=args["time"]
|
time=args["time"]
|
||||||
).upper()
|
).upper(),noPrint=True
|
||||||
)
|
)
|
||||||
return redditSearcher(
|
return redditSearcher(
|
||||||
getattr(
|
getattr(
|
||||||
@@ -233,7 +235,7 @@ def getPosts(args):
|
|||||||
sort=args["sort"],
|
sort=args["sort"],
|
||||||
multireddit=args["multireddit"],
|
multireddit=args["multireddit"],
|
||||||
time=args["time"]
|
time=args["time"]
|
||||||
).upper()
|
).upper(),noPrint=True
|
||||||
)
|
)
|
||||||
try:
|
try:
|
||||||
return redditSearcher(
|
return redditSearcher(
|
||||||
@@ -244,7 +246,7 @@ def getPosts(args):
|
|||||||
) (**keyword_params)
|
) (**keyword_params)
|
||||||
)
|
)
|
||||||
except NotFound:
|
except NotFound:
|
||||||
raise MultiredditNotFound
|
raise MultiredditNotFound("Multireddit not found")
|
||||||
|
|
||||||
elif "submitted" in args:
|
elif "submitted" in args:
|
||||||
print (
|
print (
|
||||||
@@ -254,7 +256,7 @@ def getPosts(args):
|
|||||||
sort=args["sort"],
|
sort=args["sort"],
|
||||||
user=args["user"],
|
user=args["user"],
|
||||||
time=args["time"]
|
time=args["time"]
|
||||||
).upper()
|
).upper(),noPrint=True
|
||||||
)
|
)
|
||||||
return redditSearcher(
|
return redditSearcher(
|
||||||
getattr(
|
getattr(
|
||||||
@@ -267,17 +269,17 @@ def getPosts(args):
|
|||||||
"upvoted posts of {user}\nlimit: {limit}\n".format(
|
"upvoted posts of {user}\nlimit: {limit}\n".format(
|
||||||
user=args["user"],
|
user=args["user"],
|
||||||
limit=args["limit"]
|
limit=args["limit"]
|
||||||
).upper()
|
).upper(),noPrint=True
|
||||||
)
|
)
|
||||||
try:
|
try:
|
||||||
return redditSearcher(
|
return redditSearcher(
|
||||||
reddit.redditor(args["user"]).upvoted(limit=args["limit"])
|
reddit.redditor(args["user"]).upvoted(limit=args["limit"])
|
||||||
)
|
)
|
||||||
except Forbidden:
|
except Forbidden:
|
||||||
raise InsufficientPermission
|
raise InsufficientPermission("You do not have permission to do that")
|
||||||
|
|
||||||
elif "post" in args:
|
elif "post" in args:
|
||||||
print("post: {post}\n".format(post=args["post"]).upper())
|
print("post: {post}\n".format(post=args["post"]).upper(),noPrint=True)
|
||||||
return redditSearcher(
|
return redditSearcher(
|
||||||
reddit.submission(url=args["post"]),SINGLE_POST=True
|
reddit.submission(url=args["post"]),SINGLE_POST=True
|
||||||
)
|
)
|
||||||
@@ -306,6 +308,8 @@ def redditSearcher(posts,SINGLE_POST=False):
|
|||||||
|
|
||||||
allPosts = {}
|
allPosts = {}
|
||||||
|
|
||||||
|
print("\nGETTING POSTS")
|
||||||
|
if GLOBAL.arguments.verbose: print("\n")
|
||||||
postsFile = createLogFile("POSTS")
|
postsFile = createLogFile("POSTS")
|
||||||
|
|
||||||
if SINGLE_POST:
|
if SINGLE_POST:
|
||||||
@@ -326,53 +330,65 @@ def redditSearcher(posts,SINGLE_POST=False):
|
|||||||
if result is not None:
|
if result is not None:
|
||||||
details = result
|
details = result
|
||||||
orderCount += 1
|
orderCount += 1
|
||||||
printSubmission(submission,subCount,orderCount)
|
if GLOBAL.arguments.verbose:
|
||||||
|
printSubmission(submission,subCount,orderCount)
|
||||||
subList.append(details)
|
subList.append(details)
|
||||||
|
|
||||||
postsFile.add({subCount:[details]})
|
postsFile.add({subCount:[details]})
|
||||||
|
|
||||||
else:
|
else:
|
||||||
for submission in posts:
|
try:
|
||||||
subCount += 1
|
for submission in posts:
|
||||||
|
subCount += 1
|
||||||
|
|
||||||
try:
|
if subCount % 100 == 0 and not GLOBAL.arguments.verbose:
|
||||||
details = {'postId':submission.id,
|
sys.stdout.write("• ")
|
||||||
'postTitle':submission.title,
|
sys.stdout.flush()
|
||||||
'postSubmitter':str(submission.author),
|
|
||||||
'postType':None,
|
|
||||||
'postURL':submission.url,
|
|
||||||
'postSubreddit':submission.subreddit.display_name}
|
|
||||||
except AttributeError:
|
|
||||||
continue
|
|
||||||
|
|
||||||
result = checkIfMatching(submission)
|
if subCount % 1000 == 0:
|
||||||
|
sys.stdout.write("\n"+" "*14)
|
||||||
|
sys.stdout.flush()
|
||||||
|
|
||||||
if result is not None:
|
try:
|
||||||
details = result
|
details = {'postId':submission.id,
|
||||||
orderCount += 1
|
'postTitle':submission.title,
|
||||||
printSubmission(submission,subCount,orderCount)
|
'postSubmitter':str(submission.author),
|
||||||
subList.append(details)
|
'postType':None,
|
||||||
|
'postURL':submission.url,
|
||||||
|
'postSubreddit':submission.subreddit.display_name}
|
||||||
|
except AttributeError:
|
||||||
|
continue
|
||||||
|
|
||||||
allPosts[subCount] = [details]
|
result = checkIfMatching(submission)
|
||||||
|
|
||||||
|
if result is not None:
|
||||||
|
details = result
|
||||||
|
orderCount += 1
|
||||||
|
if GLOBAL.arguments.verbose:
|
||||||
|
printSubmission(submission,subCount,orderCount)
|
||||||
|
subList.append(details)
|
||||||
|
|
||||||
|
allPosts[subCount] = [details]
|
||||||
|
except KeyboardInterrupt:
|
||||||
|
print("\nKeyboardInterrupt",noPrint=True)
|
||||||
|
|
||||||
postsFile.add(allPosts)
|
postsFile.add(allPosts)
|
||||||
|
|
||||||
if not len(subList) == 0:
|
if not len(subList) == 0:
|
||||||
print(
|
if GLOBAL.arguments.NoDownload or GLOBAL.arguments.verbose:
|
||||||
"\nTotal of {} submissions found!\n"\
|
print(
|
||||||
"{} GFYCATs, {} IMGURs, {} EROMEs, {} DIRECTs and {} SELF POSTS\n"
|
f"\n\nTotal of {len(subList)} submissions found!"
|
||||||
.format(
|
|
||||||
len(subList),
|
|
||||||
gfycatCount,
|
|
||||||
imgurCount,
|
|
||||||
eromeCount,
|
|
||||||
directCount,
|
|
||||||
selfCount
|
|
||||||
)
|
)
|
||||||
)
|
print(
|
||||||
|
f"{gfycatCount} GFYCATs, {imgurCount} IMGURs, " \
|
||||||
|
f"{eromeCount} EROMEs, {directCount} DIRECTs " \
|
||||||
|
f"and {selfCount} SELF POSTS",noPrint=True
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
print()
|
||||||
return subList
|
return subList
|
||||||
else:
|
else:
|
||||||
raise NoMatchingSubmissionFound
|
raise NoMatchingSubmissionFound("No matching submission was found")
|
||||||
|
|
||||||
def checkIfMatching(submission):
|
def checkIfMatching(submission):
|
||||||
global gfycatCount
|
global gfycatCount
|
||||||
|
|||||||
20
src/tools.py
20
src/tools.py
@@ -90,7 +90,7 @@ def createLogFile(TITLE):
|
|||||||
|
|
||||||
return FILE
|
return FILE
|
||||||
|
|
||||||
def printToFile(*args, **kwargs):
|
def printToFile(*args, noPrint=False,**kwargs):
|
||||||
"""Print to both CONSOLE and
|
"""Print to both CONSOLE and
|
||||||
CONSOLE LOG file in a folder time stampt in the name
|
CONSOLE LOG file in a folder time stampt in the name
|
||||||
"""
|
"""
|
||||||
@@ -98,15 +98,21 @@ def printToFile(*args, **kwargs):
|
|||||||
TIME = str(time.strftime("%d-%m-%Y_%H-%M-%S",
|
TIME = str(time.strftime("%d-%m-%Y_%H-%M-%S",
|
||||||
time.localtime(GLOBAL.RUN_TIME)))
|
time.localtime(GLOBAL.RUN_TIME)))
|
||||||
folderDirectory = GLOBAL.directory / "LOG_FILES" / TIME
|
folderDirectory = GLOBAL.directory / "LOG_FILES" / TIME
|
||||||
print(*args,**kwargs)
|
|
||||||
|
if not noPrint or \
|
||||||
|
GLOBAL.arguments.verbose or \
|
||||||
|
"file" in kwargs:
|
||||||
|
|
||||||
|
print(*args,**kwargs)
|
||||||
|
|
||||||
if not path.exists(folderDirectory):
|
if not path.exists(folderDirectory):
|
||||||
makedirs(folderDirectory)
|
makedirs(folderDirectory)
|
||||||
|
|
||||||
with io.open(
|
if not "file" in kwargs:
|
||||||
folderDirectory / "CONSOLE_LOG.txt","a",encoding="utf-8"
|
with io.open(
|
||||||
) as FILE:
|
folderDirectory / "CONSOLE_LOG.txt","a",encoding="utf-8"
|
||||||
print(*args, file=FILE, **kwargs)
|
) as FILE:
|
||||||
|
print(*args, file=FILE, **kwargs)
|
||||||
|
|
||||||
def nameCorrector(string):
|
def nameCorrector(string):
|
||||||
"""Swap strange characters from given string
|
"""Swap strange characters from given string
|
||||||
|
|||||||
Reference in New Issue
Block a user