2020-10-10 20:56:28 +05:30
|
|
|
import datetime
|
|
|
|
import glob
|
|
|
|
import json
|
|
|
|
import math
|
|
|
|
import os
|
|
|
|
import random
|
|
|
|
import re
|
|
|
|
import time
|
|
|
|
import urllib
|
2020-11-04 13:57:19 +05:30
|
|
|
import subprocess
|
2020-11-04 02:30:38 +05:30
|
|
|
from multiprocessing import Process
|
2020-10-10 20:56:28 +05:30
|
|
|
from concurrent.futures import as_completed
|
|
|
|
|
|
|
|
import bleach
|
|
|
|
import feedparser
|
|
|
|
import requests
|
|
|
|
from bs4 import BeautifulSoup
|
|
|
|
from flask import Response
|
2020-07-31 15:58:52 +05:30
|
|
|
from flask import render_template, flash, redirect, url_for, request, send_from_directory, Markup
|
2020-10-10 20:56:28 +05:30
|
|
|
from flask_caching import Cache
|
2020-07-13 03:13:36 +05:30
|
|
|
from flask_login import login_user, logout_user, current_user, login_required
|
2020-10-10 20:56:28 +05:30
|
|
|
from numerize import numerize
|
2020-07-14 21:15:29 +05:30
|
|
|
from requests_futures.sessions import FuturesSession
|
2020-08-30 02:05:29 +05:30
|
|
|
from werkzeug.datastructures import Headers
|
2020-10-10 20:56:28 +05:30
|
|
|
from werkzeug.urls import url_parse
|
2020-08-24 16:04:17 +05:30
|
|
|
from werkzeug.utils import secure_filename
|
2020-08-17 14:32:18 +05:30
|
|
|
from youtube_search import YoutubeSearch
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-13 03:13:36 +05:30
|
|
|
from app import app, db
|
2020-10-10 20:56:28 +05:30
|
|
|
from app.forms import LoginForm, RegistrationForm, EmptyForm, SearchForm, ChannelForm
|
|
|
|
from app.models import User, twitterPost, ytPost, Post, youtubeFollow, twitterFollow
|
2020-11-09 20:09:30 +05:30
|
|
|
|
2020-10-11 00:05:03 +05:30
|
|
|
from youtube import comments, utils, channel as ytch, search as yts
|
2020-10-10 20:56:28 +05:30
|
|
|
from youtube import watch as ytwatch
|
2020-10-30 20:46:09 +05:30
|
|
|
from youtube import video as ytvid
|
2020-11-04 02:30:38 +05:30
|
|
|
|
2020-11-09 20:09:30 +05:30
|
|
|
from nitter import feed as nitterfeed
|
|
|
|
from nitter import user as nitteruser
|
|
|
|
|
2020-09-11 14:04:24 +05:30
|
|
|
#########################################
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-09-11 14:04:24 +05:30
|
|
|
#########################################
|
2020-10-01 14:51:53 +05:30
|
|
|
|
|
|
|
cache = Cache(config={'CACHE_TYPE': 'simple'})
|
|
|
|
cache.init_app(app)
|
2020-08-30 04:45:48 +05:30
|
|
|
##########################
|
|
|
|
#### Config variables ####
|
|
|
|
##########################
|
2020-09-04 11:51:35 +05:30
|
|
|
config = json.load(open('yotter-config.json'))
|
|
|
|
##########################
|
|
|
|
#### Config variables ####
|
|
|
|
##########################
|
2020-10-10 20:56:28 +05:30
|
|
|
NITTERINSTANCE = config['nitterInstance'] # Must be https://.../
|
2020-08-30 04:45:48 +05:30
|
|
|
YOUTUBERSS = "https://www.youtube.com/feeds/videos.xml?channel_id="
|
2020-07-13 03:13:36 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-24 16:04:17 +05:30
|
|
|
##########################
|
|
|
|
#### Global variables ####
|
|
|
|
##########################
|
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
#########################
|
|
|
|
#### Twitter Logic ######
|
|
|
|
#########################
|
2020-09-09 19:27:46 +05:30
|
|
|
@app.before_request
|
|
|
|
def before_request():
|
|
|
|
if current_user.is_authenticated:
|
|
|
|
current_user.set_last_seen()
|
|
|
|
db.session.commit()
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-13 03:13:36 +05:30
|
|
|
@app.route('/')
|
|
|
|
@app.route('/index')
|
|
|
|
@login_required
|
2020-10-01 14:51:53 +05:30
|
|
|
@cache.cached(timeout=50, key_prefix='home')
|
2020-07-13 03:13:36 +05:30
|
|
|
def index():
|
2020-09-08 16:38:05 +05:30
|
|
|
return render_template('home.html', config=config)
|
2020-08-26 13:11:34 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-26 13:11:34 +05:30
|
|
|
@app.route('/twitter')
|
2020-10-01 13:17:45 +05:30
|
|
|
@app.route('/twitter/<page>')
|
2020-08-26 13:11:34 +05:30
|
|
|
@login_required
|
2020-10-01 13:17:45 +05:30
|
|
|
def twitter(page=0):
|
2020-08-24 16:04:17 +05:30
|
|
|
followingList = current_user.twitter_following_list()
|
2020-10-01 15:28:17 +05:30
|
|
|
form = EmptyForm()
|
2020-08-24 16:04:17 +05:30
|
|
|
followCount = len(followingList)
|
2020-10-01 15:28:17 +05:30
|
|
|
page = int(page)
|
2020-07-13 03:13:36 +05:30
|
|
|
avatarPath = "img/avatars/1.png"
|
2020-10-01 14:51:53 +05:30
|
|
|
|
2020-11-09 20:09:30 +05:30
|
|
|
followList = []
|
|
|
|
for f in followingList:
|
|
|
|
followList.append(f.username)
|
|
|
|
posts = []
|
2020-10-16 16:47:58 +05:30
|
|
|
|
2020-10-06 21:41:28 +05:30
|
|
|
cache_file = glob.glob("app/cache/{}_*".format(current_user.username))
|
2020-10-10 20:56:28 +05:30
|
|
|
if (len(cache_file) > 0):
|
|
|
|
time_diff = round(time.time() - os.path.getmtime(cache_file[0]))
|
2020-10-10 13:11:14 +05:30
|
|
|
else:
|
|
|
|
time_diff = 999
|
2020-11-09 20:09:30 +05:30
|
|
|
|
2020-10-06 21:41:28 +05:30
|
|
|
# If cache file is more than 1 minute old
|
|
|
|
if page == 0 and time_diff > 60:
|
2020-10-01 14:51:53 +05:30
|
|
|
if cache_file:
|
2020-10-06 21:41:28 +05:30
|
|
|
for f in cache_file:
|
2020-10-10 20:56:28 +05:30
|
|
|
os.remove(f)
|
2020-11-09 20:09:30 +05:30
|
|
|
feed = nitterfeed.get_feed(followList)
|
2020-10-01 14:51:53 +05:30
|
|
|
cache_file = "{u}_{d}.json".format(u=current_user.username, d=time.strftime("%Y%m%d-%H%M%S"))
|
|
|
|
with open("app/cache/{}".format(cache_file), 'w') as fp:
|
|
|
|
json.dump(feed, fp)
|
2020-11-09 20:09:30 +05:30
|
|
|
|
2020-10-06 21:41:28 +05:30
|
|
|
# Else, refresh feed
|
2020-10-10 20:56:28 +05:30
|
|
|
else:
|
2020-10-01 14:51:53 +05:30
|
|
|
try:
|
|
|
|
cache_file = glob.glob("app/cache/{}*".format(current_user.username))[0]
|
|
|
|
with open(cache_file, 'r') as fp:
|
|
|
|
feed = json.load(fp)
|
|
|
|
except:
|
2020-11-09 20:09:30 +05:30
|
|
|
feed = nitterfeed.get_feed(followList)
|
2020-10-01 14:51:53 +05:30
|
|
|
cache_file = "{u}_{d}.json".format(u=current_user.username, d=time.strftime("%Y%m%d-%H%M%S"))
|
|
|
|
with open("app/cache/{}".format(cache_file), 'w') as fp:
|
|
|
|
json.dump(feed, fp)
|
|
|
|
|
2020-10-01 15:28:17 +05:30
|
|
|
posts.extend(feed)
|
2020-10-01 13:39:05 +05:30
|
|
|
# Items range per page
|
2020-10-10 20:56:28 +05:30
|
|
|
page_items = page * 16
|
|
|
|
offset = page_items + 16
|
2020-10-01 13:39:05 +05:30
|
|
|
# Pagination logic
|
2020-10-10 20:56:28 +05:30
|
|
|
init_page = page - 3
|
2020-10-01 13:39:05 +05:30
|
|
|
if init_page < 0:
|
|
|
|
init_page = 0
|
2020-10-10 20:56:28 +05:30
|
|
|
total_pages = page + 5
|
|
|
|
max_pages = int(math.ceil(len(posts) / 10)) # Total number of pages.
|
2020-10-01 13:39:05 +05:30
|
|
|
if total_pages > max_pages:
|
|
|
|
total_pages = max_pages
|
|
|
|
|
|
|
|
# Posts to render
|
2020-10-02 17:23:46 +05:30
|
|
|
if posts and len(posts) > offset:
|
|
|
|
posts = posts[page_items:offset]
|
2020-10-01 13:17:45 +05:30
|
|
|
else:
|
2020-10-01 13:39:05 +05:30
|
|
|
posts = posts[page_items:]
|
2020-11-09 20:09:30 +05:30
|
|
|
return render_template('twitter.html', title='Yotter | Twitter', posts=posts, followedCount=followCount, form=form, config=config,
|
|
|
|
pages=total_pages, init_page=init_page, actual_page=page)
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-13 03:13:36 +05:30
|
|
|
|
2020-07-13 20:06:45 +05:30
|
|
|
@app.route('/savePost/<url>', methods=['POST'])
|
|
|
|
@login_required
|
|
|
|
def savePost(url):
|
2020-07-15 06:33:03 +05:30
|
|
|
savedUrl = url.replace('~', '/')
|
|
|
|
r = requests.get(savedUrl)
|
|
|
|
html = BeautifulSoup(str(r.content), "lxml")
|
2020-10-10 20:56:28 +05:30
|
|
|
post = html.body.find('div', attrs={'class': 'main-tweet'})
|
2020-07-15 06:33:03 +05:30
|
|
|
|
|
|
|
newPost = Post()
|
|
|
|
newPost.url = savedUrl
|
2020-10-10 20:56:28 +05:30
|
|
|
newPost.username = post.find('a', 'username').text.replace("@", "")
|
|
|
|
newPost.body = post.find_all('div', attrs={'class': 'tweet-content'})[0].text.encode('latin1').decode(
|
|
|
|
'unicode_escape').encode('latin1').decode('utf8')
|
|
|
|
newPost.timestamp = post.find_all('p', attrs={'class': 'tweet-published'})[0].text.encode('latin1').decode(
|
|
|
|
'unicode_escape').encode('latin1').decode('utf8')
|
2020-07-15 06:33:03 +05:30
|
|
|
newPost.user_id = current_user.id
|
|
|
|
try:
|
|
|
|
db.session.add(newPost)
|
|
|
|
db.session.commit()
|
|
|
|
except:
|
|
|
|
flash("Post could not be saved. Either it was already saved or there was an error.")
|
2020-08-27 01:43:51 +05:30
|
|
|
return redirect(request.referrer)
|
2020-07-13 20:06:45 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-15 06:33:03 +05:30
|
|
|
@app.route('/saved')
|
|
|
|
@login_required
|
|
|
|
def saved():
|
|
|
|
savedPosts = current_user.saved_posts().all()
|
2020-09-08 16:38:05 +05:30
|
|
|
return render_template('saved.html', title='Saved', savedPosts=savedPosts, config=config)
|
2020-07-15 06:33:03 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-15 06:33:03 +05:30
|
|
|
@app.route('/deleteSaved/<id>', methods=['POST'])
|
|
|
|
@login_required
|
|
|
|
def deleteSaved(id):
|
|
|
|
savedPost = Post.query.filter_by(id=id).first()
|
|
|
|
db.session.delete(savedPost)
|
|
|
|
db.session.commit()
|
|
|
|
return redirect(url_for('saved'))
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-13 03:13:36 +05:30
|
|
|
@app.route('/follow/<username>', methods=['POST'])
|
|
|
|
@login_required
|
|
|
|
def follow(username):
|
|
|
|
form = EmptyForm()
|
|
|
|
if form.validate_on_submit():
|
2020-08-24 19:01:56 +05:30
|
|
|
if followTwitterAccount(username):
|
2020-08-24 16:04:17 +05:30
|
|
|
flash("{} followed!".format(username))
|
|
|
|
return redirect(request.referrer)
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-24 19:01:56 +05:30
|
|
|
def followTwitterAccount(username):
|
2020-08-24 16:04:17 +05:30
|
|
|
if isTwitterUser(username):
|
2020-08-24 19:01:56 +05:30
|
|
|
if not current_user.is_following_tw(username):
|
|
|
|
try:
|
|
|
|
follow = twitterFollow()
|
|
|
|
follow.username = username
|
|
|
|
follow.followers.append(current_user)
|
|
|
|
db.session.add(follow)
|
|
|
|
db.session.commit()
|
|
|
|
return True
|
|
|
|
except:
|
|
|
|
flash("Twitter: Couldn't follow {}. Already followed?".format(username))
|
|
|
|
return False
|
2020-07-13 03:13:36 +05:30
|
|
|
else:
|
2020-08-24 16:04:17 +05:30
|
|
|
flash("Something went wrong... try again")
|
|
|
|
return False
|
2020-07-13 03:13:36 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-13 03:13:36 +05:30
|
|
|
@app.route('/unfollow/<username>', methods=['POST'])
|
|
|
|
@login_required
|
|
|
|
def unfollow(username):
|
|
|
|
form = EmptyForm()
|
|
|
|
if form.validate_on_submit():
|
2020-08-24 16:04:17 +05:30
|
|
|
if twUnfollow(username):
|
|
|
|
flash("{} unfollowed!".format(username))
|
|
|
|
return redirect(request.referrer)
|
2020-07-13 03:13:36 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-24 16:04:17 +05:30
|
|
|
def twUnfollow(username):
|
|
|
|
try:
|
|
|
|
user = twitterFollow.query.filter_by(username=username).first()
|
|
|
|
db.session.delete(user)
|
2020-07-13 05:40:51 +05:30
|
|
|
db.session.commit()
|
2020-08-24 16:04:17 +05:30
|
|
|
except:
|
|
|
|
flash("There was an error unfollowing the user. Try again.")
|
|
|
|
return redirect(request.referrer)
|
2020-07-13 05:40:51 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-13 05:40:51 +05:30
|
|
|
@app.route('/following')
|
|
|
|
@login_required
|
|
|
|
def following():
|
|
|
|
form = EmptyForm()
|
2020-08-24 16:04:17 +05:30
|
|
|
followCount = len(current_user.twitter_following_list())
|
|
|
|
accounts = current_user.twitter_following_list()
|
2020-10-10 20:56:28 +05:30
|
|
|
return render_template('following.html', accounts=accounts, count=followCount, form=form, config=config)
|
|
|
|
|
2020-07-13 05:40:51 +05:30
|
|
|
|
|
|
|
@app.route('/search', methods=['GET', 'POST'])
|
|
|
|
@login_required
|
|
|
|
def search():
|
|
|
|
form = SearchForm()
|
|
|
|
if form.validate_on_submit():
|
|
|
|
user = form.username.data
|
2020-08-27 13:37:48 +05:30
|
|
|
results = twitterUserSearch(user)
|
2020-07-14 22:30:50 +05:30
|
|
|
if results:
|
2020-10-10 20:56:28 +05:30
|
|
|
return render_template('search.html', form=form, results=results, config=config)
|
2020-07-13 05:40:51 +05:30
|
|
|
else:
|
2020-07-14 22:30:50 +05:30
|
|
|
flash("User {} not found...".format(user))
|
2020-08-27 13:37:48 +05:30
|
|
|
return redirect(request.referrer)
|
2020-07-13 05:40:51 +05:30
|
|
|
else:
|
2020-10-10 20:56:28 +05:30
|
|
|
return render_template('search.html', form=form, config=config)
|
|
|
|
|
2020-08-27 15:37:59 +05:30
|
|
|
@app.route('/u/<username>')
|
2020-09-26 23:47:33 +05:30
|
|
|
@app.route('/<username>')
|
2020-11-09 20:09:30 +05:30
|
|
|
@app.route('/<username>/<page>')
|
2020-07-13 03:13:36 +05:30
|
|
|
@login_required
|
2020-11-09 20:09:30 +05:30
|
|
|
def u(username, page=1):
|
|
|
|
page=int(page)
|
2020-09-26 23:48:50 +05:30
|
|
|
if username == "favicon.ico":
|
2020-10-10 20:56:28 +05:30
|
|
|
return redirect(url_for('static', filename='favicons/favicon.ico'))
|
|
|
|
form = EmptyForm()
|
|
|
|
avatarPath = "img/avatars/{}.png".format(str(random.randint(1, 12)))
|
2020-11-09 20:09:30 +05:30
|
|
|
user = nitteruser.get_user_info(username)
|
2020-08-27 01:43:51 +05:30
|
|
|
if not user:
|
2020-08-24 16:04:17 +05:30
|
|
|
flash("This user is not on Twitter.")
|
2020-08-27 01:43:51 +05:30
|
|
|
return redirect(request.referrer)
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-13 03:13:36 +05:30
|
|
|
posts = []
|
2020-11-09 20:09:30 +05:30
|
|
|
tweets=nitteruser.get_tweets(username, page)
|
|
|
|
if tweets == 'Empty feed':
|
|
|
|
posts = False
|
|
|
|
elif tweets == 'Protected feed':
|
|
|
|
posts = 'Protected'
|
|
|
|
else:
|
|
|
|
posts.extend(tweets)
|
|
|
|
|
|
|
|
if page-1 < 0:
|
|
|
|
prev_page = 0
|
|
|
|
else:
|
|
|
|
prev_page = page-1
|
|
|
|
|
|
|
|
if page > 2:
|
|
|
|
page =2
|
2020-08-27 01:43:51 +05:30
|
|
|
|
2020-11-09 20:09:30 +05:30
|
|
|
return render_template('user.html', posts=posts, user=user, form=form, config=config, page=page, prev_page=prev_page)
|
2020-07-13 03:13:36 +05:30
|
|
|
|
2020-10-11 00:05:03 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
#########################
|
|
|
|
#### Youtube Logic ######
|
|
|
|
#########################
|
2020-08-17 01:17:18 +05:30
|
|
|
@app.route('/youtube', methods=['GET', 'POST'])
|
2020-07-31 15:58:52 +05:30
|
|
|
@login_required
|
2020-08-17 01:17:18 +05:30
|
|
|
def youtube():
|
2020-08-24 16:04:17 +05:30
|
|
|
followCount = len(current_user.youtube_following_list())
|
2020-07-31 15:58:52 +05:30
|
|
|
start_time = time.time()
|
|
|
|
ids = current_user.youtube_following_list()
|
2020-08-16 18:50:05 +05:30
|
|
|
videos = getYoutubePosts(ids)
|
2020-07-31 15:58:52 +05:30
|
|
|
if videos:
|
|
|
|
videos.sort(key=lambda x: x.date, reverse=True)
|
2020-08-23 20:57:12 +05:30
|
|
|
print("--- {} seconds fetching youtube feed---".format(time.time() - start_time))
|
2020-10-10 20:56:28 +05:30
|
|
|
return render_template('youtube.html', title="Yotter | Youtube", videos=videos, followCount=followCount,
|
|
|
|
config=config)
|
|
|
|
|
2020-08-24 16:04:17 +05:30
|
|
|
|
|
|
|
@app.route('/ytfollowing', methods=['GET', 'POST'])
|
|
|
|
@login_required
|
|
|
|
def ytfollowing():
|
|
|
|
form = EmptyForm()
|
|
|
|
channelList = current_user.youtube_following_list()
|
|
|
|
channelCount = len(channelList)
|
2020-10-10 20:56:28 +05:30
|
|
|
|
|
|
|
return render_template('ytfollowing.html', form=form, channelList=channelList, channelCount=channelCount,
|
|
|
|
config=config)
|
2020-07-31 15:58:52 +05:30
|
|
|
|
2020-10-05 21:40:29 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
@app.route('/ytsearch', methods=['GET', 'POST'])
|
|
|
|
@login_required
|
|
|
|
def ytsearch():
|
|
|
|
form = ChannelForm()
|
|
|
|
button_form = EmptyForm()
|
2020-10-05 21:40:29 +05:30
|
|
|
query = request.args.get('q', None)
|
|
|
|
sort = request.args.get('s', None)
|
|
|
|
if sort != None:
|
|
|
|
sort = int(sort)
|
|
|
|
else:
|
|
|
|
sort = 0
|
|
|
|
|
|
|
|
page = request.args.get('p', None)
|
|
|
|
if page == None:
|
2020-09-11 14:04:24 +05:30
|
|
|
page = 1
|
2020-10-05 21:40:29 +05:30
|
|
|
|
|
|
|
if query:
|
2020-09-11 14:04:24 +05:30
|
|
|
autocorrect = 1
|
2020-10-10 20:56:28 +05:30
|
|
|
filters = {"time": 0, "type": 0, "duration": 0}
|
2020-10-05 21:40:29 +05:30
|
|
|
results = yts.search_by_terms(query, page, autocorrect, sort, filters)
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
next_page = "/ytsearch?q={q}&s={s}&p={p}".format(q=query, s=sort, p=int(page) + 1)
|
2020-10-05 21:40:29 +05:30
|
|
|
if int(page) == 1:
|
|
|
|
prev_page = "/ytsearch?q={q}&s={s}&p={p}".format(q=query, s=sort, p=1)
|
|
|
|
else:
|
2020-10-10 20:56:28 +05:30
|
|
|
prev_page = "/ytsearch?q={q}&s={s}&p={p}".format(q=query, s=sort, p=int(page) - 1)
|
2020-10-05 21:40:29 +05:30
|
|
|
|
2020-10-10 21:38:49 +05:30
|
|
|
for video in results['videos']:
|
|
|
|
hostname = urllib.parse.urlparse(video['videoThumb']).netloc
|
|
|
|
video['videoThumb'] = video['videoThumb'].replace("https://{}".format(hostname), "") + "&host=" + hostname
|
|
|
|
|
2020-09-27 00:07:44 +05:30
|
|
|
for channel in results['channels']:
|
2020-10-17 16:39:08 +05:30
|
|
|
if config['isInstance']:
|
2020-09-27 00:07:44 +05:30
|
|
|
channel['thumbnail'] = channel['thumbnail'].replace("~", "/")
|
|
|
|
hostName = urllib.parse.urlparse(channel['thumbnail']).netloc
|
2020-10-11 00:05:03 +05:30
|
|
|
channel['thumbnail'] = channel['thumbnail'].replace("https://{}".format(hostName),
|
|
|
|
"") + "?host=" + hostName
|
2020-10-10 20:56:28 +05:30
|
|
|
return render_template('ytsearch.html', form=form, btform=button_form, results=results,
|
|
|
|
restricted=config['restrictPublicUsage'], config=config, npage=next_page,
|
|
|
|
ppage=prev_page)
|
2020-07-31 15:58:52 +05:30
|
|
|
else:
|
2020-09-11 14:10:02 +05:30
|
|
|
return render_template('ytsearch.html', form=form, results=False)
|
2020-07-31 15:58:52 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
@app.route('/ytfollow/<channelId>', methods=['POST'])
|
|
|
|
@login_required
|
|
|
|
def ytfollow(channelId):
|
2020-10-11 00:05:03 +05:30
|
|
|
r = followYoutubeChannel(channelId)
|
2020-08-24 16:04:17 +05:30
|
|
|
return redirect(request.referrer)
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-24 16:04:17 +05:30
|
|
|
def followYoutubeChannel(channelId):
|
|
|
|
try:
|
2020-10-03 22:36:37 +05:30
|
|
|
try:
|
|
|
|
if not current_user.is_following_yt(channelId):
|
2020-11-06 19:28:43 +05:30
|
|
|
channelData = ytch.get_channel_tab_info(channelId, tab='about')
|
|
|
|
if channelData == False:
|
|
|
|
return False
|
2020-10-03 22:36:37 +05:30
|
|
|
follow = youtubeFollow()
|
|
|
|
follow.channelId = channelId
|
2020-11-06 19:28:43 +05:30
|
|
|
follow.channelName = channelData['channel_name']
|
2020-10-03 22:36:37 +05:30
|
|
|
follow.followers.append(current_user)
|
|
|
|
db.session.add(follow)
|
|
|
|
db.session.commit()
|
2020-11-06 19:28:43 +05:30
|
|
|
flash("{} followed!".format(channelData['channel_name']))
|
2020-10-03 22:36:37 +05:30
|
|
|
return True
|
|
|
|
else:
|
|
|
|
return False
|
|
|
|
except Exception as e:
|
|
|
|
print(e)
|
2020-08-24 19:01:56 +05:30
|
|
|
return False
|
2020-10-03 22:36:37 +05:30
|
|
|
except KeyError as ke:
|
|
|
|
print("KeyError: {}:'{}' could not be found".format(ke, channelId))
|
|
|
|
flash("Youtube: ChannelId '{}' is not valid".format(channelId))
|
2020-08-24 16:04:17 +05:30
|
|
|
return False
|
2020-07-31 15:58:52 +05:30
|
|
|
|
2020-10-03 22:36:37 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
@app.route('/ytunfollow/<channelId>', methods=['POST'])
|
|
|
|
@login_required
|
|
|
|
def ytunfollow(channelId):
|
2020-10-11 00:05:03 +05:30
|
|
|
unfollowYoutubeChannel(channelId)
|
2020-08-24 16:04:17 +05:30
|
|
|
return redirect(request.referrer)
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-24 16:04:17 +05:30
|
|
|
def unfollowYoutubeChannel(channelId):
|
2020-07-31 15:58:52 +05:30
|
|
|
try:
|
2020-08-24 16:04:17 +05:30
|
|
|
channel = youtubeFollow.query.filter_by(channelId=channelId).first()
|
2020-09-09 19:27:46 +05:30
|
|
|
name = channel.channelName
|
2020-07-31 15:58:52 +05:30
|
|
|
db.session.delete(channel)
|
|
|
|
db.session.commit()
|
2020-09-09 19:27:46 +05:30
|
|
|
channel = youtubeFollow.query.filter_by(channelId=channelId).first()
|
|
|
|
if channel:
|
|
|
|
db.session.delete(channel)
|
|
|
|
db.session.commit()
|
|
|
|
flash("{} unfollowed!".format(name))
|
2020-07-31 15:58:52 +05:30
|
|
|
except:
|
|
|
|
flash("There was an error unfollowing the user. Try again.")
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-23 13:50:33 +05:30
|
|
|
@app.route('/channel/<id>', methods=['GET'])
|
2020-08-27 15:37:59 +05:30
|
|
|
@app.route('/user/<id>', methods=['GET'])
|
2020-09-12 13:05:13 +05:30
|
|
|
@app.route('/c/<id>', methods=['GET'])
|
2020-08-23 13:50:33 +05:30
|
|
|
@login_required
|
|
|
|
def channel(id):
|
2020-08-23 21:06:00 +05:30
|
|
|
form = ChannelForm()
|
|
|
|
button_form = EmptyForm()
|
2020-08-23 13:50:33 +05:30
|
|
|
|
2020-10-11 00:05:03 +05:30
|
|
|
page = request.args.get('p', None)
|
|
|
|
sort = request.args.get('s', None)
|
|
|
|
if page is None:
|
|
|
|
page = 1
|
|
|
|
if sort is None:
|
|
|
|
sort = 3
|
|
|
|
|
|
|
|
data = ytch.get_channel_tab_info(id, page, sort)
|
2020-09-26 17:37:37 +05:30
|
|
|
|
2020-10-11 00:05:03 +05:30
|
|
|
for video in data['items']:
|
2020-10-17 16:39:08 +05:30
|
|
|
if config['isInstance']:
|
2020-10-11 00:05:03 +05:30
|
|
|
hostName = urllib.parse.urlparse(video['thumbnail'][1:]).netloc
|
|
|
|
video['thumbnail'] = video['thumbnail'].replace("https://{}".format(hostName), "")[1:].replace("hqdefault",
|
2020-11-04 02:30:38 +05:30
|
|
|
"mqdefault") + "&host=" + hostName
|
2020-09-26 17:37:37 +05:30
|
|
|
else:
|
2020-10-11 00:05:03 +05:30
|
|
|
video['thumbnail'] = video['thumbnail'].replace('/', '~')
|
|
|
|
|
2020-10-17 16:39:08 +05:30
|
|
|
if config['isInstance']:
|
2020-10-11 00:05:03 +05:30
|
|
|
hostName = urllib.parse.urlparse(data['avatar'][1:]).netloc
|
|
|
|
data['avatar'] = data['avatar'].replace("https://{}".format(hostName), "")[1:] + "?host=" + hostName
|
|
|
|
else:
|
|
|
|
data['avatar'] = data['avatar'].replace('/', '~')
|
|
|
|
|
|
|
|
next_page = "/channel/{q}?s={s}&p={p}".format(q=id, s=sort, p=int(page) + 1)
|
|
|
|
if int(page) == 1:
|
|
|
|
prev_page = "/channel/{q}?s={s}&p={p}".format(q=id, s=sort, p=1)
|
2020-09-26 17:37:37 +05:30
|
|
|
else:
|
2020-10-11 00:05:03 +05:30
|
|
|
prev_page = "/channel/{q}?s={s}&p={p}".format(q=id, s=sort, p=int(page) - 1)
|
2020-09-26 17:49:03 +05:30
|
|
|
|
2020-10-11 00:05:03 +05:30
|
|
|
return render_template('channel.html', form=form, btform=button_form, data=data,
|
2020-11-04 02:30:38 +05:30
|
|
|
restricted=config['restrictPublicUsage'], config=config, next_page=next_page,
|
|
|
|
prev_page=prev_page)
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-23 13:50:33 +05:30
|
|
|
|
2020-10-05 18:42:02 +05:30
|
|
|
def get_best_urls(urls):
|
|
|
|
'''Gets URLS in youtube format (format_id, url, height) and returns best ones for yotter'''
|
|
|
|
best_formats = ["22", "18", "34", "35", "36", "37", "38", "43", "44", "45", "46"]
|
2020-10-10 20:56:28 +05:30
|
|
|
best_urls = []
|
2020-10-05 18:42:02 +05:30
|
|
|
for url in urls:
|
|
|
|
for f in best_formats:
|
|
|
|
if url['format_id'] == f:
|
2020-10-10 20:56:28 +05:30
|
|
|
best_urls.append(url)
|
2020-10-05 18:42:02 +05:30
|
|
|
return best_urls
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-10-06 21:04:38 +05:30
|
|
|
def get_live_urls(urls):
|
2020-10-10 20:56:28 +05:30
|
|
|
"""Gets URLS in youtube format (format_id, url, height) and returns best ones for yotter"""
|
2020-10-06 21:04:38 +05:30
|
|
|
best_formats = ["91", "92", "93", "94", "95", "96"]
|
2020-10-10 20:56:28 +05:30
|
|
|
best_urls = []
|
2020-10-06 21:04:38 +05:30
|
|
|
for url in urls:
|
|
|
|
for f in best_formats:
|
|
|
|
if url['format_id'] == f:
|
|
|
|
best_urls.append(url)
|
|
|
|
return best_urls
|
|
|
|
|
2020-11-04 02:30:38 +05:30
|
|
|
|
2020-08-17 11:56:56 +05:30
|
|
|
@app.route('/watch', methods=['GET'])
|
2020-07-31 15:58:52 +05:30
|
|
|
@login_required
|
2020-08-17 01:17:18 +05:30
|
|
|
def watch():
|
|
|
|
id = request.args.get('v', None)
|
2020-10-30 20:46:09 +05:30
|
|
|
info = ytvid.get_info(id)
|
2020-11-04 02:30:38 +05:30
|
|
|
|
2020-10-30 20:46:09 +05:30
|
|
|
if info['error'] == False:
|
|
|
|
for format in info['formats']:
|
|
|
|
hostName = urllib.parse.urlparse(format['url']).netloc
|
|
|
|
format['url'] = format['url'].replace("https://{}".format(hostName), "") + "&host=" + hostName
|
2020-11-04 02:30:38 +05:30
|
|
|
|
2020-10-30 20:46:09 +05:30
|
|
|
for format in info['audio_formats']:
|
|
|
|
hostName = urllib.parse.urlparse(format['url']).netloc
|
|
|
|
format['url'] = format['url'].replace("https://{}".format(hostName), "") + "&host=" + hostName
|
|
|
|
|
|
|
|
# Markup description
|
|
|
|
try:
|
2020-11-04 02:30:38 +05:30
|
|
|
info['description'] = Markup(bleach.linkify(info['description'].replace("\n", "<br>"))).replace(
|
|
|
|
'www.youtube.com', config['serverName']).replace('youtube.com', config['serverName']).replace("/join",
|
|
|
|
"")
|
2020-10-30 20:46:09 +05:30
|
|
|
except AttributeError or TypeError:
|
|
|
|
print(info['description'])
|
|
|
|
|
|
|
|
# Get comments
|
|
|
|
if not info['is_live']:
|
|
|
|
videocomments = comments.video_comments(id, sort=0, offset=0, lc='', secret_key='')
|
|
|
|
videocomments = utils.post_process_comments_info(videocomments)
|
|
|
|
if videocomments is not None:
|
|
|
|
videocomments.sort(key=lambda x: x['likes'], reverse=True)
|
|
|
|
else:
|
2020-11-04 02:30:38 +05:30
|
|
|
videocomments = False
|
|
|
|
|
2020-10-30 20:46:09 +05:30
|
|
|
return render_template("video.html", info=info, title=info['title'], config=config,
|
2020-11-04 02:30:38 +05:30
|
|
|
videocomments=videocomments)
|
|
|
|
|
2020-10-30 20:46:09 +05:30
|
|
|
return render_template("video.html", info=info, title='Scheduled Video', config=config)
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
|
2020-08-30 13:35:16 +05:30
|
|
|
def markupString(string):
|
|
|
|
string = string.replace("\n\n", "<br><br>").replace("\n", "<br>")
|
|
|
|
string = bleach.linkify(string)
|
|
|
|
string = string.replace("https://youtube.com/", "")
|
|
|
|
string = string.replace("https://www.youtube.com/", "")
|
|
|
|
string = string.replace("https://twitter.com/", "/u/")
|
2020-09-08 13:36:44 +05:30
|
|
|
return Markup(string)
|
2020-08-30 13:35:16 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-30 16:10:20 +05:30
|
|
|
## PROXY videos through Yotter server to the client.
|
2020-09-07 15:51:23 +05:30
|
|
|
@app.route('/stream/<url>', methods=['GET', 'POST'])
|
2020-08-17 01:17:18 +05:30
|
|
|
@login_required
|
2020-09-07 15:51:23 +05:30
|
|
|
def stream(url):
|
2020-10-10 20:56:28 +05:30
|
|
|
# This function proxies the video stream from GoogleVideo to the client.
|
2020-09-07 15:51:23 +05:30
|
|
|
url = url.replace('YotterSlash', '/')
|
2020-09-08 13:36:44 +05:30
|
|
|
headers = Headers()
|
2020-10-10 20:56:28 +05:30
|
|
|
if (url):
|
2020-09-08 13:43:27 +05:30
|
|
|
s = requests.Session()
|
|
|
|
s.verify = True
|
2020-10-10 20:56:28 +05:30
|
|
|
req = s.get(url, stream=True)
|
2020-09-08 13:36:44 +05:30
|
|
|
headers.add('Range', request.headers['Range'])
|
2020-10-10 20:56:28 +05:30
|
|
|
headers.add('Accept-Ranges', 'bytes')
|
|
|
|
headers.add('Content-Length', str(int(req.headers['Content-Length']) + 1))
|
|
|
|
response = Response(req.iter_content(chunk_size=10 * 1024), mimetype=req.headers['Content-Type'],
|
|
|
|
content_type=req.headers['Content-Type'], direct_passthrough=True, headers=headers)
|
|
|
|
# enable browser file caching with etags
|
|
|
|
response.cache_control.public = True
|
2020-08-30 02:05:29 +05:30
|
|
|
response.cache_control.max_age = int(60000)
|
|
|
|
return response
|
2020-08-17 01:17:18 +05:30
|
|
|
else:
|
2020-08-17 11:56:56 +05:30
|
|
|
flash("Something went wrong loading the video... Try again.")
|
|
|
|
return redirect(url_for('youtube'))
|
2020-08-30 02:05:29 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-09-08 13:36:44 +05:30
|
|
|
def download_file(streamable):
|
|
|
|
with streamable as stream:
|
|
|
|
stream.raise_for_status()
|
|
|
|
for chunk in stream.iter_content(chunk_size=8192):
|
|
|
|
yield chunk
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
#########################
|
|
|
|
#### General Logic ######
|
|
|
|
#########################
|
|
|
|
@app.route('/login', methods=['GET', 'POST'])
|
|
|
|
def login():
|
|
|
|
if current_user.is_authenticated:
|
|
|
|
return redirect(url_for('index'))
|
|
|
|
form = LoginForm()
|
|
|
|
if form.validate_on_submit():
|
|
|
|
user = User.query.filter_by(username=form.username.data).first()
|
|
|
|
if user is None or not user.check_password(form.password.data):
|
|
|
|
flash('Invalid username or password')
|
|
|
|
return redirect(url_for('login'))
|
2020-09-20 13:04:53 +05:30
|
|
|
if user.username == config['admin_user']:
|
|
|
|
user.set_admin_user()
|
|
|
|
db.session.commit()
|
2020-07-31 15:58:52 +05:30
|
|
|
login_user(user, remember=form.remember_me.data)
|
|
|
|
next_page = request.args.get('next')
|
|
|
|
if not next_page or url_parse(next_page).netloc != '':
|
|
|
|
next_page = url_for('index')
|
|
|
|
return redirect(next_page)
|
2020-09-08 16:38:05 +05:30
|
|
|
return render_template('login.html', title='Sign In', form=form, config=config)
|
2020-07-31 15:58:52 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
|
|
|
# Proxy images through server
|
2020-09-04 11:12:01 +05:30
|
|
|
@app.route('/img/<url>', methods=['GET', 'POST'])
|
|
|
|
@login_required
|
|
|
|
def img(url):
|
|
|
|
pic = requests.get(url.replace("~", "/"))
|
2020-10-10 20:56:28 +05:30
|
|
|
return Response(pic, mimetype="image/png")
|
|
|
|
|
2020-09-04 11:12:01 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
@app.route('/logout')
|
|
|
|
def logout():
|
|
|
|
logout_user()
|
|
|
|
return redirect(url_for('index'))
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
@app.route('/settings')
|
|
|
|
@login_required
|
2020-10-01 14:51:53 +05:30
|
|
|
@cache.cached(timeout=50, key_prefix='settings')
|
2020-07-31 15:58:52 +05:30
|
|
|
def settings():
|
2020-09-08 16:38:05 +05:30
|
|
|
active = 0
|
|
|
|
users = db.session.query(User).all()
|
|
|
|
for u in users:
|
|
|
|
if u.last_seen == None:
|
|
|
|
u.set_last_seen()
|
2020-10-10 20:56:28 +05:30
|
|
|
db.session.commit()
|
2020-09-08 16:38:05 +05:30
|
|
|
else:
|
2020-09-08 16:44:55 +05:30
|
|
|
t = datetime.datetime.utcnow() - u.last_seen
|
2020-09-08 03:02:06 +05:30
|
|
|
s = t.total_seconds()
|
2020-10-10 20:56:28 +05:30
|
|
|
m = s / 60
|
2020-09-26 23:36:24 +05:30
|
|
|
if m < 25:
|
2020-10-10 20:56:28 +05:30
|
|
|
active = active + 1
|
2020-09-08 03:02:06 +05:30
|
|
|
|
2020-09-05 16:55:45 +05:30
|
|
|
instanceInfo = {
|
2020-10-10 20:56:28 +05:30
|
|
|
"totalUsers": db.session.query(User).count(),
|
|
|
|
"active": active,
|
2020-09-05 16:55:45 +05:30
|
|
|
}
|
2020-09-20 13:04:53 +05:30
|
|
|
return render_template('settings.html', info=instanceInfo, config=config, admin=current_user.is_admin)
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-09-30 01:48:25 +05:30
|
|
|
'''@app.route('/clear_inactive_users/<phash>')
|
2020-09-20 13:04:53 +05:30
|
|
|
@login_required
|
|
|
|
def clear_inactive_users(phash):
|
|
|
|
ahash = User.query.filter_by(username=config['admin_user']).first().password_hash
|
|
|
|
if phash == ahash:
|
|
|
|
users = db.session.query(User).all()
|
|
|
|
for u in users:
|
|
|
|
if u.username == config['admin_user']:
|
|
|
|
continue
|
|
|
|
t = datetime.datetime.utcnow() - u.last_seen
|
|
|
|
t = math.floor(t.total_seconds())
|
|
|
|
max_old_s = config['max_old_user_days']*86400
|
|
|
|
if t > max_old_s:
|
|
|
|
user = User.query.filter_by(username=u.username).first()
|
|
|
|
print("deleted "+u.username)
|
|
|
|
db.session.delete(user)
|
|
|
|
db.session.commit()
|
|
|
|
else:
|
|
|
|
flash("You must be admin for this action")
|
2020-09-30 01:48:25 +05:30
|
|
|
return redirect(request.referrer)'''
|
2020-07-31 15:58:52 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
@app.route('/export')
|
|
|
|
@login_required
|
2020-10-10 20:56:28 +05:30
|
|
|
# Export data into a JSON file. Later you can import the data.
|
2020-07-31 15:58:52 +05:30
|
|
|
def export():
|
|
|
|
a = exportData()
|
|
|
|
if a:
|
|
|
|
return send_from_directory('.', 'data_export.json', as_attachment=True)
|
|
|
|
else:
|
|
|
|
return redirect(url_for('error/405'))
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
def exportData():
|
2020-08-24 19:01:56 +05:30
|
|
|
twitterFollowing = current_user.twitter_following_list()
|
2020-07-31 15:58:52 +05:30
|
|
|
youtubeFollowing = current_user.youtube_following_list()
|
|
|
|
data = {}
|
|
|
|
data['twitter'] = []
|
|
|
|
data['youtube'] = []
|
|
|
|
|
|
|
|
for f in twitterFollowing:
|
|
|
|
data['twitter'].append({
|
|
|
|
'username': f.username
|
|
|
|
})
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
for f in youtubeFollowing:
|
|
|
|
data['youtube'].append({
|
|
|
|
'channelId': f.channelId
|
|
|
|
})
|
|
|
|
|
|
|
|
try:
|
|
|
|
with open('app/data_export.json', 'w') as outfile:
|
|
|
|
json.dump(data, outfile)
|
|
|
|
return True
|
|
|
|
except:
|
2020-10-10 20:56:28 +05:30
|
|
|
return False
|
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
|
2020-08-24 16:04:17 +05:30
|
|
|
@app.route('/importdata', methods=['GET', 'POST'])
|
|
|
|
@login_required
|
|
|
|
def importdata():
|
|
|
|
if request.method == 'POST':
|
|
|
|
# check if the post request has the file part
|
|
|
|
if 'file' not in request.files:
|
|
|
|
flash('No file part')
|
2020-08-31 16:58:57 +05:30
|
|
|
return redirect(request.referrer)
|
2020-08-24 16:04:17 +05:30
|
|
|
file = request.files['file']
|
|
|
|
# if user does not select file, browser also
|
|
|
|
# submit an empty part without filename
|
|
|
|
if file.filename == '':
|
|
|
|
flash('No selected file')
|
2020-08-31 16:58:57 +05:30
|
|
|
return redirect(request.referrer)
|
2020-10-03 22:36:37 +05:30
|
|
|
else:
|
2020-11-06 19:28:43 +05:30
|
|
|
flash("Data is being imported. You can keep using Yotter.")
|
2020-11-04 02:30:38 +05:30
|
|
|
importdataasync(file)
|
2020-08-31 16:58:57 +05:30
|
|
|
return redirect(request.referrer)
|
2020-08-24 19:01:56 +05:30
|
|
|
|
2020-08-24 16:04:17 +05:30
|
|
|
return redirect(request.referrer)
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-11-04 02:30:38 +05:30
|
|
|
def importdataasync(file):
|
|
|
|
p = Process(target=importdataforprocess, args=(file,))
|
|
|
|
p.start()
|
|
|
|
|
|
|
|
|
|
|
|
def importdataforprocess(file):
|
|
|
|
option = request.form['import_format']
|
|
|
|
if option == 'yotter':
|
|
|
|
importYotterSubscriptions(file)
|
|
|
|
elif option == 'youtube':
|
|
|
|
importYoutubeSubscriptions(file)
|
|
|
|
|
|
|
|
|
2020-09-06 13:57:14 +05:30
|
|
|
@app.route('/deleteme', methods=['GET', 'POST'])
|
|
|
|
@login_required
|
|
|
|
def deleteme():
|
|
|
|
user = User.query.filter_by(username=current_user.username).first()
|
|
|
|
db.session.delete(user)
|
|
|
|
db.session.commit()
|
|
|
|
logout_user()
|
|
|
|
return redirect(url_for('index'))
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-10-03 22:36:37 +05:30
|
|
|
def importYoutubeSubscriptions(file):
|
|
|
|
filename = secure_filename(file.filename)
|
|
|
|
try:
|
|
|
|
data = re.findall('(UC[a-zA-Z0-9_-]{22})|(?<=user/)[a-zA-Z0-9_-]+', file.read().decode('utf-8'))
|
|
|
|
for acc in data:
|
|
|
|
r = followYoutubeChannel(acc)
|
|
|
|
except Exception as e:
|
|
|
|
print(e)
|
|
|
|
flash("File is not valid.")
|
2020-09-06 13:57:14 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-09-04 18:12:35 +05:30
|
|
|
def importYotterSubscriptions(file):
|
2020-08-24 19:01:56 +05:30
|
|
|
filename = secure_filename(file.filename)
|
|
|
|
data = json.load(file)
|
|
|
|
for acc in data['twitter']:
|
|
|
|
r = followTwitterAccount(acc['username'])
|
2020-10-03 22:36:37 +05:30
|
|
|
|
2020-08-24 19:01:56 +05:30
|
|
|
for acc in data['youtube']:
|
|
|
|
r = followYoutubeChannel(acc['channelId'])
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
@app.route('/register', methods=['GET', 'POST'])
|
|
|
|
def register():
|
2020-09-04 21:51:57 +05:30
|
|
|
form = RegistrationForm()
|
2020-07-31 15:58:52 +05:30
|
|
|
if current_user.is_authenticated:
|
|
|
|
return redirect(url_for('index'))
|
|
|
|
|
2020-09-04 20:37:44 +05:30
|
|
|
REGISTRATIONS = True
|
2020-09-04 21:17:41 +05:30
|
|
|
try:
|
|
|
|
count = db.session.query(User).count()
|
|
|
|
if count >= config['maxInstanceUsers'] or config['maxInstanceUsers'] == 0:
|
|
|
|
REGISTRATIONS = False
|
|
|
|
except:
|
|
|
|
REGISTRATIONS = True
|
2020-09-04 21:51:57 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
if form.validate_on_submit():
|
2020-08-27 13:45:20 +05:30
|
|
|
if User.query.filter_by(username=form.username.data).first():
|
|
|
|
flash("This username is taken! Try with another.")
|
|
|
|
return redirect(request.referrer)
|
2020-08-27 14:26:00 +05:30
|
|
|
|
2020-08-27 13:45:20 +05:30
|
|
|
user = User(username=form.username.data)
|
|
|
|
user.set_password(form.password.data)
|
|
|
|
db.session.add(user)
|
|
|
|
db.session.commit()
|
|
|
|
flash('Congratulations, you are now a registered user!')
|
|
|
|
return redirect(url_for('login'))
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-09-08 16:38:05 +05:30
|
|
|
return render_template('register.html', title='Register', registrations=REGISTRATIONS, form=form, config=config)
|
2020-09-26 18:31:12 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-10-11 02:01:33 +05:30
|
|
|
@app.route('/status')
|
|
|
|
def status():
|
2020-09-26 18:31:12 +05:30
|
|
|
count = db.session.query(User).count()
|
|
|
|
if count >= config['maxInstanceUsers'] or config['maxInstanceUsers'] == 0:
|
2020-10-11 02:01:33 +05:30
|
|
|
filen = url_for('static', filename='img/close.png')
|
|
|
|
caniregister = False
|
2020-09-26 18:31:12 +05:30
|
|
|
else:
|
2020-10-11 02:01:33 +05:30
|
|
|
filen = url_for('static', filename='img/open.png')
|
|
|
|
caniregister = True
|
2020-07-31 15:58:52 +05:30
|
|
|
|
2020-11-05 23:45:11 +05:30
|
|
|
try:
|
|
|
|
with open ("version.txt", "r") as versionFile:
|
2020-11-05 23:55:38 +05:30
|
|
|
ver=versionFile.readlines()
|
|
|
|
hsh = ver[0]
|
|
|
|
update= ver[1]
|
|
|
|
tag=ver[2]
|
2020-11-05 23:45:11 +05:30
|
|
|
except:
|
|
|
|
try:
|
2020-11-05 23:59:01 +05:30
|
|
|
tag = str(subprocess.check_output(["git", "describe", "--tags", "--abbrev=0"]).strip())[2:-1]
|
2020-11-05 23:45:11 +05:30
|
|
|
hsh = str(subprocess.check_output(["git", "rev-parse", "--short", "HEAD"]).strip())[2:-1]
|
2020-11-05 23:51:18 +05:30
|
|
|
update = str(subprocess.check_output(["git", "log", "-1", "--format=%cd"]).strip())[2:-7]
|
2020-11-05 23:45:11 +05:30
|
|
|
except:
|
|
|
|
hsh="Unknown"
|
|
|
|
update="Unknown"
|
|
|
|
tag="Unknown"
|
2020-11-04 02:30:38 +05:30
|
|
|
return render_template('status.html', title='STATUS', count=count, max=config['maxInstanceUsers'], file=filen,
|
2020-11-05 23:45:11 +05:30
|
|
|
cani=caniregister, hash=hsh, update=update, tag=tag)
|
2020-11-04 02:30:38 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-31 15:58:52 +05:30
|
|
|
@app.route('/error/<errno>')
|
|
|
|
def error(errno):
|
2020-09-08 16:38:05 +05:30
|
|
|
return render_template('{}.html'.format(str(errno)), config=config)
|
2020-07-31 15:58:52 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-13 03:13:36 +05:30
|
|
|
def getTimeDiff(t):
|
2020-08-27 02:43:54 +05:30
|
|
|
diff = datetime.datetime.now() - datetime.datetime(*t[:6])
|
2020-07-13 03:13:36 +05:30
|
|
|
|
|
|
|
if diff.days == 0:
|
2020-07-13 20:06:45 +05:30
|
|
|
if diff.seconds > 3599:
|
2020-10-10 20:56:28 +05:30
|
|
|
timeString = "{}h".format(int((diff.seconds / 60) / 60))
|
2020-07-13 03:13:36 +05:30
|
|
|
else:
|
2020-10-10 20:56:28 +05:30
|
|
|
timeString = "{}m".format(int(diff.seconds / 60))
|
2020-07-13 03:13:36 +05:30
|
|
|
else:
|
|
|
|
timeString = "{}d".format(diff.days)
|
2020-07-13 05:40:51 +05:30
|
|
|
return timeString
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-13 05:40:51 +05:30
|
|
|
def isTwitterUser(username):
|
2020-08-30 04:45:48 +05:30
|
|
|
response = requests.get('{instance}{user}/rss'.format(instance=NITTERINSTANCE, user=username))
|
2020-08-27 01:43:51 +05:30
|
|
|
if response.status_code == 404:
|
2020-07-13 05:40:51 +05:30
|
|
|
return False
|
2020-07-14 17:54:43 +05:30
|
|
|
return True
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-27 13:37:48 +05:30
|
|
|
def twitterUserSearch(terms):
|
2020-10-10 20:56:28 +05:30
|
|
|
response = urllib.request.urlopen(
|
|
|
|
'{instance}search?f=users&q={user}'.format(instance=NITTERINSTANCE, user=urllib.parse.quote(terms))).read()
|
2020-08-27 13:37:48 +05:30
|
|
|
html = BeautifulSoup(str(response), "lxml")
|
|
|
|
|
|
|
|
results = []
|
2020-10-10 20:56:28 +05:30
|
|
|
if html.body.find('h2', attrs={'class': 'timeline-none'}):
|
2020-08-27 13:37:48 +05:30
|
|
|
return False
|
|
|
|
else:
|
2020-10-10 20:56:28 +05:30
|
|
|
html = html.body.find_all('div', attrs={'class': 'timeline-item'})
|
2020-08-27 13:37:48 +05:30
|
|
|
for item in html:
|
|
|
|
user = {
|
2020-10-10 20:56:28 +05:30
|
|
|
"fullName": item.find('a', attrs={'class': 'fullname'}).getText().encode('latin_1').decode(
|
|
|
|
'unicode_escape').encode('latin_1').decode('utf8'),
|
|
|
|
"username": item.find('a', attrs={'class': 'username'}).getText().encode('latin_1').decode(
|
|
|
|
'unicode_escape').encode('latin_1').decode('utf8'),
|
|
|
|
'avatar': "{i}{s}".format(i=NITTERINSTANCE, s=item.find('img', attrs={'class': 'avatar'})['src'][1:])
|
2020-08-27 13:37:48 +05:30
|
|
|
}
|
|
|
|
results.append(user)
|
|
|
|
return results
|
|
|
|
|
|
|
|
|
2020-08-27 01:43:51 +05:30
|
|
|
def getTwitterUserInfo(username):
|
2020-08-30 04:45:48 +05:30
|
|
|
response = urllib.request.urlopen('{instance}{user}'.format(instance=NITTERINSTANCE, user=username)).read()
|
2020-10-10 20:56:28 +05:30
|
|
|
# rssFeed = feedparser.parse(response.content)
|
2020-08-27 01:43:51 +05:30
|
|
|
|
|
|
|
html = BeautifulSoup(str(response), "lxml")
|
2020-10-10 20:56:28 +05:30
|
|
|
if html.body.find('div', attrs={'class': 'error-panel'}):
|
2020-08-27 01:43:51 +05:30
|
|
|
return False
|
|
|
|
else:
|
2020-10-10 20:56:28 +05:30
|
|
|
html = html.body.find('div', attrs={'class': 'profile-card'})
|
2020-08-27 13:37:48 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
if html.find('a', attrs={'class': 'profile-card-fullname'}):
|
|
|
|
fullName = html.find('a', attrs={'class': 'profile-card-fullname'}).getText().encode('latin1').decode(
|
|
|
|
'unicode_escape').encode('latin1').decode('utf8')
|
2020-08-27 13:37:48 +05:30
|
|
|
else:
|
|
|
|
fullName = None
|
2020-10-10 20:56:28 +05:30
|
|
|
|
|
|
|
if html.find('div', attrs={'class': 'profile-bio'}):
|
|
|
|
profileBio = html.find('div', attrs={'class': 'profile-bio'}).getText().encode('latin1').decode(
|
|
|
|
'unicode_escape').encode('latin1').decode('utf8')
|
2020-08-27 13:37:48 +05:30
|
|
|
else:
|
|
|
|
profileBio = None
|
|
|
|
|
2020-08-27 01:43:51 +05:30
|
|
|
user = {
|
2020-10-10 20:56:28 +05:30
|
|
|
"profileFullName": fullName,
|
|
|
|
"profileUsername": html.find('a', attrs={'class': 'profile-card-username'}).string.encode('latin_1').decode(
|
|
|
|
'unicode_escape').encode('latin_1').decode('utf8'),
|
|
|
|
"profileBio": profileBio,
|
|
|
|
"tweets": html.find_all('span', attrs={'class': 'profile-stat-num'})[0].string,
|
|
|
|
"following": html.find_all('span', attrs={'class': 'profile-stat-num'})[1].string,
|
|
|
|
"followers": numerize.numerize(
|
|
|
|
int(html.find_all('span', attrs={'class': 'profile-stat-num'})[2].string.replace(",", ""))),
|
|
|
|
"likes": html.find_all('span', attrs={'class': 'profile-stat-num'})[3].string,
|
|
|
|
"profilePic": "{instance}{pic}".format(instance=NITTERINSTANCE,
|
|
|
|
pic=html.find('a', attrs={'class': 'profile-card-avatar'})['href'][
|
|
|
|
1:])
|
2020-08-27 01:43:51 +05:30
|
|
|
}
|
|
|
|
return user
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-14 21:15:29 +05:30
|
|
|
def getFeed(urls):
|
|
|
|
feedPosts = []
|
|
|
|
with FuturesSession() as session:
|
2020-08-30 04:45:48 +05:30
|
|
|
futures = [session.get('{instance}{user}'.format(instance=NITTERINSTANCE, user=u.username)) for u in urls]
|
2020-07-14 21:15:29 +05:30
|
|
|
for future in as_completed(futures):
|
2020-11-06 19:28:43 +05:30
|
|
|
res= future.result().content
|
2020-08-28 02:55:08 +05:30
|
|
|
html = BeautifulSoup(res, "html.parser")
|
2020-10-10 20:56:28 +05:30
|
|
|
userFeed = html.find_all('div', attrs={'class': 'timeline-item'})
|
2020-08-28 02:55:08 +05:30
|
|
|
if userFeed != []:
|
2020-10-10 20:56:28 +05:30
|
|
|
for post in userFeed[:-1]:
|
|
|
|
date_time_str = post.find('span', attrs={'class': 'tweet-date'}).find('a')['title'].replace(",", "")
|
|
|
|
time = datetime.datetime.now() - datetime.datetime.strptime(date_time_str, '%d/%m/%Y %H:%M:%S')
|
|
|
|
if time.days >= 7:
|
|
|
|
continue
|
|
|
|
|
|
|
|
if post.find('div', attrs={'class': 'pinned'}):
|
|
|
|
if post.find('div', attrs={'class': 'pinned'}).find('span', attrs={'icon-pin'}):
|
2020-08-27 01:43:51 +05:30
|
|
|
continue
|
2020-08-28 02:55:08 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
newPost = {}
|
|
|
|
newPost["op"] = post.find('a', attrs={'class': 'username'}).text
|
|
|
|
newPost["twitterName"] = post.find('a', attrs={'class': 'fullname'}).text
|
|
|
|
newPost["timeStamp"] = date_time_str
|
|
|
|
newPost["date"] = post.find('span', attrs={'class': 'tweet-date'}).find('a').text
|
2020-11-04 02:30:38 +05:30
|
|
|
content = post.find('div', attrs={'class': 'tweet-content'})
|
2020-11-02 03:05:56 +05:30
|
|
|
newPost["content"] = Markup(str(content).replace("\n", "<br>"))
|
2020-10-10 20:56:28 +05:30
|
|
|
|
|
|
|
if post.find('div', attrs={'class': 'retweet-header'}):
|
|
|
|
newPost["username"] = post.find('div', attrs={'class': 'retweet-header'}).find('div', attrs={
|
|
|
|
'class': 'icon-container'}).text
|
|
|
|
newPost["isRT"] = True
|
|
|
|
else:
|
|
|
|
newPost["username"] = newPost["op"]
|
|
|
|
newPost["isRT"] = False
|
|
|
|
|
|
|
|
newPost["profilePic"] = NITTERINSTANCE + \
|
|
|
|
post.find('a', attrs={'class': 'tweet-avatar'}).find('img')['src'][1:]
|
|
|
|
newPost["url"] = NITTERINSTANCE + post.find('a', attrs={'class': 'tweet-link'})['href'][1:]
|
|
|
|
if post.find('div', attrs={'class': 'quote'}):
|
|
|
|
newPost["isReply"] = True
|
|
|
|
quote = post.find('div', attrs={'class': 'quote'})
|
|
|
|
if quote.find('div', attrs={'class': 'quote-text'}):
|
|
|
|
newPost["replyingTweetContent"] = Markup(quote.find('div', attrs={'class': 'quote-text'}))
|
|
|
|
|
|
|
|
if quote.find('a', attrs={'class': 'still-image'}):
|
|
|
|
newPost["replyAttachedImg"] = NITTERINSTANCE + \
|
|
|
|
quote.find('a', attrs={'class': 'still-image'})['href'][1:]
|
|
|
|
|
|
|
|
if quote.find('div', attrs={'class': 'unavailable-quote'}):
|
|
|
|
newPost["replyingUser"] = "Unavailable"
|
2020-07-14 21:15:29 +05:30
|
|
|
else:
|
2020-10-10 20:56:28 +05:30
|
|
|
try:
|
|
|
|
newPost["replyingUser"] = quote.find('a', attrs={'class': 'username'}).text
|
|
|
|
except:
|
|
|
|
newPost["replyingUser"] = "Unavailable"
|
|
|
|
post.find('div', attrs={'class': 'quote'}).decompose()
|
|
|
|
|
|
|
|
if post.find('div', attrs={'class': 'attachments'}):
|
|
|
|
if not post.find(class_='quote'):
|
|
|
|
if post.find('div', attrs={'class': 'attachments'}).find('a',
|
|
|
|
attrs={'class': 'still-image'}):
|
|
|
|
newPost["attachedImg"] = NITTERINSTANCE + \
|
|
|
|
post.find('div', attrs={'class': 'attachments'}).find('a')[
|
|
|
|
'href'][1:]
|
|
|
|
feedPosts.append(newPost)
|
2020-07-14 21:15:29 +05:30
|
|
|
return feedPosts
|
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-07-14 17:54:43 +05:30
|
|
|
def getPosts(account):
|
2020-09-02 00:33:51 +05:30
|
|
|
feedPosts = []
|
2020-10-10 20:56:28 +05:30
|
|
|
|
|
|
|
# Gather profile info.
|
2020-09-02 00:33:51 +05:30
|
|
|
rssFeed = urllib.request.urlopen('{instance}{user}'.format(instance=NITTERINSTANCE, user=account)).read()
|
2020-10-10 20:56:28 +05:30
|
|
|
# Gather feedPosts
|
2020-09-02 00:33:51 +05:30
|
|
|
res = rssFeed.decode('utf-8')
|
|
|
|
html = BeautifulSoup(res, "html.parser")
|
2020-10-10 20:56:28 +05:30
|
|
|
userFeed = html.find_all('div', attrs={'class': 'timeline-item'})
|
2020-09-02 00:33:51 +05:30
|
|
|
if userFeed != []:
|
2020-10-10 20:56:28 +05:30
|
|
|
for post in userFeed[:-1]:
|
|
|
|
date_time_str = post.find('span', attrs={'class': 'tweet-date'}).find('a')['title'].replace(",", "")
|
2020-09-02 00:33:51 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
if post.find('div', attrs={'class': 'pinned'}):
|
|
|
|
if post.find('div', attrs={'class': 'pinned'}).find('span', attrs={'icon-pin'}):
|
|
|
|
continue
|
2020-09-02 00:33:51 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
newPost = twitterPost()
|
|
|
|
newPost.op = post.find('a', attrs={'class': 'username'}).text
|
|
|
|
newPost.twitterName = post.find('a', attrs={'class': 'fullname'}).text
|
|
|
|
newPost.timeStamp = datetime.datetime.strptime(date_time_str, '%d/%m/%Y %H:%M:%S')
|
|
|
|
newPost.date = post.find('span', attrs={'class': 'tweet-date'}).find('a').text
|
|
|
|
newPost.content = Markup(post.find('div', attrs={'class': 'tweet-content'}))
|
|
|
|
|
|
|
|
if post.find('div', attrs={'class': 'retweet-header'}):
|
|
|
|
newPost.username = post.find('div', attrs={'class': 'retweet-header'}).find('div', attrs={
|
|
|
|
'class': 'icon-container'}).text
|
|
|
|
newPost.isRT = True
|
|
|
|
else:
|
|
|
|
newPost.username = newPost.op
|
|
|
|
newPost.isRT = False
|
|
|
|
|
|
|
|
newPost.profilePic = NITTERINSTANCE + post.find('a', attrs={'class': 'tweet-avatar'}).find('img')['src'][1:]
|
|
|
|
newPost.url = NITTERINSTANCE + post.find('a', attrs={'class': 'tweet-link'})['href'][1:]
|
|
|
|
if post.find('div', attrs={'class': 'quote'}):
|
|
|
|
newPost.isReply = True
|
|
|
|
quote = post.find('div', attrs={'class': 'quote'})
|
|
|
|
if quote.find('div', attrs={'class': 'quote-text'}):
|
|
|
|
newPost.replyingTweetContent = Markup(quote.find('div', attrs={'class': 'quote-text'}))
|
|
|
|
|
|
|
|
if quote.find('a', attrs={'class': 'still-image'}):
|
|
|
|
newPost.replyAttachedImg = NITTERINSTANCE + quote.find('a', attrs={'class': 'still-image'})['href'][
|
|
|
|
1:]
|
|
|
|
|
|
|
|
try:
|
|
|
|
newPost.replyingUser = quote.find('a', attrs={'class': 'username'}).text
|
|
|
|
except:
|
|
|
|
newPost.replyingUser = "Unavailable"
|
|
|
|
post.find('div', attrs={'class': 'quote'}).decompose()
|
|
|
|
|
|
|
|
if post.find('div', attrs={'class': 'attachments'}):
|
|
|
|
if not post.find(class_='quote'):
|
|
|
|
if post.find('div', attrs={'class': 'attachments'}).find('a', attrs={'class': 'still-image'}):
|
|
|
|
newPost.attachedImg = NITTERINSTANCE + \
|
|
|
|
post.find('div', attrs={'class': 'attachments'}).find('a')['href'][1:]
|
|
|
|
feedPosts.append(newPost)
|
2020-09-02 00:33:51 +05:30
|
|
|
return feedPosts
|
2020-07-27 18:30:01 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-16 18:50:05 +05:30
|
|
|
def getYoutubePosts(ids):
|
2020-07-27 18:30:01 +05:30
|
|
|
videos = []
|
|
|
|
with FuturesSession() as session:
|
2020-10-10 20:56:28 +05:30
|
|
|
futures = [session.get('https://www.youtube.com/feeds/videos.xml?channel_id={id}'.format(id=id.channelId)) for
|
|
|
|
id in ids]
|
2020-07-27 18:30:01 +05:30
|
|
|
for future in as_completed(futures):
|
|
|
|
resp = future.result()
|
2020-10-10 20:56:28 +05:30
|
|
|
rssFeed = feedparser.parse(resp.content)
|
2020-07-27 18:30:01 +05:30
|
|
|
for vid in rssFeed.entries:
|
2020-09-09 17:32:45 +05:30
|
|
|
try:
|
2020-09-10 03:55:00 +05:30
|
|
|
# Try to get time diff
|
2020-09-09 17:32:45 +05:30
|
|
|
time = datetime.datetime.now() - datetime.datetime(*vid.published_parsed[:6])
|
|
|
|
except:
|
2020-09-10 05:54:51 +05:30
|
|
|
# If youtube rss does not have parsed time, generate it. Else set time to 0.
|
|
|
|
try:
|
2020-10-10 20:56:28 +05:30
|
|
|
time = datetime.datetime.now() - datetime.datetime(
|
|
|
|
datetime.datetime.strptime(vid.published, '%y-%m-%dT%H:%M:%S+00:00'))
|
2020-09-10 05:54:51 +05:30
|
|
|
except:
|
|
|
|
time = datetime.datetime.now() - datetime.datetime.now()
|
2020-08-27 02:14:18 +05:30
|
|
|
|
2020-10-10 20:56:28 +05:30
|
|
|
if time.days >= 6:
|
2020-08-27 02:14:18 +05:30
|
|
|
continue
|
2020-10-10 20:56:28 +05:30
|
|
|
|
2020-08-16 18:50:05 +05:30
|
|
|
video = ytPost()
|
2020-09-09 17:32:45 +05:30
|
|
|
try:
|
|
|
|
video.date = vid.published_parsed
|
|
|
|
except:
|
2020-09-10 05:54:51 +05:30
|
|
|
try:
|
|
|
|
video.date = datetime.datetime.strptime(vid.published, '%y-%m-%dT%H:%M:%S+00:00').timetuple()
|
|
|
|
except:
|
|
|
|
video.date = datetime.datetime.utcnow().timetuple()
|
2020-09-09 17:32:45 +05:30
|
|
|
try:
|
|
|
|
video.timeStamp = getTimeDiff(vid.published_parsed)
|
|
|
|
except:
|
2020-09-10 05:54:51 +05:30
|
|
|
if time != 0:
|
|
|
|
video.timeStamp = "{} days".format(str(time.days))
|
|
|
|
else:
|
|
|
|
video.timeStamp = "Unknown"
|
2020-09-09 19:27:46 +05:30
|
|
|
|
2020-07-27 18:30:01 +05:30
|
|
|
video.channelName = vid.author_detail.name
|
2020-08-23 20:57:12 +05:30
|
|
|
video.channelId = vid.yt_channelid
|
2020-07-27 18:30:01 +05:30
|
|
|
video.channelUrl = vid.author_detail.href
|
2020-08-16 18:50:05 +05:30
|
|
|
video.id = vid.yt_videoid
|
2020-07-27 18:30:01 +05:30
|
|
|
video.videoTitle = vid.title
|
2020-10-17 16:39:08 +05:30
|
|
|
if config['isInstance']:
|
2020-09-15 00:09:10 +05:30
|
|
|
hostName = urllib.parse.urlparse(vid.media_thumbnail[0]['url']).netloc
|
2020-10-10 20:56:28 +05:30
|
|
|
video.videoThumb = vid.media_thumbnail[0]['url'].replace("https://{}".format(hostName), "").replace(
|
|
|
|
"hqdefault", "mqdefault") + "?host=" + hostName
|
2020-09-15 00:09:10 +05:30
|
|
|
else:
|
|
|
|
video.videoThumb = vid.media_thumbnail[0]['url'].replace('/', '~')
|
2020-07-29 14:34:00 +05:30
|
|
|
video.views = vid.media_statistics['views']
|
2020-08-16 18:50:05 +05:30
|
|
|
video.description = vid.summary_detail.value
|
2020-10-10 20:56:28 +05:30
|
|
|
video.description = re.sub(r'^https?:\/\/.*[\r\n]*', '', video.description[0:120] + "...",
|
|
|
|
flags=re.MULTILINE)
|
2020-07-27 18:30:01 +05:30
|
|
|
videos.append(video)
|
2020-11-06 15:54:59 +05:30
|
|
|
return videos
|