import os import time import html from io import BytesIO from os import path from shutil import copyfile from sys import stdout from urllib.parse import urlparse import random import subprocess import gevent import requests from PIL import Image from files.__main__ import app, cache, limiter from files.classes import * from files.helpers.actions import * from files.helpers.alerts import * from files.helpers.config.const import * from files.helpers.get import * from files.helpers.sharpen import * from files.helpers.regex import * from files.helpers.sanitize import * from files.helpers.settings import get_setting from files.helpers.slots import * from files.helpers.sorting_and_time import * from files.helpers.can_see import * from files.routes.routehelpers import execute_shadowban_viewers_and_voters from files.routes.wrappers import * from .front import frontlist from .users import userpagelisting from files.__main__ import app, limiter, redis_instance def _add_post_view(pid): db = db_session() p = db.query(Post).filter_by(id=pid).options(load_only(Post.views)).one() p.views += 1 db.add(p) try: db.commit() except: db.rollback() db.close() stdout.flush() @app.post("/publish/") @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @is_not_banned def publish(pid, v): p = get_post(pid) if not p.private: return {"message": "Post published!"} if p.author_id != v.id: abort(403) p.private = False p.created_utc = int(time.time()) g.db.add(p) notify_users = NOTIFY_USERS(f'{p.title} {p.body}', v, ghost=p.ghost, log_cost=p, followers_ping=False) if notify_users: cid, text = notif_comment2(p) if notify_users == 'everyone': alert_everyone(cid) else: for x in notify_users: add_notif(cid, x, text, pushnotif_url=p.permalink) cache.delete_memoized(frontlist) cache.delete_memoized(userpagelisting) execute_snappy(p, v) return {"message": "Post has been published successfully!"} @app.get("/submit") @app.get("/h//submit") @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def submit_get(v, sub=None): sub = get_sub_by_name(sub, graceful=True) if request.path.startswith('/h/') and not sub: abort(404) SUBS = [x[0] for x in g.db.query(Sub.name).order_by(Sub.name)] return render_template("submit.html", SUBS=SUBS, v=v, sub=sub) @app.get("/post/") @app.get("/post//") @app.get("/h//post/") @app.get("/h//post//") @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @auth_desired_with_logingate def post_id(pid, v, anything=None, sub=None): p = get_post(pid, v=v) if not can_see(v, p): abort(403) if not g.is_api_or_xhr and p.over_18 and not g.show_over_18: return render_template("errors/nsfw.html", v=v) gevent.spawn(_add_post_view, pid) if p.new: defaultsortingcomments = 'new' elif v: defaultsortingcomments = v.defaultsortingcomments else: defaultsortingcomments = "hot" sort = request.values.get("sort", defaultsortingcomments) if sort == 'saves': sort = defaultsortingcomments if not v: result = cache.get(f'post_{p.id}_{sort}') if result: calc_users() return result if v: execute_shadowban_viewers_and_voters(v, p) # shadowban check is done in sort_objects # output is needed: see comments.py comments, output = get_comments_v_properties(v, None, Comment.parent_post == p.id, Comment.level < 10) if sort == "hot": pinned = [c[0] for c in comments.filter(Comment.stickied != None).order_by(Comment.created_utc.desc())] comments = comments.filter(Comment.stickied == None) comments = comments.filter(Comment.level == 1) comments = sort_objects(sort, comments, Comment) comments = [c[0] for c in comments] else: comments = g.db.query(Comment).filter(Comment.parent_post == p.id) if sort == "hot": pinned = comments.filter(Comment.stickied != None).order_by(Comment.created_utc.desc()).all() comments = comments.filter(Comment.stickied == None) comments = comments.filter(Comment.level == 1) comments = sort_objects(sort, comments, Comment) comments = comments.all() offset = 0 ids = set() threshold = 100 if p.comment_count > threshold+25 and not (v and v.client): comments2 = [] count = 0 if p.created_utc > 1638672040: for comment in comments: comments2.append(comment) ids.add(comment.id) count += g.db.query(Comment).filter_by(parent_post=p.id, top_comment_id=comment.id).count() + 1 if count > threshold: break else: for comment in comments: comments2.append(comment) ids.add(comment.id) count += g.db.query(Comment).filter_by(parent_post=p.id, parent_comment_id=comment.id).count() + 1 if count > 20: break if len(comments) == len(comments2): offset = 0 else: offset = 1 comments = comments2 p.replies = comments if sort == "hot": pinned2 = {} for pin in pinned: if pin.level > 1: pinned2[pin.top_comment] = '' if pin.top_comment in comments: comments.remove(pin.top_comment) else: pinned2[pin] = '' p.replies = list(pinned2.keys()) + p.replies if v and v.client: return p.json template = "post.html" if (p.is_banned or p.author.shadowbanned) \ and not (v and (v.admin_level >= PERMS['POST_COMMENT_MODERATION'] or p.author_id == v.id)): template = "post_banned.html" result = render_template(template, v=v, p=p, ids=list(ids), sort=sort, render_replies=True, offset=offset, sub=p.subr, fart=get_setting('fart_mode')) if not v: cache.set(f'post_{p.id}_{sort}', result, timeout=3600) return result @app.get("/view_more///") @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @auth_desired_with_logingate def view_more(v, pid, sort, offset): p = get_post(pid, v=v) try: ids = set(int(x) for x in request.values.get("ids").split(',')) except: abort(400) if v: # shadowban check is done in sort_objects # output is needed: see comments.py comments, output = get_comments_v_properties(v, None, Comment.parent_post == pid, Comment.stickied == None, Comment.id.notin_(ids), Comment.level < 10) comments = comments.filter(Comment.level == 1) comments = sort_objects(sort, comments, Comment) comments = [c[0] for c in comments] else: comments = g.db.query(Comment).filter( Comment.parent_post == pid, Comment.level == 1, Comment.stickied == None, Comment.id.notin_(ids) ) comments = sort_objects(sort, comments, Comment) comments = comments.offset(offset).all() comments2 = [] count = 0 if p.created_utc > 1638672040: for comment in comments: comments2.append(comment) ids.add(comment.id) count += g.db.query(Comment).filter_by(parent_post=p.id, top_comment_id=comment.id).count() + 1 if count > 100: break else: for comment in comments: comments2.append(comment) ids.add(comment.id) count += g.db.query(Comment).filter_by(parent_post=p.id, parent_comment_id=comment.id).count() + 1 if count > 20: break if len(comments) == len(comments2): offset = 0 else: offset += 1 comments = comments2 return render_template("comments.html", v=v, comments=comments, p=p, ids=list(ids), render_replies=True, pid=pid, sort=sort, offset=offset) @app.get("/more_comments/") @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @auth_desired_with_logingate def more_comments(v, cid): tcid = g.db.query(Comment.top_comment_id).filter_by(id=cid).one_or_none()[0] if v: # shadowban check is done in sort_objects i think # output is needed: see comments.py comments, output = get_comments_v_properties(v, None, Comment.top_comment_id == tcid, Comment.level > 9) comments = comments.filter(Comment.parent_comment_id == cid) comments = [c[0] for c in comments] else: c = get_comment(cid) comments = c.replies(sort=request.values.get('sort')) if comments: p = comments[0].post else: p = None return render_template("comments.html", v=v, comments=comments, p=p, render_replies=True) def expand_url(post_url, fragment_url): if fragment_url.startswith("https://"): return fragment_url elif fragment_url.startswith("https://"): return f"https://{fragment_url.split('https://')[1]}" elif fragment_url.startswith('//'): return f"https:{fragment_url}" elif fragment_url.startswith('/') and '\\' not in fragment_url: parsed_url = urlparse(post_url) return f"https://{parsed_url.netloc}{fragment_url}" else: return f"{post_url}/{fragment_url}" def reddit_s_url_cleaner(url): return normalize_url(requests.get(url, headers=HEADERS, timeout=2, proxies=proxies).url) def surl_and_thumbnail_thread(post_url, post_body, post_body_html, pid, generate_thumb): #s_url dirty = False if post_url and reddit_s_url_regex.fullmatch(post_url): post_url = reddit_s_url_cleaner(post_url) dirty = True if post_body: for i in reddit_s_url_regex.finditer(post_body): old = i.group(0) new = reddit_s_url_cleaner(old) post_body = post_body.replace(old, new) post_body_html = post_body_html.replace(old, new) dirty = True if dirty: db = db_session() p = db.query(Post).filter_by(id=pid).options(load_only(Post.id)).one_or_none() p.url = post_url p.body = post_body p.body_html = post_body_html db.add(p) db.commit() db.close() stdout.flush() #thumbnail if not generate_thumb: return if post_url.startswith('/') and '\\' not in post_url: post_url = f"{SITE_FULL}{post_url}" try: x = requests.get(post_url, headers=HEADERS, timeout=5, proxies=proxies) except: return if x.status_code != 200: return if x.headers.get("Content-Type","").startswith("text/html"): soup = BeautifulSoup(x.content, 'lxml') thumb_candidate_urls = [] for tag_name in ("twitter:image", "og:image", "thumbnail"): tag = soup.find('meta', attrs={"name": tag_name, "content": True}) if not tag: tag = soup.find('meta', attrs={"property": tag_name, "content": True}) if tag: thumb_candidate_urls.append(expand_url(post_url, tag['content'])) for tag in soup.find_all("img", attrs={'src': True}): thumb_candidate_urls.append(expand_url(post_url, tag['src'])) for url in thumb_candidate_urls: try: image_req = requests.get(url, headers=HEADERS, timeout=5, proxies=proxies) except: continue if image_req.status_code >= 400: continue if not image_req.headers.get("Content-Type","").startswith("image/"): continue if image_req.headers.get("Content-Type","").startswith("image/svg"): continue with Image.open(BytesIO(image_req.content)) as i: if i.width < 30 or i.height < 30: continue break else: return elif x.headers.get("Content-Type","").startswith("image/"): image_req = x with Image.open(BytesIO(x.content)) as i: size = len(i.fp.read()) if size > 8 * 1024 * 1024: return else: return name = f'/images/{time.time()}'.replace('.','') + '.webp' with open(name, "wb") as file: for chunk in image_req.iter_content(1024): file.write(chunk) db = db_session() p = db.query(Post).filter_by(id=pid).options(load_only(Post.author_id)).one_or_none() thumburl = process_image(name, None, resize=99, uploader_id=p.author_id, db=db) if thumburl: p.thumburl = thumburl db.add(p) db.commit() db.close() stdout.flush() @app.post("/is_repost") @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def is_repost(v): not_a_repost = {'permalink': ''} if not FEATURES['REPOST_DETECTION']: return not_a_repost url = request.values.get('url') if not url or len(url) < MIN_REPOST_CHECK_URL_LENGTH: abort(400) url = normalize_url(url) url = escape_for_search(url) repost = g.db.query(Post).filter( Post.url.ilike(url), Post.deleted_utc == 0, Post.is_banned == False ).first() if repost: return {'permalink': repost.permalink} else: return not_a_repost @app.post("/submit") @app.post("/h//submit") @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit('20/day', deduct_when=lambda response: response.status_code < 400) @limiter.limit('20/day', deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @is_not_banned def submit_post(v, sub=None): url = request.values.get("url", "").strip() if '\\' in url: abort(400) title = request.values.get("title", "") title = title[:POST_TITLE_LENGTH_LIMIT].strip() body = request.values.get("body", "") body = body[:POST_BODY_LENGTH_LIMIT(g.v)].strip() if not title: abort(400, "Please enter a better title!") sub = request.values.get("sub", "").lower().replace('/h/','').strip() if SITE == 'rdrama.net' and (v.chud == 1 or v.id == 253): sub = 'chudrama' if SITE == 'rdrama.net' and v.id == 10947: sub = 'mnn' if sub == 'changelog': abort(400, "/h/changelog is archived") if sub in {'furry','vampire','racist','femboy','edgy'} and not v.client and not v.house.lower().startswith(sub): abort(400, f"You need to be a member of House {sub.capitalize()} to post in /h/{sub}") if sub and sub != 'none': sub_name = sub.strip().lower() sub = g.db.query(Sub).options(load_only(Sub.name)).filter_by(name=sub_name).one_or_none() if not sub: abort(400, f"/h/{sub_name} not found!") if not can_see(v, sub): if sub.name == 'highrollerclub': abort(403, f"Only {patron}s can post in /h/{sub}") abort(403, f"You're not allowed to post in /h/{sub}") sub = sub.name if v.exiler_username(sub): abort(400, f"You're exiled from /h/{sub}") else: sub = None if not sub and HOLE_REQUIRED: abort(400, f"You must choose a {HOLE_NAME} for your post!") if v.longpost and (len(body) < 280 or ' [](' in body or body.startswith('[](')): abort(400, "You have to type more than 280 characters!") elif v.bird and len(body) > 140: abort(400, "You have to type less than 140 characters!") embed = None if url: url = normalize_url(url) if v.admin_level < PERMS["IGNORE_DOMAIN_BAN"]: for x in g.db.query(BannedDomain): if url.startswith(x.domain): abort(400, f'Remove the banned link "{x.domain}" and try again!\nReason for link ban: "{x.reason}"') domain = tldextract.extract(url).registered_domain if domain == "twitter.com": try: embed = requests.get("https://publish.twitter.com/oembed", params={"url":url, "omit_script":"t"}, headers=HEADERS, timeout=5).json()["html"] embed = embed.replace(' 2048: abort(400, "There's a 2048 character limit for URLs!") body = process_files(request.files, v, body) body = body[:POST_BODY_LENGTH_LIMIT(v)].strip() # process_files() adds content to the body, so we need to re-strip body_for_sanitize = body if v.sharpen: body_for_sanitize = sharpen(body_for_sanitize) flag_notify = (request.values.get("notify", "on") == "on") flag_new = request.values.get("new", False, bool) or 'megathread' in title.lower() flag_over_18 = FEATURES['NSFW_MARKING'] and request.values.get("over_18", False, bool) flag_private = request.values.get("private", False, bool) flag_ghost = request.values.get("ghost", False, bool) and v.can_post_in_ghost_threads if flag_ghost: sub = None if embed and len(embed) > 1500: embed = None if embed: embed = embed.strip() if url and url.startswith(f'{SITE_FULL}/'): url = url.split(SITE_FULL)[1] if url == '': url = None flag_chudded = v.chud and sub != 'chudrama' p = Post( private=flag_private, notify=flag_notify, author_id=v.id, over_18=flag_over_18, new=flag_new, app_id=v.client.application.id if v.client else None, is_bot=(v.client is not None), url=url, body=body, embed=embed, title=title, sub=sub, ghost=flag_ghost, chudded=flag_chudded, rainbowed=bool(v.rainbow), golden=bool(v.king), queened=bool(v.queen), sharpened=bool(v.sharpen), ) title_html = filter_emojis_only(title, count_emojis=True, obj=p) if v.marseyawarded and not marseyaward_title_regex.fullmatch(title_html): abort(400, "You can only type marseys!") p.title_html = title_html body_html = sanitize(body_for_sanitize, count_emojis=True, limit_pings=100, obj=p) if v.marseyawarded and marseyaward_body_regex.search(body_html): abort(400, "You can only type marseys!") if len(body_html) > POST_BODY_HTML_LENGTH_LIMIT: abort(400, "Post body_html too long!") p.body_html = body_html g.db.add(p) g.db.flush() execute_under_siege(v, p, p.body, 'post') process_poll_options(v, p) for text in {p.body, p.title, p.url}: if execute_blackjack(v, p, text, 'post'): break vote = Vote(user_id=v.id, vote_type=1, post_id=p.id, coins=0 ) g.db.add(vote) if request.files.get('file-url') and not g.is_tor: file = request.files['file-url'] if file.content_type.startswith('image/'): name = f'/images/{time.time()}'.replace('.','') + '.webp' file.save(name) p.url = process_image(name, v) name2 = name.replace('.webp', 'r.webp') copyfile(name, name2) p.thumburl = process_image(name2, v, resize=99) elif file.content_type.startswith('video/'): p.url = process_video(file, v) name = f'/images/{time.time()}'.replace('.','') + '.webp' try: subprocess.run(["ffmpeg", "-loglevel", "quiet", "-y", "-i", p.url, "-vf", "scale='iw':-2", "-q:v", "3", "-frames:v", "1", name], check=True, timeout=30) except: if os.path.isfile(name): os.remove(name) else: p.posterurl = name name2 = name.replace('.webp', 'r.webp') copyfile(name, name2) p.thumburl = process_image(name2, v, resize=99) elif file.content_type.startswith('audio/'): p.url = process_audio(file, v) else: abort(415) if not p.private: notify_users = NOTIFY_USERS(f'{title} {body}', v, ghost=p.ghost, log_cost=p, followers_ping=False) if notify_users: cid, text = notif_comment2(p) if notify_users == 'everyone': alert_everyone(cid) else: for x in notify_users: add_notif(cid, x, text, pushnotif_url=p.permalink) if not complies_with_chud(p): p.is_banned = True p.ban_reason = "AutoJanny" body = random.choice(CHUD_MSGS).format(username=v.username, type='post', CHUD_PHRASE=v.chud_phrase) body_jannied_html = sanitize(body) c_jannied = Comment(author_id=AUTOJANNY_ID, parent_post=p.id, level=1, over_18=False, is_bot=True, app_id=None, distinguish_level=6, body=body, body_html=body_jannied_html, ghost=p.ghost ) g.db.add(c_jannied) g.db.flush() p.comment_count += 1 g.db.add(p) c_jannied.top_comment_id = c_jannied.id n = Notification(comment_id=c_jannied.id, user_id=v.id) g.db.add(n) autojanny = g.db.get(User, AUTOJANNY_ID) autojanny.comment_count += 1 g.db.add(autojanny) v.post_count += 1 g.db.add(v) execute_lawlz_actions(v, p) if (SITE == 'rdrama.net' and v.id in {2008, 3336} and not (p.sub and p.subr.stealth)) and p.sub != 'slavshit' and not p.ghost: p.stickied_utc = int(time.time()) + 28800 p.stickied = "AutoJanny" cache.delete_memoized(frontlist) cache.delete_memoized(userpagelisting) if not p.private: execute_snappy(p, v) g.db.flush() #Necessary, do NOT remove generate_thumb = (not p.thumburl and p.url and p.domain != SITE) gevent.spawn(surl_and_thumbnail_thread, p.url, p.body, p.body_html, p.id, generate_thumb) if v.client: return p.json else: p.voted = 1 return {"post_id": p.id, "success": True} @app.post("/delete/post/") @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit(DELETE_EDIT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DELETE_EDIT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def delete_post_pid(pid, v): p = get_post(pid) if p.author_id != v.id: abort(403) if not p.deleted_utc: p.deleted_utc = int(time.time()) p.is_pinned = False p.stickied = None p.stickied_utc = None g.db.add(p) cache.delete_memoized(frontlist) cache.delete_memoized(userpagelisting) v.post_count -= 1 g.db.add(v) for sort in COMMENT_SORTS.keys(): cache.delete(f'post_{p.id}_{sort}') return {"message": "Post deleted!"} @app.post("/undelete_post/") @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def undelete_post_pid(pid, v): p = get_post(pid) if p.author_id != v.id: abort(403) if p.deleted_utc: p.deleted_utc = 0 g.db.add(p) cache.delete_memoized(frontlist) cache.delete_memoized(userpagelisting) v.post_count += 1 g.db.add(v) for sort in COMMENT_SORTS.keys(): cache.delete(f'post_{p.id}_{sort}') return {"message": "Post undeleted!"} @app.post("/mark_post_nsfw/") @feature_required('NSFW_MARKING') @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def mark_post_nsfw(pid, v): p = get_post(pid) if p.author_id != v.id and v.admin_level < PERMS['POST_COMMENT_MODERATION'] and not (p.sub and v.mods(p.sub)): abort(403) if p.over_18 and v.is_permabanned: abort(403) p.over_18 = True g.db.add(p) if p.author_id != v.id: if v.admin_level >= PERMS['POST_COMMENT_MODERATION']: ma = ModAction( kind = "set_nsfw", user_id = v.id, target_post_id = p.id, ) g.db.add(ma) else: ma = SubAction( sub = p.sub, kind = "set_nsfw", user_id = v.id, target_post_id = p.id, ) g.db.add(ma) send_repeatable_notification(p.author_id, f"@{v.username} (a site admin) has marked [{p.title}](/post/{p.id}) as NSFW") return {"message": "Post has been marked as NSFW!"} @app.post("/unmark_post_nsfw/") @feature_required('NSFW_MARKING') @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def unmark_post_nsfw(pid, v): p = get_post(pid) if p.author_id != v.id and v.admin_level < PERMS['POST_COMMENT_MODERATION'] and not (p.sub and v.mods(p.sub)): abort(403) if p.over_18 and v.is_permabanned: abort(403) p.over_18 = False g.db.add(p) if p.author_id != v.id: if v.admin_level >= PERMS['POST_COMMENT_MODERATION']: ma = ModAction( kind = "unset_nsfw", user_id = v.id, target_post_id = p.id, ) g.db.add(ma) else: ma = SubAction( sub = p.sub, kind = "unset_nsfw", user_id = v.id, target_post_id = p.id, ) g.db.add(ma) send_repeatable_notification(p.author_id, f"@{v.username} (a site admin) has unmarked [{p.title}](/post/{p.id}) as NSFW") return {"message": "Post has been unmarked as NSFW!"} @app.post("/save_post/") @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def save_post(pid, v): p = get_post(pid) save = g.db.query(SaveRelationship).filter_by(user_id=v.id, post_id=p.id).one_or_none() if not save: new_save=SaveRelationship(user_id=v.id, post_id=p.id) g.db.add(new_save) cache.delete_memoized(userpagelisting) return {"message": "Post saved!"} @app.post("/unsave_post/") @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def unsave_post(pid, v): p = get_post(pid) save = g.db.query(SaveRelationship).filter_by(user_id=v.id, post_id=p.id).one_or_none() if save: g.db.delete(save) cache.delete_memoized(userpagelisting) return {"message": "Post unsaved!"} @app.post("/pin/") @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def pin_post(post_id, v): p = get_post(post_id) if p: if v.id != p.author_id: abort(403, "Only the post author can do that!") p.is_pinned = not p.is_pinned g.db.add(p) cache.delete_memoized(userpagelisting) if p.is_pinned: return {"message": "Post pinned!"} else: return {"message": "Post unpinned!"} return abort(404, "Post not found!") @app.post("/post//new") @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def set_new_sort(post_id, v): p = get_post(post_id) if not v.can_edit(p): abort(403, "Only the post author can do that!") p.new = True g.db.add(p) if v.id != p.author_id: ma = ModAction( kind = "set_new", user_id = v.id, target_post_id = p.id, ) g.db.add(ma) send_repeatable_notification(p.author_id, f"@{v.username} (a site admin) has changed the the default sorting of comments on [{p.title}](/post/{p.id}) to `new`") return {"message": "Changed the the default sorting of comments on this post to 'new'"} @app.post("/post//hot") @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DEFAULT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def unset_new_sort(post_id, v): p = get_post(post_id) if not v.can_edit(p): abort(403, "Only the post author can do that!") p.new = None g.db.add(p) if v.id != p.author_id: ma = ModAction( kind = "set_hot", user_id = v.id, target_post_id = p.id, ) g.db.add(ma) send_repeatable_notification(p.author_id, f"@{v.username} (a site admin) has changed the the default sorting of comments on [{p.title}](/post/{p.id}) to `hot`") return {"message": "Changed the the default sorting of comments on this post to 'hot'"} extensions = IMAGE_FORMATS + VIDEO_FORMATS + AUDIO_FORMATS @app.get("/submit/title") @limiter.limit("3/minute", deduct_when=lambda response: response.status_code < 400) @limiter.limit("3/minute", deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def get_post_title(v): POST_TITLE_TIMEOUT = 5 url = request.values.get("url") if not url or '\\' in url: abort(400) url = url.strip() if not url.startswith('http'): abort(400) checking_url = url.lower().split('?')[0].split('%3F')[0] if any((checking_url.endswith(f'.{x}') for x in extensions)): abort(400) try: x = gevent.with_timeout(POST_TITLE_TIMEOUT, requests.get, url, headers=HEADERS, timeout=POST_TITLE_TIMEOUT, proxies=proxies) except: abort(400) content_type = x.headers.get("Content-Type") if not content_type or "text/html" not in content_type: abort(400) # no you can't just parse html with reeeeeeeegex match = html_title_regex.search(x.text) if match and match.lastindex >= 1: title = match.group(1) else: abort(400) title = html.unescape(title) return {"url": url, "title": title} @app.post("/edit_post/") @limiter.limit('1/second', scope=rpath) @limiter.limit('1/second', scope=rpath, key_func=get_ID) @limiter.limit(DELETE_EDIT_RATELIMIT, deduct_when=lambda response: response.status_code < 400) @limiter.limit(DELETE_EDIT_RATELIMIT, deduct_when=lambda response: response.status_code < 400, key_func=get_ID) @auth_required def edit_post(pid, v): p = get_post(pid) if not v.can_edit(p): abort(403) # Disable edits on things older than 1wk unless it's a draft or editor is a jannie if time.time() - p.created_utc > 7*24*60*60 and not p.private \ and v.admin_level < PERMS["IGNORE_1WEEk_EDITING_LIMIT"] and v.id not in EXEMPT_FROM_1WEEK_EDITING_LIMIT: abort(403, "You can't edit posts older than 1 week!") title = request.values.get("title", "") title = title[:POST_TITLE_LENGTH_LIMIT].strip() body = request.values.get("body", "") body = body[:POST_BODY_LENGTH_LIMIT(g.v)].strip() if p.author.longpost and (len(body) < 280 or ' [](' in body or body.startswith('[](')): abort(403, "You have to type more than 280 characters!") elif p.author.bird and len(body) > 140: abort(403, "You have to type less than 140 characters!") if not title: abort(400, "Please enter a better title!") if not p.private: notify_users = NOTIFY_USERS(f'{title} {body}', v, oldtext=f'{p.title} {p.body}', ghost=p.ghost, log_cost=p, followers_ping=False) if notify_users: cid, text = notif_comment2(p) if notify_users == 'everyone': alert_everyone(cid) else: for x in notify_users: add_notif(cid, x, text, pushnotif_url=p.permalink) if title != p.title: title_html = filter_emojis_only(title, golden=False, obj=p) if p.author.marseyawarded and not marseyaward_title_regex.fullmatch(title_html): abort(403, "You can only type marseys!") if 'megathread' in title.lower() and 'megathread' not in p.title.lower(): p.new = True p.title = title p.title_html = title_html body = process_files(request.files, v, body) body = body[:POST_BODY_LENGTH_LIMIT(v)].strip() # process_files() may be adding stuff to the body if body != p.body: body_for_sanitize = body if p.sharpened: body_for_sanitize = sharpen(body_for_sanitize) body_html = sanitize(body_for_sanitize, golden=False, limit_pings=100, obj=p) if p.author.marseyawarded and marseyaward_body_regex.search(body_html): abort(403, "You can only type marseys!") p.body = body for text in [p.body, p.title, p.url]: if execute_blackjack(v, p, text, 'post'): break if len(body_html) > POST_BODY_HTML_LENGTH_LIMIT: abort(400, "Post body_html too long!") p.body_html = body_html process_poll_options(v, p) gevent.spawn(surl_and_thumbnail_thread, p.url, p.body, p.body_html, p.id, False) if not complies_with_chud(p): abort(403, f'You have to include "{p.author.chud_phrase}" in your post!') if v.id == p.author_id: if int(time.time()) - p.created_utc > 60 * 3: p.edited_utc = int(time.time()) else: ma=ModAction( kind="edit_post", user_id=v.id, target_post_id=p.id ) g.db.add(ma) return {"message": "Post edited successfully!"}