from urllib.parse import urlparse import mistletoe import urllib.parse import gevent from files.helpers.wrappers import * from files.helpers.sanitize import * from files.helpers.filters import * from files.helpers.markdown import * from files.helpers.session import * from files.helpers.thumbs import * from files.helpers.alerts import send_notification from files.helpers.discord import send_message from files.classes import * from flask import * from io import BytesIO from files.__main__ import app, limiter, cache from PIL import Image as PILimage from .front import frontlist site = environ.get("domain").strip() with open("snappy.txt", "r") as f: snappyquotes = f.read().split("{[para]}") @app.post("/publish/") @is_not_banned @validate_formkey def publish(pid, v): post = get_post(pid) if not post.author_id == v.id: abort(403) post.private = False g.db.add(post) cache.delete_memoized(frontlist) return "", 204 @app.get("/submit") @auth_required def submit_get(v): if v and v.is_banned and not v.unban_utc: return render_template("seized.html") return render_template("submit.html", v=v) @app.get("/post/") @app.get("/post//") @auth_desired def post_id(pid, anything=None, v=None): try: pid = int(pid) except Exception as e: pass if v: defaultsortingcomments = v.defaultsortingcomments else: defaultsortingcomments = "top" sort=request.args.get("sort", defaultsortingcomments) try: pid = int(pid) except: try: pid = int(pid, 36) except: abort(404) post = get_post(pid, v=v) if v: votes = g.db.query(CommentVote).filter_by(user_id=v.id).subquery() blocking = v.blocking.subquery() blocked = v.blocked.subquery() comms = g.db.query( Comment, votes.c.vote_type, blocking.c.id, blocked.c.id, ) if v.admin_level >=4: comms=comms.options(joinedload(Comment.oauth_app)) comms=comms.filter( Comment.parent_submission == post.id ).join( votes, votes.c.comment_id == Comment.id, isouter=True ).join( blocking, blocking.c.target_id == Comment.author_id, isouter=True ).join( blocked, blocked.c.user_id == Comment.author_id, isouter=True ) if sort == "top": comments = comms.order_by(Comment.score.desc()).all() elif sort == "bottom": comments = comms.order_by(Comment.score.asc()).all() elif sort == "new": comments = comms.order_by(Comment.created_utc.desc()).all() elif sort == "old": comments = comms.order_by(Comment.created_utc.asc()).all() elif sort == "controversial": comments = sorted(comms.all(), key=lambda x: x[0].score_disputed, reverse=True) elif sort == "random": c = comms.all() comments = random.sample(c, k=len(c)) else: abort(422) output = [] for c in comments: comment = c[0] if comment.author and comment.author.shadowbanned and not (v and v.id == comment.author_id): continue comment._voted = c[1] or 0 comment._is_blocking = c[2] or 0 comment._is_blocked = c[3] or 0 output.append(comment) post._preloaded_comments = output else: comms = g.db.query( Comment ).filter( Comment.parent_submission == post.id ) if sort == "top": comments = comms.order_by(Comment.score.desc()).all() elif sort == "bottom": comments = comms.order_by(Comment.score.asc()).all() elif sort == "new": comments = comms.order_by(Comment.created_utc.desc()).all() elif sort == "old": comments = comms.order_by(Comment.created_utc.asc()).all() elif sort == "controversial": comments = sorted(comms.all(), key=lambda x: x.score_disputed, reverse=True) elif sort == "random": c = comms.all() comments = random.sample(c, k=len(c)) else: abort(422) if random.random() < 0.1: for comment in comments: if comment.author and comment.author.shadowbanned: rand = random.randint(500,1400) vote = CommentVote(user_id=rand, vote_type=random.choice([-1, 1]), comment_id=comment.id) g.db.add(vote) try: g.db.flush() except: g.db.rollback() comment.upvotes = g.db.query(CommentVote).filter_by(comment_id=comment.id, vote_type=1).count() comment.downvotes = g.db.query(CommentVote).filter_by(comment_id=comment.id, vote_type=-1).count() g.db.add(comment) post._preloaded_comments = [x for x in comments if not (x.author and x.author.shadowbanned) or (v and v.id == x.author_id)] post.views += 1 g.db.add(post) g.db.commit() if isinstance(session.get('over_18', 0), dict): session["over_18"] = 0 if post.over_18 and not (v and v.over_18) and not session.get('over_18', 0) >= int(time.time()): if request.headers.get("Authorization"): return {"error":"Must be 18+ to view"}, 451 else: return render_template("errors/nsfw.html", v=v) post.tree_comments() if request.headers.get("Authorization"): return post.json else: return post.rendered_page(v=v, sort=sort) @app.post("/edit_post/") @is_not_banned @validate_formkey def edit_post(pid, v): p = get_post(pid) if not p.author_id == v.id: abort(403) if p.is_banned: abort(403) body = request.form.get("body", "") for i in re.finditer('^(https:\/\/.*\.(png|jpg|jpeg|gif|PNG|JPG|JPEG|GIF))', body, re.MULTILINE): body = body.replace(i.group(1), f'![]({i.group(1)})') with CustomRenderer() as renderer: body_md = renderer.render(mistletoe.Document(body)) body_html = sanitize(body_md, linkgen=True) # Run safety filter bans = filter_comment_html(body_html) if bans: ban = bans[0] reason = f"Remove the {ban.domain} link from your post and try again." if ban.reason: reason += f" {ban.reason}" #auto ban for digitally malicious content if any([x.reason==4 for x in bans]): v.ban(days=30, reason="Digitally malicious content is not allowed.") abort(403) return {"error": reason}, 403 # check spam soup = BeautifulSoup(body_html, features="html.parser") links = [x['href'] for x in soup.find_all('a') if x.get('href')] for link in links: parse_link = urlparse(link) check_url = ParseResult(scheme="https", netloc=parse_link.netloc, path=parse_link.path, params=parse_link.params, query=parse_link.query, fragment='') check_url = urlunparse(check_url) badlink = g.db.query(BadLink).filter( literal(check_url).contains( BadLink.link)).first() if badlink: if badlink.autoban: text = "Your account has been suspended for 1 day for the following reason:\n\n> Too much spam!" send_notification(1046, v, text) v.ban(days=1, reason="spam") return redirect('/notifications') else: return {"error": f"The link `{badlink.link}` is not allowed. Reason: {badlink.reason}"} p.body = body p.body_html = body_html title = request.form.get("title") p.title = title p.title_html = sanitize(title, flair=True) if int(time.time()) - p.created_utc > 60 * 3: p.edited_utc = int(time.time()) g.db.add(p) if v.agendaposter and "trans lives matter" not in body_html.lower(): p.is_banned = True p.ban_reason = "ToS Violation" g.db.add(p) c_jannied = Comment(author_id=2317, parent_submission=p.id, level=1, over_18=False, is_bot=True, app_id=None, is_pinned=True, distinguish_level=6 ) g.db.add(c_jannied) g.db.flush() body = f"""Hi @{v.username},\n\nYour post has been automatically removed because you forgot to include `trans lives matter`.\n\nDon't worry, we're here to help! We won't let you post or comment anything that doesn't express your love and acceptance towards the trans community. Feel free to resubmit your post with `trans lives matter` included. \n\n*This is an automated message; if you need help, you can message us [here](/contact).*""" with CustomRenderer(post_id=p.id) as renderer: body_md = renderer.render(mistletoe.Document(body)) body_jannied_html = sanitize(body_md) c_aux = CommentAux( id=c_jannied.id, body_html=body_jannied_html, body=body ) g.db.add(c_aux) g.db.flush() n = Notification(comment_id=c_jannied.id, user_id=v.id) g.db.add(n) notify_users = set() soup = BeautifulSoup(body_html, features="html.parser") for mention in soup.find_all("a", href=re.compile("^/@(\w+)")): username = mention["href"].split("@")[1] user = g.db.query(User).filter_by(username=username).first() if user and not v.any_block_exists(user) and user.id != v.id: notify_users.add(user) for x in notify_users: send_notification(1046, x, f"@{v.username} has mentioned you: https://{site}{p.permalink}") return redirect(p.permalink) @app.get("/submit/title") @limiter.limit("6/minute") @is_not_banned def get_post_title(v): url = request.args.get("url", None) if not url: return abort(400) #mimic chrome browser agent headers = {"User-Agent": f"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.72 Safari/537.36"} try: x = requests.get(url, headers=headers) except BaseException: return {"error": "Could not reach page"}, 400 if not x.status_code == 200: return {"error": f"Page returned {x.status_code}"}, x.status_code try: soup = BeautifulSoup(x.content, 'html.parser') data = {"url": url, "title": soup.find('title').string } return data except BaseException: return {"error": f"Could not find a title"}, 400 def thumbs(new_post): pid = new_post.id post = get_post(pid, graceful=True, session=g.db) if not post: # account for possible follower lag time.sleep(60) post = get_post(pid, session=g.db) fetch_url=post.url #get the content #mimic chrome browser agent headers={"User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.72 Safari/537.36"} try: x=requests.get(fetch_url, headers=headers) except: return False, "Unable to connect to source" if x.status_code != 200: return False, f"Source returned status {x.status_code}." #if content is image, stick with that. Otherwise, parse html. if x.headers.get("Content-Type","").startswith("text/html"): #parse html, find image, load image soup=BeautifulSoup(x.content, 'html.parser') #parse html #first, set metadata try: meta_title=soup.find('title') if meta_title: post.submission_aux.meta_title=str(meta_title.string)[:500] meta_desc = soup.find('meta', attrs={"name":"description"}) if meta_desc: post.submission_aux.meta_description=meta_desc['content'][:1000] if meta_title or meta_desc: g.db.add(post.submission_aux) g.db.commit() except Exception as e: pass #create list of urls to check thumb_candidate_urls=[] #iterate through desired meta tags meta_tags = [ "twitter:image", "og:image", "thumbnail" ] for tag_name in meta_tags: tag = soup.find( 'meta', attrs={ "name": tag_name, "content": True } ) if not tag: tag = soup.find( 'meta', attrs={ 'property': tag_name, 'content': True } ) if tag: thumb_candidate_urls.append(expand_url(post.url, tag['content'])) #parse html doc for elements for tag in soup.find_all("img", attrs={'src':True}): thumb_candidate_urls.append(expand_url(post.url, tag['src'])) #now we have a list of candidate urls to try for url in thumb_candidate_urls: try: image_req=requests.get(url, headers=headers) except: continue if image_req.status_code >= 400: continue if not image_req.headers.get("Content-Type","").startswith("image/"): continue if image_req.headers.get("Content-Type","").startswith("image/svg"): continue image = PILimage.open(BytesIO(image_req.content)) if image.width < 30 or image.height < 30: continue break else: #getting here means we are out of candidate urls (or there never were any) return False, "No usable images" elif x.headers.get("Content-Type","").startswith("image/"): #image is originally loaded fetch_url image_req=x image = PILimage.open(BytesIO(x.content)) else: print(f'Unknown content type {x.headers.get("Content-Type")}') return False, f'Unknown content type {x.headers.get("Content-Type")} for submitted content' with open("image.gif", "wb") as file: for chunk in image_req.iter_content(1024): file.write(chunk) post.thumburl = upload_file(resize=True) g.db.add(post) g.db.commit() def archiveorg(url): try: requests.get(f'https://web.archive.org/save/{url}', headers={'User-Agent': 'Mozilla/4.0 (compatible; MSIE 5.5; Windows NT)'}, timeout=100) except Exception as e: print(e) @app.post("/submit") @limiter.limit("6/minute") @is_not_banned @validate_formkey def submit_post(v): title = request.form.get("title", "").strip() title = title.strip() title = title.replace("\n", "") title = title.replace("\r", "") title = title.replace("\t", "") url = request.form.get("url", "") if url: repost = g.db.query(Submission).join(Submission.submission_aux).filter( SubmissionAux.url.ilike(url), Submission.deleted_utc == 0, Submission.is_banned == False ).first() else: repost = None if repost: return redirect(repost.permalink) if not title: if request.headers.get("Authorization"): return {"error": "Please enter a better title"}, 400 else: return render_template("submit.html", v=v, error="Please enter a better title.", title=title, url=url, body=request.form.get("body", "")), 400 elif len(title) > 500: if request.headers.get("Authorization"): return {"error": "500 character limit for titles"}, 400 else: render_template("submit.html", v=v, error="500 character limit for titles.", title=title[:500], url=url, body=request.form.get("body", "")), 400 parsed_url = urlparse(url) if not (parsed_url.scheme and parsed_url.netloc) and not request.form.get( "body") and not request.files.get("file", None): if request.headers.get("Authorization"): return {"error": "`url` or `body` parameter required."}, 400 else: return render_template("submit.html", v=v, error="Please enter a url or some text.", title=title, url=url, body=request.form.get("body", "")), 400 # sanitize title title = bleach.clean(title, tags=[]) # Force https for submitted urls if request.form.get("url"): new_url = ParseResult(scheme="https", netloc=parsed_url.netloc, path=parsed_url.path, params=parsed_url.params, query=parsed_url.query, fragment=parsed_url.fragment) url = urlunparse(new_url) else: url = "" if "i.imgur.com" in url: url = url.replace(".png", "_d.png").replace(".jpg", "_d.jpg").replace(".jpeg", "_d.jpeg") + "?maxwidth=8888" body = request.form.get("body", "") # check for duplicate dup = g.db.query(Submission).join(Submission.submission_aux).filter( Submission.author_id == v.id, Submission.deleted_utc == 0, SubmissionAux.title == title, SubmissionAux.url == url, SubmissionAux.body == body ).first() if dup: return redirect(dup.permalink) # check for domain specific rules parsed_url = urlparse(url) domain = parsed_url.netloc # check ban status domain_obj = get_domain(domain) if domain_obj: if domain_obj.reason==4: v.ban(days=30, reason="Digitally malicious content") elif domain_obj.reason==7: v.ban(reason="Sexualizing minors") if request.headers.get("Authorization"): return {"error":"ToS violation"}, 400 else: return render_template("submit.html", v=v, error="ToS Violation", title=title, url=url, body=request.form.get("body", "")), 400 if "twitter.com" in domain: embed = requests.get("https://publish.twitter.com/oembed", params={"url":url, "omit_script":"t"}).json()["html"] elif "youtu" in domain: yt_id = re.match(re.compile("^.*(youtu.be\/|v\/|u\/\w\/|embed\/|watch\?v=|shorts\/|\&v=)([^#\&\?]*).*"), url).group(2) if not yt_id or len(yt_id) != 11: embed = None else: params = parse_qs(urlparse(url).query) t = params.get('t', params.get('start', [0]))[0] if t: embed = f"https://youtube.com/embed/{yt_id}?start={t}" else: embed = f"https://youtube.com/embed/{yt_id}" elif "instagram.com" in domain: embed = requests.get("https://graph.facebook.com/v9.0/instagram_oembed", params={"url":url,"access_token":environ.get("FACEBOOK_TOKEN","").strip(),"omitscript":'true'}, headers={"User-Agent": app.config["UserAgent"]}).json()["html"] elif app.config['SERVER_NAME'] in domain: try: matches = re.match(re.compile(f"^.*{domain}/post/+\w+/(\w+)(/\w+/(\w+))?"), url) post_id = matches.group(1) comment_id = matches.group(3) if comment_id: embed = f"https://{app.config['SERVER_NAME']}/embed/comment/{comment_id}" else: embed = f"https://{app.config['SERVER_NAME']}/embed/post/{post_id}" except: embed = None else: embed = None # similarity check now = int(time.time()) cutoff = now - 60 * 60 * 24 similar_posts = g.db.query(Submission).options( lazyload('*') ).join( Submission.submission_aux ).filter( #or_( # and_( Submission.author_id == v.id, SubmissionAux.title.op('<->')(title) < app.config["SPAM_SIMILARITY_THRESHOLD"], Submission.created_utc > cutoff # ), # and_( # SubmissionAux.title.op('<->')(title) < app.config["SPAM_SIMILARITY_THRESHOLD"]/2, # Submission.created_utc > cutoff # ) #) ).all() if url: similar_urls = g.db.query(Submission).options( lazyload('*') ).join( Submission.submission_aux ).filter( #or_( # and_( Submission.author_id == v.id, SubmissionAux.url.op('<->')(url) < app.config["SPAM_URL_SIMILARITY_THRESHOLD"], Submission.created_utc > cutoff # ), # and_( # SubmissionAux.url.op('<->')(url) < app.config["SPAM_URL_SIMILARITY_THRESHOLD"]/2, # Submission.created_utc > cutoff # ) #) ).all() else: similar_urls = [] threshold = app.config["SPAM_SIMILAR_COUNT_THRESHOLD"] if v.age >= (60 * 60 * 24 * 7): threshold *= 3 elif v.age >= (60 * 60 * 24): threshold *= 2 if max(len(similar_urls), len(similar_posts)) >= threshold: text = "Your account has been suspended for 1 day for the following reason:\n\n> Too much spam!" send_notification(1046, v, text) v.ban(reason="Spamming.", days=1) for alt in v.alts: if not alt.is_suspended: alt.ban(reason="Spamming.", days=1) for post in similar_posts + similar_urls: post.is_banned = True post.is_pinned = False post.ban_reason = "Automatic spam removal. This happened because the post's creator submitted too much similar content too quickly." g.db.add(post) ma=ModAction( user_id=2317, target_submission_id=post.id, kind="ban_post", note="spam" ) g.db.add(ma) g.db.commit() return redirect("/notifications") # catch too-long body if len(str(body)) > 10000: if request.headers.get("Authorization"): return {"error":"10000 character limit for text body."}, 400 else: return render_template("submit.html", v=v, error="10000 character limit for text body.", title=title, url=url, body=request.form.get("body", "")), 400 if len(url) > 2048: if request.headers.get("Authorization"): return {"error":"2048 character limit for URLs."}, 400 else: return render_template("submit.html", v=v, error="2048 character limit for URLs.", title=title, url=url,body=request.form.get("body", "")), 400 # render text for i in re.finditer('^(https:\/\/.*\.(png|jpg|jpeg|gif|PNG|JPG|JPEG|GIF))', body, re.MULTILINE): body = body.replace(i.group(1), f'![]({i.group(1)})') with CustomRenderer() as renderer: body_md = renderer.render(mistletoe.Document(body)) body_html = sanitize(body_md, linkgen=True) # Run safety filter bans = filter_comment_html(body_html) if bans: ban = bans[0] reason = f"Remove the {ban.domain} link from your post and try again." if ban.reason: reason += f" {ban.reason}" #auto ban for digitally malicious content if any([x.reason==4 for x in bans]): v.ban(days=30, reason="Digitally malicious content is not allowed.") abort(403) if request.headers.get("Authorization"): return {"error": reason}, 403 else: return render_template("submit.html", v=v, error=reason, title=title, url=url, body=request.form.get("body", "")), 403 # check spam soup = BeautifulSoup(body_html, features="html.parser") links = [x['href'] for x in soup.find_all('a') if x.get('href')] if url: links = [url] + links for link in links: parse_link = urlparse(link) check_url = ParseResult(scheme="https", netloc=parse_link.netloc, path=parse_link.path, params=parse_link.params, query=parse_link.query, fragment='') check_url = urlunparse(check_url) badlink = g.db.query(BadLink).filter( literal(check_url).contains( BadLink.link)).first() if badlink: if badlink.autoban: text = "Your account has been suspended for 1 day for the following reason:\n\n> Too much spam!" send_notification(1046, v, text) v.ban(days=1, reason="spam") return redirect('/notifications') else: if request.headers.get("Authorization"): return {"error": f"The link `{badlink.link}` is not allowed. Reason: {badlink.reason}"}, 400 else: return render_template("submit.html", v=v, error=f"The link `{badlink.link}` is not allowed. Reason: {badlink.reason}.", title=title, url=url, body=request.form.get("body", "")), 400 # check for embeddable video domain = parsed_url.netloc new_post = Submission( private=bool(request.form.get("private","")), author_id=v.id, over_18=bool(request.form.get("over_18","")), app_id=v.client.application.id if v.client else None, is_bot = request.headers.get("X-User-Type","").lower()=="bot" ) g.db.add(new_post) g.db.flush() for rd in ["https://reddit.com/", "https://new.reddit.com/", "https://www.reddit.com/", "https://redd.it/"]: url = url.replace(rd, "https://old.reddit.com/") url = url.replace("https://mobile.twitter.com", "https://twitter.com") if url.startswith("https://old.reddit.com/") and '/comments/' in url and '?sort=' not in url: url += "?sort=controversial" title_html = sanitize(title, linkgen=True, flair=True) new_post_aux = SubmissionAux(id=new_post.id, url=url, body=body, body_html=body_html, embed_url=embed, title=title, title_html=title_html ) g.db.add(new_post_aux) g.db.flush() vote = Vote(user_id=v.id, vote_type=1, submission_id=new_post.id ) g.db.add(vote) g.db.flush() g.db.refresh(new_post) # check for uploaded image if request.files.get('file') and request.headers.get("cf-ipcountry") != "T1": #check file size if request.content_length > 16 * 1024 * 1024: g.db.rollback() abort(413) file = request.files['file'] if not file.content_type.startswith('image/'): if request.headers.get("Authorization"): return {"error": f"Image files only"}, 400 else: return render_template("submit.html", v=v, error=f"Image files only.", title=title, body=request.form.get("body", "")), 400 new_post.url = upload_file(file) g.db.add(new_post) g.db.add(new_post.submission_aux) g.db.commit() g.db.commit() # spin off thumbnail generation and csam detection as new threads if (new_post.url or request.files.get('file')) and (v.is_activated or request.headers.get('cf-ipcountry')!="T1"): thumbs(new_post) cache.delete_memoized(User.userpagelisting) g.db.commit() notify_users = set() soup = BeautifulSoup(body_html, features="html.parser") for mention in soup.find_all("a", href=re.compile("^/@(\w+)")): username = mention["href"].split("@")[1] user = g.db.query(User).filter_by(username=username).first() if user and not v.any_block_exists(user) and user.id != v.id: notify_users.add(user) for x in notify_users: send_notification(1046, x, f"@{v.username} has mentioned you: https://{site}{new_post.permalink}") if not new_post.private: for follow in v.followers: user = get_account(follow.user_id) send_notification(2360, user, f"@{v.username} has made a new post: [{title}](https://{site}{new_post.permalink})") g.db.add(new_post) g.db.commit() if v.agendaposter and "trans lives matter" not in new_post_aux.body_html.lower(): new_post.is_banned = True new_post.ban_reason = "ToS Violation" g.db.add(new_post) c_jannied = Comment(author_id=2317, parent_submission=new_post.id, level=1, over_18=False, is_bot=True, app_id=None, is_pinned=True, distinguish_level=6 ) g.db.add(c_jannied) g.db.flush() body = f"""Hi @{v.username},\n\nYour post has been automatically removed because you forgot to include `trans lives matter`.\n\nDon't worry, we're here to help! We won't let you post or comment anything that doesn't express your love and acceptance towards the trans community. Feel free to resubmit your post with `trans lives matter` included. \n\n*This is an automated message; if you need help, you can message us [here](/contact).*""" with CustomRenderer(post_id=new_post.id) as renderer: body_md = renderer.render(mistletoe.Document(body)) body_jannied_html = sanitize(body_md) c_aux = CommentAux( id=c_jannied.id, body_html=body_jannied_html, body=body ) g.db.add(c_aux) g.db.flush() n = Notification(comment_id=c_jannied.id, user_id=v.id) g.db.add(n) c = Comment(author_id=261, distinguish_level=6, parent_submission=new_post.id, level=1, over_18=False, is_bot=True, app_id=None, ) g.db.add(c) g.db.flush() if v.id == 995: body = "fuck off carp" else: body = random.choice(snappyquotes) if new_post.url: body += f"\n\n---\n\nSnapshots:\n\n* [reveddit.com](https://reveddit.com/{new_post.url})\n* [archive.org](https://web.archive.org/{new_post.url})\n* [archive.ph](https://archive.ph/?url={urllib.parse.quote(new_post.url)}&run=1) (click to archive)" gevent.spawn(archiveorg, new_post.url) with CustomRenderer(post_id=new_post.id) as renderer: body_md = renderer.render(mistletoe.Document(body)) body_html = sanitize(body_md, linkgen=True) c_aux = CommentAux( id=c.id, body_html=body_html, body=body ) g.db.add(c_aux) g.db.flush() n = Notification(comment_id=c.id, user_id=v.id) g.db.add(n) g.db.commit() send_message(f"https://{site}{new_post.permalink}") v.post_count = v.submissions.filter_by(is_banned=False, deleted_utc=0).count() g.db.add(v) cache.delete_memoized(frontlist) if request.headers.get("Authorization"): return new_post.json else: return redirect(new_post.permalink) @app.post("/delete_post/") @auth_required @validate_formkey def delete_post_pid(pid, v): post = get_post(pid) if not post.author_id == v.id: abort(403) post.deleted_utc = int(time.time()) post.is_pinned = False post.stickied = False g.db.add(post) cache.delete_memoized(frontlist) return "", 204 @app.post("/undelete_post/") @auth_required @validate_formkey def undelete_post_pid(pid, v): post = get_post(pid) if not post.author_id == v.id: abort(403) post.deleted_utc =0 g.db.add(post) cache.delete_memoized(frontlist) return "", 204 @app.post("/toggle_comment_nsfw/") @is_not_banned @validate_formkey def toggle_comment_nsfw(cid, v): comment = g.db.query(Comment).filter_by(id=cid).first() if not comment.author_id == v.id and not v.admin_level >= 3: abort(403) comment.over_18 = not comment.over_18 g.db.add(comment) return "", 204 @app.post("/toggle_post_nsfw/") @is_not_banned @validate_formkey def toggle_post_nsfw(pid, v): post = get_post(pid) if not post.author_id == v.id and not v.admin_level >= 3: abort(403) post.over_18 = not post.over_18 g.db.add(post) if post.author_id!=v.id: ma=ModAction( kind="set_nsfw" if post.over_18 else "unset_nsfw", user_id=v.id, target_submission_id=post.id, ) g.db.add(ma) return "", 204 @app.post("/save_post/") @auth_required @validate_formkey def save_post(pid, v): post=get_post(pid) new_save=SaveRelationship(user_id=v.id, submission_id=post.id, type=1) g.db.add(new_save) try: g.db.flush() except: g.db.rollback() return "", 204 @app.post("/unsave_post/") @auth_required @validate_formkey def unsave_post(pid, v): post=get_post(pid) save=g.db.query(SaveRelationship).filter_by(user_id=v.id, submission_id=post.id, type=1).first() g.db.delete(save) return "", 204