diff --git a/files/helpers/owoify.py b/files/helpers/owoify.py index 665b44c15..3c763c618 100644 --- a/files/helpers/owoify.py +++ b/files/helpers/owoify.py @@ -1,38 +1,33 @@ -import re - from owoify.structures.word import Word from owoify.utility.interleave_arrays import interleave_arrays from owoify.utility.presets import * -import files.helpers.regex as help_re -import files.helpers.sanitize as sanitize +from files.helpers.regex import * # Includes, excerpts, and modifies some functions from: # https://github.com/deadshot465/owoify-py @ owoify/owoify.py -OWO_WORD_REGEX = re.compile(r'[^\s]+') -OWO_SPACE_REGEX = re.compile(r'\s+') OWO_EXCLUDE_PATTERNS = [ - re.compile(r'\]\('), # links []() and images ![]() + owo_ignore_links_images_regex, # links []() and images ![]() # NB: May not be effective when URL part contains literal spaces vs %20 # Also relies on owoify replacements currently not affecting symbols. - sanitize.url_re, # bare links - re.compile(r':[!#@a-z0-9_\-]+:', flags=re.I|re.A), # emoji - help_re.mention_regex, # mentions - help_re.group_mention_regex, #ping group mentions - help_re.poll_regex, # polls - help_re.choice_regex, - help_re.command_regex, # markup commands - re.compile(r'\bthe\b', flags=re.I|re.A), # exclude: 'the' ↦ 'teh' + owo_ignore_emojis_regex, #emojis + owo_ignore_the_Regex, # exclude: 'the' ↦ 'teh' + sanitize_url_regex, # bare links + mention_regex, # mentions + group_mention_regex, #ping group mentions + poll_regex, # polls + choice_regex, + command_regex, # markup commands ] def owoify(source): if '`' in source or '
' in source or '' in source:
 		return source
 
-	word_matches = OWO_WORD_REGEX.findall(source)
-	space_matches = OWO_SPACE_REGEX.findall(source)
+	word_matches = owo_word_regex.findall(source)
+	space_matches = owo_space_regex.findall(source)
 
 	words = [Word(s) for s in word_matches]
 	spaces = [Word(s) for s in space_matches]
diff --git a/files/helpers/regex.py b/files/helpers/regex.py
index ac9062f5f..8bfa16763 100644
--- a/files/helpers/regex.py
+++ b/files/helpers/regex.py
@@ -177,6 +177,59 @@ search_regex_1 = re.compile(r'[\0():|&*!<>]', flags=re.A)
 search_regex_2 = re.compile(r"'", flags=re.A)
 search_regex_3 = re.compile(r'\s+', flags=re.A)
 
+###OWOIFY
+
+owo_word_regex = re.compile(r'[^\s]+', flags=re.A)
+owo_space_regex = re.compile(r'\s+', flags=re.A)
+owo_ignore_links_images_regex = re.compile(r'\]\(', flags=re.A)
+owo_ignore_emojis_regex = re.compile(r':[!#@a-z0-9_\-]+:', flags=re.I|re.A)
+owo_ignore_the_Regex = re.compile(r'\bthe\b', flags=re.I|re.A)
+
+
+###LinkifyFilter
+
+tlds = ( # Original gTLDs and ccTLDs
+	'ac','ad','ae','aero','af','ag','ai','al','am','an','ao','aq','ar','arpa','as','asia','at',
+	'au','aw','ax','az','ba','bb','bd','be','bf','bg','bh','bi','biz','bj','bm','bn','bo','br',
+	'bs','bt','bv','bw','by','bz','ca','cafe','cat','cc','cd','cf','cg','ch','ci','ck','cl',
+	'cm','cn','co','com','coop','cr','cu','cv','cx','cy','cz','de','dj','dk','dm','do','dz','ec',
+	'edu','ee','eg','er','es','et','eu','fi','fj','fk','fm','fo','fr','ga','gb','gd','ge','gf',
+	'gg','gh','gi','gl','gm','gn','gov','gp','gq','gr','gs','gt','gu','gw','gy','hk','hm','hn',
+	'hr','ht','hu','id','ie','il','im','in','info','int','io','iq','ir','is','it','je','jm','jo',
+	'jobs','jp','ke','kg','kh','ki','km','kn','kp','kr','kw','ky','kz','la','lb','lc','li','lk',
+	'lr','ls','lt','lu','lv','ly','ma','mc','md','me','mg','mh','mil','mk','ml','mm','mn','mo',
+	'mobi','mp','mq','mr','ms','mt','mu','museum','mv','mw','mx','my','mz','na','name',
+	'nc','ne','net','nf','ng','ni','nl','no','np','nr','nu','nz','om','org','pa','pe','pf','pg',
+	'ph','pk','pl','pm','pn','post','pr','pro','ps','pt','pw','py','qa','re','ro','rs','ru','rw',
+	'sa','sb','sc','sd','se','sg','sh','si','sj','sk','sl','sm','sn','so','social','sr','ss','st',
+	'su','sv','sx','sy','sz','tc','td','tel','tf','tg','th','tj','tk','tl','tm','tn','to','tp',
+	'tr','travel','tt','tv','tw','tz','ua','ug','uk','us','uy','uz','va','vc','ve','vg','vi','vn',
+	'vu','wf','ws','xn','xxx','ye','yt','yu','za','zm','zw',
+	# New gTLDs
+	'app','cleaning','club','dev','farm','florist','fun','gay','lgbt','life','lol',
+	'moe','mom','monster','new','news','online','pics','press','pub','site','blog',
+	'vip','win','world','wtf','xyz','video','host','art','media','wiki','tech',
+	'cooking','network','party','goog','markets','today','beauty','camp','top',
+	'red','city','quest','works','soy',
+	)
+
+protocols = ('http', 'https')
+
+sanitize_url_regex = re.compile(
+		r"""\(*# Match any opening parentheses.
+		\b(?"]*)?
+			# /path/zz (excluding "unsafe" chars from RFC 1738,
+			# except for ~, which happens in practice)
+		(?:\#[^#\s\|\\\^\[\]`<>"]*)?
+			# #hash (excluding "unsafe" chars from RFC 1738,
+			# except for ~, which happens in practice)
+		""".format(
+			"|".join(sorted(protocols)), "|".join(sorted(tlds))
+		),
+		re.X | re.U,
+	)
 
 ###REDDIT
 
diff --git a/files/helpers/sanitize.py b/files/helpers/sanitize.py
index 60cc2607c..a8a91335b 100644
--- a/files/helpers/sanitize.py
+++ b/files/helpers/sanitize.py
@@ -27,30 +27,6 @@ from files.helpers.const_stateful import *
 from files.helpers.regex import *
 from files.helpers.get import *
 
-TLDS = ( # Original gTLDs and ccTLDs
-	'ac','ad','ae','aero','af','ag','ai','al','am','an','ao','aq','ar','arpa','as','asia','at',
-	'au','aw','ax','az','ba','bb','bd','be','bf','bg','bh','bi','biz','bj','bm','bn','bo','br',
-	'bs','bt','bv','bw','by','bz','ca','cafe','cat','cc','cd','cf','cg','ch','ci','ck','cl',
-	'cm','cn','co','com','coop','cr','cu','cv','cx','cy','cz','de','dj','dk','dm','do','dz','ec',
-	'edu','ee','eg','er','es','et','eu','fi','fj','fk','fm','fo','fr','ga','gb','gd','ge','gf',
-	'gg','gh','gi','gl','gm','gn','gov','gp','gq','gr','gs','gt','gu','gw','gy','hk','hm','hn',
-	'hr','ht','hu','id','ie','il','im','in','info','int','io','iq','ir','is','it','je','jm','jo',
-	'jobs','jp','ke','kg','kh','ki','km','kn','kp','kr','kw','ky','kz','la','lb','lc','li','lk',
-	'lr','ls','lt','lu','lv','ly','ma','mc','md','me','mg','mh','mil','mk','ml','mm','mn','mo',
-	'mobi','mp','mq','mr','ms','mt','mu','museum','mv','mw','mx','my','mz','na','name',
-	'nc','ne','net','nf','ng','ni','nl','no','np','nr','nu','nz','om','org','pa','pe','pf','pg',
-	'ph','pk','pl','pm','pn','post','pr','pro','ps','pt','pw','py','qa','re','ro','rs','ru','rw',
-	'sa','sb','sc','sd','se','sg','sh','si','sj','sk','sl','sm','sn','so','social','sr','ss','st',
-	'su','sv','sx','sy','sz','tc','td','tel','tf','tg','th','tj','tk','tl','tm','tn','to','tp',
-	'tr','travel','tt','tv','tw','tz','ua','ug','uk','us','uy','uz','va','vc','ve','vg','vi','vn',
-	'vu','wf','ws','xn','xxx','ye','yt','yu','za','zm','zw',
-	# New gTLDs
-	'app','cleaning','club','dev','farm','florist','fun','gay','lgbt','life','lol',
-	'moe','mom','monster','new','news','online','pics','press','pub','site','blog',
-	'vip','win','world','wtf','xyz','video','host','art','media','wiki','tech',
-	'cooking','network','party','goog','markets','today','beauty','camp','top',
-	'red','city','quest','works','soy',
-	)
 
 allowed_tags = ('a','audio','b','big','blockquote','br','center','code','del','details','em','g','h1','h2','h3','h4','h5','h6','hr','i','img','li','lite-youtube','marquee','ol','p','pre','rp','rt','ruby','small','span','spoiler','strike','strong','sub','summary','sup','table','tbody','td','th','thead','tr','u','ul','video')
 
@@ -112,37 +88,6 @@ def allowed_attributes(tag, name, value):
 
 	return False
 
-def build_url_re(tlds, protocols):
-	"""Builds the url regex used by linkifier
-
-	If you want a different set of tlds or allowed protocols, pass those in
-	and stomp on the existing ``url_re``::
-
-		from bleach import linkifier
-
-		my_url_re = linkifier.build_url_re(my_tlds_list, my_protocols)
-
-		linker = LinkifyFilter(url_re=my_url_re)
-
-	"""
-	return re.compile(
-		r"""\(*# Match any opening parentheses.
-		\b(?"]*)?
-			# /path/zz (excluding "unsafe" chars from RFC 1738,
-			# except for ~, which happens in practice)
-		(?:\#[^#\s\|\\\^\[\]`<>"]*)?
-			# #hash (excluding "unsafe" chars from RFC 1738,
-			# except for ~, which happens in practice)
-		""".format(
-			"|".join(sorted(protocols)), "|".join(sorted(tlds))
-		),
-		re.VERBOSE | re.UNICODE,
-	)
-
-url_re = build_url_re(tlds=TLDS, protocols=['http', 'https'])
-
 def create_comment_duplicated(text_html):
 	new_comment = Comment(author_id=AUTOJANNY_ID,
 							parent_post=None,
@@ -582,7 +527,7 @@ def sanitize(sanitized, golden=True, limit_pings=0, showmore=False, count_emojis
 								protocols=['http', 'https'],
 								css_sanitizer=css_sanitizer,
 								filters=[partial(LinkifyFilter, skip_tags=["pre"],
-									parse_email=False, url_re=url_re)]
+									parse_email=False, url_re=sanitize_url_regex)]
 								).clean(sanitized)
 
 	#doing this here cuz of the linkifyfilter right above it (therefore unifying all link processing logic)