59 lines
1.7 KiB
Python

import re
from html.parser import HTMLParser
url_re = re.compile('<a\s*href=[\'|"](.*?)[\'"].*?>')
class HTMLFilter(HTMLParser):
text = ""
def handle_data(self, data):
self.text += data
def handle_starttag(self, tag, attrs):
if tag == "br":
self.text = self.text+"\n"
def html_filter(data):
f = HTMLFilter()
f.feed(data)
return f.text
def find_item(item, listItems):
for i in range(0, len(listItems)):
if listItems[i].id == item.id:
return i
# Check also retweets.
if item.reblog != None and item.reblog.id == listItems[i].id:
return i
return None
def is_audio_or_video(toot):
if toot.reblog != None:
return is_audio_or_video(toot.reblog)
# Checks firstly for Mastodon native videos and audios.
for media in toot.media_attachments:
if media["type"] == "video" or media["type"] == "audio":
return True
def is_image(toot):
if toot.reblog != None:
return is_audio_or_video(toot.reblog)
# Checks firstly for Mastodon native videos and audios.
for media in toot.media_attachments:
if media["type"] == "gifv" or media["type"] == "image":
return True
def get_media_urls(toot):
urls = []
for media in toot.media_attachments:
if media.get("type") == "audio" or media.get("type") == "video":
urls.append(media.get("url"))
return urls
def find_urls(toot, include_tags=False):
urls = url_re.findall(toot.content)
if include_tags == False:
for tag in toot.tags:
for url in urls[::]:
if url.lower().endswith("/tags/"+tag["name"]):
urls.remove(url)
return urls