|
|
@@ -0,0 +1,96 @@ |
|
|
|
import os.path |
|
|
|
import sys |
|
|
|
import feedparser |
|
|
|
from mastodon import Mastodon |
|
|
|
import json |
|
|
|
import requests |
|
|
|
import re |
|
|
|
import sqlite3 |
|
|
|
from datetime import datetime, date, time, timedelta |
|
|
|
|
|
|
|
if len(sys.argv) < 3: |
|
|
|
print("Usage: python3 tootbot.py twitter_account mastodon_login mastodon_passwd mastodon_instance") |
|
|
|
sys.exit(1) |
|
|
|
|
|
|
|
# sqlite db to store processed tweets (and corresponding toots ids) |
|
|
|
sql = sqlite3.connect('tootbot.db') |
|
|
|
db = sql.cursor() |
|
|
|
db.execute('''CREATE TABLE IF NOT EXISTS tweets (tweet text, toot text, twitter text, mastodon text, instance text)''') |
|
|
|
|
|
|
|
if len(sys.argv)>3: |
|
|
|
instance = sys.argv[4] |
|
|
|
else: |
|
|
|
instance = 'amicale.net' |
|
|
|
|
|
|
|
twitter = sys.argv[1] |
|
|
|
mastodon = sys.argv[2] |
|
|
|
passwd = sys.argv[3] |
|
|
|
|
|
|
|
# Create application if it does not exist |
|
|
|
if not os.path.isfile(instance+'.secret'): |
|
|
|
if Mastodon.create_app( |
|
|
|
'tootbot', |
|
|
|
api_base_url='https://'+instance, |
|
|
|
to_file = instance+'.secret' |
|
|
|
): |
|
|
|
print('tootbot app created on instance '+instance) |
|
|
|
else: |
|
|
|
print('failed to create app on instance '+instance) |
|
|
|
sys.exit(1) |
|
|
|
|
|
|
|
try: |
|
|
|
mastodon_api = Mastodon( |
|
|
|
client_id=instance+'.secret', |
|
|
|
api_base_url='https://'+instance |
|
|
|
) |
|
|
|
mastodon_api.log_in( |
|
|
|
username=mastodon, |
|
|
|
password=passwd, |
|
|
|
scopes=['read', 'write'], |
|
|
|
to_file=mastodon+".secret" |
|
|
|
) |
|
|
|
except: |
|
|
|
print("ERROR: First Login Failed!") |
|
|
|
sys.exit(1) |
|
|
|
|
|
|
|
d = feedparser.parse('http://twitrss.me/twitter_user_to_rss/?user='+twitter) |
|
|
|
|
|
|
|
for t in reversed(d.entries): |
|
|
|
# check if this tweet has been processed |
|
|
|
db.execute('SELECT * FROM tweets WHERE tweet = ? AND twitter = ? and mastodon = ? and instance = ?',(t.id, twitter, mastodon, instance)) |
|
|
|
last = db.fetchone() |
|
|
|
|
|
|
|
# process only unprocessed tweets less than 1 day old |
|
|
|
if last is None and (datetime.now()-datetime(t.published_parsed.tm_year, t.published_parsed.tm_mon, t.published_parsed.tm_mday, t.published_parsed.tm_hour, t.published_parsed.tm_min, t.published_parsed.tm_sec) < timedelta(days=1)): |
|
|
|
#h = BeautifulSoup(t.summary_detail.value, "html.parser") |
|
|
|
c = t.title |
|
|
|
toot_media = [] |
|
|
|
# get the pictures... |
|
|
|
for p in re.finditer(r"https://pbs.twimg.com/[^ \xa0\"]*", t.summary): |
|
|
|
media = requests.get(p.group(0)) |
|
|
|
media_posted = mastodon_api.media_post(media.content, mime_type=media.headers.get('content-type')) |
|
|
|
toot_media.append(media_posted['id']) |
|
|
|
|
|
|
|
# replace t.co link by original URL |
|
|
|
m = re.search(r"http[^ \xa0]*", c) |
|
|
|
if m != None: |
|
|
|
l = m.group(0) |
|
|
|
r = requests.get(l, allow_redirects=False) |
|
|
|
if r.status_code in {301,302}: |
|
|
|
c = c.replace(l,r.headers.get('Location')) |
|
|
|
|
|
|
|
# remove pic.twitter.com links |
|
|
|
m = re.search(r"pic.twitter.com[^ \xa0]*", c) |
|
|
|
if m != None: |
|
|
|
l = m.group(0) |
|
|
|
c = c.replace(l,'') |
|
|
|
|
|
|
|
# remove ellipsis |
|
|
|
c = c.replace('\xa0…','') |
|
|
|
|
|
|
|
if toot_media is not None: |
|
|
|
toot = mastodon_api.status_post(c, in_reply_to_id=None, media_ids=toot_media, sensitive=False, visibility='public', spoiler_text=None) |
|
|
|
if "id" in toot: |
|
|
|
db.execute("INSERT INTO tweets VALUES ( ? , ? , ? , ? , ? )", |
|
|
|
(t.id, toot["id"], twitter, mastodon, instance)) |
|
|
|
sql.commit() |