2018-05-18 13:41:41 -05:00
|
|
|
import binascii
|
|
|
|
import hashlib
|
|
|
|
import json
|
|
|
|
import logging
|
2018-06-16 15:02:10 -05:00
|
|
|
import mimetypes
|
|
|
|
import os
|
|
|
|
import urllib
|
2018-05-18 13:41:41 -05:00
|
|
|
from datetime import datetime
|
2018-06-16 15:02:10 -05:00
|
|
|
from functools import wraps
|
|
|
|
from typing import Any
|
|
|
|
from typing import Dict
|
|
|
|
from urllib.parse import urlencode
|
|
|
|
from urllib.parse import urlparse
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
import bleach
|
|
|
|
import mf2py
|
|
|
|
import piexif
|
2018-06-16 15:02:10 -05:00
|
|
|
import pymongo
|
|
|
|
import timeago
|
2018-05-18 13:41:41 -05:00
|
|
|
from bson.objectid import ObjectId
|
|
|
|
from flask import Flask
|
2018-06-16 15:02:10 -05:00
|
|
|
from flask import Response
|
2018-05-18 13:41:41 -05:00
|
|
|
from flask import abort
|
2018-06-16 15:02:10 -05:00
|
|
|
from flask import jsonify as flask_jsonify
|
2018-05-18 13:41:41 -05:00
|
|
|
from flask import redirect
|
|
|
|
from flask import render_template
|
2018-06-16 15:02:10 -05:00
|
|
|
from flask import request
|
2018-05-18 13:41:41 -05:00
|
|
|
from flask import session
|
|
|
|
from flask import url_for
|
2018-06-16 15:02:10 -05:00
|
|
|
from flask_wtf.csrf import CSRFProtect
|
2018-05-18 13:41:41 -05:00
|
|
|
from html2text import html2text
|
|
|
|
from itsdangerous import BadSignature
|
|
|
|
from passlib.hash import bcrypt
|
|
|
|
from u2flib_server import u2f
|
|
|
|
from werkzeug.utils import secure_filename
|
|
|
|
|
|
|
|
import activitypub
|
|
|
|
import config
|
2018-06-17 12:21:59 -05:00
|
|
|
from activitypub import embed_collection
|
2018-06-16 15:02:10 -05:00
|
|
|
from config import ADMIN_API_KEY
|
|
|
|
from config import BASE_URL
|
2018-05-18 13:41:41 -05:00
|
|
|
from config import DB
|
2018-06-16 15:02:10 -05:00
|
|
|
from config import DEBUG_MODE
|
2018-05-18 13:41:41 -05:00
|
|
|
from config import DOMAIN
|
2018-06-16 15:02:10 -05:00
|
|
|
from config import HEADERS
|
|
|
|
from config import ID
|
|
|
|
from config import JWT
|
|
|
|
from config import KEY
|
|
|
|
from config import ME
|
2018-05-18 13:41:41 -05:00
|
|
|
from config import PASS
|
2018-06-16 15:02:10 -05:00
|
|
|
from config import USERNAME
|
2018-05-21 10:04:53 -05:00
|
|
|
from config import VERSION
|
2018-05-27 04:50:09 -05:00
|
|
|
from config import _drop_db
|
2018-05-21 10:04:53 -05:00
|
|
|
from config import custom_cache_purge_hook
|
2018-06-16 15:33:51 -05:00
|
|
|
from little_boxes import activitypub as ap
|
2018-06-16 15:02:10 -05:00
|
|
|
from little_boxes.activitypub import ActivityType
|
|
|
|
from little_boxes.activitypub import clean_activity
|
2018-06-20 16:42:12 -05:00
|
|
|
from little_boxes.activitypub import get_backend
|
2018-06-17 12:21:59 -05:00
|
|
|
from little_boxes.content_helper import parse_markdown
|
|
|
|
from little_boxes.errors import ActivityNotFoundError
|
2018-06-16 15:02:10 -05:00
|
|
|
from little_boxes.errors import Error
|
2018-06-17 12:21:59 -05:00
|
|
|
from little_boxes.errors import NotFromOutboxError
|
2018-06-16 15:02:10 -05:00
|
|
|
from little_boxes.httpsig import HTTPSigAuth
|
2018-06-20 16:42:12 -05:00
|
|
|
from little_boxes.httpsig import verify_request
|
2018-06-16 15:02:10 -05:00
|
|
|
from little_boxes.webfinger import get_actor_url
|
|
|
|
from little_boxes.webfinger import get_remote_follow_template
|
|
|
|
from utils.key import get_secret_key
|
2018-06-17 13:51:23 -05:00
|
|
|
from utils.object_service import ObjectService
|
|
|
|
|
|
|
|
OBJECT_SERVICE = ACTOR_SERVICE = ObjectService()
|
|
|
|
|
|
|
|
back = activitypub.MicroblogPubBackend()
|
|
|
|
ap.use_backend(back)
|
|
|
|
|
|
|
|
MY_PERSON = ap.Person(**ME)
|
2018-06-16 14:24:53 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
app = Flask(__name__)
|
2018-06-16 15:02:10 -05:00
|
|
|
app.secret_key = get_secret_key("flask")
|
|
|
|
app.config.update(WTF_CSRF_CHECK_DEFAULT=False)
|
2018-06-01 14:54:43 -05:00
|
|
|
csrf = CSRFProtect(app)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-05-28 12:46:23 -05:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2018-05-22 17:41:37 -05:00
|
|
|
# Hook up Flask logging with gunicorn
|
2018-06-03 05:51:57 -05:00
|
|
|
root_logger = logging.getLogger()
|
2018-06-16 15:02:10 -05:00
|
|
|
if os.getenv("FLASK_DEBUG"):
|
2018-06-03 15:44:19 -05:00
|
|
|
logger.setLevel(logging.DEBUG)
|
|
|
|
root_logger.setLevel(logging.DEBUG)
|
|
|
|
else:
|
2018-06-16 15:02:10 -05:00
|
|
|
gunicorn_logger = logging.getLogger("gunicorn.error")
|
2018-06-03 15:44:19 -05:00
|
|
|
root_logger.handlers = gunicorn_logger.handlers
|
|
|
|
root_logger.setLevel(gunicorn_logger.level)
|
2018-05-22 17:41:37 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
SIG_AUTH = HTTPSigAuth(KEY)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:33:51 -05:00
|
|
|
OUTBOX = ap.Outbox(MY_PERSON)
|
2018-06-17 12:21:59 -05:00
|
|
|
INBOX = ap.Inbox(MY_PERSON)
|
2018-06-16 15:33:51 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
def verify_pass(pwd):
|
2018-06-16 15:02:10 -05:00
|
|
|
return bcrypt.verify(pwd, PASS)
|
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
@app.context_processor
|
|
|
|
def inject_config():
|
2018-06-16 15:02:10 -05:00
|
|
|
return dict(
|
|
|
|
microblogpub_version=VERSION,
|
|
|
|
config=config,
|
|
|
|
logged_in=session.get("logged_in", False),
|
|
|
|
)
|
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
@app.after_request
|
|
|
|
def set_x_powered_by(response):
|
2018-06-16 15:02:10 -05:00
|
|
|
response.headers["X-Powered-By"] = "microblog.pub"
|
2018-05-18 13:41:41 -05:00
|
|
|
return response
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
# HTML/templates helper
|
|
|
|
ALLOWED_TAGS = [
|
2018-06-16 15:02:10 -05:00
|
|
|
"a",
|
|
|
|
"abbr",
|
|
|
|
"acronym",
|
|
|
|
"b",
|
|
|
|
"blockquote",
|
|
|
|
"code",
|
|
|
|
"pre",
|
|
|
|
"em",
|
|
|
|
"i",
|
|
|
|
"li",
|
|
|
|
"ol",
|
|
|
|
"strong",
|
|
|
|
"ul",
|
|
|
|
"span",
|
|
|
|
"div",
|
|
|
|
"p",
|
|
|
|
"h1",
|
|
|
|
"h2",
|
|
|
|
"h3",
|
|
|
|
"h4",
|
|
|
|
"h5",
|
|
|
|
"h6",
|
2018-05-18 13:41:41 -05:00
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
def clean_html(html):
|
|
|
|
return bleach.clean(html, tags=ALLOWED_TAGS)
|
|
|
|
|
|
|
|
|
2018-06-16 14:24:53 -05:00
|
|
|
@app.template_filter()
|
|
|
|
def quote_plus(t):
|
|
|
|
return urllib.parse.quote_plus(t)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 14:24:53 -05:00
|
|
|
@app.template_filter()
|
|
|
|
def is_from_outbox(t):
|
2018-06-04 11:53:44 -05:00
|
|
|
return t.startswith(ID)
|
|
|
|
|
|
|
|
|
2018-06-16 14:24:53 -05:00
|
|
|
@app.template_filter()
|
|
|
|
def clean(html):
|
|
|
|
return clean_html(html)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 14:24:53 -05:00
|
|
|
@app.template_filter()
|
|
|
|
def html2plaintext(body):
|
2018-05-18 13:41:41 -05:00
|
|
|
return html2text(body)
|
|
|
|
|
|
|
|
|
|
|
|
@app.template_filter()
|
|
|
|
def domain(url):
|
|
|
|
return urlparse(url).netloc
|
|
|
|
|
|
|
|
|
|
|
|
@app.template_filter()
|
|
|
|
def get_actor(url):
|
|
|
|
if not url:
|
|
|
|
return None
|
2018-06-16 15:02:10 -05:00
|
|
|
print(f"GET_ACTOR {url}")
|
2018-05-18 13:41:41 -05:00
|
|
|
return ACTOR_SERVICE.get(url)
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
@app.template_filter()
|
|
|
|
def format_time(val):
|
|
|
|
if val:
|
2018-06-16 15:02:10 -05:00
|
|
|
return datetime.strftime(
|
|
|
|
datetime.strptime(val, "%Y-%m-%dT%H:%M:%SZ"), "%B %d, %Y, %H:%M %p"
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
return val
|
|
|
|
|
|
|
|
|
|
|
|
@app.template_filter()
|
|
|
|
def format_timeago(val):
|
|
|
|
if val:
|
|
|
|
try:
|
2018-06-16 15:02:10 -05:00
|
|
|
return timeago.format(
|
|
|
|
datetime.strptime(val, "%Y-%m-%dT%H:%M:%SZ"), datetime.utcnow()
|
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
return timeago.format(
|
|
|
|
datetime.strptime(val, "%Y-%m-%dT%H:%M:%S.%fZ"), datetime.utcnow()
|
|
|
|
)
|
2018-06-16 14:24:53 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
return val
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
def _is_img(filename):
|
|
|
|
filename = filename.lower()
|
2018-06-16 15:02:10 -05:00
|
|
|
if (
|
|
|
|
filename.endswith(".png")
|
|
|
|
or filename.endswith(".jpg")
|
|
|
|
or filename.endswith(".jpeg")
|
|
|
|
or filename.endswith(".gif")
|
|
|
|
or filename.endswith(".svg")
|
|
|
|
):
|
2018-05-18 13:41:41 -05:00
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
@app.template_filter()
|
|
|
|
def not_only_imgs(attachment):
|
|
|
|
for a in attachment:
|
2018-06-16 15:02:10 -05:00
|
|
|
if not _is_img(a["url"]):
|
2018-05-18 13:41:41 -05:00
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
@app.template_filter()
|
|
|
|
def is_img(filename):
|
|
|
|
return _is_img(filename)
|
|
|
|
|
|
|
|
|
|
|
|
def login_required(f):
|
|
|
|
@wraps(f)
|
|
|
|
def decorated_function(*args, **kwargs):
|
2018-06-16 15:02:10 -05:00
|
|
|
if not session.get("logged_in"):
|
|
|
|
return redirect(url_for("login", next=request.url))
|
2018-05-18 13:41:41 -05:00
|
|
|
return f(*args, **kwargs)
|
2018-06-16 15:02:10 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
return decorated_function
|
|
|
|
|
2018-05-21 07:41:47 -05:00
|
|
|
|
2018-05-21 07:30:52 -05:00
|
|
|
def _api_required():
|
2018-06-16 15:02:10 -05:00
|
|
|
if session.get("logged_in"):
|
|
|
|
if request.method not in ["GET", "HEAD"]:
|
2018-06-04 10:59:38 -05:00
|
|
|
# If a standard API request is made with a "login session", it must havw a CSRF token
|
|
|
|
csrf.protect()
|
2018-05-21 07:30:52 -05:00
|
|
|
return
|
|
|
|
|
|
|
|
# Token verification
|
2018-06-16 15:02:10 -05:00
|
|
|
token = request.headers.get("Authorization", "").replace("Bearer ", "")
|
2018-05-21 07:30:52 -05:00
|
|
|
if not token:
|
2018-05-29 14:36:05 -05:00
|
|
|
# IndieAuth token
|
2018-06-16 15:02:10 -05:00
|
|
|
token = request.form.get("access_token", "")
|
2018-05-21 07:30:52 -05:00
|
|
|
|
|
|
|
# Will raise a BadSignature on bad auth
|
|
|
|
payload = JWT.loads(token)
|
2018-06-16 15:02:10 -05:00
|
|
|
logger.info(f"api call by {payload}")
|
2018-05-21 07:41:47 -05:00
|
|
|
|
|
|
|
|
|
|
|
def api_required(f):
|
2018-05-18 13:41:41 -05:00
|
|
|
@wraps(f)
|
|
|
|
def decorated_function(*args, **kwargs):
|
|
|
|
try:
|
2018-05-21 07:30:52 -05:00
|
|
|
_api_required()
|
2018-05-18 13:41:41 -05:00
|
|
|
except BadSignature:
|
|
|
|
abort(401)
|
2018-05-21 07:30:52 -05:00
|
|
|
|
2018-05-21 07:41:47 -05:00
|
|
|
return f(*args, **kwargs)
|
2018-06-16 15:02:10 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
return decorated_function
|
|
|
|
|
|
|
|
|
|
|
|
def jsonify(**data):
|
2018-06-16 15:02:10 -05:00
|
|
|
if "@context" not in data:
|
|
|
|
data["@context"] = config.CTX_AS
|
2018-05-18 13:41:41 -05:00
|
|
|
return Response(
|
|
|
|
response=json.dumps(data),
|
2018-06-16 15:02:10 -05:00
|
|
|
headers={
|
|
|
|
"Content-Type": "application/json"
|
|
|
|
if app.debug
|
|
|
|
else "application/activity+json"
|
|
|
|
},
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def is_api_request():
|
2018-06-16 15:02:10 -05:00
|
|
|
h = request.headers.get("Accept")
|
2018-05-18 13:41:41 -05:00
|
|
|
if h is None:
|
|
|
|
return False
|
2018-06-16 15:02:10 -05:00
|
|
|
h = h.split(",")[0]
|
|
|
|
if h in HEADERS or h == "application/json":
|
2018-05-18 13:41:41 -05:00
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
2018-05-29 14:36:05 -05:00
|
|
|
|
|
|
|
@app.errorhandler(ValueError)
|
|
|
|
def handle_value_error(error):
|
2018-06-16 15:02:10 -05:00
|
|
|
logger.error(f"caught value error: {error!r}")
|
2018-05-29 14:36:05 -05:00
|
|
|
response = flask_jsonify(message=error.args[0])
|
|
|
|
response.status_code = 400
|
|
|
|
return response
|
|
|
|
|
|
|
|
|
|
|
|
@app.errorhandler(Error)
|
|
|
|
def handle_activitypub_error(error):
|
2018-06-16 15:02:10 -05:00
|
|
|
logger.error(f"caught activitypub error {error!r}")
|
2018-05-29 14:36:05 -05:00
|
|
|
response = flask_jsonify(error.to_dict())
|
|
|
|
response.status_code = error.status_code
|
|
|
|
return response
|
|
|
|
|
|
|
|
|
2018-06-16 14:24:53 -05:00
|
|
|
# App routes
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
#######
|
|
|
|
# Login
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
|
|
|
|
@app.route("/logout")
|
2018-05-18 13:41:41 -05:00
|
|
|
@login_required
|
|
|
|
def logout():
|
2018-06-16 15:02:10 -05:00
|
|
|
session["logged_in"] = False
|
|
|
|
return redirect("/")
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/login", methods=["POST", "GET"])
|
2018-05-18 13:41:41 -05:00
|
|
|
def login():
|
2018-06-16 15:02:10 -05:00
|
|
|
devices = [doc["device"] for doc in DB.u2f.find()]
|
2018-05-18 13:41:41 -05:00
|
|
|
u2f_enabled = True if devices else False
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.method == "POST":
|
2018-06-01 14:54:43 -05:00
|
|
|
csrf.protect()
|
2018-06-16 15:02:10 -05:00
|
|
|
pwd = request.form.get("pass")
|
2018-05-18 13:41:41 -05:00
|
|
|
if pwd and verify_pass(pwd):
|
|
|
|
if devices:
|
2018-06-16 15:02:10 -05:00
|
|
|
resp = json.loads(request.form.get("resp"))
|
2018-05-18 13:41:41 -05:00
|
|
|
print(resp)
|
|
|
|
try:
|
2018-06-16 15:02:10 -05:00
|
|
|
u2f.complete_authentication(session["challenge"], resp)
|
2018-05-18 13:41:41 -05:00
|
|
|
except ValueError as exc:
|
2018-06-16 15:02:10 -05:00
|
|
|
print("failed", exc)
|
2018-05-18 13:41:41 -05:00
|
|
|
abort(401)
|
|
|
|
return
|
|
|
|
finally:
|
2018-06-16 15:02:10 -05:00
|
|
|
session["challenge"] = None
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
session["logged_in"] = True
|
|
|
|
return redirect(request.args.get("redirect") or "/admin")
|
2018-05-18 13:41:41 -05:00
|
|
|
else:
|
|
|
|
abort(401)
|
|
|
|
|
|
|
|
payload = None
|
|
|
|
if devices:
|
|
|
|
payload = u2f.begin_authentication(ID, devices)
|
2018-06-16 15:02:10 -05:00
|
|
|
session["challenge"] = payload
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
return render_template(
|
2018-06-16 15:02:10 -05:00
|
|
|
"login.html", u2f_enabled=u2f_enabled, me=ME, payload=payload
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/remote_follow", methods=["GET", "POST"])
|
2018-05-18 13:41:41 -05:00
|
|
|
def remote_follow():
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.method == "GET":
|
|
|
|
return render_template("remote_follow.html")
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-04 10:59:38 -05:00
|
|
|
csrf.protect()
|
2018-06-21 17:23:05 -05:00
|
|
|
profile = request.form.get("profile")
|
|
|
|
if not profile.startswith("@"):
|
|
|
|
profile = f"@{profile}"
|
2018-06-16 15:02:10 -05:00
|
|
|
return redirect(
|
2018-06-21 17:23:05 -05:00
|
|
|
get_remote_follow_template(profile).format(uri=f"{USERNAME}@{DOMAIN}")
|
2018-06-16 15:02:10 -05:00
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/authorize_follow", methods=["GET", "POST"])
|
2018-05-18 13:41:41 -05:00
|
|
|
@login_required
|
|
|
|
def authorize_follow():
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.method == "GET":
|
|
|
|
return render_template(
|
|
|
|
"authorize_remote_follow.html", profile=request.args.get("profile")
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
actor = get_actor_url(request.form.get("profile"))
|
2018-05-18 13:41:41 -05:00
|
|
|
if not actor:
|
|
|
|
abort(500)
|
2018-06-16 15:02:10 -05:00
|
|
|
if DB.following.find({"remote_actor": actor}).count() > 0:
|
|
|
|
return redirect("/following")
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-17 14:54:16 -05:00
|
|
|
follow = activitypub.Follow(actor=MY_PERSON.id, object=actor)
|
2018-06-16 15:33:51 -05:00
|
|
|
OUTBOX.post(follow)
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return redirect("/following")
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/u2f/register", methods=["GET", "POST"])
|
2018-05-18 13:41:41 -05:00
|
|
|
@login_required
|
|
|
|
def u2f_register():
|
|
|
|
# TODO(tsileo): ensure no duplicates
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.method == "GET":
|
2018-05-18 13:41:41 -05:00
|
|
|
payload = u2f.begin_registration(ID)
|
2018-06-16 15:02:10 -05:00
|
|
|
session["challenge"] = payload
|
|
|
|
return render_template("u2f.html", payload=payload)
|
2018-05-18 13:41:41 -05:00
|
|
|
else:
|
2018-06-16 15:02:10 -05:00
|
|
|
resp = json.loads(request.form.get("resp"))
|
|
|
|
device, device_cert = u2f.complete_registration(session["challenge"], resp)
|
|
|
|
session["challenge"] = None
|
|
|
|
DB.u2f.insert_one({"device": device, "cert": device_cert})
|
|
|
|
return ""
|
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
#######
|
|
|
|
# Activity pub routes
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
|
|
|
|
@app.route("/")
|
2018-05-18 13:41:41 -05:00
|
|
|
def index():
|
|
|
|
if is_api_request():
|
|
|
|
return jsonify(**ME)
|
|
|
|
|
|
|
|
# FIXME(tsileo): implements pagination, also for the followers/following page
|
|
|
|
limit = 50
|
|
|
|
q = {
|
2018-06-16 15:02:10 -05:00
|
|
|
"type": "Create",
|
|
|
|
"activity.object.type": "Note",
|
|
|
|
"activity.object.inReplyTo": None,
|
|
|
|
"meta.deleted": False,
|
2018-05-18 13:41:41 -05:00
|
|
|
}
|
2018-06-16 15:02:10 -05:00
|
|
|
c = request.args.get("cursor")
|
2018-05-18 13:41:41 -05:00
|
|
|
if c:
|
2018-06-16 15:02:10 -05:00
|
|
|
q["_id"] = {"$lt": ObjectId(c)}
|
2018-06-04 10:59:38 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
outbox_data = list(
|
|
|
|
DB.outbox.find(
|
|
|
|
{"$or": [q, {"type": "Announce", "meta.undo": False}]}, limit=limit
|
|
|
|
).sort("_id", -1)
|
|
|
|
)
|
2018-06-04 10:59:38 -05:00
|
|
|
cursor = None
|
|
|
|
if outbox_data and len(outbox_data) == limit:
|
2018-06-16 15:02:10 -05:00
|
|
|
cursor = str(outbox_data[-1]["_id"])
|
2018-06-04 10:59:38 -05:00
|
|
|
|
|
|
|
for data in outbox_data:
|
2018-06-16 15:02:10 -05:00
|
|
|
if data["type"] == "Announce":
|
2018-06-04 10:59:38 -05:00
|
|
|
print(data)
|
2018-06-16 15:02:10 -05:00
|
|
|
if data["activity"]["object"].startswith("http"):
|
|
|
|
data["ref"] = {
|
|
|
|
"activity": {
|
|
|
|
"object": OBJECT_SERVICE.get(data["activity"]["object"])
|
|
|
|
},
|
|
|
|
"meta": {},
|
|
|
|
}
|
2018-06-04 10:59:38 -05:00
|
|
|
|
|
|
|
return render_template(
|
2018-06-16 15:02:10 -05:00
|
|
|
"index.html",
|
2018-06-04 10:59:38 -05:00
|
|
|
me=ME,
|
2018-06-16 15:02:10 -05:00
|
|
|
notes=DB.inbox.find(
|
|
|
|
{"type": "Create", "activity.object.type": "Note", "meta.deleted": False}
|
|
|
|
).count(),
|
2018-06-04 10:59:38 -05:00
|
|
|
followers=DB.followers.count(),
|
|
|
|
following=DB.following.count(),
|
|
|
|
outbox_data=outbox_data,
|
|
|
|
cursor=cursor,
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/with_replies")
|
2018-06-04 10:59:38 -05:00
|
|
|
def with_replies():
|
|
|
|
limit = 50
|
2018-06-16 15:02:10 -05:00
|
|
|
q = {"type": "Create", "activity.object.type": "Note", "meta.deleted": False}
|
|
|
|
c = request.args.get("cursor")
|
2018-06-04 10:59:38 -05:00
|
|
|
if c:
|
2018-06-16 15:02:10 -05:00
|
|
|
q["_id"] = {"$lt": ObjectId(c)}
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
outbox_data = list(
|
|
|
|
DB.outbox.find(
|
|
|
|
{"$or": [q, {"type": "Announce", "meta.undo": False}]}, limit=limit
|
|
|
|
).sort("_id", -1)
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
cursor = None
|
|
|
|
if outbox_data and len(outbox_data) == limit:
|
2018-06-16 15:02:10 -05:00
|
|
|
cursor = str(outbox_data[-1]["_id"])
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
for data in outbox_data:
|
2018-06-16 15:02:10 -05:00
|
|
|
if data["type"] == "Announce":
|
2018-05-18 13:41:41 -05:00
|
|
|
print(data)
|
2018-06-16 15:02:10 -05:00
|
|
|
if data["activity"]["object"].startswith("http"):
|
|
|
|
data["ref"] = {
|
|
|
|
"activity": {
|
|
|
|
"object": OBJECT_SERVICE.get(data["activity"]["object"])
|
|
|
|
},
|
|
|
|
"meta": {},
|
|
|
|
}
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
return render_template(
|
2018-06-16 15:02:10 -05:00
|
|
|
"index.html",
|
2018-05-18 13:41:41 -05:00
|
|
|
me=ME,
|
2018-06-16 15:02:10 -05:00
|
|
|
notes=DB.inbox.find(
|
|
|
|
{"type": "Create", "activity.object.type": "Note", "meta.deleted": False}
|
|
|
|
).count(),
|
2018-05-18 13:41:41 -05:00
|
|
|
followers=DB.followers.count(),
|
|
|
|
following=DB.following.count(),
|
|
|
|
outbox_data=outbox_data,
|
|
|
|
cursor=cursor,
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2018-06-03 14:28:06 -05:00
|
|
|
def _build_thread(data, include_children=True):
|
2018-06-16 15:02:10 -05:00
|
|
|
data["_requested"] = True
|
|
|
|
root_id = data["meta"].get("thread_root_parent", data["activity"]["object"]["id"])
|
2018-06-03 14:28:06 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
thread_ids = data["meta"].get("thread_parents", [])
|
2018-06-03 14:28:06 -05:00
|
|
|
if include_children:
|
2018-06-16 15:02:10 -05:00
|
|
|
thread_ids.extend(data["meta"].get("thread_children", []))
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-03 14:28:06 -05:00
|
|
|
query = {
|
2018-06-16 15:02:10 -05:00
|
|
|
"activity.object.id": {"$in": thread_ids},
|
|
|
|
"type": "Create",
|
|
|
|
"meta.deleted": False, # TODO(tsileo): handle Tombstone instead of filtering them
|
2018-06-03 14:28:06 -05:00
|
|
|
}
|
|
|
|
# Fetch the root replies, and the children
|
|
|
|
replies = [data] + list(DB.inbox.find(query)) + list(DB.outbox.find(query))
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-03 14:28:06 -05:00
|
|
|
# Index all the IDs in order to build a tree
|
|
|
|
idx = {}
|
2018-05-18 13:41:41 -05:00
|
|
|
for rep in replies:
|
2018-06-16 15:02:10 -05:00
|
|
|
rep_id = rep["activity"]["object"]["id"]
|
2018-06-03 14:28:06 -05:00
|
|
|
idx[rep_id] = rep.copy()
|
2018-06-16 15:02:10 -05:00
|
|
|
idx[rep_id]["_nodes"] = []
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-03 14:28:06 -05:00
|
|
|
# Build the tree
|
|
|
|
for rep in replies:
|
2018-06-16 15:02:10 -05:00
|
|
|
rep_id = rep["activity"]["object"]["id"]
|
2018-06-03 14:28:06 -05:00
|
|
|
if rep_id == root_id:
|
|
|
|
continue
|
2018-06-16 15:02:10 -05:00
|
|
|
reply_of = rep["activity"]["object"]["inReplyTo"]
|
|
|
|
idx[reply_of]["_nodes"].append(rep)
|
2018-06-03 14:28:06 -05:00
|
|
|
|
|
|
|
# Flatten the tree
|
|
|
|
thread = []
|
2018-06-16 15:02:10 -05:00
|
|
|
|
2018-06-03 14:28:06 -05:00
|
|
|
def _flatten(node, level=0):
|
2018-06-16 15:02:10 -05:00
|
|
|
node["_level"] = level
|
2018-06-03 14:28:06 -05:00
|
|
|
thread.append(node)
|
2018-06-16 14:24:53 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
for snode in sorted(
|
|
|
|
idx[node["activity"]["object"]["id"]]["_nodes"],
|
|
|
|
key=lambda d: d["activity"]["object"]["published"],
|
|
|
|
):
|
|
|
|
_flatten(snode, level=level + 1)
|
|
|
|
|
2018-06-03 14:28:06 -05:00
|
|
|
_flatten(idx[root_id])
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-03 14:28:06 -05:00
|
|
|
return thread
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/note/<note_id>")
|
2018-06-16 14:24:53 -05:00
|
|
|
def note_by_id(note_id):
|
2018-06-18 15:01:21 -05:00
|
|
|
data = DB.outbox.find_one({"remote_id": back.activity_url(note_id)})
|
2018-06-16 14:24:53 -05:00
|
|
|
if not data:
|
2018-06-03 14:28:06 -05:00
|
|
|
abort(404)
|
2018-06-16 15:02:10 -05:00
|
|
|
if data["meta"].get("deleted", False):
|
2018-06-03 14:28:06 -05:00
|
|
|
abort(410)
|
|
|
|
thread = _build_thread(data)
|
2018-06-03 16:11:43 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
likes = list(
|
|
|
|
DB.inbox.find(
|
|
|
|
{
|
|
|
|
"meta.undo": False,
|
|
|
|
"type": ActivityType.LIKE.value,
|
|
|
|
"$or": [
|
|
|
|
{"activity.object.id": data["activity"]["object"]["id"]},
|
|
|
|
{"activity.object": data["activity"]["object"]["id"]},
|
|
|
|
],
|
|
|
|
}
|
|
|
|
)
|
|
|
|
)
|
|
|
|
likes = [ACTOR_SERVICE.get(doc["activity"]["actor"]) for doc in likes]
|
2018-06-03 16:11:43 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
shares = list(
|
|
|
|
DB.inbox.find(
|
|
|
|
{
|
|
|
|
"meta.undo": False,
|
|
|
|
"type": ActivityType.ANNOUNCE.value,
|
|
|
|
"$or": [
|
|
|
|
{"activity.object.id": data["activity"]["object"]["id"]},
|
|
|
|
{"activity.object": data["activity"]["object"]["id"]},
|
|
|
|
],
|
|
|
|
}
|
|
|
|
)
|
|
|
|
)
|
|
|
|
shares = [ACTOR_SERVICE.get(doc["activity"]["actor"]) for doc in shares]
|
2018-06-03 16:36:16 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return render_template(
|
|
|
|
"note.html", likes=likes, shares=shares, me=ME, thread=thread, note=data
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/nodeinfo")
|
2018-06-03 03:15:11 -05:00
|
|
|
def nodeinfo():
|
|
|
|
return Response(
|
2018-06-16 15:02:10 -05:00
|
|
|
headers={
|
|
|
|
"Content-Type": "application/json; profile=http://nodeinfo.diaspora.software/ns/schema/2.0#"
|
|
|
|
},
|
|
|
|
response=json.dumps(
|
|
|
|
{
|
|
|
|
"version": "2.0",
|
|
|
|
"software": {
|
|
|
|
"name": "microblogpub",
|
|
|
|
"version": f"Microblog.pub {VERSION}",
|
|
|
|
},
|
|
|
|
"protocols": ["activitypub"],
|
|
|
|
"services": {"inbound": [], "outbound": []},
|
|
|
|
"openRegistrations": False,
|
|
|
|
"usage": {"users": {"total": 1}, "localPosts": DB.outbox.count()},
|
|
|
|
"metadata": {
|
|
|
|
"sourceCode": "https://github.com/tsileo/microblog.pub",
|
|
|
|
"nodeName": f"@{USERNAME}@{DOMAIN}",
|
|
|
|
},
|
|
|
|
}
|
|
|
|
),
|
2018-06-03 03:15:11 -05:00
|
|
|
)
|
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/.well-known/nodeinfo")
|
2018-06-03 03:15:11 -05:00
|
|
|
def wellknown_nodeinfo():
|
|
|
|
return flask_jsonify(
|
|
|
|
links=[
|
|
|
|
{
|
2018-06-16 15:02:10 -05:00
|
|
|
"rel": "http://nodeinfo.diaspora.software/ns/schema/2.0",
|
|
|
|
"href": f"{ID}/nodeinfo",
|
2018-06-03 03:15:11 -05:00
|
|
|
}
|
2018-06-16 15:02:10 -05:00
|
|
|
]
|
2018-06-03 03:15:11 -05:00
|
|
|
)
|
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/.well-known/webfinger")
|
2018-06-03 03:15:11 -05:00
|
|
|
def wellknown_webfinger():
|
2018-05-18 13:41:41 -05:00
|
|
|
"""Enable WebFinger support, required for Mastodon interopability."""
|
2018-06-16 15:02:10 -05:00
|
|
|
resource = request.args.get("resource")
|
|
|
|
if resource not in [f"acct:{USERNAME}@{DOMAIN}", ID]:
|
2018-05-18 13:41:41 -05:00
|
|
|
abort(404)
|
|
|
|
|
|
|
|
out = {
|
2018-06-16 15:02:10 -05:00
|
|
|
"subject": f"acct:{USERNAME}@{DOMAIN}",
|
2018-05-18 13:41:41 -05:00
|
|
|
"aliases": [ID],
|
|
|
|
"links": [
|
2018-06-16 15:02:10 -05:00
|
|
|
{
|
|
|
|
"rel": "http://webfinger.net/rel/profile-page",
|
|
|
|
"type": "text/html",
|
|
|
|
"href": BASE_URL,
|
|
|
|
},
|
2018-05-18 13:41:41 -05:00
|
|
|
{"rel": "self", "type": "application/activity+json", "href": ID},
|
2018-06-16 15:02:10 -05:00
|
|
|
{
|
|
|
|
"rel": "http://ostatus.org/schema/1.0/subscribe",
|
|
|
|
"template": BASE_URL + "/authorize_follow?profile={uri}",
|
|
|
|
},
|
2018-05-18 13:41:41 -05:00
|
|
|
],
|
|
|
|
}
|
|
|
|
|
|
|
|
return Response(
|
|
|
|
response=json.dumps(out),
|
2018-06-16 15:02:10 -05:00
|
|
|
headers={
|
|
|
|
"Content-Type": "application/jrd+json; charset=utf-8"
|
|
|
|
if not app.debug
|
|
|
|
else "application/json"
|
|
|
|
},
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
|
|
|
|
2018-05-28 12:46:23 -05:00
|
|
|
|
|
|
|
def add_extra_collection(raw_doc: Dict[str, Any]) -> Dict[str, Any]:
|
2018-06-16 15:02:10 -05:00
|
|
|
if raw_doc["activity"]["type"] != ActivityType.CREATE.value:
|
2018-05-29 11:59:37 -05:00
|
|
|
return raw_doc
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
raw_doc["activity"]["object"]["replies"] = embed_collection(
|
|
|
|
raw_doc.get("meta", {}).get("count_direct_reply", 0),
|
2018-06-18 15:01:21 -05:00
|
|
|
f'{raw_doc["remote_id"]}/replies',
|
2018-05-31 18:26:23 -05:00
|
|
|
)
|
2018-05-29 11:59:37 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
raw_doc["activity"]["object"]["likes"] = embed_collection(
|
2018-06-18 15:01:21 -05:00
|
|
|
raw_doc.get("meta", {}).get("count_like", 0), f'{raw_doc["remote_id"]}/likes'
|
2018-05-31 18:26:23 -05:00
|
|
|
)
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
raw_doc["activity"]["object"]["shares"] = embed_collection(
|
2018-06-18 15:01:21 -05:00
|
|
|
raw_doc.get("meta", {}).get("count_boost", 0), f'{raw_doc["remote_id"]}/shares'
|
2018-05-31 18:26:23 -05:00
|
|
|
)
|
2018-05-28 12:46:23 -05:00
|
|
|
|
|
|
|
return raw_doc
|
|
|
|
|
|
|
|
|
2018-06-04 12:10:04 -05:00
|
|
|
def remove_context(activity: Dict[str, Any]) -> Dict[str, Any]:
|
2018-06-16 15:02:10 -05:00
|
|
|
if "@context" in activity:
|
|
|
|
del activity["@context"]
|
2018-06-04 12:10:04 -05:00
|
|
|
return activity
|
|
|
|
|
|
|
|
|
|
|
|
def activity_from_doc(raw_doc: Dict[str, Any], embed: bool = False) -> Dict[str, Any]:
|
2018-05-28 12:46:23 -05:00
|
|
|
raw_doc = add_extra_collection(raw_doc)
|
2018-06-16 15:02:10 -05:00
|
|
|
activity = clean_activity(raw_doc["activity"])
|
2018-06-04 12:10:04 -05:00
|
|
|
if embed:
|
|
|
|
return remove_context(activity)
|
|
|
|
return activity
|
|
|
|
|
2018-05-28 12:46:23 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/outbox", methods=["GET", "POST"])
|
2018-06-16 14:24:53 -05:00
|
|
|
def outbox():
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.method == "GET":
|
2018-06-16 14:24:53 -05:00
|
|
|
if not is_api_request():
|
|
|
|
abort(404)
|
2018-05-18 13:41:41 -05:00
|
|
|
# TODO(tsileo): filter the outbox if not authenticated
|
|
|
|
# FIXME(tsileo): filter deleted, add query support for build_ordered_collection
|
|
|
|
q = {
|
2018-06-16 15:02:10 -05:00
|
|
|
"meta.deleted": False,
|
|
|
|
# 'type': {'$in': [ActivityType.CREATE.value, ActivityType.ANNOUNCE.value]},
|
2018-05-18 13:41:41 -05:00
|
|
|
}
|
2018-06-16 15:02:10 -05:00
|
|
|
return jsonify(
|
|
|
|
**activitypub.build_ordered_collection(
|
|
|
|
DB.outbox,
|
|
|
|
q=q,
|
|
|
|
cursor=request.args.get("cursor"),
|
|
|
|
map_func=lambda doc: activity_from_doc(doc, embed=True),
|
|
|
|
)
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
# Handle POST request
|
2018-05-21 07:30:52 -05:00
|
|
|
try:
|
|
|
|
_api_required()
|
|
|
|
except BadSignature:
|
|
|
|
abort(401)
|
2018-06-16 14:24:53 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
data = request.get_json(force=True)
|
|
|
|
print(data)
|
2018-06-17 12:21:59 -05:00
|
|
|
activity = ap.parse_activity(data)
|
|
|
|
OUTBOX.post(activity)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-05-21 10:04:53 -05:00
|
|
|
# Purge the cache if a custom hook is set, as new content was published
|
|
|
|
custom_cache_purge_hook()
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return Response(status=201, headers={"Location": activity.id})
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/outbox/<item_id>")
|
2018-05-18 13:41:41 -05:00
|
|
|
def outbox_detail(item_id):
|
2018-06-18 15:01:21 -05:00
|
|
|
doc = DB.outbox.find_one({"remote_id": back.activity_url(item_id)})
|
2018-06-16 15:02:10 -05:00
|
|
|
if doc["meta"].get("deleted", False):
|
2018-06-17 12:21:59 -05:00
|
|
|
obj = ap.parse_activity(doc["activity"])
|
2018-06-03 07:34:04 -05:00
|
|
|
resp = jsonify(**obj.get_object().get_tombstone())
|
|
|
|
resp.status_code = 410
|
|
|
|
return resp
|
2018-05-28 12:46:23 -05:00
|
|
|
return jsonify(**activity_from_doc(doc))
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/outbox/<item_id>/activity")
|
2018-05-18 13:41:41 -05:00
|
|
|
def outbox_activity(item_id):
|
2018-06-03 07:34:04 -05:00
|
|
|
# TODO(tsileo): handle Tombstone
|
2018-06-18 15:01:21 -05:00
|
|
|
data = DB.outbox.find_one(
|
|
|
|
{"remote_id": back.activity_url(item_id), "meta.deleted": False}
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
if not data:
|
|
|
|
abort(404)
|
2018-05-28 12:46:23 -05:00
|
|
|
obj = activity_from_doc(data)
|
2018-06-16 15:02:10 -05:00
|
|
|
if obj["type"] != ActivityType.CREATE.value:
|
2018-05-18 13:41:41 -05:00
|
|
|
abort(404)
|
2018-06-16 15:02:10 -05:00
|
|
|
return jsonify(**obj["object"])
|
2018-05-28 12:46:23 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/outbox/<item_id>/replies")
|
2018-05-31 18:26:23 -05:00
|
|
|
def outbox_activity_replies(item_id):
|
2018-06-03 07:34:04 -05:00
|
|
|
# TODO(tsileo): handle Tombstone
|
2018-05-31 18:26:23 -05:00
|
|
|
if not is_api_request():
|
|
|
|
abort(404)
|
2018-06-18 15:01:21 -05:00
|
|
|
data = DB.outbox.find_one(
|
|
|
|
{"remote_id": back.activity_url(item_id), "meta.deleted": False}
|
|
|
|
)
|
2018-05-31 18:26:23 -05:00
|
|
|
if not data:
|
|
|
|
abort(404)
|
2018-06-17 12:21:59 -05:00
|
|
|
obj = ap.parse_activity(data["activity"])
|
|
|
|
if obj.ACTIVITY_TYPE != ActivityType.CREATE:
|
2018-05-31 18:26:23 -05:00
|
|
|
abort(404)
|
|
|
|
|
|
|
|
q = {
|
2018-06-16 15:02:10 -05:00
|
|
|
"meta.deleted": False,
|
|
|
|
"type": ActivityType.CREATE.value,
|
|
|
|
"activity.object.inReplyTo": obj.get_object().id,
|
2018-05-31 18:26:23 -05:00
|
|
|
}
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return jsonify(
|
|
|
|
**activitypub.build_ordered_collection(
|
|
|
|
DB.inbox,
|
|
|
|
q=q,
|
|
|
|
cursor=request.args.get("cursor"),
|
|
|
|
map_func=lambda doc: doc["activity"]["object"],
|
|
|
|
col_name=f"outbox/{item_id}/replies",
|
|
|
|
first_page=request.args.get("page") == "first",
|
|
|
|
)
|
|
|
|
)
|
2018-05-31 18:26:23 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/outbox/<item_id>/likes")
|
2018-05-31 18:26:23 -05:00
|
|
|
def outbox_activity_likes(item_id):
|
2018-06-03 07:34:04 -05:00
|
|
|
# TODO(tsileo): handle Tombstone
|
2018-05-31 18:26:23 -05:00
|
|
|
if not is_api_request():
|
|
|
|
abort(404)
|
2018-06-18 15:01:21 -05:00
|
|
|
data = DB.outbox.find_one(
|
|
|
|
{"remote_id": back.activity_url(item_id), "meta.deleted": False}
|
|
|
|
)
|
2018-05-31 18:26:23 -05:00
|
|
|
if not data:
|
|
|
|
abort(404)
|
2018-06-17 12:21:59 -05:00
|
|
|
obj = ap.parse_activity(data["activity"])
|
|
|
|
if obj.ACTIVITY_TYPE != ActivityType.CREATE:
|
2018-05-31 18:26:23 -05:00
|
|
|
abort(404)
|
|
|
|
|
|
|
|
q = {
|
2018-06-16 15:02:10 -05:00
|
|
|
"meta.undo": False,
|
|
|
|
"type": ActivityType.LIKE.value,
|
|
|
|
"$or": [
|
|
|
|
{"activity.object.id": obj.get_object().id},
|
|
|
|
{"activity.object": obj.get_object().id},
|
|
|
|
],
|
2018-05-31 18:26:23 -05:00
|
|
|
}
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return jsonify(
|
|
|
|
**activitypub.build_ordered_collection(
|
|
|
|
DB.inbox,
|
|
|
|
q=q,
|
|
|
|
cursor=request.args.get("cursor"),
|
|
|
|
map_func=lambda doc: remove_context(doc["activity"]),
|
|
|
|
col_name=f"outbox/{item_id}/likes",
|
|
|
|
first_page=request.args.get("page") == "first",
|
|
|
|
)
|
|
|
|
)
|
2018-05-31 18:26:23 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/outbox/<item_id>/shares")
|
2018-05-31 18:26:23 -05:00
|
|
|
def outbox_activity_shares(item_id):
|
2018-06-03 07:34:04 -05:00
|
|
|
# TODO(tsileo): handle Tombstone
|
2018-05-31 18:26:23 -05:00
|
|
|
if not is_api_request():
|
|
|
|
abort(404)
|
2018-06-18 15:01:21 -05:00
|
|
|
data = DB.outbox.find_one(
|
|
|
|
{"remote_id": back.activity_url(item_id), "meta.deleted": False}
|
|
|
|
)
|
2018-05-31 18:26:23 -05:00
|
|
|
if not data:
|
|
|
|
abort(404)
|
2018-06-17 12:21:59 -05:00
|
|
|
obj = ap.parse_activity(data["activity"])
|
|
|
|
if obj.ACTIVITY_TYPE != ActivityType.CREATE:
|
2018-05-31 18:26:23 -05:00
|
|
|
abort(404)
|
|
|
|
|
|
|
|
q = {
|
2018-06-16 15:02:10 -05:00
|
|
|
"meta.undo": False,
|
|
|
|
"type": ActivityType.ANNOUNCE.value,
|
|
|
|
"$or": [
|
|
|
|
{"activity.object.id": obj.get_object().id},
|
|
|
|
{"activity.object": obj.get_object().id},
|
|
|
|
],
|
2018-05-31 18:26:23 -05:00
|
|
|
}
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return jsonify(
|
|
|
|
**activitypub.build_ordered_collection(
|
|
|
|
DB.inbox,
|
|
|
|
q=q,
|
|
|
|
cursor=request.args.get("cursor"),
|
|
|
|
map_func=lambda doc: remove_context(doc["activity"]),
|
|
|
|
col_name=f"outbox/{item_id}/shares",
|
|
|
|
first_page=request.args.get("page") == "first",
|
|
|
|
)
|
|
|
|
)
|
2018-05-31 18:26:23 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/admin", methods=["GET"])
|
2018-05-18 13:41:41 -05:00
|
|
|
@login_required
|
|
|
|
def admin():
|
2018-06-16 15:02:10 -05:00
|
|
|
q = {"meta.deleted": False, "meta.undo": False, "type": ActivityType.LIKE.value}
|
2018-05-18 13:41:41 -05:00
|
|
|
col_liked = DB.outbox.count(q)
|
|
|
|
|
|
|
|
return render_template(
|
2018-06-16 15:02:10 -05:00
|
|
|
"admin.html",
|
|
|
|
instances=list(DB.instances.find()),
|
|
|
|
inbox_size=DB.inbox.count(),
|
|
|
|
outbox_size=DB.outbox.count(),
|
|
|
|
object_cache_size=DB.objects_cache.count(),
|
|
|
|
actor_cache_size=DB.actors_cache.count(),
|
|
|
|
col_liked=col_liked,
|
|
|
|
col_followers=DB.followers.count(),
|
|
|
|
col_following=DB.following.count(),
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
2018-06-16 14:24:53 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/new", methods=["GET"])
|
2018-05-18 13:41:41 -05:00
|
|
|
@login_required
|
|
|
|
def new():
|
|
|
|
reply_id = None
|
2018-06-16 15:02:10 -05:00
|
|
|
content = ""
|
2018-06-03 14:28:06 -05:00
|
|
|
thread = []
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.args.get("reply"):
|
|
|
|
data = DB.inbox.find_one({"activity.object.id": request.args.get("reply")})
|
2018-06-03 14:28:06 -05:00
|
|
|
if not data:
|
2018-06-16 15:02:10 -05:00
|
|
|
data = DB.outbox.find_one({"activity.object.id": request.args.get("reply")})
|
2018-06-03 14:28:06 -05:00
|
|
|
if not data:
|
|
|
|
abort(400)
|
|
|
|
|
2018-06-17 12:21:59 -05:00
|
|
|
reply = ap.parse_activity(data["activity"])
|
2018-05-18 13:41:41 -05:00
|
|
|
reply_id = reply.id
|
2018-06-17 12:21:59 -05:00
|
|
|
if reply.ACTIVITY_TYPE == ActivityType.CREATE:
|
2018-06-03 14:28:06 -05:00
|
|
|
reply_id = reply.get_object().id
|
2018-05-18 13:41:41 -05:00
|
|
|
actor = reply.get_actor()
|
|
|
|
domain = urlparse(actor.id).netloc
|
2018-06-03 14:28:06 -05:00
|
|
|
# FIXME(tsileo): if reply of reply, fetch all participants
|
2018-06-16 15:02:10 -05:00
|
|
|
content = f"@{actor.preferredUsername}@{domain} "
|
|
|
|
thread = _build_thread(data, include_children=False)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return render_template("new.html", reply=reply_id, content=content, thread=thread)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/notifications")
|
2018-05-18 13:41:41 -05:00
|
|
|
@login_required
|
|
|
|
def notifications():
|
|
|
|
# FIXME(tsileo): implements pagination, also for the followers/following page
|
|
|
|
limit = 50
|
|
|
|
q = {
|
2018-06-16 15:02:10 -05:00
|
|
|
"type": "Create",
|
|
|
|
"activity.object.tag.type": "Mention",
|
|
|
|
"activity.object.tag.name": f"@{USERNAME}@{DOMAIN}",
|
|
|
|
"meta.deleted": False,
|
2018-05-18 13:41:41 -05:00
|
|
|
}
|
|
|
|
# TODO(tsileo): also include replies via regex on Create replyTo
|
2018-06-16 15:02:10 -05:00
|
|
|
q = {
|
|
|
|
"$or": [
|
|
|
|
q,
|
|
|
|
{"type": "Follow"},
|
|
|
|
{"type": "Accept"},
|
|
|
|
{"type": "Undo", "activity.object.type": "Follow"},
|
|
|
|
{"type": "Announce", "activity.object": {"$regex": f"^{BASE_URL}"}},
|
|
|
|
{"type": "Create", "activity.object.inReplyTo": {"$regex": f"^{BASE_URL}"}},
|
|
|
|
]
|
|
|
|
}
|
2018-05-18 13:41:41 -05:00
|
|
|
print(q)
|
2018-06-16 15:02:10 -05:00
|
|
|
c = request.args.get("cursor")
|
2018-05-18 13:41:41 -05:00
|
|
|
if c:
|
2018-06-16 15:02:10 -05:00
|
|
|
q["_id"] = {"$lt": ObjectId(c)}
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
outbox_data = list(DB.inbox.find(q, limit=limit).sort("_id", -1))
|
2018-05-18 13:41:41 -05:00
|
|
|
cursor = None
|
|
|
|
if outbox_data and len(outbox_data) == limit:
|
2018-06-16 15:02:10 -05:00
|
|
|
cursor = str(outbox_data[-1]["_id"])
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
# TODO(tsileo): fix the annonce handling, copy it from /stream
|
2018-06-16 15:02:10 -05:00
|
|
|
# for data in outbox_data:
|
2018-05-18 13:41:41 -05:00
|
|
|
# if data['type'] == 'Announce':
|
|
|
|
# print(data)
|
|
|
|
# if data['activity']['object'].startswith('http') and data['activity']['object'] in objcache:
|
|
|
|
# data['ref'] = {'activity': {'object': objcache[data['activity']['object']]}, 'meta': {}}
|
|
|
|
# out.append(data)
|
|
|
|
# else:
|
|
|
|
# out.append(data)
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return render_template("stream.html", inbox_data=outbox_data, cursor=cursor)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-05-28 12:46:23 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/key")
|
2018-05-29 14:36:05 -05:00
|
|
|
@login_required
|
|
|
|
def api_user_key():
|
|
|
|
return flask_jsonify(api_key=ADMIN_API_KEY)
|
|
|
|
|
|
|
|
|
2018-06-03 07:34:04 -05:00
|
|
|
def _user_api_arg(key: str, **kwargs):
|
2018-06-01 13:29:44 -05:00
|
|
|
"""Try to get the given key from the requests, try JSON body, form data and query arg."""
|
2018-05-29 15:16:09 -05:00
|
|
|
if request.is_json:
|
2018-06-01 13:29:44 -05:00
|
|
|
oid = request.json.get(key)
|
2018-05-29 14:36:05 -05:00
|
|
|
else:
|
2018-06-01 13:29:44 -05:00
|
|
|
oid = request.args.get(key) or request.form.get(key)
|
2018-05-29 14:36:05 -05:00
|
|
|
|
|
|
|
if not oid:
|
2018-06-16 15:02:10 -05:00
|
|
|
if "default" in kwargs:
|
|
|
|
return kwargs.get("default")
|
2018-06-03 05:50:51 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
raise ValueError(f"missing {key}")
|
2018-06-01 13:29:44 -05:00
|
|
|
|
|
|
|
return oid
|
|
|
|
|
|
|
|
|
|
|
|
def _user_api_get_note(from_outbox: bool = False):
|
2018-06-16 15:02:10 -05:00
|
|
|
oid = _user_api_arg("id")
|
2018-06-17 12:21:59 -05:00
|
|
|
note = ap.parse_activity(OBJECT_SERVICE.get(oid), expected=ActivityType.NOTE)
|
2018-06-01 13:29:44 -05:00
|
|
|
if from_outbox and not note.id.startswith(ID):
|
2018-06-16 15:02:10 -05:00
|
|
|
raise NotFromOutboxError(
|
2018-06-17 12:21:59 -05:00
|
|
|
f"cannot load {note.id}, id must be owned by the server"
|
2018-06-16 15:02:10 -05:00
|
|
|
)
|
2018-05-29 14:36:05 -05:00
|
|
|
|
2018-06-01 13:29:44 -05:00
|
|
|
return note
|
2018-05-29 14:36:05 -05:00
|
|
|
|
|
|
|
|
|
|
|
def _user_api_response(**kwargs):
|
2018-06-16 15:02:10 -05:00
|
|
|
_redirect = _user_api_arg("redirect", default=None)
|
2018-06-03 04:41:18 -05:00
|
|
|
if _redirect:
|
|
|
|
return redirect(_redirect)
|
2018-05-29 14:36:05 -05:00
|
|
|
|
|
|
|
resp = flask_jsonify(**kwargs)
|
|
|
|
resp.status_code = 201
|
|
|
|
return resp
|
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/note/delete", methods=["POST"])
|
2018-05-28 15:38:48 -05:00
|
|
|
@api_required
|
|
|
|
def api_delete():
|
2018-05-29 14:36:05 -05:00
|
|
|
"""API endpoint to delete a Note activity."""
|
2018-06-01 13:29:44 -05:00
|
|
|
note = _user_api_get_note(from_outbox=True)
|
|
|
|
|
2018-05-29 14:36:05 -05:00
|
|
|
delete = note.build_delete()
|
2018-06-17 12:21:59 -05:00
|
|
|
OUTBOX.post(delete)
|
2018-05-29 14:36:05 -05:00
|
|
|
|
|
|
|
return _user_api_response(activity=delete.id)
|
|
|
|
|
2018-05-28 15:38:48 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/boost", methods=["POST"])
|
2018-05-28 12:46:23 -05:00
|
|
|
@api_required
|
|
|
|
def api_boost():
|
2018-06-01 13:29:44 -05:00
|
|
|
note = _user_api_get_note()
|
|
|
|
|
2018-06-18 15:01:21 -05:00
|
|
|
announce = note.build_announce(MY_PERSON)
|
2018-06-17 12:21:59 -05:00
|
|
|
OUTBOX.post(announce)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-01 13:29:44 -05:00
|
|
|
return _user_api_response(activity=announce.id)
|
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/like", methods=["POST"])
|
2018-05-28 12:46:23 -05:00
|
|
|
@api_required
|
|
|
|
def api_like():
|
2018-06-01 13:29:44 -05:00
|
|
|
note = _user_api_get_note()
|
|
|
|
|
2018-06-18 15:01:21 -05:00
|
|
|
like = note.build_like(MY_PERSON)
|
2018-06-17 12:21:59 -05:00
|
|
|
OUTBOX.post(like)
|
2018-06-01 13:29:44 -05:00
|
|
|
|
|
|
|
return _user_api_response(activity=like.id)
|
2018-05-28 12:46:23 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/undo", methods=["POST"])
|
2018-05-27 07:21:06 -05:00
|
|
|
@api_required
|
|
|
|
def api_undo():
|
2018-06-16 15:02:10 -05:00
|
|
|
oid = _user_api_arg("id")
|
2018-06-18 15:01:21 -05:00
|
|
|
doc = DB.outbox.find_one(
|
|
|
|
{"$or": [{"remote_id": back.activity_url(oid)}, {"remote_id": oid}]}
|
|
|
|
)
|
2018-06-01 13:29:44 -05:00
|
|
|
if not doc:
|
2018-06-16 15:02:10 -05:00
|
|
|
raise ActivityNotFoundError(f"cannot found {oid}")
|
2018-06-01 13:29:44 -05:00
|
|
|
|
2018-06-17 12:21:59 -05:00
|
|
|
obj = ap.parse_activity(doc.get("activity"))
|
2018-06-01 13:29:44 -05:00
|
|
|
# FIXME(tsileo): detect already undo-ed and make this API call idempotent
|
|
|
|
undo = obj.build_undo()
|
2018-06-17 12:21:59 -05:00
|
|
|
OUTBOX.post(undo)
|
2018-06-01 13:29:44 -05:00
|
|
|
|
|
|
|
return _user_api_response(activity=undo.id)
|
2018-05-27 07:21:06 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/stream")
|
2018-05-18 13:41:41 -05:00
|
|
|
@login_required
|
|
|
|
def stream():
|
|
|
|
# FIXME(tsileo): implements pagination, also for the followers/following page
|
|
|
|
limit = 100
|
|
|
|
q = {
|
2018-06-16 15:02:10 -05:00
|
|
|
"type": "Create",
|
|
|
|
"activity.object.type": "Note",
|
|
|
|
"activity.object.inReplyTo": None,
|
|
|
|
"meta.deleted": False,
|
2018-05-18 13:41:41 -05:00
|
|
|
}
|
2018-06-16 15:02:10 -05:00
|
|
|
c = request.args.get("cursor")
|
2018-05-18 13:41:41 -05:00
|
|
|
if c:
|
2018-06-16 15:02:10 -05:00
|
|
|
q["_id"] = {"$lt": ObjectId(c)}
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
outbox_data = list(
|
|
|
|
DB.inbox.find({"$or": [q, {"type": "Announce"}]}, limit=limit).sort(
|
|
|
|
"activity.published", -1
|
|
|
|
)
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
cursor = None
|
|
|
|
if outbox_data and len(outbox_data) == limit:
|
2018-06-16 15:02:10 -05:00
|
|
|
cursor = str(outbox_data[-1]["_id"])
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
out = []
|
|
|
|
objcache = {}
|
2018-06-16 15:02:10 -05:00
|
|
|
cached = list(
|
|
|
|
DB.objects_cache.find({"meta.part_of_stream": True}, limit=limit * 3).sort(
|
|
|
|
"meta.announce_published", -1
|
|
|
|
)
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
for c in cached:
|
2018-06-16 15:02:10 -05:00
|
|
|
objcache[c["object_id"]] = c["cached_object"]
|
2018-05-18 13:41:41 -05:00
|
|
|
for data in outbox_data:
|
2018-06-16 15:02:10 -05:00
|
|
|
if data["type"] == "Announce":
|
|
|
|
if (
|
|
|
|
data["activity"]["object"].startswith("http")
|
|
|
|
and data["activity"]["object"] in objcache
|
|
|
|
):
|
|
|
|
data["ref"] = {
|
|
|
|
"activity": {"object": objcache[data["activity"]["object"]]},
|
|
|
|
"meta": {},
|
|
|
|
}
|
2018-05-18 13:41:41 -05:00
|
|
|
out.append(data)
|
|
|
|
else:
|
2018-06-16 15:02:10 -05:00
|
|
|
print("OMG", data)
|
2018-05-18 13:41:41 -05:00
|
|
|
else:
|
|
|
|
out.append(data)
|
2018-06-16 15:02:10 -05:00
|
|
|
return render_template("stream.html", inbox_data=out, cursor=cursor)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/inbox", methods=["GET", "POST"])
|
2018-06-16 14:24:53 -05:00
|
|
|
def inbox():
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.method == "GET":
|
2018-06-16 14:24:53 -05:00
|
|
|
if not is_api_request():
|
|
|
|
abort(404)
|
2018-05-21 07:30:52 -05:00
|
|
|
try:
|
|
|
|
_api_required()
|
|
|
|
except BadSignature:
|
|
|
|
abort(404)
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return jsonify(
|
|
|
|
**activitypub.build_ordered_collection(
|
|
|
|
DB.inbox,
|
|
|
|
q={"meta.deleted": False},
|
|
|
|
cursor=request.args.get("cursor"),
|
|
|
|
map_func=lambda doc: remove_context(doc["activity"]),
|
|
|
|
)
|
|
|
|
)
|
2018-05-21 07:30:52 -05:00
|
|
|
|
2018-06-16 14:24:53 -05:00
|
|
|
data = request.get_json(force=True)
|
2018-06-16 15:02:10 -05:00
|
|
|
logger.debug(f"req_headers={request.headers}")
|
|
|
|
logger.debug(f"raw_data={data}")
|
2018-06-20 16:42:12 -05:00
|
|
|
try:
|
|
|
|
if not verify_request(
|
|
|
|
request.method, request.path, request.headers, request.data
|
|
|
|
):
|
2018-06-16 15:02:10 -05:00
|
|
|
raise Exception("failed to verify request")
|
2018-06-02 02:07:57 -05:00
|
|
|
except Exception:
|
2018-06-16 15:02:10 -05:00
|
|
|
logger.exception(
|
|
|
|
"failed to verify request, trying to verify the payload by fetching the remote"
|
|
|
|
)
|
2018-05-21 07:41:47 -05:00
|
|
|
try:
|
2018-06-20 16:42:12 -05:00
|
|
|
data = get_backend().fetch_iri(data["id"])
|
2018-05-21 07:41:47 -05:00
|
|
|
except Exception:
|
2018-05-28 12:46:23 -05:00
|
|
|
logger.exception(f'failed to fetch remote id at {data["id"]}')
|
2018-06-02 02:07:57 -05:00
|
|
|
return Response(
|
|
|
|
status=422,
|
2018-06-16 15:02:10 -05:00
|
|
|
headers={"Content-Type": "application/json"},
|
|
|
|
response=json.dumps(
|
|
|
|
{
|
|
|
|
"error": "failed to verify request (using HTTP signatures or fetching the IRI)"
|
|
|
|
}
|
|
|
|
),
|
2018-06-02 02:07:57 -05:00
|
|
|
)
|
2018-06-17 12:21:59 -05:00
|
|
|
activity = ap.parse_activity(data)
|
2018-06-16 15:02:10 -05:00
|
|
|
logger.debug(f"inbox activity={activity}/{data}")
|
2018-06-17 12:21:59 -05:00
|
|
|
INBOX.post(activity)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return Response(status=201)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-18 15:01:21 -05:00
|
|
|
def without_id(l):
|
|
|
|
out = []
|
|
|
|
for d in l:
|
|
|
|
if "_id" in d:
|
|
|
|
del d["_id"]
|
|
|
|
out.append(d)
|
|
|
|
return out
|
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/debug", methods=["GET", "DELETE"])
|
2018-05-27 04:50:09 -05:00
|
|
|
@api_required
|
|
|
|
def api_debug():
|
|
|
|
"""Endpoint used/needed for testing, only works in DEBUG_MODE."""
|
|
|
|
if not DEBUG_MODE:
|
2018-06-16 15:02:10 -05:00
|
|
|
return flask_jsonify(message="DEBUG_MODE is off")
|
2018-05-27 04:50:09 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.method == "DELETE":
|
2018-05-27 04:50:09 -05:00
|
|
|
_drop_db()
|
2018-06-16 15:02:10 -05:00
|
|
|
return flask_jsonify(message="DB dropped")
|
2018-05-27 04:50:09 -05:00
|
|
|
|
2018-06-18 15:01:21 -05:00
|
|
|
return flask_jsonify(
|
|
|
|
inbox=DB.inbox.count(),
|
|
|
|
outbox=DB.outbox.count(),
|
|
|
|
outbox_data=without_id(DB.outbox.find()),
|
|
|
|
)
|
2018-05-27 04:50:09 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/upload", methods=["POST"])
|
2018-05-18 13:41:41 -05:00
|
|
|
@api_required
|
|
|
|
def api_upload():
|
2018-06-16 15:02:10 -05:00
|
|
|
file = request.files["file"]
|
2018-05-18 13:41:41 -05:00
|
|
|
rfilename = secure_filename(file.filename)
|
|
|
|
prefix = hashlib.sha256(os.urandom(32)).hexdigest()[:6]
|
|
|
|
mtype = mimetypes.guess_type(rfilename)[0]
|
2018-06-16 15:02:10 -05:00
|
|
|
filename = f"{prefix}_{rfilename}"
|
|
|
|
file.save(os.path.join("static", "media", filename))
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
# Remove EXIF metadata
|
2018-06-16 15:02:10 -05:00
|
|
|
if filename.lower().endswith(".jpg") or filename.lower().endswith(".jpeg"):
|
|
|
|
piexif.remove(os.path.join("static", "media", filename))
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
print("upload OK")
|
2018-05-18 13:41:41 -05:00
|
|
|
print(filename)
|
|
|
|
attachment = [
|
2018-06-16 15:02:10 -05:00
|
|
|
{
|
|
|
|
"mediaType": mtype,
|
|
|
|
"name": rfilename,
|
|
|
|
"type": "Document",
|
|
|
|
"url": BASE_URL + f"/static/media/{filename}",
|
|
|
|
}
|
2018-05-18 13:41:41 -05:00
|
|
|
]
|
|
|
|
print(attachment)
|
2018-06-16 15:02:10 -05:00
|
|
|
content = request.args.get("content")
|
|
|
|
to = request.args.get("to")
|
2018-06-17 12:21:59 -05:00
|
|
|
note = ap.Note(
|
2018-06-17 14:54:16 -05:00
|
|
|
attributedTo=MY_PERSON.id,
|
2018-06-16 15:02:10 -05:00
|
|
|
cc=[ID + "/followers"],
|
2018-06-17 12:21:59 -05:00
|
|
|
to=[to if to else ap.AS_PUBLIC],
|
2018-05-18 13:41:41 -05:00
|
|
|
content=content, # TODO(tsileo): handle markdown
|
|
|
|
attachment=attachment,
|
|
|
|
)
|
2018-06-16 15:02:10 -05:00
|
|
|
print("post_note_init")
|
2018-05-18 13:41:41 -05:00
|
|
|
print(note)
|
|
|
|
create = note.build_create()
|
|
|
|
print(create)
|
|
|
|
print(create.to_dict())
|
2018-06-17 12:21:59 -05:00
|
|
|
OUTBOX.post(create)
|
2018-06-16 15:02:10 -05:00
|
|
|
print("posted")
|
2018-06-16 14:24:53 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return Response(status=201, response="OK")
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/new_note", methods=["POST"])
|
2018-06-16 14:24:53 -05:00
|
|
|
@api_required
|
|
|
|
def api_new_note():
|
2018-06-16 15:02:10 -05:00
|
|
|
source = _user_api_arg("content")
|
2018-05-26 02:50:59 -05:00
|
|
|
if not source:
|
2018-06-16 15:02:10 -05:00
|
|
|
raise ValueError("missing content")
|
2018-06-16 14:24:53 -05:00
|
|
|
|
2018-06-01 13:29:44 -05:00
|
|
|
_reply, reply = None, None
|
|
|
|
try:
|
2018-06-16 15:02:10 -05:00
|
|
|
_reply = _user_api_arg("reply")
|
2018-06-01 13:29:44 -05:00
|
|
|
except ValueError:
|
|
|
|
pass
|
|
|
|
|
2018-06-16 14:24:53 -05:00
|
|
|
content, tags = parse_markdown(source)
|
2018-06-16 15:02:10 -05:00
|
|
|
to = request.args.get("to")
|
|
|
|
cc = [ID + "/followers"]
|
2018-06-16 14:24:53 -05:00
|
|
|
|
2018-06-01 13:29:44 -05:00
|
|
|
if _reply:
|
2018-06-17 12:21:59 -05:00
|
|
|
reply = ap.parse_activity(OBJECT_SERVICE.get(_reply))
|
2018-05-28 17:12:44 -05:00
|
|
|
cc.append(reply.attributedTo)
|
2018-06-01 13:29:44 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
for tag in tags:
|
2018-06-16 15:02:10 -05:00
|
|
|
if tag["type"] == "Mention":
|
|
|
|
cc.append(tag["href"])
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-17 12:21:59 -05:00
|
|
|
note = ap.Note(
|
2018-06-17 14:54:16 -05:00
|
|
|
attributedTo=MY_PERSON.id,
|
2018-06-16 14:24:53 -05:00
|
|
|
cc=list(set(cc)),
|
2018-06-17 12:21:59 -05:00
|
|
|
to=[to if to else ap.AS_PUBLIC],
|
2018-06-01 13:29:44 -05:00
|
|
|
content=content,
|
2018-05-18 13:41:41 -05:00
|
|
|
tag=tags,
|
2018-06-16 15:02:10 -05:00
|
|
|
source={"mediaType": "text/markdown", "content": source},
|
|
|
|
inReplyTo=reply.id if reply else None,
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
|
|
|
create = note.build_create()
|
2018-06-17 12:21:59 -05:00
|
|
|
OUTBOX.post(create)
|
2018-05-28 17:12:44 -05:00
|
|
|
|
2018-06-01 13:29:44 -05:00
|
|
|
return _user_api_response(activity=create.id)
|
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/stream")
|
2018-05-27 15:30:43 -05:00
|
|
|
@api_required
|
2018-05-18 13:41:41 -05:00
|
|
|
def api_stream():
|
|
|
|
return Response(
|
2018-06-16 15:02:10 -05:00
|
|
|
response=json.dumps(
|
|
|
|
activitypub.build_inbox_json_feed("/api/stream", request.args.get("cursor"))
|
|
|
|
),
|
|
|
|
headers={"Content-Type": "application/json"},
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
|
|
|
|
2018-05-28 17:12:44 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/block", methods=["POST"])
|
2018-05-28 17:12:44 -05:00
|
|
|
@api_required
|
|
|
|
def api_block():
|
2018-06-16 15:02:10 -05:00
|
|
|
actor = _user_api_arg("actor")
|
2018-06-01 13:29:44 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
existing = DB.outbox.find_one(
|
|
|
|
{"type": ActivityType.BLOCK.value, "activity.object": actor, "meta.undo": False}
|
|
|
|
)
|
2018-06-01 13:29:44 -05:00
|
|
|
if existing:
|
2018-06-16 15:02:10 -05:00
|
|
|
return _user_api_response(activity=existing["activity"]["id"])
|
2018-05-28 17:12:44 -05:00
|
|
|
|
2018-06-17 14:54:16 -05:00
|
|
|
block = ap.Block(actor=MY_PERSON.id, object=actor)
|
2018-06-17 12:21:59 -05:00
|
|
|
OUTBOX.post(block)
|
2018-06-01 13:29:44 -05:00
|
|
|
|
|
|
|
return _user_api_response(activity=block.id)
|
2018-05-28 17:12:44 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/api/follow", methods=["POST"])
|
2018-05-18 13:41:41 -05:00
|
|
|
@api_required
|
|
|
|
def api_follow():
|
2018-06-16 15:02:10 -05:00
|
|
|
actor = _user_api_arg("actor")
|
2018-06-01 13:29:44 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
existing = DB.following.find_one({"remote_actor": actor})
|
2018-06-01 13:29:44 -05:00
|
|
|
if existing:
|
2018-06-16 15:02:10 -05:00
|
|
|
return _user_api_response(activity=existing["activity"]["id"])
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-17 14:54:16 -05:00
|
|
|
follow = ap.Follow(actor=MY_PERSON.id, object=actor)
|
2018-06-17 12:21:59 -05:00
|
|
|
OUTBOX.post(follow)
|
2018-06-01 13:29:44 -05:00
|
|
|
|
|
|
|
return _user_api_response(activity=follow.id)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/followers")
|
2018-06-16 14:24:53 -05:00
|
|
|
def followers():
|
|
|
|
if is_api_request():
|
2018-05-18 13:41:41 -05:00
|
|
|
return jsonify(
|
|
|
|
**activitypub.build_ordered_collection(
|
|
|
|
DB.followers,
|
2018-06-16 15:02:10 -05:00
|
|
|
cursor=request.args.get("cursor"),
|
|
|
|
map_func=lambda doc: doc["remote_actor"],
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
2018-06-16 14:24:53 -05:00
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
followers = [
|
|
|
|
ACTOR_SERVICE.get(doc["remote_actor"]) for doc in DB.followers.find(limit=50)
|
|
|
|
]
|
2018-06-16 14:24:53 -05:00
|
|
|
return render_template(
|
2018-06-16 15:02:10 -05:00
|
|
|
"followers.html",
|
2018-05-18 13:41:41 -05:00
|
|
|
me=ME,
|
2018-06-16 15:02:10 -05:00
|
|
|
notes=DB.inbox.find({"object.object.type": "Note"}).count(),
|
2018-05-18 13:41:41 -05:00
|
|
|
followers=DB.followers.count(),
|
|
|
|
following=DB.following.count(),
|
|
|
|
followers_data=followers,
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/following")
|
2018-05-18 13:41:41 -05:00
|
|
|
def following():
|
|
|
|
if is_api_request():
|
|
|
|
return jsonify(
|
|
|
|
**activitypub.build_ordered_collection(
|
|
|
|
DB.following,
|
2018-06-16 15:02:10 -05:00
|
|
|
cursor=request.args.get("cursor"),
|
|
|
|
map_func=lambda doc: doc["remote_actor"],
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
2018-06-16 14:24:53 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
following = [
|
|
|
|
ACTOR_SERVICE.get(doc["remote_actor"]) for doc in DB.following.find(limit=50)
|
|
|
|
]
|
2018-05-18 13:41:41 -05:00
|
|
|
return render_template(
|
2018-06-16 15:02:10 -05:00
|
|
|
"following.html",
|
2018-05-18 13:41:41 -05:00
|
|
|
me=ME,
|
2018-06-16 15:02:10 -05:00
|
|
|
notes=DB.inbox.find({"object.object.type": "Note"}).count(),
|
2018-05-18 13:41:41 -05:00
|
|
|
followers=DB.followers.count(),
|
|
|
|
following=DB.following.count(),
|
|
|
|
following_data=following,
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/tags/<tag>")
|
2018-05-18 13:41:41 -05:00
|
|
|
def tags(tag):
|
2018-06-16 15:02:10 -05:00
|
|
|
if not DB.outbox.count(
|
|
|
|
{"activity.object.tag.type": "Hashtag", "activity.object.tag.name": "#" + tag}
|
|
|
|
):
|
2018-05-18 13:41:41 -05:00
|
|
|
abort(404)
|
|
|
|
if not is_api_request():
|
|
|
|
return render_template(
|
2018-06-16 15:02:10 -05:00
|
|
|
"tags.html",
|
2018-05-18 13:41:41 -05:00
|
|
|
tag=tag,
|
2018-06-16 15:02:10 -05:00
|
|
|
outbox_data=DB.outbox.find(
|
|
|
|
{
|
|
|
|
"type": "Create",
|
|
|
|
"activity.object.type": "Note",
|
|
|
|
"meta.deleted": False,
|
|
|
|
"activity.object.tag.type": "Hashtag",
|
|
|
|
"activity.object.tag.name": "#" + tag,
|
|
|
|
}
|
|
|
|
),
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
|
|
|
q = {
|
2018-06-16 15:02:10 -05:00
|
|
|
"meta.deleted": False,
|
|
|
|
"meta.undo": False,
|
|
|
|
"type": ActivityType.CREATE.value,
|
|
|
|
"activity.object.tag.type": "Hashtag",
|
|
|
|
"activity.object.tag.name": "#" + tag,
|
2018-05-18 13:41:41 -05:00
|
|
|
}
|
2018-06-16 15:02:10 -05:00
|
|
|
return jsonify(
|
|
|
|
**activitypub.build_ordered_collection(
|
|
|
|
DB.outbox,
|
|
|
|
q=q,
|
|
|
|
cursor=request.args.get("cursor"),
|
|
|
|
map_func=lambda doc: doc["activity"]["object"]["id"],
|
|
|
|
col_name=f"tags/{tag}",
|
|
|
|
)
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/liked")
|
2018-05-18 13:41:41 -05:00
|
|
|
def liked():
|
|
|
|
if not is_api_request():
|
|
|
|
abort(404)
|
2018-06-16 15:02:10 -05:00
|
|
|
q = {"meta.deleted": False, "meta.undo": False, "type": ActivityType.LIKE.value}
|
|
|
|
return jsonify(
|
|
|
|
**activitypub.build_ordered_collection(
|
|
|
|
DB.outbox,
|
|
|
|
q=q,
|
|
|
|
cursor=request.args.get("cursor"),
|
|
|
|
map_func=lambda doc: doc["activity"]["object"],
|
|
|
|
col_name="liked",
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
#######
|
|
|
|
# IndieAuth
|
|
|
|
|
|
|
|
|
|
|
|
def build_auth_resp(payload):
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.headers.get("Accept") == "application/json":
|
2018-05-18 13:41:41 -05:00
|
|
|
return Response(
|
|
|
|
status=200,
|
2018-06-16 15:02:10 -05:00
|
|
|
headers={"Content-Type": "application/json"},
|
2018-05-18 13:41:41 -05:00
|
|
|
response=json.dumps(payload),
|
|
|
|
)
|
|
|
|
return Response(
|
|
|
|
status=200,
|
2018-06-16 15:02:10 -05:00
|
|
|
headers={"Content-Type": "application/x-www-form-urlencoded"},
|
2018-05-18 13:41:41 -05:00
|
|
|
response=urlencode(payload),
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def _get_prop(props, name, default=None):
|
|
|
|
if name in props:
|
|
|
|
items = props.get(name)
|
|
|
|
if isinstance(items, list):
|
|
|
|
return items[0]
|
|
|
|
return items
|
|
|
|
return default
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
|
2018-05-18 13:41:41 -05:00
|
|
|
def get_client_id_data(url):
|
|
|
|
data = mf2py.parse(url=url)
|
2018-06-16 15:02:10 -05:00
|
|
|
for item in data["items"]:
|
|
|
|
if "h-x-app" in item["type"] or "h-app" in item["type"]:
|
|
|
|
props = item.get("properties", {})
|
2018-05-18 13:41:41 -05:00
|
|
|
print(props)
|
|
|
|
return dict(
|
2018-06-16 15:02:10 -05:00
|
|
|
logo=_get_prop(props, "logo"),
|
|
|
|
name=_get_prop(props, "name"),
|
|
|
|
url=_get_prop(props, "url"),
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
2018-06-16 15:02:10 -05:00
|
|
|
return dict(logo=None, name=url, url=url)
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/indieauth/flow", methods=["POST"])
|
2018-06-16 14:24:53 -05:00
|
|
|
@login_required
|
|
|
|
def indieauth_flow():
|
|
|
|
auth = dict(
|
2018-06-16 15:02:10 -05:00
|
|
|
scope=" ".join(request.form.getlist("scopes")),
|
|
|
|
me=request.form.get("me"),
|
|
|
|
client_id=request.form.get("client_id"),
|
|
|
|
state=request.form.get("state"),
|
|
|
|
redirect_uri=request.form.get("redirect_uri"),
|
|
|
|
response_type=request.form.get("response_type"),
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
code = binascii.hexlify(os.urandom(8)).decode("utf-8")
|
|
|
|
auth.update(code=code, verified=False)
|
2018-05-18 13:41:41 -05:00
|
|
|
print(auth)
|
2018-06-16 15:02:10 -05:00
|
|
|
if not auth["redirect_uri"]:
|
2018-05-18 13:41:41 -05:00
|
|
|
abort(500)
|
|
|
|
|
|
|
|
DB.indieauth.insert_one(auth)
|
|
|
|
|
|
|
|
# FIXME(tsileo): fetch client ID and validate redirect_uri
|
|
|
|
red = f'{auth["redirect_uri"]}?code={code}&state={auth["state"]}&me={auth["me"]}'
|
|
|
|
return redirect(red)
|
|
|
|
|
|
|
|
|
2018-06-16 14:24:53 -05:00
|
|
|
# @app.route('/indieauth', methods=['GET', 'POST'])
|
|
|
|
def indieauth_endpoint():
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.method == "GET":
|
|
|
|
if not session.get("logged_in"):
|
|
|
|
return redirect(url_for("login", next=request.url))
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
me = request.args.get("me")
|
2018-06-16 14:24:53 -05:00
|
|
|
# FIXME(tsileo): ensure me == ID
|
2018-06-16 15:02:10 -05:00
|
|
|
client_id = request.args.get("client_id")
|
|
|
|
redirect_uri = request.args.get("redirect_uri")
|
|
|
|
state = request.args.get("state", "")
|
|
|
|
response_type = request.args.get("response_type", "id")
|
|
|
|
scope = request.args.get("scope", "").split()
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
print("STATE", state)
|
2018-05-18 13:41:41 -05:00
|
|
|
return render_template(
|
2018-06-16 15:02:10 -05:00
|
|
|
"indieauth_flow.html",
|
2018-05-18 13:41:41 -05:00
|
|
|
client=get_client_id_data(client_id),
|
|
|
|
scopes=scope,
|
|
|
|
redirect_uri=redirect_uri,
|
|
|
|
state=state,
|
|
|
|
response_type=response_type,
|
|
|
|
client_id=client_id,
|
|
|
|
me=me,
|
|
|
|
)
|
|
|
|
|
|
|
|
# Auth verification via POST
|
2018-06-16 15:02:10 -05:00
|
|
|
code = request.form.get("code")
|
|
|
|
redirect_uri = request.form.get("redirect_uri")
|
|
|
|
client_id = request.form.get("client_id")
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
auth = DB.indieauth.find_one_and_update(
|
2018-06-16 15:02:10 -05:00
|
|
|
{
|
|
|
|
"code": code,
|
|
|
|
"redirect_uri": redirect_uri,
|
|
|
|
"client_id": client_id,
|
|
|
|
}, # }, # , 'verified': False},
|
|
|
|
{"$set": {"verified": True}},
|
|
|
|
sort=[("_id", pymongo.DESCENDING)],
|
2018-05-18 13:41:41 -05:00
|
|
|
)
|
|
|
|
print(auth)
|
|
|
|
print(code, redirect_uri, client_id)
|
|
|
|
|
|
|
|
if not auth:
|
|
|
|
abort(403)
|
|
|
|
return
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
session["logged_in"] = True
|
|
|
|
me = auth["me"]
|
|
|
|
state = auth["state"]
|
|
|
|
scope = " ".join(auth["scope"])
|
|
|
|
print("STATE", state)
|
|
|
|
return build_auth_resp({"me": me, "state": state, "scope": scope})
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
@app.route("/token", methods=["GET", "POST"])
|
2018-05-18 13:41:41 -05:00
|
|
|
def token_endpoint():
|
2018-06-16 15:02:10 -05:00
|
|
|
if request.method == "POST":
|
|
|
|
code = request.form.get("code")
|
|
|
|
me = request.form.get("me")
|
|
|
|
redirect_uri = request.form.get("redirect_uri")
|
|
|
|
client_id = request.form.get("client_id")
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
auth = DB.indieauth.find_one(
|
|
|
|
{
|
|
|
|
"code": code,
|
|
|
|
"me": me,
|
|
|
|
"redirect_uri": redirect_uri,
|
|
|
|
"client_id": client_id,
|
|
|
|
}
|
|
|
|
)
|
2018-05-18 13:41:41 -05:00
|
|
|
if not auth:
|
|
|
|
abort(403)
|
2018-06-16 15:02:10 -05:00
|
|
|
scope = " ".join(auth["scope"])
|
|
|
|
payload = dict(
|
|
|
|
me=me, client_id=client_id, scope=scope, ts=datetime.now().timestamp()
|
|
|
|
)
|
|
|
|
token = JWT.dumps(payload).decode("utf-8")
|
2018-05-18 13:41:41 -05:00
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return build_auth_resp({"me": me, "scope": scope, "access_token": token})
|
2018-05-18 13:41:41 -05:00
|
|
|
|
|
|
|
# Token verification
|
2018-06-16 15:02:10 -05:00
|
|
|
token = request.headers.get("Authorization").replace("Bearer ", "")
|
2018-05-18 13:41:41 -05:00
|
|
|
try:
|
|
|
|
payload = JWT.loads(token)
|
|
|
|
except BadSignature:
|
|
|
|
abort(403)
|
|
|
|
|
|
|
|
# TODO(tsileo): handle expiration
|
|
|
|
|
2018-06-16 15:02:10 -05:00
|
|
|
return build_auth_resp(
|
|
|
|
{
|
|
|
|
"me": payload["me"],
|
|
|
|
"scope": payload["scope"],
|
|
|
|
"client_id": payload["client_id"],
|
|
|
|
}
|
|
|
|
)
|