Compare commits
No commits in common. "280d74b4e93825ff1c2091393522c6fa9bfa2975" and "e32310da73c64575609e1b76e8f4b4ec8f8698b5" have entirely different histories.
280d74b4e9
...
e32310da73
14 changed files with 584 additions and 2547 deletions
|
@ -75,32 +75,6 @@ ME = {
|
||||||
"tag": [] # TODO tag support
|
"tag": [] # TODO tag support
|
||||||
}
|
}
|
||||||
|
|
||||||
class BaseActor:
|
|
||||||
def __init__(self, ap_actor: RawObject) -> None:
|
|
||||||
if (ap_type := ap_actor.get("type")) not in ACTOR_TYPES:
|
|
||||||
raise ValueError(f"Unexpected actor type: {ap_type}")
|
|
||||||
|
|
||||||
self._ap_actor = ap_actor
|
|
||||||
self._ap_type : str = ap_type # type: ignore
|
|
||||||
|
|
||||||
@property
|
|
||||||
def ap_actor(self) -> RawObject:
|
|
||||||
return self._ap_actor
|
|
||||||
|
|
||||||
@property
|
|
||||||
def inbox_url(self) -> str:
|
|
||||||
return self.ap_actor["inbox"]
|
|
||||||
|
|
||||||
@property
|
|
||||||
def ap_type(self) -> str:
|
|
||||||
return self._ap_type
|
|
||||||
|
|
||||||
@property
|
|
||||||
def share_inbox_url(self) -> str:
|
|
||||||
return self.ap_actor.get("endpoints", {}).get("sharedInbox") \
|
|
||||||
or self.inbox_url
|
|
||||||
|
|
||||||
|
|
||||||
class VisibilityEnum(str, enum.Enum):
|
class VisibilityEnum(str, enum.Enum):
|
||||||
PUBLIC = "public"
|
PUBLIC = "public"
|
||||||
UNLISTED = "unlisted"
|
UNLISTED = "unlisted"
|
||||||
|
@ -121,26 +95,6 @@ def handle_visibility(
|
||||||
return VisibilityEnum.DIRECT
|
return VisibilityEnum.DIRECT
|
||||||
|
|
||||||
|
|
||||||
def wrap_ap_object(ap_object: dict) -> dict:
|
|
||||||
if ap_object["type"] in ["Note"]:
|
|
||||||
|
|
||||||
if "@context" in ap_object:
|
|
||||||
del ap_object["@context"]
|
|
||||||
|
|
||||||
return {
|
|
||||||
"@context": AS_EXTENDED_CTX,
|
|
||||||
"actor": config.ID,
|
|
||||||
"to": ap_object.get("to", []),
|
|
||||||
"cc": ap_object.get("cc", []),
|
|
||||||
"id": ap_object["id"] + "/activity",
|
|
||||||
"object": ap_object,
|
|
||||||
"published": ap_object["published"],
|
|
||||||
"type": "Create",
|
|
||||||
}
|
|
||||||
|
|
||||||
return ap_object
|
|
||||||
|
|
||||||
|
|
||||||
async def post(
|
async def post(
|
||||||
url: str,
|
url: str,
|
||||||
payload : dict,
|
payload : dict,
|
||||||
|
@ -158,7 +112,6 @@ async def post(
|
||||||
auth=auth,
|
auth=auth,
|
||||||
)
|
)
|
||||||
|
|
||||||
resp.raise_for_status()
|
|
||||||
return resp
|
return resp
|
||||||
|
|
||||||
|
|
||||||
|
|
40
app/actor.py
40
app/actor.py
|
@ -1,6 +1,5 @@
|
||||||
#!/usr/bin/env python3
|
#!/usr/bin/env python3
|
||||||
import typing
|
import typing
|
||||||
import json
|
|
||||||
from loguru import logger
|
from loguru import logger
|
||||||
from app.database import AsyncSession
|
from app.database import AsyncSession
|
||||||
from app import models
|
from app import models
|
||||||
|
@ -13,6 +12,26 @@ if typing.TYPE_CHECKING:
|
||||||
import app.activitypub as ap
|
import app.activitypub as ap
|
||||||
|
|
||||||
|
|
||||||
|
class BaseActor:
|
||||||
|
def __init__(self, ap_actor: ap.RawObject) -> None:
|
||||||
|
if (ap_type := ap_actor.get("type")) not in ap.ACTOR_TYPES:
|
||||||
|
raise ValueError(f"Unexpected actor type: {ap_type}")
|
||||||
|
|
||||||
|
self._ap_actor = ap_actor
|
||||||
|
self._ap_type : str = ap_type # type: ignore
|
||||||
|
|
||||||
|
@property
|
||||||
|
def ap_actor(self) -> ap.RawObject:
|
||||||
|
return self._ap_actor
|
||||||
|
|
||||||
|
@property
|
||||||
|
def inbox_url(self) -> str:
|
||||||
|
return self.ap_actor["inbox"]
|
||||||
|
|
||||||
|
@property
|
||||||
|
def ap_type(self) -> str:
|
||||||
|
return self._ap_type
|
||||||
|
|
||||||
|
|
||||||
async def fetch_actor(
|
async def fetch_actor(
|
||||||
db_session : AsyncSession,
|
db_session : AsyncSession,
|
||||||
|
@ -30,8 +49,8 @@ async def fetch_actor(
|
||||||
ap_object = await ap.fetch(actor_id)
|
ap_object = await ap.fetch(actor_id)
|
||||||
exist_actor = await save_actor(ap_object, db_session)
|
exist_actor = await save_actor(ap_object, db_session)
|
||||||
return exist_actor
|
return exist_actor
|
||||||
else:
|
|
||||||
return exist_actor
|
return exist_actor
|
||||||
|
|
||||||
async def save_actor(
|
async def save_actor(
|
||||||
ap_object : dict,
|
ap_object : dict,
|
||||||
|
@ -60,18 +79,3 @@ def _handle (
|
||||||
handle = '@' + ap_object["preferredUsername"] + '@' + ap_id.hostname
|
handle = '@' + ap_object["preferredUsername"] + '@' + ap_id.hostname
|
||||||
|
|
||||||
return handle
|
return handle
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
async def get_public_key(
|
|
||||||
db_session: AsyncSession,
|
|
||||||
key_id: str
|
|
||||||
) -> str:
|
|
||||||
|
|
||||||
existing_actor = (
|
|
||||||
await db_session.scalars(
|
|
||||||
select(models.Actor).where(models.Actor.ap_id == key_id.split("#")[0])
|
|
||||||
)
|
|
||||||
).one_or_none()
|
|
||||||
public_key = existing_actor.ap_object["publicKey"]["publicKeyPem"]
|
|
||||||
return public_key
|
|
||||||
|
|
138
app/boxes.py
138
app/boxes.py
|
@ -5,16 +5,14 @@ import uuid
|
||||||
from sqlalchemy.orm import session
|
from sqlalchemy.orm import session
|
||||||
|
|
||||||
from app import models
|
from app import models
|
||||||
from app import ldsig
|
|
||||||
from app.database import AsyncSession
|
from app.database import AsyncSession
|
||||||
from app.models import InboxObject, OutboxObject, now
|
from app.models import InboxObject, OutboxObject, now
|
||||||
from app.activitypub import ME
|
from app.activitypub import ME
|
||||||
from app.activitypub import handle_visibility
|
from app.activitypub import handle_visibility
|
||||||
from app.config import MANUALLY_APPROVES_FOLLOWERS
|
from app.config import MANUALLY_APPROVES_FOLLOWERS
|
||||||
from app.config import BASE_URL, ID
|
from app.config import BASE_URL
|
||||||
from app.models import Actor
|
from app.models import Actor
|
||||||
from app.actor import fetch_actor
|
from app.actor import fetch_actor
|
||||||
from app.httpsig import k
|
|
||||||
|
|
||||||
import app.activitypub as ap
|
import app.activitypub as ap
|
||||||
|
|
||||||
|
@ -25,7 +23,6 @@ from sqlalchemy.orm import joinedload
|
||||||
from sqlalchemy.exc import IntegrityError
|
from sqlalchemy.exc import IntegrityError
|
||||||
from loguru import logger
|
from loguru import logger
|
||||||
from uuid import uuid4
|
from uuid import uuid4
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
@ -136,8 +133,6 @@ async def process_incoming(
|
||||||
await db_session.flush()
|
await db_session.flush()
|
||||||
await db_session.refresh(following)
|
await db_session.refresh(following)
|
||||||
return True
|
return True
|
||||||
# elif "Creat" == ap_object["type"]:
|
|
||||||
|
|
||||||
|
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
@ -178,21 +173,18 @@ async def _send_accept(
|
||||||
await db_session.rollback()
|
await db_session.rollback()
|
||||||
logger.warning("existing follower in db!")
|
logger.warning("existing follower in db!")
|
||||||
|
|
||||||
try:
|
reply_id = allocate_outbox_id()
|
||||||
reply_id = allocate_outbox_id()
|
|
||||||
|
|
||||||
url = actor.inbox_url # type: ignore
|
url = actor.inbox_url # type: ignore
|
||||||
out = {
|
out = {
|
||||||
"@context": ap.AS_CTX,
|
"@context": ap.AS_CTX,
|
||||||
"id": build_object_id(reply_id),
|
"id": build_object_id(reply_id),
|
||||||
"type": "Accept",
|
"type": "Accept",
|
||||||
"actor": ME["id"],
|
"actor": ME["id"],
|
||||||
"object": inbox_object.ap_object["id"], #type: ignore
|
"object": inbox_object.ap_object["id"], #type: ignore
|
||||||
}
|
}
|
||||||
#TODO outcoming
|
#TODO outcoming
|
||||||
await ap.post(url, out) # type: ignore
|
await ap.post(url, out) # type: ignore
|
||||||
except Exception as e:
|
|
||||||
logger.error(e)
|
|
||||||
|
|
||||||
|
|
||||||
async def _handle_undo(
|
async def _handle_undo(
|
||||||
|
@ -268,112 +260,6 @@ async def _send_follow(
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
async def _send_create(
|
|
||||||
db_session: AsyncSession,
|
|
||||||
ap_type: str,
|
|
||||||
content: str,
|
|
||||||
visibility: ap.VisibilityEnum,
|
|
||||||
published: str | None = None,
|
|
||||||
) -> bool:
|
|
||||||
object_id = build_object_id(allocate_outbox_id())
|
|
||||||
if not published:
|
|
||||||
published = now().replace(microsecond=0).isoformat().replace("+00:00", "Z")
|
|
||||||
|
|
||||||
to = []
|
|
||||||
cc = []
|
|
||||||
|
|
||||||
if visibility == ap.VisibilityEnum.PUBLIC:
|
|
||||||
to = [ap.AS_PUBLIC]
|
|
||||||
cc = [f"{BASE_URL}/followers"]
|
|
||||||
else:
|
|
||||||
raise ValueError(f"Unsupport visibility {visibility}")
|
|
||||||
|
|
||||||
ap_object = {
|
|
||||||
"@context": ap.AS_EXTENDED_CTX,
|
|
||||||
"type": ap_type,
|
|
||||||
"id": object_id,
|
|
||||||
"attributedTo": ID,
|
|
||||||
"content": content,
|
|
||||||
"to": to,
|
|
||||||
"cc": cc,
|
|
||||||
"published": published,
|
|
||||||
# "context": context,
|
|
||||||
# "conversation": context,
|
|
||||||
"url": object_id,
|
|
||||||
"tag": [],
|
|
||||||
"summary": None,
|
|
||||||
"inReplyTo": None,
|
|
||||||
"sensitive": False,
|
|
||||||
"attachment": [],
|
|
||||||
}
|
|
||||||
|
|
||||||
outbox_object = await save_to_outbox(
|
|
||||||
db_session,
|
|
||||||
object_id,
|
|
||||||
ap_object,
|
|
||||||
)
|
|
||||||
|
|
||||||
recipients = await _compute_recipients(db_session, ap_object)
|
|
||||||
ap_object = ap.wrap_ap_object(ap_object)
|
|
||||||
|
|
||||||
if ap_object["type"] == "Create":
|
|
||||||
if ap.VisibilityEnum.PUBLIC == outbox_object.visibility:
|
|
||||||
ldsig.generate_signature(ap_object, k)
|
|
||||||
|
|
||||||
for r in recipients:
|
|
||||||
await ap.post(
|
|
||||||
r,
|
|
||||||
ap_object,
|
|
||||||
)
|
|
||||||
|
|
||||||
return True
|
|
||||||
|
|
||||||
|
|
||||||
async def _compute_recipients(
|
|
||||||
db_session: AsyncSession,
|
|
||||||
ap_object: dict,
|
|
||||||
) -> set[str]:
|
|
||||||
|
|
||||||
async def process_collection(
|
|
||||||
db_session,
|
|
||||||
url) -> list[Actor]:
|
|
||||||
if url == BASE_URL + "/followers":
|
|
||||||
followers = (
|
|
||||||
(
|
|
||||||
await db_session.scalars(
|
|
||||||
select(models.Follower).options(
|
|
||||||
joinedload(models.Follower.actor)
|
|
||||||
)
|
|
||||||
)
|
|
||||||
)
|
|
||||||
.unique()
|
|
||||||
.all()
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
raise ValueError(f"{url}) not supported")
|
|
||||||
|
|
||||||
return [follower.actor for follower in followers]
|
|
||||||
|
|
||||||
_recipients = []
|
|
||||||
for field in ["to", "cc", "bcc", "bto"]:
|
|
||||||
if field in ap_object:
|
|
||||||
_recipients.extend(ap_object[field])
|
|
||||||
|
|
||||||
recipients = set()
|
|
||||||
logger.info(f"{_recipients}")
|
|
||||||
for r in _recipients:
|
|
||||||
if r in [ap.AS_PUBLIC, ID]:
|
|
||||||
continue
|
|
||||||
|
|
||||||
if r.startswith(BASE_URL):
|
|
||||||
for actor in await process_collection(db_session, r):
|
|
||||||
recipients.add(actor.share_inbox_url)
|
|
||||||
|
|
||||||
continue
|
|
||||||
|
|
||||||
return recipients
|
|
||||||
|
|
||||||
|
|
||||||
async def save_to_inbox(
|
async def save_to_inbox(
|
||||||
db_session : AsyncSession,
|
db_session : AsyncSession,
|
||||||
inbox_id : str,
|
inbox_id : str,
|
||||||
|
|
96
app/ldsig.py
96
app/ldsig.py
|
@ -1,96 +0,0 @@
|
||||||
#!/usr/bin/env python3
|
|
||||||
import base64
|
|
||||||
import hashlib
|
|
||||||
import typing
|
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
import pyld # type: ignore
|
|
||||||
from Crypto.Hash import SHA256
|
|
||||||
from Crypto.Signature import PKCS1_v1_5
|
|
||||||
from Crypto.PublicKey import RSA
|
|
||||||
from loguru import logger
|
|
||||||
from pyld import jsonld # type: ignore
|
|
||||||
|
|
||||||
from app import activitypub as ap
|
|
||||||
from app.database import AsyncSession
|
|
||||||
from app.actor import get_public_key
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
requests_loader = pyld.documentloader.requests.requests_document_loader()
|
|
||||||
|
|
||||||
|
|
||||||
def _loader(url, options={}):
|
|
||||||
# See https://github.com/digitalbazaar/pyld/issues/133
|
|
||||||
options["headers"]["Accept"] = "application/ld+json"
|
|
||||||
|
|
||||||
if url == "https://w3id.org/identity/v1":
|
|
||||||
url = (
|
|
||||||
"https://raw.githubusercontent.com/web-payments/web-payments.org"
|
|
||||||
"/master/contexts/identity-v1.jsonld"
|
|
||||||
)
|
|
||||||
return requests_loader(url, options)
|
|
||||||
|
|
||||||
|
|
||||||
pyld.jsonld.set_document_loader(_loader)
|
|
||||||
|
|
||||||
|
|
||||||
def _options_hash(doc: ap.RawObject) -> str:
|
|
||||||
doc = dict(doc["signature"])
|
|
||||||
for k in ["type", "id", "signatureValue"]:
|
|
||||||
if k in doc:
|
|
||||||
del doc[k]
|
|
||||||
doc["@context"] = "https://w3id.org/security/v1"
|
|
||||||
normalized = jsonld.normalize(
|
|
||||||
doc, {"algorithm": "URDNA2015", "format": "application/nquads"}
|
|
||||||
)
|
|
||||||
h = hashlib.new("sha256")
|
|
||||||
h.update(normalized.encode("utf-8"))
|
|
||||||
return h.hexdigest()
|
|
||||||
|
|
||||||
|
|
||||||
def _doc_hash(doc: ap.RawObject) -> str:
|
|
||||||
doc = dict(doc)
|
|
||||||
if "signature" in doc:
|
|
||||||
del doc["signature"]
|
|
||||||
normalized = jsonld.normalize(
|
|
||||||
doc, {"algorithm": "URDNA2015", "format": "application/nquads"}
|
|
||||||
)
|
|
||||||
h = hashlib.new("sha256")
|
|
||||||
h.update(normalized.encode("utf-8"))
|
|
||||||
return h.hexdigest()
|
|
||||||
|
|
||||||
|
|
||||||
async def verify_signature(
|
|
||||||
db_session: AsyncSession,
|
|
||||||
doc: ap.RawObject,
|
|
||||||
) -> bool:
|
|
||||||
if "signature" not in doc:
|
|
||||||
logger.warning("The object does contain a signature")
|
|
||||||
return False
|
|
||||||
|
|
||||||
key_id = doc["signature"]["creator"]
|
|
||||||
key = await get_public_key(db_session, key_id)
|
|
||||||
to_be_signed = _options_hash(doc) + _doc_hash(doc)
|
|
||||||
signature = doc["signature"]["signatureValue"]
|
|
||||||
pubkey = RSA.importKey(key)
|
|
||||||
signer = PKCS1_v1_5.new(pubkey)
|
|
||||||
digest = SHA256.new()
|
|
||||||
digest.update(to_be_signed.encode("utf-8"))
|
|
||||||
return signer.verify(digest, base64.b64decode(signature)) # type: ignore
|
|
||||||
|
|
||||||
|
|
||||||
def generate_signature(doc: ap.RawObject, key) -> None:
|
|
||||||
options = {
|
|
||||||
"type": "RsaSignature2017",
|
|
||||||
"creator": doc["actor"] + "#main-key",
|
|
||||||
"created": datetime.utcnow().replace(microsecond=0).isoformat() + "Z",
|
|
||||||
}
|
|
||||||
doc["signature"] = options
|
|
||||||
to_be_signed = _options_hash(doc) + _doc_hash(doc)
|
|
||||||
|
|
||||||
signer = PKCS1_v1_5.new(key)
|
|
||||||
digest = SHA256.new()
|
|
||||||
digest.update(to_be_signed.encode("utf-8"))
|
|
||||||
sig = base64.b64encode(signer.sign(digest)) # type: ignore
|
|
||||||
options["signatureValue"] = sig.decode("utf-8")
|
|
|
@ -7,7 +7,7 @@ from typing import Union
|
||||||
from app import activitypub as ap
|
from app import activitypub as ap
|
||||||
from app.database import Base
|
from app.database import Base
|
||||||
from app.database import metadata_obj
|
from app.database import metadata_obj
|
||||||
from app.activitypub import BaseActor
|
from app.actor import BaseActor
|
||||||
|
|
||||||
from sqlalchemy import Column
|
from sqlalchemy import Column
|
||||||
from sqlalchemy import Boolean
|
from sqlalchemy import Boolean
|
||||||
|
|
|
@ -1,29 +0,0 @@
|
||||||
BSD 3-Clause License
|
|
||||||
|
|
||||||
Copyright (c) 2017-2020, honmaple
|
|
||||||
All rights reserved.
|
|
||||||
|
|
||||||
Redistribution and use in source and binary forms, with or without
|
|
||||||
modification, are permitted provided that the following conditions are met:
|
|
||||||
|
|
||||||
* Redistributions of source code must retain the above copyright notice, this
|
|
||||||
list of conditions and the following disclaimer.
|
|
||||||
|
|
||||||
* Redistributions in binary form must reproduce the above copyright notice,
|
|
||||||
this list of conditions and the following disclaimer in the documentation
|
|
||||||
and/or other materials provided with the distribution.
|
|
||||||
|
|
||||||
* Neither the name of the copyright holder nor the names of its
|
|
||||||
contributors may be used to endorse or promote products derived from
|
|
||||||
this software without specific prior written permission.
|
|
||||||
|
|
||||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
||||||
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
||||||
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
||||||
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
|
||||||
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
||||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
||||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
|
||||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
|
||||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
||||||
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
@ -1,25 +0,0 @@
|
||||||
#!/usr/bin/env python
|
|
||||||
# -*- coding: utf-8 -*-
|
|
||||||
# ********************************************************************************
|
|
||||||
# Copyright © 2017-2020 jianglin
|
|
||||||
# File Name: __init__.py
|
|
||||||
# Author: jianglin
|
|
||||||
# Email: mail@honmaple.com
|
|
||||||
# Created: 2019-05-29 18:06:22 (CST)
|
|
||||||
# Last Update: Sunday 2020-08-16 19:45:09 (CST)
|
|
||||||
# By:
|
|
||||||
# Description:
|
|
||||||
# ********************************************************************************
|
|
||||||
from .document import Document
|
|
||||||
|
|
||||||
|
|
||||||
def to_text(content, **kwargs):
|
|
||||||
return Document(content, **kwargs).to_text()
|
|
||||||
|
|
||||||
|
|
||||||
def to_html(content, **kwargs):
|
|
||||||
return Document(content, **kwargs).to_html()
|
|
||||||
|
|
||||||
|
|
||||||
def to_markdown(content, **kwargs):
|
|
||||||
return Document(content, **kwargs).to_markdown()
|
|
|
@ -1,879 +0,0 @@
|
||||||
#!/usr/bin/env python
|
|
||||||
# -*- coding: utf-8 -*-
|
|
||||||
# ********************************************************************************
|
|
||||||
# Copyright © 2017-2020 jianglin
|
|
||||||
# File Name: document.py
|
|
||||||
# Author: jianglin
|
|
||||||
# Email: mail@honmaple.com
|
|
||||||
# Created: 2018-02-26 11:44:43 (CST)
|
|
||||||
# Last Update: Wednesday 2020-08-19 12:00:03 (CST)
|
|
||||||
# Description:
|
|
||||||
# ********************************************************************************
|
|
||||||
import re
|
|
||||||
from hashlib import sha1
|
|
||||||
from textwrap import dedent
|
|
||||||
|
|
||||||
from .inline import Blankline, Hr, InlineText
|
|
||||||
from .src import highlight as src_highlight
|
|
||||||
|
|
||||||
DRAWER_BEGIN_REGEXP = re.compile(r"^(\s*):(\S+):\s*$")
|
|
||||||
DRAWER_END_REGEXP = re.compile(r"^(\s*):END:\s*$")
|
|
||||||
DRAWER_PROPERTY_REGEXP = re.compile(r"^(\s*):(\S+):(\s+(.*)$|$)")
|
|
||||||
|
|
||||||
BLOCK_BEGIN_REGEXP = re.compile(r"(?i)^(\s*)#\+BEGIN_(\w+)(.*)")
|
|
||||||
BLOCK_END_REGEXP = re.compile(r"(?i)^(\s*)#\+END_(\w+)")
|
|
||||||
BLOCK_RESULT_REGEXP = re.compile(r"(?i)^(\s*)#\+RESULTS:")
|
|
||||||
BLOCK_RESULT_CONTENT_REGEXP = re.compile(r"(?:^|\s+):(\s+(.*)|$)")
|
|
||||||
|
|
||||||
TABLE_SEP_REGEXP = re.compile(r"^(\s*)(\|[+-|]*)\s*$")
|
|
||||||
TABLE_ROW_REGEXP = re.compile(r"^(\s*)(\|.*)")
|
|
||||||
TABLE_ALIGN_REGEXP = re.compile(r"^<(l|c|r)>$")
|
|
||||||
|
|
||||||
LIST_DESCRIPTIVE_REGEXP = re.compile(r"^(\s*)([+*-])\s+(.*)::(\s|$)")
|
|
||||||
LIST_UNORDER_REGEXP = re.compile(r"^(\s*)([+*-])(\s+(.*)|$)")
|
|
||||||
LIST_ORDER_REGEXP = re.compile(r"^(\s*)(([0-9]+|[a-zA-Z])[.)])(\s+(.*)|$)")
|
|
||||||
LIST_STATUS_REGEXP = re.compile(r"\[( |X|-)\]\s")
|
|
||||||
LIST_LEVEL_REGEXP = re.compile(r"(\s*)(.+)$")
|
|
||||||
|
|
||||||
HEADLINE_REGEXP = re.compile(
|
|
||||||
r"^(\*+)(?:\s+(.+?))?(?:\s+\[#(.+)\])?(\s+.*?)(?:\s+:(.+):)?$")
|
|
||||||
KEYWORD_REGEXP = re.compile(r"^(\s*)#\+([^:]+):(\s+(.*)|$)")
|
|
||||||
COMMENT_REGEXP = re.compile(r"^(\s*)#(.*)")
|
|
||||||
ATTRIBUTE_REGEXP = re.compile(r"(?:^|\s+)(:[-\w]+)\s+(.*)$")
|
|
||||||
|
|
||||||
TODO_KEYWORDS = ("DONE", "TODO")
|
|
||||||
|
|
||||||
|
|
||||||
def string_split(s, sep):
|
|
||||||
if not s:
|
|
||||||
return []
|
|
||||||
return s.split(sep)
|
|
||||||
|
|
||||||
|
|
||||||
class Parser(object):
|
|
||||||
def __init__(self, content=""):
|
|
||||||
self.lines = content.splitlines()
|
|
||||||
self.level = 0
|
|
||||||
self.element = ""
|
|
||||||
self.children = []
|
|
||||||
self.escape = True
|
|
||||||
self.needparse = True
|
|
||||||
self.parsed_nodes = (
|
|
||||||
"blankline",
|
|
||||||
"headline",
|
|
||||||
"table",
|
|
||||||
"list",
|
|
||||||
"drawer",
|
|
||||||
"block",
|
|
||||||
"block_result",
|
|
||||||
"keyword",
|
|
||||||
"hr",
|
|
||||||
)
|
|
||||||
|
|
||||||
def first_child(self):
|
|
||||||
if len(self.children) == 0:
|
|
||||||
return
|
|
||||||
return self.children[0]
|
|
||||||
|
|
||||||
def last_child(self):
|
|
||||||
if len(self.children) == 0:
|
|
||||||
return
|
|
||||||
return self.children[-1]
|
|
||||||
|
|
||||||
def add_child(self, node):
|
|
||||||
last = self.last_child()
|
|
||||||
if self.is_headline(last):
|
|
||||||
if self.is_properties(node):
|
|
||||||
last.properties = node
|
|
||||||
return
|
|
||||||
|
|
||||||
if not self.is_headline(node):
|
|
||||||
last.add_child(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
if self.is_headline(node) and node.stars > last.stars:
|
|
||||||
last.add_child(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
if self.is_table(last):
|
|
||||||
if self.is_table(node):
|
|
||||||
last.add_child(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
if self.is_list(last):
|
|
||||||
if self.is_blankline(node):
|
|
||||||
last.add_child(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
if node.level > last.level:
|
|
||||||
last.add_child(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
if self.is_list(node) and node.level == last.level:
|
|
||||||
last.add_child(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
if self.is_keyword(last):
|
|
||||||
if self.is_table(node):
|
|
||||||
node.keyword = last
|
|
||||||
|
|
||||||
if self.is_paragraph(last):
|
|
||||||
if self.is_inlinetext(node):
|
|
||||||
last.add_child(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
if self.is_inlinetext(node):
|
|
||||||
self.children.append(self.paragraph(node))
|
|
||||||
return
|
|
||||||
|
|
||||||
self.children.append(node)
|
|
||||||
|
|
||||||
def is_keyword(self, child):
|
|
||||||
return child and isinstance(child, Keyword)
|
|
||||||
|
|
||||||
def is_headline(self, child):
|
|
||||||
return child and isinstance(child, Headline)
|
|
||||||
|
|
||||||
def is_list(self, child):
|
|
||||||
return child and isinstance(child, List)
|
|
||||||
|
|
||||||
def is_table(self, child):
|
|
||||||
return child and isinstance(child, Table)
|
|
||||||
|
|
||||||
def is_src(self, child):
|
|
||||||
return child and isinstance(child, (Src, Example))
|
|
||||||
|
|
||||||
def is_inlinetext(self, child):
|
|
||||||
return child and isinstance(child, InlineText)
|
|
||||||
|
|
||||||
def is_blankline(self, child):
|
|
||||||
return child and isinstance(child, Blankline)
|
|
||||||
|
|
||||||
def is_paragraph(self, child):
|
|
||||||
return child and isinstance(child, Paragraph)
|
|
||||||
|
|
||||||
def is_properties(self, child):
|
|
||||||
return child and isinstance(child, Properties)
|
|
||||||
|
|
||||||
def inlinetext(self, text):
|
|
||||||
return InlineText(text, self.needparse, self.escape)
|
|
||||||
|
|
||||||
def paragraph(self, node):
|
|
||||||
n = Paragraph()
|
|
||||||
n.add_child(node)
|
|
||||||
return n
|
|
||||||
|
|
||||||
def _parse_paired(self, cls, index, lines):
|
|
||||||
node = cls.match(lines[index])
|
|
||||||
if not node:
|
|
||||||
return None, index
|
|
||||||
|
|
||||||
end = len(lines)
|
|
||||||
num = index + 1
|
|
||||||
while num < end:
|
|
||||||
if node.matchend(num, lines):
|
|
||||||
node.preparse(lines[index + 1:num])
|
|
||||||
return node, num
|
|
||||||
num += 1
|
|
||||||
return None, index
|
|
||||||
|
|
||||||
def _parse_nopaired(self, cls, index, lines):
|
|
||||||
node = cls.match(lines[index])
|
|
||||||
if not node:
|
|
||||||
return None, index
|
|
||||||
|
|
||||||
end = len(lines)
|
|
||||||
num = index + 1
|
|
||||||
while num < end:
|
|
||||||
if node.matchend(num, lines):
|
|
||||||
break
|
|
||||||
num += 1
|
|
||||||
node.preparse(lines[index + 1:num])
|
|
||||||
return node, num
|
|
||||||
|
|
||||||
def parse_headline(self, index, lines):
|
|
||||||
return Headline.match(lines[index]), index
|
|
||||||
|
|
||||||
def parse_list(self, index, lines):
|
|
||||||
return List.match(lines[index]), index
|
|
||||||
|
|
||||||
def parse_table(self, index, lines):
|
|
||||||
return self._parse_nopaired(Table, index, lines)
|
|
||||||
|
|
||||||
def parse_drawer(self, index, lines):
|
|
||||||
return self._parse_paired(Drawer, index, lines)
|
|
||||||
|
|
||||||
def parse_block(self, index, lines):
|
|
||||||
return self._parse_paired(Block, index, lines)
|
|
||||||
|
|
||||||
def parse_block_result(self, index, lines):
|
|
||||||
return self._parse_paired(BlockResult, index, lines)
|
|
||||||
|
|
||||||
def parse_blankline(self, index, lines):
|
|
||||||
return Blankline.match(lines[index]), index
|
|
||||||
|
|
||||||
def parse_keyword(self, index, lines):
|
|
||||||
return Keyword.match(lines[index]), index
|
|
||||||
|
|
||||||
def parse_hr(self, index, lines):
|
|
||||||
return Hr.match(lines[index]), index
|
|
||||||
|
|
||||||
def parse_inlinetext(self, index, lines):
|
|
||||||
return self.inlinetext(lines[index]), index
|
|
||||||
|
|
||||||
def parse(self, index, lines):
|
|
||||||
for b in self.parsed_nodes:
|
|
||||||
func = "parse_" + b
|
|
||||||
if not hasattr(self, func):
|
|
||||||
continue
|
|
||||||
block, num = getattr(self, func)(index, lines)
|
|
||||||
if not block:
|
|
||||||
continue
|
|
||||||
return block, num
|
|
||||||
|
|
||||||
return self.parse_inlinetext(index, lines)
|
|
||||||
|
|
||||||
def preparse(self, lines):
|
|
||||||
index = 0
|
|
||||||
while index < len(lines):
|
|
||||||
line = lines[index]
|
|
||||||
node, index = self.parse(index, lines)
|
|
||||||
if node:
|
|
||||||
node.level = len(line) - len(line.strip())
|
|
||||||
self.add_child(node)
|
|
||||||
index += 1
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
if len(self.children) == 0 and len(self.lines) > 0:
|
|
||||||
self.preparse(self.lines)
|
|
||||||
|
|
||||||
children = []
|
|
||||||
for child in self.children:
|
|
||||||
content = child.to_html()
|
|
||||||
if not content:
|
|
||||||
continue
|
|
||||||
children.append(content)
|
|
||||||
text = "\n".join(children)
|
|
||||||
if self.element:
|
|
||||||
return self.element.format(text)
|
|
||||||
return text
|
|
||||||
|
|
||||||
def __str__(self):
|
|
||||||
str_children = [str(child) for child in self.children]
|
|
||||||
return self.__class__.__name__ + '(' + ','.join(str_children) + ')'
|
|
||||||
|
|
||||||
def __repr__(self):
|
|
||||||
return self.__str__()
|
|
||||||
|
|
||||||
|
|
||||||
class Headline(Parser):
|
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
title,
|
|
||||||
stars=1,
|
|
||||||
keyword=None,
|
|
||||||
priority=None,
|
|
||||||
tags=[],
|
|
||||||
todo_keywords=TODO_KEYWORDS):
|
|
||||||
super(Headline, self).__init__()
|
|
||||||
self.title = title
|
|
||||||
self.stars = stars
|
|
||||||
self.keyword = keyword
|
|
||||||
self.priority = priority
|
|
||||||
self.tags = tags
|
|
||||||
self.properties = None
|
|
||||||
self.todo_keywords = todo_keywords
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = HEADLINE_REGEXP.match(line)
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
|
|
||||||
stars = len(match[1])
|
|
||||||
keyword = match[2] or ""
|
|
||||||
priority = match[3] or ""
|
|
||||||
|
|
||||||
if keyword and not priority:
|
|
||||||
if len(keyword) >= 4 and keyword[0:2] == "[#":
|
|
||||||
priority = keyword[2:-1]
|
|
||||||
keyword = ""
|
|
||||||
|
|
||||||
title = keyword + match[4]
|
|
||||||
keyword = ""
|
|
||||||
|
|
||||||
return cls(
|
|
||||||
title,
|
|
||||||
stars,
|
|
||||||
keyword,
|
|
||||||
priority,
|
|
||||||
string_split(match[5], ":"),
|
|
||||||
)
|
|
||||||
|
|
||||||
def id(self):
|
|
||||||
hid = 'org-{0}'.format(sha1(self.title.encode()).hexdigest()[:10])
|
|
||||||
if self.properties:
|
|
||||||
return self.properties.get("CUSTOM_ID", hid)
|
|
||||||
return hid
|
|
||||||
|
|
||||||
def toc(self):
|
|
||||||
b = ""
|
|
||||||
if self.keyword:
|
|
||||||
b = b + "<span class=\"todo\">{0}</span>".format(self.keyword)
|
|
||||||
if self.priority:
|
|
||||||
b = b + "<span class=\"priority\">{0}</span>".format(self.priority)
|
|
||||||
|
|
||||||
b = b + self.inlinetext(self.title).to_html()
|
|
||||||
|
|
||||||
for tag in self.tags:
|
|
||||||
b = b + "<span class=\"tag\">{0}</span>".format(tag)
|
|
||||||
return b.strip()
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
b = "<h{0} id=\"{1}\">{2}</h{0}>".format(
|
|
||||||
self.stars,
|
|
||||||
self.id(),
|
|
||||||
self.toc(),
|
|
||||||
)
|
|
||||||
return b + super(Headline, self).to_html()
|
|
||||||
|
|
||||||
|
|
||||||
class Drawer(Parser):
|
|
||||||
def __init__(self, name):
|
|
||||||
super(Drawer, self).__init__()
|
|
||||||
self.name = name
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = DRAWER_BEGIN_REGEXP.match(line)
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
name = match[2]
|
|
||||||
if name.upper() == "PROPERTIES":
|
|
||||||
return Properties(name)
|
|
||||||
return Drawer(name)
|
|
||||||
|
|
||||||
def matchend(self, index, lines):
|
|
||||||
return DRAWER_END_REGEXP.match(lines[index])
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
return ""
|
|
||||||
|
|
||||||
|
|
||||||
class Properties(Drawer):
|
|
||||||
def __init__(self, name):
|
|
||||||
super(Properties, self).__init__(name)
|
|
||||||
self.properties = {}
|
|
||||||
|
|
||||||
def parse(self, index, lines):
|
|
||||||
match = DRAWER_PROPERTY_REGEXP.match(lines[index])
|
|
||||||
if match:
|
|
||||||
self.properties[match[2].upper()] = match[4]
|
|
||||||
return None, index
|
|
||||||
|
|
||||||
def get(self, key, default=None):
|
|
||||||
return self.properties.get(key, default)
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
return ""
|
|
||||||
|
|
||||||
|
|
||||||
class Block(Parser):
|
|
||||||
def __init__(self, name, params=""):
|
|
||||||
super(Block, self).__init__()
|
|
||||||
self.name = name
|
|
||||||
self.params = params
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = BLOCK_BEGIN_REGEXP.match(line)
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
|
|
||||||
name = match[2].lower()
|
|
||||||
if name == "src":
|
|
||||||
return Src(*match[3].strip().split(" ", 1))
|
|
||||||
if name == "example":
|
|
||||||
return Example(match[3])
|
|
||||||
if name == "center":
|
|
||||||
return Center(match[3])
|
|
||||||
if name == "verse":
|
|
||||||
return Verse(match[3])
|
|
||||||
if name == "quote":
|
|
||||||
return Quote(match[3])
|
|
||||||
if name == "export":
|
|
||||||
return Export(*match[3].strip().split(" ", 1))
|
|
||||||
return cls(name, match[3])
|
|
||||||
|
|
||||||
def matchend(self, index, lines):
|
|
||||||
match = BLOCK_END_REGEXP.match(lines[index])
|
|
||||||
return match and match[2].lower() == self.name
|
|
||||||
|
|
||||||
|
|
||||||
class Center(Block):
|
|
||||||
def __init__(self, params=""):
|
|
||||||
super(Center, self).__init__("center", params)
|
|
||||||
self.element = "<div style=\"text-align: center;\">\n{0}\n</div>"
|
|
||||||
|
|
||||||
|
|
||||||
class Verse(Block):
|
|
||||||
def __init__(self, params=""):
|
|
||||||
super(Verse, self).__init__("verse", params)
|
|
||||||
self.element = "<p class=\"verse\">\n{0}\n</p>"
|
|
||||||
|
|
||||||
def add_child(self, node):
|
|
||||||
self.children.append(node)
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
children = [child.to_html() for child in self.children]
|
|
||||||
return self.element.format("<br />".join(children))
|
|
||||||
|
|
||||||
|
|
||||||
class Quote(Block):
|
|
||||||
def __init__(self, params=""):
|
|
||||||
super(Quote, self).__init__("quote", params)
|
|
||||||
self.element = "<blockquote>\n{0}\n</blockquote>"
|
|
||||||
|
|
||||||
|
|
||||||
class Export(Block):
|
|
||||||
def __init__(self, language="", params=""):
|
|
||||||
super(Export, self).__init__("export", params)
|
|
||||||
self.language = language
|
|
||||||
self.escape = self.language.upper() != "HTML"
|
|
||||||
self.parsed_nodes = ()
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
if not self.escape:
|
|
||||||
return super(Export, self).to_html()
|
|
||||||
return ""
|
|
||||||
|
|
||||||
|
|
||||||
class Src(Block):
|
|
||||||
def __init__(self, language="", params="", highlight=False):
|
|
||||||
super(Src, self).__init__("src", params)
|
|
||||||
self.language = language
|
|
||||||
self.highlight_code = highlight
|
|
||||||
self.element = "<pre class=\"src src-{0}\">\n{1}\n</pre>"
|
|
||||||
self.needparse = False
|
|
||||||
self.escape = False
|
|
||||||
self.parsed_nodes = ()
|
|
||||||
|
|
||||||
def add_child(self, node):
|
|
||||||
self.children.append(node)
|
|
||||||
|
|
||||||
def highlight(self, language, text):
|
|
||||||
return src_highlight(language, text)
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
text = "\n".join([child.to_html() for child in self.children])
|
|
||||||
if self.highlight_code:
|
|
||||||
return self.highlight(self.language, dedent(text))
|
|
||||||
if not self.language:
|
|
||||||
return "<pre>\n{0}\n</pre>".format(dedent(text))
|
|
||||||
return self.element.format(self.language, dedent(text))
|
|
||||||
|
|
||||||
|
|
||||||
class Example(Src):
|
|
||||||
def __init__(self, params="", highlight=False):
|
|
||||||
super(Example, self).__init__("example", params, highlight)
|
|
||||||
self.name = "example"
|
|
||||||
|
|
||||||
|
|
||||||
class BlockResult(Parser):
|
|
||||||
def __init__(self):
|
|
||||||
super(BlockResult, self).__init__()
|
|
||||||
self.element = "<pre class=\"example\">\n{0}\n</pre>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = BLOCK_RESULT_REGEXP.match(line)
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
return cls()
|
|
||||||
|
|
||||||
def matchend(self, index, lines):
|
|
||||||
return not BLOCK_RESULT_CONTENT_REGEXP.match(lines[index])
|
|
||||||
|
|
||||||
def parse(self, index, lines):
|
|
||||||
match = BLOCK_RESULT_CONTENT_REGEXP.match(lines[index])
|
|
||||||
return self.inlinetext(match[2]), index
|
|
||||||
|
|
||||||
|
|
||||||
class ListItem(Parser):
|
|
||||||
def __init__(self, status=None, checkbox="HTML"):
|
|
||||||
super(ListItem, self).__init__()
|
|
||||||
self.status = status
|
|
||||||
self.checkbox = checkbox
|
|
||||||
self.element = "<li>\n{0}\n</li>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
status = None
|
|
||||||
content = line
|
|
||||||
status_match = LIST_STATUS_REGEXP.match(line)
|
|
||||||
if status_match:
|
|
||||||
status, content = status_match[1], content[len("[ ] "):]
|
|
||||||
|
|
||||||
node = cls(status)
|
|
||||||
node.add_child(node.inlinetext(content))
|
|
||||||
return node
|
|
||||||
|
|
||||||
def set_status(self):
|
|
||||||
if not self.checkbox:
|
|
||||||
return
|
|
||||||
|
|
||||||
if self.checkbox == "HTML":
|
|
||||||
if self.status == "X":
|
|
||||||
node = self.inlinetext(
|
|
||||||
'<input type="checkbox" checked="checked" />')
|
|
||||||
else:
|
|
||||||
node = self.inlinetext('<input type="checkbox" />')
|
|
||||||
node.needparse = False
|
|
||||||
node.escape = False
|
|
||||||
else:
|
|
||||||
node = self.inlinetext("=[{0}]=".format(self.status))
|
|
||||||
|
|
||||||
if not self.children:
|
|
||||||
self.children.append(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
self.children[0].children = [node] + self.children[0].children
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
if self.status is not None:
|
|
||||||
self.set_status()
|
|
||||||
return super(ListItem, self).to_html()
|
|
||||||
|
|
||||||
|
|
||||||
class DescriptiveItem(ListItem):
|
|
||||||
def __init__(self, title="", status=""):
|
|
||||||
super(DescriptiveItem, self).__init__(title, status)
|
|
||||||
self.element = "<dt>\n{0}\n</dt>"
|
|
||||||
|
|
||||||
|
|
||||||
class List(Parser):
|
|
||||||
def __init__(self, items=[]):
|
|
||||||
super(List, self).__init__()
|
|
||||||
self.children = items
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = UnorderList.match(line)
|
|
||||||
if match:
|
|
||||||
return match
|
|
||||||
|
|
||||||
match = OrderList.match(line)
|
|
||||||
if match:
|
|
||||||
return match
|
|
||||||
|
|
||||||
return Descriptive.match(line)
|
|
||||||
|
|
||||||
def add_child(self, node):
|
|
||||||
if self.is_list(node) and node.level == self.level:
|
|
||||||
self.children.append(node.children[0])
|
|
||||||
return
|
|
||||||
last = self.last_child()
|
|
||||||
last.add_child(node)
|
|
||||||
|
|
||||||
|
|
||||||
class Descriptive(List):
|
|
||||||
def __init__(self, items=[]):
|
|
||||||
super(Descriptive, self).__init__(items)
|
|
||||||
self.element = "<dd>\n{0}\n</dd>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = LIST_DESCRIPTIVE_REGEXP.match(line)
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
title = DescriptiveItem.match(match[3])
|
|
||||||
return cls([title])
|
|
||||||
|
|
||||||
|
|
||||||
class UnorderList(List):
|
|
||||||
def __init__(self, items=[]):
|
|
||||||
super(UnorderList, self).__init__(items)
|
|
||||||
self.element = "<ul>\n{0}\n</ul>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = LIST_UNORDER_REGEXP.match(line)
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
title = ListItem.match(match[4])
|
|
||||||
return cls([title])
|
|
||||||
|
|
||||||
|
|
||||||
class OrderList(List):
|
|
||||||
def __init__(self, items=[]):
|
|
||||||
super(OrderList, self).__init__(items)
|
|
||||||
self.element = "<ol>\n{0}\n</ol>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = LIST_ORDER_REGEXP.match(line)
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
title = ListItem.match(match[4])
|
|
||||||
return cls([title])
|
|
||||||
|
|
||||||
|
|
||||||
class TableColumn(Parser):
|
|
||||||
def __init__(self, content="", header=False):
|
|
||||||
super(TableColumn, self).__init__(content)
|
|
||||||
self.header = header
|
|
||||||
self.parsed_nodes = ()
|
|
||||||
|
|
||||||
def add_child(self, child):
|
|
||||||
self.children.append(child)
|
|
||||||
|
|
||||||
def reset(self):
|
|
||||||
self.header = True
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
self.element = "<th>{0}</th>" if self.header else "<td>{0}</td>"
|
|
||||||
return super(TableColumn, self).to_html()
|
|
||||||
|
|
||||||
|
|
||||||
class TableRow(Parser):
|
|
||||||
def __init__(self, header=False):
|
|
||||||
super(TableRow, self).__init__()
|
|
||||||
self.is_sep = False
|
|
||||||
self.header = header
|
|
||||||
self.element = "<tr>\n{0}\n</tr>"
|
|
||||||
self.parsed_nodes = ("tablecolumn", )
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = TABLE_ROW_REGEXP.match(line)
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
|
|
||||||
row = cls()
|
|
||||||
row.is_sep = bool(TABLE_SEP_REGEXP.match(line))
|
|
||||||
row.preparse(match[2].strip("|").split("|"))
|
|
||||||
return row
|
|
||||||
|
|
||||||
def add_child(self, child):
|
|
||||||
self.children.append(child)
|
|
||||||
|
|
||||||
def parse_tablecolumn(self, index, lines):
|
|
||||||
return TableColumn(lines[index].strip(), self.header), index
|
|
||||||
|
|
||||||
def reset(self):
|
|
||||||
self.header = True
|
|
||||||
for column in self.children:
|
|
||||||
column.reset()
|
|
||||||
|
|
||||||
|
|
||||||
class Table(Parser):
|
|
||||||
def __init__(self, keyword=None):
|
|
||||||
super(Table, self).__init__()
|
|
||||||
self.element = "<table>\n{0}\n</table>"
|
|
||||||
self.keyword = keyword
|
|
||||||
self.parsed_nodes = ("tablerow", )
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
row = TableRow.match(line)
|
|
||||||
if not row:
|
|
||||||
return
|
|
||||||
|
|
||||||
table = cls()
|
|
||||||
if row.is_sep:
|
|
||||||
return table
|
|
||||||
table.add_child(row)
|
|
||||||
return table
|
|
||||||
|
|
||||||
def matchend(self, index, lines):
|
|
||||||
return not TABLE_ROW_REGEXP.match(lines[index])
|
|
||||||
|
|
||||||
def reset(self):
|
|
||||||
first = self.first_child()
|
|
||||||
if first and first.header:
|
|
||||||
return
|
|
||||||
for row in self.children:
|
|
||||||
row.reset()
|
|
||||||
|
|
||||||
def add_child(self, child):
|
|
||||||
if child.is_sep:
|
|
||||||
return self.reset()
|
|
||||||
self.children.append(child)
|
|
||||||
|
|
||||||
def parse_tablerow(self, index, lines):
|
|
||||||
return TableRow.match(lines[index]), index
|
|
||||||
|
|
||||||
|
|
||||||
class Keyword(Parser):
|
|
||||||
def __init__(self, key, value=""):
|
|
||||||
super(Keyword, self).__init__()
|
|
||||||
self.key = key
|
|
||||||
self.value = value
|
|
||||||
|
|
||||||
def options(self):
|
|
||||||
results = {}
|
|
||||||
for line in self.value.split(" "):
|
|
||||||
if not line:
|
|
||||||
continue
|
|
||||||
m = line.split(":", 1)
|
|
||||||
k = m[0]
|
|
||||||
if not k:
|
|
||||||
continue
|
|
||||||
results[k] = "" if len(m) == 1 else m[1]
|
|
||||||
return results
|
|
||||||
|
|
||||||
def properties(self):
|
|
||||||
results = {}
|
|
||||||
line = self.value.strip()
|
|
||||||
if not line:
|
|
||||||
return results
|
|
||||||
m = line.split(" ", 1)
|
|
||||||
k = m[0]
|
|
||||||
if not k:
|
|
||||||
return results
|
|
||||||
results[k] = "" if len(m) == 1 else m[1]
|
|
||||||
return results
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = KEYWORD_REGEXP.match(line)
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
return cls(match[2], match[4])
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
return ""
|
|
||||||
|
|
||||||
|
|
||||||
class Paragraph(Parser):
|
|
||||||
def __init__(self, content=""):
|
|
||||||
super(Paragraph, self).__init__(content)
|
|
||||||
self.element = "<p>\n{0}\n</p>"
|
|
||||||
self.parsed_nodes = ()
|
|
||||||
|
|
||||||
def add_child(self, node):
|
|
||||||
self.children.append(node)
|
|
||||||
|
|
||||||
|
|
||||||
class Section(Parser):
|
|
||||||
def __init__(self, headline):
|
|
||||||
super(Section, self).__init__()
|
|
||||||
self.headline = headline
|
|
||||||
|
|
||||||
@property
|
|
||||||
def stars(self):
|
|
||||||
return self.headline.stars
|
|
||||||
|
|
||||||
def add_child(self, node):
|
|
||||||
last = self.last_child()
|
|
||||||
if not last:
|
|
||||||
self.children.append(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
if node.stars > last.stars:
|
|
||||||
last.add_child(node)
|
|
||||||
return
|
|
||||||
self.children.append(node)
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
text = "<li>"
|
|
||||||
text += "<a href=\"#{0}\">{1}</a>".format(
|
|
||||||
self.headline.id(),
|
|
||||||
self.headline.toc(),
|
|
||||||
)
|
|
||||||
if not self.children:
|
|
||||||
return text + "</li>"
|
|
||||||
|
|
||||||
text += "\n<ul>\n{0}\n</ul>\n</li>".format(
|
|
||||||
"\n".join([child.to_html() for child in self.children]))
|
|
||||||
return text
|
|
||||||
|
|
||||||
|
|
||||||
class Toc(Parser):
|
|
||||||
def __init__(self):
|
|
||||||
super(Toc, self).__init__()
|
|
||||||
self.element = (
|
|
||||||
'<div id="table-of-contents">'
|
|
||||||
'<h2>Table of Contents</h2>'
|
|
||||||
'<div id="text-table-of-contents">'
|
|
||||||
'\n<ul>\n{0}\n</ul>\n</div></div>')
|
|
||||||
|
|
||||||
def add_child(self, node):
|
|
||||||
last = self.last_child()
|
|
||||||
if not last:
|
|
||||||
self.children.append(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
if node.stars > last.stars:
|
|
||||||
last.add_child(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
if node.stars < last.stars:
|
|
||||||
last.add_child(node)
|
|
||||||
return
|
|
||||||
|
|
||||||
self.children.append(node)
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
if not self.children:
|
|
||||||
return ""
|
|
||||||
return super(Toc, self).to_html()
|
|
||||||
|
|
||||||
|
|
||||||
class Document(Parser):
|
|
||||||
def __init__(self, content, offset=0, highlight=False, **options):
|
|
||||||
super(Document, self).__init__(content)
|
|
||||||
self.offset = offset
|
|
||||||
self.highlight = highlight
|
|
||||||
self.options = options
|
|
||||||
self.properties = {}
|
|
||||||
self.toc = Toc()
|
|
||||||
|
|
||||||
def _is_true(self, value):
|
|
||||||
return value in ("true", "t", "1", True, 1)
|
|
||||||
|
|
||||||
def section(self, node):
|
|
||||||
return Section(node)
|
|
||||||
|
|
||||||
def parse_keyword(self, index, lines):
|
|
||||||
block, index = super(Document, self).parse_keyword(index, lines)
|
|
||||||
if not block:
|
|
||||||
return block, index
|
|
||||||
|
|
||||||
if block.key == "OPTIONS":
|
|
||||||
self.options.update(**block.options())
|
|
||||||
elif block.key == "PROPERTY":
|
|
||||||
self.properties.update(**block.properties())
|
|
||||||
else:
|
|
||||||
self.properties[block.key] = block.value
|
|
||||||
return block, index
|
|
||||||
|
|
||||||
def parse_headline(self, index, lines):
|
|
||||||
block, index = super(Document, self).parse_headline(index, lines)
|
|
||||||
if not block:
|
|
||||||
return block, index
|
|
||||||
block.stars = block.stars + self.offset
|
|
||||||
|
|
||||||
todo_keywords = self.properties.get("TODO")
|
|
||||||
if todo_keywords:
|
|
||||||
block.todo_keywords = todo_keywords.split(" ")
|
|
||||||
s = block.title.split(" ", 1)
|
|
||||||
if len(s) > 1 and s[0] in block.todo_keywords:
|
|
||||||
block.keyword = s[0]
|
|
||||||
block.title = s[1]
|
|
||||||
self.toc.add_child(self.section(block))
|
|
||||||
return block, index
|
|
||||||
|
|
||||||
def parse_block(self, index, lines):
|
|
||||||
block, index = super(Document, self).parse_block(index, lines)
|
|
||||||
if not block:
|
|
||||||
return block, index
|
|
||||||
if self.is_src(block):
|
|
||||||
block.highlight_code = self.highlight
|
|
||||||
return block, index
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
text = super(Document, self).to_html()
|
|
||||||
if self._is_true(self.options.get("toc")):
|
|
||||||
return self.toc.to_html() + "\n" + text
|
|
||||||
return text
|
|
|
@ -1,432 +0,0 @@
|
||||||
#!/usr/bin/env python
|
|
||||||
# -*- coding: utf-8 -*-
|
|
||||||
# ********************************************************************************
|
|
||||||
# Copyright © 2017-2020 jianglin
|
|
||||||
# File Name: inline.py
|
|
||||||
# Author: jianglin
|
|
||||||
# Email: mail@honmaple.com
|
|
||||||
# Created: 2018-02-26 11:41:22 (CST)
|
|
||||||
# Last Update: Tuesday 2020-08-18 17:21:40 (CST)
|
|
||||||
# By:
|
|
||||||
# Description:
|
|
||||||
# ********************************************************************************
|
|
||||||
import re
|
|
||||||
import os
|
|
||||||
|
|
||||||
# _inline_regexp = r"(^|.*?(?<![/\\])){0}(.+?(?<![/\\])){0}(.*?|$)"
|
|
||||||
_inline_regexp = r"(^|.*?(?<![/\\])){0}(.+?(?<![/\\])){0}(.*?|$)"
|
|
||||||
|
|
||||||
BOLD_REGEXP = re.compile(_inline_regexp.format('\\*'))
|
|
||||||
CODE_REGEXP = re.compile(_inline_regexp.format('(?:\\=|`)'))
|
|
||||||
ITALIC_REGEXP = re.compile(_inline_regexp.format('(?:\\*\\*|\\/)'))
|
|
||||||
DELETE_REGEXP = re.compile(_inline_regexp.format('\\+'))
|
|
||||||
VERBATIM_REGEXP = re.compile(_inline_regexp.format('~'))
|
|
||||||
UNDERLINE_REGEXP = re.compile(_inline_regexp.format('_'))
|
|
||||||
|
|
||||||
PERCENT_REGEXP = re.compile(r"\[(\d+/\d+|\d+%)\]")
|
|
||||||
|
|
||||||
HR_REGEXP = re.compile(r"^\s*\-{5,}\s*")
|
|
||||||
FN_REGEXP = re.compile(r"(^|.*?(?<![/\\]))(\[fn:(.+?)\])(.*?|$)")
|
|
||||||
IMG_REGEXP = re.compile(r"^[.](png|gif|jpe?g|svg|tiff?)$")
|
|
||||||
LINK_REGEXP = re.compile(r'\[\[(.+?)\](?:\[(.+?)\])?\]')
|
|
||||||
VIDEO_REGEXP = re.compile(r"^[.](webm|mp4)$")
|
|
||||||
|
|
||||||
NEWLINE_REGEXP = re.compile(r"(^|.*?(?<![/\\]))(\\\\(\s*)$)")
|
|
||||||
BLANKLINE_REGEXP = re.compile(r"^(\s*)$")
|
|
||||||
|
|
||||||
TIMESTAMP_REGEXP = re.compile(
|
|
||||||
r"^<(\d{4}-\d{2}-\d{2})( [A-Za-z]+)?( \d{2}:\d{2})?( \+\d+[dwmy])?>")
|
|
||||||
|
|
||||||
_html_escape = (
|
|
||||||
("&", "&"),
|
|
||||||
("'", "'"),
|
|
||||||
("<", "<"),
|
|
||||||
(">", ">"),
|
|
||||||
("\"", """),
|
|
||||||
)
|
|
||||||
|
|
||||||
# https://github.com/tsroten/zhon/blob/develop/zhon/hanzi.py
|
|
||||||
_chinese_non_stops = (
|
|
||||||
# Fullwidth ASCII variants
|
|
||||||
'\uFF02\uFF03\uFF04\uFF05\uFF06\uFF07\uFF08\uFF09\uFF0A\uFF0B\uFF0C\uFF0D'
|
|
||||||
'\uFF0F\uFF1A\uFF1B\uFF1C\uFF1D\uFF1E\uFF20\uFF3B\uFF3C\uFF3D\uFF3E\uFF3F'
|
|
||||||
'\uFF40\uFF5B\uFF5C\uFF5D\uFF5E\uFF5F\uFF60'
|
|
||||||
|
|
||||||
# Halfwidth CJK punctuation
|
|
||||||
'\uFF62\uFF63\uFF64'
|
|
||||||
|
|
||||||
# CJK symbols and punctuation
|
|
||||||
'\u3000\u3001\u3003'
|
|
||||||
|
|
||||||
# CJK angle and corner brackets
|
|
||||||
'\u3008\u3009\u300A\u300B\u300C\u300D\u300E\u300F\u3010\u3011'
|
|
||||||
|
|
||||||
# CJK brackets and symbols/punctuation
|
|
||||||
'\u3014\u3015\u3016\u3017\u3018\u3019\u301A\u301B\u301C\u301D\u301E\u301F'
|
|
||||||
|
|
||||||
# Other CJK symbols
|
|
||||||
'\u3030'
|
|
||||||
|
|
||||||
# Special CJK indicators
|
|
||||||
'\u303E\u303F'
|
|
||||||
|
|
||||||
# Dashes
|
|
||||||
'\u2013\u2014'
|
|
||||||
|
|
||||||
# Quotation marks and apostrophe
|
|
||||||
'\u2018\u2019\u201B\u201C\u201D\u201E\u201F'
|
|
||||||
|
|
||||||
# General punctuation
|
|
||||||
'\u2026\u2027'
|
|
||||||
|
|
||||||
# Overscores and underscores
|
|
||||||
'\uFE4F'
|
|
||||||
|
|
||||||
# Small form variants
|
|
||||||
'\uFE51\uFE54'
|
|
||||||
|
|
||||||
# Latin punctuation
|
|
||||||
'\u00B7')
|
|
||||||
|
|
||||||
_chinese_stops = (
|
|
||||||
'\uFF01' # Fullwidth exclamation mark
|
|
||||||
'\uFF1F' # Fullwidth question mark
|
|
||||||
'\uFF61' # Halfwidth ideographic full stop
|
|
||||||
'\u3002' # Ideographic full stop
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def html_escape(text):
|
|
||||||
for e in _html_escape:
|
|
||||||
text = text.replace(e[0], e[1])
|
|
||||||
return text
|
|
||||||
|
|
||||||
|
|
||||||
def match_chinese(ch):
|
|
||||||
if '\u4e00' <= ch <= '\u9fff':
|
|
||||||
return True
|
|
||||||
if ch in _chinese_stops:
|
|
||||||
return True
|
|
||||||
return ch in _chinese_non_stops
|
|
||||||
|
|
||||||
|
|
||||||
def match_emphasis(cls, regexp, line, index):
|
|
||||||
match = regexp.match(line, index)
|
|
||||||
if not match:
|
|
||||||
return None, index
|
|
||||||
|
|
||||||
end = match.end()
|
|
||||||
|
|
||||||
if index != 0:
|
|
||||||
prechar = line[index - 1]
|
|
||||||
border = prechar != " " and prechar not in "-({'\""
|
|
||||||
if border and not match_chinese(prechar):
|
|
||||||
return None, index
|
|
||||||
|
|
||||||
if end < len(line):
|
|
||||||
endchar = line[end]
|
|
||||||
border = endchar != " " and endchar not in "-.,:!?;'\")}["
|
|
||||||
if border and not match_chinese(endchar):
|
|
||||||
return None, index
|
|
||||||
return cls(match[2]), end - 1
|
|
||||||
|
|
||||||
|
|
||||||
class InlineParser(object):
|
|
||||||
def __init__(self, content=""):
|
|
||||||
self.content = content
|
|
||||||
self.children = []
|
|
||||||
self.element = ""
|
|
||||||
|
|
||||||
def add_child(self, child):
|
|
||||||
self.children.append(child)
|
|
||||||
|
|
||||||
def parse_code(self, index, lines):
|
|
||||||
return Code.match(lines, index)
|
|
||||||
|
|
||||||
def parse_bold(self, index, lines):
|
|
||||||
return Bold.match(lines, index)
|
|
||||||
|
|
||||||
def parse_italic(self, index, lines):
|
|
||||||
return Italic.match(lines, index)
|
|
||||||
|
|
||||||
def parse_delete(self, index, lines):
|
|
||||||
return Delete.match(lines, index)
|
|
||||||
|
|
||||||
def parse_verbatim(self, index, lines):
|
|
||||||
return Verbatim.match(lines, index)
|
|
||||||
|
|
||||||
def parse_underline(self, index, lines):
|
|
||||||
return Underline.match(lines, index)
|
|
||||||
|
|
||||||
def parse_percent(self, index, lines):
|
|
||||||
return Percent.match(lines, index)
|
|
||||||
|
|
||||||
def parse_link(self, index, lines):
|
|
||||||
return Link.match(lines, index)
|
|
||||||
|
|
||||||
def parse_fn(self, index, lines):
|
|
||||||
return Fn.match(lines, index)
|
|
||||||
|
|
||||||
def parse_newline(self, index, lines):
|
|
||||||
return Newline.match(lines, index)
|
|
||||||
|
|
||||||
def parse(self, index, lines):
|
|
||||||
chars = (
|
|
||||||
("=", "code"),
|
|
||||||
("`", "code"),
|
|
||||||
("~", "verbatim"),
|
|
||||||
("_", "underline"),
|
|
||||||
("+", "delete"),
|
|
||||||
("/", "italic"),
|
|
||||||
("**", "italic"),
|
|
||||||
("*", "bold"),
|
|
||||||
("[[", "link"),
|
|
||||||
("[", "percent"),
|
|
||||||
("\\", "newline"),
|
|
||||||
)
|
|
||||||
char_map = dict(chars)
|
|
||||||
single_char = lines[index]
|
|
||||||
double_char = lines[index:index + 2]
|
|
||||||
for char in chars:
|
|
||||||
c1 = len(char[0]) == 1 and char[0] == single_char
|
|
||||||
c2 = len(char[0]) == 2 and char[0] == double_char
|
|
||||||
|
|
||||||
if c1 or c2:
|
|
||||||
node, num = getattr(self, "parse_" + char_map[char[0]])(
|
|
||||||
index, lines)
|
|
||||||
if node:
|
|
||||||
return node, num
|
|
||||||
|
|
||||||
if lines[index:index + 3] == "[fn":
|
|
||||||
node, num = self.parse_fn(index, lines)
|
|
||||||
if node:
|
|
||||||
return node, num
|
|
||||||
|
|
||||||
child = self.last_child()
|
|
||||||
if child and isinstance(child, Text):
|
|
||||||
child.content += single_char
|
|
||||||
return None, index
|
|
||||||
return Text(single_char), index
|
|
||||||
|
|
||||||
def last_child(self):
|
|
||||||
if len(self.children) == 0:
|
|
||||||
return
|
|
||||||
return self.children[-1]
|
|
||||||
|
|
||||||
def preparse(self, lines):
|
|
||||||
index = 0
|
|
||||||
while index < len(lines):
|
|
||||||
block, index = self.parse(index, lines)
|
|
||||||
index += 1
|
|
||||||
if not block:
|
|
||||||
continue
|
|
||||||
self.add_child(block)
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
if len(self.children) == 0 and self.content:
|
|
||||||
self.preparse(self.content)
|
|
||||||
|
|
||||||
text = "".join([child.to_html() for child in self.children])
|
|
||||||
if self.element:
|
|
||||||
return self.element.format(text)
|
|
||||||
return text
|
|
||||||
|
|
||||||
def __str__(self):
|
|
||||||
return '{}({})'.format(self.__class__.__name__, self.content.strip())
|
|
||||||
|
|
||||||
def __repr__(self):
|
|
||||||
return self.__str__()
|
|
||||||
|
|
||||||
|
|
||||||
class Text(InlineParser):
|
|
||||||
def to_html(self):
|
|
||||||
return self.content
|
|
||||||
|
|
||||||
|
|
||||||
class Newline(InlineParser):
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
match = NEWLINE_REGEXP.match(line, index)
|
|
||||||
if not match:
|
|
||||||
return None, index
|
|
||||||
return cls(), match.end() - 1
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
return "<br/>"
|
|
||||||
|
|
||||||
|
|
||||||
class Bold(InlineParser):
|
|
||||||
def __init__(self, content):
|
|
||||||
super(Bold, self).__init__(content)
|
|
||||||
self.element = "<b>{0}</b>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
return match_emphasis(cls, BOLD_REGEXP, line, index)
|
|
||||||
|
|
||||||
|
|
||||||
class Code(InlineParser):
|
|
||||||
def __init__(self, content):
|
|
||||||
super(Code, self).__init__(content)
|
|
||||||
self.element = "<code>{0}</code>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
return match_emphasis(cls, CODE_REGEXP, line, index)
|
|
||||||
|
|
||||||
|
|
||||||
class Italic(InlineParser):
|
|
||||||
def __init__(self, content):
|
|
||||||
super(Italic, self).__init__(content)
|
|
||||||
self.element = "<i>{0}</i>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
return match_emphasis(cls, ITALIC_REGEXP, line, index)
|
|
||||||
|
|
||||||
|
|
||||||
class Delete(InlineParser):
|
|
||||||
def __init__(self, content):
|
|
||||||
super(Delete, self).__init__(content)
|
|
||||||
self.element = "<del>{0}</del>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
return match_emphasis(cls, DELETE_REGEXP, line, index)
|
|
||||||
|
|
||||||
|
|
||||||
class Verbatim(InlineParser):
|
|
||||||
def __init__(self, content):
|
|
||||||
super(Verbatim, self).__init__(content)
|
|
||||||
self.element = "<code>{0}</code>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
return match_emphasis(cls, VERBATIM_REGEXP, line, index)
|
|
||||||
|
|
||||||
|
|
||||||
class Underline(InlineParser):
|
|
||||||
def __init__(self, content):
|
|
||||||
super(Underline, self).__init__(content)
|
|
||||||
self.element = "<span style=\"text-decoration:underline\">{0}</span>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
return match_emphasis(cls, UNDERLINE_REGEXP, line, index)
|
|
||||||
|
|
||||||
|
|
||||||
class Percent(InlineParser):
|
|
||||||
def __init__(self, content):
|
|
||||||
super(Percent, self).__init__(content)
|
|
||||||
self.element = "<code>[{0}]</code>"
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
match = PERCENT_REGEXP.match(line, index)
|
|
||||||
if not match:
|
|
||||||
return None, index
|
|
||||||
return cls(match[1]), match.end()
|
|
||||||
|
|
||||||
|
|
||||||
class Link(InlineParser):
|
|
||||||
def __init__(self, url, desc=None):
|
|
||||||
super(Link, self).__init__(url)
|
|
||||||
self.desc = desc
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
match = LINK_REGEXP.match(line, index)
|
|
||||||
if not match:
|
|
||||||
return None, index
|
|
||||||
return cls(match[1], match[2]), match.end()
|
|
||||||
|
|
||||||
def is_img(self):
|
|
||||||
_, ext = os.path.splitext(self.content)
|
|
||||||
return not self.desc and IMG_REGEXP.match(ext)
|
|
||||||
|
|
||||||
def is_vedio(self):
|
|
||||||
_, ext = os.path.splitext(self.content)
|
|
||||||
return not self.desc and VIDEO_REGEXP.match(ext)
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
if self.is_img():
|
|
||||||
return "<img src=\"{0}\"/>".format(self.content)
|
|
||||||
if self.is_vedio():
|
|
||||||
return "<video src=\"{0}\">{0}</video>".format(self.content)
|
|
||||||
if self.desc:
|
|
||||||
return '<a href="{0}">{1}</a>'.format(self.content, self.desc)
|
|
||||||
return '<a href="{0}">{1}</a>'.format(self.content, self.content)
|
|
||||||
|
|
||||||
|
|
||||||
class Fn(InlineParser):
|
|
||||||
def __init__(self, content):
|
|
||||||
super(Fn, self).__init__(content)
|
|
||||||
self.element = '<sup><a id="fnr:{0}" class="footref" href="#fn.{0}">{0}</a></sup>'
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
match = FN_REGEXP.match(line, index)
|
|
||||||
if not match:
|
|
||||||
return None, index
|
|
||||||
return cls(match[3]), match.end()
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
return self.element.format(self.content)
|
|
||||||
|
|
||||||
|
|
||||||
class Timestamp(InlineParser):
|
|
||||||
def __init__(self, date="", time="", interval=None):
|
|
||||||
super(Timestamp, self).__init__()
|
|
||||||
self.date = date
|
|
||||||
self.time = time
|
|
||||||
self.interval = interval
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line, index):
|
|
||||||
match = TIMESTAMP_REGEXP.match(line, index)
|
|
||||||
if not match:
|
|
||||||
return None, index
|
|
||||||
return cls(match[1], match[3], match[4]), match.end()
|
|
||||||
|
|
||||||
|
|
||||||
class Blankline(InlineParser):
|
|
||||||
def __init__(self):
|
|
||||||
super(Blankline, self).__init__()
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
match = BLANKLINE_REGEXP.match(line)
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
return cls()
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
return ""
|
|
||||||
|
|
||||||
|
|
||||||
class Hr(InlineParser):
|
|
||||||
def __init__(self):
|
|
||||||
super(Hr, self).__init__()
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def match(cls, line):
|
|
||||||
if HR_REGEXP.match(line):
|
|
||||||
return cls()
|
|
||||||
return
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
return ""
|
|
||||||
|
|
||||||
|
|
||||||
class InlineText(InlineParser):
|
|
||||||
def __init__(self, content="", needparse=True, escape=True):
|
|
||||||
super(InlineText, self).__init__(content)
|
|
||||||
self.needparse = needparse
|
|
||||||
self.escape = escape
|
|
||||||
|
|
||||||
def to_html(self):
|
|
||||||
if self.escape:
|
|
||||||
self.content = html_escape(self.content)
|
|
||||||
if not self.needparse:
|
|
||||||
return self.content
|
|
||||||
return super(InlineText, self).to_html()
|
|
|
@ -1,30 +0,0 @@
|
||||||
#!/usr/bin/env python
|
|
||||||
# -*- coding: utf-8 -*-
|
|
||||||
# ********************************************************************************
|
|
||||||
# Copyright © 2017-2020 jianglin
|
|
||||||
# File Name: src.py
|
|
||||||
# Author: jianglin
|
|
||||||
# Email: mail@honmaple.com
|
|
||||||
# Created: 2018-02-26 12:41:22 (CST)
|
|
||||||
# Last Update: Sunday 2020-08-16 19:45:32 (CST)
|
|
||||||
# By:
|
|
||||||
# Description:
|
|
||||||
# ********************************************************************************
|
|
||||||
try:
|
|
||||||
import pygments
|
|
||||||
from pygments import lexers
|
|
||||||
from pygments import formatters
|
|
||||||
except ImportError:
|
|
||||||
pygments = None
|
|
||||||
|
|
||||||
|
|
||||||
def highlight(language, text):
|
|
||||||
if pygments is None:
|
|
||||||
return text
|
|
||||||
|
|
||||||
try:
|
|
||||||
lexer = lexers.get_lexer_by_name(language)
|
|
||||||
except pygments.util.ClassNotFound:
|
|
||||||
lexer = lexers.guess_lexer(text)
|
|
||||||
formatter = formatters.HtmlFormatter()
|
|
||||||
return pygments.highlight(text, lexer, formatter)
|
|
|
@ -9,7 +9,6 @@ from app.config import AP_CONTENT_TYPE, USER_AGENT
|
||||||
|
|
||||||
from app.database import AsyncSession
|
from app.database import AsyncSession
|
||||||
from app.database import get_db_session
|
from app.database import get_db_session
|
||||||
from app.actor import fetch_actor
|
|
||||||
from sqlalchemy import select
|
from sqlalchemy import select
|
||||||
|
|
||||||
|
|
||||||
|
@ -57,11 +56,20 @@ async def inbox_prechecker(
|
||||||
except Exception:
|
except Exception:
|
||||||
logger.exception("Failed to precheck delete activity")
|
logger.exception("Failed to precheck delete activity")
|
||||||
|
|
||||||
actor_id = payload["actor"]
|
actor_url = payload["actor"]
|
||||||
_actor = await fetch_actor(db_session, actor_id)
|
async with httpx.AsyncClient() as client:
|
||||||
|
resp = await client.get(
|
||||||
|
actor_url,
|
||||||
|
headers={
|
||||||
|
"User-Agent": USER_AGENT,
|
||||||
|
"Accept": AP_CONTENT_TYPE,
|
||||||
|
},
|
||||||
|
follow_redirects=True,
|
||||||
|
)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
pubkey = _actor.ap_actor["publicKey"]["publicKeyPem"]
|
_actor = resp.json()
|
||||||
|
pubkey = _actor["publicKey"]["publicKeyPem"]
|
||||||
except json.JSONDecodeError:
|
except json.JSONDecodeError:
|
||||||
raise ValueError
|
raise ValueError
|
||||||
except KeyError:
|
except KeyError:
|
||||||
|
|
1352
poetry.lock
generated
1352
poetry.lock
generated
File diff suppressed because it is too large
Load diff
|
@ -21,8 +21,6 @@ pydantic = "^1.10.2"
|
||||||
tomli-w = "^1.0.0"
|
tomli-w = "^1.0.0"
|
||||||
invoke = "^2.0.0"
|
invoke = "^2.0.0"
|
||||||
cffi = "^1.15.1"
|
cffi = "^1.15.1"
|
||||||
pyld = "^2.0.3"
|
|
||||||
requests = "^2.28.2"
|
|
||||||
|
|
||||||
|
|
||||||
[tool.poetry.group.dev.dependencies]
|
[tool.poetry.group.dev.dependencies]
|
||||||
|
|
43
tasks.py
43
tasks.py
|
@ -6,19 +6,10 @@ import asyncio
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|
||||||
from prompt_toolkit import prompt
|
from prompt_toolkit import prompt
|
||||||
from prompt_toolkit.key_binding import KeyBindings
|
|
||||||
from invoke import Context # type: ignore
|
from invoke import Context # type: ignore
|
||||||
from invoke import run # type: ignore
|
from invoke import run # type: ignore
|
||||||
from invoke import task # type: ignore
|
from invoke import task # type: ignore
|
||||||
|
|
||||||
_kb = KeyBindings()
|
|
||||||
|
|
||||||
|
|
||||||
@_kb.add("c-@")
|
|
||||||
def _(event):
|
|
||||||
"""Save multi-line buffer on CTRL + space"""
|
|
||||||
event.current_buffer.validate_and_handle()
|
|
||||||
|
|
||||||
@task
|
@task
|
||||||
def config(ctx):
|
def config(ctx):
|
||||||
from Crypto.PublicKey import RSA
|
from Crypto.PublicKey import RSA
|
||||||
|
@ -101,6 +92,7 @@ def accept_follow(ctx):
|
||||||
async def _do():
|
async def _do():
|
||||||
async with async_session() as db_session: #type: ignore
|
async with async_session() as db_session: #type: ignore
|
||||||
try:
|
try:
|
||||||
|
|
||||||
exist_request = (
|
exist_request = (
|
||||||
await db_session.scalars(
|
await db_session.scalars(
|
||||||
select(models.IncomingActivity).where(
|
select(models.IncomingActivity).where(
|
||||||
|
@ -135,36 +127,5 @@ def accept_follow(ctx):
|
||||||
|
|
||||||
print("Done!")
|
print("Done!")
|
||||||
|
|
||||||
|
|
||||||
asyncio.run(_do())
|
asyncio.run(_do())
|
||||||
|
|
||||||
|
|
||||||
@task
|
|
||||||
def send_note(ctx):
|
|
||||||
from app.database import async_session
|
|
||||||
from app.activitypub import VisibilityEnum
|
|
||||||
from app.boxes import _send_create
|
|
||||||
from app.orgpython import to_html
|
|
||||||
|
|
||||||
|
|
||||||
content = prompt(
|
|
||||||
(
|
|
||||||
"note contents, in org mode, "
|
|
||||||
"use [CTRL] + [SPACE] to submit:\n"
|
|
||||||
),
|
|
||||||
key_bindings=_kb,
|
|
||||||
multiline=True,
|
|
||||||
)
|
|
||||||
content = to_html(content)
|
|
||||||
|
|
||||||
async def _dodo():
|
|
||||||
async with async_session() as db_session: #type: ignore
|
|
||||||
|
|
||||||
await _send_create(
|
|
||||||
db_session,
|
|
||||||
"Note",
|
|
||||||
content,
|
|
||||||
VisibilityEnum.PUBLIC
|
|
||||||
)
|
|
||||||
print("Done!")
|
|
||||||
|
|
||||||
asyncio.run(_dodo())
|
|
||||||
|
|
Loading…
Reference in a new issue