forked from kenkeiras/org-rw
Apply autoformatting with black.
This commit is contained in:
parent
d5f8d76aeb
commit
1d71d1a3c3
@ -63,6 +63,7 @@ INACTIVE_TIME_STAMP_RE = re.compile(r"\[{}\]".format(BASE_TIME_STAMP_RE))
|
||||
# BASE_TIME_RANGE_RE = (r'(?P<start_year>\d{4})-(?P<start_month>\d{2})-(?P<start_day>\d{2}) (?P<start_dow>[^ ]+)((?P<start_hour>\d{1,2}):(?P<start_minute>\d{1,2}))?',
|
||||
# r'(?P<end_year>\d{4})-(?P<end_month>\d{2})-(?P<end_day>\d{2}) (?P<end_dow>[^ ]+)((?P<end_hour>\d{1,2}):(?P<end_minute>\d{1,2}))?')
|
||||
|
||||
|
||||
def get_tokens(value):
|
||||
if isinstance(value, Text):
|
||||
return value.contents
|
||||
@ -70,6 +71,7 @@ def get_tokens(value):
|
||||
return [value.line]
|
||||
raise Exception("Unknown how to get tokens from: {}".format(value))
|
||||
|
||||
|
||||
def get_links_from_content(content):
|
||||
in_link = False
|
||||
in_description = False
|
||||
@ -85,7 +87,7 @@ def get_links_from_content(content):
|
||||
elif tok.tok_type == LinkTokenType.CLOSE:
|
||||
in_link = False
|
||||
in_description = False
|
||||
yield Link(''.join(link_value), ''.join(link_description))
|
||||
yield Link("".join(link_value), "".join(link_description))
|
||||
link_value = []
|
||||
link_description = []
|
||||
elif isinstance(tok, str) and in_link:
|
||||
@ -94,8 +96,25 @@ def get_links_from_content(content):
|
||||
else:
|
||||
link_value.append(tok)
|
||||
|
||||
|
||||
class Headline:
|
||||
def __init__(self, start_line, depth, orig, properties, keywords, priority_start, priority, title_start, title, tags_start, tags, contents, children, structural):
|
||||
def __init__(
|
||||
self,
|
||||
start_line,
|
||||
depth,
|
||||
orig,
|
||||
properties,
|
||||
keywords,
|
||||
priority_start,
|
||||
priority,
|
||||
title_start,
|
||||
title,
|
||||
tags_start,
|
||||
tags,
|
||||
contents,
|
||||
children,
|
||||
structural,
|
||||
):
|
||||
self.start_line = start_line
|
||||
self.depth = depth
|
||||
self.orig = orig
|
||||
@ -115,6 +134,7 @@ class Headline:
|
||||
for content in self.contents:
|
||||
yield from get_links_from_content(content)
|
||||
|
||||
|
||||
RawLine = collections.namedtuple("RawLine", ("linenum", "line"))
|
||||
Keyword = collections.namedtuple(
|
||||
"Keyword", ("linenum", "match", "key", "value", "options")
|
||||
@ -140,6 +160,7 @@ class MarkerType(Enum):
|
||||
UNDERLINED_MODE = 0b10000
|
||||
VERBATIM_MODE = 0b100000
|
||||
|
||||
|
||||
MARKERS = {
|
||||
"*": MarkerType.BOLD_MODE,
|
||||
"~": MarkerType.CODE_MODE,
|
||||
@ -157,14 +178,17 @@ for tok, mode in MARKERS.items():
|
||||
MarkerToken = collections.namedtuple("MarkerToken", ("closing", "tok_type"))
|
||||
LinkToken = collections.namedtuple("LinkToken", ("tok_type"))
|
||||
|
||||
|
||||
class LinkTokenType(Enum):
|
||||
OPEN_LINK = 3
|
||||
OPEN_DESCRIPTION = 5
|
||||
CLOSE = 4
|
||||
|
||||
|
||||
BEGIN_PROPERTIES = "OPEN_PROPERTIES"
|
||||
END_PROPERTIES = "CLOSE_PROPERTIES"
|
||||
|
||||
|
||||
def token_from_type(tok_type):
|
||||
return ModeToMarker[tok_type]
|
||||
|
||||
@ -258,9 +282,9 @@ class Link:
|
||||
|
||||
def get_raw(self):
|
||||
if self.description:
|
||||
return '[[{}][{}]]'.format(self.value, self.description)
|
||||
return "[[{}][{}]]".format(self.value, self.description)
|
||||
else:
|
||||
return '[[{}]]'.format(self.value)
|
||||
return "[[{}]]".format(self.value)
|
||||
|
||||
|
||||
class Text:
|
||||
@ -278,16 +302,16 @@ class Text:
|
||||
contents.append(chunk)
|
||||
elif isinstance(chunk, LinkToken):
|
||||
if chunk.tok_type == LinkTokenType.OPEN_LINK:
|
||||
contents.append('[[')
|
||||
contents.append("[[")
|
||||
elif chunk.tok_type == LinkTokenType.OPEN_DESCRIPTION:
|
||||
contents.append('][')
|
||||
contents.append("][")
|
||||
else:
|
||||
assert chunk.tok_type == LinkTokenType.CLOSE
|
||||
contents.append(']]')
|
||||
contents.append("]]")
|
||||
else:
|
||||
assert isinstance(chunk, MarkerToken)
|
||||
contents.append(token_from_type(chunk.tok_type))
|
||||
return ''.join(contents)
|
||||
return "".join(contents)
|
||||
|
||||
|
||||
class Bold:
|
||||
@ -417,62 +441,67 @@ def tokenize_contents(contents: str):
|
||||
tokens.append((TOKEN_TYPE_TEXT, "".join(text)))
|
||||
text = []
|
||||
|
||||
|
||||
cursor = enumerate(contents)
|
||||
for i, char in cursor:
|
||||
has_changed = False
|
||||
|
||||
# Possible link opening
|
||||
if char == '[':
|
||||
if (len(contents) > i + 3
|
||||
if char == "[":
|
||||
if (
|
||||
len(contents) > i + 3
|
||||
# At least 3 characters more to open and close a link
|
||||
and contents[i + 1] == '['):
|
||||
close = contents.find(']', i)
|
||||
and contents[i + 1] == "["
|
||||
):
|
||||
close = contents.find("]", i)
|
||||
|
||||
if close != -1 and contents[close + 1] == ']':
|
||||
if close != -1 and contents[close + 1] == "]":
|
||||
# Link with no description
|
||||
cut_string()
|
||||
|
||||
in_link = True
|
||||
tokens.append((TOKEN_TYPE_OPEN_LINK, None))
|
||||
assert '[' == (next(cursor)[1])
|
||||
assert "[" == (next(cursor)[1])
|
||||
last_link_start = i
|
||||
continue
|
||||
if close != -1 and contents[close + 1] == '[':
|
||||
if close != -1 and contents[close + 1] == "[":
|
||||
# Link with description?
|
||||
|
||||
close = contents.find(']', close + 1)
|
||||
if close != -1 and contents[close + 1] == ']':
|
||||
close = contents.find("]", close + 1)
|
||||
if close != -1 and contents[close + 1] == "]":
|
||||
# No match here means this is not an Org link
|
||||
cut_string()
|
||||
|
||||
in_link = True
|
||||
tokens.append((TOKEN_TYPE_OPEN_LINK, None))
|
||||
assert '[' == (next(cursor)[1])
|
||||
assert "[" == (next(cursor)[1])
|
||||
last_link_start = i
|
||||
continue
|
||||
|
||||
# Possible link close or open of description
|
||||
if char == ']' and in_link:
|
||||
if contents[i + 1] == ']':
|
||||
if char == "]" and in_link:
|
||||
if contents[i + 1] == "]":
|
||||
cut_string()
|
||||
|
||||
tokens.append((TOKEN_TYPE_CLOSE_LINK, None))
|
||||
assert ']' == (next(cursor)[1])
|
||||
assert "]" == (next(cursor)[1])
|
||||
in_link = False
|
||||
in_link_description = False
|
||||
continue
|
||||
|
||||
if contents[i + 1] == '[' and not in_link_description:
|
||||
if contents[i + 1] == "[" and not in_link_description:
|
||||
cut_string()
|
||||
|
||||
tokens.append((TOKEN_TYPE_OPEN_DESCRIPTION, None))
|
||||
assert '[' == (next(cursor)[1])
|
||||
assert "[" == (next(cursor)[1])
|
||||
continue
|
||||
|
||||
raise Exception("Link cannot contain ']' not followed by '[' or ']'. Starting with {}".format(contents[last_link_start:i + 10]))
|
||||
raise Exception(
|
||||
"Link cannot contain ']' not followed by '[' or ']'. Starting with {}".format(
|
||||
contents[last_link_start : i + 10]
|
||||
)
|
||||
)
|
||||
|
||||
if (in_link and not in_link_description):
|
||||
if in_link and not in_link_description:
|
||||
# Link's pointer have no formatting
|
||||
pass
|
||||
|
||||
|
@ -1,5 +1,15 @@
|
||||
from .org_dom import (Bold, Code, Headline, Italic, Line, RawLine, Strike,
|
||||
Text, Underlined, Verbatim)
|
||||
from .org_dom import (
|
||||
Bold,
|
||||
Code,
|
||||
Headline,
|
||||
Italic,
|
||||
Line,
|
||||
RawLine,
|
||||
Strike,
|
||||
Text,
|
||||
Underlined,
|
||||
Verbatim,
|
||||
)
|
||||
|
||||
|
||||
def get_hl_raw_contents(doc: Headline) -> str:
|
||||
@ -8,7 +18,7 @@ def get_hl_raw_contents(doc: Headline) -> str:
|
||||
for content in doc.contents:
|
||||
lines.append(get_raw_contents(content))
|
||||
|
||||
raw = ''.join(lines)
|
||||
raw = "".join(lines)
|
||||
return raw
|
||||
|
||||
|
||||
@ -19,9 +29,11 @@ def get_rawline_contents(doc: RawLine) -> str:
|
||||
def get_span_contents(doc: Line) -> str:
|
||||
return doc.get_raw()
|
||||
|
||||
|
||||
def get_text_contents(doc: Text) -> str:
|
||||
return doc.get_raw()
|
||||
|
||||
|
||||
def get_raw_contents(doc) -> str:
|
||||
if isinstance(doc, Headline):
|
||||
return get_hl_raw_contents(doc)
|
||||
@ -30,8 +42,8 @@ def get_raw_contents(doc) -> str:
|
||||
if isinstance(doc, Line):
|
||||
return get_span_contents(doc)
|
||||
if isinstance(doc, list):
|
||||
return ''.join([get_raw_contents(chunk) for chunk in doc])
|
||||
return "".join([get_raw_contents(chunk) for chunk in doc])
|
||||
if isinstance(doc, (Text, Bold, Code, Italic, Strike, Underlined, Verbatim)):
|
||||
return doc.get_raw()
|
||||
print('Unhandled type: ' + str(doc))
|
||||
raise NotImplementedError('Unhandled type: ' + str(doc))
|
||||
print("Unhandled type: " + str(doc))
|
||||
raise NotImplementedError("Unhandled type: " + str(doc))
|
||||
|
18
setup.py
18
setup.py
@ -1,15 +1,15 @@
|
||||
from setuptools import setup
|
||||
|
||||
setup(
|
||||
name='org-dom',
|
||||
version='0.0.1',
|
||||
description=
|
||||
'Library to de/serialize org-files and manipulate them in a DOM-like manner.',
|
||||
author='kenkeiras',
|
||||
author_email='kenkeiras@codigoparallevar.com',
|
||||
license='Apache License 2.0',
|
||||
packages=['org_dom'],
|
||||
name="org-dom",
|
||||
version="0.0.1",
|
||||
description="Library to de/serialize org-files and manipulate them in a DOM-like manner.",
|
||||
author="kenkeiras",
|
||||
author_email="kenkeiras@codigoparallevar.com",
|
||||
license="Apache License 2.0",
|
||||
packages=["org_dom"],
|
||||
scripts=[],
|
||||
include_package_data=False,
|
||||
install_requires=[],
|
||||
zip_safe=True)
|
||||
zip_safe=True,
|
||||
)
|
||||
|
@ -4,37 +4,60 @@ import unittest
|
||||
from datetime import datetime as DT
|
||||
|
||||
from org_dom import dumps, load, loads
|
||||
from utils.dom_assertions import (BOLD, CODE, HL, ITALIC, SPAN, STRIKE,
|
||||
UNDERLINED, VERBATIM, WEB_LINK, Dom, Tokens)
|
||||
|
||||
from utils.dom_assertions import (
|
||||
BOLD,
|
||||
CODE,
|
||||
HL,
|
||||
ITALIC,
|
||||
SPAN,
|
||||
STRIKE,
|
||||
UNDERLINED,
|
||||
VERBATIM,
|
||||
WEB_LINK,
|
||||
Dom,
|
||||
Tokens,
|
||||
)
|
||||
|
||||
DIR = os.path.dirname(os.path.abspath(__file__))
|
||||
|
||||
|
||||
class TestSerde(unittest.TestCase):
|
||||
def test_simple_file_01(self):
|
||||
with open(os.path.join(DIR, '01-simple.org')) as f:
|
||||
with open(os.path.join(DIR, "01-simple.org")) as f:
|
||||
doc = load(f)
|
||||
|
||||
ex = Dom(props=[('TITLE', '01-Simple'),
|
||||
('DESCRIPTION', 'Simple org file'),
|
||||
('TODO', 'TODO(t) PAUSED(p) | DONE(d)')],
|
||||
children=(HL(
|
||||
'First level',
|
||||
ex = Dom(
|
||||
props=[
|
||||
('ID', '01-simple-first-level-id'),
|
||||
('CREATED', DT(2020, 1, 1, 1, 1)),
|
||||
("TITLE", "01-Simple"),
|
||||
("DESCRIPTION", "Simple org file"),
|
||||
("TODO", "TODO(t) PAUSED(p) | DONE(d)"),
|
||||
],
|
||||
content=' First level content\n',
|
||||
children=(
|
||||
HL(
|
||||
"First level",
|
||||
props=[
|
||||
("ID", "01-simple-first-level-id"),
|
||||
("CREATED", DT(2020, 1, 1, 1, 1)),
|
||||
],
|
||||
content=" First level content\n",
|
||||
children=[
|
||||
HL('Second level',
|
||||
props=[('ID', '01-simple-second-level-id')],
|
||||
content='\n Second level content\n',
|
||||
HL(
|
||||
"Second level",
|
||||
props=[("ID", "01-simple-second-level-id")],
|
||||
content="\n Second level content\n",
|
||||
children=[
|
||||
HL('Third level',
|
||||
props=[('ID', '01-simple-third-level-id')],
|
||||
content='\n Third level content\n')
|
||||
])
|
||||
])))
|
||||
HL(
|
||||
"Third level",
|
||||
props=[("ID", "01-simple-third-level-id")],
|
||||
content="\n Third level content\n",
|
||||
)
|
||||
],
|
||||
)
|
||||
],
|
||||
)
|
||||
),
|
||||
)
|
||||
|
||||
ex.assert_matches(self, doc)
|
||||
|
||||
@ -55,44 +78,75 @@ class TestSerde(unittest.TestCase):
|
||||
self.assertEqual(dumps(doc), orig)
|
||||
|
||||
def test_markup_file_02(self):
|
||||
self.maxDiff = 10000
|
||||
with open(os.path.join(DIR, '02-markup.org')) as f:
|
||||
with open(os.path.join(DIR, "02-markup.org")) as f:
|
||||
doc = load(f)
|
||||
|
||||
ex = Dom(props=[('TITLE', '02-Markup'),
|
||||
('DESCRIPTION', 'Simple org file to test markup'),
|
||||
('TODO', 'TODO(t) PAUSED(p) | DONE(d)')],
|
||||
children=(HL('First level',
|
||||
ex = Dom(
|
||||
props=[
|
||||
('ID', '02-markup-first-level-id'),
|
||||
('CREATED', DT(2020, 1, 1, 1, 1)),
|
||||
("TITLE", "02-Markup"),
|
||||
("DESCRIPTION", "Simple org file to test markup"),
|
||||
("TODO", "TODO(t) PAUSED(p) | DONE(d)"),
|
||||
],
|
||||
children=(
|
||||
HL(
|
||||
"First level",
|
||||
props=[
|
||||
("ID", "02-markup-first-level-id"),
|
||||
("CREATED", DT(2020, 1, 1, 1, 1)),
|
||||
],
|
||||
content=[
|
||||
SPAN(" This is a ", BOLD("bold phrase"),
|
||||
".\n"),
|
||||
SPAN(" This is a ", BOLD("bold phrase"), ".\n"),
|
||||
SPAN("\n"),
|
||||
SPAN(" This is a ",
|
||||
VERBATIM("verbatim phrase"), ".\n"),
|
||||
SPAN(" This is a ", VERBATIM("verbatim phrase"), ".\n"),
|
||||
SPAN("\n"),
|
||||
SPAN(" This is a ", ITALIC("italic phrase"),
|
||||
".\n"),
|
||||
SPAN(" This is a ", ITALIC("italic phrase"), ".\n"),
|
||||
SPAN("\n"),
|
||||
SPAN(" This is a ",
|
||||
STRIKE("strike-through phrase"), ".\n"),
|
||||
SPAN(" This is a ", STRIKE("strike-through phrase"), ".\n"),
|
||||
SPAN("\n"),
|
||||
SPAN(" This is a ",
|
||||
UNDERLINED("underlined phrase"), ".\n"),
|
||||
SPAN(" This is a ", UNDERLINED("underlined phrase"), ".\n"),
|
||||
SPAN("\n"),
|
||||
SPAN(" This is a ", CODE("code phrase"),
|
||||
".\n"),
|
||||
|
||||
SPAN(" This is a ", CODE("code phrase"), ".\n"),
|
||||
SPAN("\n"),
|
||||
SPAN(" This is a nested ", BOLD(["bold ", VERBATIM(["verbatim ", ITALIC(["italic ", STRIKE(["strike ", UNDERLINED(["underlined ", CODE("code ."), " ."]), " ."]), " ."]), " ."]), " ."])),
|
||||
SPAN(
|
||||
" This is a nested ",
|
||||
BOLD(
|
||||
[
|
||||
"bold ",
|
||||
VERBATIM(
|
||||
[
|
||||
"verbatim ",
|
||||
ITALIC(
|
||||
[
|
||||
"italic ",
|
||||
STRIKE(
|
||||
[
|
||||
"strike ",
|
||||
UNDERLINED(
|
||||
[
|
||||
"underlined ",
|
||||
CODE("code ."),
|
||||
" .",
|
||||
]
|
||||
),
|
||||
" .",
|
||||
]
|
||||
),
|
||||
" .",
|
||||
]
|
||||
),
|
||||
" .",
|
||||
]
|
||||
),
|
||||
" .",
|
||||
]
|
||||
),
|
||||
),
|
||||
SPAN("\n"),
|
||||
|
||||
SPAN("\n"),
|
||||
# THIS IS INTERLEAVED, not nested
|
||||
SPAN([" This is a interleaved ",
|
||||
SPAN(
|
||||
[
|
||||
" This is a interleaved ",
|
||||
Tokens.BOLD_START,
|
||||
"bold ",
|
||||
Tokens.VERBATIM_START,
|
||||
@ -116,58 +170,88 @@ class TestSerde(unittest.TestCase):
|
||||
Tokens.UNDERLINED_END,
|
||||
" .",
|
||||
Tokens.CODE_END,
|
||||
"\n"]),
|
||||
|
||||
"\n",
|
||||
]
|
||||
),
|
||||
SPAN("\n"),
|
||||
SPAN(" This is a _ non-underlined phrase because an incorrectly placed content _.\n"),
|
||||
SPAN(
|
||||
" This is a _ non-underlined phrase because an incorrectly placed content _.\n"
|
||||
),
|
||||
SPAN("\n"),
|
||||
|
||||
SPAN(" This is a _ non-underlined phrase because an incorrectly placed content beginning_.\n"),
|
||||
SPAN(
|
||||
" This is a _ non-underlined phrase because an incorrectly placed content beginning_.\n"
|
||||
),
|
||||
SPAN("\n"),
|
||||
|
||||
SPAN(""),
|
||||
SPAN(" This is a _non-underlined phrase because an incorrectly placed content end _.\n"),
|
||||
SPAN(
|
||||
" This is a _non-underlined phrase because an incorrectly placed content end _.\n"
|
||||
),
|
||||
SPAN("\n"),
|
||||
|
||||
SPAN(""),
|
||||
SPAN(" This is a _non-underlined phrase because the lack of an end.\n"),
|
||||
SPAN(
|
||||
" This is a _non-underlined phrase because the lack of an end.\n"
|
||||
),
|
||||
SPAN("\n"),
|
||||
|
||||
SPAN("\n"),
|
||||
SPAN(" This is a _non-underlined phrase because an empty line between beginning and\n"),
|
||||
SPAN(
|
||||
" This is a _non-underlined phrase because an empty line between beginning and\n"
|
||||
),
|
||||
SPAN("\n"),
|
||||
|
||||
SPAN(""),
|
||||
SPAN(" end._\n"),
|
||||
])))
|
||||
],
|
||||
)
|
||||
),
|
||||
)
|
||||
|
||||
ex.assert_matches(self, doc)
|
||||
|
||||
def test_links_file_03(self):
|
||||
with open(os.path.join(DIR, '03-links.org')) as f:
|
||||
with open(os.path.join(DIR, "03-links.org")) as f:
|
||||
doc = load(f)
|
||||
|
||||
links = list(doc.get_links())
|
||||
self.assertEqual(len(links), 2)
|
||||
self.assertEqual(links[0].value, 'https://codigoparallevar.com/1')
|
||||
self.assertEqual(links[0].description, 'web link')
|
||||
self.assertEqual(links[0].value, "https://codigoparallevar.com/1")
|
||||
self.assertEqual(links[0].description, "web link")
|
||||
|
||||
self.assertEqual(links[1].value, 'https://codigoparallevar.com/2')
|
||||
self.assertEqual(links[1].description, 'web link')
|
||||
ex = Dom(props=[('TITLE', '03-Links'),
|
||||
('DESCRIPTION', 'Simple org file to test links'),
|
||||
('TODO', 'TODO(t) PAUSED(p) | DONE(d)')],
|
||||
children=(HL('First level',
|
||||
self.assertEqual(links[1].value, "https://codigoparallevar.com/2")
|
||||
self.assertEqual(links[1].description, "web link")
|
||||
ex = Dom(
|
||||
props=[
|
||||
('ID', '03-markup-first-level-id'),
|
||||
('CREATED', DT(2020, 1, 1, 1, 1)),
|
||||
("TITLE", "03-Links"),
|
||||
("DESCRIPTION", "Simple org file to test links"),
|
||||
("TODO", "TODO(t) PAUSED(p) | DONE(d)"),
|
||||
],
|
||||
children=(
|
||||
HL(
|
||||
"First level",
|
||||
props=[
|
||||
("ID", "03-markup-first-level-id"),
|
||||
("CREATED", DT(2020, 1, 1, 1, 1)),
|
||||
],
|
||||
content=[
|
||||
SPAN(" This is a ", WEB_LINK("web link", "https://codigoparallevar.com/1"),
|
||||
".\n"),
|
||||
SPAN(
|
||||
" This is a ",
|
||||
WEB_LINK("web link", "https://codigoparallevar.com/1"),
|
||||
".\n",
|
||||
),
|
||||
SPAN("\n"),
|
||||
SPAN(" This is a ", ITALIC(["italized ", WEB_LINK("web link", "https://codigoparallevar.com/2")]),
|
||||
".\n"),
|
||||
])))
|
||||
SPAN(
|
||||
" This is a ",
|
||||
ITALIC(
|
||||
[
|
||||
"italized ",
|
||||
WEB_LINK(
|
||||
"web link", "https://codigoparallevar.com/2"
|
||||
),
|
||||
]
|
||||
),
|
||||
".\n",
|
||||
),
|
||||
],
|
||||
)
|
||||
),
|
||||
)
|
||||
|
||||
ex.assert_matches(self, doc)
|
||||
|
@ -2,8 +2,17 @@ import collections
|
||||
import unittest
|
||||
from datetime import datetime
|
||||
|
||||
from org_dom import (Bold, Code, Italic, Line, Strike, Text, Underlined,
|
||||
Verbatim, get_raw_contents)
|
||||
from org_dom import (
|
||||
Bold,
|
||||
Code,
|
||||
Italic,
|
||||
Line,
|
||||
Strike,
|
||||
Text,
|
||||
Underlined,
|
||||
Verbatim,
|
||||
get_raw_contents,
|
||||
)
|
||||
|
||||
|
||||
def timestamp_to_datetime(ts):
|
||||
|
Loading…
Reference in New Issue
Block a user