Fix text parsing issues #9
@ -103,7 +103,7 @@ PLANNING_RE = re.compile(
|
|||||||
r")+\s*"
|
r")+\s*"
|
||||||
)
|
)
|
||||||
LIST_ITEM_RE = re.compile(
|
LIST_ITEM_RE = re.compile(
|
||||||
r"(?P<indentation>\s*)((?P<bullet>[*\-+])|((?P<counter>\d|[a-zA-Z])(?P<counter_sep>[.)]))) ((?P<checkbox_indentation>\s*)\[(?P<checkbox_value>[ Xx])\])?((?P<tag_indentation>\s*)(?P<tag>.*?)::)?(?P<content>.*)"
|
r"(?P<indentation>\s*)((?P<bullet>[*\-+])|((?P<counter>\d|[a-zA-Z])(?P<counter_sep>[.)]))) ((?P<checkbox_indentation>\s*)\[(?P<checkbox_value>[ Xx])\])?((?P<tag_indentation>\s*)((?P<tag>.*?)\s::))?(?P<content>.*)"
|
||||||
)
|
)
|
||||||
|
|
||||||
IMPLICIT_LINK_RE = re.compile(r"(https?:[^<> ]*[a-zA-Z0-9])")
|
IMPLICIT_LINK_RE = re.compile(r"(https?:[^<> ]*[a-zA-Z0-9])")
|
||||||
@ -1911,7 +1911,12 @@ def tokenize_contents(contents: str) -> List[TokenItems]:
|
|||||||
continue
|
continue
|
||||||
|
|
||||||
# Possible link close or open of description
|
# Possible link close or open of description
|
||||||
if char == "]" and len(contents) > i + 1 and in_link:
|
if (
|
||||||
|
char == "]"
|
||||||
|
and len(contents) > i + 1
|
||||||
|
and in_link
|
||||||
|
and contents[i + 1] in "]["
|
||||||
|
):
|
||||||
if contents[i + 1] == "]":
|
if contents[i + 1] == "]":
|
||||||
cut_string()
|
cut_string()
|
||||||
|
|
||||||
@ -1962,6 +1967,7 @@ def tokenize_contents(contents: str) -> List[TokenItems]:
|
|||||||
cut_string()
|
cut_string()
|
||||||
tokens.append((TOKEN_TYPE_CLOSE_MARKER, char))
|
tokens.append((TOKEN_TYPE_CLOSE_MARKER, char))
|
||||||
has_changed = True
|
has_changed = True
|
||||||
|
closes.remove(i)
|
||||||
|
|
||||||
if not has_changed:
|
if not has_changed:
|
||||||
text.append(char)
|
text.append(char)
|
||||||
|
Loading…
Reference in New Issue
Block a user