Fix text parsing issues #9
@ -1911,7 +1911,7 @@ def tokenize_contents(contents: str) -> List[TokenItems]:
|
|||||||
continue
|
continue
|
||||||
|
|
||||||
# Possible link close or open of description
|
# Possible link close or open of description
|
||||||
if char == "]" and len(contents) > i + 1 and in_link:
|
if char == "]" and len(contents) > i + 1 and in_link and contents[i + 1] in "][":
|
||||||
if contents[i + 1] == "]":
|
if contents[i + 1] == "]":
|
||||||
cut_string()
|
cut_string()
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user