Fix text parsing issues #9
@ -1962,6 +1962,7 @@ def tokenize_contents(contents: str) -> List[TokenItems]:
|
|||||||
cut_string()
|
cut_string()
|
||||||
tokens.append((TOKEN_TYPE_CLOSE_MARKER, char))
|
tokens.append((TOKEN_TYPE_CLOSE_MARKER, char))
|
||||||
has_changed = True
|
has_changed = True
|
||||||
|
closes.remove(i)
|
||||||
|
|
||||||
if not has_changed:
|
if not has_changed:
|
||||||
text.append(char)
|
text.append(char)
|
||||||
|
Loading…
Reference in New Issue
Block a user