Fix text parsing issues #9
1 changed files with 1 additions and 0 deletions
|
@ -1962,6 +1962,7 @@ def tokenize_contents(contents: str) -> List[TokenItems]:
|
||||||
cut_string()
|
cut_string()
|
||||||
tokens.append((TOKEN_TYPE_CLOSE_MARKER, char))
|
tokens.append((TOKEN_TYPE_CLOSE_MARKER, char))
|
||||||
has_changed = True
|
has_changed = True
|
||||||
|
closes.remove(i)
|
||||||
|
|
||||||
if not has_changed:
|
if not has_changed:
|
||||||
text.append(char)
|
text.append(char)
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue