forked from kenkeiras/org-rw
Properly track which tokens are used for closing formats.
This commit is contained in:
parent
490b36887a
commit
f31c64c242
1 changed files with 1 additions and 0 deletions
|
@ -1962,6 +1962,7 @@ def tokenize_contents(contents: str) -> List[TokenItems]:
|
|||
cut_string()
|
||||
tokens.append((TOKEN_TYPE_CLOSE_MARKER, char))
|
||||
has_changed = True
|
||||
closes.remove(i)
|
||||
|
||||
if not has_changed:
|
||||
text.append(char)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue