From 7c83462a616ae5ddeaa5e9777da50d51b5e3ff2c Mon Sep 17 00:00:00 2001 From: fsp Date: Tue, 18 Mar 2014 13:56:29 +0800 Subject: [PATCH] Removed duplication from ignored_tokens in lexer --- jinja2/lexer.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/jinja2/lexer.py b/jinja2/lexer.py index 36042e01..d7c7e5ec 100644 --- a/jinja2/lexer.py +++ b/jinja2/lexer.py @@ -136,8 +136,8 @@ operator_re = re.compile('(%s)' % '|'.join(re.escape(x) for x in ignored_tokens = frozenset([TOKEN_COMMENT_BEGIN, TOKEN_COMMENT, TOKEN_COMMENT_END, TOKEN_WHITESPACE, - TOKEN_WHITESPACE, TOKEN_LINECOMMENT_BEGIN, - TOKEN_LINECOMMENT_END, TOKEN_LINECOMMENT]) + TOKEN_LINECOMMENT_BEGIN, TOKEN_LINECOMMENT_END, + TOKEN_LINECOMMENT]) ignore_if_empty = frozenset([TOKEN_WHITESPACE, TOKEN_DATA, TOKEN_COMMENT, TOKEN_LINECOMMENT]) -- 2.47.2