From 8240d962ae0f09119fde7b1575924068f02c6d8c Mon Sep 17 00:00:00 2001 From: Victor Uriarte Date: Thu, 2 Jun 2016 12:23:21 -0700 Subject: Replace iter(range(len(...))) with enumerate --- sqlparse/compat.py | 2 -- sqlparse/lexer.py | 9 ++++----- 2 files changed, 4 insertions(+), 7 deletions(-) (limited to 'sqlparse') diff --git a/sqlparse/compat.py b/sqlparse/compat.py index 84d0c96..c1aacf6 100644 --- a/sqlparse/compat.py +++ b/sqlparse/compat.py @@ -18,7 +18,6 @@ if PY3: return str(s) - range = range text_type = str string_types = (str,) from io import StringIO @@ -33,7 +32,6 @@ elif PY2: return unicode(s, encoding) - range = xrange text_type = unicode string_types = (basestring,) from StringIO import StringIO diff --git a/sqlparse/lexer.py b/sqlparse/lexer.py index d2ae8f6..bb7fb48 100644 --- a/sqlparse/lexer.py +++ b/sqlparse/lexer.py @@ -16,7 +16,7 @@ import re from sqlparse import tokens from sqlparse.keywords import SQL_REGEX -from sqlparse.compat import StringIO, string_types, text_type, range +from sqlparse.compat import StringIO, string_types, text_type from sqlparse.utils import consume @@ -67,9 +67,8 @@ class Lexer(object): except UnicodeDecodeError: text = text.decode('unicode-escape') - iterable = iter(range(len(text))) - - for pos in iterable: + iterable = enumerate(text) + for pos, char in iterable: for rexmatch, action, new_state in statetokens: m = rexmatch(text, pos) @@ -93,7 +92,7 @@ class Lexer(object): consume(iterable, m.end() - pos - 1) break else: - yield tokens.Error, text[pos] + yield tokens.Error, char def tokenize(sql, encoding=None): -- cgit v1.2.1