diff options
Diffstat (limited to 'tests/test_sql.py')
-rw-r--r-- | tests/test_sql.py | 203 |
1 files changed, 101 insertions, 102 deletions
diff --git a/tests/test_sql.py b/tests/test_sql.py index 6be34006..efd63be6 100644 --- a/tests/test_sql.py +++ b/tests/test_sql.py @@ -3,10 +3,11 @@ Pygments SQL lexers tests ~~~~~~~~~~~~~~~~~~~~~~~~~ - :copyright: Copyright 2006-2016 by the Pygments team, see AUTHORS. + :copyright: Copyright 2006-2019 by the Pygments team, see AUTHORS. :license: BSD, see LICENSE for details. """ -import unittest + +import pytest from pygments.lexers.sql import name_between_bracket_re, \ name_between_backtick_re, tsql_go_re, tsql_declare_re, \ @@ -15,104 +16,102 @@ from pygments.lexers.sql import name_between_bracket_re, \ from pygments.token import Comment, Name, Number, Punctuation, Whitespace -class TransactSqlLexerTest(unittest.TestCase): - - def setUp(self): - self.lexer = TransactSqlLexer() - - def _assertAreTokensOfType(self, examples, expected_token_type): - for test_number, example in enumerate(examples.split(), 1): - token_count = 0 - for token_type, token_value in self.lexer.get_tokens(example): - if token_type != Whitespace: - token_count += 1 - self.assertEqual( - token_type, expected_token_type, - 'token_type #%d for %s is be %s but must be %s' % - (test_number, token_value, token_type, expected_token_type)) - self.assertEqual( - token_count, 1, - '%s must yield exactly 1 token instead of %d' % - (example, token_count)) - - def _assertTokensMatch(self, text, expected_tokens_without_trailing_newline): - actual_tokens = tuple(self.lexer.get_tokens(text)) - if (len(actual_tokens) >= 1) and (actual_tokens[-1] == (Whitespace, '\n')): - actual_tokens = tuple(actual_tokens[:-1]) - self.assertEqual( - expected_tokens_without_trailing_newline, actual_tokens, - 'text must yield expected tokens: %s' % text) - - def test_can_lex_float(self): - self._assertAreTokensOfType( - '1. 1.e1 .1 1.2 1.2e3 1.2e+3 1.2e-3 1e2', Number.Float) - self._assertTokensMatch( - '1e2.1e2', - ((Number.Float, '1e2'), (Number.Float, '.1e2')) +@pytest.fixture(scope='module') +def lexer(): + yield TransactSqlLexer() + + +def _assert_are_tokens_of_type(lexer, examples, expected_token_type): + for test_number, example in enumerate(examples.split(), 1): + token_count = 0 + for token_type, token_value in lexer.get_tokens(example): + if token_type != Whitespace: + token_count += 1 + assert token_type == expected_token_type, \ + 'token_type #%d for %s is be %s but must be %s' % \ + (test_number, token_value, token_type, expected_token_type) + assert token_count == 1, \ + '%s must yield exactly 1 token instead of %d' % \ + (example, token_count) + + +def _assert_tokens_match(lexer, text, expected_tokens_without_trailing_newline): + actual_tokens = tuple(lexer.get_tokens(text)) + if (len(actual_tokens) >= 1) and (actual_tokens[-1] == (Whitespace, '\n')): + actual_tokens = tuple(actual_tokens[:-1]) + assert expected_tokens_without_trailing_newline == actual_tokens, \ + 'text must yield expected tokens: %s' % text + + +def test_can_lex_float(lexer): + _assert_are_tokens_of_type(lexer, + '1. 1.e1 .1 1.2 1.2e3 1.2e+3 1.2e-3 1e2', + Number.Float) + _assert_tokens_match(lexer, + '1e2.1e2', + ((Number.Float, '1e2'), (Number.Float, '.1e2'))) + + +def test_can_reject_almost_float(lexer): + _assert_tokens_match(lexer, '.e1', ((Punctuation, '.'), (Name, 'e1'))) + + +def test_can_lex_integer(lexer): + _assert_are_tokens_of_type(lexer, '1 23 456', Number.Integer) + + +def test_can_lex_names(lexer): + _assert_are_tokens_of_type(lexer, + u'thingy thingy123 _thingy _ _123 Ähnliches Müll #temp1 ##temp2', + Name) + + +def test_can_lex_comments(lexer): + _assert_tokens_match(lexer, '--\n', ((Comment.Single, '--\n'),)) + _assert_tokens_match(lexer, '/**/', ( + (Comment.Multiline, '/*'), (Comment.Multiline, '*/') + )) + _assert_tokens_match(lexer, '/*/**/*/', ( + (Comment.Multiline, '/*'), + (Comment.Multiline, '/*'), + (Comment.Multiline, '*/'), + (Comment.Multiline, '*/'), + )) + + +def test_can_match_analyze_text_res(): + assert ['`a`', '`bc`'] == \ + name_between_backtick_re.findall('select `a`, `bc` from some') + assert ['[a]', '[bc]'] == \ + name_between_bracket_re.findall('select [a], [bc] from some') + assert tsql_declare_re.search('--\nDeClaRe @some int;') + assert tsql_go_re.search('select 1\ngo\n--') + assert tsql_variable_re.search('create procedure dbo.usp_x @a int, @b int') + + +def test_can_analyze_text(): + mysql_lexer = MySqlLexer() + sql_lexer = SqlLexer() + tsql_lexer = TransactSqlLexer() + code_to_expected_lexer_map = { + 'select `a`, `bc` from some': mysql_lexer, + 'select a, bc from some': sql_lexer, + 'select [a], [bc] from some': tsql_lexer, + '-- `a`, `bc`\nselect [a], [bc] from some': tsql_lexer, + '-- `a`, `bc`\nselect [a], [bc] from some; go': tsql_lexer, + } + sql_lexers = set(code_to_expected_lexer_map.values()) + for code, expected_lexer in code_to_expected_lexer_map.items(): + ratings_and_lexers = list((lexer.analyse_text(code), lexer.name) for lexer in sql_lexers) + best_rating, best_lexer_name = sorted(ratings_and_lexers, reverse=True)[0] + expected_rating = expected_lexer.analyse_text(code) + message = ( + 'lexer must be %s (rating %.2f) instead of ' + '%s (rating %.2f) for analyse_text() on code:\n%s') % ( + expected_lexer.name, + expected_rating, + best_lexer_name, + best_rating, + code ) - - def test_can_reject_almost_float(self): - self._assertTokensMatch( - '.e1', - ((Punctuation, '.'), (Name, 'e1'))) - - def test_can_lex_integer(self): - self._assertAreTokensOfType( - '1 23 456', Number.Integer) - - def test_can_lex_names(self): - self._assertAreTokensOfType( - u'thingy thingy123 _thingy _ _123 Ähnliches Müll #temp1 ##temp2', Name) - - def test_can_lex_comments(self): - self._assertTokensMatch('--\n', ((Comment.Single, '--\n'),)) - self._assertTokensMatch('/**/', ( - (Comment.Multiline, '/*'), (Comment.Multiline, '*/') - )) - self._assertTokensMatch('/*/**/*/', ( - (Comment.Multiline, '/*'), - (Comment.Multiline, '/*'), - (Comment.Multiline, '*/'), - (Comment.Multiline, '*/'), - )) - - -class SqlAnalyzeTextTest(unittest.TestCase): - def test_can_match_analyze_text_res(self): - self.assertEqual(['`a`', '`bc`'], - name_between_backtick_re.findall('select `a`, `bc` from some')) - self.assertEqual(['[a]', '[bc]'], - name_between_bracket_re.findall('select [a], [bc] from some')) - self.assertTrue(tsql_declare_re.search('--\nDeClaRe @some int;')) - self.assertTrue(tsql_go_re.search('select 1\ngo\n--')) - self.assertTrue(tsql_variable_re.search( - 'create procedure dbo.usp_x @a int, @b int')) - - def test_can_analyze_text(self): - mysql_lexer = MySqlLexer() - sql_lexer = SqlLexer() - tsql_lexer = TransactSqlLexer() - code_to_expected_lexer_map = { - 'select `a`, `bc` from some': mysql_lexer, - 'select a, bc from some': sql_lexer, - 'select [a], [bc] from some': tsql_lexer, - '-- `a`, `bc`\nselect [a], [bc] from some': tsql_lexer, - '-- `a`, `bc`\nselect [a], [bc] from some; go': tsql_lexer, - } - sql_lexers = set(code_to_expected_lexer_map.values()) - for code, expected_lexer in code_to_expected_lexer_map.items(): - ratings_and_lexers = list((lexer.analyse_text(code), lexer.name) for lexer in sql_lexers) - best_rating, best_lexer_name = sorted(ratings_and_lexers, reverse=True)[0] - expected_rating = expected_lexer.analyse_text(code) - message = ( - 'lexer must be %s (rating %.2f) instead of ' - '%s (rating %.2f) for analyse_text() on code:\n%s') % ( - expected_lexer.name, - expected_rating, - best_lexer_name, - best_rating, - code - ) - self.assertEqual( - expected_lexer.name, best_lexer_name, message - ) + assert expected_lexer.name == best_lexer_name, message |