""" Pygments SQL lexers tests ~~~~~~~~~~~~~~~~~~~~~~~~~ :copyright: Copyright 2006-2023 by the Pygments team, see AUTHORS. :license: BSD, see LICENSE for details. """ import pytest from pygments.lexers.sql import name_between_bracket_re, \ name_between_backtick_re, tsql_go_re, tsql_declare_re, \ tsql_variable_re, MySqlLexer, TransactSqlLexer from pygments.token import Comment, Name, Number, Punctuation, Whitespace @pytest.fixture(scope='module') def lexer(): yield TransactSqlLexer() def _assert_are_tokens_of_type(lexer, examples, expected_token_type): for test_number, example in enumerate(examples.split(), 1): token_count = 0 for token_type, token_value in lexer.get_tokens(example): if token_type != Whitespace: token_count += 1 assert token_type == expected_token_type, \ 'token_type #%d for %s is be %s but must be %s' % \ (test_number, token_value, token_type, expected_token_type) assert token_count == 1, \ '%s must yield exactly 1 token instead of %d' % \ (example, token_count) def _assert_tokens_match(lexer, text, expected_tokens_without_trailing_newline): actual_tokens = tuple(lexer.get_tokens(text)) if (len(actual_tokens) >= 1) and (actual_tokens[-1] == (Whitespace, '\n')): actual_tokens = tuple(actual_tokens[:-1]) assert expected_tokens_without_trailing_newline == actual_tokens, \ 'text must yield expected tokens: %s' % text def test_can_lex_float(lexer): _assert_are_tokens_of_type(lexer, '1. 1.e1 .1 1.2 1.2e3 1.2e+3 1.2e-3 1e2', Number.Float) _assert_tokens_match(lexer, '1e2.1e2', ((Number.Float, '1e2'), (Number.Float, '.1e2'))) def test_can_reject_almost_float(lexer): _assert_tokens_match(lexer, '.e1', ((Punctuation, '.'), (Name, 'e1'))) def test_can_lex_integer(lexer): _assert_are_tokens_of_type(lexer, '1 23 456', Number.Integer) def test_can_lex_names(lexer): _assert_are_tokens_of_type( lexer, 'thingy thingy123 _thingy _ _123 Ähnliches Müll #temp1 ##temp2', Name) def test_can_lex_comments(lexer): _assert_tokens_match(lexer, '--\n', ((Comment.Single, '--\n'),)) _assert_tokens_match(lexer, '/**/', ( (Comment.Multiline, '/*'), (Comment.Multiline, '*/') )) _assert_tokens_match(lexer, '/*/**/*/', ( (Comment.Multiline, '/*'), (Comment.Multiline, '/*'), (Comment.Multiline, '*/'), (Comment.Multiline, '*/'), )) def test_can_match_analyze_text_res(): assert ['`a`', '`bc`'] == \ name_between_backtick_re.findall('select `a`, `bc` from some') assert ['[a]', '[bc]'] == \ name_between_bracket_re.findall('select [a], [bc] from some') assert tsql_declare_re.search('--\nDeClaRe @some int;') assert tsql_go_re.search('select 1\ngo\n--') assert tsql_variable_re.search('create procedure dbo.usp_x @a int, @b int') def test_can_analyze_text(): mysql_lexer = MySqlLexer() tsql_lexer = TransactSqlLexer() code_to_expected_lexer_map = { 'select `a`, `bc` from some': mysql_lexer, 'select [a], [bc] from some': tsql_lexer, '-- `a`, `bc`\nselect [a], [bc] from some': tsql_lexer, '-- `a`, `bc`\nselect [a], [bc] from some; go': tsql_lexer, } sql_lexers = set(code_to_expected_lexer_map.values()) for code, expected_lexer in code_to_expected_lexer_map.items(): ratings_and_lexers = list((lexer.analyse_text(code), lexer.name) for lexer in sql_lexers) best_rating, best_lexer_name = sorted(ratings_and_lexers, reverse=True)[0] expected_rating = expected_lexer.analyse_text(code) message = ( 'lexer must be %s (rating %.2f) instead of ' '%s (rating %.2f) for analyse_text() on code:\n%s') % ( expected_lexer.name, expected_rating, best_lexer_name, best_rating, code ) assert expected_lexer.name == best_lexer_name, message