diff options
Diffstat (limited to 'tests/test_basic.py')
-rw-r--r-- | tests/test_basic.py | 72 |
1 files changed, 0 insertions, 72 deletions
diff --git a/tests/test_basic.py b/tests/test_basic.py deleted file mode 100644 index 4488d34e..00000000 --- a/tests/test_basic.py +++ /dev/null @@ -1,72 +0,0 @@ -""" - Pygments Basic lexers tests - ~~~~~~~~~~~~~~~~~~~~~~~~~~~ - - :copyright: Copyright 2006-2021 by the Pygments team, see AUTHORS. - :license: BSD, see LICENSE for details. -""" - -import pytest - -from pygments.lexers.basic import VBScriptLexer -from pygments.token import Error, Name, Number, Punctuation, String, Whitespace - - -@pytest.fixture(scope='module') -def lexer(): - yield VBScriptLexer() - - -def assert_are_tokens_of_type(lexer, examples, expected_token_type): - for test_number, example in enumerate(examples.split(), 1): - token_count = 0 - for token_type, token_value in lexer.get_tokens(example): - if token_type != Whitespace: - token_count += 1 - assert token_type == expected_token_type, \ - 'token_type #%d for %s is be %s but must be %s' % \ - (test_number, token_value, token_type, expected_token_type) - assert token_count == 1, \ - '%s must yield exactly 1 token instead of %d' % (example, token_count) - - -def assert_tokens_match(lexer, text, expected_tokens_without_trailing_newline): - actual_tokens = tuple(lexer.get_tokens(text)) - if (len(actual_tokens) >= 1) and (actual_tokens[-1] == (Whitespace, '\n')): - actual_tokens = tuple(actual_tokens[:-1]) - assert expected_tokens_without_trailing_newline == actual_tokens, \ - 'text must yield expected tokens: %s' % text - - -def test_can_lex_float(lexer): - assert_are_tokens_of_type(lexer, - '1. 1.e1 .1 1.2 1.2e3 1.2e+3 1.2e-3 1e2', - Number.Float) - assert_tokens_match(lexer, - '1e2.1e2', - ((Number.Float, '1e2'), (Number.Float, '.1e2'))) - - -def test_can_reject_almost_float(lexer): - assert_tokens_match(lexer, '.e1', ((Punctuation, '.'), (Name, 'e1'))) - - -def test_can_lex_integer(lexer): - assert_are_tokens_of_type(lexer, '1 23 456', Number.Integer) - - -def test_can_lex_names(lexer): - assert_are_tokens_of_type(lexer, 'thingy thingy123 _thingy _123', Name) - - -def test_can_recover_after_unterminated_string(lexer): - assert_tokens_match(lexer, - '"x\nx', - ((String.Double, '"'), (String.Double, 'x'), - (Error, '\n'), (Name, 'x'))) - - -def test_can_recover_from_invalid_character(lexer): - assert_tokens_match(lexer, - 'a;bc\nd', - ((Name, 'a'), (Error, ';bc\n'), (Name, 'd'))) |