summaryrefslogtreecommitdiff
path: root/tests/test_r.py
diff options
context:
space:
mode:
authorOleh Prypin <oleh@pryp.in>2021-01-18 20:39:02 +0100
committerGitHub <noreply@github.com>2021-01-18 20:39:02 +0100
commitf0445be718da83541ea3401aad882f3937147263 (patch)
tree26bd361a410d8bea33ce259321fad63e7f3c61af /tests/test_r.py
parent423c44a451db7e5f63147b1c1519661d745fc43a (diff)
downloadpygments-git-f0445be718da83541ea3401aad882f3937147263.tar.gz
Replace tests that assert on token output with auto-updatable samples (#1649)
Diffstat (limited to 'tests/test_r.py')
-rw-r--r--tests/test_r.py111
1 files changed, 0 insertions, 111 deletions
diff --git a/tests/test_r.py b/tests/test_r.py
deleted file mode 100644
index ffd3f576..00000000
--- a/tests/test_r.py
+++ /dev/null
@@ -1,111 +0,0 @@
-"""
- R Tests
- ~~~~~~~
-
- :copyright: Copyright 2006-2021 by the Pygments team, see AUTHORS.
- :license: BSD, see LICENSE for details.
-"""
-
-import pytest
-
-from pygments.lexers import SLexer
-from pygments.token import Token, Name, Punctuation
-
-
-@pytest.fixture(scope='module')
-def lexer():
- yield SLexer()
-
-
-def test_call(lexer):
- fragment = 'f(1, a)\n'
- tokens = [
- (Name.Function, 'f'),
- (Punctuation, '('),
- (Token.Literal.Number, '1'),
- (Punctuation, ','),
- (Token.Text, ' '),
- (Token.Name, 'a'),
- (Punctuation, ')'),
- (Token.Text, '\n'),
- ]
- assert list(lexer.get_tokens(fragment)) == tokens
-
-
-def test_name1(lexer):
- fragment = '._a_2.c'
- tokens = [
- (Name, '._a_2.c'),
- (Token.Text, '\n'),
- ]
- assert list(lexer.get_tokens(fragment)) == tokens
-
-
-def test_name2(lexer):
- # Invalid names are valid if backticks are used
- fragment = '`.1 blah`'
- tokens = [
- (Name, '`.1 blah`'),
- (Token.Text, '\n'),
- ]
- assert list(lexer.get_tokens(fragment)) == tokens
-
-
-def test_name3(lexer):
- # Internal backticks can be escaped
- fragment = '`.1 \\` blah`'
- tokens = [
- (Name, '`.1 \\` blah`'),
- (Token.Text, '\n'),
- ]
- assert list(lexer.get_tokens(fragment)) == tokens
-
-
-def test_custom_operator(lexer):
- fragment = '7 % and % 8'
- tokens = [
- (Token.Literal.Number, '7'),
- (Token.Text, ' '),
- (Token.Operator, '% and %'),
- (Token.Text, ' '),
- (Token.Literal.Number, '8'),
- (Token.Text, '\n'),
- ]
- assert list(lexer.get_tokens(fragment)) == tokens
-
-
-def test_indexing(lexer):
- fragment = 'a[1]'
- tokens = [
- (Token.Name, 'a'),
- (Token.Punctuation, '['),
- (Token.Literal.Number, '1'),
- (Token.Punctuation, ']'),
- (Token.Text, '\n'),
- ]
- assert list(lexer.get_tokens(fragment)) == tokens
-
-
-def test_dot_name(lexer):
- fragment = '. <- 1'
- tokens = [
- (Token.Name, '.'),
- (Token.Text, ' '),
- (Token.Operator, '<-'),
- (Token.Text, ' '),
- (Token.Literal.Number, '1'),
- (Token.Text, '\n')
- ]
- assert list(lexer.get_tokens(fragment)) == tokens
-
-
-def test_dot_indexing(lexer):
- fragment = '.[1]'
- tokens = [
- (Token.Name, '.'),
- (Token.Punctuation, '['),
- (Token.Literal.Number, '1'),
- (Token.Punctuation, ']'),
- (Token.Text, '\n'),
- ]
- assert list(lexer.get_tokens(fragment)) == tokens