diff options
author | Matthäus G. Chajdas <dev@anteru.net> | 2021-02-14 09:07:56 +0100 |
---|---|---|
committer | Matthäus G. Chajdas <dev@anteru.net> | 2021-02-14 09:07:56 +0100 |
commit | f9bc5a047492153e7e129e916f68c4f3f32c3f36 (patch) | |
tree | 387d83ff411243c9852b039188235a3f56562bbf | |
parent | a2579d3ab55857c1881c172698498ed088b51607 (diff) | |
download | pygments-git-f9bc5a047492153e7e129e916f68c4f3f32c3f36.tar.gz |
Cleanup test_ruby after merge.
-rw-r--r-- | tests/test_ruby.py | 133 |
1 files changed, 1 insertions, 132 deletions
diff --git a/tests/test_ruby.py b/tests/test_ruby.py index 9697bea8..a272926d 100644 --- a/tests/test_ruby.py +++ b/tests/test_ruby.py @@ -9,7 +9,7 @@ import pytest -from pygments.token import Name, Number, Operator, Text, Token +from pygments.token import Name from pygments.lexers.ruby import RubyLexer @@ -18,137 +18,6 @@ def lexer(): yield RubyLexer() -def test_range_syntax1(lexer): - fragment = u'1..3\n' - tokens = [ - (Number.Integer, u'1'), - (Operator, u'..'), - (Number.Integer, u'3'), - (Text, u'\n'), - ] - assert list(lexer.get_tokens(fragment)) == tokens - - -def test_range_syntax2(lexer): - fragment = u'1...3\n' - tokens = [ - (Number.Integer, u'1'), - (Operator, u'...'), - (Number.Integer, u'3'), - (Text, u'\n'), - ] - assert list(lexer.get_tokens(fragment)) == tokens - - -def test_range_syntax3(lexer): - fragment = u'1 .. 3\n' - tokens = [ - (Number.Integer, u'1'), - (Text, u' '), - (Operator, u'..'), - (Text, u' '), - (Number.Integer, u'3'), - (Text, u'\n'), - ] - assert list(lexer.get_tokens(fragment)) == tokens - - -def test_interpolation_nested_curly(lexer): - fragment = ( - u'"A#{ (3..5).group_by { |x| x/2}.map ' - u'do |k,v| "#{k}" end.join }" + "Z"\n') - - tokens = [ - (Token.Literal.String.Double, u'"'), - (Token.Literal.String.Double, u'A'), - (Token.Literal.String.Interpol, u'#{'), - (Token.Text, u' '), - (Token.Punctuation, u'('), - (Token.Literal.Number.Integer, u'3'), - (Token.Operator, u'..'), - (Token.Literal.Number.Integer, u'5'), - (Token.Punctuation, u')'), - (Token.Operator, u'.'), - (Token.Name, u'group_by'), - (Token.Text, u' '), - (Token.Literal.String.Interpol, u'{'), - (Token.Text, u' '), - (Token.Operator, u'|'), - (Token.Name, u'x'), - (Token.Operator, u'|'), - (Token.Text, u' '), - (Token.Name, u'x'), - (Token.Operator, u'/'), - (Token.Literal.Number.Integer, u'2'), - (Token.Literal.String.Interpol, u'}'), - (Token.Operator, u'.'), - (Token.Name, u'map'), - (Token.Text, u' '), - (Token.Keyword, u'do'), - (Token.Text, u' '), - (Token.Operator, u'|'), - (Token.Name, u'k'), - (Token.Punctuation, u','), - (Token.Name, u'v'), - (Token.Operator, u'|'), - (Token.Text, u' '), - (Token.Literal.String.Double, u'"'), - (Token.Literal.String.Interpol, u'#{'), - (Token.Name, u'k'), - (Token.Literal.String.Interpol, u'}'), - (Token.Literal.String.Double, u'"'), - (Token.Text, u' '), - (Token.Keyword, u'end'), - (Token.Operator, u'.'), - (Token.Name, u'join'), - (Token.Text, u' '), - (Token.Literal.String.Interpol, u'}'), - (Token.Literal.String.Double, u'"'), - (Token.Text, u' '), - (Token.Operator, u'+'), - (Token.Text, u' '), - (Token.Literal.String.Double, u'"'), - (Token.Literal.String.Double, u'Z'), - (Token.Literal.String.Double, u'"'), - (Token.Text, u'\n'), - ] - assert list(lexer.get_tokens(fragment)) == tokens - - -def test_operator_methods(lexer): - fragment = u'x.==4\n' - tokens = [ - (Token.Name, u'x'), - (Token.Operator, u'.'), - (Token.Name.Operator, u'=='), - (Token.Literal.Number.Integer, u'4'), - (Token.Text, u'\n'), - ] - assert list(lexer.get_tokens(fragment)) == tokens - - -def test_escaped_bracestring(lexer): - fragment = u'str.gsub(%r{\\\\\\\\}, "/")\n' - tokens = [ - (Token.Name, u'str'), - (Token.Operator, u'.'), - (Token.Name, u'gsub'), - (Token.Punctuation, u'('), - (Token.Literal.String.Regex, u'%r{'), - (Token.Literal.String.Regex, u'\\\\'), - (Token.Literal.String.Regex, u'\\\\'), - (Token.Literal.String.Regex, u'}'), - (Token.Punctuation, u','), - (Token.Text, u' '), - (Token.Literal.String.Double, u'"'), - (Token.Literal.String.Double, u'/'), - (Token.Literal.String.Double, u'"'), - (Token.Punctuation, u')'), - (Token.Text, u'\n'), - ] - assert list(lexer.get_tokens(fragment)) == tokens - - @pytest.mark.parametrize( 'method_name', ( |