115 lines
4.1 KiB
Python
115 lines
4.1 KiB
Python
"""
|
|
Pygments SQL lexers tests
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
:copyright: Copyright 2006-2023 by the Pygments team, see AUTHORS.
|
|
:license: BSD, see LICENSE for details.
|
|
"""
|
|
|
|
import pytest
|
|
|
|
from pygments.lexers.sql import name_between_bracket_re, \
|
|
name_between_backtick_re, tsql_go_re, tsql_declare_re, \
|
|
tsql_variable_re, MySqlLexer, TransactSqlLexer
|
|
|
|
from pygments.token import Comment, Name, Number, Punctuation, Whitespace
|
|
|
|
|
|
@pytest.fixture(scope='module')
|
|
def lexer():
|
|
yield TransactSqlLexer()
|
|
|
|
|
|
def _assert_are_tokens_of_type(lexer, examples, expected_token_type):
|
|
for test_number, example in enumerate(examples.split(), 1):
|
|
token_count = 0
|
|
for token_type, token_value in lexer.get_tokens(example):
|
|
if token_type != Whitespace:
|
|
token_count += 1
|
|
assert token_type == expected_token_type, \
|
|
'token_type #%d for %s is be %s but must be %s' % \
|
|
(test_number, token_value, token_type, expected_token_type)
|
|
assert token_count == 1, \
|
|
'%s must yield exactly 1 token instead of %d' % \
|
|
(example, token_count)
|
|
|
|
|
|
def _assert_tokens_match(lexer, text, expected_tokens_without_trailing_newline):
|
|
actual_tokens = tuple(lexer.get_tokens(text))
|
|
if (len(actual_tokens) >= 1) and (actual_tokens[-1] == (Whitespace, '\n')):
|
|
actual_tokens = tuple(actual_tokens[:-1])
|
|
assert expected_tokens_without_trailing_newline == actual_tokens, \
|
|
'text must yield expected tokens: %s' % text
|
|
|
|
|
|
def test_can_lex_float(lexer):
|
|
_assert_are_tokens_of_type(lexer,
|
|
'1. 1.e1 .1 1.2 1.2e3 1.2e+3 1.2e-3 1e2',
|
|
Number.Float)
|
|
_assert_tokens_match(lexer,
|
|
'1e2.1e2',
|
|
((Number.Float, '1e2'), (Number.Float, '.1e2')))
|
|
|
|
|
|
def test_can_reject_almost_float(lexer):
|
|
_assert_tokens_match(lexer, '.e1', ((Punctuation, '.'), (Name, 'e1')))
|
|
|
|
|
|
def test_can_lex_integer(lexer):
|
|
_assert_are_tokens_of_type(lexer, '1 23 456', Number.Integer)
|
|
|
|
|
|
def test_can_lex_names(lexer):
|
|
_assert_are_tokens_of_type(
|
|
lexer,
|
|
'thingy thingy123 _thingy _ _123 Ähnliches Müll #temp1 ##temp2',
|
|
Name)
|
|
|
|
|
|
def test_can_lex_comments(lexer):
|
|
_assert_tokens_match(lexer, '--\n', ((Comment.Single, '--\n'),))
|
|
_assert_tokens_match(lexer, '/**/', (
|
|
(Comment.Multiline, '/*'), (Comment.Multiline, '*/')
|
|
))
|
|
_assert_tokens_match(lexer, '/*/**/*/', (
|
|
(Comment.Multiline, '/*'),
|
|
(Comment.Multiline, '/*'),
|
|
(Comment.Multiline, '*/'),
|
|
(Comment.Multiline, '*/'),
|
|
))
|
|
|
|
|
|
def test_can_match_analyze_text_res():
|
|
assert ['`a`', '`bc`'] == \
|
|
name_between_backtick_re.findall('select `a`, `bc` from some')
|
|
assert ['[a]', '[bc]'] == \
|
|
name_between_bracket_re.findall('select [a], [bc] from some')
|
|
assert tsql_declare_re.search('--\nDeClaRe @some int;')
|
|
assert tsql_go_re.search('select 1\ngo\n--')
|
|
assert tsql_variable_re.search('create procedure dbo.usp_x @a int, @b int')
|
|
|
|
|
|
def test_can_analyze_text():
|
|
mysql_lexer = MySqlLexer()
|
|
tsql_lexer = TransactSqlLexer()
|
|
code_to_expected_lexer_map = {
|
|
'select `a`, `bc` from some': mysql_lexer,
|
|
'select [a], [bc] from some': tsql_lexer,
|
|
'-- `a`, `bc`\nselect [a], [bc] from some': tsql_lexer,
|
|
'-- `a`, `bc`\nselect [a], [bc] from some; go': tsql_lexer,
|
|
}
|
|
sql_lexers = set(code_to_expected_lexer_map.values())
|
|
for code, expected_lexer in code_to_expected_lexer_map.items():
|
|
ratings_and_lexers = list((lexer.analyse_text(code), lexer.name) for lexer in sql_lexers)
|
|
best_rating, best_lexer_name = sorted(ratings_and_lexers, reverse=True)[0]
|
|
expected_rating = expected_lexer.analyse_text(code)
|
|
message = (
|
|
'lexer must be %s (rating %.2f) instead of '
|
|
'%s (rating %.2f) for analyse_text() on code:\n%s') % (
|
|
expected_lexer.name,
|
|
expected_rating,
|
|
best_lexer_name,
|
|
best_rating,
|
|
code
|
|
)
|
|
assert expected_lexer.name == best_lexer_name, message
|