|
4 | 4 |
|
5 | 5 | namespace Doctrine\SqlFormatter\Tests;
|
6 | 6 |
|
| 7 | +use Doctrine\SqlFormatter\Cursor; |
| 8 | +use Doctrine\SqlFormatter\Token; |
7 | 9 | use Doctrine\SqlFormatter\Tokenizer;
|
8 |
| -use PHPUnit\Framework\Attributes\DoesNotPerformAssertions; |
| 10 | +use Generator; |
| 11 | +use PHPUnit\Framework\Attributes\DataProvider; |
9 | 12 | use PHPUnit\Framework\TestCase;
|
10 | 13 | use ReflectionClass;
|
11 | 14 |
|
12 | 15 | use function array_filter;
|
| 16 | +use function implode; |
13 | 17 | use function preg_match;
|
| 18 | +use function serialize; |
14 | 19 | use function sort;
|
15 | 20 | use function strtoupper;
|
16 | 21 |
|
@@ -58,9 +63,131 @@ public function testKeywordsReservedAreSingleUpperWord(): void
|
58 | 63 | self::assertSame([], $kwsDiff);
|
59 | 64 | }
|
60 | 65 |
|
61 |
| - #[DoesNotPerformAssertions] |
62 |
| - public function testThereAreNoRegressions(): void |
| 66 | + /** @param list<Token> $expectedTokens */ |
| 67 | + public static function assertEqualsTokens(array $expectedTokens, Cursor $cursor): void |
63 | 68 | {
|
64 |
| - (new Tokenizer())->tokenize('*/'); |
| 69 | + $tokens = []; |
| 70 | + |
| 71 | + $cursor = $cursor->subCursor(); |
| 72 | + |
| 73 | + while ($token = $cursor->next()) { |
| 74 | + $tokens[] = $token; |
| 75 | + } |
| 76 | + |
| 77 | + if (serialize($tokens) === serialize($expectedTokens)) { // optimize self::assertEquals() for large inputs |
| 78 | + self::assertTrue(true); |
| 79 | + } else { |
| 80 | + self::assertEquals($expectedTokens, $tokens); |
| 81 | + } |
| 82 | + } |
| 83 | + |
| 84 | + /** @param list<Token> $expectedTokens */ |
| 85 | + #[DataProvider('tokenizeData')] |
| 86 | + #[DataProvider('tokenizeLongConcatData')] |
| 87 | + public function testTokenize(array $expectedTokens, string $sql): void |
| 88 | + { |
| 89 | + self::assertEqualsTokens($expectedTokens, (new Tokenizer())->tokenize($sql)); |
| 90 | + } |
| 91 | + |
| 92 | + /** @return Generator<mixed[]> */ |
| 93 | + public static function tokenizeData(): Generator |
| 94 | + { |
| 95 | + yield 'empty' => [ |
| 96 | + [], |
| 97 | + '', |
| 98 | + ]; |
| 99 | + |
| 100 | + yield 'basic' => [ |
| 101 | + [ |
| 102 | + new Token(Token::TOKEN_TYPE_RESERVED_TOPLEVEL, 'select'), |
| 103 | + new Token(Token::TOKEN_TYPE_WHITESPACE, ' '), |
| 104 | + new Token(Token::TOKEN_TYPE_NUMBER, '1'), |
| 105 | + ], |
| 106 | + 'select 1', |
| 107 | + ]; |
| 108 | + |
| 109 | + yield 'there are no regressions' => [ |
| 110 | + [ |
| 111 | + new Token(Token::TOKEN_TYPE_BOUNDARY, '*'), |
| 112 | + new Token(Token::TOKEN_TYPE_BOUNDARY, '/'), |
| 113 | + ], |
| 114 | + '*/', |
| 115 | + ]; |
| 116 | + |
| 117 | + yield 'unclosed quoted string' => [ |
| 118 | + [ |
| 119 | + new Token(Token::TOKEN_TYPE_QUOTE, '\'foo...'), |
| 120 | + ], |
| 121 | + '\'foo...', |
| 122 | + ]; |
| 123 | + |
| 124 | + yield 'unclosed block comment' => [ |
| 125 | + [ |
| 126 | + new Token(Token::TOKEN_TYPE_BLOCK_COMMENT, '/* foo...'), |
| 127 | + ], |
| 128 | + '/* foo...', |
| 129 | + ]; |
| 130 | + } |
| 131 | + |
| 132 | + /** @return Generator<mixed[]> */ |
| 133 | + public static function tokenizeLongConcatData(): Generator |
| 134 | + { |
| 135 | + $count = 2_000; |
| 136 | + |
| 137 | + $sqlParts = []; |
| 138 | + for ($i = 0; $i < $count; $i++) { |
| 139 | + $sqlParts[] = 'cast(\'foo' . $i . '\' as blob)'; |
| 140 | + } |
| 141 | + |
| 142 | + $concat = 'concat(' . implode(', ', $sqlParts) . ')'; |
| 143 | + $sql = 'select iif(' . $concat . ' = ' . $concat . ', 10, 20) x'; |
| 144 | + |
| 145 | + $expectedTokens = [ |
| 146 | + new Token(Token::TOKEN_TYPE_RESERVED_TOPLEVEL, 'select'), |
| 147 | + new Token(Token::TOKEN_TYPE_WHITESPACE, ' '), |
| 148 | + new Token(Token::TOKEN_TYPE_WORD, 'iif'), |
| 149 | + new Token(Token::TOKEN_TYPE_BOUNDARY, '('), |
| 150 | + ]; |
| 151 | + |
| 152 | + for ($j = 0; $j < 2; $j++) { |
| 153 | + if ($j !== 0) { |
| 154 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_WHITESPACE, ' '); |
| 155 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_BOUNDARY, '='); |
| 156 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_WHITESPACE, ' '); |
| 157 | + } |
| 158 | + |
| 159 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_RESERVED, 'concat'); |
| 160 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_BOUNDARY, '('); |
| 161 | + |
| 162 | + for ($i = 0; $i < $count; $i++) { |
| 163 | + if ($i !== 0) { |
| 164 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_BOUNDARY, ','); |
| 165 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_WHITESPACE, ' '); |
| 166 | + } |
| 167 | + |
| 168 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_RESERVED, 'cast'); |
| 169 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_BOUNDARY, '('); |
| 170 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_QUOTE, '\'foo' . $i . '\''); |
| 171 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_WHITESPACE, ' '); |
| 172 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_RESERVED, 'as'); |
| 173 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_WHITESPACE, ' '); |
| 174 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_WORD, 'blob'); |
| 175 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_BOUNDARY, ')'); |
| 176 | + } |
| 177 | + |
| 178 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_BOUNDARY, ')'); |
| 179 | + } |
| 180 | + |
| 181 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_BOUNDARY, ','); |
| 182 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_WHITESPACE, ' '); |
| 183 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_NUMBER, '10'); |
| 184 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_BOUNDARY, ','); |
| 185 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_WHITESPACE, ' '); |
| 186 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_NUMBER, '20'); |
| 187 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_BOUNDARY, ')'); |
| 188 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_WHITESPACE, ' '); |
| 189 | + $expectedTokens[] = new Token(Token::TOKEN_TYPE_WORD, 'x'); |
| 190 | + |
| 191 | + yield 'long concat' => [$expectedTokens, $sql]; |
65 | 192 | }
|
66 | 193 | }
|
0 commit comments