PageRenderTime 71ms CodeModel.GetById 26ms RepoModel.GetById 1ms app.codeStats 0ms

/pygments/pygments/lexers/sql.py

https://bitbucket.org/tcorll/butlertest
Python | 559 lines | 490 code | 13 blank | 56 comment | 8 complexity | dfaa59192cd6b86c9f385127417162f7 MD5 | raw file
Possible License(s): Apache-2.0, GPL-2.0, LGPL-3.0, BSD-2-Clause
  1. # -*- coding: utf-8 -*-
  2. """
  3. pygments.lexers.sql
  4. ~~~~~~~~~~~~~~~~~~~
  5. Lexers for various SQL dialects and related interactive sessions.
  6. Postgres specific lexers:
  7. `PostgresLexer`
  8. A SQL lexer for the PostgreSQL dialect. Differences w.r.t. the SQL
  9. lexer are:
  10. - keywords and data types list parsed from the PG docs (run the
  11. `_postgres_builtins` module to update them);
  12. - Content of $-strings parsed using a specific lexer, e.g. the content
  13. of a PL/Python function is parsed using the Python lexer;
  14. - parse PG specific constructs: E-strings, $-strings, U&-strings,
  15. different operators and punctuation.
  16. `PlPgsqlLexer`
  17. A lexer for the PL/pgSQL language. Adds a few specific construct on
  18. top of the PG SQL lexer (such as <<label>>).
  19. `PostgresConsoleLexer`
  20. A lexer to highlight an interactive psql session:
  21. - identifies the prompt and does its best to detect the end of command
  22. in multiline statement where not all the lines are prefixed by a
  23. prompt, telling them apart from the output;
  24. - highlights errors in the output and notification levels;
  25. - handles psql backslash commands.
  26. The ``tests/examplefiles`` contains a few test files with data to be
  27. parsed by these lexers.
  28. :copyright: Copyright 2006-2012 by the Pygments team, see AUTHORS.
  29. :license: BSD, see LICENSE for details.
  30. """
  31. import re
  32. from pygments.lexer import Lexer, RegexLexer, do_insertions, bygroups
  33. from pygments.token import Punctuation, \
  34. Text, Comment, Operator, Keyword, Name, String, Number, Generic
  35. from pygments.lexers import get_lexer_by_name, ClassNotFound
  36. from pygments.lexers._postgres_builtins import KEYWORDS, DATATYPES, \
  37. PSEUDO_TYPES, PLPGSQL_KEYWORDS
  38. __all__ = ['PostgresLexer', 'PlPgsqlLexer', 'PostgresConsoleLexer',
  39. 'SqlLexer', 'MySqlLexer', 'SqliteConsoleLexer']
  40. line_re = re.compile('.*?\n')
  41. language_re = re.compile(r"\s+LANGUAGE\s+'?(\w+)'?", re.IGNORECASE)
  42. def language_callback(lexer, match):
  43. """Parse the content of a $-string using a lexer
  44. The lexer is chosen looking for a nearby LANGUAGE.
  45. """
  46. l = None
  47. m = language_re.match(lexer.text[match.end():match.end()+100])
  48. if m is not None:
  49. l = lexer._get_lexer(m.group(1))
  50. else:
  51. m = list(language_re.finditer(
  52. lexer.text[max(0, match.start()-100):match.start()]))
  53. if m:
  54. l = lexer._get_lexer(m[-1].group(1))
  55. if l:
  56. yield (match.start(1), String, match.group(1))
  57. for x in l.get_tokens_unprocessed(match.group(2)):
  58. yield x
  59. yield (match.start(3), String, match.group(3))
  60. else:
  61. yield (match.start(), String, match.group())
  62. class PostgresBase(object):
  63. """Base class for Postgres-related lexers.
  64. This is implemented as a mixin to avoid the Lexer metaclass kicking in.
  65. this way the different lexer don't have a common Lexer ancestor. If they
  66. had, _tokens could be created on this ancestor and not updated for the
  67. other classes, resulting e.g. in PL/pgSQL parsed as SQL. This shortcoming
  68. seem to suggest that regexp lexers are not really subclassable.
  69. """
  70. def get_tokens_unprocessed(self, text, *args):
  71. # Have a copy of the entire text to be used by `language_callback`.
  72. self.text = text
  73. for x in super(PostgresBase, self).get_tokens_unprocessed(
  74. text, *args):
  75. yield x
  76. def _get_lexer(self, lang):
  77. if lang.lower() == 'sql':
  78. return get_lexer_by_name('postgresql', **self.options)
  79. tries = [ lang ]
  80. if lang.startswith('pl'):
  81. tries.append(lang[2:])
  82. if lang.endswith('u'):
  83. tries.append(lang[:-1])
  84. if lang.startswith('pl') and lang.endswith('u'):
  85. tries.append(lang[2:-1])
  86. for l in tries:
  87. try:
  88. return get_lexer_by_name(l, **self.options)
  89. except ClassNotFound:
  90. pass
  91. else:
  92. # TODO: better logging
  93. # print >>sys.stderr, "language not found:", lang
  94. return None
  95. class PostgresLexer(PostgresBase, RegexLexer):
  96. """
  97. Lexer for the PostgreSQL dialect of SQL.
  98. *New in Pygments 1.5.*
  99. """
  100. name = 'PostgreSQL SQL dialect'
  101. aliases = ['postgresql', 'postgres']
  102. mimetypes = ['text/x-postgresql']
  103. flags = re.IGNORECASE
  104. tokens = {
  105. 'root': [
  106. (r'\s+', Text),
  107. (r'--.*?\n', Comment.Single),
  108. (r'/\*', Comment.Multiline, 'multiline-comments'),
  109. (r'(' + '|'.join([s.replace(" ", "\s+")
  110. for s in DATATYPES + PSEUDO_TYPES])
  111. + r')\b', Name.Builtin),
  112. (r'(' + '|'.join(KEYWORDS) + r')\b', Keyword),
  113. (r'[+*/<>=~!@#%^&|`?-]+', Operator),
  114. (r'::', Operator), # cast
  115. (r'\$\d+', Name.Variable),
  116. (r'([0-9]*\.[0-9]*|[0-9]+)(e[+-]?[0-9]+)?', Number.Float),
  117. (r'[0-9]+', Number.Integer),
  118. (r"(E|U&)?'(''|[^'])*'", String.Single),
  119. (r'(U&)?"(""|[^"])*"', String.Name), # quoted identifier
  120. (r'(?s)(\$[^\$]*\$)(.*?)(\1)', language_callback),
  121. (r'[a-zA-Z_][a-zA-Z0-9_]*', Name),
  122. # psql variable in SQL
  123. (r""":(['"]?)[a-z][a-z0-9_]*\b\1""", Name.Variable),
  124. (r'[;:()\[\]\{\},\.]', Punctuation),
  125. ],
  126. 'multiline-comments': [
  127. (r'/\*', Comment.Multiline, 'multiline-comments'),
  128. (r'\*/', Comment.Multiline, '#pop'),
  129. (r'[^/\*]+', Comment.Multiline),
  130. (r'[/*]', Comment.Multiline)
  131. ],
  132. }
  133. class PlPgsqlLexer(PostgresBase, RegexLexer):
  134. """
  135. Handle the extra syntax in Pl/pgSQL language.
  136. *New in Pygments 1.5.*
  137. """
  138. name = 'PL/pgSQL'
  139. aliases = ['plpgsql']
  140. mimetypes = ['text/x-plpgsql']
  141. flags = re.IGNORECASE
  142. tokens = dict((k, l[:]) for (k, l) in PostgresLexer.tokens.iteritems())
  143. # extend the keywords list
  144. for i, pattern in enumerate(tokens['root']):
  145. if pattern[1] == Keyword:
  146. tokens['root'][i] = (
  147. r'(' + '|'.join(KEYWORDS + PLPGSQL_KEYWORDS) + r')\b',
  148. Keyword)
  149. del i
  150. break
  151. else:
  152. assert 0, "SQL keywords not found"
  153. # Add specific PL/pgSQL rules (before the SQL ones)
  154. tokens['root'][:0] = [
  155. (r'\%[a-z][a-z0-9_]*\b', Name.Builtin), # actually, a datatype
  156. (r':=', Operator),
  157. (r'\<\<[a-z][a-z0-9_]*\>\>', Name.Label),
  158. (r'\#[a-z][a-z0-9_]*\b', Keyword.Pseudo), # #variable_conflict
  159. ]
  160. class PsqlRegexLexer(PostgresBase, RegexLexer):
  161. """
  162. Extend the PostgresLexer adding support specific for psql commands.
  163. This is not a complete psql lexer yet as it lacks prompt support
  164. and output rendering.
  165. """
  166. name = 'PostgreSQL console - regexp based lexer'
  167. aliases = [] # not public
  168. flags = re.IGNORECASE
  169. tokens = dict((k, l[:]) for (k, l) in PostgresLexer.tokens.iteritems())
  170. tokens['root'].append(
  171. (r'\\[^\s]+', Keyword.Pseudo, 'psql-command'))
  172. tokens['psql-command'] = [
  173. (r'\n', Text, 'root'),
  174. (r'\s+', Text),
  175. (r'\\[^\s]+', Keyword.Pseudo),
  176. (r""":(['"]?)[a-z][a-z0-9_]*\b\1""", Name.Variable),
  177. (r"'(''|[^'])*'", String.Single),
  178. (r"`([^`])*`", String.Backtick),
  179. (r"[^\s]+", String.Symbol),
  180. ]
  181. re_prompt = re.compile(r'^(\S.*?)??[=\-\(\$\'\"][#>]')
  182. re_psql_command = re.compile(r'\s*\\')
  183. re_end_command = re.compile(r';\s*(--.*?)?$')
  184. re_psql_command = re.compile(r'(\s*)(\\.+?)(\s+)$')
  185. re_error = re.compile(r'(ERROR|FATAL):')
  186. re_message = re.compile(
  187. r'((?:DEBUG|INFO|NOTICE|WARNING|ERROR|'
  188. r'FATAL|HINT|DETAIL|CONTEXT|LINE [0-9]+):)(.*?\n)')
  189. class lookahead(object):
  190. """Wrap an iterator and allow pushing back an item."""
  191. def __init__(self, x):
  192. self.iter = iter(x)
  193. self._nextitem = None
  194. def __iter__(self):
  195. return self
  196. def send(self, i):
  197. self._nextitem = i
  198. return i
  199. def next(self):
  200. if self._nextitem is not None:
  201. ni = self._nextitem
  202. self._nextitem = None
  203. return ni
  204. return self.iter.next()
  205. class PostgresConsoleLexer(Lexer):
  206. """
  207. Lexer for psql sessions.
  208. *New in Pygments 1.5.*
  209. """
  210. name = 'PostgreSQL console (psql)'
  211. aliases = ['psql', 'postgresql-console', 'postgres-console']
  212. mimetypes = ['text/x-postgresql-psql']
  213. def get_tokens_unprocessed(self, data):
  214. sql = PsqlRegexLexer(**self.options)
  215. lines = lookahead(line_re.findall(data))
  216. # prompt-output cycle
  217. while 1:
  218. # consume the lines of the command: start with an optional prompt
  219. # and continue until the end of command is detected
  220. curcode = ''
  221. insertions = []
  222. while 1:
  223. try:
  224. line = lines.next()
  225. except StopIteration:
  226. # allow the emission of partially collected items
  227. # the repl loop will be broken below
  228. break
  229. # Identify a shell prompt in case of psql commandline example
  230. if line.startswith('$') and not curcode:
  231. lexer = get_lexer_by_name('console', **self.options)
  232. for x in lexer.get_tokens_unprocessed(line):
  233. yield x
  234. break
  235. # Identify a psql prompt
  236. mprompt = re_prompt.match(line)
  237. if mprompt is not None:
  238. insertions.append((len(curcode),
  239. [(0, Generic.Prompt, mprompt.group())]))
  240. curcode += line[len(mprompt.group()):]
  241. else:
  242. curcode += line
  243. # Check if this is the end of the command
  244. # TODO: better handle multiline comments at the end with
  245. # a lexer with an external state?
  246. if re_psql_command.match(curcode) \
  247. or re_end_command.search(curcode):
  248. break
  249. # Emit the combined stream of command and prompt(s)
  250. for item in do_insertions(insertions,
  251. sql.get_tokens_unprocessed(curcode)):
  252. yield item
  253. # Emit the output lines
  254. out_token = Generic.Output
  255. while 1:
  256. line = lines.next()
  257. mprompt = re_prompt.match(line)
  258. if mprompt is not None:
  259. # push the line back to have it processed by the prompt
  260. lines.send(line)
  261. break
  262. mmsg = re_message.match(line)
  263. if mmsg is not None:
  264. if mmsg.group(1).startswith("ERROR") \
  265. or mmsg.group(1).startswith("FATAL"):
  266. out_token = Generic.Error
  267. yield (mmsg.start(1), Generic.Strong, mmsg.group(1))
  268. yield (mmsg.start(2), out_token, mmsg.group(2))
  269. else:
  270. yield (0, out_token, line)
  271. class SqlLexer(RegexLexer):
  272. """
  273. Lexer for Structured Query Language. Currently, this lexer does
  274. not recognize any special syntax except ANSI SQL.
  275. """
  276. name = 'SQL'
  277. aliases = ['sql']
  278. filenames = ['*.sql']
  279. mimetypes = ['text/x-sql']
  280. flags = re.IGNORECASE
  281. tokens = {
  282. 'root': [
  283. (r'\s+', Text),
  284. (r'--.*?\n', Comment.Single),
  285. (r'/\*', Comment.Multiline, 'multiline-comments'),
  286. (r'(ABORT|ABS|ABSOLUTE|ACCESS|ADA|ADD|ADMIN|AFTER|AGGREGATE|'
  287. r'ALIAS|ALL|ALLOCATE|ALTER|ANALYSE|ANALYZE|AND|ANY|ARE|AS|'
  288. r'ASC|ASENSITIVE|ASSERTION|ASSIGNMENT|ASYMMETRIC|AT|ATOMIC|'
  289. r'AUTHORIZATION|AVG|BACKWARD|BEFORE|BEGIN|BETWEEN|BITVAR|'
  290. r'BIT_LENGTH|BOTH|BREADTH|BY|C|CACHE|CALL|CALLED|CARDINALITY|'
  291. r'CASCADE|CASCADED|CASE|CAST|CATALOG|CATALOG_NAME|CHAIN|'
  292. r'CHARACTERISTICS|CHARACTER_LENGTH|CHARACTER_SET_CATALOG|'
  293. r'CHARACTER_SET_NAME|CHARACTER_SET_SCHEMA|CHAR_LENGTH|CHECK|'
  294. r'CHECKED|CHECKPOINT|CLASS|CLASS_ORIGIN|CLOB|CLOSE|CLUSTER|'
  295. r'COALSECE|COBOL|COLLATE|COLLATION|COLLATION_CATALOG|'
  296. r'COLLATION_NAME|COLLATION_SCHEMA|COLUMN|COLUMN_NAME|'
  297. r'COMMAND_FUNCTION|COMMAND_FUNCTION_CODE|COMMENT|COMMIT|'
  298. r'COMMITTED|COMPLETION|CONDITION_NUMBER|CONNECT|CONNECTION|'
  299. r'CONNECTION_NAME|CONSTRAINT|CONSTRAINTS|CONSTRAINT_CATALOG|'
  300. r'CONSTRAINT_NAME|CONSTRAINT_SCHEMA|CONSTRUCTOR|CONTAINS|'
  301. r'CONTINUE|CONVERSION|CONVERT|COPY|CORRESPONTING|COUNT|'
  302. r'CREATE|CREATEDB|CREATEUSER|CROSS|CUBE|CURRENT|CURRENT_DATE|'
  303. r'CURRENT_PATH|CURRENT_ROLE|CURRENT_TIME|CURRENT_TIMESTAMP|'
  304. r'CURRENT_USER|CURSOR|CURSOR_NAME|CYCLE|DATA|DATABASE|'
  305. r'DATETIME_INTERVAL_CODE|DATETIME_INTERVAL_PRECISION|DAY|'
  306. r'DEALLOCATE|DECLARE|DEFAULT|DEFAULTS|DEFERRABLE|DEFERRED|'
  307. r'DEFINED|DEFINER|DELETE|DELIMITER|DELIMITERS|DEREF|DESC|'
  308. r'DESCRIBE|DESCRIPTOR|DESTROY|DESTRUCTOR|DETERMINISTIC|'
  309. r'DIAGNOSTICS|DICTIONARY|DISCONNECT|DISPATCH|DISTINCT|DO|'
  310. r'DOMAIN|DROP|DYNAMIC|DYNAMIC_FUNCTION|DYNAMIC_FUNCTION_CODE|'
  311. r'EACH|ELSE|ENCODING|ENCRYPTED|END|END-EXEC|EQUALS|ESCAPE|EVERY|'
  312. r'EXCEPT|ESCEPTION|EXCLUDING|EXCLUSIVE|EXEC|EXECUTE|EXISTING|'
  313. r'EXISTS|EXPLAIN|EXTERNAL|EXTRACT|FALSE|FETCH|FINAL|FIRST|FOR|'
  314. r'FORCE|FOREIGN|FORTRAN|FORWARD|FOUND|FREE|FREEZE|FROM|FULL|'
  315. r'FUNCTION|G|GENERAL|GENERATED|GET|GLOBAL|GO|GOTO|GRANT|GRANTED|'
  316. r'GROUP|GROUPING|HANDLER|HAVING|HIERARCHY|HOLD|HOST|IDENTITY|'
  317. r'IGNORE|ILIKE|IMMEDIATE|IMMUTABLE|IMPLEMENTATION|IMPLICIT|IN|'
  318. r'INCLUDING|INCREMENT|INDEX|INDITCATOR|INFIX|INHERITS|INITIALIZE|'
  319. r'INITIALLY|INNER|INOUT|INPUT|INSENSITIVE|INSERT|INSTANTIABLE|'
  320. r'INSTEAD|INTERSECT|INTO|INVOKER|IS|ISNULL|ISOLATION|ITERATE|JOIN|'
  321. r'KEY|KEY_MEMBER|KEY_TYPE|LANCOMPILER|LANGUAGE|LARGE|LAST|'
  322. r'LATERAL|LEADING|LEFT|LENGTH|LESS|LEVEL|LIKE|LIMIT|LISTEN|LOAD|'
  323. r'LOCAL|LOCALTIME|LOCALTIMESTAMP|LOCATION|LOCATOR|LOCK|LOWER|'
  324. r'MAP|MATCH|MAX|MAXVALUE|MESSAGE_LENGTH|MESSAGE_OCTET_LENGTH|'
  325. r'MESSAGE_TEXT|METHOD|MIN|MINUTE|MINVALUE|MOD|MODE|MODIFIES|'
  326. r'MODIFY|MONTH|MORE|MOVE|MUMPS|NAMES|NATIONAL|NATURAL|NCHAR|'
  327. r'NCLOB|NEW|NEXT|NO|NOCREATEDB|NOCREATEUSER|NONE|NOT|NOTHING|'
  328. r'NOTIFY|NOTNULL|NULL|NULLABLE|NULLIF|OBJECT|OCTET_LENGTH|OF|OFF|'
  329. r'OFFSET|OIDS|OLD|ON|ONLY|OPEN|OPERATION|OPERATOR|OPTION|OPTIONS|'
  330. r'OR|ORDER|ORDINALITY|OUT|OUTER|OUTPUT|OVERLAPS|OVERLAY|OVERRIDING|'
  331. r'OWNER|PAD|PARAMETER|PARAMETERS|PARAMETER_MODE|PARAMATER_NAME|'
  332. r'PARAMATER_ORDINAL_POSITION|PARAMETER_SPECIFIC_CATALOG|'
  333. r'PARAMETER_SPECIFIC_NAME|PARAMATER_SPECIFIC_SCHEMA|PARTIAL|'
  334. r'PASCAL|PENDANT|PLACING|PLI|POSITION|POSTFIX|PRECISION|PREFIX|'
  335. r'PREORDER|PREPARE|PRESERVE|PRIMARY|PRIOR|PRIVILEGES|PROCEDURAL|'
  336. r'PROCEDURE|PUBLIC|READ|READS|RECHECK|RECURSIVE|REF|REFERENCES|'
  337. r'REFERENCING|REINDEX|RELATIVE|RENAME|REPEATABLE|REPLACE|RESET|'
  338. r'RESTART|RESTRICT|RESULT|RETURN|RETURNED_LENGTH|'
  339. r'RETURNED_OCTET_LENGTH|RETURNED_SQLSTATE|RETURNS|REVOKE|RIGHT|'
  340. r'ROLE|ROLLBACK|ROLLUP|ROUTINE|ROUTINE_CATALOG|ROUTINE_NAME|'
  341. r'ROUTINE_SCHEMA|ROW|ROWS|ROW_COUNT|RULE|SAVE_POINT|SCALE|SCHEMA|'
  342. r'SCHEMA_NAME|SCOPE|SCROLL|SEARCH|SECOND|SECURITY|SELECT|SELF|'
  343. r'SENSITIVE|SERIALIZABLE|SERVER_NAME|SESSION|SESSION_USER|SET|'
  344. r'SETOF|SETS|SHARE|SHOW|SIMILAR|SIMPLE|SIZE|SOME|SOURCE|SPACE|'
  345. r'SPECIFIC|SPECIFICTYPE|SPECIFIC_NAME|SQL|SQLCODE|SQLERROR|'
  346. r'SQLEXCEPTION|SQLSTATE|SQLWARNINIG|STABLE|START|STATE|STATEMENT|'
  347. r'STATIC|STATISTICS|STDIN|STDOUT|STORAGE|STRICT|STRUCTURE|STYPE|'
  348. r'SUBCLASS_ORIGIN|SUBLIST|SUBSTRING|SUM|SYMMETRIC|SYSID|SYSTEM|'
  349. r'SYSTEM_USER|TABLE|TABLE_NAME| TEMP|TEMPLATE|TEMPORARY|TERMINATE|'
  350. r'THAN|THEN|TIMESTAMP|TIMEZONE_HOUR|TIMEZONE_MINUTE|TO|TOAST|'
  351. r'TRAILING|TRANSATION|TRANSACTIONS_COMMITTED|'
  352. r'TRANSACTIONS_ROLLED_BACK|TRANSATION_ACTIVE|TRANSFORM|'
  353. r'TRANSFORMS|TRANSLATE|TRANSLATION|TREAT|TRIGGER|TRIGGER_CATALOG|'
  354. r'TRIGGER_NAME|TRIGGER_SCHEMA|TRIM|TRUE|TRUNCATE|TRUSTED|TYPE|'
  355. r'UNCOMMITTED|UNDER|UNENCRYPTED|UNION|UNIQUE|UNKNOWN|UNLISTEN|'
  356. r'UNNAMED|UNNEST|UNTIL|UPDATE|UPPER|USAGE|USER|'
  357. r'USER_DEFINED_TYPE_CATALOG|USER_DEFINED_TYPE_NAME|'
  358. r'USER_DEFINED_TYPE_SCHEMA|USING|VACUUM|VALID|VALIDATOR|VALUES|'
  359. r'VARIABLE|VERBOSE|VERSION|VIEW|VOLATILE|WHEN|WHENEVER|WHERE|'
  360. r'WITH|WITHOUT|WORK|WRITE|YEAR|ZONE)\b', Keyword),
  361. (r'(ARRAY|BIGINT|BINARY|BIT|BLOB|BOOLEAN|CHAR|CHARACTER|DATE|'
  362. r'DEC|DECIMAL|FLOAT|INT|INTEGER|INTERVAL|NUMBER|NUMERIC|REAL|'
  363. r'SERIAL|SMALLINT|VARCHAR|VARYING|INT8|SERIAL8|TEXT)\b',
  364. Name.Builtin),
  365. (r'[+*/<>=~!@#%^&|`?-]', Operator),
  366. (r'[0-9]+', Number.Integer),
  367. # TODO: Backslash escapes?
  368. (r"'(''|[^'])*'", String.Single),
  369. (r'"(""|[^"])*"', String.Symbol), # not a real string literal in ANSI SQL
  370. (r'[a-zA-Z_][a-zA-Z0-9_]*', Name),
  371. (r'[;:()\[\],\.]', Punctuation)
  372. ],
  373. 'multiline-comments': [
  374. (r'/\*', Comment.Multiline, 'multiline-comments'),
  375. (r'\*/', Comment.Multiline, '#pop'),
  376. (r'[^/\*]+', Comment.Multiline),
  377. (r'[/*]', Comment.Multiline)
  378. ]
  379. }
  380. class MySqlLexer(RegexLexer):
  381. """
  382. Special lexer for MySQL.
  383. """
  384. name = 'MySQL'
  385. aliases = ['mysql']
  386. mimetypes = ['text/x-mysql']
  387. flags = re.IGNORECASE
  388. tokens = {
  389. 'root': [
  390. (r'\s+', Text),
  391. (r'(#|--\s+).*?\n', Comment.Single),
  392. (r'/\*', Comment.Multiline, 'multiline-comments'),
  393. (r'[0-9]+', Number.Integer),
  394. (r'[0-9]*\.[0-9]+(e[+-][0-9]+)', Number.Float),
  395. # TODO: add backslash escapes
  396. (r"'(''|[^'])*'", String.Single),
  397. (r'"(""|[^"])*"', String.Double),
  398. (r"`(``|[^`])*`", String.Symbol),
  399. (r'[+*/<>=~!@#%^&|`?-]', Operator),
  400. (r'\b(tinyint|smallint|mediumint|int|integer|bigint|date|'
  401. r'datetime|time|bit|bool|tinytext|mediumtext|longtext|text|'
  402. r'tinyblob|mediumblob|longblob|blob|float|double|double\s+'
  403. r'precision|real|numeric|dec|decimal|timestamp|year|char|'
  404. r'varchar|varbinary|varcharacter|enum|set)(\b\s*)(\()?',
  405. bygroups(Keyword.Type, Text, Punctuation)),
  406. (r'\b(add|all|alter|analyze|and|as|asc|asensitive|before|between|'
  407. r'bigint|binary|blob|both|by|call|cascade|case|change|char|'
  408. r'character|check|collate|column|condition|constraint|continue|'
  409. r'convert|create|cross|current_date|current_time|'
  410. r'current_timestamp|current_user|cursor|database|databases|'
  411. r'day_hour|day_microsecond|day_minute|day_second|dec|decimal|'
  412. r'declare|default|delayed|delete|desc|describe|deterministic|'
  413. r'distinct|distinctrow|div|double|drop|dual|each|else|elseif|'
  414. r'enclosed|escaped|exists|exit|explain|fetch|float|float4|float8'
  415. r'|for|force|foreign|from|fulltext|grant|group|having|'
  416. r'high_priority|hour_microsecond|hour_minute|hour_second|if|'
  417. r'ignore|in|index|infile|inner|inout|insensitive|insert|int|'
  418. r'int1|int2|int3|int4|int8|integer|interval|into|is|iterate|'
  419. r'join|key|keys|kill|leading|leave|left|like|limit|lines|load|'
  420. r'localtime|localtimestamp|lock|long|loop|low_priority|match|'
  421. r'minute_microsecond|minute_second|mod|modifies|natural|'
  422. r'no_write_to_binlog|not|numeric|on|optimize|option|optionally|'
  423. r'or|order|out|outer|outfile|precision|primary|procedure|purge|'
  424. r'raid0|read|reads|real|references|regexp|release|rename|repeat|'
  425. r'replace|require|restrict|return|revoke|right|rlike|schema|'
  426. r'schemas|second_microsecond|select|sensitive|separator|set|'
  427. r'show|smallint|soname|spatial|specific|sql|sql_big_result|'
  428. r'sql_calc_found_rows|sql_small_result|sqlexception|sqlstate|'
  429. r'sqlwarning|ssl|starting|straight_join|table|terminated|then|'
  430. r'to|trailing|trigger|undo|union|unique|unlock|unsigned|update|'
  431. r'usage|use|using|utc_date|utc_time|utc_timestamp|values|'
  432. r'varying|when|where|while|with|write|x509|xor|year_month|'
  433. r'zerofill)\b', Keyword),
  434. # TODO: this list is not complete
  435. (r'\b(auto_increment|engine|charset|tables)\b', Keyword.Pseudo),
  436. (r'(true|false|null)', Name.Constant),
  437. (r'([a-zA-Z_][a-zA-Z0-9_]*)(\s*)(\()',
  438. bygroups(Name.Function, Text, Punctuation)),
  439. (r'[a-zA-Z_][a-zA-Z0-9_]*', Name),
  440. (r'@[A-Za-z0-9]*[._]*[A-Za-z0-9]*', Name.Variable),
  441. (r'[;:()\[\],\.]', Punctuation)
  442. ],
  443. 'multiline-comments': [
  444. (r'/\*', Comment.Multiline, 'multiline-comments'),
  445. (r'\*/', Comment.Multiline, '#pop'),
  446. (r'[^/\*]+', Comment.Multiline),
  447. (r'[/*]', Comment.Multiline)
  448. ]
  449. }
  450. class SqliteConsoleLexer(Lexer):
  451. """
  452. Lexer for example sessions using sqlite3.
  453. *New in Pygments 0.11.*
  454. """
  455. name = 'sqlite3con'
  456. aliases = ['sqlite3']
  457. filenames = ['*.sqlite3-console']
  458. mimetypes = ['text/x-sqlite3-console']
  459. def get_tokens_unprocessed(self, data):
  460. sql = SqlLexer(**self.options)
  461. curcode = ''
  462. insertions = []
  463. for match in line_re.finditer(data):
  464. line = match.group()
  465. if line.startswith('sqlite> ') or line.startswith(' ...> '):
  466. insertions.append((len(curcode),
  467. [(0, Generic.Prompt, line[:8])]))
  468. curcode += line[8:]
  469. else:
  470. if curcode:
  471. for item in do_insertions(insertions,
  472. sql.get_tokens_unprocessed(curcode)):
  473. yield item
  474. curcode = ''
  475. insertions = []
  476. if line.startswith('SQL error: '):
  477. yield (match.start(), Generic.Traceback, line)
  478. else:
  479. yield (match.start(), Generic.Output, line)
  480. if curcode:
  481. for item in do_insertions(insertions,
  482. sql.get_tokens_unprocessed(curcode)):
  483. yield item