3RNN/Lib/site-packages/pygments/lexers/nix.py
2024-05-26 19:49:15 +02:00

145 lines
4.3 KiB
Python

"""
pygments.lexers.nix
~~~~~~~~~~~~~~~~~~~
Lexers for the NixOS Nix language.
:copyright: Copyright 2006-2024 by the Pygments team, see AUTHORS.
:license: BSD, see LICENSE for details.
"""
import re
from pygments.lexer import RegexLexer, include
from pygments.token import Text, Comment, Operator, Keyword, Name, String, \
Number, Punctuation, Literal
__all__ = ['NixLexer']
class NixLexer(RegexLexer):
"""
For the Nix language.
"""
name = 'Nix'
url = 'http://nixos.org/nix/'
aliases = ['nixos', 'nix']
filenames = ['*.nix']
mimetypes = ['text/x-nix']
version_added = '2.0'
keywords = ['rec', 'with', 'let', 'in', 'inherit', 'assert', 'if',
'else', 'then', '...']
builtins = ['import', 'abort', 'baseNameOf', 'dirOf', 'isNull', 'builtins',
'map', 'removeAttrs', 'throw', 'toString', 'derivation']
operators = ['++', '+', '?', '.', '!', '//', '==', '/',
'!=', '&&', '||', '->', '=', '<', '>', '*', '-']
punctuations = ["(", ")", "[", "]", ";", "{", "}", ":", ",", "@"]
tokens = {
'root': [
# comments starting with #
(r'#.*$', Comment.Single),
# multiline comments
(r'/\*', Comment.Multiline, 'comment'),
# whitespace
(r'\s+', Text),
# keywords
('({})'.format('|'.join(re.escape(entry) + '\\b' for entry in keywords)), Keyword),
# highlight the builtins
('({})'.format('|'.join(re.escape(entry) + '\\b' for entry in builtins)),
Name.Builtin),
(r'\b(true|false|null)\b', Name.Constant),
# floats
(r'-?(\d+\.\d*|\.\d+)([eE][-+]?\d+)?', Number.Float),
# integers
(r'-?[0-9]+', Number.Integer),
# paths
(r'[\w.+-]*(\/[\w.+-]+)+', Literal),
(r'~(\/[\w.+-]+)+', Literal),
(r'\<[\w.+-]+(\/[\w.+-]+)*\>', Literal),
# operators
('({})'.format('|'.join(re.escape(entry) for entry in operators)),
Operator),
# word operators
(r'\b(or|and)\b', Operator.Word),
(r'\{', Punctuation, 'block'),
# punctuations
('({})'.format('|'.join(re.escape(entry) for entry in punctuations)), Punctuation),
# strings
(r'"', String.Double, 'doublequote'),
(r"''", String.Multiline, 'multiline'),
# urls
(r'[a-zA-Z][a-zA-Z0-9\+\-\.]*\:[\w%/?:@&=+$,\\.!~*\'-]+', Literal),
# names of variables
(r'[\w-]+(?=\s*=)', String.Symbol),
(r'[a-zA-Z_][\w\'-]*', Text),
(r"\$\{", String.Interpol, 'antiquote'),
],
'comment': [
(r'[^/*]+', Comment.Multiline),
(r'/\*', Comment.Multiline, '#push'),
(r'\*/', Comment.Multiline, '#pop'),
(r'[*/]', Comment.Multiline),
],
'multiline': [
(r"''(\$|'|\\n|\\r|\\t|\\)", String.Escape),
(r"''", String.Multiline, '#pop'),
(r'\$\{', String.Interpol, 'antiquote'),
(r"[^'\$]+", String.Multiline),
(r"\$[^\{']", String.Multiline),
(r"'[^']", String.Multiline),
(r"\$(?=')", String.Multiline),
],
'doublequote': [
(r'\\(\\|"|\$|n)', String.Escape),
(r'"', String.Double, '#pop'),
(r'\$\{', String.Interpol, 'antiquote'),
(r'[^"\\\$]+', String.Double),
(r'\$[^\{"]', String.Double),
(r'\$(?=")', String.Double),
(r'\\', String.Double),
],
'antiquote': [
(r"\}", String.Interpol, '#pop'),
# TODO: we should probably escape also here ''${ \${
(r"\$\{", String.Interpol, '#push'),
include('root'),
],
'block': [
(r"\}", Punctuation, '#pop'),
include('root'),
],
}
def analyse_text(text):
rv = 0.0
# TODO: let/in
if re.search(r'import.+?<[^>]+>', text):
rv += 0.4
if re.search(r'mkDerivation\s+(\(|\{|rec)', text):
rv += 0.4
if re.search(r'=\s+mkIf\s+', text):
rv += 0.4
if re.search(r'\{[a-zA-Z,\s]+\}:', text):
rv += 0.1
return rv