Uploaded Test files
This commit is contained in:
parent
f584ad9d97
commit
2e81cb7d99
16627 changed files with 2065359 additions and 102444 deletions
263
venv/Lib/site-packages/pygments/lexers/tnt.py
Normal file
263
venv/Lib/site-packages/pygments/lexers/tnt.py
Normal file
|
@ -0,0 +1,263 @@
|
|||
# -*- coding: utf-8 -*-
|
||||
"""
|
||||
pygments.lexers.tnt
|
||||
~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
Lexer for Typographic Number Theory.
|
||||
|
||||
:copyright: Copyright 2006-2020 by the Pygments team, see AUTHORS.
|
||||
:license: BSD, see LICENSE for details.
|
||||
"""
|
||||
|
||||
import re
|
||||
|
||||
from pygments.lexer import Lexer
|
||||
from pygments.token import Text, Comment, Operator, Keyword, Name, Number, \
|
||||
Punctuation, Error
|
||||
|
||||
__all__ = ['TNTLexer']
|
||||
|
||||
|
||||
class TNTLexer(Lexer):
|
||||
"""
|
||||
Lexer for Typographic Number Theory, as described in the book
|
||||
Gödel, Escher, Bach, by Douglas R. Hofstadter,
|
||||
or as summarized here:
|
||||
https://github.com/Kenny2github/language-tnt/blob/master/README.md#summary-of-tnt
|
||||
|
||||
.. versionadded:: 2.7
|
||||
"""
|
||||
|
||||
name = 'Typographic Number Theory'
|
||||
aliases = ['tnt']
|
||||
filenames = ['*.tnt']
|
||||
|
||||
cur = []
|
||||
|
||||
LOGIC = set('⊃→]&∧^|∨Vv')
|
||||
OPERATORS = set('+.⋅*')
|
||||
VARIABLES = set('abcde')
|
||||
PRIMES = set("'′")
|
||||
NEGATORS = set('~!')
|
||||
QUANTIFIERS = set('AE∀∃')
|
||||
NUMBERS = set('0123456789')
|
||||
WHITESPACE = set('\t \v\n')
|
||||
|
||||
RULES = re.compile('''(?xi)
|
||||
joining | separation | double-tilde | fantasy\\ rule
|
||||
| carry[- ]over(?:\\ of)?(?:\\ line)?\\ ([0-9]+) | detachment
|
||||
| contrapositive | De\\ Morgan | switcheroo
|
||||
| specification | generalization | interchange
|
||||
| existence | symmetry | transitivity
|
||||
| add\\ S | drop\\ S | induction
|
||||
| axiom\\ ([1-5]) | premise | push | pop
|
||||
''')
|
||||
LINENOS = re.compile(r'(?:[0-9]+)(?:(?:, ?|,? and )(?:[0-9]+))*')
|
||||
COMMENT = re.compile(r'\[[^\n\]]+\]')
|
||||
|
||||
def __init__(self, *args, **kwargs):
|
||||
Lexer.__init__(self, *args, **kwargs)
|
||||
self.cur = []
|
||||
|
||||
def whitespace(self, start, text, required=False):
|
||||
"""Tokenize whitespace."""
|
||||
end = start
|
||||
try:
|
||||
while text[end] in self.WHITESPACE:
|
||||
end += 1
|
||||
except IndexError:
|
||||
end = len(text)
|
||||
if required:
|
||||
assert end != start
|
||||
if end != start:
|
||||
self.cur.append((start, Text, text[start:end]))
|
||||
return end
|
||||
|
||||
def variable(self, start, text):
|
||||
"""Tokenize a variable."""
|
||||
assert text[start] in self.VARIABLES
|
||||
end = start+1
|
||||
while text[end] in self.PRIMES:
|
||||
end += 1
|
||||
self.cur.append((start, Name.Variable, text[start:end]))
|
||||
return end
|
||||
|
||||
def term(self, start, text):
|
||||
"""Tokenize a term."""
|
||||
if text[start] == 'S': # S...S(...) or S...0
|
||||
end = start+1
|
||||
while text[end] == 'S':
|
||||
end += 1
|
||||
self.cur.append((start, Number.Integer, text[start:end]))
|
||||
return self.term(end, text)
|
||||
if text[start] == '0': # the singleton 0
|
||||
self.cur.append((start, Number.Integer, text[start]))
|
||||
return start+1
|
||||
if text[start] in self.VARIABLES: # a''...
|
||||
return self.variable(start, text)
|
||||
if text[start] == '(': # (...+...)
|
||||
self.cur.append((start, Punctuation, text[start]))
|
||||
start = self.term(start+1, text)
|
||||
assert text[start] in self.OPERATORS
|
||||
self.cur.append((start, Operator, text[start]))
|
||||
start = self.term(start+1, text)
|
||||
assert text[start] == ')'
|
||||
self.cur.append((start, Punctuation, text[start]))
|
||||
return start+1
|
||||
raise AssertionError # no matches
|
||||
|
||||
def formula(self, start, text):
|
||||
"""Tokenize a formula."""
|
||||
if text[start] in self.NEGATORS: # ~<...>
|
||||
end = start+1
|
||||
while text[end] in self.NEGATORS:
|
||||
end += 1
|
||||
self.cur.append((start, Operator, text[start:end]))
|
||||
return self.formula(end, text)
|
||||
if text[start] in self.QUANTIFIERS: # Aa:<...>
|
||||
self.cur.append((start, Keyword.Declaration, text[start]))
|
||||
start = self.variable(start+1, text)
|
||||
assert text[start] == ':'
|
||||
self.cur.append((start, Punctuation, text[start]))
|
||||
return self.formula(start+1, text)
|
||||
if text[start] == '<': # <...&...>
|
||||
self.cur.append((start, Punctuation, text[start]))
|
||||
start = self.formula(start+1, text)
|
||||
assert text[start] in self.LOGIC
|
||||
self.cur.append((start, Operator, text[start]))
|
||||
start = self.formula(start+1, text)
|
||||
assert text[start] == '>'
|
||||
self.cur.append((start, Punctuation, text[start]))
|
||||
return start+1
|
||||
# ...=...
|
||||
start = self.term(start, text)
|
||||
assert text[start] == '='
|
||||
self.cur.append((start, Operator, text[start]))
|
||||
start = self.term(start+1, text)
|
||||
return start
|
||||
|
||||
def rule(self, start, text):
|
||||
"""Tokenize a rule."""
|
||||
match = self.RULES.match(text, start)
|
||||
assert match is not None
|
||||
groups = sorted(match.regs[1:]) # exclude whole match
|
||||
for group in groups:
|
||||
if group[0] >= 0: # this group matched
|
||||
self.cur.append((start, Keyword, text[start:group[0]]))
|
||||
self.cur.append((group[0], Number.Integer,
|
||||
text[group[0]:group[1]]))
|
||||
if group[1] != match.end():
|
||||
self.cur.append((group[1], Keyword,
|
||||
text[group[1]:match.end()]))
|
||||
break
|
||||
else:
|
||||
self.cur.append((start, Keyword, text[start:match.end()]))
|
||||
return match.end()
|
||||
|
||||
def lineno(self, start, text):
|
||||
"""Tokenize a line referral."""
|
||||
end = start
|
||||
while text[end] not in self.NUMBERS:
|
||||
end += 1
|
||||
self.cur.append((start, Punctuation, text[start]))
|
||||
self.cur.append((start+1, Text, text[start+1:end]))
|
||||
start = end
|
||||
match = self.LINENOS.match(text, start)
|
||||
assert match is not None
|
||||
assert text[match.end()] == ')'
|
||||
self.cur.append((match.start(), Number.Integer, match.group(0)))
|
||||
self.cur.append((match.end(), Punctuation, text[match.end()]))
|
||||
return match.end() + 1
|
||||
|
||||
def error_till_line_end(self, start, text):
|
||||
"""Mark everything from ``start`` to the end of the line as Error."""
|
||||
end = start
|
||||
try:
|
||||
while text[end] != '\n': # there's whitespace in rules
|
||||
end += 1
|
||||
except IndexError:
|
||||
end = len(text)
|
||||
if end != start:
|
||||
self.cur.append((start, Error, text[start:end]))
|
||||
end = self.whitespace(end, text)
|
||||
return end
|
||||
|
||||
def get_tokens_unprocessed(self, text):
|
||||
"""Returns a list of TNT tokens."""
|
||||
self.cur = []
|
||||
start = end = self.whitespace(0, text)
|
||||
while start <= end < len(text):
|
||||
try:
|
||||
# try line number
|
||||
while text[end] in self.NUMBERS:
|
||||
end += 1
|
||||
if end != start: # actual number present
|
||||
self.cur.append((start, Number.Integer, text[start:end]))
|
||||
# whitespace is required after a line number
|
||||
orig = len(self.cur)
|
||||
try:
|
||||
start = end = self.whitespace(end, text, True)
|
||||
except AssertionError:
|
||||
del self.cur[orig:]
|
||||
start = end = self.error_till_line_end(end, text)
|
||||
continue
|
||||
# at this point it could be a comment
|
||||
match = self.COMMENT.match(text, start)
|
||||
if match is not None:
|
||||
self.cur.append((start, Comment, text[start:match.end()]))
|
||||
start = end = match.end()
|
||||
# anything after the closing bracket is invalid
|
||||
start = end = self.error_till_line_end(start, text)
|
||||
# do not attempt to process the rest
|
||||
continue
|
||||
del match
|
||||
if text[start] in '[]': # fantasy push or pop
|
||||
self.cur.append((start, Keyword, text[start]))
|
||||
start += 1
|
||||
end += 1
|
||||
else:
|
||||
# one formula, possibly containing subformulae
|
||||
orig = len(self.cur)
|
||||
try:
|
||||
start = end = self.formula(start, text)
|
||||
except AssertionError: # not well-formed
|
||||
del self.cur[orig:]
|
||||
while text[end] not in self.WHITESPACE:
|
||||
end += 1
|
||||
self.cur.append((start, Error, text[start:end]))
|
||||
start = end
|
||||
# skip whitespace after formula
|
||||
orig = len(self.cur)
|
||||
try:
|
||||
start = end = self.whitespace(end, text, True)
|
||||
except AssertionError:
|
||||
del self.cur[orig:]
|
||||
start = end = self.error_till_line_end(start, text)
|
||||
continue
|
||||
# rule proving this formula a theorem
|
||||
orig = len(self.cur)
|
||||
try:
|
||||
start = end = self.rule(start, text)
|
||||
except AssertionError:
|
||||
del self.cur[orig:]
|
||||
start = end = self.error_till_line_end(start, text)
|
||||
continue
|
||||
# skip whitespace after rule
|
||||
start = end = self.whitespace(end, text)
|
||||
# line marker
|
||||
if text[start] == '(':
|
||||
orig = len(self.cur)
|
||||
try:
|
||||
start = end = self.lineno(start, text)
|
||||
except AssertionError:
|
||||
del self.cur[orig:]
|
||||
start = end = self.error_till_line_end(start, text)
|
||||
continue
|
||||
start = end = self.whitespace(start, text)
|
||||
except IndexError:
|
||||
try:
|
||||
del self.cur[orig:]
|
||||
except NameError:
|
||||
pass # if orig was never defined, fine
|
||||
self.error_till_line_end(start, text)
|
||||
return self.cur
|
Loading…
Add table
Add a link
Reference in a new issue