Bladeren bron

Tokenizer now supports a very, very limited template syntax.

tags/v0.1
Ben Kurtovic 12 jaren geleden
bovenliggende
commit
32d99c3c75
2 gewijzigde bestanden met toevoegingen van 24 en 9 verwijderingen
  1. +8
    -8
      mwparserfromhell/parser/builder.py
  2. +16
    -1
      mwparserfromhell/parser/tokenizer.py

+ 8
- 8
mwparserfromhell/parser/builder.py Bestand weergeven

@@ -59,7 +59,7 @@ class Builder(object):
value = self._pop()
return Parameter(key, value, showkey)
else:
self._write(self._handle_token())
self._write(self._handle_token(token))

def _handle_template(self):
params = []
@@ -81,7 +81,7 @@ class Builder(object):
name = self._pop()
return Template(name, params)
else:
self._write(self._handle_token())
self._write(self._handle_token(token))

def _handle_entity(self):
token = self._tokens.pop(0)
@@ -103,7 +103,7 @@ class Builder(object):
title = self._pop()
return Heading(title, level)
else:
self._write(self._handle_token())
self._write(self._handle_token(token))

def _handle_attribute(self):
name, quoted = None, False
@@ -122,7 +122,7 @@ class Builder(object):
return Attribute(name, self._pop(), quoted)
return Attribute(self._pop(), quoted=quoted)
else:
self._write(self._handle_token())
self._write(self._handle_token(token))

def _handle_tag(self, token):
type_, showtag = token.type, token.showtag
@@ -146,10 +146,9 @@ class Builder(object):
return Tag(type_, tag, contents, attrs, showtag, False,
open_pad, token.padding)
else:
self._write(self._handle_token())
self._write(self._handle_token(token))

def _handle_token(self):
token = self._tokens.pop(0)
def _handle_token(self, token):
if isinstance(token, tokens.Text):
return Text(token.text)
elif isinstance(token, tokens.TemplateOpen):
@@ -165,5 +164,6 @@ class Builder(object):
self._tokens = tokenlist
self._push()
while self._tokens:
self._write(self._handle_token())
node = self._handle_token(self._tokens.pop(0))
self._write(node)
return self._pop()

+ 16
- 1
mwparserfromhell/parser/tokenizer.py Bestand weergeven

@@ -24,6 +24,9 @@ from . import tokens

__all__ = ["Tokenizer"]

class BadRoute(Exception):
pass

class Tokenizer(object):
START = object()
END = object()
@@ -33,7 +36,7 @@ class Tokenizer(object):
self._head = 0
self._stacks = []

self._modifiers = []
self._context = []

def _push(self):
self._stacks.append([])
@@ -66,6 +69,18 @@ class Tokenizer(object):
while True:
if self._read() in (stop, self.END):
return self._pop()
elif self._read(0) == "{" and self._read(1) == "{":
reset = self._head
self._head += 2
try:
template = self._parse_until("}")
except BadRoute:
self._head = reset
self._write(tokens.Text(text=self._read()))
else:
self._write(tokens.TemplateOpen())
self._stacks[-1] += template
self._write(tokens.TemplateClose())
else:
self._write(tokens.Text(text=self._read()))
self._head += 1


Laden…
Annuleren
Opslaan