From 36180a9e473c74fceaf4587b26806815bc313dbc Mon Sep 17 00:00:00 2001 From: Ben Kurtovic Date: Wed, 14 Aug 2013 02:38:03 -0400 Subject: [PATCH] To clarify usage, emit_text() -> emit_char() and emit_string() -> emit_text() --- mwparserfromhell/parser/tokenizer.c | 96 ++++++++++++++++++------------------- 1 file changed, 48 insertions(+), 48 deletions(-) diff --git a/mwparserfromhell/parser/tokenizer.c b/mwparserfromhell/parser/tokenizer.c index 62e8599..eff000a 100644 --- a/mwparserfromhell/parser/tokenizer.c +++ b/mwparserfromhell/parser/tokenizer.c @@ -102,9 +102,9 @@ static void Textbuffer_dealloc(Textbuffer* self) } /* - Write text to the given textbuffer. + Write a Unicode codepoint to the given textbuffer. */ -static int Textbuffer_write(Textbuffer** this, Py_UNICODE text) +static int Textbuffer_write(Textbuffer** this, Py_UNICODE code) { Textbuffer* self = *this; @@ -115,7 +115,7 @@ static int Textbuffer_write(Textbuffer** this, Py_UNICODE text) new->next = self; *this = self = new; } - self->data[self->size] = text; + self->data[self->size] = code; self->size++; return 0; } @@ -369,22 +369,22 @@ static int Tokenizer_emit_first(Tokenizer* self, PyObject* token) } /* - Write text to the current textbuffer. + Write a Unicode codepoint to the current textbuffer. */ -static int Tokenizer_emit_text(Tokenizer* self, Py_UNICODE text) +static int Tokenizer_emit_char(Tokenizer* self, Py_UNICODE code) { - return Textbuffer_write(&(self->topstack->textbuffer), text); + return Textbuffer_write(&(self->topstack->textbuffer), code); } /* Write a string of text to the current textbuffer. */ -static int Tokenizer_emit_string(Tokenizer* self, const char* text) +static int Tokenizer_emit_text(Tokenizer* self, const char* text) { int i = 0; while (text[i]) { - if (Tokenizer_emit_text(self, text[i])) + if (Tokenizer_emit_char(self, text[i])) return -1; i++; } @@ -456,7 +456,7 @@ static int Tokenizer_emit_text_then_stack(Tokenizer* self, const char* text) { PyObject* stack = Tokenizer_pop(self); - if (Tokenizer_emit_string(self, text)) { + if (Tokenizer_emit_text(self, text)) { Py_DECREF(stack); return -1; } @@ -783,7 +783,7 @@ static int Tokenizer_parse_wikilink(Tokenizer* self) if (BAD_ROUTE) { RESET_ROUTE(); self->head = reset; - if (Tokenizer_emit_string(self, "[[")) + if (Tokenizer_emit_text(self, "[[")) return -1; return 0; } @@ -869,7 +869,7 @@ static int Tokenizer_parse_heading(Tokenizer* self) RESET_ROUTE(); self->head = reset + best - 1; for (i = 0; i < best; i++) { - if (Tokenizer_emit_text(self, *"=")) + if (Tokenizer_emit_char(self, *"=")) return -1; } self->global ^= GL_HEADING; @@ -907,7 +907,7 @@ static int Tokenizer_parse_heading(Tokenizer* self) if (heading->level < best) { diff = best - heading->level; for (i = 0; i < diff; i++) { - if (Tokenizer_emit_text(self, *"=")) { + if (Tokenizer_emit_char(self, *"=")) { Py_DECREF(heading->title); free(heading); return -1; @@ -958,7 +958,7 @@ static HeadingData* Tokenizer_handle_heading_end(Tokenizer* self) if (level < best) { diff = best - level; for (i = 0; i < diff; i++) { - if (Tokenizer_emit_text(self, *"=")) + if (Tokenizer_emit_char(self, *"=")) return NULL; } } @@ -966,7 +966,7 @@ static HeadingData* Tokenizer_handle_heading_end(Tokenizer* self) } else { for (i = 0; i < best; i++) { - if (Tokenizer_emit_text(self, *"=")) { + if (Tokenizer_emit_char(self, *"=")) { Py_DECREF(after->title); free(after); return NULL; @@ -1182,7 +1182,7 @@ static int Tokenizer_parse_entity(Tokenizer* self) if (BAD_ROUTE) { RESET_ROUTE(); self->head = reset; - if (Tokenizer_emit_text(self, *"&")) + if (Tokenizer_emit_char(self, *"&")) return -1; return 0; } @@ -1210,7 +1210,7 @@ static int Tokenizer_parse_comment(Tokenizer* self) if (BAD_ROUTE) { RESET_ROUTE(); self->head = reset; - if (Tokenizer_emit_string(self, "