|
- # -*- coding: utf-8 -*-
- #
- # Copyright (C) 2009-2013 Ben Kurtovic <ben.kurtovic@verizon.net>
- #
- # Permission is hereby granted, free of charge, to any person obtaining a copy
- # of this software and associated documentation files (the "Software"), to deal
- # in the Software without restriction, including without limitation the rights
- # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
- # copies of the Software, and to permit persons to whom the Software is
- # furnished to do so, subject to the following conditions:
- #
- # The above copyright notice and this permission notice shall be included in
- # all copies or substantial portions of the Software.
- #
- # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
- # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
- # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
- # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
- # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
- # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
- # SOFTWARE.
-
- from datetime import datetime
- import re
- from os.path import expanduser
- from threading import Lock
- from time import sleep
-
- import oursql
-
- from earwigbot import exceptions
- from earwigbot import wiki
- from earwigbot.tasks import Task
-
- class AFCStatistics(Task):
- """A task to generate statistics for WikiProject Articles for Creation.
-
- Statistics are stored in a MySQL database ("u_earwig_afc_statistics")
- accessed with oursql. Statistics are synchronied with the live database
- every four minutes and saved once an hour, on the hour, to self.pagename.
- In the live bot, this is "Template:AFC statistics".
- """
- name = "afc_statistics"
- number = 2
-
- # Chart status number constants:
- CHART_NONE = 0
- CHART_PEND = 1
- CHART_DRAFT = 2
- CHART_REVIEW = 3
- CHART_ACCEPT = 4
- CHART_DECLINE = 5
- CHART_MISPLACE = 6
-
- def setup(self):
- self.cfg = cfg = self.config.tasks.get(self.name, {})
-
- # Set some wiki-related attributes:
- self.pagename = cfg.get("page", "Template:AFC statistics")
- self.pending_cat = cfg.get("pending", "Pending AfC submissions")
- self.ignore_list = cfg.get("ignoreList", [])
- default_summary = "Updating statistics for [[WP:WPAFC|WikiProject Articles for creation]]."
- self.summary = self.make_summary(cfg.get("summary", default_summary))
-
- # Templates used in chart generation:
- templates = cfg.get("templates", {})
- self.tl_header = templates.get("header", "AFC statistics/header")
- self.tl_row = templates.get("row", "AFC statistics/row")
- self.tl_footer = templates.get("footer", "AFC statistics/footer")
-
- # Connection data for our SQL database:
- kwargs = cfg.get("sql", {})
- kwargs["read_default_file"] = expanduser("~/.my.cnf")
- self.conn_data = kwargs
- self.db_access_lock = Lock()
-
- def run(self, **kwargs):
- """Entry point for a task event.
-
- Depending on the kwargs passed, we will either synchronize our local
- statistics database with the site (self.sync()) or save it to the wiki
- (self.save()). We will additionally create an SQL connection with our
- local database.
- """
- action = kwargs.get("action")
- if not self.db_access_lock.acquire(False): # Non-blocking
- if action == "sync":
- self.logger.info("A sync is already ongoing; aborting")
- return
- self.logger.info("Waiting for database access lock")
- self.db_access_lock.acquire()
-
- try:
- self.site = self.bot.wiki.get_site()
- self.conn = oursql.connect(**self.conn_data)
- try:
- if action == "save":
- self.save(kwargs)
- elif action == "sync":
- self.sync(kwargs)
- elif action == "update":
- self.update(kwargs)
- finally:
- self.conn.close()
- finally:
- self.db_access_lock.release()
-
- def save(self, kwargs):
- """Save our local statistics to the wiki.
-
- After checking for emergency shutoff, the statistics chart is compiled,
- and then saved to self.pagename using self.summary iff it has changed
- since last save.
- """
- self.logger.info("Saving chart")
- if kwargs.get("fromIRC"):
- summary = self.summary + " (!earwigbot)"
- else:
- if self.shutoff_enabled():
- return
- summary = self.summary
-
- statistics = self.compile_charts()
-
- page = self.site.get_page(self.pagename)
- text = page.get()
- newtext = re.sub(u"<!-- stat begin -->(.*?)<!-- stat end -->",
- "<!-- stat begin -->\n" + statistics + "\n<!-- stat end -->",
- text, flags=re.DOTALL)
- if newtext == text:
- self.logger.info("Chart unchanged; not saving")
- return # Don't edit the page if we're not adding anything
-
- newtext = re.sub("<!-- sig begin -->(.*?)<!-- sig end -->",
- "<!-- sig begin -->~~~ at ~~~~~<!-- sig end -->",
- newtext)
- page.edit(newtext, summary, minor=True, bot=True)
- self.logger.info(u"Chart saved to [[{0}]]".format(page.title))
-
- def compile_charts(self):
- """Compile and return all statistics information from our local db."""
- stats = ""
- with self.conn.cursor() as cursor:
- cursor.execute("SELECT * FROM chart")
- for chart in cursor:
- stats += self.compile_chart(chart) + "\n"
- return stats[:-1] # Drop the last newline
-
- def compile_chart(self, chart_info):
- """Compile and return a single statistics chart."""
- chart_id, chart_title, special_title = chart_info
-
- chart = self.tl_header + "|" + chart_title
- if special_title:
- chart += "|" + special_title
- chart = "{{" + chart + "}}"
-
- query = "SELECT * FROM page JOIN row ON page_id = row_id WHERE row_chart = ?"
- with self.conn.cursor(oursql.DictCursor) as cursor:
- cursor.execute(query, (chart_id,))
- for page in cursor:
- chart += "\n" + self.compile_chart_row(page)
-
- chart += "\n{{" + self.tl_footer + "}}"
- return chart
-
- def compile_chart_row(self, page):
- """Compile and return a single chart row.
-
- 'page' is a dict of page information, taken as a row from the page
- table, where keys are column names and values are their cell contents.
- """
- row = u"{0}|s={page_status}|t={page_title}|h={page_short}|z={page_size}|"
- if page["page_special_oldid"]:
- row += "sr={page_special_user}|sd={page_special_time}|si={page_special_oldid}|"
- row += "mr={page_modify_user}|md={page_modify_time}|mi={page_modify_oldid}"
-
- page["page_special_time"] = self.format_time(page["page_special_time"])
- page["page_modify_time"] = self.format_time(page["page_modify_time"])
-
- if page["page_notes"]:
- row += "|n=1{page_notes}"
-
- return "{{" + row.format(self.tl_row, **page) + "}}"
-
- def format_time(self, dt):
- """Format a datetime into the standard MediaWiki timestamp format."""
- return dt.strftime("%H:%M, %d %b %Y")
-
- def sync(self, kwargs):
- """Synchronize our local statistics database with the site.
-
- Syncing involves, in order, updating tracked submissions that have
- been changed since last sync (self.update_tracked()), adding pending
- submissions that are not tracked (self.add_untracked()), and removing
- old submissions from the database (self.delete_old()).
-
- The sync will be canceled if SQL replication lag is greater than 600
- seconds, because this will lead to potential problems and outdated
- data, not to mention putting demand on an already overloaded server.
- Giving sync the kwarg "ignore_replag" will go around this restriction.
- """
- self.logger.info("Starting sync")
-
- replag = self.site.get_replag()
- self.logger.debug("Server replag is {0}".format(replag))
- if replag > 600 and not kwargs.get("ignore_replag"):
- msg = "Sync canceled as replag ({0} secs) is greater than ten minutes"
- self.logger.warn(msg.format(replag))
- return
-
- with self.conn.cursor() as cursor:
- self.update_tracked(cursor)
- self.add_untracked(cursor)
- self.delete_old(cursor)
-
- self.logger.info("Sync completed")
-
- def update_tracked(self, cursor):
- """Update tracked submissions that have been changed since last sync.
-
- This is done by iterating through every page in our database and
- comparing our stored latest revision ID with the actual latest revision
- ID from an SQL query. If they differ, we will update our information
- about the page (self.update_page()).
-
- If the page does not exist, we will remove it from our database with
- self.untrack_page().
- """
- self.logger.debug("Updating tracked submissions")
- query1 = "SELECT page_id, page_title, page_modify_oldid FROM page"
- query2 = """SELECT page_latest, page_title, page_namespace FROM page
- WHERE page_id = ?"""
- cursor.execute(query1)
-
- for pageid, title, oldid in cursor:
- result = list(self.site.sql_query(query2, (pageid,)))
- if not result:
- self.untrack_page(cursor, pageid)
- continue
-
- real_oldid = result[0][0]
- if oldid != real_oldid:
- msg = u"Updating page [[{0}]] (id: {1}) @ {2}"
- self.logger.debug(msg.format(title, pageid, oldid))
- self.logger.debug(" {0} -> {1}".format(oldid, real_oldid))
- base = result[0][1].decode("utf8").replace("_", " ")
- ns = self.site.namespace_id_to_name(result[0][2])
- if ns:
- real_title = u":".join((ns, base))
- else:
- real_title = base
- try:
- self.update_page(cursor, pageid, real_title)
- except Exception:
- e = u"Error updating page [[{0}]] (id: {1})"
- self.logger.exception(e.format(real_title, pageid))
-
- def add_untracked(self, cursor):
- """Add pending submissions that are not yet tracked.
-
- This is done by compiling a list of all currently tracked submissions
- and iterating through all members of self.pending_cat via SQL. If a
- page in the pending category is not tracked and is not in
- self.ignore_list, we will track it with self.track_page().
- """
- self.logger.debug("Adding untracked pending submissions")
- cursor.execute("SELECT page_id FROM page")
- tracked = [i[0] for i in cursor.fetchall()]
-
- category = self.site.get_category(self.pending_cat)
- for page in category.get_members():
- title, pageid = page.title, page.pageid
- if title in self.ignore_list:
- continue
- if pageid not in tracked:
- msg = u"Tracking page [[{0}]] (id: {1})".format(title, pageid)
- self.logger.debug(msg)
- try:
- self.track_page(cursor, pageid, title)
- except Exception:
- e = u"Error tracking page [[{0}]] (id: {1})"
- self.logger.exception(e.format(title, pageid))
-
- def delete_old(self, cursor):
- """Remove old submissions from the database.
-
- "Old" is defined as a submission that has been declined or accepted
- more than 36 hours ago. Pending submissions cannot be "old".
- """
- self.logger.debug("Removing old submissions from chart")
- query = """DELETE FROM page, row USING page JOIN row
- ON page_id = row_id WHERE row_chart IN (?, ?)
- AND ADDTIME(page_special_time, '36:00:00') < NOW()"""
- cursor.execute(query, (self.CHART_ACCEPT, self.CHART_DECLINE))
-
- def update(self, kwargs):
- """Update a page by name, regardless of whether anything has changed.
-
- Mainly intended as a command to be used via IRC, e.g.:
- !tasks start afc_statistics action=update page=Foobar
- """
- title = kwargs.get("page")
- if not title:
- return
-
- title = title.replace("_", " ").decode("utf8")
- query = "SELECT page_id, page_modify_oldid FROM page WHERE page_title = ?"
- with self.conn.cursor() as cursor:
- cursor.execute(query, (title,))
- try:
- pageid, oldid = cursor.fetchall()[0]
- except IndexError:
- msg = u"Page [[{0}]] not found in database".format(title)
- self.logger.error(msg)
-
- msg = u"Updating page [[{0}]] (id: {1}) @ {2}"
- self.logger.info(msg.format(title, pageid, oldid))
- self.update_page(cursor, pageid, title)
-
- def untrack_page(self, cursor, pageid):
- """Remove a page, given by ID, from our database."""
- self.logger.debug("Untracking page (id: {0})".format(pageid))
- query = """DELETE FROM page, row USING page JOIN row
- ON page_id = row_id WHERE page_id = ?"""
- cursor.execute(query, (pageid,))
-
- def track_page(self, cursor, pageid, title):
- """Update hook for when page is not in our database.
-
- A variety of SQL queries are used to gather information about the page,
- which is then saved to our database.
- """
- content = self.get_content(title)
- if content is None:
- msg = u"Could not get page content for [[{0}]]".format(title)
- self.logger.error(msg)
- return
-
- namespace = self.site.get_page(title).namespace
- status, chart = self.get_status_and_chart(content, namespace)
- if chart == self.CHART_NONE:
- msg = u"Could not find a status for [[{0}]]".format(title)
- self.logger.warn(msg)
- return
-
- short = self.get_short_title(title)
- size = self.get_size(content)
- m_user, m_time, m_id = self.get_modify(pageid)
- s_user, s_time, s_id = self.get_special(pageid, chart)
- notes = self.get_notes(chart, content, m_time, s_user)
-
- query1 = "INSERT INTO row VALUES (?, ?)"
- query2 = "INSERT INTO page VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)"
- cursor.execute(query1, (pageid, chart))
- cursor.execute(query2, (pageid, status, title, short, size, notes,
- m_user, m_time, m_id, s_user, s_time, s_id))
-
- def update_page(self, cursor, pageid, title):
- """Update hook for when page is already in our database.
-
- A variety of SQL queries are used to gather information about the page,
- which is compared against our stored information. Differing information
- is then updated.
- """
- content = self.get_content(title)
- if content is None:
- msg = u"Could not get page content for [[{0}]]".format(title)
- self.logger.error(msg)
- return
-
- namespace = self.site.get_page(title).namespace
- status, chart = self.get_status_and_chart(content, namespace)
- if chart == self.CHART_NONE:
- self.untrack_page(cursor, pageid)
- return
-
- query = "SELECT * FROM page JOIN row ON page_id = row_id WHERE page_id = ?"
- with self.conn.cursor(oursql.DictCursor) as dict_cursor:
- dict_cursor.execute(query, (pageid,))
- result = dict_cursor.fetchall()[0]
-
- size = self.get_size(content)
- m_user, m_time, m_id = self.get_modify(pageid)
-
- if title != result["page_title"]:
- self.update_page_title(cursor, result, pageid, title)
-
- if m_id != result["page_modify_oldid"]:
- self.update_page_modify(cursor, result, pageid, size, m_user,
- m_time, m_id)
-
- if status != result["page_status"]:
- special = self.update_page_status(cursor, result, pageid, status,
- chart)
- s_user = special[0]
- else:
- s_user = result["page_special_user"]
-
- notes = self.get_notes(chart, content, m_time, s_user)
- if notes != result["page_notes"]:
- self.update_page_notes(cursor, result, pageid, notes)
-
- def update_page_title(self, cursor, result, pageid, title):
- """Update the title and short_title of a page in our database."""
- query = "UPDATE page SET page_title = ?, page_short = ? WHERE page_id = ?"
- short = self.get_short_title(title)
- cursor.execute(query, (title, short, pageid))
-
- msg = u" {0}: title: {1} -> {2}"
- self.logger.debug(msg.format(pageid, result["page_title"], title))
-
- def update_page_modify(self, cursor, result, pageid, size, m_user, m_time, m_id):
- """Update the last modified information of a page in our database."""
- query = """UPDATE page SET page_size = ?, page_modify_user = ?,
- page_modify_time = ?, page_modify_oldid = ?
- WHERE page_id = ?"""
- cursor.execute(query, (size, m_user, m_time, m_id, pageid))
-
- msg = u" {0}: modify: {1} / {2} / {3} -> {4} / {5} / {6}"
- msg = msg.format(pageid, result["page_modify_user"],
- result["page_modify_time"],
- result["page_modify_oldid"], m_user, m_time, m_id)
- self.logger.debug(msg)
-
- def update_page_status(self, cursor, result, pageid, status, chart):
- """Update the status and "specialed" information of a page."""
- query1 = """UPDATE page JOIN row ON page_id = row_id
- SET page_status = ?, row_chart = ? WHERE page_id = ?"""
- query2 = """UPDATE page SET page_special_user = ?,
- page_special_time = ?, page_special_oldid = ?
- WHERE page_id = ?"""
- cursor.execute(query1, (status, chart, pageid))
-
- msg = " {0}: status: {1} ({2}) -> {3} ({4})"
- self.logger.debug(msg.format(pageid, result["page_status"],
- result["row_chart"], status, chart))
-
- s_user, s_time, s_id = self.get_special(pageid, chart)
- if s_id != result["page_special_oldid"]:
- cursor.execute(query2, (s_user, s_time, s_id, pageid))
- msg = u"{0}: special: {1} / {2} / {3} -> {4} / {5} / {6}"
- msg = msg.format(pageid, result["page_special_user"],
- result["page_special_time"],
- result["page_special_oldid"], s_user, s_time, s_id)
- self.logger.debug(msg)
-
- return s_user, s_time, s_id
-
- def update_page_notes(self, cursor, result, pageid, notes):
- """Update the notes (or warnings) of a page in our database."""
- query = "UPDATE page SET page_notes = ? WHERE page_id = ?"
- cursor.execute(query, (notes, pageid))
- msg = " {0}: notes: {1} -> {2}"
- self.logger.debug(msg.format(pageid, result["page_notes"], notes))
-
- def get_content(self, title):
- """Get the current content of a page by title from the API.
-
- The page's current revision ID is retrieved from SQL, and then
- an API query is made to get its content. This is the only API query
- used in the task's code.
- """
- query = "SELECT page_latest FROM page WHERE page_title = ? AND page_namespace = ?"
- try:
- namespace, base = title.split(":", 1)
- except ValueError:
- base = title
- ns = wiki.NS_MAIN
- else:
- try:
- ns = self.site.namespace_name_to_id(namespace)
- except exceptions.NamespaceNotFoundError:
- base = title
- ns = wiki.NS_MAIN
-
- result = self.site.sql_query(query, (base.replace(" ", "_"), ns))
- try:
- revid = int(list(result)[0][0])
- except IndexError:
- return None
- return self.get_revision_content(revid)
-
- def get_revision_content(self, revid, tries=1):
- """Get the content of a revision by ID from the API."""
- res = self.site.api_query(action="query", prop="revisions",
- revids=revid, rvprop="content")
- try:
- return res["query"]["pages"].values()[0]["revisions"][0]["*"]
- except KeyError:
- if tries > 0:
- sleep(5)
- return self.get_revision_content(revid, tries=tries - 1)
-
- def get_status_and_chart(self, content, namespace):
- """Determine the status and chart number of an AFC submission.
-
- The methodology used here is the same one I've been using for years
- (see also commands.afc_report), but with the new draft system taken
- into account. The order here is important: if there is more than one
- {{AFC submission}} template on a page, we need to know which one to
- use (revision history search to find the most recent isn't a viable
- idea :P).
- """
- statuses = self.get_statuses(content)
-
- if "R" in statuses:
- status, chart = "r", self.CHART_REVIEW
- elif "H" in statuses:
- status, chart = "p", self.CHART_DRAFT
- elif "P" in statuses:
- status, chart = "p", self.CHART_PEND
- elif "T" in statuses:
- status, chart = None, self.CHART_NONE
- elif "D" in statuses:
- status, chart = "d", self.CHART_DECLINE
- else:
- status, chart = None, self.CHART_NONE
-
- if namespace == wiki.NS_MAIN:
- if not statuses:
- status, chart = "a", self.CHART_ACCEPT
- else:
- status, chart = None, self.CHART_MISPLACE
-
- return status, chart
-
- def get_statuses(self, content):
- """Return a list of all AFC submission statuses in a page's text."""
- re_has_templates = "\{\{[aA][fF][cC] submission\s*(\}\}|\||/)"
- re_template = "\{\{[aA][fF][cC] submission\s*(.*?)\}\}"
- re_remove_embed = "(\{\{[aA][fF][cC] submission\s*(.*?))\{\{(.*?)\}\}(.*?)\}\}"
- valid = ["R", "H", "P", "T", "D"]
- subtemps = {
- "/reviewing": "R",
- "/onhold": "H",
- "/pending": "P",
- "/draft": "T",
- "/declined": "D"
- }
- statuses = []
-
- while re.search(re_has_templates, content):
- status = "P"
- match = re.search(re_template, content, re.S)
- if not match:
- return statuses
- temp = match.group(1)
- limit = 0
- while "{{" in temp and limit < 50:
- content = re.sub(re_remove_embed, "\\1\\4}}", content, 1, re.S)
- match = re.search(re_template, content, re.S)
- temp = match.group(1)
- limit += 1
- params = temp.split("|")
- try:
- subtemp, params = params[0].strip(), params[1:]
- except IndexError:
- status = "P"
- params = []
- else:
- if subtemp:
- status = subtemps.get(subtemp)
- params = []
- for param in params:
- param = param.strip().upper()
- if "=" in param:
- key, value = param.split("=", 1)
- if key.strip() == "1":
- status = value if value in valid else "P"
- break
- else:
- status = param if param in valid else "P"
- break
- statuses.append(status)
- content = re.sub(re_template, "", content, 1, re.S)
-
- return statuses
-
- def get_short_title(self, title):
- """Shorten a title so we can display it in a chart using less space.
-
- Basically, this just means removing the "Wikipedia talk:Articles for
- creation" part from the beginning. If it is longer than 50 characters,
- we'll shorten it down to 47 and add an poor-man's ellipsis at the end.
- """
- short = re.sub("Wikipedia(\s*talk)?\:Articles\sfor\screation\/", "", title)
- if len(short) > 50:
- short = short[:47] + "..."
- return short
-
- def get_size(self, content):
- """Return a page's size in a short, pretty format."""
- return "{0} kB".format(round(len(content) / 1000.0, 1))
-
- def get_modify(self, pageid):
- """Return information about a page's last edit ("modification").
-
- This consists of the most recent editor, modification time, and the
- lastest revision ID.
- """
- query = """SELECT rev_user_text, rev_timestamp, rev_id FROM revision
- JOIN page ON rev_id = page_latest WHERE page_id = ?"""
- result = self.site.sql_query(query, (pageid,))
- m_user, m_time, m_id = list(result)[0]
- timestamp = datetime.strptime(m_time, "%Y%m%d%H%M%S")
- return m_user.decode("utf8"), timestamp, m_id
-
- def get_special(self, pageid, chart):
- """Return information about a page's "special" edit.
-
- I tend to use the term "special" as a verb a lot, which is bound to
- cause confusion. It is merely a short way of saying "the edit in which
- a declined submission was declined, an accepted submission was
- accepted, a submission in review was set as such, a pending submission
- was submitted, and a "misplaced" submission was created."
-
- This "information" consists of the special edit's editor, its time, and
- its revision ID. If the page's status is not something that involves
- "special"-ing, we will return None for all three. The same will be
- returned if we cannot determine when the page was "special"-ed, or if
- it was "special"-ed more than 100 edits ago.
- """
- if chart == self.CHART_NONE:
- return None, None, None
- elif chart == self.CHART_MISPLACE:
- return self.get_create(pageid)
- elif chart == self.CHART_ACCEPT:
- search_for = None
- search_not = ["R", "H", "P", "T", "D"]
- elif chart == self.CHART_DRAFT:
- search_for = "H"
- search_not = []
- elif chart == self.CHART_PEND:
- search_for = "P"
- search_not = []
- elif chart == self.CHART_REVIEW:
- search_for = "R"
- search_not = []
- elif chart == self.CHART_DECLINE:
- search_for = "D"
- search_not = ["R", "H", "P", "T"]
-
- query = """SELECT rev_user_text, rev_timestamp, rev_id
- FROM revision WHERE rev_page = ? ORDER BY rev_id DESC"""
- result = self.site.sql_query(query, (pageid,))
-
- counter = 0
- last = (None, None, None)
- for user, ts, revid in result:
- counter += 1
- if counter > 50:
- msg = "Exceeded 50 content lookups while determining special for page (id: {0}, chart: {1})"
- self.logger.warn(msg.format(pageid, chart))
- return None, None, None
- try:
- content = self.get_revision_content(revid)
- except exceptions.APIError:
- msg = "API error interrupted SQL query in get_special() for page (id: {0}, chart: {1})"
- self.logger.exception(msg.format(pageid, chart))
- return None, None, None
- statuses = self.get_statuses(content)
- matches = [s in statuses for s in search_not]
- if search_for:
- if search_for not in statuses or any(matches):
- return last
- else:
- if any(matches):
- return last
- timestamp = datetime.strptime(ts, "%Y%m%d%H%M%S")
- last = (user.decode("utf8"), timestamp, revid)
-
- return last
-
- def get_create(self, pageid):
- """Return information about a page's first edit ("creation").
-
- This consists of the page creator, creation time, and the earliest
- revision ID.
- """
- query = """SELECT rev_user_text, rev_timestamp, rev_id
- FROM revision WHERE rev_id =
- (SELECT MIN(rev_id) FROM revision WHERE rev_page = ?)"""
- result = self.site.sql_query(query, (pageid,))
- c_user, c_time, c_id = list(result)[0]
- timestamp = datetime.strptime(c_time, "%Y%m%d%H%M%S")
- return c_user.decode("utf8"), timestamp, c_id
-
- def get_notes(self, chart, content, m_time, s_user):
- """Return any special notes or warnings about this page.
-
- copyvio: submission is a suspected copyright violation
- unsourced: submission lacks references completely
- no-inline: submission has no inline citations
- short: submission is less than a kilobyte in length
- resubmit: submission was resubmitted after a previous decline
- old: submission has not been touched in > 4 days
- blocked: submitter is currently blocked
- """
- notes = ""
-
- ignored_charts = [self.CHART_NONE, self.CHART_ACCEPT, self.CHART_DECLINE]
- if chart in ignored_charts:
- return notes
-
- copyvios = self.config.tasks.get("afc_copyvios", {})
- regex = "\{\{\s*" + copyvios.get("template", "AfC suspected copyvio")
- if re.search(regex, content):
- notes += "|nc=1" # Submission is a suspected copyvio
-
- if not re.search("\<ref\s*(.*?)\>(.*?)\</ref\>", content, re.I | re.S):
- regex = "(https?:)|\[//(?!{0})([^ \]\\t\\n\\r\\f\\v]+?)"
- sitedomain = re.escape(self.site.domain)
- if re.search(regex.format(sitedomain), content, re.I | re.S):
- notes += "|ni=1" # Submission has no inline citations
- else:
- notes += "|nu=1" # Submission is completely unsourced
-
- if len(content) < 1000:
- notes += "|ns=1" # Submission is short
-
- statuses = self.get_statuses(content)
- if "D" in statuses and chart != self.CHART_MISPLACE:
- notes += "|nr=1" # Submission was resubmitted
-
- time_since_modify = (datetime.utcnow() - m_time).total_seconds()
- max_time = 4 * 24 * 60 * 60
- if time_since_modify > max_time:
- notes += "|no=1" # Submission hasn't been touched in over 4 days
-
- if chart in [self.CHART_PEND, self.CHART_DRAFT] and s_user:
- submitter = self.site.get_user(s_user)
- try:
- if submitter.blockinfo:
- notes += "|nb=1" # Submitter is blocked
- except exceptions.UserNotFoundError: # Likely an IP
- pass
-
- return notes
|