from collections import namedtuple
import os, os.path
import pkg_resources
+import random
import re
+import shutil
+import unicodedata
from sqlalchemy.sql import func
import whoosh
import whoosh.filedb.fileindex
import whoosh.index
from whoosh.qparser import QueryParser
+import whoosh.scoring
import whoosh.spelling
from pokedex.db import connect
import pokedex.db.tables as tables
from pokedex.roomaji import romanize
+__all__ = ['open_index', 'lookup', 'random_lookup']
+
+INTERMEDIATE_LOOKUP_RESULTS = 25
+MAX_LOOKUP_RESULTS = 10
+
# Dictionary of table name => table class.
# Need the table name so we can get the class from the table name after we
# retrieve something from the index
]:
indexed_tables[cls.__tablename__] = cls
-# Dictionary of extra keys to file types of objects under, e.g. Pokémon can
-# also be looked up purely by number
-extra_keys = {
- tables.Move: [
- lambda row: u"move %d" % row.id,
- ],
- tables.Pokemon: [
- lambda row: unicode(row.id),
- ],
-}
+def normalize(name):
+ """Strips irrelevant formatting junk from name input.
+
+ Specifically: everything is lowercased, and accents are removed.
+ """
+ # http://stackoverflow.com/questions/517923/what-is-the-best-way-to-remove-accents-in-a-python-unicode-string
+ # Makes sense to me. Decompose by Unicode rules, then remove combining
+ # characters, then recombine. I'm explicitly doing it this way instead of
+ # testing combining() because Korean characters apparently decompose! But
+ # the results are considered letters, not combining characters, so testing
+ # for Mn works well, and combining them again makes them look right.
+ nkfd_form = unicodedata.normalize('NFKD', unicode(name))
+ name = u"".join(c for c in nkfd_form
+ if unicodedata.category(c) != 'Mn')
+ name = unicodedata.normalize('NFC', name)
+
+ name = name.strip()
+ name = name.lower()
+
+ return name
+
def open_index(directory=None, session=None, recreate=False):
"""Opens the whoosh index stored in the named directory and returns (index,
# Defaults
if not directory:
directory = pkg_resources.resource_filename('pokedex',
- 'data/whoosh_index')
+ 'data/whoosh-index')
if not session:
session = connect()
# Apparently not a real index. Fall out of the if and create it
pass
+ # Delete and start over if we're going to bail anyway.
+ if directory_exists and recreate:
+ # Be safe and only delete if it looks like a whoosh index, i.e.,
+ # everything starts with _
+ if all(f[0] == '_' for f in os.listdir(directory)):
+ shutil.rmtree(directory)
+ directory_exists = False
+
if not directory_exists:
os.mkdir(directory)
- # Create index
+ ### Create index
schema = whoosh.fields.Schema(
name=whoosh.fields.ID(stored=True),
- table=whoosh.fields.STORED,
- row_id=whoosh.fields.STORED,
+ table=whoosh.fields.ID(stored=True),
+ row_id=whoosh.fields.ID(stored=True),
language=whoosh.fields.STORED,
+ iso3166=whoosh.fields.STORED,
+ display_name=whoosh.fields.STORED, # non-lowercased name
+ forme_name=whoosh.fields.ID,
)
index = whoosh.index.create_in(directory, schema=schema, indexname='MAIN')
writer = index.writer()
# Index every name in all our tables of interest
+ # speller_entries becomes a list of (word, score) tuples; the score is 2
+ # for English names, 1.5 for Roomaji, and 1 for everything else. I think
+ # this biases the results in the direction most people expect, especially
+ # when e.g. German names are very similar to English names
speller_entries = []
for cls in indexed_tables.values():
q = session.query(cls)
- # Only index base Pokémon formes
- if hasattr(cls, 'forme_base_pokemon_id'):
- q = q.filter_by(forme_base_pokemon_id=None)
-
for row in q.yield_per(5):
- row_key = dict(table=cls.__tablename__, row_id=row.id)
-
- name = row.name.lower()
- writer.add_document(name=name, **row_key)
- speller_entries.append(name)
-
- for extra_key_func in extra_keys.get(cls, []):
- extra_key = extra_key_func(row)
- writer.add_document(name=extra_key, **row_key)
+ # XXX need to give forme_name a dummy value because I can't search
+ # for explicitly empty fields. boo.
+ row_key = dict(table=unicode(cls.__tablename__),
+ row_id=unicode(row.id),
+ forme_name=u'XXX')
+
+ def add(name, language, iso3166, score):
+ normalized_name = normalize(name)
+ writer.add_document(name=normalized_name, display_name=name,
+ language=language,
+ iso3166=iso3166,
+ **row_key)
+ speller_entries.append((normalized_name, score))
+
+ # If this is a form, mark it as such
+ if getattr(row, 'forme_base_pokemon_id', None):
+ row_key['forme_name'] = row.forme_name
+
+ name = row.name
+ add(name, None, u'us', 1)
# Pokemon also get other languages
- if cls == tables.Pokemon:
- for foreign_name in row.foreign_names:
- name = foreign_name.name.lower()
- writer.add_document(name=name,
- language=foreign_name.language.name,
- **row_key)
- speller_entries.append(name)
-
- if foreign_name.language.name == 'Japanese':
- # Add Roomaji too
- roomaji = romanize(foreign_name.name).lower()
- writer.add_document(name=roomaji,
- language='Roomaji',
- **row_key)
- speller_entries.append(roomaji)
-
+ for foreign_name in getattr(row, 'foreign_names', []):
+ moonspeak = foreign_name.name
+ if name == moonspeak:
+ # Don't add the English name again as a different language;
+ # no point and it makes spell results confusing
+ continue
+
+ add(moonspeak, foreign_name.language.name,
+ foreign_name.language.iso3166,
+ 3)
+
+ # Add Roomaji too
+ if foreign_name.language.name == 'Japanese':
+ roomaji = romanize(foreign_name.name)
+ add(roomaji, u'Roomaji', u'jp', 8)
writer.commit()
# at once, as every call to add_* does a commit(), and those seem to be
# expensive
speller = whoosh.spelling.SpellChecker(index.storage)
- speller.add_words(speller_entries)
+ speller.add_scored_words(speller_entries)
return index, speller
-LookupResult = namedtuple('LookupResult', ['object', 'language', 'exact'])
-def lookup(name, session=None, indices=None, exact_only=False):
+class LanguageWeighting(whoosh.scoring.Weighting):
+ """A scoring class that forces otherwise-equal English results to come
+ before foreign results.
+ """
+
+ def score(self, searcher, fieldnum, text, docnum, weight, QTF=1):
+ doc = searcher.stored_fields(docnum)
+ if doc['language'] == None:
+ # English (well, "default"); leave it at 1
+ return weight
+ elif doc['language'] == u'Roomaji':
+ # Give Roomaji a bit of a boost, as it's most likely to be searched
+ return weight * 0.95
+ else:
+ # Everything else can drop down the totem pole
+ return weight * 0.9
+
+rx_is_number = re.compile('^\d+$')
+
+LookupResult = namedtuple('LookupResult',
+ ['object', 'name', 'language', 'iso3166', 'exact'])
+
+def _parse_table_name(name):
+ """Takes a singular table name, table name, or table object and returns the
+ table name.
+
+ Returns None for a bogus name.
+ """
+ if hasattr(name, '__tablename__'):
+ return getattr(name, '__tablename__')
+ elif name in indexed_tables:
+ return name
+ elif name + 's' in indexed_tables:
+ return name + 's'
+ else:
+ # Bogus. Be nice and return dummy
+ return None
+
+def _whoosh_records_to_results(records, session, exact=True):
+ """Converts a list of whoosh's indexed records to LookupResult tuples
+ containing database objects.
+ """
+ # XXX this 'exact' thing is getting kinda leaky. would like a better way
+ # to handle it, since only lookup() cares about fuzzy results
+ seen = {}
+ results = []
+ for record in records:
+ # Skip dupes
+ seen_key = record['table'], record['row_id']
+ if seen_key in seen:
+ continue
+ seen[seen_key] = True
+
+ cls = indexed_tables[record['table']]
+ obj = session.query(cls).get(record['row_id'])
+
+ results.append(LookupResult(object=obj,
+ name=record['display_name'],
+ language=record['language'],
+ iso3166=record['iso3166'],
+ exact=exact))
+
+ return results
+
+
+def lookup(input, valid_types=[], session=None, indices=None, exact_only=False):
"""Attempts to find some sort of object, given a database session and name.
- Returns a list of named (object, language, exact) tuples. `object` is a
- database object, `language` is the name of the language in which the name
- was found, and `exact` is True iff this was an exact match.
+ Returns a list of named (object, name, language, iso3166, exact) tuples.
+ `object` is a database object, `name` is the name under which the object
+ was found, `language` and `iso3166` are the name and country code of the
+ language in which the name was found, and `exact` is True iff this was an
+ exact match.
This function currently ONLY does fuzzy matching if there are no exact
matches.
- Formes are not returned; "Shaymin" will return only grass Shaymin.
+ Formes are not returned unless requested; "Shaymin" will return only grass
+ Shaymin.
- Recognizes:
- - Pokémon names: "Eevee"
+ Extraneous whitespace is removed with extreme prejudice.
- `name`
+ Recognizes:
+ - Names: "Eevee", "Surf", "Run Away", "Payapa Berry", etc.
+ - Foreign names: "Iibui", "Eivui"
+ - Fuzzy names in whatever language: "Evee", "Ibui"
+ - IDs: "133", "192", "250"
+ Also:
+ - Type restrictions. "type:psychic" will only return the type. This is
+ how to make ID lookup useful. Multiple type specs can be entered with
+ commas, as "move,item:1". If `valid_types` are provided, any type prefix
+ will be ignored.
+ - Alternate formes can be specified merely like "wash rotom".
+
+ `input`
Name of the thing to look for.
+ `valid_types`
+ A list of table objects or names, e.g., `['pokemon', 'moves']`. If
+ this is provided, only results in one of the given tables will be
+ returned.
+
`session`
A database session to use for retrieving objects. As with get_index,
if this is not provided, a connection to the default database will be
else:
index, speller = open_index()
- name = unicode(name)
-
+ name = normalize(input)
exact = True
-
- # Look for exact name. A Term object does an exact match, so we don't have
- # to worry about a query parser tripping on weird characters in the input
+ form = None
+
+ # Remove any type prefix (pokemon:133) before constructing a query
+ if ':' in name:
+ prefix_chunk, name = name.split(':', 1)
+ name = name.strip()
+
+ if not valid_types:
+ # Only use types from the query string if none were explicitly
+ # provided
+ prefixes = prefix_chunk.split(',')
+ valid_types = [_.strip() for _ in prefixes]
+
+ # Random lookup
+ if name == 'random':
+ return random_lookup(indices=(index, speller),
+ session=session,
+ valid_types=valid_types)
+
+ # Do different things depending what the query looks like
+ # Note: Term objects do an exact match, so we don't have to worry about a
+ # query parser tripping on weird characters in the input
+ if '*' in name or '?' in name:
+ exact_only = True
+ query = whoosh.query.Wildcard(u'name', name)
+ elif rx_is_number.match(name):
+ # Don't spell-check numbers!
+ exact_only = True
+ query = whoosh.query.Term(u'row_id', name)
+ else:
+ # Not an integer
+ query = whoosh.query.Term(u'name', name) \
+ & whoosh.query.Term(u'forme_name', u'XXX')
+
+ # If there's a space in the input, this might be a form
+ if ' ' in name:
+ form, formless_name = name.split(' ', 1)
+ form_query = whoosh.query.Term(u'name', formless_name) \
+ & whoosh.query.Term(u'forme_name', form)
+ query = query | form_query
+
+ ### Filter by type of object
+ type_terms = []
+ for valid_type in valid_types:
+ table_name = _parse_table_name(valid_type)
+ if table_name:
+ # Quietly ignore bogus valid_types; more likely to DTRT
+ type_terms.append(whoosh.query.Term(u'table', table_name))
+
+ if type_terms:
+ query = query & whoosh.query.Or(type_terms)
+
+
+ ### Actual searching
searcher = index.searcher()
- query = whoosh.query.Term('name', name.lower())
- results = searcher.search(query)
+ searcher.weighting = LanguageWeighting() # XXX kosher? docs say search()
+ # takes a weighting kw but it
+ # certainly does not
+ results = searcher.search(query, limit=INTERMEDIATE_LOOKUP_RESULTS)
# Look for some fuzzy matches if necessary
if not exact_only and not results:
exact = False
results = []
- for suggestion in speller.suggest(name, 10):
+ for suggestion in speller.suggest(name, INTERMEDIATE_LOOKUP_RESULTS):
query = whoosh.query.Term('name', suggestion)
results.extend(searcher.search(query))
### Convert results to db objects
- objects = []
- seen = {}
- for result in results:
- # Skip dupe results
- seen_key = result['table'], result['row_id']
- if seen_key in seen:
- continue
- seen[seen_key] = True
+ objects = _whoosh_records_to_results(results, session, exact=exact)
+
+ # Only return up to 10 matches; beyond that, something is wrong.
+ # We strip out duplicate entries above, so it's remotely possible that we
+ # should have more than 10 here and lost a few. The speller returns 25 to
+ # give us some padding, and should avoid that problem. Not a big deal if
+ # we lose the 25th-most-likely match anyway.
+ return objects[:MAX_LOOKUP_RESULTS]
+
- cls = indexed_tables[result['table']]
- obj = session.query(cls).get(result['row_id'])
- objects.append(LookupResult(obj, result['language'], exact))
+def random_lookup(valid_types=[], session=None, indices=None):
+ """Takes similar arguments as `lookup()`, but returns a random lookup
+ result from one of the provided `valid_types`.
+ """
+
+ tables = []
+ for valid_type in valid_types:
+ table_name = _parse_table_name(valid_type)
+ if table_name:
+ tables.append(indexed_tables[table_name])
+
+ if not tables:
+ # n.b.: It's possible we got a list of valid_types and none of them
+ # were valid, but this function is guaranteed to return *something*, so
+ # it politely selects from the entire index isntead
+ tables = indexed_tables.values()
+
+ # Rather than create an array of many hundred items and pick randomly from
+ # it, just pick a number up to the total number of potential items, then
+ # pick randomly from that, and partition the whole range into chunks.
+ # This also avoids the slight problem that the index contains more rows
+ # (for languages) for some items than others.
+ # XXX ought to cache this (in the index?) if possible
+ total = 0
+ partitions = []
+ for table in tables:
+ count = session.query(table).count()
+ total += count
+ partitions.append((table, count))
+
+ n = random.randint(1, total)
+ while n > partitions[0][1]:
+ n -= partitions[0][1]
+ partitions.pop(0)
+
+ return lookup(unicode(n), valid_types=[ partitions[0][0] ],
+ indices=indices, session=session)
+
+def prefix_lookup(prefix, session=None, indices=None):
+ """Returns terms starting with the given exact prefix.
+
+ No special magic is currently done with the name; type prefixes are not
+ recognized.
+
+ `session` and `indices` are treated as with `lookup()`.
+ """
+
+ if not session:
+ session = connect()
+
+ if indices:
+ index, speller = indices
+ else:
+ index, speller = open_index()
+
+ query = whoosh.query.Prefix(u'name', normalize(prefix))
+
+ searcher = index.searcher()
+ searcher.weighting = LanguageWeighting()
+ results = searcher.search(query) # XXX , limit=MAX_LOOKUP_RESULTS)
- return objects
+ return _whoosh_records_to_results(results, session)