"""
Wrappers for lucene Fields and Documents.
"""
from future_builtins import map
import datetime, calendar
import operator
import warnings
import lucene
from java.lang import Double, Float, Long, Number, Object
from java.util import Arrays, HashSet
from org.apache.lucene import document, index, search, util
from org.apache.lucene.search import grouping
from .queries import Query
[docs]class Field(document.FieldType):
"""Saved parameters which can generate lucene Fields given values.
:param name: name of field
:param boost: boost factor
:param stored, indexed, settings: lucene FieldType attributes
"""
attrs = set(name[3].lower() + name[4:] for name in dir(document.FieldType) if name.startswith('set'))
def __init__(self, name, stored=False, indexed=True, boost=1.0, **settings):
document.FieldType.__init__(self)
self.name, self.boost = name, boost
if set(['store', 'index', 'termvector']).intersection(settings):
warnings.warn('lucene Field.{Store,Index,TermVector} deprecated, use FieldType attibutes', DeprecationWarning)
store = settings.get('store', False)
if isinstance(store, bool):
store = 'yes' if store else 'no'
store = document.Field.Store.valueOf(store.upper())
index = settings.get('index', 'analyzed')
if isinstance(index, bool):
index = document.Field.Index.toIndex(index, settings.get('analyzed', False), settings.get('omitNorms', False))
else:
index = document.Field.Index.valueOf(index.upper())
termvector = settings.get('termvector', False)
if isinstance(termvector, bool):
termvector = document.Field.TermVector.toTermVector(termvector, settings.get('withOffsets', False), settings.get('withPositions', False))
else:
termvector = document.Field.TermVector.valueOf(termvector.upper())
ft = document.Field.translateFieldType(store, index, termvector)
self.update(**Field.settings.fget(ft))
else:
self.update(stored=stored, indexed=indexed, **settings)
def update(self, docValueType='', indexOptions='', numericType='', **settings):
if docValueType:
self.setDocValueType(getattr(index.FieldInfo.DocValuesType, docValueType.upper()))
if indexOptions:
self.setIndexOptions(getattr(index.FieldInfo.IndexOptions, indexOptions.upper()))
if numericType:
self.setNumericType(getattr(document.FieldType.NumericType, numericType.upper()))
for name in settings:
getattr(self, 'set' + name[:1].upper() + name[1:])(settings[name])
@property
[docs] def settings(self):
"dict representation of settings"
defaults = document.FieldType()
result = {'indexed': self.indexed()}
for name in Field.attrs:
value = getattr(self, name)()
if value != getattr(defaults, name)():
result[name] = value if isinstance(value, int) else str(value)
return result
[docs] def items(self, *values):
"Generate lucene Fields suitable for adding to a document."
for value in values:
field = document.Field(self.name, value, self)
field.setBoost(self.boost)
yield field
[docs]class MapField(Field):
"""Field which applies a function across its values.
:param func: callable
"""
def __init__(self, name, func, **kwargs):
Field.__init__(self, name, **kwargs)
self.func = func
[docs] def items(self, *values):
"Generate fields with mapped values."
return Field.items(self, *map(self.func, values))
[docs]class NestedField(Field):
"""Field which indexes every component into its own field.
Original value may be stored for convenience.
:param sep: field separator used on name and values
"""
def __init__(self, name, sep='.', tokenized=False, **kwargs):
if set(['store', 'index', 'termvector']).intersection(kwargs):
kwargs.setdefault('index', True)
Field.__init__(self, name, tokenized=tokenized, **kwargs)
self.sep = sep
self.names = tuple(self.values(name))
[docs] def values(self, value):
"Generate component field values in order."
value = value.split(self.sep)
for index in range(1, len(value) + 1):
yield self.sep.join(value[:index])
[docs] def items(self, *values):
"Generate indexed component fields."
for value in values:
for index, text in enumerate(self.values(value)):
yield document.Field(self.names[index], text, self)
[docs] def prefix(self, value):
"Return prefix query of the closest possible prefixed field."
index = value.count(self.sep)
return Query.prefix(self.names[index], value)
[docs] def range(self, start, stop, lower=True, upper=False):
"Return range query of the closest possible prefixed field."
index = max(value.count(self.sep) for value in (start, stop) if value is not None)
return Query.range(self.names[index], start, stop, lower, upper)
[docs]class NumericField(Field):
"""Field which indexes numbers in a prefix tree.
:param name: name of field
:param type: optional int, float, or lucene NumericType string
"""
def __init__(self, name, type=None, step=0, stored=False, tokenized=False, **kwargs):
if type:
kwargs['numericType'] = {int: 'long', float: 'double'}.get(type, str(type))
Field.__init__(self, name, stored=stored, tokenized=tokenized, **kwargs)
if step:
warnings.warn('step deprecated, use numericPrecisionStep', DeprecationWarning)
self.setNumericPrecisionStep(step)
[docs] def items(self, *values):
"Generate lucene NumericFields suitable for adding to a document."
if not self.numericType():
cls, = set(map(type, values))
self.update(numericType='double' if issubclass(cls, float) else 'long')
for value in values:
if isinstance(value, float):
yield document.DoubleField(self.name, value, self)
else:
yield document.LongField(self.name, long(value), self)
def numeric(self, cls, start, stop, lower, upper):
if isinstance(start, float) or isinstance(stop, float):
start, stop = (value if value is None else Double(value) for value in (start, stop))
return cls.newDoubleRange(self.name, self.numericPrecisionStep(), start, stop, lower, upper)
if start is not None:
start = None if start < Long.MIN_VALUE else Long(long(start))
if stop is not None:
stop = None if stop > Long.MAX_VALUE else Long(long(stop))
return cls.newLongRange(self.name, self.numericPrecisionStep(), start, stop, lower, upper)
[docs] def range(self, start, stop, lower=True, upper=False):
"Return lucene NumericRangeQuery."
return self.numeric(search.NumericRangeQuery, start, stop, lower, upper)
[docs] def term(self, value):
"Return range query to match single term."
return self.range(value, value, upper=True)
[docs] def filter(self, start, stop, lower=True, upper=False):
"Return lucene NumericRangeFilter."
return self.numeric(search.NumericRangeFilter, start, stop, lower, upper)
[docs]class DateTimeField(NumericField):
"""Field which indexes datetimes as a NumericField of timestamps.
Supports datetimes, dates, and any prefix of time tuples.
"""
def __init__(self, name, **kwargs):
NumericField.__init__(self, name, type=float, **kwargs)
[docs] def timestamp(self, date):
"Return utc timestamp from date or time tuple."
if isinstance(date, datetime.date):
return calendar.timegm(date.timetuple()) + getattr(date, 'microsecond', 0) * 1e-6
return float(calendar.timegm(tuple(date) + (None, 1, 1, 0, 0, 0)[len(date):]))
[docs] def items(self, *dates):
"Generate lucene NumericFields of timestamps."
return NumericField.items(self, *map(self.timestamp, dates))
[docs] def range(self, start, stop, lower=True, upper=False):
"Return NumericRangeQuery of timestamps."
start, stop = (date and self.timestamp(date) for date in (start, stop))
return NumericField.range(self, start, stop, lower, upper)
[docs] def prefix(self, date):
"Return range query which matches the date prefix."
if isinstance(date, datetime.date):
date = date.timetuple()[:6 if isinstance(date, datetime.datetime) else 3]
if len(date) == 2 and date[1] == 12: # month must be valid
return self.range(date, (date[0]+1, 1))
return self.range(date, tuple(date[:-1]) + (date[-1]+1,))
[docs] def duration(self, date, days=0, **delta):
"""Return date range query within time span of date.
:param date: origin date or tuple
:param days,delta: timedelta parameters
"""
if not isinstance(date, datetime.date):
date = datetime.datetime(*(tuple(date) + (None, 1, 1)[len(date):]))
delta = datetime.timedelta(days, **delta)
return self.range(*sorted([date, date + delta]), upper=True)
[docs] def within(self, days=0, weeks=0, utc=True, **delta):
"""Return date range query within current time and delta.
If the delta is an exact number of days, then dates will be used.
:param days,weeks: number of days to offset from today
:param utc: optionally use utc instead of local time
:param delta: additional timedelta parameters
"""
date = datetime.datetime.utcnow() if utc else datetime.datetime.now()
if not (isinstance(days + weeks, float) or delta):
date = date.date()
return self.duration(date, days, weeks=weeks, **delta)
[docs]class Document(dict):
"Multimapping of field names to values, but default getters return the first value."
def __init__(self, doc):
for field in doc.getFields():
value = convert(field.numericValue() or field.stringValue() or field.binaryValue())
self.setdefault(field.name(), []).append(value)
def __getitem__(self, name):
return dict.__getitem__(self, name)[0]
def get(self, name, default=None):
return dict.get(self, name, [default])[0]
[docs] def getlist(self, name):
"Return list of all values for given field."
return dict.get(self, name, [])
[docs] def dict(self, *names, **defaults):
"""Return dict representation of document.
:param names: names of multi-valued fields to return as a list
:param defaults: include only given fields, using default values as necessary
"""
defaults.update((name, self[name]) for name in (defaults or self) if name in self)
defaults.update((name, self.getlist(name)) for name in names)
return defaults
def convert(value):
"Return python object from java Object."
if util.BytesRef.instance_(value):
return util.BytesRef.cast_(value).utf8ToString()
if not Number.instance_(value):
return value.toString() if Object.instance_(value) else value
value = Number.cast_(value)
return value.doubleValue() if Float.instance_(value) or Double.instance_(value) else int(value.longValue())
[docs]class Hit(Document):
"A Document from a search result, with :attr:`id`, :attr:`score`, and optional sort :attr:`keys`."
def __init__(self, doc, id, score, keys=()):
Document.__init__(self, doc)
self.id, self.score = id, score
self.keys = tuple(map(convert, keys))
[docs] def dict(self, *names, **defaults):
"Return dict representation of document with __id__, __score__, and any sort __keys__."
result = Document.dict(self, *names, **defaults)
result.update(__id__=self.id, __score__=self.score)
if self.keys:
result['__keys__'] = self.keys
return result
[docs]class Hits(object):
"""Search results: lazily evaluated and memory efficient.
Provides a read-only sequence interface to hit objects.
:param searcher: `IndexSearcher`_ which can retrieve documents
:param scoredocs: lucene ScoreDocs
:param count: total number of hits
:param maxscore: maximum score
:param fields: optional field selectors
"""
def __init__(self, searcher, scoredocs, count=None, maxscore=None, fields=None):
self.searcher, self.scoredocs = searcher, scoredocs
self.count, self.maxscore = count, maxscore
self.fields = fields
[docs] def select(self, *fields):
"Only load selected fields."
self.fields = HashSet(Arrays.asList(fields))
[docs] def __len__(self):
return len(self.scoredocs)
[docs] def __getitem__(self, index):
if isinstance(index, slice):
start, stop, step = index.indices(len(self))
assert step == 1, 'slice step is not supported'
scoredocs = self.scoredocs[start:stop] if stop - start < len(self) else self.scoredocs
return type(self)(self.searcher, scoredocs, self.count, self.maxscore, self.fields)
scoredoc = self.scoredocs[index]
keys = search.FieldDoc.cast_(scoredoc).fields if search.FieldDoc.instance_(scoredoc) else ()
doc = self.searcher.doc(scoredoc.doc, self.fields)
return Hit(doc, scoredoc.doc, scoredoc.score, keys)
@property
def ids(self):
return map(operator.attrgetter('doc'), self.scoredocs)
@property
def scores(self):
return map(operator.attrgetter('score'), self.scoredocs)
[docs] def items(self):
"Generate zipped ids and scores."
return map(operator.attrgetter('doc', 'score'), self.scoredocs)
[docs] def groupby(self, func):
"Return ordered list of `Hits`_ grouped by value of function applied to doc ids."
groups = {}
for scoredoc in self.scoredocs:
value = func(scoredoc.doc)
try:
group = groups[value]
except KeyError:
group = groups[value] = type(self)(self.searcher, [], fields=self.fields)
group.index, group.value = len(groups), value
group.scoredocs.append(scoredoc)
return sorted(groups.values(), key=lambda group: group.__dict__.pop('index'))
[docs] def filter(self, func):
"Return `Hits`_ filtered by function applied to doc ids."
scoredocs = [scoredoc for scoredoc in self.scoredocs if func(scoredoc.doc)]
return type(self)(self.searcher, scoredocs, fields=self.fields)
[docs] def sorted(self, key, reverse=False):
"Return `Hits`_ sorted by key function applied to doc ids."
scoredocs = sorted(self.scoredocs, key=lambda scoredoc: key(scoredoc.doc), reverse=reverse)
return type(self)(self.searcher, scoredocs, self.count, self.maxscore, self.fields)
[docs]class Grouping(object):
"""Delegated lucene SearchGroups with optimized faceting.
:param searcher: `IndexSearcher`_ which can retrieve documents
:param field: unique field name to group by
:param query: lucene Query to select groups
:param count: maximum number of groups
:param sort: lucene Sort to order groups
"""
def __init__(self, searcher, field, query=None, count=None, sort=None):
warnings.warn('Grouping deprecated, use GroupingSearch instead', DeprecationWarning)
self.searcher, self.field = searcher, field
self.query = query or search.MatchAllDocsQuery()
self.sort = sort or search.Sort.RELEVANCE
if count is None:
collector = grouping.term.TermAllGroupsCollector(field)
search.IndexSearcher.search(self.searcher, self.query, collector)
count = collector.groupCount
collector = grouping.term.TermFirstPassGroupingCollector(field, self.sort, count)
search.IndexSearcher.search(self.searcher, self.query, collector)
self.searchgroups = collector.getTopGroups(0, False).of_(grouping.SearchGroup)
[docs] def __len__(self):
return self.searchgroups.size()
[docs] def __iter__(self):
for searchgroup in self.searchgroups:
yield convert(searchgroup.groupValue)
[docs] def facets(self, filter):
"Generate field values and counts which match given filter."
collector = grouping.term.TermSecondPassGroupingCollector(self.field, self.searchgroups, self.sort, self.sort, 1, False, False, False)
search.IndexSearcher.search(self.searcher, self.query, filter, collector)
for groupdocs in collector.getTopGroups(0).groups:
yield convert(groupdocs.groupValue), groupdocs.totalHits
[docs] def groups(self, count=1, sort=None, scores=False, maxscore=False):
"""Generate grouped `Hits`_ from second pass grouping collector.
:param count: maximum number of docs per group
:param sort: lucene Sort to order docs within group
:param scores: compute scores for candidate results
:param maxscore: compute maximum score of all results
"""
sort = sort or self.sort
if sort == search.Sort.RELEVANCE:
scores = maxscore = True
collector = grouping.term.TermSecondPassGroupingCollector(self.field, self.searchgroups, self.sort, sort, count, scores, maxscore, True)
search.IndexSearcher.search(self.searcher, self.query, collector)
for groupdocs in collector.getTopGroups(0).groups:
hits = Hits(self.searcher, groupdocs.scoreDocs, groupdocs.totalHits, groupdocs.maxScore, getattr(self, 'fields', None))
hits.value = convert(groupdocs.groupValue)
yield hits
[docs]class GroupingSearch(grouping.GroupingSearch):
"""Inherited lucene GroupingSearch with optimized faceting.
:param field: unique field name to group by
:param sort: lucene Sort to order groups and docs
:param cache: use unlimited caching
:param attrs: additional attributes to set
"""
select = Hits.__dict__['select']
def __init__(self, field, sort=None, cache=True, **attrs):
grouping.GroupingSearch.__init__(self, field)
self.field = field
if sort:
self.groupSort = self.sortWithinGroup = sort
self.fillSortFields = True
if cache:
self.setCachingInMB(float('inf'), True)
for name in attrs:
getattr(type(self), name).__set__(self, attrs[name])
[docs] def __len__(self):
return self.allMatchingGroups.size()
[docs] def __iter__(self):
return map(convert, self.allMatchingGroups)
def search(self, searcher, query, filter, count):
if count is None:
count = sum(search.FieldCache.DEFAULT.getTermsIndex(reader, self.field).valueCount for reader in searcher.readers)
return grouping.GroupingSearch.search(self, searcher, filter, query, 0, count).groups
[docs] def facets(self, searcher, filter=None):
"Generate field values and counts which match given filter."
for groupdocs in self.search(searcher, search.MatchAllDocsQuery(), filter, None):
yield convert(groupdocs.groupValue), groupdocs.totalHits
[docs] def groups(self, searcher, query, filter=None, count=None):
"Generate grouped `Hits`_ from search parameters."
for groupdocs in self.search(searcher, query, filter, count):
hits = Hits(searcher, groupdocs.scoreDocs, groupdocs.totalHits, groupdocs.maxScore, getattr(self, 'fields', None))
hits.value = convert(groupdocs.groupValue)
yield hits