blob: be5d052a698c20f4981b4486b4e02eae92ee3ff3 [file] [log] [blame]
# Copyright 2018 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""Build indexing and search."""
import heapq
import logging
import random
import re
from google.appengine.ext import db
from google.appengine.ext import ndb
from protorpc import messages
from components import utils
from proto import common_pb2
import buildtags
import config
import errors
import metrics
import model
import user
MAX_RETURN_BUILDS = 1000
RE_TAG_INDEX_SEARCH_CURSOR = re.compile(r'^id>\d+$')
def fix_max_builds(max_builds):
"""Fixes a page size."""
max_builds = max_builds or 100
if not isinstance(max_builds, int):
raise errors.InvalidInputError('max_builds must be an integer')
if max_builds < 0:
raise errors.InvalidInputError('max_builds must be positive')
return min(MAX_RETURN_BUILDS, max_builds)
@ndb.tasklet
def fetch_page_async(query, page_size, start_cursor, predicate=None):
"""Fetches a page of Build entities."""
assert query
assert isinstance(page_size, int)
assert start_cursor is None or isinstance(start_cursor, basestring)
curs = None
if start_cursor:
try:
curs = ndb.Cursor(urlsafe=start_cursor)
except db.BadValueError as ex:
msg = 'Bad cursor "%s": %s' % (start_cursor, ex)
logging.warning(msg)
raise errors.InvalidInputError(msg)
entities = []
skipped = 0
pages = 0
started = utils.utcnow()
while len(entities) < page_size:
# It is important not to request more than needed in query.fetch_page,
# otherwise the cursor we return to the user skips fetched, but not returned
# entities, and the user will never see them.
to_fetch = page_size - len(entities)
page, curs, more = yield query.fetch_page_async(to_fetch, start_cursor=curs)
pages += 1
for entity in page:
if predicate and not predicate(entity): # pragma: no cover
skipped += 1
continue
entities.append(entity)
if len(entities) >= page_size:
break
if not more:
break
logging.debug(
'fetch_page: fetched %d pages in %dms, skipped %d entities', pages,
(utils.utcnow() - started).total_seconds() * 1000, skipped
)
curs_str = None
if more:
curs_str = curs.urlsafe()
raise ndb.Return(entities, curs_str)
@ndb.tasklet
def check_acls_async(bucket_ids, inc_metric=None):
"""Checks access to the buckets.
Raises an error if the current identity doesn't have access to any of the
buckets.
"""
assert bucket_ids
bucket_ids = sorted(set(bucket_ids))
for bucket_id in bucket_ids:
config.validate_bucket_id(bucket_id)
futs = [user.can_search_builds_async(b) for b in bucket_ids]
for bucket_id, fut in zip(bucket_ids, futs):
if not (yield fut):
if inc_metric: # pragma: no cover
inc_metric.increment(fields={'bucket': bucket_id})
raise user.current_identity_cannot(
'search builds in bucket %s', bucket_id
)
class StatusFilter(messages.Enum):
# A build must have status model.BuildStatus.SCHEDULED.
SCHEDULED = model.BuildStatus.SCHEDULED.number
# A build must have status model.BuildStatus.STARTED.
STARTED = model.BuildStatus.STARTED.number
# A build must have status model.BuildStatus.COMPLETED.
COMPLETED = model.BuildStatus.COMPLETED.number
# A build must have status model.BuildStatus.SCHEDULED or
# model.BuildStatus.STARTED.
INCOMPLETE = 10
class Query(object):
"""Argument for search. Mutable."""
def __init__(
self,
project=None,
bucket_ids=None,
tags=None,
status=None,
result=None,
failure_reason=None,
cancelation_reason=None,
created_by=None,
max_builds=None,
start_cursor=None,
retry_of=None,
canary=None,
create_time_low=None,
create_time_high=None,
build_low=None,
build_high=None,
include_experimental=None
):
"""Initializes Query.
Args:
project (str): project id to search in.
Mutually exclusive with bucket_ids.
bucket_ids (list of str): a list of bucket_ids to search in.
A build must be in one of the buckets.
Mutually exclusive with project.
tags (list of str): a list of tags that a build must have.
All of the |tags| must be present in a build.
status (StatusFilter or common_pb2.Status): build status.
result (model.BuildResult): build result.
failure_reason (model.FailureReason): failure reason.
cancelation_reason (model.CancelationReason): build cancelation reason.
created_by (str): identity who created a build.
max_builds (int): maximum number of builds to return.
start_cursor (string): a value of "next" cursor returned by previous
search_by_tags call. If not None, return next builds in the query.
retry_of (int): value of retry_of attribute.
canary (bool): if not None, value of "canary" field.
Search by canary_preference is not supported.
create_time_low (datetime.datetime): if not None, minimum value of
create_time attribute. Inclusive.
create_time_high (datetime.datetime): if not None, maximum value of
create_time attribute. Exclusive.
build_low (int): if not None, id of the minimum build. Inclusive.
build_high (int): if not None, id of the maximal build. Exclusive.
include_experimental (bool): if true, search results will include
experimental builds. Otherwise, experimental builds will be excluded.
"""
self.project = project
self.bucket_ids = bucket_ids
self.tags = tags
self.status = status
self.result = result
self.failure_reason = failure_reason
self.cancelation_reason = cancelation_reason
self.created_by = created_by
self.retry_of = retry_of
self.canary = canary
self.create_time_low = create_time_low
self.create_time_high = create_time_high
self.build_low = build_low
self.build_high = build_high
self.max_builds = max_builds
self.start_cursor = start_cursor
self.include_experimental = include_experimental
def copy(self):
return Query(**self.__dict__)
def __eq__(self, other): # pragma: no cover
# "pragma: no cover" because this code is executed
# by mock module, not service_test
# pylint: disable=unidiomatic-typecheck
return type(self) == type(other) and self.__dict__ == other.__dict__
def __ne__(self, other): # pragma: no cover
# "pragma: no cover" because this code is executed
# by mock module, not service_test
return not self.__eq__(other)
def __repr__(self):
return repr(self.__dict__)
@property
def status_is_v2(self):
return isinstance(self.status, int)
def validate(self):
"""Raises errors.InvalidInputError if self is invalid."""
assert isinstance(self.status, (type(None), StatusFilter, int)), self.status
assert isinstance(self.bucket_ids, (type(None), list)), self.bucket_ids
assert not (self.bucket_ids and self.project)
buildtags.validate_tags(self.tags, 'search')
create_time_range = (
self.create_time_low is not None or self.create_time_high is not None
)
build_range = self.build_low is not None or self.build_high is not None
if create_time_range and build_range:
raise errors.InvalidInputError(
'create_time_low and create_time_high are mutually exclusive with '
'build_low and build_high'
)
def get_create_time_order_build_id_range(self):
"""Returns low/high build id range for results ordered by creation time.
Low boundary is inclusive. High boundary is exclusive.
Assumes self is valid.
"""
if self.build_low is not None or self.build_high is not None:
return (self.build_low, self.build_high)
else:
return model.build_id_range(self.create_time_low, self.create_time_high)
@ndb.tasklet
def search_async(q):
"""Searches for builds.
Args:
q (Query): the query.
Returns:
A tuple:
builds (list of Build): query result.
next_cursor (string): cursor for the next page.
None if there are no more builds.
Raises:
errors.InvalidInputError if q is invalid.
"""
q.validate()
q = q.copy()
if (q.create_time_low is not None and
q.create_time_low < model.BEGINING_OF_THE_WORLD):
q.create_time_low = None
if q.create_time_high is not None:
if q.create_time_high <= model.BEGINING_OF_THE_WORLD:
raise ndb.Return([], None)
if (q.create_time_low is not None and
q.create_time_low >= q.create_time_high):
raise ndb.Return([], None)
q.tags = q.tags or []
q.max_builds = fix_max_builds(q.max_builds)
q.created_by = user.parse_identity(q.created_by)
q.status = q.status if q.status != common_pb2.STATUS_UNSPECIFIED else None
if not q.bucket_ids and q.retry_of is not None:
retry_of_build = yield model.Build.get_by_id_async(q.retry_of)
if retry_of_build:
q.bucket_ids = [retry_of_build.bucket_id]
if q.bucket_ids:
yield check_acls_async(q.bucket_ids)
q.bucket_ids = set(q.bucket_ids)
is_tag_index_cursor = (
q.start_cursor and RE_TAG_INDEX_SEARCH_CURSOR.match(q.start_cursor)
)
can_use_tag_index = (
indexed_tags(q.tags) and (not q.start_cursor or is_tag_index_cursor)
)
if is_tag_index_cursor and not can_use_tag_index:
raise errors.InvalidInputError('invalid cursor')
can_use_query_search = not q.start_cursor or not is_tag_index_cursor
assert can_use_tag_index or can_use_query_search
# Try searching using tag index.
if can_use_tag_index:
try:
search_start_time = utils.utcnow()
results = yield _tag_index_search_async(q)
logging.info(
'tag index search took %dms',
(utils.utcnow() - search_start_time).total_seconds() * 1000
)
raise ndb.Return(results)
except errors.TagIndexIncomplete:
if not can_use_query_search:
raise
logging.info('falling back to querying')
# Searching using datastore query.
assert can_use_query_search
search_start_time = utils.utcnow()
results = yield _query_search_async(q)
logging.info(
'query search took %dms',
(utils.utcnow() - search_start_time).total_seconds() * 1000
)
raise ndb.Return(results)
def _between(value, low, high): # pragma: no cover
# low is inclusive, high is exclusive
if low is not None and value < low:
return False
if high is not None and value >= high:
return False
return True
@ndb.tasklet
def _query_search_async(q):
"""Searches for builds using NDB query. For args doc, see search().
Assumes:
- q is valid
- if bool(bucket_ids), permissions are checked.
"""
if not q.bucket_ids:
q.bucket_ids = yield user.get_accessible_buckets_async()
if q.bucket_ids is None:
# User has access to all buckets.
pass
else:
if q.project:
def get_project_id(bucket_id):
project_id, _ = config.parse_bucket_id(bucket_id)
return project_id
# Note: get_accessible_buckets_async is memcached per user for 10m.
q.bucket_id = {
bid for bid in q.bucket_ids if get_project_id(bid) == q.project
}
if not q.bucket_ids:
raise ndb.Return([], None)
# (q.bucket_ids is None) means the requester has access to all buckets.
assert q.bucket_ids is None or q.bucket_ids
check_buckets_locally = q.retry_of is not None
dq = model.Build.query()
for t in q.tags:
dq = dq.filter(model.Build.tags == t)
filter_if = lambda p, v: dq if v is None else dq.filter(p == v)
expected_statuses_v1 = None
if q.status_is_v2:
dq = dq.filter(model.Build.status_v2 == q.status)
elif q.status == StatusFilter.INCOMPLETE:
expected_statuses_v1 = (
model.BuildStatus.SCHEDULED, model.BuildStatus.STARTED
)
dq = dq.filter(model.Build.incomplete == True)
elif q.status is not None:
s = model.BuildStatus.lookup_by_number(q.status.number)
expected_statuses_v1 = (s,)
dq = dq.filter(model.Build.status == s)
dq = filter_if(model.Build.result, q.result)
dq = filter_if(model.Build.failure_reason, q.failure_reason)
dq = filter_if(model.Build.cancelation_reason, q.cancelation_reason)
dq = filter_if(model.Build.created_by, q.created_by)
dq = filter_if(model.Build.retry_of, q.retry_of)
dq = filter_if(model.Build.canary, q.canary)
if not q.include_experimental:
dq = dq.filter(model.Build.experimental == False)
if q.bucket_ids and not check_buckets_locally:
dq = dq.filter(model.Build.bucket_id.IN(q.bucket_ids))
id_low, id_high = q.get_create_time_order_build_id_range()
if id_low is not None:
dq = dq.filter(model.Build.key >= ndb.Key(model.Build, id_low))
if id_high is not None:
dq = dq.filter(model.Build.key < ndb.Key(model.Build, id_high))
dq = dq.order(model.Build.key)
def local_predicate(build):
if q.status_is_v2:
if build.status_v2 != q.status: # pragma: no cover
return False
elif expected_statuses_v1 and build.status not in expected_statuses_v1:
return False # pragma: no cover
if q.bucket_ids and build.bucket_id not in q.bucket_ids:
return False
if not _between(build.create_time, q.create_time_low, q.create_time_high):
return False # pragma: no cover
return True
raise ndb.Return((
yield fetch_page_async(
dq, q.max_builds, q.start_cursor, predicate=local_predicate
)
))
@ndb.non_transactional
@ndb.tasklet
def _populate_tag_index_entry_bucket_id(indexes):
"""Populates indexes[i].entries[j].bucket_id."""
to_migrate = {
i for i, idx in enumerate(indexes)
if any(not e.bucket_id for e in idx.entries)
}
if not to_migrate:
return
build_ids = sorted({
e.build_id
for i in to_migrate
for e in indexes[i].entries
if not e.bucket_id
})
builds = yield ndb.get_multi_async(
ndb.Key(model.Build, bid) for bid in build_ids
)
bucket_ids = {
build_id: build.bucket_id if build else None
for build_id, build in zip(build_ids, builds)
}
@ndb.transactional_tasklet
def txn_async(key):
idx = yield key.get_async()
new_entries = []
for e in idx.entries:
e.bucket_id = e.bucket_id or bucket_ids[e.build_id]
if e.bucket_id:
new_entries.append(e)
else: # pragma: no cover | pycoverage is confused
# Such build does not exist.
# Note: add_to_tag_index_async adds new entries with bucket_id.
# This code runs only for old TagIndeEntries, so there is no race.
pass
idx.entries = new_entries
yield idx.put_async()
raise ndb.Return(idx)
futs = [(i, txn_async(indexes[i].key)) for i in to_migrate]
for i, fut in futs:
indexes[i] = fut.get_result()
@ndb.tasklet
def _tag_index_search_async(q):
"""Searches for builds using TagIndex entities. For args doc, see search().
Assumes:
- arguments are valid
- if bool(q.bucket_ids), permissions are checked.
Raises:
errors.TagIndexIncomplete if the tag index is complete and cannot be used.
"""
assert q.tags
assert not q.bucket_ids or isinstance(q.bucket_ids, set)
# Choose a tag to search by.
all_indexed_tags = indexed_tags(q.tags)
assert all_indexed_tags
indexed_tag = all_indexed_tags[0] # choose the most selective tag.
indexed_tag_key = buildtags.parse(indexed_tag)[0]
# Exclude the indexed tag from the tag filter.
q = q.copy()
q.tags = q.tags[:]
q.tags.remove(indexed_tag)
# Determine build id range we are considering.
# id_low is inclusive, id_high is exclusive.
id_low, id_high = q.get_create_time_order_build_id_range()
id_low = 0 if id_low is None else id_low
id_high = (1 << 64) - 1 if id_high is None else id_high
if q.start_cursor:
# The cursor is a minimum build id, exclusive. Such cursor is resilient
# to duplicates and additions of index entries to beginning or end.
assert RE_TAG_INDEX_SEARCH_CURSOR.match(q.start_cursor)
min_id_exclusive = int(q.start_cursor[len('id>'):])
id_low = max(id_low, min_id_exclusive + 1)
if id_low >= id_high:
raise ndb.Return([], None)
# Load index entries and put them to a min-heap, sorted by build_id.
entry_heap = [] # tuples (build_id, TagIndexEntry).
indexes = yield ndb.get_multi_async(TagIndex.all_shard_keys(indexed_tag))
indexes = [idx for idx in indexes if idx]
yield _populate_tag_index_entry_bucket_id(indexes)
for idx in indexes:
if idx.permanently_incomplete:
raise errors.TagIndexIncomplete(
'TagIndex(%s) is incomplete' % idx.key.id()
)
for e in idx.entries:
if id_low <= e.build_id < id_high:
entry_heap.append((e.build_id, e))
if not entry_heap:
raise ndb.Return([], None)
heapq.heapify(entry_heap)
# If buckets were not specified explicitly, permissions were not checked
# earlier. In this case, check permissions for each build.
check_permissions = not q.bucket_ids
has_access_cache = {}
# scalar_filters maps a name of a model.Build attribute to a filter value.
# Applies only to non-repeated fields.
scalar_filters = [
('result', q.result),
('failure_reason', q.failure_reason),
('cancelation_reason', q.cancelation_reason),
('created_by', q.created_by),
('retry_of', q.retry_of),
('canary', q.canary),
# TODO(crbug.com/851036): use e.bucket_id to filter by project before
# fetching a build.
('project', q.project),
]
scalar_filters = [(a, v) for a, v in scalar_filters if v is not None]
if q.status_is_v2:
scalar_filters.append(('status_v2', q.status))
elif q.status == StatusFilter.INCOMPLETE:
scalar_filters.append(('incomplete', True))
elif q.status is not None:
scalar_filters.append(
('status', model.BuildStatus.lookup_by_number(q.status.number))
)
# Find the builds.
result = [] # ordered by build id by ascending.
last_considered_entry = None
skipped_entries = 0
inconsistent_entries = 0
eof = False
while len(result) < q.max_builds:
fetch_count = q.max_builds - len(result)
entries_to_fetch = [] # ordered by build id by ascending.
while entry_heap:
_, e = heapq.heappop(entry_heap)
prev = last_considered_entry
last_considered_entry = e
if prev and prev.build_id == e.build_id:
# Tolerate duplicates.
continue
# If we filter by bucket, check it here without fetching the build.
# This is not a security check.
if q.bucket_ids and e.bucket_id not in q.bucket_ids:
continue
if check_permissions:
has = has_access_cache.get(e.bucket_id)
if has is None:
has = yield user.can_search_builds_async(e.bucket_id)
has_access_cache[e.bucket_id] = has
if not has:
continue
entries_to_fetch.append(e)
if len(entries_to_fetch) >= fetch_count:
break
if not entries_to_fetch:
eof = True
break
builds = yield ndb.get_multi_async(
ndb.Key(model.Build, e.build_id) for e in entries_to_fetch
)
for e, b in zip(entries_to_fetch, builds):
# Check for inconsistent entries.
if not (b and b.bucket_id == e.bucket_id and indexed_tag in b.tags):
logging.warning('entry with build_id %d is inconsistent', e.build_id)
inconsistent_entries += 1
continue
# Check user-supplied filters.
if any(getattr(b, a) != v for a, v in scalar_filters):
skipped_entries += 1
continue
if not _between(b.create_time, q.create_time_low, q.create_time_high):
continue # pragma: no cover
if any(t not in b.tags for t in q.tags):
skipped_entries += 1
continue
if b.experimental and not q.include_experimental:
continue
result.append(b)
metrics.TAG_INDEX_SEARCH_SKIPPED_BUILDS.add(
skipped_entries, fields={'tag': indexed_tag_key}
)
metrics.TAG_INDEX_INCONSISTENT_ENTRIES.add(
inconsistent_entries, fields={'tag': indexed_tag_key}
)
# Return the results.
next_cursor = None
if not eof and last_considered_entry:
next_cursor = 'id>%d' % last_considered_entry.build_id
raise ndb.Return(result, next_cursor)
def indexed_tags(tags):
"""Returns a list of tags that must be indexed.
The order of returned tags is from more selective to less selective.
"""
if not tags:
return []
return sorted(
set(t for t in tags if t.startswith(('buildset:', 'build_address:')))
)
def update_tag_indexes_async(builds):
"""Updates tag indexes for the builds.
For each new build, for each indexed tag, add an entry to a tag index.
"""
index_entries = {}
for b in builds:
for t in set(indexed_tags(b.tags)):
index_entries.setdefault(t, []).append(
TagIndexEntry(
build_id=b.key.id(),
bucket_id=b.bucket_id,
)
)
return [
add_to_tag_index_async(tag, entries)
for tag, entries in index_entries.iteritems()
]
def add_to_tag_index_async(tag, new_entries):
"""Adds index entries to the tag index.
new_entries must be a list of TagIndexEntry and not have duplicate builds ids.
"""
if not new_entries: # pragma: no cover
return
build_ids = {e.build_id for e in new_entries}
assert len(build_ids) == len(new_entries), new_entries
@ndb.transactional_tasklet
def txn_async():
idx_key = TagIndex.random_shard_key(tag)
idx = (yield idx_key.get_async()) or TagIndex(key=idx_key)
if idx.permanently_incomplete:
return
# Avoid going beyond 1Mb entity size limit by limiting the number of entries
new_size = len(idx.entries) + len(new_entries)
if new_size > TagIndex.MAX_ENTRY_COUNT:
idx.permanently_incomplete = True
idx.entries = []
else:
logging.debug(
'adding %d entries to TagIndex(%s)', len(new_entries), idx_key.id()
)
idx.entries.extend(new_entries)
yield idx.put_async()
return txn_async()
class TagIndexEntry(ndb.Model):
"""A single entry in a TagIndex, references a build."""
created_time = ndb.DateTimeProperty(auto_now_add=True)
# ID of the build.
build_id = ndb.IntegerProperty(indexed=False)
# Bucket id of the build.
# Same format as model.Build.bucket_id.
bucket_id = ndb.StringProperty(indexed=False)
class TagIndex(ndb.Model):
"""A custom index of builds by a tag.
Entity key:
Entity id has format "<tag_key>:<tag_value>" or
":<shard_index>:<tag_key>:<tag_value>" for positive values of shard_index.
TagIndex has no parent.
"""
MAX_ENTRY_COUNT = 1000
SHARD_COUNT = 16 # This value MUST NOT decrease.
# if incomplete, this TagIndex should not be used in search.
# It is set to True if there are more than MAX_ENTRY_COUNT builds
# for this tag.
permanently_incomplete = ndb.BooleanProperty(indexed=False)
# entries is a superset of all builds that have the tag equal to the id of
# this entity. It may contain references to non-existent builds or builds that
# do not actually have this tag; such builds must be ignored.
entries = ndb.LocalStructuredProperty(
TagIndexEntry, repeated=True, indexed=False
)
@classmethod
def make_key(cls, shard_index, tag):
"""Returns a TagIndex entity key."""
assert shard_index >= 0
assert not tag.startswith(':')
iid = tag if shard_index == 0 else ':%d:%s' % (shard_index, tag)
return ndb.Key(TagIndex, iid)
@classmethod
def all_shard_keys(cls, tag): # pragma: no cover
return [cls.make_key(i, tag) for i in xrange(cls.SHARD_COUNT)]
@classmethod
def random_shard_key(cls, tag):
"""Returns a TagIndex entity key of a random shard."""
return cls.make_key(cls.random_shard_index(), tag)
@classmethod
def random_shard_index(cls): # pragma: no cover
return random.randint(0, cls.SHARD_COUNT - 1)