Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(599)

Unified Diff: filter/txnBuf/query_merger.go

Issue 1309803004: Add transaction buffer filter. (Closed) Base URL: https://github.com/luci/gae.git@add_query_support
Patch Set: rebase Created 5 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: filter/txnBuf/query_merger.go
diff --git a/filter/txnBuf/query_merger.go b/filter/txnBuf/query_merger.go
new file mode 100644
index 0000000000000000000000000000000000000000..1e4ff44ccf92891208ee46d27eea9307c3ff9583
--- /dev/null
+++ b/filter/txnBuf/query_merger.go
@@ -0,0 +1,295 @@
+// Copyright 2015 The Chromium Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style license that can be
+// found in the LICENSE file.
+
+package txnBuf
+
+import (
+ "bytes"
+ "sort"
+ "sync"
+
+ "github.com/luci/gae/impl/memory"
+ ds "github.com/luci/gae/service/datastore"
+ "github.com/luci/gae/service/datastore/serialize"
+ "github.com/luci/luci-go/common/stringset"
+)
+
+// queryToIter takes a FinalizedQuery and returns an iterator function which
+// will produce either *items or errors.
+//
+// - d is the raw datastore to run this query on
+// - dedup is a function which will return true if the given key has already
+// been used in the merged result set. If nil, then this check is skipped.
+// - filter is a function which will return true if the given key appears
+// in the in-memory result set. This means that all entries for this key
+// will show up from the in-memory query and don't need to be merged. This
+// also prevents deleted entities from showing up in a query.
+func queryToIter(stopChan chan struct{}, fq *ds.FinalizedQuery, d ds.RawInterface, dedup, filter func(string) bool) func() (*item, error) {
+ c := make(chan *item)
+
+ go func() {
+ defer close(c)
+
+ err := d.Run(fq, func(k *ds.Key, pm ds.PropertyMap, _ ds.CursorCB) (keepGoing bool) {
+ i := &item{key: k, data: pm}
+ encKey := i.getEncKey()
+ if filter != nil && filter(encKey) {
+ return true
+ }
+ if dedup != nil && dedup(encKey) {
Vadim Sh. 2015/09/30 01:11:28 why have both dedup & filter if they are used iden
iannucci 2015/09/30 02:00:10 Hm... ok.
+ return true
+ }
+
+ select {
+ case c <- i:
+ return true
+ case <-stopChan:
+ return false
+ }
+ })
+ if err != nil {
+ c <- &item{err: err}
+ }
+ }()
+
+ return func() (*item, error) {
+ for {
+ itm := <-c
+ if itm == nil {
+ return nil, nil
+ }
+ if itm.err != nil {
Vadim Sh. 2015/09/30 01:11:28 not sure it matters, but iteration function return
iannucci 2015/09/30 02:00:10 doesn't matter, we bail on the first error.
+ return nil, itm.err
+ }
+ return itm, nil
+ }
+ }
+}
+
+// adjustQuery applies various mutations to the query to make it suitable for
+// merging. In general, this removes limits and offsets the 'distinct' modifier,
+// and it ensures that if there are sort orders which won't appear in the
+// result data that the query is transformed into a projection query which
+// contains all of the data. A non-projection query will never be transformed
+// in this way.
+func adjustQuery(fq *ds.FinalizedQuery) (*ds.FinalizedQuery, error) {
+ q := fq.Original()
+
+ // The limit and offset must be done in-memory because otherwise we may
+ // request too few entities from the underlying store if many matching
+ // entities have been deleted in the buffered transaction.
+ q = q.Limit(-1)
+ q = q.Offset(-1)
+
+ // distinction must be done in-memory, because otherwise there's no way
+ // to merge in the effect of the in-flight changes (because there's no way
+ // to push back to the datastore "yeah, I know you told me that the (1, 2)
+ // result came from `/Bob,1`, but would you mind pretending that it didn't
+ // and tell me next the one instead?
+ if fq.Distinct() {
+ q = q.Distinct(false)
+ }
+
+ // since we need to merge results, we must have all order-related fields
+ // in each result. The only time we wouldn't have all the data available would
+ // be for a keys-only or projection query. To fix this, we convert all
+ // Projection and KeysOnly queries to project on /all/ Orders.
+ //
+ // FinalizedQuery already guarantees that all projected fields show up in
+ // the Orders, but the projected fields could be a subset of the orders.
+ //
+ // Additionally on a keys-only query, any orders other than __key__ require
+ // conversion of this query to a projection query including those orders in
+ // order to merge the results correctly.
+ //
+ // In both cases, the resulting objects returned to the higher layers of
+ // the stack will only include the information requested by the user; keysonly
+ // queries will discard all PropertyMap data, and projection queries will
+ // discard any field data that the user didn't ask for.
+ orders := fq.Orders()
+ if len(fq.Project()) > 0 || (fq.KeysOnly() && len(orders) > 1) {
+ q = q.KeysOnly(false)
+
+ for _, o := range orders {
+ if o.Property == "__key__" {
+ continue
+ }
+ q = q.Project(o.Property)
+ }
+ }
+
+ return q.Finalize()
+}
+
+// runMergedQueries executes a user query `fq` against the parent datastore as
+// well as the in-memory datastore, calling `cb` with the merged result set.
+//
+// It's expected that the caller of this function will apply limit and offset
+// if the query contains those restrictions. This may convert the query to
+// an expanded projection query with more data than the user asked for. It's the
+// caller's responsibility to prune away the extra data.
+//
+// See also `dsTxnBuf.Run()`.
+func runMergedQueries(fq *ds.FinalizedQuery, state *txnBufState, cb func(k *ds.Key, data ds.PropertyMap) bool) error {
+ toRun, err := adjustQuery(fq)
+ if err != nil {
+ return err
+ }
+
+ stopChan := make(chan struct{})
+ defer close(stopChan)
+
+ cmpLower, cmpUpper := memory.GetBinaryBounds(fq)
+ cmpOrder := fq.Orders()
+ cmpFn := func(i *item) string {
+ return i.getCmpRow(cmpLower, cmpUpper, cmpOrder)
+ }
+
+ dedup := stringset.Set(nil)
+ distinct := stringset.Set(nil)
+ distinctOrder := []ds.IndexColumn(nil)
+ if len(fq.Project()) > 0 { // the original query was a projection query
+ if fq.Distinct() {
+ // it was a distinct projection query, so we need to dedup by distinct
+ // options.
+ distinct = stringset.New(0)
+ proj := fq.Project()
+ distinctOrder = make([]ds.IndexColumn, len(proj))
+ for i, p := range proj {
+ distinctOrder[i].Property = p
+ }
+ }
+ } else {
+ // the original was a normal or keysonly query, so we need to dedup by keys
+ dedup = stringset.New(0)
+ }
+
+ // need lock around dedup since it's not threadsafe and we read/write it in
+ // different goroutines. No lock is needed around state.entState.has because
+ // the whole transaction is locked during this query and so the entState is
+ // effectively read-only.
+ dedupLock := sync.Mutex{}
+ dedupFn := (func(string) bool)(nil)
+ if dedup != nil {
+ dedupFn = func(val string) bool {
+ dedupLock.Lock()
+ defer dedupLock.Unlock()
+ return dedup.Has(val)
+ }
+ }
+
+ parItemGet := queryToIter(stopChan, toRun, state.parentDS, dedupFn, state.entState.has)
+ memItemGet := queryToIter(stopChan, toRun, state.memDS, dedupFn, nil)
+
+ pitm, err := parItemGet()
+ if err != nil {
+ return err
+ }
+
+ mitm, err := memItemGet()
+ if err != nil {
+ return err
+ }
+
+ for {
+ usePitm := pitm != nil
+ if pitm != nil && mitm != nil {
+ usePitm = cmpFn(pitm) < cmpFn(mitm)
+ } else if pitm == nil && mitm == nil {
+ break
+ }
+
+ toUse := (*item)(nil)
+ if usePitm {
+ toUse = pitm
+ if pitm, err = parItemGet(); err != nil {
+ return err
+ }
+ } else {
+ toUse = mitm
+ if mitm, err = memItemGet(); err != nil {
+ return err
+ }
+ }
+
+ if dedup != nil {
+ encKey := toUse.getEncKey()
+ dedupLock.Lock()
+ added := dedup.Add(encKey)
+ dedupLock.Unlock()
+ if !added {
+ continue
+ }
+ }
+ if distinct != nil {
+ // NOTE: We know that toUse will not be used after this point for
+ // comparison purposes, so re-use its cmpRow property for our distinct
+ // filter here.
+ toUse.cmpRow = ""
+ if !distinct.Add(toUse.getCmpRow(nil, nil, distinctOrder)) {
+ continue
+ }
+ }
+ if !cb(toUse.key, toUse.data) {
+ break
+ }
+ }
+
+ return nil
+}
+
+// toComparableString computes the byte-sortable 'order' string for the given
+// key/PropertyMap.
+//
+// * start/end are byte sequences which are the inequality bounds of the
+// query, if any. These are a serialized datastore.Property. If the
+// inequality column is inverted, then start and end are also inverted and
+// swapped with each other.
+// * order is the list of sort orders in the actual executing queries.
+// * k / pm are the data to derive a sortable string for.
+//
+// The result of this function is the series of serialized properties, one per
+// order column, which represent this key/pm's first entry in the composite
+// index that would point to it (e.g. the one with `order` sort orders).
+func toComparableString(start, end []byte, order []ds.IndexColumn, k *ds.Key, pm ds.PropertyMap) (row, key []byte) {
+ doCmp := true
+ soFar := []byte{}
+ ps := serialize.PropertyMapPartially(k, nil)
+ for _, ord := range order {
+ row, ok := ps[ord.Property]
+ if !ok {
+ if vals, ok := pm[ord.Property]; ok {
+ row = serialize.PropertySlice(vals)
+ }
+ }
+ sort.Sort(row)
+ foundOne := false
+ for _, serialized := range row {
+ if ord.Descending {
+ serialized = serialize.Invert(serialized)
+ }
+ if doCmp {
+ maybe := serialize.Join(soFar, serialized)
+ cmp := bytes.Compare(maybe, start)
+ if cmp >= 0 {
+ foundOne = true
+ soFar = maybe
+ doCmp = len(soFar) < len(start)
+ break
+ }
+ } else {
+ foundOne = true
+ soFar = serialize.Join(soFar, serialized)
+ break
+ }
+ }
+ if !foundOne {
+ return nil, nil
+ }
+ }
+ if end != nil && bytes.Compare(soFar, end) >= 0 {
+ return nil, nil
+ }
+ return soFar, ps["__key__"][0]
+}

Powered by Google App Engine
This is Rietveld 408576698