1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
|
/**
* Copyright (C) 2018-present MongoDB, Inc.
*
* This program is free software: you can redistribute it and/or modify
* it under the terms of the Server Side Public License, version 1,
* as published by MongoDB, Inc.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* Server Side Public License for more details.
*
* You should have received a copy of the Server Side Public License
* along with this program. If not, see
* <http://www.mongodb.com/licensing/server-side-public-license>.
*
* As a special exception, the copyright holders give permission to link the
* code of portions of this program with the OpenSSL library under certain
* conditions as described in each individual source file and distribute
* linked combinations including the program with the OpenSSL library. You
* must comply with the Server Side Public License in all respects for
* all of the code used other than as permitted herein. If you modify file(s)
* with this exception, you may extend this exception to your version of the
* file(s), but you are not obligated to do so. If you do not wish to do so,
* delete this exception statement from your version. If you delete this
* exception statement from all source files in the program, then also delete
* it in the license file.
*/
#include "mongo/platform/basic.h"
#include "mongo/db/query/planner_wildcard_helpers.h"
#include <vector>
#include "mongo/bson/util/builder.h"
#include "mongo/db/exec/projection_executor_utils.h"
#include "mongo/db/index/wildcard_key_generator.h"
#include "mongo/db/index_names.h"
#include "mongo/db/query/index_bounds.h"
#include "mongo/logv2/log.h"
#define MONGO_LOGV2_DEFAULT_COMPONENT ::mongo::logv2::LogComponent::kQuery
namespace mongo {
namespace wildcard_planning {
namespace {
/**
* Returns a new key pattern object with '$_path' and finds the wildcard field name.
*/
BSONObj makeNewKeyPattern(const IndexEntry* index, StringData* wildcardFieldName) {
BSONObjBuilder newPattern;
size_t idx = 0;
for (auto elem : index->keyPattern) {
if (idx == index->wildcardFieldPos) {
newPattern.append(BSON("$_path" << 1).firstElement());
*wildcardFieldName = elem.fieldNameStringData();
}
newPattern.append(elem);
idx++;
}
return newPattern.obj();
}
/**
* Compares the path 'fieldNameOrArrayIndexPath' to 'staticComparisonPath', ignoring any array
* indices present in the former if they are not present in the latter. The 'multikeyPathComponents'
* set contains the path positions that are known to be arrays; only numerical path components that
* immediately follow an array field are considered array indices. If 'fieldNameOrArrayIndexPath' is
* 'a.0.b', it will match 'staticComparisonPath' 'a.0.b', and it will also match 'a.b' but only if
* 'multikeyPathComponents' indicates that 'a' is an array.
*/
bool fieldNameOrArrayIndexPathMatches(const FieldRef& fieldNameOrArrayIndexPath,
const FieldRef& staticComparisonPath,
const MultikeyComponents& multikeyPathComponents) {
// Can't be equal if 'staticComparisonPath' has more parts than 'fieldNameOrArrayIndexPath'.
if (staticComparisonPath.numParts() > fieldNameOrArrayIndexPath.numParts()) {
return false;
}
size_t offset = 0;
for (size_t i = 0; i < fieldNameOrArrayIndexPath.numParts(); ++i) {
if (i - offset >= staticComparisonPath.numParts()) {
return false;
}
if (fieldNameOrArrayIndexPath.getPart(i) == staticComparisonPath.getPart(i - offset)) {
continue;
} else if (multikeyPathComponents.count(i - 1) &&
fieldNameOrArrayIndexPath.isNumericPathComponentStrict(i)) {
++offset;
continue;
}
return false;
}
// Ensure that we matched the entire 'staticComparisonPath' dotted string.
return fieldNameOrArrayIndexPath.numParts() == staticComparisonPath.numParts() + offset;
}
/**
* Returns true if 'multikeyPathSet' contains a FieldRef that matches 'pathToLookup' exactly, or
* matches 'pathToLookup' when the latter's array indices are ignored.
*/
bool fieldNameOrArrayIndexPathSetContains(const std::set<FieldRef>& multikeyPathSet,
const MultikeyComponents& multikeyPathComponents,
const FieldRef& pathToLookup) {
// Fast-path check for an exact match. If there is no exact match and 'pathToLookup' has no
// numeric path components, then 'multikeyPathSet' does not contain the path.
if (multikeyPathSet.count(pathToLookup)) {
return true;
} else if (!pathToLookup.hasNumericPathComponents()) {
return false;
}
// Determine whether any of the 'multikeyPathSet' entries match 'pathToLookup' under relaxed
// fieldname-or-array-index constraints.
return std::any_of(
multikeyPathSet.begin(), multikeyPathSet.end(), [&](const auto& multikeyPath) {
return fieldNameOrArrayIndexPathMatches(
pathToLookup, multikeyPath, multikeyPathComponents);
});
}
/**
* Returns the positions of all path components in 'queryPath' that may be interpreted as array
* indices by the query system. We obtain this list by finding all multikey path components that
* have a numerical path component immediately after. Note that the 'queryPath' argument may be a
* prefix of the full path used to generate 'multikeyPaths', and so we must avoid checking path
* components beyond the end of 'queryPath'.
*/
std::vector<size_t> findArrayIndexPathComponents(const MultikeyComponents& multikeyPaths,
const FieldRef& queryPath) {
std::vector<size_t> arrayIndices;
for (auto i : multikeyPaths) {
if (i < queryPath.numParts() - 1 && queryPath.isNumericPathComponentStrict(i + 1)) {
arrayIndices.push_back(i + 1);
}
}
return arrayIndices;
}
/**
* Returns a FieldRef of the full dotted field, minus the parts at indices listed in
* 'skipComponents'.
*/
FieldRef pathWithoutSpecifiedComponents(const FieldRef& path,
const std::set<size_t>& skipComponents) {
// If 'skipComponents' is empty, we return 'path' immediately.
if (skipComponents.empty()) {
return path;
}
FieldRef result;
for (size_t index = 0; index < path.numParts(); ++index) {
if (!skipComponents.count(index)) {
result.appendPart(path.getPart(index));
}
}
return result;
}
/**
* Returns a MultikeyPaths which indicates which components of 'indexedPath' are multikey, by
* looking up multikeyness in 'multikeyPathSet'.
*/
MultikeyPaths buildMultiKeyPathsForExpandedWildcardIndexEntry(
const BSONObj& keyPattern,
const FieldRef& indexedPath,
const std::set<FieldRef>& multikeyPathSet) {
MultikeyPaths multikeyPaths{};
for (const auto& field : keyPattern) {
if (WildcardNames::isWildcardFieldName(field.fieldNameStringData())) {
FieldRef pathToLookup;
MultikeyComponents mkComponents;
for (size_t i = 0; i < indexedPath.numParts(); ++i) {
pathToLookup.appendPart(indexedPath.getPart(i));
if (fieldNameOrArrayIndexPathSetContains(
multikeyPathSet, mkComponents, pathToLookup)) {
mkComponents.insert(i);
}
}
multikeyPaths.emplace_back(mkComponents);
} else {
multikeyPaths.emplace_back();
}
}
return multikeyPaths;
}
std::set<FieldRef> generateFieldNameOrArrayIndexPathSet(const MultikeyComponents& multikeyPaths,
const FieldRef& queryPath,
bool requiresSubpathBounds) {
// We iterate over the power set of array index positions to generate all necessary paths.
// The algorithm is unavoidably O(n2^n), but we enforce that 'n' is never more than single
// digits during the planner's index selection phase.
const auto potentialArrayIndices = findArrayIndexPathComponents(multikeyPaths, queryPath);
invariant(potentialArrayIndices.size() <= kWildcardMaxArrayIndexTraversalDepth);
invariant(potentialArrayIndices.size() < sizeof(size_t) * 8u);
// We iterate over every value [0..2^n), where 'n' is the size of 'potentialArrayIndices',
// treating each value as a 'bitMask' of 'n' bits. Each bit in 'bitMask' represents the
// entry at the equivalent position in the 'potentialArrayIndices' vector. When a given bit
// is set, we treat the corresponding numeric path component as an array index, and generate
// a path that omits it. When a bit is not set, we treat the numeric path component as a
// literal fieldname, and we generate a path that includes it. Because we iterate over every
// value [0..2^n), we ensure that we generate every combination of 'n' bits, and therefore
// every possible fieldname and array index path.
std::set<FieldRef> paths;
for (size_t bitMask = 0; bitMask < (size_t{1} << potentialArrayIndices.size()); ++bitMask) {
std::set<size_t> arrayIndicesToSkip;
for (size_t i = 0; i < potentialArrayIndices.size(); ++i) {
if (bitMask & (size_t{1} << i)) {
arrayIndicesToSkip.insert(potentialArrayIndices[i]);
}
}
// Add the path to the FieldRef set, and obtain an iterator pointing to the new entry.
const auto result =
paths.emplace(pathWithoutSpecifiedComponents(queryPath, arrayIndicesToSkip));
// If any path in the set prefixes another, then the bounds generated will overlap (and
// thus, be invalid). So, we must make sure that the new path does not prefix and is not
// prefixed by any existing entries in the set. If any such prefixes do exist, we must
// remove the subpath(s) and retain only the shortest prefix path, since the bounds it
// generates will be a superset of all the paths generated by the removed entries.
if (requiresSubpathBounds && result.second) {
const auto currentPathItr = result.first;
// If the new path is a subpath of an existing entry, remove the new path.
if (currentPathItr != paths.begin() &&
std::prev(currentPathItr)->isPrefixOf(*currentPathItr)) {
paths.erase(currentPathItr);
continue;
}
// If existing paths are subpaths of the new entry, remove the old paths.
while (std::next(currentPathItr) != paths.end() &&
currentPathItr->isPrefixOf(*std::next(currentPathItr))) {
paths.erase(std::next(currentPathItr));
}
}
}
return paths;
}
/**
* Returns false if 'queryPath' includes any numerical path components which render it unanswerable
* by the $** index, true otherwise. Specifically, the $** index cannot answer the query if any
* of the following scenarios occur:
*
* - The query path traverses through more than 'kWildcardMaxArrayIndexTraversalDepth' nested arrays
* via explicit array indices.
* - The query path has multiple successive positional components that come immediately after a
* multikey path component.
* - The query path lies along a $** projection through an array index.
*
* For an example of the latter case, say that our query path is 'a.0.b', our projection includes
* {'a.0': 1}, and 'a' is multikey. The query semantics will match 'a.0.b' by either field name or
* array index against the documents, but because the $** index always projects numeric path
* components strictly as field names, the projection {'a.0': 1} cannot correctly support this
* query.
*
* To see why, consider the document {a: [1, 2, 3]}. Query {'a.0': 1} will match this document, but
* the projection {'a.0': 1} will produce output document {a: []}, and so we will not index any of
* the values [1, 2, 3] for 'a'.
*/
bool validateNumericPathComponents(const MultikeyPaths& multikeyPaths,
const std::set<FieldRef>& includedPaths,
const FieldRef& queryPath) {
// Find the position of the Wildcard's MultikeyComponents in the paths, we assume that the
// wildcard field is the only one that can be multikey.
auto wildcardComponent = std::find_if(multikeyPaths.begin(),
multikeyPaths.end(),
[](const MultikeyComponents& c) { return !c.empty(); });
if (wildcardComponent == multikeyPaths.end()) {
// If no MultikeyComponents just return.
return true;
}
// Find the positions of all multikey path components in 'queryPath' that have a numerical path
// component immediately after. For a queryPath of 'a.2.b' this will return position 0; that is,
// 'a'. If no such multikey path was found, we are clear to proceed with planning.
const auto arrayIndices = findArrayIndexPathComponents(*wildcardComponent, queryPath);
if (arrayIndices.empty()) {
return true;
}
// To support $** fieldname-or-array-index semantics, the planner must generate the power set of
// all paths with and without array indices. Because this is O(2^n), we decline to answer
// queries that traverse more than 8 levels of array indices.
if (arrayIndices.size() > kWildcardMaxArrayIndexTraversalDepth) {
LOGV2_DEBUG(20955,
2,
"Declining to answer query on a field with $** index, as it traverses through "
"more than the maximum permitted depth of nested array indices",
"field"_attr = queryPath.dottedField(),
"maxNestedArrayIndices"_attr = kWildcardMaxArrayIndexTraversalDepth);
return false;
}
// Prevent the query from attempting to use a wildcard index if there are multiple successive
// positional path components that follow a multikey path component. For example, the path
// "a.0.1" cannot use a wildcard index if "a" is multikey.
//
// This restriction stems from the fact that wildcard indices do not recursively index nested
// arrays. The document {a: [[3, 4]]}, for instance, will have a single index key containing
// the array [3, 4] rather than individual index keys for 3 and 4. If "a" is known to be
// multikey, then a user could issue a query like {"a.0.1": {$eq: 4}} to attempt to match by
// position within a nested array. The access planner is not able to generate useful bounds for
// such positional queries over nested arrays.
//
// We have already found all positional path components that are immediately preceded by a
// multikey path component. All that remains is to bail out if any of these positional path
// components are followed by another positional component.
for (auto&& positionalComponentIndex : arrayIndices) {
auto adjacentIndex = positionalComponentIndex + 1;
if (adjacentIndex < queryPath.numParts() &&
queryPath.isNumericPathComponentStrict(adjacentIndex)) {
// There are two adjacent positional components, so this query might need to match by
// position in a nested array. The query cannot be answered using a wildcard index.
return false;
}
}
// If 'includedPaths' is empty, then either the $** projection is an exclusion, or no explicit
// projection was provided. In either case, it is not possible for the query path to lie along
// an array index projection, and so we are safe to proceed with planning.
if (includedPaths.empty()) {
return true;
}
// Find the $** projected field which prefixes or is equal to the query path. If 'includedPaths'
// is non-empty then we are guaranteed that exactly one entry will prefix the query path, since
// (a) if no such inclusion exists, an IndexEntry would not have been created for this path, and
// (b) conflicting paths, such as 'a.b' and 'a.b.c', are not permitted in projections.
auto includePath = std::find_if(
includedPaths.begin(), includedPaths.end(), [&queryPath](const auto& includedPath) {
return includedPath.isPrefixOfOrEqualTo(queryPath);
});
invariant(std::next(includePath) == includedPaths.end() || *std::next(includePath) > queryPath);
// If the projectedPath responsible for including this queryPath prefixes it up to and including
// the numerical array index field, then the queryPath lies along a projection through the array
// index, and we cannot support the query for the reasons outlined above.
return arrayIndices[0] >= includePath->numParts();
}
/**
* Queries whose bounds overlap the Object type bracket may require special handling, since the $**
* index does not index complete objects but instead only contains the leaves along each of its
* subpaths. Since we ban all object-value queries except those on the empty object {}, this will
* typically only be relevant for bounds involving MinKey and MaxKey, such as {$exists: true}.
*/
bool boundsOverlapObjectTypeBracket(const OrderedIntervalList& oil) {
// Create an Interval representing the subrange ({}, []) of the object type bracket. We exclude
// both ends of the bracket because $** indexes support queries on empty objects and arrays.
static const Interval objectTypeBracketBounds = []() {
BSONObjBuilder objBracketBounds;
objBracketBounds.appendMinForType("", BSONType::Object);
objBracketBounds.appendMaxForType("", BSONType::Object);
return IndexBoundsBuilder::makeRangeInterval(objBracketBounds.obj(),
BoundInclusion::kExcludeBothStartAndEndKeys);
}();
// Determine whether any of the ordered intervals overlap with the object type bracket. Because
// Interval's various bounds-comparison methods all depend upon the bounds being in ascending
// order, we reverse the direction of the input OIL if necessary here.
const bool isDescending = (oil.computeDirection() == Interval::Direction::kDirectionDescending);
const auto& oilAscending = (isDescending ? oil.reverseClone() : oil);
// Iterate through each of the OIL's intervals. If the current interval precedes the bracket, we
// must check the next interval in sequence. If the interval succeeds the bracket then we can
// stop checking. If we neither precede nor succeed the object type bracket, then they overlap.
for (const auto& interval : oilAscending.intervals) {
switch (interval.compare(objectTypeBracketBounds)) {
case Interval::IntervalComparison::INTERVAL_PRECEDES_COULD_UNION:
case Interval::IntervalComparison::INTERVAL_PRECEDES:
// Break out of the switch and proceed to check the next interval.
break;
case Interval::IntervalComparison::INTERVAL_SUCCEEDS:
return false;
default:
return true;
}
}
// If we're here, then all the OIL's bounds precede the object type bracket.
return false;
}
/**
* Returns expanded wildcard key pattern with a wildcard field replaced by the given expandField and
* the position of the replaced wildcard field.
*/
std::pair<BSONObj, size_t> expandWildcardIndexKeyPattern(const BSONObj& wildcardKeyPattern,
StringData expandFieldName) {
int wildcardFieldPos = -1;
int fieldPos = 0;
BSONObjBuilder builder{};
for (const auto& field : wildcardKeyPattern) {
const auto& fieldName = field.fieldNameStringData();
if (WildcardNames::isWildcardFieldName(fieldName)) {
tassert(7246500,
str::stream()
<< "Wildcard Index's key pattern must contain exactly one wildcard field: '"
<< wildcardKeyPattern << "'.",
wildcardFieldPos < 0);
builder.appendAs(field, expandFieldName);
wildcardFieldPos = fieldPos;
} else {
builder.append(field);
}
++fieldPos;
}
tassert(7246501,
str::stream() << "Wildcard Index's key pattern must contain one wildcard field: '"
<< wildcardKeyPattern << "'.",
wildcardFieldPos >= 0);
return std::make_pair(builder.obj(), static_cast<size_t>(wildcardFieldPos));
}
boost::optional<IndexEntry> createExpandedIndexEntry(const IndexEntry& wildcardIndex,
const std::string& fieldName,
const std::set<FieldRef>& includedPaths) {
// Convert string 'fieldName' into a FieldRef, to better facilitate the subsequent checks.
auto queryPath = FieldRef{fieldName};
// $** indices hold multikey metadata directly in the index keys, rather than in the index
// catalog. In turn, the index key data is used to produce a set of multikey paths
// in-memory. Here we convert this set of all multikey paths into a MultikeyPaths vector
// which will indicate to the downstream planning code which components of 'fieldName' are
// multikey.
auto multikeyPaths = buildMultiKeyPathsForExpandedWildcardIndexEntry(
wildcardIndex.keyPattern, queryPath, wildcardIndex.multikeyPathSet);
// Check whether a query on the current fieldpath is answerable by the $** index, given any
// numerical path components that may be present in the path string.
if (!validateNumericPathComponents(multikeyPaths, includedPaths, queryPath)) {
return boost::none;
}
// The expanded IndexEntry is only considered multikey if the particular path represented by
// this IndexEntry has a multikey path component. For instance, suppose we have index {$**:
// 1} with "a" as the only multikey path. If we have a query on paths "a.b" and "c.d", then
// we will generate two expanded index entries: one for "a.b" and "c.d". The "a.b" entry
// will be marked as multikey because "a" is multikey, whereas the "c.d" entry will not be
// marked as multikey.
tassert(7246506,
"multikeyPaths size must be equal to the number of the key pattern fields.",
multikeyPaths.size() == static_cast<size_t>(wildcardIndex.keyPattern.nFields()));
auto [expandedKeyPattern, wildcardFieldPos] =
expandWildcardIndexKeyPattern(wildcardIndex.keyPattern, fieldName);
const bool isMultikey = !multikeyPaths[wildcardFieldPos].empty();
IndexEntry entry(std::move(expandedKeyPattern),
IndexType::INDEX_WILDCARD,
IndexDescriptor::kLatestIndexVersion,
isMultikey,
std::move(multikeyPaths),
// Expanded index entries always use the fixed-size multikey paths
// representation, so we purposefully discard 'multikeyPathSet'.
{},
true, // sparse
false, // unique
{wildcardIndex.identifier.catalogName, fieldName},
wildcardIndex.filterExpr,
wildcardIndex.infoObj,
wildcardIndex.collator,
wildcardIndex.indexPathProjection,
wildcardFieldPos);
return entry;
}
/**
* Determines if an expanded index entry can satisfy a query on a wildcard field with a FETCH
* (for e.g., it may only be able to answer a query on the prefix if the wildcard field is being
* queried with an incompatible $not predicate).
*
* Note: we could just use 'index.keyPattern' here for this check, but then we would have to iterate
* through the entire pattern to get to the field at 'wildcardPos'.
*/
bool canOnlyAnswerWildcardPrefixQuery(const IndexEntry& index, const IndexBounds& bounds) {
tassert(7444000, "Expected a wildcard index.", index.type == INDEX_WILDCARD);
tassert(7444001,
"A wildcard index should always have a virtual $_path field at wildcardFieldPos - 1.",
bounds.fields[index.wildcardFieldPos - 1].name == "$_path"_sd);
if (index.wildcardFieldPos == 1) {
// This is either a single-field wildcard index, or a compound wildcard index without a
// prefix.
return false;
}
// If the index entry was not expanded to include a second $_path field, we cannot answer a
// query on a wildcard field with an IXSCAN + FETCH if the predicate itself is, for e.g. an
// ineligible $not query, because we won't retrieve documents where the wildcard field is
// missing from the IXSCAN.
return bounds.fields[index.wildcardFieldPos].name != "$_path"_sd;
}
} // namespace
void expandWildcardIndexEntry(const IndexEntry& wildcardIndex,
const stdx::unordered_set<std::string>& fields,
std::vector<IndexEntry>* out) {
tassert(7246502, "out parameter cannot be null", out);
tassert(7246503,
"expandWildcardIndexEntry expected only WildcardIndexes",
wildcardIndex.type == INDEX_WILDCARD);
// (Ignore FCV check): This is intentional because we want clusters which have wildcard indexes
// still be able to use the feature even if the FCV is downgraded.
if (!feature_flags::gFeatureFlagCompoundWildcardIndexes.isEnabledAndIgnoreFCVUnsafe()) {
// Should only have one field of the form {"path.$**" : 1}.
tassert(7246511,
"Wildcard Index's key pattern must always have length 1 for non-compound Wildcard "
"Indexes",
wildcardIndex.keyPattern.nFields() == 1);
tassert(7246512,
"Wildcard Index's field name must end with the wildcard suffix '$**'",
wildcardIndex.keyPattern.firstElement().fieldNameStringData().endsWith("$**"));
}
// $** indexes do not keep the multikey metadata inside the index catalog entry, as the amount
// of metadata is not bounded. We do not expect IndexEntry objects for $** indexes to have a
// fixed-size vector of multikey metadata until after they are expanded.
tassert(7246504,
"multikeyPaths must be empty for Wildcard Indexes",
wildcardIndex.multikeyPaths.empty());
// Obtain the projection executor from the parent wildcard IndexEntry.
auto* wildcardProjection = wildcardIndex.indexPathProjection;
tassert(
7246505, "wildcardProjection must be non-null for Wildcard Indexes", wildcardProjection);
const auto projectedFields =
projection_executor_utils::applyProjectionToFields(wildcardProjection->exec(), fields);
const static auto kEmptySet = std::set<FieldRef>{};
const auto& includedPaths =
wildcardProjection->exhaustivePaths() ? *wildcardProjection->exhaustivePaths() : kEmptySet;
for (auto&& fieldName : projectedFields) {
auto entry = createExpandedIndexEntry(wildcardIndex, fieldName, includedPaths);
if (entry == boost::none) {
continue;
}
tassert(7246507,
"'$_path' is reserved fieldname for Wildcard Indexes",
"$_path"_sd != fieldName);
out->push_back(*entry);
}
// If this wildcard index cannot be expanded because the wildcard field is not relevant. We
// should also check whether the regular fields is able to answer the query or not. That is - if
// any field of the regular fields in a compound wildcard index is in 'fields', then we should
// also generate an expanded wildcard 'IndexEntry' for later index analysis.
// (Ignore FCV check): This is intentional because we want clusters which have wildcard indexes
// still be able to use the feature even if the FCV is downgraded.
if (feature_flags::gFeatureFlagCompoundWildcardIndexes.isEnabledAndIgnoreFCVUnsafe()) {
bool shouldExpand = false;
for (auto elem : wildcardIndex.keyPattern) {
auto fieldName = elem.fieldNameStringData();
if (WildcardNames::isWildcardFieldName(fieldName)) {
break;
}
if (fields.count(fieldName.toString())) {
shouldExpand = true;
break;
}
}
// This expanded IndexEntry is for queries on the non-wildcard prefix of a compound wildcard
// index, the wildcard component is not required. We use the reserved path, "$_path", to
// instruct the query planner to generate "all values" index bounds and not consider this
// field in supporting any sort operation.
if (shouldExpand) {
auto entry = createExpandedIndexEntry(wildcardIndex, "$_path", {} /* paths included */);
out->push_back(*entry);
}
}
}
bool canOnlyAnswerWildcardPrefixQuery(
const std::vector<std::unique_ptr<QuerySolutionNode>>& ixscanNodes) {
return std::any_of(ixscanNodes.begin(), ixscanNodes.end(), [](const auto& node) {
if (node->getType() == StageType::STAGE_IXSCAN) {
const auto* ixScanNode = static_cast<IndexScanNode*>(node.get());
const auto& index = ixScanNode->index;
if (index.type == INDEX_WILDCARD &&
canOnlyAnswerWildcardPrefixQuery(index, ixScanNode->bounds)) {
return true;
}
}
return false;
});
}
BoundsTightness translateWildcardIndexBoundsAndTightness(
const IndexEntry& index,
BoundsTightness tightnessIn,
OrderedIntervalList* oil,
interval_evaluation_tree::Builder* ietBuilder) {
// This method should only ever be called for a $** IndexEntry. We expect to be called during
// planning, *before* finishWildcardIndexScanNode has been invoked. The IndexEntry should thus
// only have a single keyPattern field and multikeyPath entry, but this is sufficient to
// determine whether it will be necessary to adjust the tightness.
invariant(index.type == IndexType::INDEX_WILDCARD);
// (Ignore FCV check): This is intentional because we want clusters which have wildcard indexes
// still be able to use the feature even if the FCV is downgraded.
if (!feature_flags::gFeatureFlagCompoundWildcardIndexes.isEnabledAndIgnoreFCVUnsafe()) {
invariant(index.keyPattern.nFields() == 1);
invariant(index.multikeyPaths.size() == 1);
}
invariant(oil);
// If 'oil' was not filled the filter type may not be supported, but we can still use this
// wildcard index for queries on prefix fields. The index bounds for the wildcard field will be
// filled later to include all values. Therefore, we should use INEXACT_FETCH to avoid false
// positives.
if (oil->name.empty()) {
return BoundsTightness::INEXACT_FETCH;
}
// If our bounds include any objects -- anything in the range ({}, []) -- then we will need to
// use subpath bounds; that is, we will add the interval ["path.","path/") at the point where we
// finalize the index scan. If the subpath interval is required but the bounds do not already
// run from MinKey to MaxKey, then we must expand them to [MinKey, MaxKey]. Consider the case
// where out bounds are [[MinKey, undefined), (null, MaxKey]] as generated by {$ne: null}. Our
// result set should include documents such as {a: {b: null}}; however, the wildcard index key
// for this object will be {"": "a.b", "": null}, which means that the original bounds would
// skip this document. We must also set the tightness to INEXACT_FETCH to avoid false positives.
if (boundsOverlapObjectTypeBracket(*oil) && !oil->intervals.front().isMinToMax()) {
oil->intervals = {IndexBoundsBuilder::allValues()};
if (ietBuilder) {
// We need to replace a previously added interval in the IET builder with a new
// all-values interval.
tassert(
6944102, "Cannot pop an element from an empty IET builder", !ietBuilder->isEmpty());
ietBuilder->pop();
ietBuilder->addConst(*oil);
}
return BoundsTightness::INEXACT_FETCH;
}
auto wildcardElt = getWildcardField(index);
// If the query passes through any array indices, we must always fetch and filter the documents.
const auto arrayIndicesTraversedByQuery = findArrayIndexPathComponents(
index.multikeyPaths[index.wildcardFieldPos], FieldRef{wildcardElt.fieldName()});
// If the list of array indices we traversed is non-empty, set the tightness to INEXACT_FETCH.
return (arrayIndicesTraversedByQuery.empty() ? tightnessIn : BoundsTightness::INEXACT_FETCH);
}
void finalizeWildcardIndexScanConfiguration(
IndexScanNode* scan, std::vector<interval_evaluation_tree::Builder>* ietBuilders) {
IndexEntry* index = &scan->index;
IndexBounds* bounds = &scan->bounds;
// We should only ever reach this point when processing a $** index. Sanity check the arguments.
invariant(index && index->type == IndexType::INDEX_WILDCARD);
// (Ignore FCV check): This is intentional because we want clusters which have wildcard indexes
// still be able to use the feature even if the FCV is downgraded.
if (!feature_flags::gFeatureFlagCompoundWildcardIndexes.isEnabledAndIgnoreFCVUnsafe()) {
invariant(index->keyPattern.nFields() == 1);
invariant(index->multikeyPaths.size() == 1);
invariant(bounds && bounds->fields.size() == 1);
invariant(bounds->fields.front().name == index->keyPattern.firstElementFieldName());
tassert(6536700,
"IET Builders list must be size of 1 or empty for wildcard indexes",
ietBuilders->empty() || ietBuilders->size() == 1);
}
// For $** indexes, the IndexEntry key pattern is {..., 'path.to.field': 1, ...} but the actual
// keys in the index are of the form {..., '$_path': 1, 'path.to.field': 1, ...}, where the
// value of the wildcard field in each key is 'path.to.field'. We push a new entry into the
// bounds vector for the leading '$_path' bound here. We also push corresponding fields into the
// IndexScanNode's keyPattern and its multikeyPaths vector.
index->multikeyPaths.insert(index->multikeyPaths.begin() + index->wildcardFieldPos,
MultikeyComponents{});
bounds->fields.insert(bounds->fields.begin() + index->wildcardFieldPos, {"$_path"});
StringData wildcardFieldName;
index->keyPattern = makeNewKeyPattern(index, &wildcardFieldName);
if (!ietBuilders->empty()) {
auto wildcardIt = ietBuilders->begin();
std::advance(wildcardIt, index->wildcardFieldPos);
ietBuilders->emplace(wildcardIt);
}
// Update the position as we insert "$_path" prior to the wildcard field.
index->wildcardFieldPos++;
// If the wildcard field is "$_path", the index is used to answer query only on the non-wildcard
// prefix of a compound wildcard index. The bounds for both "$_path" fields should be
// "[MinKey, MaxKey]". Because the wildcard field can generate multiple keys for one single
// document, we should also instruct the IXSCAN to dedup keys.
if (wildcardFieldName == "$_path"_sd) {
bounds->fields[index->wildcardFieldPos - 1].intervals.push_back(
IndexBoundsBuilder::allValues());
bounds->fields[index->wildcardFieldPos].intervals.push_back(
IndexBoundsBuilder::allValues());
bounds->fields[index->wildcardFieldPos].name = "$_path";
scan->shouldDedup = true;
return;
}
// Create a FieldRef to perform any necessary manipulations on the query path string.
FieldRef queryPath{wildcardFieldName};
auto& multikeyPaths = index->multikeyPaths[index->wildcardFieldPos];
// If the bounds overlap the object type bracket or the wildcard field's bounds were not filled,
// then we must retrieve all documents which include the given path. We must therefore add
// bounds that encompass all its subpaths, specifically the interval ["path.","path/") on
// "$_path".
const bool requiresSubpathBounds = bounds->fields[index->wildcardFieldPos].name.empty() ||
boundsOverlapObjectTypeBracket(bounds->fields[index->wildcardFieldPos]);
// Account for fieldname-or-array-index semantics. $** indexes do not explicitly encode array
// indices in their keys, so if this query traverses one or more multikey fields via an array
// index (e.g. query 'a.0.b' where 'a' is an array), then we must generate bounds on all array-
// and non-array permutations of the path in order to produce INEXACT_FETCH bounds.
auto paths =
generateFieldNameOrArrayIndexPathSet(multikeyPaths, queryPath, requiresSubpathBounds);
// Add a $_path point-interval for each path that needs to be traversed in the index. If subpath
// bounds are required, then we must add a further range interval on ["path.","path/").
static const char subPathStart = '.', subPathEnd = static_cast<char>('.' + 1);
auto& pathIntervals = bounds->fields[index->wildcardFieldPos - 1].intervals;
for (const auto& fieldPath : paths) {
auto path = fieldPath.dottedField().toString();
pathIntervals.push_back(IndexBoundsBuilder::makePointInterval(path));
if (requiresSubpathBounds) {
pathIntervals.push_back(IndexBoundsBuilder::makeRangeInterval(
path + subPathStart, path + subPathEnd, BoundInclusion::kIncludeStartKeyOnly));
// Queries which scan subpaths for a single wildcard index should be deduped. The index
// bounds may include multiple keys associated with the same document. Therefore, we
// instruct the IXSCAN to dedup keys which point to the same object.
scan->shouldDedup = true;
}
}
}
bool isWildcardObjectSubpathScan(const IndexScanNode* node) {
// If the node is not a $** index scan, return false immediately.
if (!node || node->index.type != IndexType::INDEX_WILDCARD) {
return false;
}
// We expect consistent arguments, representing a $** index which has already been finalized.
// (Ignore FCV check): This is intentional because we want clusters which have wildcard indexes
// still be able to use the feature even if the FCV is downgraded.
if (!feature_flags::gFeatureFlagCompoundWildcardIndexes.isEnabledAndIgnoreFCVUnsafe()) {
invariant(node->index.keyPattern.nFields() == 2);
invariant(node->index.multikeyPaths.size() == 2);
invariant(node->bounds.fields.size() == 2);
invariant(node->bounds.fields.front().name ==
node->index.keyPattern.firstElementFieldName());
invariant(node->bounds.fields.back().name ==
std::next(node->index.keyPattern.begin())->fieldName());
}
// Check the bounds on the query field for any intersections with the object type bracket.
return boundsOverlapObjectTypeBracket(node->bounds.fields[node->index.wildcardFieldPos]);
}
BSONElement getWildcardField(const IndexEntry& index) {
uassert(7246601, "The index is not a wildcard index", index.type == IndexType::INDEX_WILDCARD);
BSONObjIterator it(index.keyPattern);
BSONElement wildcardElt = it.next();
for (size_t i = 0; i < index.wildcardFieldPos; ++i) {
invariant(it.more());
wildcardElt = it.next();
}
return wildcardElt;
}
} // namespace wildcard_planning
} // namespace mongo
|