summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorantirez <antirez@gmail.com>2012-11-30 15:41:26 +0100
committerantirez <antirez@gmail.com>2012-11-30 16:36:42 +0100
commitf50e658455f6455ac443e185e5fc738ef15093b3 (patch)
tree974395fec7f8c402348b4f3118ed7ff4ddad32f3
parentb4abbaf755d723d33b0e81880fb035ab88f3544f (diff)
downloadredis-f50e658455f6455ac443e185e5fc738ef15093b3.tar.gz
SDIFF is now able to select between two algorithms for speed.
SDIFF used an algorithm that was O(N) where N is the total number of elements of all the sets involved in the operation. The algorithm worked like that: ALGORITHM 1: 1) For the first set, add all the members to an auxiliary set. 2) For all the other sets, remove all the members of the set from the auxiliary set. So it is an O(N) algorithm where N is the total number of elements in all the sets involved in the diff operation. Cristobal Viedma suggested to modify the algorithm to the following: ALGORITHM 2: 1) Iterate all the elements of the first set. 2) For every element, check if the element also exists in all the other remaining sets. 3) Add the element to the auxiliary set only if it does not exist in any of the other sets. The complexity of this algorithm on the worst case is O(N*M) where N is the size of the first set and M the total number of sets involved in the operation. However when there are elements in common, with this algorithm we stop the computation for a given element as long as we find a duplicated element into another set. I (antirez) added an additional step to algorithm 2 to make it faster, that is to sort the set to subtract from the biggest to the smallest, so that it is more likely to find a duplicate in a larger sets that are checked before the smaller ones. WHAT IS BETTER? None of course, for instance if the first set is much larger than the other sets the second algorithm does a lot more work compared to the first algorithm. Similarly if the first set is much smaller than the other sets, the original algorithm will less work. So this commit makes Redis able to guess the number of operations required by each algorithm, and select the best at runtime according to the input received. However, since the second algorithm has better constant times and can do less work if there are duplicated elements, an advantage is given to the second algorithm.
-rw-r--r--src/t_set.c113
-rw-r--r--tests/unit/type/set.tcl7
2 files changed, 101 insertions, 19 deletions
diff --git a/src/t_set.c b/src/t_set.c
index 46a0c6ee5..01ac92aa4 100644
--- a/src/t_set.c
+++ b/src/t_set.c
@@ -565,6 +565,14 @@ int qsortCompareSetsByCardinality(const void *s1, const void *s2) {
return setTypeSize(*(robj**)s1)-setTypeSize(*(robj**)s2);
}
+/* This is used by SDIFF and in this case we can receive NULL that should
+ * be handled as empty sets. */
+int qsortCompareSetsByRevCardinality(const void *s1, const void *s2) {
+ robj *o1 = *(robj**)s1, *o2 = *(robj**)s2;
+
+ return (o2 ? setTypeSize(o2) : 0) - (o1 ? setTypeSize(o1) : 0);
+}
+
void sinterGenericCommand(redisClient *c, robj **setkeys, unsigned long setnum, robj *dstkey) {
robj **sets = zmalloc(sizeof(robj*)*setnum);
setTypeIterator *si;
@@ -712,6 +720,7 @@ void sunionDiffGenericCommand(redisClient *c, robj **setkeys, int setnum, robj *
setTypeIterator *si;
robj *ele, *dstset = NULL;
int j, cardinality = 0;
+ int diff_algo = 1;
for (j = 0; j < setnum; j++) {
robj *setobj = dstkey ?
@@ -728,34 +737,106 @@ void sunionDiffGenericCommand(redisClient *c, robj **setkeys, int setnum, robj *
sets[j] = setobj;
}
+ /* Select what DIFF algorithm to use.
+ *
+ * Algorithm 1 is O(N*M) where N is the size of the element first set
+ * and M the total number of sets.
+ *
+ * Algorithm 2 is O(N) where N is the total number of elements in all
+ * the sets.
+ *
+ * We compute what is the best bet with the current input here. */
+ if (op == REDIS_OP_DIFF && sets[0]) {
+ long long algo_one_work = 0, algo_two_work = 0;
+
+ for (j = 0; j < setnum; j++) {
+ if (sets[j] == NULL) continue;
+
+ algo_one_work += setTypeSize(sets[0]);
+ algo_two_work += setTypeSize(sets[j]);
+ }
+
+ /* Algorithm 1 has better constant times and performs less operations
+ * if there are elements in common. Give it some advantage. */
+ algo_one_work /= 2;
+ diff_algo = (algo_one_work <= algo_two_work) ? 1 : 2;
+
+ if (diff_algo == 1 && setnum > 1) {
+ /* With algorithm 1 it is better to order the sets to subtract
+ * by decreasing size, so that we are more likely to find
+ * duplicated elements ASAP. */
+ qsort(sets+1,setnum-1,sizeof(robj*),
+ qsortCompareSetsByRevCardinality);
+ }
+ }
+
/* We need a temp set object to store our union. If the dstkey
* is not NULL (that is, we are inside an SUNIONSTORE operation) then
* this set object will be the resulting object to set into the target key*/
dstset = createIntsetObject();
- /* Iterate all the elements of all the sets, add every element a single
- * time to the result set */
- for (j = 0; j < setnum; j++) {
- if (op == REDIS_OP_DIFF && j == 0 && !sets[j]) break; /* result set is empty */
- if (!sets[j]) continue; /* non existing keys are like empty sets */
+ if (op == REDIS_OP_UNION) {
+ /* Union is trivial, just add every element of every set to the
+ * temporary set. */
+ for (j = 0; j < setnum; j++) {
+ if (!sets[j]) continue; /* non existing keys are like empty sets */
- si = setTypeInitIterator(sets[j]);
+ si = setTypeInitIterator(sets[j]);
+ while((ele = setTypeNextObject(si)) != NULL) {
+ if (setTypeAdd(dstset,ele)) cardinality++;
+ decrRefCount(ele);
+ }
+ setTypeReleaseIterator(si);
+ }
+ } else if (op == REDIS_OP_DIFF && sets[0] && diff_algo == 1) {
+ /* DIFF Algorithm 1:
+ *
+ * We perform the diff by iterating all the elements of the first set,
+ * and only adding it to the target set if the element does not exist
+ * into all the other sets.
+ *
+ * This way we perform at max N*M operations, where N is the size of
+ * the first set, and M the number of sets. */
+ si = setTypeInitIterator(sets[0]);
while((ele = setTypeNextObject(si)) != NULL) {
- if (op == REDIS_OP_UNION || j == 0) {
- if (setTypeAdd(dstset,ele)) {
- cardinality++;
- }
- } else if (op == REDIS_OP_DIFF) {
- if (setTypeRemove(dstset,ele)) {
- cardinality--;
- }
+ for (j = 1; j < setnum; j++) {
+ if (!sets[j]) continue; /* no key is an empty set. */
+ if (setTypeIsMember(sets[j],ele)) break;
+ }
+ if (j == setnum) {
+ /* There is no other set with this element. Add it. */
+ setTypeAdd(dstset,ele);
+ cardinality++;
}
decrRefCount(ele);
}
setTypeReleaseIterator(si);
+ } else if (op == REDIS_OP_DIFF && sets[0] && diff_algo == 2) {
+ /* DIFF Algorithm 2:
+ *
+ * Add all the elements of the first set to the auxiliary set.
+ * Then remove all the elements of all the next sets from it.
+ *
+ * This is O(N) where N is the sum of all the elements in every
+ * set. */
+ for (j = 0; j < setnum; j++) {
+ if (!sets[j]) continue; /* non existing keys are like empty sets */
+
+ si = setTypeInitIterator(sets[j]);
+ while((ele = setTypeNextObject(si)) != NULL) {
+ if (j == 0) {
+ if (setTypeAdd(dstset,ele)) cardinality++;
+ } else {
+ if (setTypeRemove(dstset,ele)) cardinality--;
+ }
+ decrRefCount(ele);
+ }
+ setTypeReleaseIterator(si);
- /* Exit when result set is empty. */
- if (op == REDIS_OP_DIFF && cardinality == 0) break;
+ /* Exit if result set is empty as any additional removal
+ * of elements will have no effect. */
+ if (cardinality == 0) break;
+ }
}
/* Output the content of the resulting set, if not in STORE mode */
diff --git a/tests/unit/type/set.tcl b/tests/unit/type/set.tcl
index 9603eb568..a77759e5d 100644
--- a/tests/unit/type/set.tcl
+++ b/tests/unit/type/set.tcl
@@ -199,9 +199,10 @@ start_server {
test "SDIFFSTORE with three sets - $type" {
r sdiffstore setres set1 set4 set5
- # The type is determined by type of the first key to diff against.
- # See the implementation for more information.
- assert_encoding $type setres
+ # When we start with intsets, we should always end with intsets.
+ if {$type eq {intset}} {
+ assert_encoding intset setres
+ }
assert_equal {1 2 3 4} [lsort [r smembers setres]]
}
}