summaryrefslogtreecommitdiff
path: root/jstests/sharding/auth_add_shard.js
blob: 8435c768c4f7a208e2ad8f033321da77ffa771cd (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
// SERVER-5124
// The puporse of this test is to test authentication when adding/removing a shard. The test sets
// up a sharded system, then adds/removes a shard.
(function() {
'use strict';

// login method to login into the database
function login(userObj) {
    var authResult = mongos.getDB(userObj.db).auth(userObj.username, userObj.password); 
    printjson(authResult);
}

// admin user object
var adminUser = { db: "admin", username: "foo", password: "bar" };

//set up a 2 shard cluster with keyfile
var st = new ShardingTest({ name: "auth_add_shard1", shards: 1,
                            mongos: 1, keyFile: "jstests/libs/key1" });

var mongos = st.s0;
var admin = mongos.getDB("admin");

print("1 shard system setup");

//add the admin user
print("adding user");
mongos.getDB(adminUser.db).createUser({ user: adminUser.username,
                                        pwd: adminUser.password,
                                        roles: jsTest.adminUserRoles});

//login as admin user
login(adminUser);

assert.eq(1, st.config.shards.count() , "initial server count wrong");

//start a mongod with NO keyfile
var conn = MongoRunner.runMongod({});
print(conn);

// --------------- Test 1 --------------------
// Add shard to the existing cluster (should fail because it was added without a keyfile)
printjson(assert.commandFailed(admin.runCommand({ addShard: conn.host })));

// stop mongod
MongoRunner.stopMongod(conn);

//--------------- Test 2 --------------------
//start mongod again, this time with keyfile 
var conn = MongoRunner.runMongod({keyFile: "jstests/libs/key1"});
//try adding the new shard
assert.commandWorked(admin.runCommand({ addShard: conn.host }));

//Add some data
var db = mongos.getDB("foo");
var collA = mongos.getCollection("foo.bar");

// enable sharding on a collection
assert.commandWorked(admin.runCommand({ enableSharding: "" + collA.getDB() }));
st.ensurePrimaryShard("foo", "shard0000");

assert.commandWorked(admin.runCommand({ shardCollection: "" + collA, key: { _id: 1 } }));

// add data to the sharded collection
for (var i = 0; i < 4; i++) {
  db.bar.save({ _id: i });
  assert.commandWorked(admin.runCommand({ split: "" + collA, middle: { _id: i } }));
}

// move a chunk 
assert.commandWorked(admin.runCommand({ moveChunk: "foo.bar", find: { _id: 1 }, to: "shard0001" }));

//verify the chunk was moved
admin.runCommand({ flushRouterConfig: 1 });

var config = mongos.getDB("config");
st.printShardingStatus(true);

// start balancer before removing the shard
st.startBalancer();

//--------------- Test 3 --------------------
// now drain the shard
assert.commandWorked(admin.runCommand({removeShard: conn.host}));

// give it some time to drain
assert.soon(function() {
    var result = admin.runCommand({removeShard: conn.host});
    printjson(result);

    return result.ok && result.state == "completed";
}, "failed to drain shard completely", 5 * 60 * 1000);

assert.eq(1, st.config.shards.count() , "removed server still appears in count");

MongoRunner.stopMongod(conn);

st.stop();

})();