1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68
|
// Tests group using slaveOk
var st = new ShardingTest( testName = "groupSlaveOk",
numShards = 1,
verboseLevel = 0,
numMongos = 1,
{ rs : true,
rs0 : { nodes : 2 }
})
var rst = st._rs[0].test
// Insert data into replica set
var conn = new Mongo( st.s.host )
conn.setLogLevel( 3 )
var coll = conn.getCollection( "test.groupSlaveOk" )
coll.drop()
for( var i = 0; i < 300; i++ ){
coll.insert( { i : i % 10 } )
}
// Make sure the writes get through, otherwise we can continue to error these one-at-a-time
coll.getDB().getLastError()
st.printShardingStatus()
// Wait for client to update itself and replication to finish
rst.awaitReplication()
var primary = rst.getPrimary()
var sec = rst.getSecondary()
// Data now inserted... stop the master, since only two in set, other will still be secondary
rst.stop( rst.getMaster(), undefined, true )
printjson( rst.status() )
// Wait for the mongos to recognize the slave
ReplSetTest.awaitRSClientHosts( conn, sec, { ok : true, secondary : true } )
// Need to check slaveOk=true first, since slaveOk=false will destroy conn in pool when
// master is down
conn.setSlaveOk()
// Should not throw exception, since slaveOk'd
assert.eq( 10, coll.group({ key : { i : true } ,
reduce : function( obj, ctx ){ ctx.count += 1 } ,
initial : { count : 0 } }).length )
try {
conn.setSlaveOk( false )
coll.group({ key : { i : true } ,
reduce : function( obj, ctx ){ ctx.count += 1 } ,
initial : { count : 0 } })
print( "Should not reach here!" )
printjson( coll.getDB().getLastError() )
assert( false )
}
catch( e ){
print( "Non-slaveOk'd connection failed." )
}
// Finish
st.stop()
|