mongo/jstests/sharding/query/explain/explain_cmd.js

190 lines
6.8 KiB
JavaScript

/* Tests for the mongos explain command.
* @tags: [
* requires_fcv_80
* ]
*/
import {ChangeStreamTest} from "jstests/libs/query/change_stream_util.js";
import {ShardingTest} from "jstests/libs/shardingtest.js";
// Create a cluster with 3 shards.
let st = new ShardingTest({shards: 2});
var db = st.s.getDB("test");
let explain;
assert.commandWorked(db.adminCommand({enableSharding: db.getName(), primaryShard: st.shard1.shardName}));
// Setup a collection that will be sharded. The shard key will be 'a'. There's also an index on
// 'b'.
let collSharded = db.getCollection("mongos_explain_cmd");
collSharded.drop();
collSharded.createIndex({a: 1});
collSharded.createIndex({b: 1});
db.adminCommand({shardCollection: collSharded.getFullName(), key: {a: 1}});
// Pre-split the collection to ensure that both shards have chunks. Explicitly
// move chunks since the balancer is disabled.
assert.commandWorked(db.adminCommand({split: collSharded.getFullName(), middle: {a: 1}}));
printjson(db.adminCommand({moveChunk: collSharded.getFullName(), find: {a: 1}, to: st.shard0.shardName}));
assert.commandWorked(db.adminCommand({split: collSharded.getFullName(), middle: {a: 2}}));
printjson(db.adminCommand({moveChunk: collSharded.getFullName(), find: {a: 2}, to: st.shard1.shardName}));
// Put data on each shard.
for (let i = 0; i < 3; i++) {
collSharded.insert({_id: i, a: i, b: 1});
}
st.printShardingStatus();
// Test a scatter-gather count command.
assert.eq(3, collSharded.count({b: 1}));
// Explain the scatter-gather count.
explain = db.runCommand({explain: {count: collSharded.getName(), query: {b: 1}}, verbosity: "allPlansExecution"});
// Validate some basic properties of the result.
printjson(explain);
assert.commandWorked(explain);
assert("queryPlanner" in explain);
assert("executionStats" in explain);
assert.eq(2, explain.queryPlanner.winningPlan.shards.length);
assert.eq(2, explain.executionStats.executionStages.shards.length);
assert("serverInfo" in explain, explain);
assert.hasFields(explain.serverInfo, ["host", "port", "version", "gitVersion"]);
// An explain of a command that doesn't exist should fail gracefully.
explain = db.runCommand({explain: {nonexistent: collSharded.getName(), query: {b: 1}}, verbosity: "allPlansExecution"});
printjson(explain);
assert.commandFailed(explain);
// -------
// Setup a collection that is not sharded.
let collUnsharded = db.getCollection("mongos_explain_cmd_unsharded");
collUnsharded.drop();
collUnsharded.createIndex({a: 1});
collUnsharded.createIndex({b: 1});
for (let i = 0; i < 3; i++) {
collUnsharded.insert({_id: i, a: i, b: 1});
}
assert.eq(3, collUnsharded.count({b: 1}));
// -------
// Explain a delete operation and verify that it hits all shards without the shard key
explain = db.runCommand({
explain: {delete: collSharded.getName(), deletes: [{q: {b: 1}, limit: 0}]},
verbosity: "allPlansExecution",
});
assert.commandWorked(explain, tojson(explain));
assert.eq(explain.queryPlanner.winningPlan.stage, "SHARD_WRITE");
let shards = explain.queryPlanner.winningPlan.shards;
assert.eq(shards.length, 2);
for (let shard of shards) {
const shardStage = shard.winningPlan.stage;
assert(shardStage === "DELETE" || shardStage === "BATCHED_DELETE");
assert(shard.hasOwnProperty("explainVersion"));
}
// Check that the deletes didn't actually happen.
assert.eq(3, collSharded.count({b: 1}));
// Explain a delete operation and verify that it hits only one shard with the shard key
explain = db.runCommand({
explain: {delete: collSharded.getName(), deletes: [{q: {a: 1}, limit: 0}]},
verbosity: "allPlansExecution",
});
assert.commandWorked(explain, tojson(explain));
assert.eq(explain.queryPlanner.winningPlan.shards.length, 1);
// Check that the deletes didn't actually happen.
assert.eq(3, collSharded.count({b: 1}));
// Check that we fail gracefully if we try to do an explain of a write batch that has more
// than one operation in it.
explain = db.runCommand({
explain: {
delete: collSharded.getName(),
deletes: [
{q: {a: 1}, limit: 1},
{q: {a: 2}, limit: 1},
],
},
verbosity: "allPlansExecution",
});
assert.commandFailed(explain, tojson(explain));
// Explain a multi upsert operation and verify that it hits all shards
explain = db.runCommand({
explain: {update: collSharded.getName(), updates: [{q: {}, u: {$set: {b: 10}}, multi: true}]},
verbosity: "allPlansExecution",
});
assert.commandWorked(explain, tojson(explain));
assert.eq(explain.queryPlanner.winningPlan.shards.length, 2);
assert.eq(explain.queryPlanner.winningPlan.stage, "SHARD_WRITE");
shards = explain.queryPlanner.winningPlan.shards;
assert.eq(shards.length, 2);
for (let shard of shards) {
const shardStage = shard.winningPlan.stage;
assert.eq(shardStage, "UPDATE");
assert(shard.hasOwnProperty("explainVersion"));
}
// Check that the update didn't actually happen.
assert.eq(0, collSharded.count({b: 10}));
// Explain an upsert operation and verify that it hits only a single shard
explain = db.runCommand({
explain: {update: collSharded.getName(), updates: [{q: {a: 10}, u: {a: 10}, upsert: true}]},
verbosity: "allPlansExecution",
});
assert.commandWorked(explain, tojson(explain));
assert.eq(explain.queryPlanner.winningPlan.shards.length, 1);
// Check that the upsert didn't actually happen.
assert.eq(0, collSharded.count({a: 10}));
// Sharded updateOne that does not target a single shard can now be executed with a two phase
// write protocol that will target at most 1 matching document.
// Explain an upsert operation which cannot be targeted and verify that it is successful.
explain = db.runCommand({
explain: {update: collSharded.getName(), updates: [{q: {b: 10}, u: {b: 10}, upsert: true}]},
verbosity: "allPlansExecution",
});
assert(explain.queryPlanner);
assert(explain.executionStats);
assert.eq(explain.queryPlanner.winningPlan.stage, "SHARD_WRITE");
assert.eq(explain.queryPlanner.winningPlan.inputStage.winningPlan.stage, "SHARD_MERGE");
assert.eq(explain.executionStats.executionStages.stage, "SHARD_WRITE");
assert.eq(explain.executionStats.inputStage.executionStages.stage, "SHARD_MERGE");
// Check that the upsert didn't actually happen.
assert.eq(0, collSharded.count({b: 10}));
// Explain a changeStream, ensure an error is thrown under snapshot read concern.
const session = db.getMongo().startSession();
const sessionDB = session.getDatabase(db.getName());
const cst = new ChangeStreamTest(sessionDB);
assert.throwsWithCode(() => {
explain = cst.startWatchingChanges({
pipeline: [{$changeStream: {}}],
collection: "coll",
aggregateOptions: {
explain: true,
readConcern: {level: "snapshot"},
txnNumber: NumberLong(0),
autocommit: false,
startTransaction: true,
},
});
}, ErrorCodes.OperationNotSupportedInTransaction);
cst.cleanUp();
st.stop();