mirror of
https://github.com/DeNNiiInc/dbgate.git
synced 2026-04-18 04:26:01 +00:00
354 lines
10 KiB
JavaScript
354 lines
10 KiB
JavaScript
const _ = require('lodash');
|
|
const stream = require('stream');
|
|
const isPromise = require('is-promise');
|
|
const driverBase = require('../frontend/driver');
|
|
const Analyser = require('./Analyser');
|
|
const MongoClient = require('mongodb').MongoClient;
|
|
const ObjectId = require('mongodb').ObjectId;
|
|
const AbstractCursor = require('mongodb').AbstractCursor;
|
|
const createBulkInsertStream = require('./createBulkInsertStream');
|
|
|
|
function transformMongoData(row) {
|
|
return _.cloneDeepWith(row, (x) => {
|
|
if (x && x.constructor == ObjectId) return { $oid: x.toString() };
|
|
});
|
|
}
|
|
|
|
async function readCursor(cursor, options) {
|
|
options.recordset({ __isDynamicStructure: true });
|
|
await cursor.forEach((row) => {
|
|
options.row(transformMongoData(row));
|
|
});
|
|
}
|
|
|
|
function convertObjectId(condition) {
|
|
return _.cloneDeepWith(condition, (x) => {
|
|
if (x && x.$oid) return ObjectId(x.$oid);
|
|
});
|
|
}
|
|
|
|
function findArrayResult(resValue) {
|
|
if (!_.isPlainObject(resValue)) return null;
|
|
const arrays = _.values(resValue).filter((x) => _.isArray(x));
|
|
if (arrays.length == 1) return arrays[0];
|
|
return null;
|
|
}
|
|
|
|
async function getScriptableDb(pool) {
|
|
const db = pool.__getDatabase();
|
|
const collections = await db.listCollections().toArray();
|
|
for (const collection of collections) {
|
|
db[collection.name] = db.collection(collection.name);
|
|
}
|
|
return db;
|
|
}
|
|
|
|
/** @type {import('dbgate-types').EngineDriver} */
|
|
const driver = {
|
|
...driverBase,
|
|
analyserClass: Analyser,
|
|
async connect({ server, port, user, password, database, useDatabaseUrl, databaseUrl, ssl }) {
|
|
let mongoUrl;
|
|
if (useDatabaseUrl) {
|
|
// change port to ssh tunnel port
|
|
const url = new URL(databaseUrl);
|
|
url.port = port;
|
|
mongoUrl = url.href;
|
|
} else {
|
|
mongoUrl = user
|
|
? `mongodb://${user}:${password}@${server}:${port}`
|
|
: `mongodb://${server}:${port}`;
|
|
}
|
|
|
|
const options = {
|
|
useUnifiedTopology: true,
|
|
};
|
|
if (ssl) {
|
|
options.tls = true;
|
|
options.tlsCAFile = ssl.ca;
|
|
options.tlsCertificateKeyFile = ssl.cert || ssl.key;
|
|
options.tlsCertificateKeyFilePassword = ssl.password;
|
|
options.tlsAllowInvalidCertificates = !ssl.rejectUnauthorized;
|
|
}
|
|
|
|
const pool = new MongoClient(mongoUrl, options);
|
|
await pool.connect();
|
|
// const pool = await MongoClient.connect(mongoUrl);
|
|
pool.__getDatabase = database ? () => pool.db(database) : () => pool.db();
|
|
pool.__databaseName = database;
|
|
return pool;
|
|
},
|
|
// @ts-ignore
|
|
async query(pool, sql) {
|
|
return {
|
|
rows: [],
|
|
columns: [],
|
|
};
|
|
},
|
|
async script(pool, sql) {
|
|
let func;
|
|
func = eval(`(db,ObjectId) => { ${sql} }`);
|
|
const db = await getScriptableDb(pool);
|
|
func(db, ObjectId);
|
|
},
|
|
async stream(pool, sql, options) {
|
|
let func;
|
|
try {
|
|
func = eval(`(db,ObjectId) => ${sql}`);
|
|
} catch (err) {
|
|
options.info({
|
|
message: 'Error compiling expression: ' + err.message,
|
|
time: new Date(),
|
|
severity: 'error',
|
|
});
|
|
options.done();
|
|
return;
|
|
}
|
|
const db = await getScriptableDb(pool);
|
|
|
|
let exprValue;
|
|
try {
|
|
exprValue = func(db, ObjectId);
|
|
} catch (err) {
|
|
options.info({
|
|
message: 'Error evaluating expression: ' + err.message,
|
|
time: new Date(),
|
|
severity: 'error',
|
|
});
|
|
options.done();
|
|
return;
|
|
}
|
|
|
|
if (exprValue instanceof AbstractCursor) {
|
|
await readCursor(exprValue, options);
|
|
} else if (isPromise(exprValue)) {
|
|
try {
|
|
const resValue = await exprValue;
|
|
|
|
options.info({
|
|
message: 'Command succesfully executed',
|
|
time: new Date(),
|
|
severity: 'info',
|
|
});
|
|
try {
|
|
options.info({
|
|
message: `Result: ${JSON.stringify(resValue)}`,
|
|
time: new Date(),
|
|
severity: 'info',
|
|
});
|
|
} catch (err) {
|
|
options.info({
|
|
message: `Result: ${resValue}`,
|
|
time: new Date(),
|
|
severity: 'info',
|
|
});
|
|
}
|
|
|
|
const arrayRes = findArrayResult(resValue);
|
|
if (arrayRes) {
|
|
options.recordset({ __isDynamicStructure: true });
|
|
for (const row of arrayRes) {
|
|
options.row(row);
|
|
}
|
|
}
|
|
} catch (err) {
|
|
options.info({
|
|
message: 'Error when running command: ' + err.message,
|
|
time: new Date(),
|
|
severity: 'error',
|
|
});
|
|
}
|
|
}
|
|
|
|
options.done();
|
|
},
|
|
async readQuery(pool, sql, structure) {
|
|
try {
|
|
const json = JSON.parse(sql);
|
|
if (json && json.pureName) {
|
|
sql = `db.${json.pureName}.find()`;
|
|
}
|
|
} catch (err) {
|
|
// query is not JSON serialized collection name
|
|
}
|
|
|
|
// const pass = new stream.PassThrough({
|
|
// objectMode: true,
|
|
// highWaterMark: 100,
|
|
// });
|
|
|
|
func = eval(`(db,ObjectId) => ${sql}`);
|
|
const db = await getScriptableDb(pool);
|
|
exprValue = func(db, ObjectId);
|
|
|
|
// return directly stream without header row
|
|
return exprValue.stream();
|
|
|
|
// pass.write(structure || { __isDynamicStructure: true });
|
|
// exprValue.on('data', (row) => pass.write(row));
|
|
// exprValue.on('end', () => pass.end());
|
|
|
|
// return pass;
|
|
},
|
|
async writeTable(pool, name, options) {
|
|
return createBulkInsertStream(this, stream, pool, name, options);
|
|
},
|
|
async getVersion(pool) {
|
|
const status = await pool.__getDatabase().admin().serverInfo();
|
|
return {
|
|
...status,
|
|
versionText: `MongoDB ${status.version}`,
|
|
};
|
|
},
|
|
async listDatabases(pool) {
|
|
const res = await pool.__getDatabase().admin().listDatabases();
|
|
return res.databases;
|
|
},
|
|
async readCollection(pool, options) {
|
|
try {
|
|
const collection = pool.__getDatabase().collection(options.pureName);
|
|
if (options.countDocuments) {
|
|
const count = await collection.countDocuments(convertObjectId(options.condition) || {});
|
|
return { count };
|
|
} else if (options.aggregate) {
|
|
let cursor = await collection.aggregate(options.aggregate);
|
|
const rows = await cursor.toArray();
|
|
return { rows: rows.map(transformMongoData) };
|
|
} else {
|
|
// console.log('options.condition', JSON.stringify(options.condition, undefined, 2));
|
|
let cursor = await collection.find(convertObjectId(options.condition) || {});
|
|
if (options.sort) cursor = cursor.sort(options.sort);
|
|
if (options.skip) cursor = cursor.skip(options.skip);
|
|
if (options.limit) cursor = cursor.limit(options.limit);
|
|
const rows = await cursor.toArray();
|
|
return { rows: rows.map(transformMongoData) };
|
|
}
|
|
} catch (err) {
|
|
return { errorMessage: err.message };
|
|
}
|
|
},
|
|
async updateCollection(pool, changeSet) {
|
|
const res = {
|
|
inserted: [],
|
|
updated: [],
|
|
deleted: [],
|
|
replaced: [],
|
|
};
|
|
try {
|
|
const db = pool.__getDatabase();
|
|
for (const insert of changeSet.inserts) {
|
|
const collection = db.collection(insert.pureName);
|
|
const document = {
|
|
...insert.document,
|
|
...insert.fields,
|
|
};
|
|
const resdoc = await collection.insert(convertObjectId(document));
|
|
res.inserted.push(resdoc._id);
|
|
}
|
|
for (const update of changeSet.updates) {
|
|
const collection = db.collection(update.pureName);
|
|
if (update.document) {
|
|
const document = {
|
|
...update.document,
|
|
...update.fields,
|
|
};
|
|
const doc = await collection.findOne(convertObjectId(update.condition));
|
|
if (doc) {
|
|
const resdoc = await collection.replaceOne(convertObjectId(update.condition), {
|
|
...convertObjectId(document),
|
|
_id: doc._id,
|
|
});
|
|
res.replaced.push(resdoc._id);
|
|
}
|
|
} else {
|
|
const resdoc = await collection.updateOne(convertObjectId(update.condition), {
|
|
$set: convertObjectId(update.fields),
|
|
});
|
|
res.updated.push(resdoc._id);
|
|
}
|
|
}
|
|
for (const del of changeSet.deletes) {
|
|
const collection = db.collection(del.pureName);
|
|
const resdoc = await collection.deleteOne(convertObjectId(del.condition));
|
|
res.deleted.push(resdoc._id);
|
|
}
|
|
return res;
|
|
} catch (err) {
|
|
return { errorMessage: err.message };
|
|
}
|
|
},
|
|
|
|
async createDatabase(pool, name) {
|
|
const db = pool.db(name);
|
|
await db.createCollection('collection1');
|
|
},
|
|
|
|
async dropDatabase(pool, name) {
|
|
const db = pool.db(name);
|
|
await db.dropDatabase();
|
|
},
|
|
|
|
async loadFieldValues(pool, name, field, search) {
|
|
try {
|
|
const collection = pool.__getDatabase().collection(name.pureName);
|
|
// console.log('options.condition', JSON.stringify(options.condition, undefined, 2));
|
|
|
|
const pipelineMatch = [];
|
|
|
|
if (search) {
|
|
const tokens = _.compact(search.split(' ').map((x) => x.trim()));
|
|
if (tokens.length > 0) {
|
|
pipelineMatch.push({
|
|
$match: {
|
|
$and: tokens.map((token) => ({
|
|
[field]: {
|
|
$regex: `.*${token}.*`,
|
|
$options: 'i',
|
|
},
|
|
})),
|
|
},
|
|
});
|
|
}
|
|
}
|
|
|
|
let cursor = await collection.aggregate([
|
|
...pipelineMatch,
|
|
{
|
|
$group: { _id: '$' + field },
|
|
},
|
|
{
|
|
$sort: { _id: 1 },
|
|
},
|
|
{
|
|
$limit: 100,
|
|
},
|
|
]);
|
|
const rows = await cursor.toArray();
|
|
return _.uniqBy(
|
|
rows.map(transformMongoData).map(({ _id }) => {
|
|
if (_.isArray(_id) || _.isPlainObject(_id)) return { value: null };
|
|
return { value: _id };
|
|
}),
|
|
(x) => x.value
|
|
);
|
|
} catch (err) {
|
|
return { errorMessage: err.message };
|
|
}
|
|
},
|
|
|
|
readJsonQuery(pool, select, structure) {
|
|
const { collection, condition, sort } = select;
|
|
|
|
const db = pool.__getDatabase();
|
|
const res = db
|
|
.collection(collection)
|
|
.find(condition || {})
|
|
.sort(sort || {})
|
|
.stream();
|
|
|
|
return res;
|
|
},
|
|
};
|
|
|
|
module.exports = driver;
|