Ghost/core/server/models/base/bulk-operations.js
Nazar Gargol 95044e3ba0 Added error handling for failed member imports
no issue

- When bulk insert fails there is no transactional logic to revert
related records form being inserted. Also, previously there were no
attempts to "retry" the insert.
- To avoid complex retry logic, an iterative one-by-one insert retry
approach was taken. If this becomes a bottleneck in the future, the
retry algorithm could be improved.
- To avoid a lot of code duplication refactored model's `bulkAdd` & `bulkDestroy`
methods to use 'bulk-operations' module.
- Updated error handling and logging for bulk delete operations. It's very
unlikely for error to happen here,  but still need to make sure there is
a proper logging in place to trace back the failure.
- Added debug logs. This should improve debugging experience and
performance measurements.
- Added handling for unrecognized errors. Handling inspired by current unrecognized
error handling by ghost importer -10e5d5f3d4/core/server/data/importer/importers/data/base.js (L148-L154)
2020-08-26 17:11:35 +12:00

92 lines
2.3 KiB
JavaScript

const _ = require('lodash');
const errors = require('@tryghost/errors');
const db = require('../../data/db');
const logging = require('../../../shared/logging');
const CHUNK_SIZE = 100;
async function insertChunkSequential(table, chunk, result) {
for (const record of chunk) {
try {
await db.knex(table).insert(record);
result.successful += 1;
} catch (err) {
err.errorDetails = record;
result.errors.push(err);
result.unsuccessfulRecords.push(record);
result.unsuccessful += 1;
}
}
}
async function insertChunk(table, chunk, result) {
try {
await db.knex(table).insert(chunk);
result.successful += chunk.length;
} catch (err) {
await insertChunkSequential(table, chunk, result);
}
}
async function insert(table, data) {
const result = {
successful: 0,
unsuccessful: 0,
unsuccessfulRecords: [],
errors: []
};
for (const chunk of _.chunk(data, CHUNK_SIZE)) {
await insertChunk(table, chunk, result);
}
return result;
}
async function delChunkSequential(table, chunk, result) {
for (const id of chunk) {
try {
await db.knex(table).where('id', id).del();
result.successful += 1;
} catch (err) {
const importError = new errors.DataImportError({
message: `Failed to remove entry from ${table}`,
context: `Entry id: ${id}`,
err: err
});
logging.error(importError);
result.errors.push(importError);
result.unsuccessfulIds.push(id);
result.unsuccessful += 1;
}
}
}
async function delChunk(table, chunk, result) {
try {
await db.knex(table).whereIn('id', chunk).del();
result.successful += chunk.length;
} catch (err) {
await delChunkSequential(table, chunk, result);
}
}
async function del(table, ids) {
const result = {
successful: 0,
unsuccessful: 0,
unsuccessfulIds: [],
errors: []
};
for (const chunk of _.chunk(ids, CHUNK_SIZE)) {
await delChunk(table, chunk, result);
}
return result;
}
module.exports.insert = insert;
module.exports.del = del;