* chore: Refactor collection export to use FileOperations processor and task * Tweak options
130 lines
3.4 KiB
TypeScript
130 lines
3.4 KiB
TypeScript
import fs from "fs";
|
|
import invariant from "invariant";
|
|
import { truncate } from "lodash";
|
|
import ExportFailureEmail from "@server/emails/templates/ExportFailureEmail";
|
|
import ExportSuccessEmail from "@server/emails/templates/ExportSuccessEmail";
|
|
import Logger from "@server/logging/logger";
|
|
import { Collection, Event, FileOperation, Team, User } from "@server/models";
|
|
import { FileOperationState } from "@server/models/FileOperation";
|
|
import { uploadToS3FromBuffer } from "@server/utils/s3";
|
|
import { archiveCollections } from "@server/utils/zip";
|
|
import BaseTask, { TaskPriority } from "./BaseTask";
|
|
|
|
type Props = {
|
|
fileOperationId: string;
|
|
};
|
|
|
|
export default class ExportMarkdownZipTask extends BaseTask<Props> {
|
|
/**
|
|
* Runs the export task.
|
|
*
|
|
* @param props The props
|
|
*/
|
|
public async perform({ fileOperationId }: Props) {
|
|
const fileOperation = await FileOperation.findByPk(fileOperationId);
|
|
invariant(fileOperation, "fileOperation not found");
|
|
|
|
const [team, user] = await Promise.all([
|
|
Team.findByPk(fileOperation.teamId),
|
|
User.findByPk(fileOperation.userId),
|
|
]);
|
|
invariant(team, "team operation not found");
|
|
invariant(user, "user operation not found");
|
|
|
|
const collectionIds = fileOperation.collectionId
|
|
? [fileOperation.collectionId]
|
|
: await user.collectionIds();
|
|
|
|
const collections = await Collection.findAll({
|
|
where: {
|
|
id: collectionIds,
|
|
},
|
|
});
|
|
|
|
try {
|
|
Logger.info("task", `ExportTask processing data for ${fileOperationId}`);
|
|
|
|
await this.updateFileOperation(
|
|
fileOperation,
|
|
FileOperationState.Creating
|
|
);
|
|
|
|
const filePath = await archiveCollections(collections);
|
|
|
|
Logger.info("task", `ExportTask uploading data for ${fileOperationId}`);
|
|
|
|
await this.updateFileOperation(
|
|
fileOperation,
|
|
FileOperationState.Uploading
|
|
);
|
|
|
|
const fileBuffer = await fs.promises.readFile(filePath);
|
|
const url = await uploadToS3FromBuffer(
|
|
fileBuffer,
|
|
"application/zip",
|
|
fileOperation.key,
|
|
"private"
|
|
);
|
|
|
|
await this.updateFileOperation(
|
|
fileOperation,
|
|
FileOperationState.Complete,
|
|
undefined,
|
|
url
|
|
);
|
|
|
|
await ExportSuccessEmail.schedule({
|
|
to: user.email,
|
|
id: fileOperation.id,
|
|
teamUrl: team.url,
|
|
});
|
|
} catch (error) {
|
|
await this.updateFileOperation(
|
|
fileOperation,
|
|
FileOperationState.Error,
|
|
error
|
|
);
|
|
await ExportFailureEmail.schedule({
|
|
to: user.email,
|
|
teamUrl: team.url,
|
|
});
|
|
throw error;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Update the state of the underlying FileOperation in the database and send
|
|
* an event to the client.
|
|
*
|
|
* @param fileOperation The FileOperation to update
|
|
*/
|
|
private async updateFileOperation(
|
|
fileOperation: FileOperation,
|
|
state: FileOperationState,
|
|
error?: Error,
|
|
url?: string
|
|
) {
|
|
await fileOperation.update({
|
|
state,
|
|
url,
|
|
error: error ? truncate(error.message, { length: 255 }) : undefined,
|
|
});
|
|
await Event.schedule({
|
|
name: "fileOperations.update",
|
|
modelId: fileOperation.id,
|
|
teamId: fileOperation.teamId,
|
|
actorId: fileOperation.userId,
|
|
});
|
|
}
|
|
|
|
/**
|
|
* Job options such as priority and retry strategy, as defined by Bull.
|
|
*/
|
|
public get options() {
|
|
return {
|
|
priority: TaskPriority.Background,
|
|
attempts: 2,
|
|
};
|
|
}
|
|
}
|