Compare commits
8 Commits
a14c976019
...
8085a3fbf0
Author | SHA1 | Date |
---|---|---|
laozhoubuluo | 8085a3fbf0 | |
naskya | 82c98ae72f | |
naskya | 5b3f93457b | |
naskya | 4d9c0f8e7b | |
naskya | bf2b624bc9 | |
naskya | 5261eb24b6 | |
naskya | d440e9b388 | |
老周部落 | 094f705d4f |
|
@ -8,10 +8,11 @@ services:
|
|||
|
||||
workflow:
|
||||
rules:
|
||||
- if: $CI_PIPELINE_SOURCE == 'merge_request_event'
|
||||
- if: $CI_PROJECT_PATH == 'firefish/firefish'
|
||||
when: always
|
||||
- if: $CI_MERGE_REQUEST_TARGET_BRANCH_NAME == 'main'
|
||||
when: never
|
||||
- if: $CI_MERGE_REQUEST_PROJECT_PATH == 'firefish/firefish'
|
||||
when: always
|
||||
- when: never
|
||||
|
||||
cache:
|
||||
paths:
|
||||
|
@ -23,6 +24,8 @@ cache:
|
|||
|
||||
stages:
|
||||
- test
|
||||
- build
|
||||
- dependency
|
||||
|
||||
variables:
|
||||
POSTGRES_DB: 'firefish_db'
|
||||
|
@ -36,7 +39,6 @@ variables:
|
|||
|
||||
default:
|
||||
before_script:
|
||||
- mkdir -p "${CARGO_HOME}"
|
||||
- apt-get update && apt-get -y upgrade
|
||||
- apt-get -y --no-install-recommends install curl
|
||||
- curl -fsSL 'https://deb.nodesource.com/setup_18.x' | bash -
|
||||
|
@ -48,10 +50,71 @@ default:
|
|||
- export PGPASSWORD="${POSTGRES_PASSWORD}"
|
||||
- psql --host postgres --user "${POSTGRES_USER}" --dbname "${POSTGRES_DB}" --command 'CREATE EXTENSION pgroonga'
|
||||
|
||||
build_and_cargo_unit_test:
|
||||
build_test:
|
||||
stage: test
|
||||
script:
|
||||
- pnpm install --frozen-lockfile
|
||||
- pnpm run build:debug
|
||||
- pnpm run migrate
|
||||
|
||||
container_image_build:
|
||||
stage: build
|
||||
image: docker.io/debian:bookworm-slim
|
||||
services: []
|
||||
before_script: []
|
||||
rules:
|
||||
- if: $CI_COMMIT_BRANCH == 'develop'
|
||||
script:
|
||||
- apt-get update && apt-get -y upgrade
|
||||
- apt-get install -y --no-install-recommends buildah ca-certificates
|
||||
- buildah login --username "${CI_REGISTRY_USER}" --password "${CI_REGISTRY_PASSWORD}" "${CI_REGISTRY}"
|
||||
- buildah build --security-opt seccomp=unconfined --cap-add all --tag "${CI_REGISTRY}/${CI_PROJECT_PATH}/develop:not-for-production" --platform linux/amd64 .
|
||||
- buildah push "${CI_REGISTRY}/${CI_PROJECT_PATH}/develop:not-for-production" "docker://${CI_REGISTRY}/${CI_PROJECT_PATH}/develop:not-for-production"
|
||||
|
||||
cargo_unit_test:
|
||||
stage: test
|
||||
rules:
|
||||
- if: $CI_PIPELINE_SOURCE == 'merge_request_event'
|
||||
changes:
|
||||
paths:
|
||||
- packages/backend-rs/**/*
|
||||
- packages/macro-rs/**/*
|
||||
- Cargo.toml
|
||||
- Cargo.lock
|
||||
- if: $CI_MERGE_REQUEST_TARGET_BRANCH_NAME == 'main'
|
||||
when: never
|
||||
script:
|
||||
- cargo check --features napi
|
||||
- pnpm install --frozen-lockfile
|
||||
- mkdir packages/backend-rs/built
|
||||
- cp packages/backend-rs/index.js packages/backend-rs/built/index.js
|
||||
- cp packages/backend-rs/index.d.ts packages/backend-rs/built/index.d.ts
|
||||
- pnpm --filter='!backend-rs' run build:debug
|
||||
- cargo test
|
||||
|
||||
cargo_clippy:
|
||||
stage: test
|
||||
rules:
|
||||
- if: $CI_PIPELINE_SOURCE == 'merge_request_event'
|
||||
changes:
|
||||
paths:
|
||||
- packages/backend-rs/**/*
|
||||
- packages/macro-rs/**/*
|
||||
- Cargo.toml
|
||||
- Cargo.lock
|
||||
- if: $CI_MERGE_REQUEST_TARGET_BRANCH_NAME == 'main'
|
||||
when: never
|
||||
script:
|
||||
- cargo clippy -- -D warnings
|
||||
|
||||
renovate:
|
||||
stage: dependency
|
||||
image:
|
||||
name: docker.io/renovate/renovate:37-slim
|
||||
entrypoint: [""]
|
||||
rules:
|
||||
- if: $RENOVATE && $CI_PIPELINE_SOURCE == 'schedule'
|
||||
services: []
|
||||
before_script: []
|
||||
script:
|
||||
- renovate --platform gitlab --token "${API_TOKEN}" --endpoint "${CI_SERVER_URL}/api/v4" "${CI_PROJECT_PATH}"
|
||||
|
|
|
@ -74,6 +74,34 @@ mentions: "Mentions"
|
|||
directNotes: "Direct messages"
|
||||
cw: "Content warning"
|
||||
importAndExport: "Import/Export Data"
|
||||
importAndExportWarn: "The Import/Export Data feature is an experimental feature and
|
||||
implementation may change at any time without prior notice.\n
|
||||
Due to differences in the exported data of different software versions, the actual
|
||||
conditions of the import program, and the server health of the exported data link,
|
||||
the imported data may be incomplete or the access permissions may not be set
|
||||
correctly (for example, there is no access permission mark in the
|
||||
Mastodon/Akkoma/Pleroma exported data, so all posts makes public after import),
|
||||
so please be sure to check the imported data carefully integrity and configure
|
||||
the correct access permissions for it."
|
||||
importAndExportInfo: "Since some data cannot be obtained after the original account is
|
||||
frozen or the original server goes offline, it is strongly recommendedthat you import
|
||||
the data before the original account is frozen (migrated, logged out) or the original
|
||||
server goes offline.\n
|
||||
If the original account is frozen or the original server is offline but you have the
|
||||
original images, you can try uploading them to the network disk before importing the
|
||||
data, which may help with data import.\n
|
||||
Since some data is obtained from its server using your current account when importing
|
||||
data, data that the current account does not have permission to access will be regarded
|
||||
as broken. Please make adjustments including but not limited to access permissions,
|
||||
Manually following accounts and other methods allow the current account to obtain
|
||||
relevant data, so that the import program can normally obtain the data it needs to
|
||||
obtain to help you import.\n
|
||||
Since it is impossible to confirm whether the broken link content posted by someone other
|
||||
than you is posted by him/her, if there is broken link content posted by others in the
|
||||
discussion thread, the related content and subsequent replies will not be imported.\n
|
||||
Since data import is greatly affected by network communication, it is recommended that you
|
||||
pay attention to data recovery after a period of time. If the data is still not restored,
|
||||
you can try importing the same backup file again and try again."
|
||||
import: "Import"
|
||||
export: "Export"
|
||||
files: "Files"
|
||||
|
|
|
@ -61,6 +61,16 @@ mention: "提及"
|
|||
mentions: "提及"
|
||||
directNotes: "私信"
|
||||
importAndExport: "导入 / 导出数据"
|
||||
importAndExportWarn: "导入 / 导出数据功能是一项实验性功能,实现可能会随时变化而无预先通知。\n
|
||||
由于不同软件不同版本的导出数据、导入程序实际情况以及导出数据链接的服务器运行状况不同,导入的数据可能会不完整或未被正确设置访问权限
|
||||
(例如 Mastodon/Akkoma/Pleroma 导出数据内无访问权限标记,因此所有帖子导入后均为公开状态),因此请务必谨慎核对导入数据的完整性,
|
||||
并为其配置正确的访问权限。"
|
||||
importAndExportInfo: "由于原账号冻结或者原服务器下线后部分数据无法获取,因此强烈建议您在原账号冻结(迁移、注销)或原服务器下线前导入数据。\n
|
||||
在原账号冻结或者原服务器下线但您拥有原始图片的情况下,可以尝试在导入数据之前将其上传到网盘上,可能对数据导入有所帮助。\n
|
||||
由于导入数据时部分数据是使用您当前账号到其服务器上获取,因此当前账号无权访问的数据会视为断链。请通过包括但不限于访问权限调整、
|
||||
手动关注账户等方式让当前帐号可以获取到相关数据,以便导入程序能够正常获取到需要获取的数据从而帮助您进行导入。\n
|
||||
由于无法确认非您本人发表的断链内容的是否由其本人发表,因此如果讨论串内有其他人发表的断链内容,则相关内容以及后续回复不会被导入。\n
|
||||
由于数据导入受网络通信影响较大,因此建议您一段时间之后再关注数据恢复情况。如果数据仍未恢复可以尝试再次导入同样的备份文件重试一次。"
|
||||
import: "导入"
|
||||
export: "导出"
|
||||
files: "文件"
|
||||
|
|
|
@ -335,6 +335,7 @@ export function createImportMastoPostJob(
|
|||
user: ThinUser,
|
||||
post: any,
|
||||
signatureCheck: boolean,
|
||||
parent: Note | null = null,
|
||||
) {
|
||||
return dbQueue.add(
|
||||
"importMastoPost",
|
||||
|
@ -342,6 +343,7 @@ export function createImportMastoPostJob(
|
|||
user: user,
|
||||
post: post,
|
||||
signatureCheck: signatureCheck,
|
||||
parent: parent,
|
||||
},
|
||||
{
|
||||
removeOnComplete: true,
|
||||
|
|
|
@ -11,6 +11,7 @@ import type { Poll } from "@/models/entities/poll.js";
|
|||
import type { DbUserJobData } from "@/queue/types.js";
|
||||
import { createTemp } from "@/misc/create-temp.js";
|
||||
import { inspect } from "node:util";
|
||||
import { config } from "@/config.js";
|
||||
|
||||
const logger = queueLogger.createSubLogger("export-notes");
|
||||
|
||||
|
@ -131,5 +132,6 @@ async function serialize(
|
|||
visibility: note.visibility,
|
||||
visibleUserIds: note.visibleUserIds,
|
||||
localOnly: note.localOnly,
|
||||
objectUrl: `${config.url}/notes/${note.id}`,
|
||||
};
|
||||
}
|
||||
|
|
|
@ -1,12 +1,14 @@
|
|||
import * as Post from "@/misc/post.js";
|
||||
import create from "@/services/note/create.js";
|
||||
import { NoteFiles, Users } from "@/models/index.js";
|
||||
import Resolver from "@/remote/activitypub/resolver.js";
|
||||
import { DriveFiles, NoteFiles, Users } from "@/models/index.js";
|
||||
import type { DbUserImportMastoPostJobData } from "@/queue/types.js";
|
||||
import { queueLogger } from "../../logger.js";
|
||||
import { uploadFromUrl } from "@/services/drive/upload-from-url.js";
|
||||
import type { DriveFile } from "@/models/entities/drive-file.js";
|
||||
import type Bull from "bull";
|
||||
import { createImportCkPostJob } from "@/queue/index.js";
|
||||
import { resolveNote } from "@/remote/activitypub/models/note.js";
|
||||
import { Notes, NoteEdits } from "@/models/index.js";
|
||||
import type { Note } from "@/models/entities/note.js";
|
||||
import { genId } from "backend-rs";
|
||||
|
@ -23,20 +25,37 @@ export async function importCkPost(
|
|||
return;
|
||||
}
|
||||
const post = job.data.post;
|
||||
/*
|
||||
if (post.replyId != null) {
|
||||
done();
|
||||
return;
|
||||
const parent = job.data.parent;
|
||||
const isRenote = post.renoteId !== null;
|
||||
let reply: Note | null = null;
|
||||
let renote: Note | null = null;
|
||||
job.progress(20);
|
||||
if (!isRenote && post.replyId !== null) {
|
||||
if (
|
||||
!parent &&
|
||||
typeof post.objectUrl !== "undefined" &&
|
||||
post.objectUrl !== null
|
||||
) {
|
||||
const resolver = new Resolver();
|
||||
const originalNote = await resolver.resolve(post.objectUrl);
|
||||
reply = await resolveNote(originalNote.inReplyTo);
|
||||
} else {
|
||||
reply = post.replyId !== null ? parent : null;
|
||||
}
|
||||
}
|
||||
if (post.renoteId != null) {
|
||||
done();
|
||||
return;
|
||||
// renote also need resolve original note
|
||||
if (
|
||||
isRenote &&
|
||||
!parent &&
|
||||
typeof post.objectUrl !== "undefined" &&
|
||||
post.objectUrl !== null
|
||||
) {
|
||||
const resolver = new Resolver();
|
||||
const originalNote = await resolver.resolve(post.objectUrl);
|
||||
renote = await resolveNote(originalNote.quoteUrl);
|
||||
} else {
|
||||
renote = isRenote ? parent : null;
|
||||
}
|
||||
if (post.visibility !== "public") {
|
||||
done();
|
||||
return;
|
||||
}
|
||||
*/
|
||||
const urls = (post.files || [])
|
||||
.map((x: any) => x.url)
|
||||
.filter((x: String) => x.startsWith("http"));
|
||||
|
@ -49,7 +68,17 @@ export async function importCkPost(
|
|||
});
|
||||
files.push(file);
|
||||
} catch (e) {
|
||||
logger.info(`Skipped adding file to drive: ${url}`);
|
||||
// try to get the same md5 file from user drive
|
||||
const md5 = post.files.map((x: any) => x.url).find(url).md5;
|
||||
const much = await DriveFiles.findOneBy({
|
||||
md5: md5,
|
||||
userId: user.id,
|
||||
});
|
||||
if (much) {
|
||||
files.push(much);
|
||||
} else {
|
||||
logger.info(`Skipped adding file to drive: ${url}`);
|
||||
}
|
||||
}
|
||||
}
|
||||
const { text, cw, localOnly, createdAt, visibility } = Post.parse(post);
|
||||
|
@ -88,8 +117,8 @@ export async function importCkPost(
|
|||
files: files.length === 0 ? undefined : files,
|
||||
poll: undefined,
|
||||
text: text || undefined,
|
||||
reply: post.replyId ? job.data.parent : null,
|
||||
renote: post.renoteId ? job.data.parent : null,
|
||||
reply,
|
||||
renote,
|
||||
cw: cw,
|
||||
localOnly,
|
||||
visibility: visibility,
|
||||
|
|
|
@ -10,6 +10,7 @@ import type { DriveFile } from "@/models/entities/drive-file.js";
|
|||
import { Notes, NoteEdits } from "@/models/index.js";
|
||||
import type { Note } from "@/models/entities/note.js";
|
||||
import { genId } from "backend-rs";
|
||||
import { createImportMastoPostJob } from "@/queue/index.js";
|
||||
|
||||
const logger = queueLogger.createSubLogger("import-masto-post");
|
||||
|
||||
|
@ -23,12 +24,17 @@ export async function importMastoPost(
|
|||
return;
|
||||
}
|
||||
const post = job.data.post;
|
||||
const parent = job.data.parent;
|
||||
const isRenote = post.type === "Announce";
|
||||
let reply: Note | null = null;
|
||||
let renote: Note | null = null;
|
||||
job.progress(20);
|
||||
if (!isRenote && post.object.inReplyTo != null) {
|
||||
reply = await resolveNote(post.object.inReplyTo);
|
||||
if (parent == null) {
|
||||
reply = await resolveNote(post.object.inReplyTo);
|
||||
} else {
|
||||
reply = parent;
|
||||
}
|
||||
}
|
||||
// renote also need resolve original note
|
||||
if (isRenote) {
|
||||
|
@ -135,4 +141,14 @@ export async function importMastoPost(
|
|||
done();
|
||||
|
||||
logger.info("Imported");
|
||||
if (post.childNotes) {
|
||||
for (const child of post.childNotes) {
|
||||
createImportMastoPostJob(
|
||||
job.data.user,
|
||||
child,
|
||||
job.data.signatureCheck,
|
||||
note,
|
||||
);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -40,7 +40,10 @@ export async function importPosts(
|
|||
file.url,
|
||||
job.data.user.id,
|
||||
);
|
||||
for (const post of outbox.orderedItems) {
|
||||
logger.info("Parsing mastodon style posts");
|
||||
const arr = recreateChainForMastodon(outbox.orderedItems);
|
||||
logger.debug(JSON.stringify(arr, null, 2));
|
||||
for (const post of arr) {
|
||||
createImportMastoPostJob(job.data.user, post, job.data.signatureCheck);
|
||||
}
|
||||
} catch (e) {
|
||||
|
@ -60,12 +63,15 @@ export async function importPosts(
|
|||
if (Array.isArray(parsed)) {
|
||||
logger.info("Parsing *key posts");
|
||||
const arr = recreateChain(parsed);
|
||||
logger.debug(JSON.stringify(arr, null, 2));
|
||||
for (const post of arr) {
|
||||
createImportCkPostJob(job.data.user, post, job.data.signatureCheck);
|
||||
}
|
||||
} else if (parsed instanceof Object) {
|
||||
logger.info("Parsing Mastodon posts");
|
||||
for (const post of parsed.orderedItems) {
|
||||
const arr = recreateChainForMastodon(parsed.orderedItems);
|
||||
logger.debug(JSON.stringify(arr, null, 2));
|
||||
for (const post of arr) {
|
||||
createImportMastoPostJob(job.data.user, post, job.data.signatureCheck);
|
||||
}
|
||||
}
|
||||
|
@ -96,9 +102,56 @@ function recreateChain(arr: any[]): any {
|
|||
let parent = null;
|
||||
if (note.replyId != null) {
|
||||
parent = lookup[`${note.replyId}`];
|
||||
// Accept URL, let import process to resolveNote
|
||||
if (
|
||||
!parent &&
|
||||
typeof note.objectUrl !== "undefined" &&
|
||||
note.objectUrl.startsWith("http")
|
||||
) {
|
||||
notesTree.push(note);
|
||||
}
|
||||
}
|
||||
if (note.renoteId != null) {
|
||||
parent = lookup[`${note.renoteId}`];
|
||||
// Accept URL, let import process to resolveNote
|
||||
if (
|
||||
!parent &&
|
||||
typeof note.objectUrl !== "undefined" &&
|
||||
note.objectUrl.startsWith("http")
|
||||
) {
|
||||
notesTree.push(note);
|
||||
}
|
||||
}
|
||||
|
||||
if (parent) {
|
||||
parent.childNotes.push(note);
|
||||
}
|
||||
}
|
||||
return notesTree;
|
||||
}
|
||||
|
||||
function recreateChainForMastodon(arr: any[]): any {
|
||||
type NotesMap = {
|
||||
[id: string]: any;
|
||||
};
|
||||
const notesTree: any[] = [];
|
||||
const lookup: NotesMap = {};
|
||||
for (const note of arr) {
|
||||
lookup[`${note.id}`] = note;
|
||||
note.childNotes = [];
|
||||
if (note.object.inReplyTo == null) {
|
||||
notesTree.push(note);
|
||||
}
|
||||
}
|
||||
for (const note of arr) {
|
||||
let parent = null;
|
||||
if (note.object.inReplyTo != null) {
|
||||
const inReplyToIdForLookup = `${note.object.inReplyTo}/activity`;
|
||||
parent = lookup[`${inReplyToIdForLookup}`];
|
||||
// Accept URL, let import process to resolveNote
|
||||
if (!parent && note.object.inReplyTo.startsWith("http")) {
|
||||
notesTree.push(note);
|
||||
}
|
||||
}
|
||||
|
||||
if (parent) {
|
||||
|
|
|
@ -1,5 +1,14 @@
|
|||
<template>
|
||||
<div class="_formRoot">
|
||||
<FormSection>
|
||||
<template #label>{{ i18n.ts.importAndExport }}</template>
|
||||
<FormInfo warn class="_formBlock">{{
|
||||
i18n.ts.importAndExportWarn
|
||||
}}</FormInfo>
|
||||
<FormInfo class="_formBlock">{{
|
||||
i18n.ts.importAndExportInfo
|
||||
}}</FormInfo>
|
||||
</FormSection>
|
||||
<FormSection>
|
||||
<template #label>{{ i18n.ts._exportOrImport.allNotes }}</template>
|
||||
<FormFolder>
|
||||
|
@ -177,6 +186,7 @@
|
|||
|
||||
<script lang="ts" setup>
|
||||
import { ref } from "vue";
|
||||
import FormInfo from "@/components/MkInfo.vue";
|
||||
import MkButton from "@/components/MkButton.vue";
|
||||
import FormSection from "@/components/form/section.vue";
|
||||
import FormFolder from "@/components/form/folder.vue";
|
||||
|
|
|
@ -0,0 +1,15 @@
|
|||
{
|
||||
"$schema": "https://docs.renovatebot.com/renovate-schema.json",
|
||||
"extends": ["config:base"],
|
||||
"rangeStrategy": "bump",
|
||||
"branchConcurrentLimit": 5,
|
||||
"enabledManagers": ["npm", "cargo"],
|
||||
"baseBranches": ["develop"],
|
||||
"lockFileMaintenance": {
|
||||
"enabled": true,
|
||||
"recreateWhen": "always",
|
||||
"rebaseStalePrs": true,
|
||||
"branchTopic": "lock-file-maintenance",
|
||||
"commitMessageAction": "Lock file maintenance"
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue