mirror of
https://github.com/ankitects/anki.git
synced 2025-09-18 14:02:21 -04:00

* Implement colpkg exporting on backend * Use exporting logic in backup.rs * Refactor exporting.rs * Add backend function to export collection * Refactor backend/collection.rs * Use backend for colpkg exporting * Don't use default zip compression for media * Add exporting progress * Refactor media file writing * Write dummy collections * Localize dummy collection note * Minimize dummy db size * Use `NamedTempFile::new()` instead of `new_in` * Drop redundant v2 dummy collection * COLLECTION_VERSION -> PACKAGE_VERSION * Split `lock_collection()` into two to drop flag * Expose new colpkg in GUI * Improve dummy collection message * Please type checker * importing-colpkg-too-new -> exporting-... * Compress the media map in the v3 package (dae) On collections with lots of media, it can grow into megabytes. Also return an error in extract_media_file_names(), instead of masking it as an optional. * Store media map as a vector in the v3 package (dae) This compresses better (eg 280kb original, 100kb hashmap, 42kb vec) In the colpkg import case we don't need random access. When importing an apkg, we will need to be able to fetch file data for a given media filename, but the existing map doesn't help us there, as we need filename->index, not index->filename. * Ensure folders in the media dir don't break the file mapping (dae)
481 lines
15 KiB
Python
481 lines
15 KiB
Python
# Copyright: Ankitects Pty Ltd and contributors
|
|
# License: GNU AGPL, version 3 or later; http://www.gnu.org/licenses/agpl.html
|
|
|
|
# pylint: disable=invalid-name
|
|
|
|
from __future__ import annotations
|
|
|
|
import json
|
|
import os
|
|
import re
|
|
import shutil
|
|
import threading
|
|
import time
|
|
import unicodedata
|
|
import zipfile
|
|
from io import BufferedWriter
|
|
from typing import Any, Optional, Sequence
|
|
from zipfile import ZipFile
|
|
|
|
from anki import hooks
|
|
from anki.cards import CardId
|
|
from anki.collection import Collection
|
|
from anki.decks import DeckId
|
|
from anki.utils import ids2str, namedtmp, split_fields, strip_html
|
|
|
|
|
|
class Exporter:
|
|
includeHTML: bool | None = None
|
|
ext: Optional[str] = None
|
|
includeTags: Optional[bool] = None
|
|
includeSched: Optional[bool] = None
|
|
includeMedia: Optional[bool] = None
|
|
|
|
def __init__(
|
|
self,
|
|
col: Collection,
|
|
did: Optional[DeckId] = None,
|
|
cids: Optional[list[CardId]] = None,
|
|
) -> None:
|
|
self.col = col.weakref()
|
|
self.did = did
|
|
self.cids = cids
|
|
|
|
@staticmethod
|
|
def key(col: Collection) -> str:
|
|
return ""
|
|
|
|
def doExport(self, path) -> None:
|
|
raise Exception("not implemented")
|
|
|
|
def exportInto(self, path: str) -> None:
|
|
self._escapeCount = 0
|
|
file = open(path, "wb")
|
|
self.doExport(file)
|
|
file.close()
|
|
|
|
def processText(self, text: str) -> str:
|
|
if self.includeHTML is False:
|
|
text = self.stripHTML(text)
|
|
|
|
text = self.escapeText(text)
|
|
|
|
return text
|
|
|
|
def escapeText(self, text: str) -> str:
|
|
"Escape newlines, tabs, CSS and quotechar."
|
|
# fixme: we should probably quote fields with newlines
|
|
# instead of converting them to spaces
|
|
text = text.replace("\n", " ")
|
|
text = text.replace("\r", "")
|
|
text = text.replace("\t", " " * 8)
|
|
text = re.sub("(?i)<style>.*?</style>", "", text)
|
|
text = re.sub(r"\[\[type:[^]]+\]\]", "", text)
|
|
if '"' in text or "'" in text:
|
|
text = '"' + text.replace('"', '""') + '"'
|
|
return text
|
|
|
|
def stripHTML(self, text: str) -> str:
|
|
# very basic conversion to text
|
|
s = text
|
|
s = re.sub(r"(?i)<(br ?/?|div|p)>", " ", s)
|
|
s = re.sub(r"\[sound:[^]]+\]", "", s)
|
|
s = strip_html(s)
|
|
s = re.sub(r"[ \n\t]+", " ", s)
|
|
s = s.strip()
|
|
return s
|
|
|
|
def cardIds(self) -> Any:
|
|
if self.cids is not None:
|
|
cids = self.cids
|
|
elif not self.did:
|
|
cids = self.col.db.list("select id from cards")
|
|
else:
|
|
cids = self.col.decks.cids(self.did, children=True)
|
|
self.count = len(cids)
|
|
return cids
|
|
|
|
|
|
# Cards as TSV
|
|
######################################################################
|
|
|
|
|
|
class TextCardExporter(Exporter):
|
|
|
|
ext = ".txt"
|
|
includeHTML = True
|
|
|
|
def __init__(self, col) -> None:
|
|
Exporter.__init__(self, col)
|
|
|
|
@staticmethod
|
|
def key(col: Collection) -> str:
|
|
return col.tr.exporting_cards_in_plain_text()
|
|
|
|
def doExport(self, file) -> None:
|
|
ids = sorted(self.cardIds())
|
|
strids = ids2str(ids)
|
|
|
|
def esc(s):
|
|
# strip off the repeated question in answer if exists
|
|
s = re.sub("(?si)^.*<hr id=answer>\n*", "", s)
|
|
return self.processText(s)
|
|
|
|
out = ""
|
|
for cid in ids:
|
|
c = self.col.get_card(cid)
|
|
out += esc(c.question())
|
|
out += "\t" + esc(c.answer()) + "\n"
|
|
file.write(out.encode("utf-8"))
|
|
|
|
|
|
# Notes as TSV
|
|
######################################################################
|
|
|
|
|
|
class TextNoteExporter(Exporter):
|
|
|
|
ext = ".txt"
|
|
includeTags = True
|
|
includeHTML = True
|
|
|
|
def __init__(self, col: Collection) -> None:
|
|
Exporter.__init__(self, col)
|
|
self.includeID = False
|
|
|
|
@staticmethod
|
|
def key(col: Collection) -> str:
|
|
return col.tr.exporting_notes_in_plain_text()
|
|
|
|
def doExport(self, file: BufferedWriter) -> None:
|
|
cardIds = self.cardIds()
|
|
data = []
|
|
for id, flds, tags in self.col.db.execute(
|
|
"""
|
|
select guid, flds, tags from notes
|
|
where id in
|
|
(select nid from cards
|
|
where cards.id in %s)"""
|
|
% ids2str(cardIds)
|
|
):
|
|
row = []
|
|
# note id
|
|
if self.includeID:
|
|
row.append(str(id))
|
|
# fields
|
|
row.extend([self.processText(f) for f in split_fields(flds)])
|
|
# tags
|
|
if self.includeTags:
|
|
row.append(tags.strip())
|
|
data.append("\t".join(row))
|
|
self.count = len(data)
|
|
out = "\n".join(data)
|
|
file.write(out.encode("utf-8"))
|
|
|
|
|
|
# Anki decks
|
|
######################################################################
|
|
# media files are stored in self.mediaFiles, but not exported.
|
|
|
|
|
|
class AnkiExporter(Exporter):
|
|
|
|
ext = ".anki2"
|
|
includeSched: bool | None = False
|
|
includeMedia = True
|
|
|
|
def __init__(self, col: Collection) -> None:
|
|
Exporter.__init__(self, col)
|
|
|
|
@staticmethod
|
|
def key(col: Collection) -> str:
|
|
return col.tr.exporting_anki_20_deck()
|
|
|
|
def deckIds(self) -> list[DeckId]:
|
|
if self.cids:
|
|
return self.col.decks.for_card_ids(self.cids)
|
|
elif self.did:
|
|
return self.src.decks.deck_and_child_ids(self.did)
|
|
else:
|
|
return []
|
|
|
|
def exportInto(self, path: str) -> None:
|
|
# sched info+v2 scheduler not compatible w/ older clients
|
|
self._v2sched = self.col.sched_ver() != 1 and self.includeSched
|
|
|
|
# create a new collection at the target
|
|
try:
|
|
os.unlink(path)
|
|
except OSError:
|
|
pass
|
|
self.dst = Collection(path)
|
|
self.src = self.col
|
|
# find cards
|
|
cids = self.cardIds()
|
|
# copy cards, noting used nids
|
|
nids = {}
|
|
data: list[Sequence] = []
|
|
for row in self.src.db.execute(
|
|
"select * from cards where id in " + ids2str(cids)
|
|
):
|
|
# clear flags
|
|
row = list(row)
|
|
row[-2] = 0
|
|
nids[row[1]] = True
|
|
data.append(row)
|
|
self.dst.db.executemany(
|
|
"insert into cards values (?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?)", data
|
|
)
|
|
# notes
|
|
strnids = ids2str(list(nids.keys()))
|
|
notedata = []
|
|
for row in self.src.db.all("select * from notes where id in " + strnids):
|
|
# remove system tags if not exporting scheduling info
|
|
if not self.includeSched:
|
|
row = list(row)
|
|
row[5] = self.removeSystemTags(row[5])
|
|
notedata.append(row)
|
|
self.dst.db.executemany(
|
|
"insert into notes values (?,?,?,?,?,?,?,?,?,?,?)", notedata
|
|
)
|
|
# models used by the notes
|
|
mids = self.dst.db.list("select distinct mid from notes where id in " + strnids)
|
|
# card history and revlog
|
|
if self.includeSched:
|
|
data = self.src.db.all("select * from revlog where cid in " + ids2str(cids))
|
|
self.dst.db.executemany(
|
|
"insert into revlog values (?,?,?,?,?,?,?,?,?)", data
|
|
)
|
|
else:
|
|
# need to reset card state
|
|
self.dst.sched.reset_cards(cids)
|
|
# models - start with zero
|
|
self.dst.mod_schema(check=False)
|
|
self.dst.models.remove_all_notetypes()
|
|
for m in self.src.models.all():
|
|
if int(m["id"]) in mids:
|
|
self.dst.models.update(m)
|
|
# decks
|
|
dids = self.deckIds()
|
|
dconfs = {}
|
|
for d in self.src.decks.all():
|
|
if str(d["id"]) == "1":
|
|
continue
|
|
if dids and d["id"] not in dids:
|
|
continue
|
|
if not d["dyn"] and d["conf"] != 1:
|
|
if self.includeSched:
|
|
dconfs[d["conf"]] = True
|
|
if not self.includeSched:
|
|
# scheduling not included, so reset deck settings to default
|
|
d = dict(d)
|
|
d["conf"] = 1
|
|
self.dst.decks.update(d)
|
|
# copy used deck confs
|
|
for dc in self.src.decks.all_config():
|
|
if dc["id"] in dconfs:
|
|
self.dst.decks.update_config(dc)
|
|
# find used media
|
|
media = {}
|
|
self.mediaDir = self.src.media.dir()
|
|
if self.includeMedia:
|
|
for row in notedata:
|
|
flds = row[6]
|
|
mid = row[2]
|
|
for file in self.src.media.files_in_str(mid, flds):
|
|
# skip files in subdirs
|
|
if file != os.path.basename(file):
|
|
continue
|
|
media[file] = True
|
|
if self.mediaDir:
|
|
for fname in os.listdir(self.mediaDir):
|
|
path = os.path.join(self.mediaDir, fname)
|
|
if os.path.isdir(path):
|
|
continue
|
|
if fname.startswith("_"):
|
|
# Scan all models in mids for reference to fname
|
|
for m in self.src.models.all():
|
|
if int(m["id"]) in mids:
|
|
if self._modelHasMedia(m, fname):
|
|
media[fname] = True
|
|
break
|
|
self.mediaFiles = list(media.keys())
|
|
self.dst.crt = self.src.crt
|
|
# todo: tags?
|
|
self.count = self.dst.card_count()
|
|
self.postExport()
|
|
self.dst.close(downgrade=True)
|
|
|
|
def postExport(self) -> None:
|
|
# overwrite to apply customizations to the deck before it's closed,
|
|
# such as update the deck description
|
|
pass
|
|
|
|
def removeSystemTags(self, tags: str) -> Any:
|
|
return self.src.tags.rem_from_str("marked leech", tags)
|
|
|
|
def _modelHasMedia(self, model, fname) -> bool:
|
|
# First check the styling
|
|
if fname in model["css"]:
|
|
return True
|
|
# If no reference to fname then check the templates as well
|
|
for t in model["tmpls"]:
|
|
if fname in t["qfmt"] or fname in t["afmt"]:
|
|
return True
|
|
return False
|
|
|
|
|
|
# Packaged Anki decks
|
|
######################################################################
|
|
|
|
|
|
class AnkiPackageExporter(AnkiExporter):
|
|
|
|
ext = ".apkg"
|
|
|
|
def __init__(self, col: Collection) -> None:
|
|
AnkiExporter.__init__(self, col)
|
|
|
|
@staticmethod
|
|
def key(col: Collection) -> str:
|
|
return col.tr.exporting_anki_deck_package()
|
|
|
|
def exportInto(self, path: str) -> None:
|
|
# open a zip file
|
|
z = zipfile.ZipFile(
|
|
path, "w", zipfile.ZIP_DEFLATED, allowZip64=True, strict_timestamps=False
|
|
)
|
|
media = self.doExport(z, path)
|
|
# media map
|
|
z.writestr("media", json.dumps(media))
|
|
z.close()
|
|
|
|
def doExport(self, z: ZipFile, path: str) -> dict[str, str]: # type: ignore
|
|
# export into the anki2 file
|
|
colfile = path.replace(".apkg", ".anki2")
|
|
AnkiExporter.exportInto(self, colfile)
|
|
if not self._v2sched:
|
|
z.write(colfile, "collection.anki2")
|
|
else:
|
|
# prevent older clients from accessing
|
|
# pylint: disable=unreachable
|
|
self._addDummyCollection(z)
|
|
z.write(colfile, "collection.anki21")
|
|
|
|
# and media
|
|
self.prepareMedia()
|
|
media = self._exportMedia(z, self.mediaFiles, self.mediaDir)
|
|
# tidy up intermediate files
|
|
os.unlink(colfile)
|
|
p = path.replace(".apkg", ".media.db2")
|
|
if os.path.exists(p):
|
|
os.unlink(p)
|
|
os.chdir(self.mediaDir)
|
|
shutil.rmtree(path.replace(".apkg", ".media"))
|
|
return media
|
|
|
|
def _exportMedia(self, z: ZipFile, files: list[str], fdir: str) -> dict[str, str]:
|
|
media = {}
|
|
for c, file in enumerate(files):
|
|
cStr = str(c)
|
|
file = hooks.media_file_filter(file)
|
|
mpath = os.path.join(fdir, file)
|
|
if os.path.isdir(mpath):
|
|
continue
|
|
if os.path.exists(mpath):
|
|
if re.search(r"\.svg$", file, re.IGNORECASE):
|
|
z.write(mpath, cStr, zipfile.ZIP_DEFLATED)
|
|
else:
|
|
z.write(mpath, cStr, zipfile.ZIP_STORED)
|
|
media[cStr] = unicodedata.normalize("NFC", file)
|
|
hooks.media_files_did_export(c)
|
|
|
|
return media
|
|
|
|
def prepareMedia(self) -> None:
|
|
# chance to move each file in self.mediaFiles into place before media
|
|
# is zipped up
|
|
pass
|
|
|
|
# create a dummy collection to ensure older clients don't try to read
|
|
# data they don't understand
|
|
def _addDummyCollection(self, zip) -> None:
|
|
path = namedtmp("dummy.anki2")
|
|
c = Collection(path)
|
|
n = c.newNote()
|
|
n.fields[0] = "This file requires a newer version of Anki."
|
|
c.addNote(n)
|
|
c.save()
|
|
c.close(downgrade=True)
|
|
|
|
zip.write(path, "collection.anki2")
|
|
os.unlink(path)
|
|
|
|
|
|
# Collection package
|
|
######################################################################
|
|
|
|
|
|
class AnkiCollectionPackageExporter(AnkiPackageExporter):
|
|
|
|
ext = ".colpkg"
|
|
verbatim = True
|
|
includeSched = None
|
|
LEGACY = True
|
|
|
|
def __init__(self, col):
|
|
AnkiPackageExporter.__init__(self, col)
|
|
|
|
@staticmethod
|
|
def key(col: Collection) -> str:
|
|
return col.tr.exporting_anki_collection_package()
|
|
|
|
def exportInto(self, path: str) -> None:
|
|
"""Export collection. Caller must re-open afterwards."""
|
|
|
|
def exporting_media() -> bool:
|
|
return any(
|
|
hook.__name__ == "exported_media"
|
|
for hook in hooks.media_files_did_export._hooks
|
|
)
|
|
|
|
def progress() -> None:
|
|
while exporting_media():
|
|
progress = self.col._backend.latest_progress()
|
|
if progress.HasField("exporting"):
|
|
hooks.media_files_did_export(progress.exporting)
|
|
time.sleep(0.1)
|
|
|
|
threading.Thread(target=progress).start()
|
|
self.col.export_collection(path, self.includeMedia, self.LEGACY)
|
|
|
|
|
|
class AnkiCollectionPackage21bExporter(AnkiCollectionPackageExporter):
|
|
LEGACY = False
|
|
|
|
@staticmethod
|
|
def key(_col: Collection) -> str:
|
|
return "Anki 2.1.50+ Collection Package"
|
|
|
|
|
|
# Export modules
|
|
##########################################################################
|
|
|
|
|
|
def exporters(col: Collection) -> list[tuple[str, Any]]:
|
|
def id(obj):
|
|
if callable(obj.key):
|
|
key_str = obj.key(col)
|
|
else:
|
|
key_str = obj.key
|
|
return (f"{key_str} (*{obj.ext})", obj)
|
|
|
|
exps = [
|
|
id(AnkiCollectionPackageExporter),
|
|
id(AnkiCollectionPackage21bExporter),
|
|
id(AnkiPackageExporter),
|
|
id(TextNoteExporter),
|
|
id(TextCardExporter),
|
|
]
|
|
hooks.exporters_list_created(exps)
|
|
return exps
|