|
|
|
@ -14,17 +14,26 @@
|
|
|
|
|
# along with Korman. If not, see <http://www.gnu.org/licenses/>. |
|
|
|
|
|
|
|
|
|
import bpy |
|
|
|
|
from PyHSPlasma import * |
|
|
|
|
|
|
|
|
|
from collections import defaultdict |
|
|
|
|
from contextlib import contextmanager |
|
|
|
|
import itertools |
|
|
|
|
from pathlib import Path |
|
|
|
|
import re |
|
|
|
|
from xml.sax.saxutils import escape as xml_escape |
|
|
|
|
import weakref |
|
|
|
|
|
|
|
|
|
from .explosions import NonfatalExportError |
|
|
|
|
from .. import korlib |
|
|
|
|
from . import logger |
|
|
|
|
from pathlib import Path |
|
|
|
|
from PyHSPlasma import * |
|
|
|
|
import weakref |
|
|
|
|
from xml.sax.saxutils import escape as xml_escape |
|
|
|
|
|
|
|
|
|
_SP_LANGUAGES = {"English", "French", "German", "Italian", "Spanish"} |
|
|
|
|
|
|
|
|
|
# Detects if there are any Plasma esHTML tags in the translated data. If so, we store |
|
|
|
|
# as CDATA instead of XML encoding the entry. |
|
|
|
|
_ESHTML_REGEX = re.compile("<.+>") |
|
|
|
|
|
|
|
|
|
class LocalizationConverter: |
|
|
|
|
def __init__(self, exporter=None, **kwargs): |
|
|
|
|
if exporter is not None: |
|
|
|
@ -37,20 +46,16 @@ class LocalizationConverter:
|
|
|
|
|
self._age_name = kwargs.get("age_name") |
|
|
|
|
self._path = kwargs.get("path") |
|
|
|
|
self._version = kwargs.get("version") |
|
|
|
|
self._journals = {} |
|
|
|
|
self._strings = {} |
|
|
|
|
|
|
|
|
|
def add_journal(self, name, language, text_id, indent=0): |
|
|
|
|
if text_id.is_modified: |
|
|
|
|
self._report.warn("Journal '{}' translation for '{}' is modified on the disk but not reloaded in Blender.", |
|
|
|
|
name, language, indent=indent) |
|
|
|
|
journal = self._journals.setdefault(name, {}) |
|
|
|
|
journal[language] = text_id.as_string() |
|
|
|
|
self._strings = defaultdict(lambda: defaultdict(dict)) |
|
|
|
|
|
|
|
|
|
def add_string(self, set_name, element_name, language, value): |
|
|
|
|
trans_set = self._strings.setdefault(set_name, {}) |
|
|
|
|
trans_element = trans_set.setdefault(element_name, {}) |
|
|
|
|
trans_element[language] = value |
|
|
|
|
def add_string(self, set_name, element_name, language, value, indent=0): |
|
|
|
|
self._report.msg("Accepted '{}' translation for '{}'.", element_name, language, indent=indent) |
|
|
|
|
if isinstance(value, bpy.types.Text): |
|
|
|
|
if value.is_modified: |
|
|
|
|
self._report.warn("'{}' translation for '{}' is modified on the disk but not reloaded in Blender.", |
|
|
|
|
element_name, language, indent=indent) |
|
|
|
|
value = value.as_string() |
|
|
|
|
self._strings[set_name][element_name][language] = value |
|
|
|
|
|
|
|
|
|
@contextmanager |
|
|
|
|
def _generate_file(self, filename, **kwargs): |
|
|
|
@ -68,12 +73,12 @@ class LocalizationConverter:
|
|
|
|
|
finally: |
|
|
|
|
handle.close() |
|
|
|
|
|
|
|
|
|
def _generate_journal_texts(self): |
|
|
|
|
def _generate_text_files(self): |
|
|
|
|
age_name = self._age_name |
|
|
|
|
|
|
|
|
|
def write_journal_file(language, file_name, contents): |
|
|
|
|
try: |
|
|
|
|
def write_text_file(language, file_name, contents): |
|
|
|
|
with self._generate_file(dirname="ageresources", filename=file_name) as stream: |
|
|
|
|
try: |
|
|
|
|
stream.write(contents.encode("windows-1252")) |
|
|
|
|
except UnicodeEncodeError: |
|
|
|
|
self._report.warn("Translation '{}': Contents contains characters that cannot be used in this version of Plasma. They will appear as a '?' in game.", |
|
|
|
@ -84,8 +89,9 @@ class LocalizationConverter:
|
|
|
|
|
stream.write(contents.encode("windows-1252", "replace")) |
|
|
|
|
return True |
|
|
|
|
|
|
|
|
|
for journal_name, translations in self._journals.items(): |
|
|
|
|
self._report.msg("Copying Journal '{}'", journal_name, indent=1) |
|
|
|
|
locs = itertools.chain(self._strings["Journals"].items(), self._strings["DynaTexts"].items()) |
|
|
|
|
for journal_name, translations in locs: |
|
|
|
|
self._report.msg("Copying localization '{}'", journal_name, indent=1) |
|
|
|
|
for language_name, value in translations.items(): |
|
|
|
|
if language_name not in _SP_LANGUAGES: |
|
|
|
|
self._report.warn("Translation '{}' will not be used because it is not supported in this version of Plasma.", |
|
|
|
@ -93,7 +99,7 @@ class LocalizationConverter:
|
|
|
|
|
continue |
|
|
|
|
suffix = "_{}".format(language_name.lower()) if language_name != "English" else "" |
|
|
|
|
file_name = "{}--{}{}.txt".format(age_name, journal_name, suffix) |
|
|
|
|
write_journal_file(language_name, file_name, value) |
|
|
|
|
write_text_file(language_name, file_name, value) |
|
|
|
|
|
|
|
|
|
# Ensure that default (read: "English") journal is available |
|
|
|
|
if "English" not in translations: |
|
|
|
@ -102,55 +108,69 @@ class LocalizationConverter:
|
|
|
|
|
if language_name is not None: |
|
|
|
|
file_name = "{}--{}.txt".format(age_name, journal_name) |
|
|
|
|
# If you manage to screw up this badly... Well, I am very sorry. |
|
|
|
|
if write_journal_file(language_name, file_name, value): |
|
|
|
|
if write_text_file(language_name, file_name, value): |
|
|
|
|
self._report.warn("No 'English' translation available, so '{}' will be used as the default", |
|
|
|
|
language_name, indent=2) |
|
|
|
|
else: |
|
|
|
|
self._report.port("No 'English' nor any other suitable default translation available", indent=2) |
|
|
|
|
|
|
|
|
|
def _generate_loc_file(self): |
|
|
|
|
# Only generate this junk if needed |
|
|
|
|
if not self._strings and not self._journals: |
|
|
|
|
def _generate_loc_files(self): |
|
|
|
|
if not self._strings: |
|
|
|
|
return |
|
|
|
|
|
|
|
|
|
method = bpy.context.scene.world.plasma_age.localization_method |
|
|
|
|
if method == "single_file": |
|
|
|
|
self._generate_loc_file("{}.loc".format(self._age_name), self._strings) |
|
|
|
|
elif method in {"database", "database_back_compat"}: |
|
|
|
|
# Where the strings are set -> element -> language: str, we want language -> set -> element: str |
|
|
|
|
# This is so we can mimic pfLocalizationEditor's <agename>English.loc pathing. |
|
|
|
|
database = defaultdict(lambda: defaultdict(dict)) |
|
|
|
|
for set_name, elements in self._strings.items(): |
|
|
|
|
for element_name, translations in elements.items(): |
|
|
|
|
for language_name, value in translations.items(): |
|
|
|
|
database[language_name][set_name][element_name] = value |
|
|
|
|
|
|
|
|
|
for language_name, sets in database.items(): |
|
|
|
|
self._generate_loc_file("{}{}.loc".format(self._age_name, language_name), sets, language_name) |
|
|
|
|
|
|
|
|
|
# Generate an empty localization file to defeat any old ones from Korman 0.11 (and lower) |
|
|
|
|
if method == "database_back_compat": |
|
|
|
|
self._generate_loc_file("{}.loc".format(self._age_name), {}) |
|
|
|
|
else: |
|
|
|
|
raise RuntimeError("Unexpected localization method {}".format(method)) |
|
|
|
|
|
|
|
|
|
def _generate_loc_file(self, filename, sets, language_name=None): |
|
|
|
|
def write_line(value, *args, **kwargs): |
|
|
|
|
# tabs suck, then you die... |
|
|
|
|
whitespace = " " * kwargs.pop("indent", 0) |
|
|
|
|
if args or kwargs: |
|
|
|
|
value = value.format(*args, **kwargs) |
|
|
|
|
line = "".join((whitespace, value, "\n")) |
|
|
|
|
stream.write(line.encode("utf-16_le")) |
|
|
|
|
stream.write(line.encode("utf-8")) |
|
|
|
|
|
|
|
|
|
age_name = self._age_name |
|
|
|
|
enc = plEncryptedStream.kEncAes if self._version == pvEoa else None |
|
|
|
|
file_name = "{}.loc".format(age_name) |
|
|
|
|
with self._generate_file(file_name, enc=enc) as stream: |
|
|
|
|
# UTF-16 little endian byte order mark |
|
|
|
|
stream.write(b"\xFF\xFE") |
|
|
|
|
def iter_element(element): |
|
|
|
|
if language_name is None: |
|
|
|
|
yield from element.items() |
|
|
|
|
else: |
|
|
|
|
yield language_name, element |
|
|
|
|
|
|
|
|
|
write_line("<?xml version=\"1.0\" encoding=\"utf-16\"?>") |
|
|
|
|
enc = plEncryptedStream.kEncAes if self._version == pvEoa else None |
|
|
|
|
with self._generate_file(filename, enc=enc) as stream: |
|
|
|
|
write_line("<?xml version=\"1.0\" encoding=\"utf-8\"?>") |
|
|
|
|
write_line("<localizations>") |
|
|
|
|
write_line("<age name=\"{}\">", age_name, indent=1) |
|
|
|
|
write_line("<age name=\"{}\">", self._age_name, indent=1) |
|
|
|
|
|
|
|
|
|
# Arbitrary strings defined by something like a GUI or a node tree |
|
|
|
|
for set_name, elements in self._strings.items(): |
|
|
|
|
for set_name, elements in sets.items(): |
|
|
|
|
write_line("<set name=\"{}\">", set_name, indent=2) |
|
|
|
|
for element_name, translations in elements.items(): |
|
|
|
|
for element_name, value in elements.items(): |
|
|
|
|
write_line("<element name=\"{}\">", element_name, indent=3) |
|
|
|
|
for language_name, value in translations.items(): |
|
|
|
|
write_line("<translation language=\"{language}\">{translation}</translation>", |
|
|
|
|
language=language_name, translation=xml_escape(value), indent=4) |
|
|
|
|
write_line("</element>", indent=3) |
|
|
|
|
write_line("</set>", indent=2) |
|
|
|
|
|
|
|
|
|
# Journals |
|
|
|
|
if self._journals: |
|
|
|
|
write_line("<set name=\"Journals\">", indent=2) |
|
|
|
|
for journal_name, translations in self._journals.items(): |
|
|
|
|
write_line("<element name=\"{}\">", journal_name, indent=3) |
|
|
|
|
for language_name, value in translations.items(): |
|
|
|
|
for translation_language, translation_value in iter_element(value): |
|
|
|
|
if _ESHTML_REGEX.search(translation_value): |
|
|
|
|
encoded_value = "<![CDATA[{}]]>".format(translation_value) |
|
|
|
|
else: |
|
|
|
|
encoded_value = xml_escape(translation_value) |
|
|
|
|
write_line("<translation language=\"{language}\">{translation}</translation>", |
|
|
|
|
language=language_name, translation=xml_escape(value), indent=4) |
|
|
|
|
language=translation_language, translation=encoded_value, indent=4) |
|
|
|
|
write_line("</element>", indent=3) |
|
|
|
|
write_line("</set>", indent=2) |
|
|
|
|
|
|
|
|
@ -164,7 +184,7 @@ class LocalizationConverter:
|
|
|
|
|
loc_path = str(Path(self._path) / "dat" / "{}.loc".format(self._age_name)) |
|
|
|
|
log = logger.ExportVerboseLogger if age_props.verbose else logger.ExportProgressLogger |
|
|
|
|
with korlib.ConsoleToggler(age_props.show_console), log(loc_path) as self._report: |
|
|
|
|
self._report.progress_add_step("Harvesting Journals") |
|
|
|
|
self._report.progress_add_step("Harvesting Translations") |
|
|
|
|
self._report.progress_add_step("Generating Localization") |
|
|
|
|
self._report.progress_start("Exporting Localization Data") |
|
|
|
|
|
|
|
|
@ -176,20 +196,23 @@ class LocalizationConverter:
|
|
|
|
|
self._report.raise_errors() |
|
|
|
|
|
|
|
|
|
def _run_harvest_journals(self): |
|
|
|
|
from ..properties.modifiers import TranslationMixin |
|
|
|
|
|
|
|
|
|
objects = bpy.context.scene.objects |
|
|
|
|
self._report.progress_advance() |
|
|
|
|
self._report.progress_range = len(objects) |
|
|
|
|
inc_progress = self._report.progress_increment |
|
|
|
|
|
|
|
|
|
for i in objects: |
|
|
|
|
journal = i.plasma_modifiers.journalbookmod |
|
|
|
|
if journal.enabled: |
|
|
|
|
translations = [j for j in journal.journal_translations if j.text_id is not None] |
|
|
|
|
for mod_type in filter(None, (getattr(j, "pl_id", None) for j in TranslationMixin.__subclasses__())): |
|
|
|
|
modifier = getattr(i.plasma_modifiers, mod_type) |
|
|
|
|
if modifier.enabled: |
|
|
|
|
translations = [j for j in modifier.translations if j.text_id is not None] |
|
|
|
|
if not translations: |
|
|
|
|
self._report.error("Journal '{}': No content translations available. The journal will not be exported.", |
|
|
|
|
self._report.error("'{}': No content translations available. The localization will not be exported.", |
|
|
|
|
i.name, indent=2) |
|
|
|
|
for j in translations: |
|
|
|
|
self.add_journal(journal.key_name, j.language, j.text_id, indent=1) |
|
|
|
|
self.add_string(modifier.localization_set, modifier.key_name, j.language, j.text_id, indent=1) |
|
|
|
|
inc_progress() |
|
|
|
|
|
|
|
|
|
def _run_generate(self): |
|
|
|
@ -198,6 +221,6 @@ class LocalizationConverter:
|
|
|
|
|
|
|
|
|
|
def save(self): |
|
|
|
|
if self._version > pvPots: |
|
|
|
|
self._generate_loc_file() |
|
|
|
|
self._generate_loc_files() |
|
|
|
|
else: |
|
|
|
|
self._generate_journal_texts() |
|
|
|
|
self._generate_text_files() |
|
|
|
|