#!/usr/bin/env python3 # -*- coding: utf-8 -*- # # Script to generate Minetest translation template files and update # translation files. # # Copyright (C) 2019 Joachim Stolberg, 2020 FaceDeer, 2020 Louis Royer, # 2023 Wuzzy. # License: LGPLv2.1 or later (see LICENSE file for details) import os, fnmatch, re, shutil, errno from sys import argv as _argv from sys import stderr as _stderr # Running params params = {"recursive": False, "help": False, "verbose": False, "folders": [], "old-file": False, "break-long-lines": False, "print-source": False, "truncate-unused": False, } # Available CLI options options = {"recursive": ['--recursive', '-r'], "help": ['--help', '-h'], "verbose": ['--verbose', '-v'], "old-file": ['--old-file', '-o'], "break-long-lines": ['--break-long-lines', '-b'], "print-source": ['--print-source', '-p'], "truncate-unused": ['--truncate-unused', '-t'], } # Strings longer than this will have extra space added between # them in the translation files to make it easier to distinguish their # beginnings and endings at a glance doublespace_threshold = 80 # These symbols mark comment lines showing the source file name. # A comment may look like "##[ init.lua ]##". symbol_source_prefix = "##[" symbol_source_suffix = "]##" # comment to mark the section of old/unused strings comment_unused = "##### not used anymore #####" def set_params_folders(tab: list): '''Initialize params["folders"] from CLI arguments.''' # Discarding argument 0 (tool name) for param in tab[1:]: stop_param = False for option in options: if param in options[option]: stop_param = True break if not stop_param: params["folders"].append(os.path.abspath(param)) def set_params(tab: list): '''Initialize params from CLI arguments.''' for option in options: for option_name in options[option]: if option_name in tab: params[option] = True break def print_help(name): '''Prints some help message.''' print(f'''SYNOPSIS {name} [OPTIONS] [PATHS...] DESCRIPTION {', '.join(options["help"])} prints this help message {', '.join(options["recursive"])} run on all subfolders of paths given {', '.join(options["old-file"])} create *.old files {', '.join(options["break-long-lines"])} add extra line breaks before and after long strings {', '.join(options["print-source"])} add comments denoting the source file {', '.join(options["verbose"])} add output information {', '.join(options["truncate-unused"])} delete unused strings from files ''') def main(): '''Main function''' set_params(_argv) set_params_folders(_argv) if params["help"]: print_help(_argv[0]) else: # Add recursivity message print("Running ", end='') if params["recursive"]: print("recursively ", end='') # Running if len(params["folders"]) >= 2: print("on folder list:", params["folders"]) for f in params["folders"]: if params["recursive"]: run_all_subfolders(f) else: update_folder(f) elif len(params["folders"]) == 1: print("on folder", params["folders"][0]) if params["recursive"]: run_all_subfolders(params["folders"][0]) else: update_folder(params["folders"][0]) else: print("on folder", os.path.abspath("./")) if params["recursive"]: run_all_subfolders(os.path.abspath("./")) else: update_folder(os.path.abspath("./")) # Group 2 will be the string, groups 1 and 3 will be the delimiters (" or ') # See https://stackoverflow.com/questions/46967465/regex-match-text-in-either-single-or-double-quote pattern_lua_quoted = re.compile( r'(?:^|[\.=,{\(\s])' # Look for beginning of file or anything that isn't a function identifier r'N?F?S\s*\(\s*' # Matches S, FS, NS or NFS function call r'(["\'])((?:\\\1|(?:(?!\1)).)*)(\1)' # Quoted string r'[\s,\)]', # End of call or argument re.DOTALL) # Handles the [[ ... ]] string delimiters pattern_lua_bracketed = re.compile( r'(?:^|[\.=,{\(\s])' # Same as for pattern_lua_quoted r'N?F?S\s*\(\s*' # Same as for pattern_lua_quoted r'\[\[(.*?)\]\]' # [[ ... ]] string delimiters r'[\s,\)]', # Same as for pattern_lua_quoted re.DOTALL) # Handles "concatenation" .. " of strings" pattern_concat = re.compile(r'["\'][\s]*\.\.[\s]*["\']', re.DOTALL) # Handles a translation line in *.tr file. # Group 1 is the source string left of the equals sign. # Group 2 is the translated string, right of the equals sign. pattern_tr = re.compile( r'(.*)' # Source string # the separating equals sign, if NOT preceded by @, unless # that @ is preceded by another @ r'(?:(?2.5 if exc.errno == errno.EEXIST and os.path.isdir(path): pass else: raise # Converts the template dictionary to a text to be written as a file # dKeyStrings is a dictionary of localized string to source file sets # dOld is a dictionary of existing translations and comments from # the previous version of this text def strings_to_text(dkeyStrings, dOld, mod_name, header_comments, textdomain): # if textdomain is specified, insert it at the top if textdomain != None: lOut = [textdomain] # argument is full textdomain line # otherwise, use mod name as textdomain automatically else: lOut = [f"# textdomain: {mod_name}"] if header_comments is not None: lOut.append(header_comments) dGroupedBySource = {} for key in dkeyStrings: sourceList = list(dkeyStrings[key]) sourceString = "\n".join(sourceList) listForSource = dGroupedBySource.get(sourceString, []) listForSource.append(key) dGroupedBySource[sourceString] = listForSource lSourceKeys = list(dGroupedBySource.keys()) lSourceKeys.sort() for source in lSourceKeys: localizedStrings = dGroupedBySource[source] if params["print-source"]: if lOut[-1] != "": lOut.append("") lOut.append(source) for localizedString in localizedStrings: val = dOld.get(localizedString, {}) translation = val.get("translation", "") comment = val.get("comment") if params["break-long-lines"] and len(localizedString) > doublespace_threshold and not lOut[-1] == "": lOut.append("") if comment != None and comment != "" and not comment.startswith("# textdomain:"): lOut.append(comment) lOut.append(f"{localizedString}={translation}") if params["break-long-lines"] and len(localizedString) > doublespace_threshold: lOut.append("") unusedExist = False if not params["truncate-unused"]: for key in dOld: if key not in dkeyStrings: val = dOld[key] translation = val.get("translation") comment = val.get("comment") # only keep an unused translation if there was translated # text or a comment associated with it if translation != None and (translation != "" or comment): if not unusedExist: unusedExist = True lOut.append("\n\n" + comment_unused + "\n") if params["break-long-lines"] and len(key) > doublespace_threshold and not lOut[-1] == "": lOut.append("") if comment != None: lOut.append(comment) lOut.append(f"{key}={translation}") if params["break-long-lines"] and len(key) > doublespace_threshold: lOut.append("") return "\n".join(lOut) + '\n' # Writes a template.txt file # dkeyStrings is the dictionary returned by generate_template def write_template(templ_file, dkeyStrings, mod_name): # read existing template file to preserve comments existing_template = import_tr_file(templ_file) text = strings_to_text(dkeyStrings, existing_template[0], mod_name, existing_template[2], existing_template[3]) mkdir_p(os.path.dirname(templ_file)) with open(templ_file, "wt", encoding='utf-8') as template_file: template_file.write(text) # Gets all translatable strings from a lua file def read_lua_file_strings(lua_file): lOut = [] with open(lua_file, encoding='utf-8') as text_file: text = text_file.read() text = re.sub(pattern_concat, "", text) strings = [] for s in pattern_lua_quoted.findall(text): strings.append(s[1]) for s in pattern_lua_bracketed.findall(text): strings.append(s) for s in strings: found_bad = pattern_bad_luastring.search(s) if found_bad: print("SYNTAX ERROR: Unescaped '@' in Lua string: " + s) continue s = s.replace('\\"', '"') s = s.replace("\\'", "'") s = s.replace("\n", "@n") s = s.replace("\\n", "@n") s = s.replace("=", "@=") lOut.append(s) return lOut # Gets strings from an existing translation file # returns both a dictionary of translations # and the full original source text so that the new text # can be compared to it for changes. # Returns also header comments in the third return value. def import_tr_file(tr_file): dOut = {} text = None in_header = True header_comments = None textdomain = None if os.path.exists(tr_file): with open(tr_file, "r", encoding='utf-8') as existing_file : # save the full text to allow for comparison # of the old version with the new output text = existing_file.read() existing_file.seek(0) # a running record of the current comment block # we're inside, to allow preceeding multi-line comments # to be retained for a translation line latest_comment_block = None for line in existing_file.readlines(): line = line.rstrip('\n') # "##### not used anymore #####" comment if line == comment_unused: # Always delete the 'not used anymore' comment. # It will be re-added to the file if neccessary. latest_comment_block = None if header_comments != None: in_header = False continue # Comment lines elif line.startswith("#"): # Source file comments: ##[ file.lua ]## if line.startswith(symbol_source_prefix) and line.endswith(symbol_source_suffix): # This line marks the end of header comments. if params["print-source"]: in_header = False # Remove those comments; they may be added back automatically. continue # Store first occurance of textdomain # discard all subsequent textdomain lines if line.startswith("# textdomain:"): if textdomain == None: textdomain = line continue elif in_header: # Save header comments (normal comments at top of file) if not header_comments: header_comments = line else: header_comments = header_comments + "\n" + line else: # Save normal comments if line.startswith("# textdomain:") and textdomain == None: textdomain = line elif not latest_comment_block: latest_comment_block = line else: latest_comment_block = latest_comment_block + "\n" + line continue match = pattern_tr.match(line) if match: # this line is a translated line outval = {} outval["translation"] = match.group(2) if latest_comment_block: # if there was a comment, record that. outval["comment"] = latest_comment_block latest_comment_block = None in_header = False dOut[match.group(1)] = outval return (dOut, text, header_comments, textdomain) # like os.walk but returns sorted filenames def sorted_os_walk(folder): tuples = [] t = 0 for root, dirs, files in os.walk(folder): tuples.append( (root, dirs, files) ) t = t + 1 tuples = sorted(tuples) paths_and_files = [] f = 0 for tu in tuples: root = tu[0] dirs = tu[1] files = tu[2] files = sorted(files, key=str.lower) for filename in files: paths_and_files.append( (os.path.join(root, filename), filename) ) f = f + 1 return paths_and_files # Walks all lua files in the mod folder, collects translatable strings, # and writes it to a template.txt file # Returns a dictionary of localized strings to source file lists # that can be used with the strings_to_text function. def generate_template(folder, mod_name): dOut = {} paths_and_files = sorted_os_walk(folder) for paf in paths_and_files: fullpath_filename = paf[0] filename = paf[1] if fnmatch.fnmatch(filename, "*.lua"): found = read_lua_file_strings(fullpath_filename) if params["verbose"]: print(f"{fullpath_filename}: {str(len(found))} translatable strings") for s in found: sources = dOut.get(s, set()) sources.add(os.path.relpath(fullpath_filename, start=folder)) dOut[s] = sources if len(dOut) == 0: return None # Convert source file set to list, sort it and add comment symbols. # Needed because a set is unsorted and might result in unpredictable. # output orders if any source string appears in multiple files. for d in dOut: sources = dOut.get(d, set()) sources = sorted(list(sources), key=str.lower) newSources = [] for i in sources: i = "/".join(os.path.split(i)).lstrip("/") newSources.append(f"{symbol_source_prefix} {i} {symbol_source_suffix}") dOut[d] = newSources templ_file = os.path.join(folder, "locale/template.txt") write_template(templ_file, dOut, mod_name) return dOut # Updates an existing .tr file, copying the old one to a ".old" file # if any changes have happened # dNew is the data used to generate the template, it has all the # currently-existing localized strings def update_tr_file(dNew, mod_name, tr_file): if params["verbose"]: print(f"updating {tr_file}") tr_import = import_tr_file(tr_file) dOld = tr_import[0] textOld = tr_import[1] textNew = strings_to_text(dNew, dOld, mod_name, tr_import[2], tr_import[3]) if textOld and textOld != textNew: print(f"{tr_file} has changed.") if params["old-file"]: shutil.copyfile(tr_file, f"{tr_file}.old") with open(tr_file, "w", encoding='utf-8') as new_tr_file: new_tr_file.write(textNew) # Updates translation files for the mod in the given folder def update_mod(folder): if not os.path.exists(os.path.join(folder, "init.lua")): print(f"Mod folder {folder} is missing init.lua, aborting.") exit(1) assert not is_modpack(folder) modname = get_modname(folder) print(f"Updating translations for {modname}") data = generate_template(folder, modname) if data == None: print(f"No translatable strings found in {modname}") else: for tr_file in get_existing_tr_files(folder): update_tr_file(data, modname, os.path.join(folder, "locale/", tr_file)) def is_modpack(folder): return os.path.exists(os.path.join(folder, "modpack.txt")) or os.path.exists(os.path.join(folder, "modpack.conf")) def is_game(folder): return os.path.exists(os.path.join(folder, "game.conf")) and os.path.exists(os.path.join(folder, "mods")) # Determines if the folder being pointed to is a game, mod or a mod pack # and then runs update_mod accordingly def update_folder(folder): if is_game(folder): run_all_subfolders(os.path.join(folder, "mods")) elif is_modpack(folder): run_all_subfolders(folder) else: update_mod(folder) print("Done.") def run_all_subfolders(folder): for modfolder in [f.path for f in os.scandir(folder) if f.is_dir() and not f.name.startswith('.')]: update_folder(modfolder) main()