Translating LaTeX documents into different languages using DeepL API
up vote
7
down vote
favorite
I'm coming from C and C++ and have often troubles using Python's full potential. This is a script I wrote to translate LaTeX documents into different languages. A parser replaces all latex specific syntax with hashes and then sends it all to deepl. The pydeepl wrapper is from Github.
What do you think of this code?
import sys
import re
import pydeepl
from random import choice
from tqdm import tqdm
def make_xlat(*args, **kwds):
adict = dict(*args, **kwds)
rx = re.compile('|'.join(map(re.escape, adict)))
def one_xlat(match):
return adict[match.group(0)]
def xlat(text):
return rx.sub(one_xlat, text)
return xlat
if __name__ == "__main__":
fileInputName = sys.argv[1]
#fileInputName = "Introduction.tex"
fileOutName = fileInputName.split('.')[0]+"_trans.tex"
fileIn = open(fileInputName, "r")
fileOut = open(fileOutName, "w")
fileStr = fileIn.read()
print("Starting hashing...")
#replace commands like begin{*}, end{*}, tabs etc. with hashes
searchObj1 = re.findall( r"\begin{w+}|t| |r|\end{w+}|\usepackage{w+}|\newcommand{w+}|\include{.*}|\input{w+}|\w+[.*}|%.*", fileStr)
#random number for every found command + a prefix which hopefully doens't appear in text. Used to skip lines later, which don't need translation
list1 = ['X#X' + str(choice(range(1111, 9999, 1))) for x in searchObj1]
#make a dictionary out of hashes
d1 = dict(zip(searchObj1,list1))
translate = make_xlat(d1)
hashedText = translate(fileStr)
#replace all latex commands (starting with a backslash) with hashes
searchObj2 = re.findall( r"\w+",hashedText)
#random number + prefix again
list2 = ['X#X' + str(choice(range(1111, 9999, 1))) for x in searchObj2]
#make a dictionary
d2 = dict(zip(searchObj2,list2))
translate = make_xlat(d2)
hashedText = translate(hashedText)
#print(hashedText)
#fileOut.write(translate(hashedText))
d1.update(d2) # combine dictionaries
#with open('hash_dict.json', 'w') as f:
#json.dump(d1, f)
print("Hashing done. Starting translation...")
translated = ''
for line in tqdm(hashedText.splitlines()): #tqdm is a progressbar
#print(line)
if line.startswith("X#X") and len(line) == 7:
translated = translated + line + 'n'
continue
if line == 'n':
translated = translated + 'n'
elif line == '':
translated = translated + 'n'
else:
translated = translated+pydeepl.translate(line, "DE", "EN")+'n'
#translated = translated+pydeepl.translate(hashedText, "DE", "EN")
#print(translated)
d1Inv = {val:key for (key, val) in d1.items()} #swap dictionary
translate2 = make_xlat(d1Inv)
fileStrOut = translate2(translated)
#print(fileStrOut)
fileOut.write(fileStrOut)
print("success")
fileIn.close()
fileOut.close()
EDIT 1: One flaw I already know is that I'm sending sentences by sentences which makes it really slow. However, sending all lines at once produces errors which I couldn't get rid of. Deepl scrambles stuff around and dehashign doesn't work anymore.
python api tex
New contributor
add a comment |
up vote
7
down vote
favorite
I'm coming from C and C++ and have often troubles using Python's full potential. This is a script I wrote to translate LaTeX documents into different languages. A parser replaces all latex specific syntax with hashes and then sends it all to deepl. The pydeepl wrapper is from Github.
What do you think of this code?
import sys
import re
import pydeepl
from random import choice
from tqdm import tqdm
def make_xlat(*args, **kwds):
adict = dict(*args, **kwds)
rx = re.compile('|'.join(map(re.escape, adict)))
def one_xlat(match):
return adict[match.group(0)]
def xlat(text):
return rx.sub(one_xlat, text)
return xlat
if __name__ == "__main__":
fileInputName = sys.argv[1]
#fileInputName = "Introduction.tex"
fileOutName = fileInputName.split('.')[0]+"_trans.tex"
fileIn = open(fileInputName, "r")
fileOut = open(fileOutName, "w")
fileStr = fileIn.read()
print("Starting hashing...")
#replace commands like begin{*}, end{*}, tabs etc. with hashes
searchObj1 = re.findall( r"\begin{w+}|t| |r|\end{w+}|\usepackage{w+}|\newcommand{w+}|\include{.*}|\input{w+}|\w+[.*}|%.*", fileStr)
#random number for every found command + a prefix which hopefully doens't appear in text. Used to skip lines later, which don't need translation
list1 = ['X#X' + str(choice(range(1111, 9999, 1))) for x in searchObj1]
#make a dictionary out of hashes
d1 = dict(zip(searchObj1,list1))
translate = make_xlat(d1)
hashedText = translate(fileStr)
#replace all latex commands (starting with a backslash) with hashes
searchObj2 = re.findall( r"\w+",hashedText)
#random number + prefix again
list2 = ['X#X' + str(choice(range(1111, 9999, 1))) for x in searchObj2]
#make a dictionary
d2 = dict(zip(searchObj2,list2))
translate = make_xlat(d2)
hashedText = translate(hashedText)
#print(hashedText)
#fileOut.write(translate(hashedText))
d1.update(d2) # combine dictionaries
#with open('hash_dict.json', 'w') as f:
#json.dump(d1, f)
print("Hashing done. Starting translation...")
translated = ''
for line in tqdm(hashedText.splitlines()): #tqdm is a progressbar
#print(line)
if line.startswith("X#X") and len(line) == 7:
translated = translated + line + 'n'
continue
if line == 'n':
translated = translated + 'n'
elif line == '':
translated = translated + 'n'
else:
translated = translated+pydeepl.translate(line, "DE", "EN")+'n'
#translated = translated+pydeepl.translate(hashedText, "DE", "EN")
#print(translated)
d1Inv = {val:key for (key, val) in d1.items()} #swap dictionary
translate2 = make_xlat(d1Inv)
fileStrOut = translate2(translated)
#print(fileStrOut)
fileOut.write(fileStrOut)
print("success")
fileIn.close()
fileOut.close()
EDIT 1: One flaw I already know is that I'm sending sentences by sentences which makes it really slow. However, sending all lines at once produces errors which I couldn't get rid of. Deepl scrambles stuff around and dehashign doesn't work anymore.
python api tex
New contributor
1
Instead of sending the whole text in 1 go or line per line, you could try to find a middleground and send chunks of a few lines at a time
– Maarten Fabré
Dec 5 at 13:47
Thought about that too. The problem is I cant guarantee it will work correctly. It's a tradeoff.
– Mr.Sh4nnon
Dec 5 at 13:52
add a comment |
up vote
7
down vote
favorite
up vote
7
down vote
favorite
I'm coming from C and C++ and have often troubles using Python's full potential. This is a script I wrote to translate LaTeX documents into different languages. A parser replaces all latex specific syntax with hashes and then sends it all to deepl. The pydeepl wrapper is from Github.
What do you think of this code?
import sys
import re
import pydeepl
from random import choice
from tqdm import tqdm
def make_xlat(*args, **kwds):
adict = dict(*args, **kwds)
rx = re.compile('|'.join(map(re.escape, adict)))
def one_xlat(match):
return adict[match.group(0)]
def xlat(text):
return rx.sub(one_xlat, text)
return xlat
if __name__ == "__main__":
fileInputName = sys.argv[1]
#fileInputName = "Introduction.tex"
fileOutName = fileInputName.split('.')[0]+"_trans.tex"
fileIn = open(fileInputName, "r")
fileOut = open(fileOutName, "w")
fileStr = fileIn.read()
print("Starting hashing...")
#replace commands like begin{*}, end{*}, tabs etc. with hashes
searchObj1 = re.findall( r"\begin{w+}|t| |r|\end{w+}|\usepackage{w+}|\newcommand{w+}|\include{.*}|\input{w+}|\w+[.*}|%.*", fileStr)
#random number for every found command + a prefix which hopefully doens't appear in text. Used to skip lines later, which don't need translation
list1 = ['X#X' + str(choice(range(1111, 9999, 1))) for x in searchObj1]
#make a dictionary out of hashes
d1 = dict(zip(searchObj1,list1))
translate = make_xlat(d1)
hashedText = translate(fileStr)
#replace all latex commands (starting with a backslash) with hashes
searchObj2 = re.findall( r"\w+",hashedText)
#random number + prefix again
list2 = ['X#X' + str(choice(range(1111, 9999, 1))) for x in searchObj2]
#make a dictionary
d2 = dict(zip(searchObj2,list2))
translate = make_xlat(d2)
hashedText = translate(hashedText)
#print(hashedText)
#fileOut.write(translate(hashedText))
d1.update(d2) # combine dictionaries
#with open('hash_dict.json', 'w') as f:
#json.dump(d1, f)
print("Hashing done. Starting translation...")
translated = ''
for line in tqdm(hashedText.splitlines()): #tqdm is a progressbar
#print(line)
if line.startswith("X#X") and len(line) == 7:
translated = translated + line + 'n'
continue
if line == 'n':
translated = translated + 'n'
elif line == '':
translated = translated + 'n'
else:
translated = translated+pydeepl.translate(line, "DE", "EN")+'n'
#translated = translated+pydeepl.translate(hashedText, "DE", "EN")
#print(translated)
d1Inv = {val:key for (key, val) in d1.items()} #swap dictionary
translate2 = make_xlat(d1Inv)
fileStrOut = translate2(translated)
#print(fileStrOut)
fileOut.write(fileStrOut)
print("success")
fileIn.close()
fileOut.close()
EDIT 1: One flaw I already know is that I'm sending sentences by sentences which makes it really slow. However, sending all lines at once produces errors which I couldn't get rid of. Deepl scrambles stuff around and dehashign doesn't work anymore.
python api tex
New contributor
I'm coming from C and C++ and have often troubles using Python's full potential. This is a script I wrote to translate LaTeX documents into different languages. A parser replaces all latex specific syntax with hashes and then sends it all to deepl. The pydeepl wrapper is from Github.
What do you think of this code?
import sys
import re
import pydeepl
from random import choice
from tqdm import tqdm
def make_xlat(*args, **kwds):
adict = dict(*args, **kwds)
rx = re.compile('|'.join(map(re.escape, adict)))
def one_xlat(match):
return adict[match.group(0)]
def xlat(text):
return rx.sub(one_xlat, text)
return xlat
if __name__ == "__main__":
fileInputName = sys.argv[1]
#fileInputName = "Introduction.tex"
fileOutName = fileInputName.split('.')[0]+"_trans.tex"
fileIn = open(fileInputName, "r")
fileOut = open(fileOutName, "w")
fileStr = fileIn.read()
print("Starting hashing...")
#replace commands like begin{*}, end{*}, tabs etc. with hashes
searchObj1 = re.findall( r"\begin{w+}|t| |r|\end{w+}|\usepackage{w+}|\newcommand{w+}|\include{.*}|\input{w+}|\w+[.*}|%.*", fileStr)
#random number for every found command + a prefix which hopefully doens't appear in text. Used to skip lines later, which don't need translation
list1 = ['X#X' + str(choice(range(1111, 9999, 1))) for x in searchObj1]
#make a dictionary out of hashes
d1 = dict(zip(searchObj1,list1))
translate = make_xlat(d1)
hashedText = translate(fileStr)
#replace all latex commands (starting with a backslash) with hashes
searchObj2 = re.findall( r"\w+",hashedText)
#random number + prefix again
list2 = ['X#X' + str(choice(range(1111, 9999, 1))) for x in searchObj2]
#make a dictionary
d2 = dict(zip(searchObj2,list2))
translate = make_xlat(d2)
hashedText = translate(hashedText)
#print(hashedText)
#fileOut.write(translate(hashedText))
d1.update(d2) # combine dictionaries
#with open('hash_dict.json', 'w') as f:
#json.dump(d1, f)
print("Hashing done. Starting translation...")
translated = ''
for line in tqdm(hashedText.splitlines()): #tqdm is a progressbar
#print(line)
if line.startswith("X#X") and len(line) == 7:
translated = translated + line + 'n'
continue
if line == 'n':
translated = translated + 'n'
elif line == '':
translated = translated + 'n'
else:
translated = translated+pydeepl.translate(line, "DE", "EN")+'n'
#translated = translated+pydeepl.translate(hashedText, "DE", "EN")
#print(translated)
d1Inv = {val:key for (key, val) in d1.items()} #swap dictionary
translate2 = make_xlat(d1Inv)
fileStrOut = translate2(translated)
#print(fileStrOut)
fileOut.write(fileStrOut)
print("success")
fileIn.close()
fileOut.close()
EDIT 1: One flaw I already know is that I'm sending sentences by sentences which makes it really slow. However, sending all lines at once produces errors which I couldn't get rid of. Deepl scrambles stuff around and dehashign doesn't work anymore.
python api tex
python api tex
New contributor
New contributor
edited Dec 5 at 10:15
New contributor
asked Dec 5 at 6:46
Mr.Sh4nnon
385
385
New contributor
New contributor
1
Instead of sending the whole text in 1 go or line per line, you could try to find a middleground and send chunks of a few lines at a time
– Maarten Fabré
Dec 5 at 13:47
Thought about that too. The problem is I cant guarantee it will work correctly. It's a tradeoff.
– Mr.Sh4nnon
Dec 5 at 13:52
add a comment |
1
Instead of sending the whole text in 1 go or line per line, you could try to find a middleground and send chunks of a few lines at a time
– Maarten Fabré
Dec 5 at 13:47
Thought about that too. The problem is I cant guarantee it will work correctly. It's a tradeoff.
– Mr.Sh4nnon
Dec 5 at 13:52
1
1
Instead of sending the whole text in 1 go or line per line, you could try to find a middleground and send chunks of a few lines at a time
– Maarten Fabré
Dec 5 at 13:47
Instead of sending the whole text in 1 go or line per line, you could try to find a middleground and send chunks of a few lines at a time
– Maarten Fabré
Dec 5 at 13:47
Thought about that too. The problem is I cant guarantee it will work correctly. It's a tradeoff.
– Mr.Sh4nnon
Dec 5 at 13:52
Thought about that too. The problem is I cant guarantee it will work correctly. It's a tradeoff.
– Mr.Sh4nnon
Dec 5 at 13:52
add a comment |
2 Answers
2
active
oldest
votes
up vote
5
down vote
accepted
While using
sys.argv
is fine for a first version of a program, you should give it a proper CLI. It would be nice if multiple latex files could be passed (since it is common to split a large project into multiple files) and also to specify the source and target languages. For this you can useargparse
:
import argparse
def parse_args(args=None):
parser = argparse.ArgumentParser()
parser.add_argument("--from", default="DE", help="Language of the source document(s)")
parser.add_argument("--to", default="EN", help="Language of the target document")
parser.add_argument("file", nargs="+", help="Path(s) to the latex file(s)")
return parser.parse_args(args)
if __name__ == "__main__":
args = parse_args()
print(args.to, args.from, args.file)
...
This even gives you a nice usage message when calling the script with the option
-h
or--help
:
usage: script_name.py [-h] [--from FROM] [--to TO] file [file ...]
positional arguments:
file Path(s) to the latex file(s)
optional arguments:
-h, --help show this help message and exit
--from FROM Language of the source document(s)
--to TO Language of the target document
Now, let's get to your "hashing". When randomly selecting from $H = 9999 - 1111 = 8888$ values, it takes on average $Q(H) = sqrt{tfrac{pi}{2}H}approx 118.16$ draws until you get a number twice. In other words, if you have more than about 120 elements, you will overwrite some. And that number is actually less than half of that because you separately hash
begin
andend
.
One way to avoid collisions is to just keep on counting up. For this you could use
itertools.count
, which produces an infinite stream of increasing numbers, starting with the argument (or0
if not given), when being iterated:
from itertools import count
counter = count(1111)
assert len(searchObj1) + len(searchObj2) <= 8888, "Too many objects to hash"
list1 = ['X#X{}'.format(i) for _, i in zip(searchObj1, counter)]
list2 = ['X#X{}'.format(i) for _, i in zip(searchObj2, counter)]
Another way to make it less likely is to extend the range. And at that point you might as well use the built-in
hash
function:
list1 = ['X#X{}'.format(hash(x) for x in searchObj1]
In this case you would need to relax the
and len(line) == 7
requirement later.
When opening files, you should use the
with
keyword to ensure they are properly closed, even if an exception occurs somewhere in the block:
with open(fileInputName) as fileIn, open(fileOutName, "w") as fileOut:
...
You should also think about better names. Python has an official style-guide, PEP8, which recommend using
lower_case
instead ofcamelCase
. In addition,list1
,searchObj1
,d2
are all not very good names.
Doing string addition can be very costly, because in Python strings are immutable. This means that when you do
str_a + str_b
, Python has to allocate a new string object of lengthlen(str_a) + len(str_b)
and then copy the content of both strings into that new array. This takes more time the longer the strings are and the more often you do it. Since you are adding strings of the length of a full document and are doing so in a loop, this can be very slow.
Instead, build a
list
andstr.join
it at the end:
translated =
for line in fileIn:
#print(line)
if line.startswith("X#X") and len(line) == 7:
translated.append(line)
elif not line.strip():
translated.append('')
else:
translated.append(pydeepl.translate(line, args.from, args.to))
translated = 'n'.join(translated)
Note that I directly iterate over the file (since files are iterable), meaning that this program is less limited by the amount of memory available. However this means that the progress bar does not work anymore (and I therefore removed it here), because it needs to know how many lines there are. You can use something like this to add it back: https://blog.nelsonliu.me/2016/07/29/progress-bars-for-python-file-reading-with-tqdm/
Thank you very much for those suggestions. I'll definitely implement all of them.
– Mr.Sh4nnon
Dec 5 at 11:07
add a comment |
up vote
3
down vote
naming
both the naming convention is wrong. PEP-008 standardizes on snake_case
for variables and functions, as the naming itself is unclear. What is rx
? It is a re
pattern. Then call it so. adict
is indeed a dict, but it contains the hashes, so also call it like that.
map
map
is a useful tool, but not a lot of people use it or know it. Especially since the advent of generator expressions, it is also hardly ever needed to use. I find a generator expression a lot simpler to read:
def make_xlat(hashes: dict) -> typing.Callable[[str], str]:
pattern = re.compile("|".join(re.escape(key) for key in hashes))
def one_xlat(match):
return hashes[match.group(0)]
def xlat(text: str):
return pattern.sub(one_xlat, text)
return xlat
the hashes
You use a convoluted way str(choice(range(1111, 9999, 1)))
to generate a random number. This will lead to collisions. The easiest way to tackle this, is to use a generator, which keeps what numbers are given out already.
def get_random_ids(begin=0, end=9999):
yield from random.sample(range(begin, end + 1), end - begin + 1)
The downside is this materializes a list
with all the elements. For the 10000 numbers, this is still okay, but if you need a number with 10 digits, this starts to count. An alternative here would be something like this:
def get_random_ids(begin=0, end=9999):
seen = set()
while True:
number = random.randint(begin, end)
if number in seen:
continue
seen.add(number)
yield number
But this only helps if you need a limited number of id's with a long length
To help you with testing, it helps if you can supply a seed to put the pseudo-random generator in the same state every time you test a particular piece
def get_random_ids(begin=0, end=9999, seed=None, ):
"""
generates unique random integers between `begin` and `end`
The random generator can be seeded with `seed`
"""
if seed is not None:
random.seed(seed)
yield from random.sample(range(begin, end + 1), end - begin + 1)
pathlib.Path
If instead of with the bare filenames, you convert them into a Path
, reading and writing becomes a lot easier:
input_file = Path("latex_sample.tex")
input_text = input_file.read_text()
then to output to the hashed text, you can use with_suffix
or with_name
hashed_file = input_file.with_suffix(".hash.tex")
hashed_file.write_text(hashed_text)
regex pattern
you have a very long regex pattern. In your make_xlat
function you assemble one on the fly with '|'.join
. You can do that here as well
commands = (
r"\begin{w+}",
r"t",
" ",
"r",
r"\end{w+}",
r"\usepackage{w+}",
r"\newcommand{w+}",
r"\include{.*}",
r"\input{w+}",
r"\w+[.*}",
r"%.*",
r"\w+",
)
search_pattern = re.compile("|".join(commands))
This way you can easily add commands or add comments for the more obscure commands
the translation
You go over the hashed text line per line, depending on some condition translate, and then do a lot of sting concatenation
This can be done a lot simpler with a generator:
def translate(text: str, lang_in="DE", lang_out="EN"):
hash_pattern = re.compile(r"^X#Xd{4}$")
for line in text.splitlines():
if line in {"", "n"} or hash_pattern.match(line):
yield line
else:
yield pydeepl.translate(line, lang_in, lang_out)
translation = "n".join(translate(hashed_text, lang_in="DE", lang_out="EN"))
Instead of checking startswith
and len
, I used a regular expression
split the work
This code:
- reads a texts
- hashes the commands
- translates the text
- dehashes the commands
- writes to an output file
- save the hashes to a file
You only split of a part of the hashing of the commands. It is simpler, clearer and easier to test if you also split of the rest
def hash_commands(
input_text: str, random_seed=None
) -> typing.Tuple[str, typing.Mapping]:
commands = (
r"\begin{w+}",
r"t",
" ",
"r",
r"\end{w+}",
r"\usepackage{w+}",
r"\newcommand{w+}",
r"\include{.*}",
r"\input{w+}",
r"\w+[.*}",
r"%.*",
r"\w+",
)
search_pattern = re.compile("|".join(commands))
ids = get_random_ids(seed=random_seed)
matches = search_pattern.findall(input_text)
hashes = {
command: f"X#X{id:04}"
for command, id in zip(matches, ids)
}
translate = make_xlat(hashes)
hashed_text = translate(input_text)
return hashed_text, hashes
takes a text, and hashes the commands. A way to generalize this would be to make the commands
tuple an argument, or split of the generation of the hashes
dict to a separate function, and use hashes
as an argument to the hash_commands
function.
def dehash_text(hashed_text: str, hashes: typing.Mapping):
hashes_inverse = {hash: command for command, hash in hashes.items()}
translate = make_xlat(hashes_inverse)
return translate(hashed_text)
does the inverse.
def save_hashes(hashes, out_file):
hashes_inverse = {hash: command for command, hash in hashes.items()}
json.dump(
{"hashes": hashes, "inverse": hashes_inverse}, out_file, indent=4
)
and then the main logic become very simple and easy to understand
if __name__ == "__main__":
input_file = Path("latex_sample.tex")
input_text = input_file.read_text()
hashed_text, hashes = hash_commands(input_text=input_text, random_seed=42)
hash_file = Path("hash_dict.json")
with hash_file.open("w") as hash_filehandle:
save_hashes(hashes, hash_filehandle)
hashed_file = input_file.with_suffix(".hash.tex")
hashed_file.write_text(hashed_text)
translation = "n".join(translate(hashed_text, lang_in="DE", lang_out="EN"))
translation_dehashed = dehash_text(translation, hashes)
# print(translation_dehashed)
output_file = input_file.with_suffix(".trans.tex")
output_file.write_text(translation_dehashed)
I included an output of the hashed text, so you can see whether the function correctly recognises the commands. That way you see it doesn't recognise the usepackage{tree-dvips}
Thanks! learned a lot while reading this. Will add this this evening!
– Mr.Sh4nnon
Dec 5 at 13:53
add a comment |
2 Answers
2
active
oldest
votes
2 Answers
2
active
oldest
votes
active
oldest
votes
active
oldest
votes
up vote
5
down vote
accepted
While using
sys.argv
is fine for a first version of a program, you should give it a proper CLI. It would be nice if multiple latex files could be passed (since it is common to split a large project into multiple files) and also to specify the source and target languages. For this you can useargparse
:
import argparse
def parse_args(args=None):
parser = argparse.ArgumentParser()
parser.add_argument("--from", default="DE", help="Language of the source document(s)")
parser.add_argument("--to", default="EN", help="Language of the target document")
parser.add_argument("file", nargs="+", help="Path(s) to the latex file(s)")
return parser.parse_args(args)
if __name__ == "__main__":
args = parse_args()
print(args.to, args.from, args.file)
...
This even gives you a nice usage message when calling the script with the option
-h
or--help
:
usage: script_name.py [-h] [--from FROM] [--to TO] file [file ...]
positional arguments:
file Path(s) to the latex file(s)
optional arguments:
-h, --help show this help message and exit
--from FROM Language of the source document(s)
--to TO Language of the target document
Now, let's get to your "hashing". When randomly selecting from $H = 9999 - 1111 = 8888$ values, it takes on average $Q(H) = sqrt{tfrac{pi}{2}H}approx 118.16$ draws until you get a number twice. In other words, if you have more than about 120 elements, you will overwrite some. And that number is actually less than half of that because you separately hash
begin
andend
.
One way to avoid collisions is to just keep on counting up. For this you could use
itertools.count
, which produces an infinite stream of increasing numbers, starting with the argument (or0
if not given), when being iterated:
from itertools import count
counter = count(1111)
assert len(searchObj1) + len(searchObj2) <= 8888, "Too many objects to hash"
list1 = ['X#X{}'.format(i) for _, i in zip(searchObj1, counter)]
list2 = ['X#X{}'.format(i) for _, i in zip(searchObj2, counter)]
Another way to make it less likely is to extend the range. And at that point you might as well use the built-in
hash
function:
list1 = ['X#X{}'.format(hash(x) for x in searchObj1]
In this case you would need to relax the
and len(line) == 7
requirement later.
When opening files, you should use the
with
keyword to ensure they are properly closed, even if an exception occurs somewhere in the block:
with open(fileInputName) as fileIn, open(fileOutName, "w") as fileOut:
...
You should also think about better names. Python has an official style-guide, PEP8, which recommend using
lower_case
instead ofcamelCase
. In addition,list1
,searchObj1
,d2
are all not very good names.
Doing string addition can be very costly, because in Python strings are immutable. This means that when you do
str_a + str_b
, Python has to allocate a new string object of lengthlen(str_a) + len(str_b)
and then copy the content of both strings into that new array. This takes more time the longer the strings are and the more often you do it. Since you are adding strings of the length of a full document and are doing so in a loop, this can be very slow.
Instead, build a
list
andstr.join
it at the end:
translated =
for line in fileIn:
#print(line)
if line.startswith("X#X") and len(line) == 7:
translated.append(line)
elif not line.strip():
translated.append('')
else:
translated.append(pydeepl.translate(line, args.from, args.to))
translated = 'n'.join(translated)
Note that I directly iterate over the file (since files are iterable), meaning that this program is less limited by the amount of memory available. However this means that the progress bar does not work anymore (and I therefore removed it here), because it needs to know how many lines there are. You can use something like this to add it back: https://blog.nelsonliu.me/2016/07/29/progress-bars-for-python-file-reading-with-tqdm/
Thank you very much for those suggestions. I'll definitely implement all of them.
– Mr.Sh4nnon
Dec 5 at 11:07
add a comment |
up vote
5
down vote
accepted
While using
sys.argv
is fine for a first version of a program, you should give it a proper CLI. It would be nice if multiple latex files could be passed (since it is common to split a large project into multiple files) and also to specify the source and target languages. For this you can useargparse
:
import argparse
def parse_args(args=None):
parser = argparse.ArgumentParser()
parser.add_argument("--from", default="DE", help="Language of the source document(s)")
parser.add_argument("--to", default="EN", help="Language of the target document")
parser.add_argument("file", nargs="+", help="Path(s) to the latex file(s)")
return parser.parse_args(args)
if __name__ == "__main__":
args = parse_args()
print(args.to, args.from, args.file)
...
This even gives you a nice usage message when calling the script with the option
-h
or--help
:
usage: script_name.py [-h] [--from FROM] [--to TO] file [file ...]
positional arguments:
file Path(s) to the latex file(s)
optional arguments:
-h, --help show this help message and exit
--from FROM Language of the source document(s)
--to TO Language of the target document
Now, let's get to your "hashing". When randomly selecting from $H = 9999 - 1111 = 8888$ values, it takes on average $Q(H) = sqrt{tfrac{pi}{2}H}approx 118.16$ draws until you get a number twice. In other words, if you have more than about 120 elements, you will overwrite some. And that number is actually less than half of that because you separately hash
begin
andend
.
One way to avoid collisions is to just keep on counting up. For this you could use
itertools.count
, which produces an infinite stream of increasing numbers, starting with the argument (or0
if not given), when being iterated:
from itertools import count
counter = count(1111)
assert len(searchObj1) + len(searchObj2) <= 8888, "Too many objects to hash"
list1 = ['X#X{}'.format(i) for _, i in zip(searchObj1, counter)]
list2 = ['X#X{}'.format(i) for _, i in zip(searchObj2, counter)]
Another way to make it less likely is to extend the range. And at that point you might as well use the built-in
hash
function:
list1 = ['X#X{}'.format(hash(x) for x in searchObj1]
In this case you would need to relax the
and len(line) == 7
requirement later.
When opening files, you should use the
with
keyword to ensure they are properly closed, even if an exception occurs somewhere in the block:
with open(fileInputName) as fileIn, open(fileOutName, "w") as fileOut:
...
You should also think about better names. Python has an official style-guide, PEP8, which recommend using
lower_case
instead ofcamelCase
. In addition,list1
,searchObj1
,d2
are all not very good names.
Doing string addition can be very costly, because in Python strings are immutable. This means that when you do
str_a + str_b
, Python has to allocate a new string object of lengthlen(str_a) + len(str_b)
and then copy the content of both strings into that new array. This takes more time the longer the strings are and the more often you do it. Since you are adding strings of the length of a full document and are doing so in a loop, this can be very slow.
Instead, build a
list
andstr.join
it at the end:
translated =
for line in fileIn:
#print(line)
if line.startswith("X#X") and len(line) == 7:
translated.append(line)
elif not line.strip():
translated.append('')
else:
translated.append(pydeepl.translate(line, args.from, args.to))
translated = 'n'.join(translated)
Note that I directly iterate over the file (since files are iterable), meaning that this program is less limited by the amount of memory available. However this means that the progress bar does not work anymore (and I therefore removed it here), because it needs to know how many lines there are. You can use something like this to add it back: https://blog.nelsonliu.me/2016/07/29/progress-bars-for-python-file-reading-with-tqdm/
Thank you very much for those suggestions. I'll definitely implement all of them.
– Mr.Sh4nnon
Dec 5 at 11:07
add a comment |
up vote
5
down vote
accepted
up vote
5
down vote
accepted
While using
sys.argv
is fine for a first version of a program, you should give it a proper CLI. It would be nice if multiple latex files could be passed (since it is common to split a large project into multiple files) and also to specify the source and target languages. For this you can useargparse
:
import argparse
def parse_args(args=None):
parser = argparse.ArgumentParser()
parser.add_argument("--from", default="DE", help="Language of the source document(s)")
parser.add_argument("--to", default="EN", help="Language of the target document")
parser.add_argument("file", nargs="+", help="Path(s) to the latex file(s)")
return parser.parse_args(args)
if __name__ == "__main__":
args = parse_args()
print(args.to, args.from, args.file)
...
This even gives you a nice usage message when calling the script with the option
-h
or--help
:
usage: script_name.py [-h] [--from FROM] [--to TO] file [file ...]
positional arguments:
file Path(s) to the latex file(s)
optional arguments:
-h, --help show this help message and exit
--from FROM Language of the source document(s)
--to TO Language of the target document
Now, let's get to your "hashing". When randomly selecting from $H = 9999 - 1111 = 8888$ values, it takes on average $Q(H) = sqrt{tfrac{pi}{2}H}approx 118.16$ draws until you get a number twice. In other words, if you have more than about 120 elements, you will overwrite some. And that number is actually less than half of that because you separately hash
begin
andend
.
One way to avoid collisions is to just keep on counting up. For this you could use
itertools.count
, which produces an infinite stream of increasing numbers, starting with the argument (or0
if not given), when being iterated:
from itertools import count
counter = count(1111)
assert len(searchObj1) + len(searchObj2) <= 8888, "Too many objects to hash"
list1 = ['X#X{}'.format(i) for _, i in zip(searchObj1, counter)]
list2 = ['X#X{}'.format(i) for _, i in zip(searchObj2, counter)]
Another way to make it less likely is to extend the range. And at that point you might as well use the built-in
hash
function:
list1 = ['X#X{}'.format(hash(x) for x in searchObj1]
In this case you would need to relax the
and len(line) == 7
requirement later.
When opening files, you should use the
with
keyword to ensure they are properly closed, even if an exception occurs somewhere in the block:
with open(fileInputName) as fileIn, open(fileOutName, "w") as fileOut:
...
You should also think about better names. Python has an official style-guide, PEP8, which recommend using
lower_case
instead ofcamelCase
. In addition,list1
,searchObj1
,d2
are all not very good names.
Doing string addition can be very costly, because in Python strings are immutable. This means that when you do
str_a + str_b
, Python has to allocate a new string object of lengthlen(str_a) + len(str_b)
and then copy the content of both strings into that new array. This takes more time the longer the strings are and the more often you do it. Since you are adding strings of the length of a full document and are doing so in a loop, this can be very slow.
Instead, build a
list
andstr.join
it at the end:
translated =
for line in fileIn:
#print(line)
if line.startswith("X#X") and len(line) == 7:
translated.append(line)
elif not line.strip():
translated.append('')
else:
translated.append(pydeepl.translate(line, args.from, args.to))
translated = 'n'.join(translated)
Note that I directly iterate over the file (since files are iterable), meaning that this program is less limited by the amount of memory available. However this means that the progress bar does not work anymore (and I therefore removed it here), because it needs to know how many lines there are. You can use something like this to add it back: https://blog.nelsonliu.me/2016/07/29/progress-bars-for-python-file-reading-with-tqdm/
While using
sys.argv
is fine for a first version of a program, you should give it a proper CLI. It would be nice if multiple latex files could be passed (since it is common to split a large project into multiple files) and also to specify the source and target languages. For this you can useargparse
:
import argparse
def parse_args(args=None):
parser = argparse.ArgumentParser()
parser.add_argument("--from", default="DE", help="Language of the source document(s)")
parser.add_argument("--to", default="EN", help="Language of the target document")
parser.add_argument("file", nargs="+", help="Path(s) to the latex file(s)")
return parser.parse_args(args)
if __name__ == "__main__":
args = parse_args()
print(args.to, args.from, args.file)
...
This even gives you a nice usage message when calling the script with the option
-h
or--help
:
usage: script_name.py [-h] [--from FROM] [--to TO] file [file ...]
positional arguments:
file Path(s) to the latex file(s)
optional arguments:
-h, --help show this help message and exit
--from FROM Language of the source document(s)
--to TO Language of the target document
Now, let's get to your "hashing". When randomly selecting from $H = 9999 - 1111 = 8888$ values, it takes on average $Q(H) = sqrt{tfrac{pi}{2}H}approx 118.16$ draws until you get a number twice. In other words, if you have more than about 120 elements, you will overwrite some. And that number is actually less than half of that because you separately hash
begin
andend
.
One way to avoid collisions is to just keep on counting up. For this you could use
itertools.count
, which produces an infinite stream of increasing numbers, starting with the argument (or0
if not given), when being iterated:
from itertools import count
counter = count(1111)
assert len(searchObj1) + len(searchObj2) <= 8888, "Too many objects to hash"
list1 = ['X#X{}'.format(i) for _, i in zip(searchObj1, counter)]
list2 = ['X#X{}'.format(i) for _, i in zip(searchObj2, counter)]
Another way to make it less likely is to extend the range. And at that point you might as well use the built-in
hash
function:
list1 = ['X#X{}'.format(hash(x) for x in searchObj1]
In this case you would need to relax the
and len(line) == 7
requirement later.
When opening files, you should use the
with
keyword to ensure they are properly closed, even if an exception occurs somewhere in the block:
with open(fileInputName) as fileIn, open(fileOutName, "w") as fileOut:
...
You should also think about better names. Python has an official style-guide, PEP8, which recommend using
lower_case
instead ofcamelCase
. In addition,list1
,searchObj1
,d2
are all not very good names.
Doing string addition can be very costly, because in Python strings are immutable. This means that when you do
str_a + str_b
, Python has to allocate a new string object of lengthlen(str_a) + len(str_b)
and then copy the content of both strings into that new array. This takes more time the longer the strings are and the more often you do it. Since you are adding strings of the length of a full document and are doing so in a loop, this can be very slow.
Instead, build a
list
andstr.join
it at the end:
translated =
for line in fileIn:
#print(line)
if line.startswith("X#X") and len(line) == 7:
translated.append(line)
elif not line.strip():
translated.append('')
else:
translated.append(pydeepl.translate(line, args.from, args.to))
translated = 'n'.join(translated)
Note that I directly iterate over the file (since files are iterable), meaning that this program is less limited by the amount of memory available. However this means that the progress bar does not work anymore (and I therefore removed it here), because it needs to know how many lines there are. You can use something like this to add it back: https://blog.nelsonliu.me/2016/07/29/progress-bars-for-python-file-reading-with-tqdm/
answered Dec 5 at 10:52
Graipher
23k53384
23k53384
Thank you very much for those suggestions. I'll definitely implement all of them.
– Mr.Sh4nnon
Dec 5 at 11:07
add a comment |
Thank you very much for those suggestions. I'll definitely implement all of them.
– Mr.Sh4nnon
Dec 5 at 11:07
Thank you very much for those suggestions. I'll definitely implement all of them.
– Mr.Sh4nnon
Dec 5 at 11:07
Thank you very much for those suggestions. I'll definitely implement all of them.
– Mr.Sh4nnon
Dec 5 at 11:07
add a comment |
up vote
3
down vote
naming
both the naming convention is wrong. PEP-008 standardizes on snake_case
for variables and functions, as the naming itself is unclear. What is rx
? It is a re
pattern. Then call it so. adict
is indeed a dict, but it contains the hashes, so also call it like that.
map
map
is a useful tool, but not a lot of people use it or know it. Especially since the advent of generator expressions, it is also hardly ever needed to use. I find a generator expression a lot simpler to read:
def make_xlat(hashes: dict) -> typing.Callable[[str], str]:
pattern = re.compile("|".join(re.escape(key) for key in hashes))
def one_xlat(match):
return hashes[match.group(0)]
def xlat(text: str):
return pattern.sub(one_xlat, text)
return xlat
the hashes
You use a convoluted way str(choice(range(1111, 9999, 1)))
to generate a random number. This will lead to collisions. The easiest way to tackle this, is to use a generator, which keeps what numbers are given out already.
def get_random_ids(begin=0, end=9999):
yield from random.sample(range(begin, end + 1), end - begin + 1)
The downside is this materializes a list
with all the elements. For the 10000 numbers, this is still okay, but if you need a number with 10 digits, this starts to count. An alternative here would be something like this:
def get_random_ids(begin=0, end=9999):
seen = set()
while True:
number = random.randint(begin, end)
if number in seen:
continue
seen.add(number)
yield number
But this only helps if you need a limited number of id's with a long length
To help you with testing, it helps if you can supply a seed to put the pseudo-random generator in the same state every time you test a particular piece
def get_random_ids(begin=0, end=9999, seed=None, ):
"""
generates unique random integers between `begin` and `end`
The random generator can be seeded with `seed`
"""
if seed is not None:
random.seed(seed)
yield from random.sample(range(begin, end + 1), end - begin + 1)
pathlib.Path
If instead of with the bare filenames, you convert them into a Path
, reading and writing becomes a lot easier:
input_file = Path("latex_sample.tex")
input_text = input_file.read_text()
then to output to the hashed text, you can use with_suffix
or with_name
hashed_file = input_file.with_suffix(".hash.tex")
hashed_file.write_text(hashed_text)
regex pattern
you have a very long regex pattern. In your make_xlat
function you assemble one on the fly with '|'.join
. You can do that here as well
commands = (
r"\begin{w+}",
r"t",
" ",
"r",
r"\end{w+}",
r"\usepackage{w+}",
r"\newcommand{w+}",
r"\include{.*}",
r"\input{w+}",
r"\w+[.*}",
r"%.*",
r"\w+",
)
search_pattern = re.compile("|".join(commands))
This way you can easily add commands or add comments for the more obscure commands
the translation
You go over the hashed text line per line, depending on some condition translate, and then do a lot of sting concatenation
This can be done a lot simpler with a generator:
def translate(text: str, lang_in="DE", lang_out="EN"):
hash_pattern = re.compile(r"^X#Xd{4}$")
for line in text.splitlines():
if line in {"", "n"} or hash_pattern.match(line):
yield line
else:
yield pydeepl.translate(line, lang_in, lang_out)
translation = "n".join(translate(hashed_text, lang_in="DE", lang_out="EN"))
Instead of checking startswith
and len
, I used a regular expression
split the work
This code:
- reads a texts
- hashes the commands
- translates the text
- dehashes the commands
- writes to an output file
- save the hashes to a file
You only split of a part of the hashing of the commands. It is simpler, clearer and easier to test if you also split of the rest
def hash_commands(
input_text: str, random_seed=None
) -> typing.Tuple[str, typing.Mapping]:
commands = (
r"\begin{w+}",
r"t",
" ",
"r",
r"\end{w+}",
r"\usepackage{w+}",
r"\newcommand{w+}",
r"\include{.*}",
r"\input{w+}",
r"\w+[.*}",
r"%.*",
r"\w+",
)
search_pattern = re.compile("|".join(commands))
ids = get_random_ids(seed=random_seed)
matches = search_pattern.findall(input_text)
hashes = {
command: f"X#X{id:04}"
for command, id in zip(matches, ids)
}
translate = make_xlat(hashes)
hashed_text = translate(input_text)
return hashed_text, hashes
takes a text, and hashes the commands. A way to generalize this would be to make the commands
tuple an argument, or split of the generation of the hashes
dict to a separate function, and use hashes
as an argument to the hash_commands
function.
def dehash_text(hashed_text: str, hashes: typing.Mapping):
hashes_inverse = {hash: command for command, hash in hashes.items()}
translate = make_xlat(hashes_inverse)
return translate(hashed_text)
does the inverse.
def save_hashes(hashes, out_file):
hashes_inverse = {hash: command for command, hash in hashes.items()}
json.dump(
{"hashes": hashes, "inverse": hashes_inverse}, out_file, indent=4
)
and then the main logic become very simple and easy to understand
if __name__ == "__main__":
input_file = Path("latex_sample.tex")
input_text = input_file.read_text()
hashed_text, hashes = hash_commands(input_text=input_text, random_seed=42)
hash_file = Path("hash_dict.json")
with hash_file.open("w") as hash_filehandle:
save_hashes(hashes, hash_filehandle)
hashed_file = input_file.with_suffix(".hash.tex")
hashed_file.write_text(hashed_text)
translation = "n".join(translate(hashed_text, lang_in="DE", lang_out="EN"))
translation_dehashed = dehash_text(translation, hashes)
# print(translation_dehashed)
output_file = input_file.with_suffix(".trans.tex")
output_file.write_text(translation_dehashed)
I included an output of the hashed text, so you can see whether the function correctly recognises the commands. That way you see it doesn't recognise the usepackage{tree-dvips}
Thanks! learned a lot while reading this. Will add this this evening!
– Mr.Sh4nnon
Dec 5 at 13:53
add a comment |
up vote
3
down vote
naming
both the naming convention is wrong. PEP-008 standardizes on snake_case
for variables and functions, as the naming itself is unclear. What is rx
? It is a re
pattern. Then call it so. adict
is indeed a dict, but it contains the hashes, so also call it like that.
map
map
is a useful tool, but not a lot of people use it or know it. Especially since the advent of generator expressions, it is also hardly ever needed to use. I find a generator expression a lot simpler to read:
def make_xlat(hashes: dict) -> typing.Callable[[str], str]:
pattern = re.compile("|".join(re.escape(key) for key in hashes))
def one_xlat(match):
return hashes[match.group(0)]
def xlat(text: str):
return pattern.sub(one_xlat, text)
return xlat
the hashes
You use a convoluted way str(choice(range(1111, 9999, 1)))
to generate a random number. This will lead to collisions. The easiest way to tackle this, is to use a generator, which keeps what numbers are given out already.
def get_random_ids(begin=0, end=9999):
yield from random.sample(range(begin, end + 1), end - begin + 1)
The downside is this materializes a list
with all the elements. For the 10000 numbers, this is still okay, but if you need a number with 10 digits, this starts to count. An alternative here would be something like this:
def get_random_ids(begin=0, end=9999):
seen = set()
while True:
number = random.randint(begin, end)
if number in seen:
continue
seen.add(number)
yield number
But this only helps if you need a limited number of id's with a long length
To help you with testing, it helps if you can supply a seed to put the pseudo-random generator in the same state every time you test a particular piece
def get_random_ids(begin=0, end=9999, seed=None, ):
"""
generates unique random integers between `begin` and `end`
The random generator can be seeded with `seed`
"""
if seed is not None:
random.seed(seed)
yield from random.sample(range(begin, end + 1), end - begin + 1)
pathlib.Path
If instead of with the bare filenames, you convert them into a Path
, reading and writing becomes a lot easier:
input_file = Path("latex_sample.tex")
input_text = input_file.read_text()
then to output to the hashed text, you can use with_suffix
or with_name
hashed_file = input_file.with_suffix(".hash.tex")
hashed_file.write_text(hashed_text)
regex pattern
you have a very long regex pattern. In your make_xlat
function you assemble one on the fly with '|'.join
. You can do that here as well
commands = (
r"\begin{w+}",
r"t",
" ",
"r",
r"\end{w+}",
r"\usepackage{w+}",
r"\newcommand{w+}",
r"\include{.*}",
r"\input{w+}",
r"\w+[.*}",
r"%.*",
r"\w+",
)
search_pattern = re.compile("|".join(commands))
This way you can easily add commands or add comments for the more obscure commands
the translation
You go over the hashed text line per line, depending on some condition translate, and then do a lot of sting concatenation
This can be done a lot simpler with a generator:
def translate(text: str, lang_in="DE", lang_out="EN"):
hash_pattern = re.compile(r"^X#Xd{4}$")
for line in text.splitlines():
if line in {"", "n"} or hash_pattern.match(line):
yield line
else:
yield pydeepl.translate(line, lang_in, lang_out)
translation = "n".join(translate(hashed_text, lang_in="DE", lang_out="EN"))
Instead of checking startswith
and len
, I used a regular expression
split the work
This code:
- reads a texts
- hashes the commands
- translates the text
- dehashes the commands
- writes to an output file
- save the hashes to a file
You only split of a part of the hashing of the commands. It is simpler, clearer and easier to test if you also split of the rest
def hash_commands(
input_text: str, random_seed=None
) -> typing.Tuple[str, typing.Mapping]:
commands = (
r"\begin{w+}",
r"t",
" ",
"r",
r"\end{w+}",
r"\usepackage{w+}",
r"\newcommand{w+}",
r"\include{.*}",
r"\input{w+}",
r"\w+[.*}",
r"%.*",
r"\w+",
)
search_pattern = re.compile("|".join(commands))
ids = get_random_ids(seed=random_seed)
matches = search_pattern.findall(input_text)
hashes = {
command: f"X#X{id:04}"
for command, id in zip(matches, ids)
}
translate = make_xlat(hashes)
hashed_text = translate(input_text)
return hashed_text, hashes
takes a text, and hashes the commands. A way to generalize this would be to make the commands
tuple an argument, or split of the generation of the hashes
dict to a separate function, and use hashes
as an argument to the hash_commands
function.
def dehash_text(hashed_text: str, hashes: typing.Mapping):
hashes_inverse = {hash: command for command, hash in hashes.items()}
translate = make_xlat(hashes_inverse)
return translate(hashed_text)
does the inverse.
def save_hashes(hashes, out_file):
hashes_inverse = {hash: command for command, hash in hashes.items()}
json.dump(
{"hashes": hashes, "inverse": hashes_inverse}, out_file, indent=4
)
and then the main logic become very simple and easy to understand
if __name__ == "__main__":
input_file = Path("latex_sample.tex")
input_text = input_file.read_text()
hashed_text, hashes = hash_commands(input_text=input_text, random_seed=42)
hash_file = Path("hash_dict.json")
with hash_file.open("w") as hash_filehandle:
save_hashes(hashes, hash_filehandle)
hashed_file = input_file.with_suffix(".hash.tex")
hashed_file.write_text(hashed_text)
translation = "n".join(translate(hashed_text, lang_in="DE", lang_out="EN"))
translation_dehashed = dehash_text(translation, hashes)
# print(translation_dehashed)
output_file = input_file.with_suffix(".trans.tex")
output_file.write_text(translation_dehashed)
I included an output of the hashed text, so you can see whether the function correctly recognises the commands. That way you see it doesn't recognise the usepackage{tree-dvips}
Thanks! learned a lot while reading this. Will add this this evening!
– Mr.Sh4nnon
Dec 5 at 13:53
add a comment |
up vote
3
down vote
up vote
3
down vote
naming
both the naming convention is wrong. PEP-008 standardizes on snake_case
for variables and functions, as the naming itself is unclear. What is rx
? It is a re
pattern. Then call it so. adict
is indeed a dict, but it contains the hashes, so also call it like that.
map
map
is a useful tool, but not a lot of people use it or know it. Especially since the advent of generator expressions, it is also hardly ever needed to use. I find a generator expression a lot simpler to read:
def make_xlat(hashes: dict) -> typing.Callable[[str], str]:
pattern = re.compile("|".join(re.escape(key) for key in hashes))
def one_xlat(match):
return hashes[match.group(0)]
def xlat(text: str):
return pattern.sub(one_xlat, text)
return xlat
the hashes
You use a convoluted way str(choice(range(1111, 9999, 1)))
to generate a random number. This will lead to collisions. The easiest way to tackle this, is to use a generator, which keeps what numbers are given out already.
def get_random_ids(begin=0, end=9999):
yield from random.sample(range(begin, end + 1), end - begin + 1)
The downside is this materializes a list
with all the elements. For the 10000 numbers, this is still okay, but if you need a number with 10 digits, this starts to count. An alternative here would be something like this:
def get_random_ids(begin=0, end=9999):
seen = set()
while True:
number = random.randint(begin, end)
if number in seen:
continue
seen.add(number)
yield number
But this only helps if you need a limited number of id's with a long length
To help you with testing, it helps if you can supply a seed to put the pseudo-random generator in the same state every time you test a particular piece
def get_random_ids(begin=0, end=9999, seed=None, ):
"""
generates unique random integers between `begin` and `end`
The random generator can be seeded with `seed`
"""
if seed is not None:
random.seed(seed)
yield from random.sample(range(begin, end + 1), end - begin + 1)
pathlib.Path
If instead of with the bare filenames, you convert them into a Path
, reading and writing becomes a lot easier:
input_file = Path("latex_sample.tex")
input_text = input_file.read_text()
then to output to the hashed text, you can use with_suffix
or with_name
hashed_file = input_file.with_suffix(".hash.tex")
hashed_file.write_text(hashed_text)
regex pattern
you have a very long regex pattern. In your make_xlat
function you assemble one on the fly with '|'.join
. You can do that here as well
commands = (
r"\begin{w+}",
r"t",
" ",
"r",
r"\end{w+}",
r"\usepackage{w+}",
r"\newcommand{w+}",
r"\include{.*}",
r"\input{w+}",
r"\w+[.*}",
r"%.*",
r"\w+",
)
search_pattern = re.compile("|".join(commands))
This way you can easily add commands or add comments for the more obscure commands
the translation
You go over the hashed text line per line, depending on some condition translate, and then do a lot of sting concatenation
This can be done a lot simpler with a generator:
def translate(text: str, lang_in="DE", lang_out="EN"):
hash_pattern = re.compile(r"^X#Xd{4}$")
for line in text.splitlines():
if line in {"", "n"} or hash_pattern.match(line):
yield line
else:
yield pydeepl.translate(line, lang_in, lang_out)
translation = "n".join(translate(hashed_text, lang_in="DE", lang_out="EN"))
Instead of checking startswith
and len
, I used a regular expression
split the work
This code:
- reads a texts
- hashes the commands
- translates the text
- dehashes the commands
- writes to an output file
- save the hashes to a file
You only split of a part of the hashing of the commands. It is simpler, clearer and easier to test if you also split of the rest
def hash_commands(
input_text: str, random_seed=None
) -> typing.Tuple[str, typing.Mapping]:
commands = (
r"\begin{w+}",
r"t",
" ",
"r",
r"\end{w+}",
r"\usepackage{w+}",
r"\newcommand{w+}",
r"\include{.*}",
r"\input{w+}",
r"\w+[.*}",
r"%.*",
r"\w+",
)
search_pattern = re.compile("|".join(commands))
ids = get_random_ids(seed=random_seed)
matches = search_pattern.findall(input_text)
hashes = {
command: f"X#X{id:04}"
for command, id in zip(matches, ids)
}
translate = make_xlat(hashes)
hashed_text = translate(input_text)
return hashed_text, hashes
takes a text, and hashes the commands. A way to generalize this would be to make the commands
tuple an argument, or split of the generation of the hashes
dict to a separate function, and use hashes
as an argument to the hash_commands
function.
def dehash_text(hashed_text: str, hashes: typing.Mapping):
hashes_inverse = {hash: command for command, hash in hashes.items()}
translate = make_xlat(hashes_inverse)
return translate(hashed_text)
does the inverse.
def save_hashes(hashes, out_file):
hashes_inverse = {hash: command for command, hash in hashes.items()}
json.dump(
{"hashes": hashes, "inverse": hashes_inverse}, out_file, indent=4
)
and then the main logic become very simple and easy to understand
if __name__ == "__main__":
input_file = Path("latex_sample.tex")
input_text = input_file.read_text()
hashed_text, hashes = hash_commands(input_text=input_text, random_seed=42)
hash_file = Path("hash_dict.json")
with hash_file.open("w") as hash_filehandle:
save_hashes(hashes, hash_filehandle)
hashed_file = input_file.with_suffix(".hash.tex")
hashed_file.write_text(hashed_text)
translation = "n".join(translate(hashed_text, lang_in="DE", lang_out="EN"))
translation_dehashed = dehash_text(translation, hashes)
# print(translation_dehashed)
output_file = input_file.with_suffix(".trans.tex")
output_file.write_text(translation_dehashed)
I included an output of the hashed text, so you can see whether the function correctly recognises the commands. That way you see it doesn't recognise the usepackage{tree-dvips}
naming
both the naming convention is wrong. PEP-008 standardizes on snake_case
for variables and functions, as the naming itself is unclear. What is rx
? It is a re
pattern. Then call it so. adict
is indeed a dict, but it contains the hashes, so also call it like that.
map
map
is a useful tool, but not a lot of people use it or know it. Especially since the advent of generator expressions, it is also hardly ever needed to use. I find a generator expression a lot simpler to read:
def make_xlat(hashes: dict) -> typing.Callable[[str], str]:
pattern = re.compile("|".join(re.escape(key) for key in hashes))
def one_xlat(match):
return hashes[match.group(0)]
def xlat(text: str):
return pattern.sub(one_xlat, text)
return xlat
the hashes
You use a convoluted way str(choice(range(1111, 9999, 1)))
to generate a random number. This will lead to collisions. The easiest way to tackle this, is to use a generator, which keeps what numbers are given out already.
def get_random_ids(begin=0, end=9999):
yield from random.sample(range(begin, end + 1), end - begin + 1)
The downside is this materializes a list
with all the elements. For the 10000 numbers, this is still okay, but if you need a number with 10 digits, this starts to count. An alternative here would be something like this:
def get_random_ids(begin=0, end=9999):
seen = set()
while True:
number = random.randint(begin, end)
if number in seen:
continue
seen.add(number)
yield number
But this only helps if you need a limited number of id's with a long length
To help you with testing, it helps if you can supply a seed to put the pseudo-random generator in the same state every time you test a particular piece
def get_random_ids(begin=0, end=9999, seed=None, ):
"""
generates unique random integers between `begin` and `end`
The random generator can be seeded with `seed`
"""
if seed is not None:
random.seed(seed)
yield from random.sample(range(begin, end + 1), end - begin + 1)
pathlib.Path
If instead of with the bare filenames, you convert them into a Path
, reading and writing becomes a lot easier:
input_file = Path("latex_sample.tex")
input_text = input_file.read_text()
then to output to the hashed text, you can use with_suffix
or with_name
hashed_file = input_file.with_suffix(".hash.tex")
hashed_file.write_text(hashed_text)
regex pattern
you have a very long regex pattern. In your make_xlat
function you assemble one on the fly with '|'.join
. You can do that here as well
commands = (
r"\begin{w+}",
r"t",
" ",
"r",
r"\end{w+}",
r"\usepackage{w+}",
r"\newcommand{w+}",
r"\include{.*}",
r"\input{w+}",
r"\w+[.*}",
r"%.*",
r"\w+",
)
search_pattern = re.compile("|".join(commands))
This way you can easily add commands or add comments for the more obscure commands
the translation
You go over the hashed text line per line, depending on some condition translate, and then do a lot of sting concatenation
This can be done a lot simpler with a generator:
def translate(text: str, lang_in="DE", lang_out="EN"):
hash_pattern = re.compile(r"^X#Xd{4}$")
for line in text.splitlines():
if line in {"", "n"} or hash_pattern.match(line):
yield line
else:
yield pydeepl.translate(line, lang_in, lang_out)
translation = "n".join(translate(hashed_text, lang_in="DE", lang_out="EN"))
Instead of checking startswith
and len
, I used a regular expression
split the work
This code:
- reads a texts
- hashes the commands
- translates the text
- dehashes the commands
- writes to an output file
- save the hashes to a file
You only split of a part of the hashing of the commands. It is simpler, clearer and easier to test if you also split of the rest
def hash_commands(
input_text: str, random_seed=None
) -> typing.Tuple[str, typing.Mapping]:
commands = (
r"\begin{w+}",
r"t",
" ",
"r",
r"\end{w+}",
r"\usepackage{w+}",
r"\newcommand{w+}",
r"\include{.*}",
r"\input{w+}",
r"\w+[.*}",
r"%.*",
r"\w+",
)
search_pattern = re.compile("|".join(commands))
ids = get_random_ids(seed=random_seed)
matches = search_pattern.findall(input_text)
hashes = {
command: f"X#X{id:04}"
for command, id in zip(matches, ids)
}
translate = make_xlat(hashes)
hashed_text = translate(input_text)
return hashed_text, hashes
takes a text, and hashes the commands. A way to generalize this would be to make the commands
tuple an argument, or split of the generation of the hashes
dict to a separate function, and use hashes
as an argument to the hash_commands
function.
def dehash_text(hashed_text: str, hashes: typing.Mapping):
hashes_inverse = {hash: command for command, hash in hashes.items()}
translate = make_xlat(hashes_inverse)
return translate(hashed_text)
does the inverse.
def save_hashes(hashes, out_file):
hashes_inverse = {hash: command for command, hash in hashes.items()}
json.dump(
{"hashes": hashes, "inverse": hashes_inverse}, out_file, indent=4
)
and then the main logic become very simple and easy to understand
if __name__ == "__main__":
input_file = Path("latex_sample.tex")
input_text = input_file.read_text()
hashed_text, hashes = hash_commands(input_text=input_text, random_seed=42)
hash_file = Path("hash_dict.json")
with hash_file.open("w") as hash_filehandle:
save_hashes(hashes, hash_filehandle)
hashed_file = input_file.with_suffix(".hash.tex")
hashed_file.write_text(hashed_text)
translation = "n".join(translate(hashed_text, lang_in="DE", lang_out="EN"))
translation_dehashed = dehash_text(translation, hashes)
# print(translation_dehashed)
output_file = input_file.with_suffix(".trans.tex")
output_file.write_text(translation_dehashed)
I included an output of the hashed text, so you can see whether the function correctly recognises the commands. That way you see it doesn't recognise the usepackage{tree-dvips}
edited Dec 5 at 13:58
answered Dec 5 at 13:40
Maarten Fabré
4,384417
4,384417
Thanks! learned a lot while reading this. Will add this this evening!
– Mr.Sh4nnon
Dec 5 at 13:53
add a comment |
Thanks! learned a lot while reading this. Will add this this evening!
– Mr.Sh4nnon
Dec 5 at 13:53
Thanks! learned a lot while reading this. Will add this this evening!
– Mr.Sh4nnon
Dec 5 at 13:53
Thanks! learned a lot while reading this. Will add this this evening!
– Mr.Sh4nnon
Dec 5 at 13:53
add a comment |
Mr.Sh4nnon is a new contributor. Be nice, and check out our Code of Conduct.
Mr.Sh4nnon is a new contributor. Be nice, and check out our Code of Conduct.
Mr.Sh4nnon is a new contributor. Be nice, and check out our Code of Conduct.
Mr.Sh4nnon is a new contributor. Be nice, and check out our Code of Conduct.
Thanks for contributing an answer to Code Review Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fcodereview.stackexchange.com%2fquestions%2f209049%2ftranslating-latex-documents-into-different-languages-using-deepl-api%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
1
Instead of sending the whole text in 1 go or line per line, you could try to find a middleground and send chunks of a few lines at a time
– Maarten Fabré
Dec 5 at 13:47
Thought about that too. The problem is I cant guarantee it will work correctly. It's a tradeoff.
– Mr.Sh4nnon
Dec 5 at 13:52