Prefer with context manager to open files
This commit is contained in:
parent
f093c8bd79
commit
5dc5615bfd
@ -234,9 +234,8 @@ class CLIDbManager(object):
|
||||
except:
|
||||
version = (0, 0, 0)
|
||||
if os.path.isfile(path_name):
|
||||
file = open(path_name, 'r', encoding='utf8')
|
||||
name = file.readline().strip()
|
||||
file.close()
|
||||
with open(path_name, 'r', encoding='utf8') as file:
|
||||
name = file.readline().strip()
|
||||
|
||||
(tval, last) = time_val(dirpath)
|
||||
(enable, stock_id) = self.icon_values(dirpath, self.active,
|
||||
@ -293,9 +292,8 @@ class CLIDbManager(object):
|
||||
name_list = [ name[0] for name in self.current_names ]
|
||||
title = find_next_db_name(name_list)
|
||||
|
||||
name_file = open(path_name, "w", encoding='utf8')
|
||||
name_file.write(title)
|
||||
name_file.close()
|
||||
with open(path_name, "w", encoding='utf8') as name_file:
|
||||
name_file.write(title)
|
||||
|
||||
if create_db:
|
||||
# write the version number into metadata
|
||||
@ -409,9 +407,8 @@ class CLIDbManager(object):
|
||||
dirpath = os.path.join(dbdir, dpath)
|
||||
path_name = os.path.join(dirpath, NAME_FILE)
|
||||
if os.path.isfile(path_name):
|
||||
file = open(path_name, 'r', encoding='utf8')
|
||||
name = file.readline().strip()
|
||||
file.close()
|
||||
with open(path_name, 'r', encoding='utf8') as file:
|
||||
name = file.readline().strip()
|
||||
if re.match("^" + dbname + "$", name):
|
||||
match_list.append((name, dirpath))
|
||||
if len(match_list) == 0:
|
||||
@ -438,12 +435,10 @@ class CLIDbManager(object):
|
||||
Returns old_name, new_name if success, None, None if no success
|
||||
"""
|
||||
try:
|
||||
name_file = open(filepath, "r", encoding='utf8')
|
||||
old_text=name_file.read()
|
||||
name_file.close()
|
||||
name_file = open(filepath, "w", encoding='utf8')
|
||||
name_file.write(new_text)
|
||||
name_file.close()
|
||||
with open(filepath, "r", encoding='utf8') as name_file:
|
||||
old_text=name_file.read()
|
||||
with open(filepath, "w", encoding='utf8') as name_file:
|
||||
name_file.write(new_text)
|
||||
except (OSError, IOError) as msg:
|
||||
CLIDbManager.ERROR(_("Could not rename Family Tree"),
|
||||
str(msg))
|
||||
@ -543,11 +538,10 @@ def find_locker_name(dirpath):
|
||||
"""
|
||||
try:
|
||||
fname = os.path.join(dirpath, "lock")
|
||||
ifile = open(fname, 'r', encoding='utf8')
|
||||
username = ifile.read().strip()
|
||||
# feature request 2356: avoid genitive form
|
||||
last = _("Locked by %s") % username
|
||||
ifile.close()
|
||||
with open(fname, 'r', encoding='utf8') as ifile:
|
||||
username = ifile.read().strip()
|
||||
# feature request 2356: avoid genitive form
|
||||
last = _("Locked by %s") % username
|
||||
except (OSError, IOError, UnicodeDecodeError):
|
||||
last = _("Unknown")
|
||||
return last
|
||||
|
@ -203,25 +203,22 @@ class DbState(Callback):
|
||||
dirpath = os.path.join(dbdir, dpath)
|
||||
path_name = os.path.join(dirpath, "name.txt")
|
||||
if os.path.isfile(path_name):
|
||||
file = open(path_name, 'r', encoding='utf8')
|
||||
name = file.readline().strip()
|
||||
file.close()
|
||||
with open(path_name, 'r', encoding='utf8') as file:
|
||||
name = file.readline().strip()
|
||||
if dbname == name:
|
||||
locked = False
|
||||
locked_by = None
|
||||
backend = None
|
||||
fname = os.path.join(dirpath, "database.txt")
|
||||
if os.path.isfile(fname):
|
||||
ifile = open(fname, 'r', encoding='utf8')
|
||||
backend = ifile.read().strip()
|
||||
ifile.close()
|
||||
with open(fname, 'r', encoding='utf8') as ifile:
|
||||
backend = ifile.read().strip()
|
||||
else:
|
||||
backend = "bsddb"
|
||||
try:
|
||||
fname = os.path.join(dirpath, "lock")
|
||||
ifile = open(fname, 'r', encoding='utf8')
|
||||
locked_by = ifile.read().strip()
|
||||
locked = True
|
||||
with open(fname, 'r', encoding='utf8') as ifile:
|
||||
locked_by = ifile.read().strip()
|
||||
ifile.close()
|
||||
except (OSError, IOError):
|
||||
pass
|
||||
|
@ -103,9 +103,8 @@ class FilterList(object):
|
||||
if os.path.isfile(self.file):
|
||||
parser = make_parser()
|
||||
parser.setContentHandler(FilterParser(self))
|
||||
the_file = open(self.file, 'r', encoding='utf8')
|
||||
parser.parse(the_file)
|
||||
the_file.close()
|
||||
with open(self.file, 'r', encoding='utf8') as the_file:
|
||||
parser.parse(the_file)
|
||||
except (IOError, OSError):
|
||||
print("IO/OSError in _filterlist.py")
|
||||
except SAXParseException:
|
||||
|
@ -603,9 +603,8 @@ class GVDotDoc(GVDocBase):
|
||||
if self._filename[-3:] != ".gv":
|
||||
self._filename += ".gv"
|
||||
|
||||
dotfile = open(self._filename, "wb")
|
||||
dotfile.write(self._dot.getvalue())
|
||||
dotfile.close()
|
||||
with open(self._filename, "wb") as dotfile:
|
||||
dotfile.write(self._dot.getvalue())
|
||||
|
||||
#-------------------------------------------------------------------------------
|
||||
#
|
||||
|
@ -146,31 +146,30 @@ class StyleSheetList(object):
|
||||
"""
|
||||
Saves the current StyleSheet definitions to the associated file.
|
||||
"""
|
||||
xml_file = open(self.__file, "w")
|
||||
xml_file.write('<?xml version="1.0" encoding="utf-8"?>\n')
|
||||
xml_file.write('<stylelist>\n')
|
||||
with open(self.__file, "w") as xml_file:
|
||||
xml_file.write('<?xml version="1.0" encoding="utf-8"?>\n')
|
||||
xml_file.write('<stylelist>\n')
|
||||
|
||||
for name in sorted(self.map.keys()): # enable diff of archived copies
|
||||
if name == "default":
|
||||
continue
|
||||
sheet = self.map[name]
|
||||
xml_file.write('<sheet name="%s">\n' % escxml(name))
|
||||
for name in sorted(self.map.keys()): # enable diff of archived copies
|
||||
if name == "default":
|
||||
continue
|
||||
sheet = self.map[name]
|
||||
xml_file.write('<sheet name="%s">\n' % escxml(name))
|
||||
|
||||
for p_name in sheet.get_paragraph_style_names():
|
||||
self.write_paragraph_style(xml_file, sheet, p_name)
|
||||
for p_name in sheet.get_paragraph_style_names():
|
||||
self.write_paragraph_style(xml_file, sheet, p_name)
|
||||
|
||||
for t_name in sheet.get_table_style_names():
|
||||
self.write_table_style(xml_file, sheet, t_name)
|
||||
for t_name in sheet.get_table_style_names():
|
||||
self.write_table_style(xml_file, sheet, t_name)
|
||||
|
||||
for c_name in sheet.get_cell_style_names():
|
||||
self.write_cell_style(xml_file, sheet, c_name)
|
||||
for c_name in sheet.get_cell_style_names():
|
||||
self.write_cell_style(xml_file, sheet, c_name)
|
||||
|
||||
for g_name in sheet.get_draw_style_names():
|
||||
self.write_graphics_style(xml_file, sheet, g_name)
|
||||
for g_name in sheet.get_draw_style_names():
|
||||
self.write_graphics_style(xml_file, sheet, g_name)
|
||||
|
||||
xml_file.write('</sheet>\n')
|
||||
xml_file.write('</stylelist>\n')
|
||||
xml_file.close()
|
||||
xml_file.write('</sheet>\n')
|
||||
xml_file.write('</stylelist>\n')
|
||||
|
||||
def write_paragraph_style(self, xml_file, sheet, p_name):
|
||||
|
||||
@ -275,9 +274,8 @@ class StyleSheetList(object):
|
||||
if os.path.isfile(self.__file):
|
||||
parser = make_parser()
|
||||
parser.setContentHandler(SheetParser(self))
|
||||
the_file = open(self.__file)
|
||||
parser.parse(the_file)
|
||||
the_file.close()
|
||||
with open(self.__file) as the_file:
|
||||
parser.parse(the_file)
|
||||
except (IOError, OSError, SAXParseException):
|
||||
pass
|
||||
|
||||
|
@ -458,68 +458,67 @@ class BookList(object):
|
||||
"""
|
||||
Saves the current BookList to the associated file.
|
||||
"""
|
||||
f = open(self.file, "w")
|
||||
f.write("<?xml version=\"1.0\" encoding=\"utf-8\"?>\n")
|
||||
f.write('<booklist>\n')
|
||||
for name in sorted(self.bookmap): # enable a diff of archived copies
|
||||
book = self.get_book(name)
|
||||
dbname = book.get_dbname()
|
||||
f.write(' <book name="%s" database="%s">\n' % (name, dbname) )
|
||||
for item in book.get_item_list():
|
||||
f.write(' <item name="%s" trans_name="%s">\n' %
|
||||
(item.get_name(), item.get_translated_name() ) )
|
||||
options = item.option_class.handler.options_dict
|
||||
for option_name in sorted(options.keys()): # enable a diff
|
||||
option_value = options[option_name]
|
||||
if isinstance(option_value, (list, tuple)):
|
||||
f.write(' <option name="%s" value="" '
|
||||
'length="%d">\n' % (
|
||||
escape(option_name),
|
||||
len(options[option_name]) ) )
|
||||
for list_index in range(len(option_value)):
|
||||
option_type = type_name(option_value[list_index])
|
||||
value = escape(str(option_value[list_index]))
|
||||
with open(self.file, "w") as f:
|
||||
f.write("<?xml version=\"1.0\" encoding=\"utf-8\"?>\n")
|
||||
f.write('<booklist>\n')
|
||||
for name in sorted(self.bookmap): # enable a diff of archived copies
|
||||
book = self.get_book(name)
|
||||
dbname = book.get_dbname()
|
||||
f.write(' <book name="%s" database="%s">\n' % (name, dbname) )
|
||||
for item in book.get_item_list():
|
||||
f.write(' <item name="%s" trans_name="%s">\n' %
|
||||
(item.get_name(), item.get_translated_name() ) )
|
||||
options = item.option_class.handler.options_dict
|
||||
for option_name in sorted(options.keys()): # enable a diff
|
||||
option_value = options[option_name]
|
||||
if isinstance(option_value, (list, tuple)):
|
||||
f.write(' <option name="%s" value="" '
|
||||
'length="%d">\n' % (
|
||||
escape(option_name),
|
||||
len(options[option_name]) ) )
|
||||
for list_index in range(len(option_value)):
|
||||
option_type = type_name(option_value[list_index])
|
||||
value = escape(str(option_value[list_index]))
|
||||
value = value.replace('"', '"')
|
||||
f.write(' <listitem number="%d" type="%s" '
|
||||
'value="%s"/>\n' % (
|
||||
list_index,
|
||||
option_type,
|
||||
value ) )
|
||||
f.write(' </option>\n')
|
||||
else:
|
||||
option_type = type_name(option_value)
|
||||
value = escape(str(option_value))
|
||||
value = value.replace('"', '"')
|
||||
f.write(' <listitem number="%d" type="%s" '
|
||||
f.write(' <option name="%s" type="%s" '
|
||||
'value="%s"/>\n' % (
|
||||
list_index,
|
||||
escape(option_name),
|
||||
option_type,
|
||||
value ) )
|
||||
f.write(' </option>\n')
|
||||
else:
|
||||
option_type = type_name(option_value)
|
||||
value = escape(str(option_value))
|
||||
value = value.replace('"', '"')
|
||||
f.write(' <option name="%s" type="%s" '
|
||||
'value="%s"/>\n' % (
|
||||
escape(option_name),
|
||||
option_type,
|
||||
value) )
|
||||
value) )
|
||||
|
||||
f.write(' <style name="%s"/>\n' % item.get_style_name() )
|
||||
f.write(' </item>\n')
|
||||
if book.get_paper_name():
|
||||
f.write(' <paper name="%s"/>\n' % book.get_paper_name() )
|
||||
if book.get_orientation() is not None: # 0 is legal
|
||||
f.write(' <orientation value="%s"/>\n' %
|
||||
book.get_orientation() )
|
||||
if book.get_paper_metric() is not None: # 0 is legal
|
||||
f.write(' <metric value="%s"/>\n' % book.get_paper_metric() )
|
||||
if book.get_custom_paper_size():
|
||||
size = book.get_custom_paper_size()
|
||||
f.write(' <size value="%f %f"/>\n' % (size[0], size[1]) )
|
||||
if book.get_margins():
|
||||
for pos in range(len(book.get_margins())):
|
||||
f.write(' <margin number="%s" value="%f"/>\n' %
|
||||
(pos, book.get_margin(pos)) )
|
||||
if book.get_format_name():
|
||||
f.write(' <format name="%s"/>\n' % book.get_format_name() )
|
||||
if book.get_output():
|
||||
f.write(' <output name="%s"/>\n' % book.get_output() )
|
||||
f.write(' </book>\n')
|
||||
f.write(' <style name="%s"/>\n' % item.get_style_name() )
|
||||
f.write(' </item>\n')
|
||||
if book.get_paper_name():
|
||||
f.write(' <paper name="%s"/>\n' % book.get_paper_name() )
|
||||
if book.get_orientation() is not None: # 0 is legal
|
||||
f.write(' <orientation value="%s"/>\n' %
|
||||
book.get_orientation() )
|
||||
if book.get_paper_metric() is not None: # 0 is legal
|
||||
f.write(' <metric value="%s"/>\n' % book.get_paper_metric() )
|
||||
if book.get_custom_paper_size():
|
||||
size = book.get_custom_paper_size()
|
||||
f.write(' <size value="%f %f"/>\n' % (size[0], size[1]) )
|
||||
if book.get_margins():
|
||||
for pos in range(len(book.get_margins())):
|
||||
f.write(' <margin number="%s" value="%f"/>\n' %
|
||||
(pos, book.get_margin(pos)) )
|
||||
if book.get_format_name():
|
||||
f.write(' <format name="%s"/>\n' % book.get_format_name() )
|
||||
if book.get_output():
|
||||
f.write(' <output name="%s"/>\n' % book.get_output() )
|
||||
f.write(' </book>\n')
|
||||
|
||||
f.write('</booklist>\n')
|
||||
f.close()
|
||||
f.write('</booklist>\n')
|
||||
|
||||
def parse(self):
|
||||
"""
|
||||
|
@ -504,9 +504,8 @@ class OptionListCollection(_options.OptionListCollection):
|
||||
if os.path.isfile(self.filename):
|
||||
p = make_parser()
|
||||
p.setContentHandler(OptionParser(self))
|
||||
the_file = open(self.filename, encoding="utf-8")
|
||||
p.parse(the_file)
|
||||
the_file.close()
|
||||
with open(self.filename, encoding="utf-8") as the_file:
|
||||
p.parse(the_file)
|
||||
except (IOError, OSError, SAXParseException):
|
||||
pass
|
||||
|
||||
@ -1000,9 +999,8 @@ class DocOptionListCollection(_options.OptionListCollection):
|
||||
if os.path.isfile(self.filename):
|
||||
p = make_parser()
|
||||
p.setContentHandler(DocOptionParser(self))
|
||||
the_file = open(self.filename, encoding="utf-8")
|
||||
p.parse(the_file)
|
||||
the_file.close()
|
||||
with open(self.filename, encoding="utf-8") as the_file:
|
||||
p.parse(the_file)
|
||||
except (IOError, OSError, SAXParseException):
|
||||
pass
|
||||
|
||||
|
@ -279,9 +279,8 @@ def resize_to_jpeg_buffer(source, size, crop=None):
|
||||
scaled = img.scale_simple(int(size[0]), int(size[1]), GdkPixbuf.InterpType.BILINEAR)
|
||||
os.close(filed)
|
||||
scaled.savev(dest, "jpeg", "", "")
|
||||
ofile = open(dest, mode='rb')
|
||||
data = ofile.read()
|
||||
ofile.close()
|
||||
with open(dest, mode='rb') as ofile:
|
||||
data = ofile.read()
|
||||
try:
|
||||
os.unlink(dest)
|
||||
except:
|
||||
|
@ -436,8 +436,7 @@ class DbBsddb(DbBsddbRead, DbWriteBase, UpdateCallback):
|
||||
|
||||
def __log_error(self):
|
||||
mypath = os.path.join(self.get_save_path(),DBRECOVFN)
|
||||
ofile = open(mypath, "w")
|
||||
ofile.close()
|
||||
with open(mypath, "w") as ofile:
|
||||
try:
|
||||
clear_lock_file(self.get_save_path())
|
||||
except:
|
||||
|
@ -285,10 +285,8 @@ class HtmlDoc(BaseDoc, TextDoc):
|
||||
Copy support files to the datadir that needs to hold them
|
||||
"""
|
||||
#css of textdoc styles
|
||||
tdfile = open(os.path.join(self._backend.datadirfull(),
|
||||
_TEXTDOCSCREEN), 'w')
|
||||
tdfile.write(self.style_declaration)
|
||||
tdfile.close()
|
||||
with open(os.path.join(self._backend.datadirfull(), _TEXTDOCSCREEN), 'w') as tdfile:
|
||||
tdfile.write(self.style_declaration)
|
||||
#css file
|
||||
if self.css_filename:
|
||||
#we do an extra check in case file does not exist, eg cli call
|
||||
|
@ -30,12 +30,11 @@ from test import test_util
|
||||
test_util.path_append_parent()
|
||||
|
||||
def get_potfile(filename):
|
||||
fp = open(filename, "r")
|
||||
retvals = []
|
||||
for line in fp:
|
||||
if line and line[0] != "#":
|
||||
retvals.append(line.strip())
|
||||
fp.close()
|
||||
with open(filename, "r") as fp:
|
||||
retvals = []
|
||||
for line in fp:
|
||||
if line and line[0] != "#":
|
||||
retvals.append(line.strip())
|
||||
return retvals
|
||||
|
||||
# POTFILES.skip
|
||||
@ -55,9 +54,8 @@ class TestPOT(unittest.TestCase):
|
||||
realpath = (dir + "/" + file)
|
||||
pathfile = realpath[3:]
|
||||
if os.path.exists(realpath):
|
||||
fp = open(realpath, "r")
|
||||
lines = fp.read()
|
||||
fp.close()
|
||||
with open(realpath, "r") as fp:
|
||||
lines = fp.read()
|
||||
found = False
|
||||
for search in searches:
|
||||
if search in lines:
|
||||
@ -88,9 +86,8 @@ class TestMake(unittest.TestCase):
|
||||
if pathfile[3:] in excluded_files:
|
||||
self.assertTrue(True, "exclude '%s'" % pathfile)
|
||||
elif os.path.exists(makefile):
|
||||
fp = open(makefile, "r")
|
||||
lines = fp.read()
|
||||
fp.close()
|
||||
with open(makefile, "r") as fp:
|
||||
lines = fp.read()
|
||||
self.assertTrue(filename in lines, "'%s' not in %s/Makefile.in" %
|
||||
(filename, path))
|
||||
else:
|
||||
@ -107,9 +104,8 @@ class TestGetText(unittest.TestCase):
|
||||
def helper(self, pofile, searches):
|
||||
if not os.path.exists("../../" + pofile):
|
||||
self.assertTrue(False, "'%s' is in POTFILES.in and does not exist" % pofile)
|
||||
fp = open("../../" + pofile, "r")
|
||||
lines = fp.read()
|
||||
fp.close()
|
||||
with open("../../" + pofile, "r") as fp:
|
||||
lines = fp.read()
|
||||
found = False
|
||||
for search in searches:
|
||||
found = (search in lines) or found
|
||||
|
228
po/update_po.py
228
po/update_po.py
@ -166,23 +166,22 @@ def TipsParse(filename, mark):
|
||||
"Editor."
|
||||
'''
|
||||
|
||||
tips = open('../data/tips.xml.in.h', 'w')
|
||||
marklist = root.iter(mark)
|
||||
for key in marklist:
|
||||
tip = ElementTree.tostring(key, encoding="UTF-8", method="xml")
|
||||
if sys.version_info[0] < 3:
|
||||
tip = tip.replace("<?xml version='1.0' encoding='UTF-8'?>", "")
|
||||
tip = tip.replace('\n<_tip number="%(number)s">' % key.attrib, "")
|
||||
else: # python3 support
|
||||
tip = tip.decode("utf-8")
|
||||
tip = tip.replace('<_tip number="%(number)s">' % key.attrib, "")
|
||||
tip = tip.replace("<br />", "<br/>")
|
||||
#tip = tip.replace("\n</_tip>\n", "</_tip>\n") # special case tip 7
|
||||
#tip = tip.replace("\n<b>", "<b>") # special case tip 18
|
||||
tip = tip.replace("</_tip>\n\n", "")
|
||||
tip = tip.replace('"', '"')
|
||||
tips.write('char *s = N_("%s");\n' % tip)
|
||||
tips.close()
|
||||
with open('../data/tips.xml.in.h', 'w') as tips:
|
||||
marklist = root.iter(mark)
|
||||
for key in marklist:
|
||||
tip = ElementTree.tostring(key, encoding="UTF-8", method="xml")
|
||||
if sys.version_info[0] < 3:
|
||||
tip = tip.replace("<?xml version='1.0' encoding='UTF-8'?>", "")
|
||||
tip = tip.replace('\n<_tip number="%(number)s">' % key.attrib, "")
|
||||
else: # python3 support
|
||||
tip = tip.decode("utf-8")
|
||||
tip = tip.replace('<_tip number="%(number)s">' % key.attrib, "")
|
||||
tip = tip.replace("<br />", "<br/>")
|
||||
#tip = tip.replace("\n</_tip>\n", "</_tip>\n") # special case tip 7
|
||||
#tip = tip.replace("\n<b>", "<b>") # special case tip 18
|
||||
tip = tip.replace("</_tip>\n\n", "")
|
||||
tip = tip.replace('"', '"')
|
||||
tips.write('char *s = N_("%s");\n' % tip)
|
||||
print ('Wrote ../data/tips.xml.in.h')
|
||||
root.clear()
|
||||
|
||||
@ -215,15 +214,14 @@ def HolidaysParse(filename, mark):
|
||||
msgid "Yom Kippur"
|
||||
'''
|
||||
|
||||
holidays = open('../data/holidays.xml.in.h', 'w')
|
||||
for key in ellist:
|
||||
if key.attrib.get(mark):
|
||||
line = key.attrib
|
||||
string = line.items
|
||||
# mapping via the line dict (_name is the key)
|
||||
name = 'char *s = N_("%(_name)s");\n' % line
|
||||
holidays.write(name)
|
||||
holidays.close()
|
||||
with open('../data/holidays.xml.in.h', 'w') as holidays:
|
||||
for key in ellist:
|
||||
if key.attrib.get(mark):
|
||||
line = key.attrib
|
||||
string = line.items
|
||||
# mapping via the line dict (_name is the key)
|
||||
name = 'char *s = N_("%(_name)s");\n' % line
|
||||
holidays.write(name)
|
||||
print ('Wrote ../data/holidays.xml.in.h')
|
||||
root.clear()
|
||||
|
||||
@ -262,20 +260,19 @@ def XmlParse(filename, mark):
|
||||
</p>
|
||||
'''
|
||||
|
||||
head = open(filename + '.h', 'w')
|
||||
|
||||
for key in root.iter():
|
||||
if key.tag == '{http://www.freedesktop.org/standards/shared-mime-info}%s' % mark:
|
||||
comment = 'char *s = N_("%s");\n' % key.text
|
||||
head.write(comment)
|
||||
|
||||
if root.tag == 'application':
|
||||
with open(filename + '.h', 'w') as head:
|
||||
|
||||
for key in root.iter():
|
||||
if key.tag == mark:
|
||||
if key.tag == '{http://www.freedesktop.org/standards/shared-mime-info}%s' % mark:
|
||||
comment = 'char *s = N_("%s");\n' % key.text
|
||||
head.write(comment)
|
||||
|
||||
head.close()
|
||||
|
||||
if root.tag == 'application':
|
||||
for key in root.iter():
|
||||
if key.tag == mark:
|
||||
comment = 'char *s = N_("%s");\n' % key.text
|
||||
head.write(comment)
|
||||
|
||||
print ('Wrote %s' % filename)
|
||||
root.clear()
|
||||
|
||||
@ -301,22 +298,19 @@ def DesktopParse(filename):
|
||||
perform genealogical research and analysis"
|
||||
'''
|
||||
|
||||
desktop = open('../data/gramps.desktop.in.h', 'w')
|
||||
|
||||
f = open(filename)
|
||||
lines = [file.strip() for file in f]
|
||||
f.close()
|
||||
|
||||
for line in lines:
|
||||
if line[0] == '_':
|
||||
for i in range(len(line)):
|
||||
if line[i] == '=':
|
||||
val = 'char *s = N_("%s");\n' % line[i+1:len(line)]
|
||||
desktop.write(val)
|
||||
|
||||
desktop.close()
|
||||
with open('../data/gramps.desktop.in.h', 'w') as desktop:
|
||||
|
||||
with open(filename) as f:
|
||||
lines = [file.strip() for file in f]
|
||||
|
||||
for line in lines:
|
||||
if line[0] == '_':
|
||||
for i in range(len(line)):
|
||||
if line[i] == '=':
|
||||
val = 'char *s = N_("%s");\n' % line[i+1:len(line)]
|
||||
desktop.write(val)
|
||||
|
||||
print ('Wrote ../data/gramps.desktop.in.h')
|
||||
|
||||
|
||||
def KeyParse(filename, mark):
|
||||
"""
|
||||
@ -342,29 +336,26 @@ def KeyParse(filename, mark):
|
||||
msgid "Gramps XML database"
|
||||
msgid "GEDCOM"
|
||||
'''
|
||||
|
||||
key = open('../data/gramps.keys.in.h', 'w')
|
||||
|
||||
f = open(filename)
|
||||
lines = [file for file in f]
|
||||
f.close()
|
||||
|
||||
temp = []
|
||||
|
||||
for line in lines:
|
||||
for i in range(len(line)):
|
||||
if line[i:i+12] == mark:
|
||||
temp.append(line.strip())
|
||||
|
||||
for t in temp:
|
||||
for i in range(len(t)):
|
||||
if t[i] == '=':
|
||||
val = 'char *s = N_("%s");\n' % t[i+1:len(t)]
|
||||
key.write(val)
|
||||
|
||||
key.close()
|
||||
|
||||
with open('../data/gramps.keys.in.h', 'w') as key:
|
||||
|
||||
with open(filename) as f:
|
||||
lines = [file for file in f]
|
||||
|
||||
temp = []
|
||||
|
||||
for line in lines:
|
||||
for i in range(len(line)):
|
||||
if line[i:i+12] == mark:
|
||||
temp.append(line.strip())
|
||||
|
||||
for t in temp:
|
||||
for i in range(len(t)):
|
||||
if t[i] == '=':
|
||||
val = 'char *s = N_("%s");\n' % t[i+1:len(t)]
|
||||
key.write(val)
|
||||
|
||||
print ('Wrote ../data/gramps.keys.in.h')
|
||||
|
||||
|
||||
def main():
|
||||
"""
|
||||
@ -467,15 +458,14 @@ def create_filesfile():
|
||||
dir = os.getcwd()
|
||||
topdir = os.path.normpath(os.path.join(dir, '..', 'gramps'))
|
||||
lentopdir = len(topdir)
|
||||
f = open('POTFILES.in')
|
||||
infiles = dict(['../' + file.strip(), None] for file in f if file.strip()
|
||||
and not file[0]=='#')
|
||||
f.close()
|
||||
f = open('POTFILES.skip')
|
||||
notinfiles = dict(['../' + file.strip(), None] for file in f if file
|
||||
and not file[0]=='#')
|
||||
f.close()
|
||||
|
||||
with open('POTFILES.in') as f:
|
||||
infiles = dict(['../' + file.strip(), None] for file in f if file.strip()
|
||||
and not file[0]=='#')
|
||||
|
||||
with open('POTFILES.skip') as f:
|
||||
notinfiles = dict(['../' + file.strip(), None] for file in f if file
|
||||
and not file[0]=='#')
|
||||
|
||||
for (dirpath, dirnames, filenames) in os.walk(topdir):
|
||||
root, subdir = os.path.split(dirpath)
|
||||
if subdir.startswith("."):
|
||||
@ -499,11 +489,10 @@ def create_filesfile():
|
||||
if full_filename[lentopdir:] in notinfiles:
|
||||
infiles['../gramps' + full_filename[lentopdir:]] = None
|
||||
#now we write out all the files in form ../gramps/filename
|
||||
f = open('tmpfiles', 'w')
|
||||
for file in sorted(infiles.keys()):
|
||||
f.write(file)
|
||||
f.write('\n')
|
||||
f.close()
|
||||
with open('tmpfiles', 'w') as f:
|
||||
for file in sorted(infiles.keys()):
|
||||
f.write(file)
|
||||
f.write('\n')
|
||||
|
||||
def listing(name, extensionlist):
|
||||
"""
|
||||
@ -511,22 +500,19 @@ def listing(name, extensionlist):
|
||||
Parsing from a textual file (gramps) is faster and easy for maintenance.
|
||||
Like POTFILES.in and POTFILES.skip
|
||||
"""
|
||||
|
||||
f = open('tmpfiles')
|
||||
files = [file.strip() for file in f if file and not file[0]=='#']
|
||||
f.close()
|
||||
|
||||
temp = open(name, 'w')
|
||||
|
||||
for entry in files:
|
||||
for ext in extensionlist:
|
||||
if entry.endswith(ext):
|
||||
temp.write(entry)
|
||||
temp.write('\n')
|
||||
break
|
||||
|
||||
temp.close()
|
||||
|
||||
|
||||
with open('tmpfiles') as f:
|
||||
files = [file.strip() for file in f if file and not file[0]=='#']
|
||||
|
||||
with open(name, 'w') as temp:
|
||||
|
||||
for entry in files:
|
||||
for ext in extensionlist:
|
||||
if entry.endswith(ext):
|
||||
temp.write(entry)
|
||||
temp.write('\n')
|
||||
break
|
||||
|
||||
def headers():
|
||||
"""
|
||||
Look at existing C file format headers.
|
||||
@ -565,14 +551,13 @@ def extract_xml():
|
||||
XmlParse('../data/gramps.appdata.xml.in', '_p')
|
||||
DesktopParse('../data/gramps.desktop.in')
|
||||
KeyParse('../data/gramps.keys.in', '_description')
|
||||
|
||||
|
||||
def create_template():
|
||||
"""
|
||||
Create a new file for template, if it does not exist.
|
||||
"""
|
||||
template = open('gramps.pot', 'w')
|
||||
template.close()
|
||||
|
||||
with open('gramps.pot', 'w') as template:
|
||||
|
||||
def extract_glade():
|
||||
"""
|
||||
Extract messages from a temp file with all .glade
|
||||
@ -622,21 +607,20 @@ def extract_gtkbuilder():
|
||||
msgid "At least one rule must apply"
|
||||
msgid "Exactly one rule must apply"
|
||||
'''
|
||||
|
||||
|
||||
files = ['../gramps/plugins/importer/importgedcom.glade', '../gramps/gui/glade/rule.glade']
|
||||
temp = open('gtklist.h', 'w')
|
||||
|
||||
for filename in files:
|
||||
tree = ElementTree.parse(filename)
|
||||
root = tree.getroot()
|
||||
for line in root.iter():
|
||||
att = line.attrib
|
||||
if att == {'id': '0', 'translatable': 'yes'}:
|
||||
col = 'char *s = N_("%s");\n' % line.text
|
||||
temp.write(col)
|
||||
root.clear()
|
||||
|
||||
temp.close()
|
||||
with open('gtklist.h', 'w') as temp:
|
||||
|
||||
for filename in files:
|
||||
tree = ElementTree.parse(filename)
|
||||
root = tree.getroot()
|
||||
for line in root.iter():
|
||||
att = line.attrib
|
||||
if att == {'id': '0', 'translatable': 'yes'}:
|
||||
col = 'char *s = N_("%s");\n' % line.text
|
||||
temp.write(col)
|
||||
root.clear()
|
||||
|
||||
print ('Wrote gtklist.h')
|
||||
|
||||
def retrieve():
|
||||
|
Loading…
x
Reference in New Issue
Block a user