2007-06-15 Don Allingham <don@gramps-project.org>
* src/GrampsDbUtils/_Backup.py: Optimization and documentation. svn: r8557
This commit is contained in:
parent
b5f51cf15f
commit
e9860079c3
@ -1,3 +1,6 @@
|
|||||||
|
2007-06-15 Don Allingham <don@gramps-project.org>
|
||||||
|
* src/GrampsDbUtils/_Backup.py: Optimization and documentation.
|
||||||
|
|
||||||
2007-06-14 Alex Roitman <shura@gramps-project.org>
|
2007-06-14 Alex Roitman <shura@gramps-project.org>
|
||||||
* src/GrampsDbUtils/_Backup.py (__do_restore): Minor optimization.
|
* src/GrampsDbUtils/_Backup.py (__do_restore): Minor optimization.
|
||||||
* src/GrampsDb/_GrampsDbBase.py (__get_from_handle): Pass txn.
|
* src/GrampsDb/_GrampsDbBase.py (__get_from_handle): Pass txn.
|
||||||
|
@ -19,7 +19,30 @@
|
|||||||
#
|
#
|
||||||
|
|
||||||
"""
|
"""
|
||||||
Provides backup and restore functions for a database
|
Description
|
||||||
|
===========
|
||||||
|
|
||||||
|
This module Provides backup and restore functions for a database. The
|
||||||
|
backup function saves the data into backup files, while the restore
|
||||||
|
function loads the data back into a database.
|
||||||
|
|
||||||
|
You should only restore the data into an empty database.
|
||||||
|
|
||||||
|
Implementation
|
||||||
|
==============
|
||||||
|
|
||||||
|
Not all of the database tables need to be backed up, since many are
|
||||||
|
automatically generated from the others. The tables that are backed up
|
||||||
|
are the primary tables and the metadata table.
|
||||||
|
|
||||||
|
The database consists of a table of "pickled" tuples. Each of the
|
||||||
|
primary tables is "walked", and the pickled tuple is extracted, and
|
||||||
|
written to the backup file.
|
||||||
|
|
||||||
|
Restoring the data is just as simple. The backup file is parsed an
|
||||||
|
entry at a time, and inserted into the associated database table. The
|
||||||
|
derived tables are built automatically as the items are entered into
|
||||||
|
db.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
#-------------------------------------------------------------------------
|
#-------------------------------------------------------------------------
|
||||||
@ -48,7 +71,7 @@ import cPickle as pickle
|
|||||||
|
|
||||||
LOG = logging.getLogger(".Backup")
|
LOG = logging.getLogger(".Backup")
|
||||||
|
|
||||||
def export(database):
|
def backup(database):
|
||||||
"""
|
"""
|
||||||
Exports the database to a set of backup files. These files consist
|
Exports the database to a set of backup files. These files consist
|
||||||
of the pickled database tables, one file for each table.
|
of the pickled database tables, one file for each table.
|
||||||
@ -64,6 +87,28 @@ def export(database):
|
|||||||
except (OSError, IOError), msg:
|
except (OSError, IOError), msg:
|
||||||
ErrorDialog(_("Error saving backup data"), str(msg))
|
ErrorDialog(_("Error saving backup data"), str(msg))
|
||||||
|
|
||||||
|
def __mk_backup_name(database, base):
|
||||||
|
"""
|
||||||
|
Returns the backup name of the database table
|
||||||
|
|
||||||
|
@param database: database instance
|
||||||
|
@type database: GrampsDbDir
|
||||||
|
@param base: base name of the table
|
||||||
|
@type base: str
|
||||||
|
"""
|
||||||
|
return os.path.join(database.get_save_path(), base + ".gbkp")
|
||||||
|
|
||||||
|
def __mk_tmp_name(database, base):
|
||||||
|
"""
|
||||||
|
Returns the temporary backup name of the database table
|
||||||
|
|
||||||
|
@param database: database instance
|
||||||
|
@type database: GrampsDbDir
|
||||||
|
@param base: base name of the table
|
||||||
|
@type base: str
|
||||||
|
"""
|
||||||
|
return os.path.join(database.get_save_path(), base + ".gbkp.new")
|
||||||
|
|
||||||
def __do_export(database):
|
def __do_export(database):
|
||||||
"""
|
"""
|
||||||
Loop through each table of the database, saving the pickled data
|
Loop through each table of the database, saving the pickled data
|
||||||
@ -74,7 +119,7 @@ def __do_export(database):
|
|||||||
"""
|
"""
|
||||||
try:
|
try:
|
||||||
for (base, tbl) in __build_tbl_map(database):
|
for (base, tbl) in __build_tbl_map(database):
|
||||||
backup_name = os.path.join(database.get_save_path(), base + ".gbkp.new")
|
backup_name = __mk_tmp_name(database, base)
|
||||||
backup_table = open(backup_name, 'wb')
|
backup_table = open(backup_name, 'wb')
|
||||||
|
|
||||||
cursor = tbl.cursor()
|
cursor = tbl.cursor()
|
||||||
@ -88,8 +133,8 @@ def __do_export(database):
|
|||||||
return
|
return
|
||||||
|
|
||||||
for (base, tbl) in __build_tbl_map(database):
|
for (base, tbl) in __build_tbl_map(database):
|
||||||
new_name = os.path.join(database.get_save_path(), base + ".gbkp")
|
new_name = __mk_backup_name(database, base)
|
||||||
old_name = new_name + ".new"
|
old_name = __mk_tmp_name(database, base)
|
||||||
if os.path.isfile(new_name):
|
if os.path.isfile(new_name):
|
||||||
os.unlink(new_name)
|
os.unlink(new_name)
|
||||||
os.rename(old_name, new_name)
|
os.rename(old_name, new_name)
|
||||||
@ -119,26 +164,53 @@ def __do_restore(database):
|
|||||||
@type database: GrampsDbDir
|
@type database: GrampsDbDir
|
||||||
"""
|
"""
|
||||||
for (base, tbl) in __build_tbl_map(database):
|
for (base, tbl) in __build_tbl_map(database):
|
||||||
backup_name = os.path.join(database.get_save_path(), base + ".gbkp")
|
backup_name = __mk_backup_name(database, base)
|
||||||
backup_table = open(backup_name, 'rb')
|
backup_table = open(backup_name, 'rb')
|
||||||
|
|
||||||
|
if database.UseTXN:
|
||||||
|
__load_tbl_txn(database, backup_table, tbl)
|
||||||
|
else:
|
||||||
|
__load_tbl_no_txn(backup_table, tbl)
|
||||||
|
|
||||||
|
database.rebuild_secondary()
|
||||||
|
|
||||||
|
def __load_tbl_no_txn(backup_table, tbl):
|
||||||
|
"""
|
||||||
|
Returns the temporary backup name of the database table
|
||||||
|
|
||||||
|
@param backup_table: file containing the backup data
|
||||||
|
@type backup_table: file
|
||||||
|
@param tbl: Berkeley db database table
|
||||||
|
@type tbl: Berkeley db database table
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
while True:
|
||||||
|
data = pickle.load(backup_table)
|
||||||
|
tbl.put(data[0], data[1], txn=None)
|
||||||
|
except EOFError:
|
||||||
|
tbl.sync()
|
||||||
|
backup_table.close()
|
||||||
|
|
||||||
|
def __load_tbl_txn(database, backup_table, tbl):
|
||||||
|
"""
|
||||||
|
Returns the temporary backup name of the database table
|
||||||
|
|
||||||
|
@param database: database instance
|
||||||
|
@type database: GrampsDbDir
|
||||||
|
@param backup_table: file containing the backup data
|
||||||
|
@type backup_table: file
|
||||||
|
@param tbl: Berkeley db database table
|
||||||
|
@type tbl: Berkeley db database table
|
||||||
|
"""
|
||||||
try:
|
try:
|
||||||
while True:
|
while True:
|
||||||
data = pickle.load(backup_table)
|
data = pickle.load(backup_table)
|
||||||
if database.UseTXN:
|
|
||||||
txn = database.env.txn_begin()
|
txn = database.env.txn_begin()
|
||||||
tbl.put(data[0], data[1], txn=txn)
|
tbl.put(data[0], data[1], txn=txn)
|
||||||
txn.commit()
|
txn.commit()
|
||||||
else:
|
|
||||||
tbl.put(data[0], data[1], txn=None)
|
|
||||||
except EOFError:
|
except EOFError:
|
||||||
if not database.UseTXN:
|
|
||||||
tbl.sync()
|
|
||||||
|
|
||||||
backup_table.close()
|
backup_table.close()
|
||||||
|
|
||||||
database.rebuild_secondary()
|
|
||||||
|
|
||||||
def __build_tbl_map(database):
|
def __build_tbl_map(database):
|
||||||
"""
|
"""
|
||||||
Builds a table map of names to database tables.
|
Builds a table map of names to database tables.
|
||||||
|
@ -533,7 +533,7 @@ class ViewManager:
|
|||||||
self.uistate.set_busy_cursor(1)
|
self.uistate.set_busy_cursor(1)
|
||||||
self.uistate.progress.show()
|
self.uistate.progress.show()
|
||||||
self.uistate.push_message(self.state, _("Autobackup..."))
|
self.uistate.push_message(self.state, _("Autobackup..."))
|
||||||
GrampsDbUtils.Backup.export(self.state.db)
|
GrampsDbUtils.Backup.backup(self.state.db)
|
||||||
self.uistate.set_busy_cursor(0)
|
self.uistate.set_busy_cursor(0)
|
||||||
self.uistate.progress.hide()
|
self.uistate.progress.hide()
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user