Imported Upstream version 4.6.2
This commit is contained in:
875
ipaserver/install/ipa_restore.py
Normal file
875
ipaserver/install/ipa_restore.py
Normal file
@@ -0,0 +1,875 @@
|
||||
# Authors: Rob Crittenden <rcritten@redhat.com>
|
||||
#
|
||||
# Copyright (C) 2013 Red Hat
|
||||
# see file 'COPYING' for use and warranty information
|
||||
#
|
||||
# This program is free software; you can redistribute it and/or modify
|
||||
# it under the terms of the GNU General Public License as published by
|
||||
# the Free Software Foundation, either version 3 of the License, or
|
||||
# (at your option) any later version.
|
||||
#
|
||||
# This program is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||
# GNU General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU General Public License
|
||||
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
#
|
||||
|
||||
import logging
|
||||
import os
|
||||
import shutil
|
||||
import tempfile
|
||||
import time
|
||||
import pwd
|
||||
import ldif
|
||||
import itertools
|
||||
|
||||
import six
|
||||
# pylint: disable=import-error
|
||||
if six.PY3:
|
||||
# The SafeConfigParser class has been renamed to ConfigParser in Py3
|
||||
from configparser import ConfigParser as SafeConfigParser
|
||||
else:
|
||||
from ConfigParser import SafeConfigParser
|
||||
# pylint: enable=import-error
|
||||
|
||||
from ipaclient.install.client import update_ipa_nssdb
|
||||
from ipalib import api, errors
|
||||
from ipalib.constants import FQDN
|
||||
from ipapython import version, ipautil
|
||||
from ipapython.ipautil import run, user_input
|
||||
from ipapython import admintool
|
||||
from ipapython.dn import DN
|
||||
from ipaserver.install.replication import (wait_for_task, ReplicationManager,
|
||||
get_cs_replication_manager)
|
||||
from ipaserver.install import installutils
|
||||
from ipaserver.install import dsinstance, httpinstance, cainstance, krbinstance
|
||||
from ipapython import ipaldap
|
||||
import ipapython.errors
|
||||
from ipaplatform.constants import constants
|
||||
from ipaplatform.tasks import tasks
|
||||
from ipaplatform import services
|
||||
from ipaplatform.paths import paths
|
||||
|
||||
try:
|
||||
from ipaserver.install import adtrustinstance
|
||||
except ImportError:
|
||||
adtrustinstance = None
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def recursive_chown(path, uid, gid):
|
||||
'''
|
||||
Change ownership of all files and directories in a path.
|
||||
'''
|
||||
for root, dirs, files in os.walk(path):
|
||||
for dir in dirs:
|
||||
os.chown(os.path.join(root, dir), uid, gid)
|
||||
os.chmod(os.path.join(root, dir), 0o750)
|
||||
for file in files:
|
||||
os.chown(os.path.join(root, file), uid, gid)
|
||||
os.chmod(os.path.join(root, file), 0o640)
|
||||
|
||||
|
||||
def decrypt_file(tmpdir, filename, keyring):
|
||||
source = filename
|
||||
(dest, ext) = os.path.splitext(filename)
|
||||
|
||||
if ext != '.gpg':
|
||||
raise admintool.ScriptError('Trying to decrypt a non-gpg file')
|
||||
|
||||
dest = os.path.basename(dest)
|
||||
dest = os.path.join(tmpdir, dest)
|
||||
|
||||
args = [paths.GPG,
|
||||
'--batch',
|
||||
'-o', dest]
|
||||
|
||||
if keyring is not None:
|
||||
args.append('--no-default-keyring')
|
||||
args.append('--keyring')
|
||||
args.append(keyring + '.pub')
|
||||
args.append('--secret-keyring')
|
||||
args.append(keyring + '.sec')
|
||||
|
||||
args.append('-d')
|
||||
args.append(source)
|
||||
|
||||
result = run(args, raiseonerr=False)
|
||||
if result.returncode != 0:
|
||||
raise admintool.ScriptError('gpg failed: %s' % result.error_log)
|
||||
|
||||
return dest
|
||||
|
||||
|
||||
class RemoveRUVParser(ldif.LDIFParser):
|
||||
def __init__(self, input_file, writer):
|
||||
ldif.LDIFParser.__init__(self, input_file)
|
||||
self.writer = writer
|
||||
|
||||
def handle(self, dn, entry):
|
||||
objectclass = None
|
||||
nsuniqueid = None
|
||||
|
||||
for name, value in entry.items():
|
||||
name = name.lower()
|
||||
if name == 'objectclass':
|
||||
objectclass = [x.lower() for x in value]
|
||||
elif name == 'nsuniqueid':
|
||||
nsuniqueid = [x.lower() for x in value]
|
||||
|
||||
if (
|
||||
objectclass and nsuniqueid and
|
||||
b'nstombstone' in objectclass and
|
||||
b'ffffffff-ffffffff-ffffffff-ffffffff' in nsuniqueid
|
||||
):
|
||||
logger.debug("Removing RUV entry %s", dn)
|
||||
return
|
||||
|
||||
self.writer.unparse(dn, entry)
|
||||
|
||||
|
||||
class Restore(admintool.AdminTool):
|
||||
command_name = 'ipa-restore'
|
||||
log_file_name = paths.IPARESTORE_LOG
|
||||
|
||||
usage = "%prog [options] backup"
|
||||
|
||||
description = "Restore IPA files and databases."
|
||||
|
||||
# directories and files listed here will be removed from filesystem before
|
||||
# files from backup are copied
|
||||
DIRS_TO_BE_REMOVED = [
|
||||
paths.DNSSEC_TOKENS_DIR,
|
||||
]
|
||||
|
||||
FILES_TO_BE_REMOVED = []
|
||||
|
||||
def __init__(self, options, args):
|
||||
super(Restore, self).__init__(options, args)
|
||||
self._conn = None
|
||||
|
||||
@classmethod
|
||||
def add_options(cls, parser):
|
||||
super(Restore, cls).add_options(parser, debug_option=True)
|
||||
|
||||
parser.add_option("-p", "--password", dest="password",
|
||||
help="Directory Manager password")
|
||||
parser.add_option("--gpg-keyring", dest="gpg_keyring",
|
||||
help="The gpg key name to be used")
|
||||
parser.add_option("--data", dest="data_only", action="store_true",
|
||||
default=False, help="Restore only the data")
|
||||
parser.add_option("--online", dest="online", action="store_true",
|
||||
default=False, help="Perform the LDAP restores online, for data only.")
|
||||
parser.add_option("--instance", dest="instance",
|
||||
help="The 389-ds instance to restore (defaults to all found)")
|
||||
parser.add_option("--backend", dest="backend",
|
||||
help="The backend to restore within the instance or instances")
|
||||
parser.add_option('--no-logs', dest="no_logs", action="store_true",
|
||||
default=False, help="Do not restore log files from the backup")
|
||||
parser.add_option('-U', '--unattended', dest="unattended",
|
||||
action="store_true", default=False,
|
||||
help="Unattended restoration never prompts the user")
|
||||
|
||||
|
||||
def setup_logging(self, log_file_mode='a'):
|
||||
super(Restore, self).setup_logging(log_file_mode='a')
|
||||
|
||||
|
||||
def validate_options(self):
|
||||
parser = self.option_parser
|
||||
options = self.options
|
||||
super(Restore, self).validate_options(needs_root=True)
|
||||
|
||||
if len(self.args) < 1:
|
||||
parser.error("must provide the backup to restore")
|
||||
elif len(self.args) > 1:
|
||||
parser.error("must provide exactly one name for the backup")
|
||||
|
||||
dirname = self.args[0]
|
||||
if not os.path.isabs(dirname):
|
||||
dirname = os.path.join(paths.IPA_BACKUP_DIR, dirname)
|
||||
if not os.path.isdir(dirname):
|
||||
parser.error("must provide path to backup directory")
|
||||
|
||||
if options.gpg_keyring:
|
||||
if (not os.path.exists(options.gpg_keyring + '.pub') or
|
||||
not os.path.exists(options.gpg_keyring + '.sec')):
|
||||
parser.error("no such key %s" % options.gpg_keyring)
|
||||
|
||||
|
||||
def ask_for_options(self):
|
||||
options = self.options
|
||||
super(Restore, self).ask_for_options()
|
||||
|
||||
# get the directory manager password
|
||||
self.dirman_password = options.password
|
||||
if not options.password:
|
||||
if not options.unattended:
|
||||
self.dirman_password = installutils.read_password(
|
||||
"Directory Manager (existing master)",
|
||||
confirm=False, validate=False)
|
||||
if self.dirman_password is None:
|
||||
raise admintool.ScriptError(
|
||||
"Directory Manager password required")
|
||||
|
||||
|
||||
def run(self):
|
||||
options = self.options
|
||||
super(Restore, self).run()
|
||||
|
||||
self.backup_dir = self.args[0]
|
||||
if not os.path.isabs(self.backup_dir):
|
||||
self.backup_dir = os.path.join(paths.IPA_BACKUP_DIR, self.backup_dir)
|
||||
|
||||
logger.info("Preparing restore from %s on %s",
|
||||
self.backup_dir, FQDN)
|
||||
|
||||
self.header = os.path.join(self.backup_dir, 'header')
|
||||
|
||||
try:
|
||||
self.read_header()
|
||||
except IOError as e:
|
||||
raise admintool.ScriptError("Cannot read backup metadata: %s" % e)
|
||||
|
||||
if options.data_only:
|
||||
restore_type = 'DATA'
|
||||
else:
|
||||
restore_type = self.backup_type
|
||||
|
||||
# These checks would normally be in the validate method but
|
||||
# we need to know the type of backup we're dealing with.
|
||||
if restore_type == 'FULL':
|
||||
if options.online:
|
||||
raise admintool.ScriptError(
|
||||
"File restoration cannot be done online")
|
||||
if options.instance or options.backend:
|
||||
raise admintool.ScriptError(
|
||||
"Restore must be in data-only mode when restoring a "
|
||||
"specific instance or backend")
|
||||
else:
|
||||
installutils.check_server_configuration()
|
||||
|
||||
self.init_api()
|
||||
|
||||
if options.instance:
|
||||
instance_dir = (paths.VAR_LIB_SLAPD_INSTANCE_DIR_TEMPLATE %
|
||||
options.instance)
|
||||
if not os.path.exists(instance_dir):
|
||||
raise admintool.ScriptError(
|
||||
"Instance %s does not exist" % options.instance)
|
||||
|
||||
self.instances = [options.instance]
|
||||
|
||||
if options.backend:
|
||||
for instance in self.instances:
|
||||
db_dir = (paths.SLAPD_INSTANCE_DB_DIR_TEMPLATE %
|
||||
(instance, options.backend))
|
||||
if os.path.exists(db_dir):
|
||||
break
|
||||
else:
|
||||
raise admintool.ScriptError(
|
||||
"Backend %s does not exist" % options.backend)
|
||||
|
||||
self.backends = [options.backend]
|
||||
|
||||
for instance, backend in itertools.product(self.instances,
|
||||
self.backends):
|
||||
db_dir = (paths.SLAPD_INSTANCE_DB_DIR_TEMPLATE %
|
||||
(instance, backend))
|
||||
if os.path.exists(db_dir):
|
||||
break
|
||||
else:
|
||||
raise admintool.ScriptError(
|
||||
"Cannot restore a data backup into an empty system")
|
||||
|
||||
logger.info("Performing %s restore from %s backup",
|
||||
restore_type, self.backup_type)
|
||||
|
||||
if self.backup_host != FQDN:
|
||||
raise admintool.ScriptError(
|
||||
"Host name %s does not match backup name %s" %
|
||||
(FQDN, self.backup_host))
|
||||
|
||||
if self.backup_ipa_version != str(version.VERSION):
|
||||
logger.warning(
|
||||
"Restoring data from a different release of IPA.\n"
|
||||
"Data is version %s.\n"
|
||||
"Server is running %s.",
|
||||
self.backup_ipa_version, str(version.VERSION))
|
||||
if (not options.unattended and
|
||||
not user_input("Continue to restore?", False)):
|
||||
raise admintool.ScriptError("Aborted")
|
||||
|
||||
pent = pwd.getpwnam(constants.DS_USER)
|
||||
|
||||
# Temporary directory for decrypting files before restoring
|
||||
self.top_dir = tempfile.mkdtemp("ipa")
|
||||
os.chown(self.top_dir, pent.pw_uid, pent.pw_gid)
|
||||
os.chmod(self.top_dir, 0o750)
|
||||
self.dir = os.path.join(self.top_dir, "ipa")
|
||||
os.mkdir(self.dir)
|
||||
os.chmod(self.dir, 0o750)
|
||||
os.chown(self.dir, pent.pw_uid, pent.pw_gid)
|
||||
|
||||
cwd = os.getcwd()
|
||||
|
||||
logger.info("Temporary setting umask to 022")
|
||||
old_umask = os.umask(0o022)
|
||||
try:
|
||||
dirsrv = services.knownservices.dirsrv
|
||||
|
||||
self.extract_backup(options.gpg_keyring)
|
||||
|
||||
if restore_type == 'FULL':
|
||||
self.restore_default_conf()
|
||||
self.init_api(confdir=self.dir + paths.ETC_IPA)
|
||||
|
||||
databases = []
|
||||
for instance in self.instances:
|
||||
for backend in self.backends:
|
||||
database = (instance, backend)
|
||||
ldiffile = os.path.join(self.dir, '%s-%s.ldif' % database)
|
||||
if os.path.exists(ldiffile):
|
||||
databases.append(database)
|
||||
|
||||
if options.instance:
|
||||
for instance, backend in databases:
|
||||
if instance == options.instance:
|
||||
break
|
||||
else:
|
||||
raise admintool.ScriptError(
|
||||
"Instance %s not found in backup" % options.instance)
|
||||
|
||||
if options.backend:
|
||||
for instance, backend in databases:
|
||||
if backend == options.backend:
|
||||
break
|
||||
else:
|
||||
raise admintool.ScriptError(
|
||||
"Backend %s not found in backup" % options.backend)
|
||||
|
||||
# Big fat warning
|
||||
if (not options.unattended and
|
||||
not user_input("Restoring data will overwrite existing live data. Continue to restore?", False)):
|
||||
raise admintool.ScriptError("Aborted")
|
||||
|
||||
logger.info(
|
||||
"Each master will individually need to be re-initialized or")
|
||||
logger.info(
|
||||
"re-created from this one. The replication agreements on")
|
||||
logger.info(
|
||||
"masters running IPA 3.1 or earlier will need to be manually")
|
||||
logger.info(
|
||||
"re-enabled. See the man page for details.")
|
||||
|
||||
logger.info("Disabling all replication.")
|
||||
self.disable_agreements()
|
||||
|
||||
if restore_type != 'FULL':
|
||||
if not options.online:
|
||||
logger.info('Stopping Directory Server')
|
||||
dirsrv.stop(capture_output=False)
|
||||
else:
|
||||
logger.info('Starting Directory Server')
|
||||
dirsrv.start(capture_output=False)
|
||||
else:
|
||||
logger.info('Stopping IPA services')
|
||||
result = run(['ipactl', 'stop'], raiseonerr=False)
|
||||
if result.returncode not in [0, 6]:
|
||||
logger.warning('Stopping IPA failed: %s', result.error_log)
|
||||
|
||||
self.restore_selinux_booleans()
|
||||
|
||||
http = httpinstance.HTTPInstance()
|
||||
|
||||
# We do either a full file restore or we restore data.
|
||||
if restore_type == 'FULL':
|
||||
self.remove_old_files()
|
||||
self.cert_restore_prepare()
|
||||
self.file_restore(options.no_logs)
|
||||
self.cert_restore()
|
||||
if 'CA' in self.backup_services:
|
||||
self.__create_dogtag_log_dirs()
|
||||
|
||||
# Always restore the data from ldif
|
||||
# We need to restore both userRoot and ipaca.
|
||||
for instance, backend in databases:
|
||||
self.ldif2db(instance, backend, online=options.online)
|
||||
|
||||
if restore_type != 'FULL':
|
||||
if not options.online:
|
||||
logger.info('Starting Directory Server')
|
||||
dirsrv.start(capture_output=False)
|
||||
else:
|
||||
# restore access controll configuration
|
||||
auth_backup_path = os.path.join(paths.VAR_LIB_IPA, 'auth_backup')
|
||||
if os.path.exists(auth_backup_path):
|
||||
tasks.restore_auth_configuration(auth_backup_path)
|
||||
# explicitly enable then disable the pki tomcatd service to
|
||||
# re-register its instance. FIXME, this is really wierd.
|
||||
services.knownservices.pki_tomcatd.enable()
|
||||
services.knownservices.pki_tomcatd.disable()
|
||||
|
||||
logger.info('Restarting GSS-proxy')
|
||||
gssproxy = services.service('gssproxy', api)
|
||||
gssproxy.reload_or_restart()
|
||||
logger.info('Starting IPA services')
|
||||
run(['ipactl', 'start'])
|
||||
logger.info('Restarting SSSD')
|
||||
sssd = services.service('sssd', api)
|
||||
sssd.restart()
|
||||
http.remove_httpd_ccaches()
|
||||
# have the daemons pick up their restored configs
|
||||
run([paths.SYSTEMCTL, "--system", "daemon-reload"])
|
||||
finally:
|
||||
try:
|
||||
os.chdir(cwd)
|
||||
except Exception as e:
|
||||
logger.error('Cannot change directory to %s: %s', cwd, e)
|
||||
shutil.rmtree(self.top_dir)
|
||||
logger.info("Restoring umask to %s", old_umask)
|
||||
os.umask(old_umask)
|
||||
|
||||
|
||||
def get_connection(self):
|
||||
'''
|
||||
Create an ldapi connection and bind to it using autobind as root.
|
||||
'''
|
||||
instance_name = installutils.realm_to_serverid(api.env.realm)
|
||||
|
||||
if not services.knownservices.dirsrv.is_running(instance_name):
|
||||
raise admintool.ScriptError(
|
||||
"directory server instance is not running/configured"
|
||||
)
|
||||
|
||||
if self._conn is not None:
|
||||
return self._conn
|
||||
|
||||
ldap_uri = ipaldap.get_ldap_uri(protocol='ldapi', realm=api.env.realm)
|
||||
self._conn = ipaldap.LDAPClient(ldap_uri)
|
||||
|
||||
try:
|
||||
self._conn.external_bind()
|
||||
except Exception as e:
|
||||
raise admintool.ScriptError('Unable to bind to LDAP server: %s'
|
||||
% e)
|
||||
return self._conn
|
||||
|
||||
|
||||
def disable_agreements(self):
|
||||
'''
|
||||
Find all replication agreements on all masters and disable them.
|
||||
|
||||
Warn very loudly about any agreements/masters we cannot contact.
|
||||
'''
|
||||
try:
|
||||
conn = self.get_connection()
|
||||
except Exception as e:
|
||||
logger.error('Unable to get connection, skipping disabling '
|
||||
'agreements: %s', e)
|
||||
return
|
||||
masters = []
|
||||
dn = DN(('cn', 'masters'), ('cn', 'ipa'), ('cn', 'etc'), api.env.basedn)
|
||||
try:
|
||||
entries = conn.get_entries(dn, conn.SCOPE_ONELEVEL)
|
||||
except Exception as e:
|
||||
raise admintool.ScriptError(
|
||||
"Failed to read master data: %s" % e)
|
||||
else:
|
||||
masters = [ent.single_value['cn'] for ent in entries]
|
||||
|
||||
for master in masters:
|
||||
if master == api.env.host:
|
||||
continue
|
||||
|
||||
try:
|
||||
repl = ReplicationManager(api.env.realm, master,
|
||||
self.dirman_password)
|
||||
except Exception as e:
|
||||
logger.critical("Unable to disable agreement on %s: %s",
|
||||
master, e)
|
||||
continue
|
||||
|
||||
master_dn = DN(('cn', master), ('cn', 'masters'), ('cn', 'ipa'), ('cn', 'etc'), api.env.basedn)
|
||||
try:
|
||||
services = repl.conn.get_entries(master_dn,
|
||||
repl.conn.SCOPE_ONELEVEL)
|
||||
except errors.NotFound:
|
||||
continue
|
||||
|
||||
services_cns = [s.single_value['cn'] for s in services]
|
||||
|
||||
host_entries = repl.find_ipa_replication_agreements()
|
||||
hosts = [rep.single_value.get('nsds5replicahost')
|
||||
for rep in host_entries]
|
||||
|
||||
for host in hosts:
|
||||
logger.info('Disabling replication agreement on %s to %s',
|
||||
master, host)
|
||||
repl.disable_agreement(host)
|
||||
|
||||
if 'CA' in services_cns:
|
||||
try:
|
||||
repl = get_cs_replication_manager(api.env.realm, master,
|
||||
self.dirman_password)
|
||||
except Exception as e:
|
||||
logger.critical("Unable to disable agreement on %s: %s",
|
||||
master, e)
|
||||
continue
|
||||
|
||||
host_entries = repl.find_ipa_replication_agreements()
|
||||
hosts = [rep.single_value.get('nsds5replicahost')
|
||||
for rep in host_entries]
|
||||
for host in hosts:
|
||||
logger.info('Disabling CA replication agreement on %s to '
|
||||
'%s', master, host)
|
||||
repl.hostnames = [master, host]
|
||||
repl.disable_agreement(host)
|
||||
|
||||
|
||||
def ldif2db(self, instance, backend, online=True):
|
||||
'''
|
||||
Restore a LDIF backup of the data in this instance.
|
||||
|
||||
If executed online create a task and wait for it to complete.
|
||||
'''
|
||||
logger.info('Restoring from %s in %s', backend, instance)
|
||||
|
||||
cn = time.strftime('import_%Y_%m_%d_%H_%M_%S')
|
||||
dn = DN(('cn', cn), ('cn', 'import'), ('cn', 'tasks'), ('cn', 'config'))
|
||||
|
||||
ldifdir = paths.SLAPD_INSTANCE_LDIF_DIR_TEMPLATE % instance
|
||||
ldifname = '%s-%s.ldif' % (instance, backend)
|
||||
ldiffile = os.path.join(ldifdir, ldifname)
|
||||
srcldiffile = os.path.join(self.dir, ldifname)
|
||||
|
||||
if not os.path.exists(ldifdir):
|
||||
pent = pwd.getpwnam(constants.DS_USER)
|
||||
os.mkdir(ldifdir)
|
||||
os.chmod(ldifdir, 0o770)
|
||||
os.chown(ldifdir, pent.pw_uid, pent.pw_gid)
|
||||
|
||||
ipautil.backup_file(ldiffile)
|
||||
with open(ldiffile, 'w') as out_file:
|
||||
ldif_writer = ldif.LDIFWriter(out_file)
|
||||
with open(srcldiffile, 'rb') as in_file:
|
||||
ldif_parser = RemoveRUVParser(in_file, ldif_writer)
|
||||
ldif_parser.parse()
|
||||
|
||||
# Make sure the modified ldiffile is owned by DS_USER
|
||||
pent = pwd.getpwnam(constants.DS_USER)
|
||||
os.chown(ldiffile, pent.pw_uid, pent.pw_gid)
|
||||
|
||||
if online:
|
||||
conn = self.get_connection()
|
||||
ent = conn.make_entry(
|
||||
dn,
|
||||
{
|
||||
'objectClass': ['top', 'extensibleObject'],
|
||||
'cn': [cn],
|
||||
'nsFilename': [ldiffile],
|
||||
'nsUseOneFile': ['true'],
|
||||
}
|
||||
)
|
||||
ent['nsInstance'] = [backend]
|
||||
|
||||
try:
|
||||
conn.add_entry(ent)
|
||||
except Exception as e:
|
||||
logger.error("Unable to bind to LDAP server: %s", e)
|
||||
return
|
||||
|
||||
logger.info("Waiting for LDIF to finish")
|
||||
wait_for_task(conn, dn)
|
||||
else:
|
||||
try:
|
||||
os.makedirs(paths.VAR_LOG_DIRSRV_INSTANCE_TEMPLATE % instance)
|
||||
except OSError as e:
|
||||
pass
|
||||
|
||||
args = [paths.LDIF2DB,
|
||||
'-Z', instance,
|
||||
'-i', ldiffile,
|
||||
'-n', backend]
|
||||
result = run(args, raiseonerr=False)
|
||||
if result.returncode != 0:
|
||||
logger.critical("ldif2db failed: %s", result.error_log)
|
||||
|
||||
|
||||
def bak2db(self, instance, backend, online=True):
|
||||
'''
|
||||
Restore a BAK backup of the data and changelog in this instance.
|
||||
|
||||
If backend is None then all backends are restored.
|
||||
|
||||
If executed online create a task and wait for it to complete.
|
||||
|
||||
instance here is a loaded term. It can mean either a separate
|
||||
389-ds install instance or a separate 389-ds backend. We only need
|
||||
to treat ipaca specially.
|
||||
'''
|
||||
if backend is not None:
|
||||
logger.info('Restoring %s in %s', backend, instance)
|
||||
else:
|
||||
logger.info('Restoring %s', instance)
|
||||
|
||||
cn = time.strftime('restore_%Y_%m_%d_%H_%M_%S')
|
||||
|
||||
dn = DN(('cn', cn), ('cn', 'restore'), ('cn', 'tasks'), ('cn', 'config'))
|
||||
|
||||
if online:
|
||||
conn = self.get_connection()
|
||||
ent = conn.make_entry(
|
||||
dn,
|
||||
{
|
||||
'objectClass': ['top', 'extensibleObject'],
|
||||
'cn': [cn],
|
||||
'nsArchiveDir': [os.path.join(self.dir, instance)],
|
||||
'nsDatabaseType': ['ldbm database'],
|
||||
}
|
||||
)
|
||||
if backend is not None:
|
||||
ent['nsInstance'] = [backend]
|
||||
|
||||
try:
|
||||
conn.add_entry(ent)
|
||||
except Exception as e:
|
||||
raise admintool.ScriptError('Unable to bind to LDAP server: %s'
|
||||
% e)
|
||||
|
||||
logger.info("Waiting for restore to finish")
|
||||
wait_for_task(conn, dn)
|
||||
else:
|
||||
args = [paths.BAK2DB,
|
||||
'-Z', instance,
|
||||
os.path.join(self.dir, instance)]
|
||||
if backend is not None:
|
||||
args.append('-n')
|
||||
args.append(backend)
|
||||
result = run(args, raiseonerr=False)
|
||||
if result.returncode != 0:
|
||||
logger.critical("bak2db failed: %s", result.error_log)
|
||||
|
||||
|
||||
def restore_default_conf(self):
|
||||
'''
|
||||
Restore paths.IPA_DEFAULT_CONF to temporary directory.
|
||||
|
||||
Primary purpose of this method is to get cofiguration for api
|
||||
finalization when restoring ipa after uninstall.
|
||||
'''
|
||||
cwd = os.getcwd()
|
||||
os.chdir(self.dir)
|
||||
args = ['tar',
|
||||
'--xattrs',
|
||||
'--selinux',
|
||||
'-xzf',
|
||||
os.path.join(self.dir, 'files.tar'),
|
||||
paths.IPA_DEFAULT_CONF[1:],
|
||||
]
|
||||
|
||||
result = run(args, raiseonerr=False)
|
||||
if result.returncode != 0:
|
||||
logger.critical('Restoring %s failed: %s',
|
||||
paths.IPA_DEFAULT_CONF, result.error_log)
|
||||
os.chdir(cwd)
|
||||
|
||||
def remove_old_files(self):
|
||||
"""
|
||||
Removes all directories, files or temporal files that should be
|
||||
removed before backup files are copied, to prevent errors.
|
||||
"""
|
||||
for d in self.DIRS_TO_BE_REMOVED:
|
||||
try:
|
||||
shutil.rmtree(d)
|
||||
except OSError as e:
|
||||
if e.errno != 2: # 2: dir does not exist
|
||||
logger.warning("Could not remove directory: %s (%s)", d, e)
|
||||
|
||||
for f in self.FILES_TO_BE_REMOVED:
|
||||
try:
|
||||
os.remove(f)
|
||||
except OSError as e:
|
||||
if e.errno != 2: # 2: file does not exist
|
||||
logger.warning("Could not remove file: %s (%s)", f, e)
|
||||
|
||||
def file_restore(self, nologs=False):
|
||||
'''
|
||||
Restore all the files in the tarball.
|
||||
|
||||
This MUST be done offline because we directly backup the 389-ds
|
||||
databases.
|
||||
'''
|
||||
logger.info("Restoring files")
|
||||
cwd = os.getcwd()
|
||||
os.chdir('/')
|
||||
args = ['tar',
|
||||
'--xattrs',
|
||||
'--selinux',
|
||||
'-xzf',
|
||||
os.path.join(self.dir, 'files.tar')
|
||||
]
|
||||
if nologs:
|
||||
args.append('--exclude')
|
||||
args.append('var/log')
|
||||
|
||||
result = run(args, raiseonerr=False)
|
||||
if result.returncode != 0:
|
||||
logger.critical('Restoring files failed: %s', result.error_log)
|
||||
|
||||
os.chdir(cwd)
|
||||
|
||||
|
||||
def read_header(self):
|
||||
'''
|
||||
Read the backup file header that contains the meta data about
|
||||
this particular backup.
|
||||
'''
|
||||
config = SafeConfigParser()
|
||||
config.read(self.header)
|
||||
|
||||
self.backup_type = config.get('ipa', 'type')
|
||||
self.backup_time = config.get('ipa', 'time')
|
||||
self.backup_host = config.get('ipa', 'host')
|
||||
self.backup_ipa_version = config.get('ipa', 'ipa_version')
|
||||
self.backup_version = config.get('ipa', 'version')
|
||||
# pylint: disable=no-member
|
||||
# we can assume that returned object is string and it has .split()
|
||||
# method
|
||||
self.backup_services = config.get('ipa', 'services').split(',')
|
||||
# pylint: enable=no-member
|
||||
|
||||
|
||||
def extract_backup(self, keyring=None):
|
||||
'''
|
||||
Extract the contents of the tarball backup into a temporary location,
|
||||
decrypting if necessary.
|
||||
'''
|
||||
|
||||
encrypt = False
|
||||
filename = None
|
||||
if self.backup_type == 'FULL':
|
||||
filename = os.path.join(self.backup_dir, 'ipa-full.tar')
|
||||
else:
|
||||
filename = os.path.join(self.backup_dir, 'ipa-data.tar')
|
||||
if not os.path.exists(filename):
|
||||
if not os.path.exists(filename + '.gpg'):
|
||||
raise admintool.ScriptError('Unable to find backup file in %s' % self.backup_dir)
|
||||
else:
|
||||
filename = filename + '.gpg'
|
||||
encrypt = True
|
||||
|
||||
if encrypt:
|
||||
logger.info('Decrypting %s', filename)
|
||||
filename = decrypt_file(self.dir, filename, keyring)
|
||||
|
||||
os.chdir(self.dir)
|
||||
|
||||
args = ['tar',
|
||||
'--xattrs',
|
||||
'--selinux',
|
||||
'-xzf',
|
||||
filename,
|
||||
'.'
|
||||
]
|
||||
run(args)
|
||||
|
||||
pent = pwd.getpwnam(constants.DS_USER)
|
||||
os.chown(self.top_dir, pent.pw_uid, pent.pw_gid)
|
||||
recursive_chown(self.dir, pent.pw_uid, pent.pw_gid)
|
||||
|
||||
if encrypt:
|
||||
# We can remove the decoded tarball
|
||||
os.unlink(filename)
|
||||
|
||||
def __create_dogtag_log_dirs(self):
|
||||
"""
|
||||
If we are doing a full restore and the dogtag log directories do
|
||||
not exist then tomcat will fail to start.
|
||||
|
||||
The directory is different depending on whether we have a d9-based
|
||||
or a d10-based installation.
|
||||
"""
|
||||
dirs = []
|
||||
# dogtag 10
|
||||
if (os.path.exists(paths.VAR_LIB_PKI_TOMCAT_DIR) and
|
||||
not os.path.exists(paths.TOMCAT_TOPLEVEL_DIR)):
|
||||
dirs += [paths.TOMCAT_TOPLEVEL_DIR,
|
||||
paths.TOMCAT_CA_DIR,
|
||||
paths.TOMCAT_CA_ARCHIVE_DIR,
|
||||
paths.TOMCAT_SIGNEDAUDIT_DIR]
|
||||
|
||||
try:
|
||||
pent = pwd.getpwnam(constants.PKI_USER)
|
||||
except KeyError:
|
||||
logger.debug("No %s user exists, skipping CA directory creation",
|
||||
constants.PKI_USER)
|
||||
return
|
||||
logger.debug('Creating log directories for dogtag')
|
||||
for dir in dirs:
|
||||
try:
|
||||
logger.debug('Creating %s', dir)
|
||||
os.mkdir(dir)
|
||||
os.chmod(dir, 0o770)
|
||||
os.chown(dir, pent.pw_uid, pent.pw_gid)
|
||||
tasks.restore_context(dir)
|
||||
except Exception as e:
|
||||
# This isn't so fatal as to side-track the restore
|
||||
logger.error('Problem with %s: %s', dir, e)
|
||||
|
||||
def restore_selinux_booleans(self):
|
||||
bools = dict(constants.SELINUX_BOOLEAN_HTTPD)
|
||||
if 'ADTRUST' in self.backup_services:
|
||||
if adtrustinstance:
|
||||
bools.update(constants.SELINUX_BOOLEAN_ADTRUST)
|
||||
else:
|
||||
logger.error(
|
||||
'The AD trust package was not found, '
|
||||
'not setting SELinux booleans.')
|
||||
try:
|
||||
tasks.set_selinux_booleans(bools)
|
||||
except ipapython.errors.SetseboolError as e:
|
||||
logger.error('%s', e)
|
||||
|
||||
def cert_restore_prepare(self):
|
||||
cainstance.CAInstance().stop_tracking_certificates()
|
||||
httpinstance.HTTPInstance().stop_tracking_certificates()
|
||||
try:
|
||||
dsinstance.DsInstance().stop_tracking_certificates(
|
||||
installutils.realm_to_serverid(api.env.realm))
|
||||
except (OSError, IOError):
|
||||
# When IPA is not installed, DS NSS DB does not exist
|
||||
pass
|
||||
|
||||
krbinstance.KrbInstance().stop_tracking_certs()
|
||||
|
||||
for basename in ('cert8.db', 'key3.db', 'secmod.db', 'pwdfile.txt'):
|
||||
filename = os.path.join(paths.IPA_NSSDB_DIR, basename)
|
||||
try:
|
||||
ipautil.backup_file(filename)
|
||||
except OSError as e:
|
||||
logger.error("Failed to backup %s: %s", filename, e)
|
||||
|
||||
tasks.remove_ca_certs_from_systemwide_ca_store()
|
||||
|
||||
def cert_restore(self):
|
||||
try:
|
||||
update_ipa_nssdb()
|
||||
except RuntimeError as e:
|
||||
logger.error("%s", e)
|
||||
|
||||
tasks.reload_systemwide_ca_store()
|
||||
|
||||
services.knownservices.certmonger.restart()
|
||||
|
||||
def init_api(self, **overrides):
|
||||
overrides.setdefault('confdir', paths.ETC_IPA)
|
||||
api.bootstrap(in_server=True, context='restore', **overrides)
|
||||
api.finalize()
|
||||
|
||||
self.instances = [installutils.realm_to_serverid(api.env.realm)]
|
||||
self.backends = ['userRoot', 'ipaca']
|
||||
Reference in New Issue
Block a user