mirror of
https://salsa.debian.org/freeipa-team/freeipa.git
synced 2024-12-27 09:21:59 -06:00
fea7163e87
Reviewed-By: Petr Viktorin <pviktori@redhat.com>
1544 lines
57 KiB
Python
1544 lines
57 KiB
Python
# Authors: Karl MacMillan <kmacmillan@mentalrootkit.com>
|
|
#
|
|
# Copyright (C) 2007 Red Hat
|
|
# see file 'COPYING' for use and warranty information
|
|
#
|
|
# This program is free software; you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
#
|
|
|
|
import time
|
|
import datetime
|
|
import sys
|
|
import os
|
|
|
|
import ldap
|
|
|
|
from ipalib import api, errors
|
|
from ipalib.constants import CACERT
|
|
from ipapython import services as ipaservices
|
|
from ipapython.ipa_log_manager import *
|
|
from ipapython import ipautil, dogtag, ipaldap
|
|
from ipapython.dn import DN
|
|
|
|
# the default container used by AD for user entries
|
|
WIN_USER_CONTAINER = DN(('cn', 'Users'))
|
|
# the default container used by IPA for user entries
|
|
IPA_USER_CONTAINER = DN(('cn', 'users'), ('cn', 'accounts'))
|
|
PORT = 636
|
|
TIMEOUT = 120
|
|
REPL_MAN_DN = DN(('cn', 'replication manager'), ('cn', 'config'))
|
|
DNA_DN = DN(('cn', 'Posix IDs'), ('cn', 'Distributed Numeric Assignment Plugin'), ('cn', 'plugins'), ('cn', 'config'))
|
|
|
|
IPA_REPLICA = 1
|
|
WINSYNC = 2
|
|
|
|
# List of attributes that need to be excluded from replication initialization.
|
|
TOTAL_EXCLUDES = ('entryusn',
|
|
'krblastsuccessfulauth',
|
|
'krblastfailedauth',
|
|
'krbloginfailedcount')
|
|
|
|
# List of attributes that need to be excluded from normal replication.
|
|
EXCLUDES = ('memberof', 'idnssoaserial') + TOTAL_EXCLUDES
|
|
|
|
# List of attributes that are not updated on empty replication
|
|
STRIP_ATTRS = ('modifiersName',
|
|
'modifyTimestamp',
|
|
'internalModifiersName',
|
|
'internalModifyTimestamp')
|
|
|
|
|
|
def replica_conn_check(master_host, host_name, realm, check_ca,
|
|
dogtag_master_ds_port, admin_password=None):
|
|
"""
|
|
Check the ports used by the replica both locally and remotely to be sure
|
|
that replication will work.
|
|
|
|
Does not return a value, will sys.exit() on failure.
|
|
"""
|
|
print "Run connection check to master"
|
|
args = ["/usr/sbin/ipa-replica-conncheck", "--master", master_host,
|
|
"--auto-master-check", "--realm", realm,
|
|
"--principal", "admin",
|
|
"--hostname", host_name]
|
|
nolog=tuple()
|
|
|
|
if admin_password:
|
|
args.extend(["--password", admin_password])
|
|
nolog=(admin_password,)
|
|
|
|
if check_ca and dogtag_master_ds_port == dogtag.Dogtag9Constants.DS_PORT:
|
|
args.append('--check-ca')
|
|
(stdin, stderr, returncode) = ipautil.run(
|
|
args, raiseonerr=False, capture_output=False, nolog=nolog)
|
|
|
|
if returncode != 0:
|
|
sys.exit("Connection check failed!" +
|
|
"\nPlease fix your network settings according to error messages above." +
|
|
"\nIf the check results are not valid it can be skipped with --skip-conncheck parameter.")
|
|
else:
|
|
print "Connection check OK"
|
|
|
|
def enable_replication_version_checking(hostname, realm, dirman_passwd):
|
|
"""
|
|
Check the replication version checking plugin. If it is not
|
|
enabled then enable it and restart 389-ds. If it is enabled
|
|
the do nothing.
|
|
"""
|
|
conn = ipaldap.IPAdmin(hostname, realm=realm, ldapi=True)
|
|
if dirman_passwd:
|
|
conn.do_simple_bind(bindpw=dirman_passwd)
|
|
else:
|
|
conn.do_sasl_gssapi_bind()
|
|
entry = conn.get_entry(DN(('cn', 'IPA Version Replication'),
|
|
('cn', 'plugins'),
|
|
('cn', 'config')))
|
|
if entry.single_value.get('nsslapd-pluginenabled') == 'off':
|
|
conn.modify_s(entry.dn, [(ldap.MOD_REPLACE, 'nsslapd-pluginenabled', 'on')])
|
|
conn.unbind()
|
|
serverid = "-".join(realm.split("."))
|
|
ipaservices.knownservices.dirsrv.restart(instance_name=serverid)
|
|
else:
|
|
conn.unbind()
|
|
|
|
|
|
def wait_for_task(conn, dn):
|
|
"""Check task status
|
|
|
|
Task is complete when the nsTaskExitCode attr is set.
|
|
|
|
:return: the task's return code
|
|
"""
|
|
assert isinstance(dn, DN)
|
|
attrlist = [
|
|
'nsTaskLog', 'nsTaskStatus', 'nsTaskExitCode', 'nsTaskCurrentItem',
|
|
'nsTaskTotalItems']
|
|
while True:
|
|
entry = conn.get_entry(dn, attrlist)
|
|
if entry.single_value.get('nsTaskExitCode'):
|
|
exit_code = int(entry.single_value['nsTaskExitCode'])
|
|
break
|
|
time.sleep(1)
|
|
return exit_code
|
|
|
|
|
|
def wait_for_entry(connection, entry, timeout=7200, attr='', quiet=True):
|
|
"""Wait for entry and/or attr to show up"""
|
|
|
|
filter = "(objectclass=*)"
|
|
attrlist = []
|
|
if attr:
|
|
filter = "(%s=*)" % attr
|
|
attrlist.append(attr)
|
|
timeout += int(time.time())
|
|
|
|
dn = entry.dn
|
|
|
|
if not quiet:
|
|
sys.stdout.write("Waiting for %s %s:%s " % (connection, dn, attr))
|
|
sys.stdout.flush()
|
|
entry = None
|
|
while not entry and int(time.time()) < timeout:
|
|
try:
|
|
[entry] = connection.get_entries(
|
|
dn, ldap.SCOPE_BASE, filter, attrlist)
|
|
except errors.NotFound:
|
|
pass # no entry yet
|
|
except Exception, e: # badness
|
|
print "\nError reading entry", dn, e
|
|
break
|
|
if not entry:
|
|
if not quiet:
|
|
sys.stdout.write(".")
|
|
sys.stdout.flush()
|
|
time.sleep(1)
|
|
|
|
if not entry and int(time.time()) > timeout:
|
|
print "\nwait_for_entry timeout for %s for %s" % (connection, dn)
|
|
elif entry and not quiet:
|
|
print "\nThe waited for entry is:", entry
|
|
elif not entry:
|
|
print "\nError: could not read entry %s from %s" % (dn, connection)
|
|
|
|
|
|
class ReplicationManager(object):
|
|
"""Manage replication agreements between DS servers, and sync
|
|
agreements with Windows servers"""
|
|
def __init__(self, realm, hostname, dirman_passwd, port=PORT, starttls=False, conn=None):
|
|
self.hostname = hostname
|
|
self.port = port
|
|
self.dirman_passwd = dirman_passwd
|
|
self.realm = realm
|
|
self.starttls = starttls
|
|
self.suffix = ipautil.realm_to_suffix(realm)
|
|
self.need_memberof_fixup = False
|
|
|
|
# The caller is allowed to pass in an existing IPAdmin connection.
|
|
# Open a new one if not provided
|
|
if conn is None:
|
|
# If we are passed a password we'll use it as the DM password
|
|
# otherwise we'll do a GSSAPI bind.
|
|
if starttls:
|
|
self.conn = ipaldap.IPAdmin(
|
|
hostname, port=port, cacert=CACERT, protocol='ldap',
|
|
start_tls=True)
|
|
else:
|
|
self.conn = ipaldap.IPAdmin(hostname, port=port, cacert=CACERT)
|
|
if dirman_passwd:
|
|
self.conn.do_simple_bind(bindpw=dirman_passwd)
|
|
else:
|
|
self.conn.do_sasl_gssapi_bind()
|
|
else:
|
|
self.conn = conn
|
|
|
|
self.repl_man_passwd = dirman_passwd
|
|
|
|
# these are likely constant, but you could change them
|
|
# at runtime if you really want
|
|
self.repl_man_dn = REPL_MAN_DN
|
|
self.repl_man_cn = "replication manager"
|
|
|
|
def _get_replica_id(self, conn, master_conn):
|
|
"""
|
|
Returns the replica ID which is unique for each backend.
|
|
|
|
conn is the connection we are trying to get the replica ID for.
|
|
master_conn is the master we are going to replicate with.
|
|
"""
|
|
# First see if there is already one set
|
|
dn = self.replica_dn()
|
|
assert isinstance(dn, DN)
|
|
try:
|
|
replica = conn.get_entry(dn)
|
|
except errors.NotFound:
|
|
pass
|
|
else:
|
|
if replica.single_value.get('nsDS5ReplicaId'):
|
|
return int(replica.single_value['nsDS5ReplicaId'])
|
|
|
|
# Ok, either the entry doesn't exist or the attribute isn't set
|
|
# so get it from the other master
|
|
retval = -1
|
|
dn = DN(('cn','replication'),('cn','etc'), self.suffix)
|
|
try:
|
|
replica = master_conn.get_entry(dn)
|
|
except errors.NotFound:
|
|
root_logger.debug("Unable to retrieve nsDS5ReplicaId from remote server")
|
|
raise
|
|
else:
|
|
if replica.single_value.get('nsDS5ReplicaId') is None:
|
|
root_logger.debug("Unable to retrieve nsDS5ReplicaId from remote server")
|
|
raise RuntimeError("Unable to retrieve nsDS5ReplicaId from remote server")
|
|
|
|
# Now update the value on the master
|
|
retval = int(replica.single_value['nsDS5ReplicaId'])
|
|
mod = [(ldap.MOD_REPLACE, 'nsDS5ReplicaId', str(retval + 1))]
|
|
|
|
try:
|
|
master_conn.modify_s(dn, mod)
|
|
except Exception, e:
|
|
root_logger.debug("Problem updating nsDS5ReplicaID %s" % e)
|
|
raise
|
|
|
|
return retval
|
|
|
|
def get_agreement_filter(self, agreement_types=None, host=None):
|
|
"""
|
|
Get an LDAP replication agreement filter with a possibility to filter
|
|
the agreements by their type and a host
|
|
"""
|
|
if agreement_types is None:
|
|
agreement_types = (IPA_REPLICA, WINSYNC)
|
|
elif not isinstance(agreement_types, (list, tuple)):
|
|
agreement_types = (agreement_types,)
|
|
|
|
agreement_types_filters = []
|
|
if IPA_REPLICA in agreement_types:
|
|
agreement_types_filters.append('(&(objectclass=nsds5ReplicationAgreement)(nsDS5ReplicaRoot=%s))'
|
|
% self.suffix)
|
|
if WINSYNC in agreement_types:
|
|
agreement_types_filters.append('(objectclass=nsDSWindowsReplicationAgreement)')
|
|
if len(agreement_types_filters) > 1:
|
|
agreement_filter = '(|%s)' % ''.join(agreement_types_filters)
|
|
else:
|
|
agreement_filter = ''.join(agreement_types_filters)
|
|
|
|
if host is not None:
|
|
agreement_filter = '(&%s(nsDS5ReplicaHost=%s))' % (agreement_filter, host)
|
|
|
|
return agreement_filter
|
|
|
|
def find_replication_agreements(self):
|
|
"""
|
|
The replication agreements are stored in
|
|
cn="$SUFFIX",cn=mapping tree,cn=config
|
|
|
|
FIXME: Rather than failing with a read error if a user tries
|
|
to read this it simply returns zero entries. We need to use
|
|
GER to determine if we are allowed to read this to return a proper
|
|
response. For now just return "No entries" even if the user may
|
|
not be allowed to see them.
|
|
"""
|
|
filt = self.get_agreement_filter()
|
|
try:
|
|
ents = self.conn.get_entries(
|
|
DN(('cn', 'mapping tree'), ('cn', 'config')),
|
|
ldap.SCOPE_SUBTREE, filt)
|
|
except errors.NotFound:
|
|
ents = []
|
|
return ents
|
|
|
|
def find_ipa_replication_agreements(self):
|
|
"""
|
|
The replication agreements are stored in
|
|
cn="$SUFFIX",cn=mapping tree,cn=config
|
|
|
|
Return the list of hosts we have replication agreements.
|
|
"""
|
|
|
|
filt = self.get_agreement_filter(IPA_REPLICA)
|
|
try:
|
|
ents = self.conn.get_entries(
|
|
DN(('cn', 'mapping tree'), ('cn', 'config')),
|
|
ldap.SCOPE_SUBTREE, filt)
|
|
except errors.NotFound:
|
|
ents = []
|
|
|
|
return ents
|
|
|
|
def get_replication_agreement(self, hostname):
|
|
"""
|
|
The replication agreements are stored in
|
|
cn="$SUFFIX",cn=mapping tree,cn=config
|
|
|
|
Get the replication agreement for a specific host.
|
|
|
|
Returns None if not found.
|
|
"""
|
|
|
|
filt = self.get_agreement_filter(host=hostname)
|
|
try:
|
|
entries = self.conn.get_entries(
|
|
DN(('cn', 'mapping tree'), ('cn', 'config')),
|
|
ldap.SCOPE_SUBTREE, filt)
|
|
except errors.NotFound:
|
|
return None
|
|
|
|
if len(entries) == 0:
|
|
return None
|
|
else:
|
|
return entries[0] # There can be only one
|
|
|
|
def add_replication_manager(self, conn, dn, pw):
|
|
"""
|
|
Create a pseudo user to use for replication.
|
|
"""
|
|
assert isinstance(dn, DN)
|
|
rdn_attr = dn[0].attr
|
|
rdn_val = dn[0].value
|
|
|
|
ent = conn.make_entry(
|
|
dn,
|
|
{
|
|
'objectclass': ["top", "person"],
|
|
rdn_attr: [rdn_val],
|
|
'userpassword': [pw],
|
|
'sn': ["replication manager pseudo user"],
|
|
}
|
|
)
|
|
|
|
try:
|
|
conn.add_entry(ent)
|
|
except errors.DuplicateEntry:
|
|
conn.modify_s(dn, [(ldap.MOD_REPLACE, "userpassword", pw)])
|
|
pass
|
|
|
|
def delete_replication_manager(self, conn, dn=REPL_MAN_DN):
|
|
assert isinstance(dn, DN)
|
|
try:
|
|
conn.delete_entry(dn)
|
|
except errors.NotFound:
|
|
pass
|
|
|
|
def get_replica_type(self, master=True):
|
|
if master:
|
|
return "3"
|
|
else:
|
|
return "2"
|
|
|
|
def replica_dn(self):
|
|
return DN(('cn','replica'),('cn',self.suffix),('cn','mapping tree'),('cn','config'))
|
|
|
|
def replica_config(self, conn, replica_id, replica_binddn):
|
|
assert isinstance(replica_binddn, DN)
|
|
dn = self.replica_dn()
|
|
assert isinstance(dn, DN)
|
|
|
|
try:
|
|
entry = conn.get_entry(dn)
|
|
managers = entry.get('nsDS5ReplicaBindDN')
|
|
for m in managers:
|
|
if replica_binddn == DN(m):
|
|
return
|
|
# Add the new replication manager
|
|
mod = [(ldap.MOD_ADD, 'nsDS5ReplicaBindDN', replica_binddn)]
|
|
conn.modify_s(dn, mod)
|
|
|
|
# replication is already configured
|
|
return
|
|
except errors.NotFound:
|
|
pass
|
|
|
|
replica_type = self.get_replica_type()
|
|
|
|
entry = conn.make_entry(
|
|
dn,
|
|
objectclass=["top", "nsds5replica", "extensibleobject"],
|
|
cn=["replica"],
|
|
nsds5replicaroot=[str(self.suffix)],
|
|
nsds5replicaid=[str(replica_id)],
|
|
nsds5replicatype=[replica_type],
|
|
nsds5flags=["1"],
|
|
nsds5replicabinddn=[replica_binddn],
|
|
nsds5replicalegacyconsumer=["off"],
|
|
)
|
|
conn.add_entry(entry)
|
|
|
|
def setup_changelog(self, conn):
|
|
ent = conn.get_entry(
|
|
DN(
|
|
('cn', 'config'), ('cn', 'ldbm database'),
|
|
('cn', 'plugins'), ('cn', 'config')),
|
|
['nsslapd-directory'])
|
|
dbdir = os.path.dirname(ent.single_value.get('nsslapd-directory'))
|
|
|
|
entry = conn.make_entry(
|
|
DN(('cn', 'changelog5'), ('cn', 'config')),
|
|
{
|
|
'objectclass': ["top", "extensibleobject"],
|
|
'cn': ["changelog5"],
|
|
'nsslapd-changelogdir': [os.path.join(dbdir, "cldb")],
|
|
}
|
|
)
|
|
try:
|
|
conn.add_entry(entry)
|
|
except errors.DuplicateEntry:
|
|
return
|
|
|
|
def setup_chaining_backend(self, conn):
|
|
chaindn = DN(('cn', 'chaining database'), ('cn', 'plugins'), ('cn', 'config'))
|
|
benamebase = "chaindb"
|
|
urls = [self.to_ldap_url(conn)]
|
|
cn = ""
|
|
benum = 1
|
|
done = False
|
|
while not done:
|
|
try:
|
|
cn = benamebase + str(benum) # e.g. localdb1
|
|
dn = DN(('cn', cn), chaindn)
|
|
entry = conn.make_entry(
|
|
dn,
|
|
{
|
|
'objectclass': [
|
|
'top', 'extensibleObject', 'nsBackendInstance'],
|
|
'cn': [cn],
|
|
'nsslapd-suffix': [str(self.suffix)],
|
|
'nsfarmserverurl': urls,
|
|
'nsmultiplexorbinddn': [self.repl_man_dn],
|
|
'nsmultiplexorcredentials': [self.repl_man_passwd],
|
|
}
|
|
)
|
|
self.conn.add_entry(entry)
|
|
done = True
|
|
except errors.DuplicateEntry:
|
|
benum += 1
|
|
except errors.ExecutionError, e:
|
|
print "Could not add backend entry " + dn, e
|
|
raise
|
|
|
|
return cn
|
|
|
|
def to_ldap_url(self, conn):
|
|
return "ldap://%s/" % ipautil.format_netloc(conn.host, conn.port)
|
|
|
|
def setup_chaining_farm(self, conn):
|
|
try:
|
|
conn.modify_s(self.suffix, [(ldap.MOD_ADD, 'aci',
|
|
[ "(targetattr = \"*\")(version 3.0; acl \"Proxied authorization for database links\"; allow (proxy) userdn = \"ldap:///%s\";)" % self.repl_man_dn ])])
|
|
except ldap.TYPE_OR_VALUE_EXISTS:
|
|
root_logger.debug("proxy aci already exists in suffix %s on %s" % (self.suffix, conn.host))
|
|
|
|
def get_mapping_tree_entry(self):
|
|
try:
|
|
entries = self.conn.get_entries(
|
|
DN(('cn', 'mapping tree'), ('cn', 'config')),
|
|
ldap.SCOPE_ONELEVEL,
|
|
"(cn=\"%s\")" % (self.suffix))
|
|
# TODO: Check we got only one entry
|
|
return entries[0]
|
|
except errors.NotFound:
|
|
root_logger.debug(
|
|
"failed to find mapping tree entry for %s", self.suffix)
|
|
raise
|
|
|
|
|
|
def enable_chain_on_update(self, bename):
|
|
mtent = self.get_mapping_tree_entry()
|
|
dn = mtent.dn
|
|
|
|
plgent = self.conn.get_entry(
|
|
DN(('cn', 'Multimaster Replication Plugin'), ('cn', 'plugins'),
|
|
('cn', 'config')),
|
|
['nsslapd-pluginPath'])
|
|
path = plgent.single_value.get('nsslapd-pluginPath')
|
|
|
|
mod = [(ldap.MOD_REPLACE, 'nsslapd-state', 'backend'),
|
|
(ldap.MOD_ADD, 'nsslapd-backend', bename),
|
|
(ldap.MOD_ADD, 'nsslapd-distribution-plugin', path),
|
|
(ldap.MOD_ADD, 'nsslapd-distribution-funct', 'repl_chain_on_update')]
|
|
|
|
try:
|
|
self.conn.modify_s(dn, mod)
|
|
except ldap.TYPE_OR_VALUE_EXISTS:
|
|
root_logger.debug("chainOnUpdate already enabled for %s" % self.suffix)
|
|
|
|
def setup_chain_on_update(self, other_conn):
|
|
chainbe = self.setup_chaining_backend(other_conn)
|
|
self.enable_chain_on_update(chainbe)
|
|
|
|
def add_passsync_user(self, conn, password):
|
|
pass_dn = DN(('uid', 'passsync'), ('cn', 'sysaccounts'), ('cn', 'etc'), self.suffix)
|
|
print "The user for the Windows PassSync service is %s" % pass_dn
|
|
try:
|
|
conn.get_entry(pass_dn)
|
|
print "Windows PassSync entry exists, not resetting password"
|
|
return
|
|
except errors.NotFound:
|
|
pass
|
|
|
|
# The user doesn't exist, add it
|
|
entry = conn.make_entry(
|
|
pass_dn,
|
|
objectclass=["account", "simplesecurityobject"],
|
|
uid=["passsync"],
|
|
userPassword=[password],
|
|
)
|
|
conn.add_entry(entry)
|
|
|
|
# Add it to the list of users allowed to bypass password policy
|
|
extop_dn = DN(('cn', 'ipa_pwd_extop'), ('cn', 'plugins'), ('cn', 'config'))
|
|
entry = conn.get_entry(extop_dn)
|
|
pass_mgrs = entry.get('passSyncManagersDNs')
|
|
if not pass_mgrs:
|
|
pass_mgrs = []
|
|
if not isinstance(pass_mgrs, list):
|
|
pass_mgrs = [pass_mgrs]
|
|
pass_mgrs.append(pass_dn)
|
|
mod = [(ldap.MOD_REPLACE, 'passSyncManagersDNs', pass_mgrs)]
|
|
conn.modify_s(extop_dn, mod)
|
|
|
|
# And finally grant it permission to write passwords
|
|
mod = [(ldap.MOD_ADD, 'aci',
|
|
['(targetattr = "userPassword || krbPrincipalKey || sambaLMPassword || sambaNTPassword || passwordHistory")(version 3.0; acl "Windows PassSync service can write passwords"; allow (write) userdn="ldap:///%s";)' % pass_dn])]
|
|
try:
|
|
conn.modify_s(self.suffix, mod)
|
|
except ldap.TYPE_OR_VALUE_EXISTS:
|
|
root_logger.debug("passsync aci already exists in suffix %s on %s" % (self.suffix, conn.host))
|
|
|
|
def setup_winsync_agmt(self, entry, win_subtree=None):
|
|
if win_subtree is None:
|
|
win_subtree = DN(WIN_USER_CONTAINER, self.ad_suffix)
|
|
ds_subtree = DN(IPA_USER_CONTAINER, self.suffix)
|
|
windomain = ipautil.suffix_to_realm(self.suffix)
|
|
|
|
entry["objectclass"] = ["nsDSWindowsReplicationAgreement"]
|
|
entry["nsds7WindowsReplicaSubtree"] = [win_subtree]
|
|
entry["nsds7DirectoryReplicaSubtree"] = [ds_subtree]
|
|
# for now, just sync users and ignore groups
|
|
entry["nsds7NewWinUserSyncEnabled"] = ['true']
|
|
entry["nsds7NewWinGroupSyncEnabled"] = ['false']
|
|
entry["nsds7WindowsDomain"] = [windomain]
|
|
|
|
def agreement_dn(self, hostname, master=None):
|
|
"""
|
|
IPA agreement use the same dn on both sides, dogtag does not.
|
|
master is not used for IPA agreements but for dogtag it will
|
|
tell which side we want.
|
|
"""
|
|
cn = "meTo%s" % (hostname)
|
|
dn = DN(('cn', cn), self.replica_dn())
|
|
|
|
return (cn, dn)
|
|
|
|
def setup_agreement(self, a_conn, b_hostname, port=389,
|
|
repl_man_dn=None, repl_man_passwd=None,
|
|
iswinsync=False, win_subtree=None, isgssapi=False,
|
|
master=None):
|
|
"""
|
|
master is used to determine which side of the agreement we are
|
|
creating. This is only needed for dogtag replication agreements
|
|
which use a different name on each side. If master is None then
|
|
isn't a dogtag replication agreement.
|
|
"""
|
|
|
|
if repl_man_dn is not None:
|
|
assert isinstance(repl_man_dn, DN)
|
|
|
|
cn, dn = self.agreement_dn(b_hostname, master=master)
|
|
try:
|
|
a_conn.get_entry(dn)
|
|
return
|
|
except errors.NotFound:
|
|
pass
|
|
|
|
entry = a_conn.make_entry(
|
|
dn,
|
|
objectclass=["nsds5replicationagreement"],
|
|
cn=[cn],
|
|
nsds5replicahost=[b_hostname],
|
|
nsds5replicaport=[str(port)],
|
|
nsds5replicatimeout=[str(TIMEOUT)],
|
|
nsds5replicaroot=[str(self.suffix)],
|
|
description=["me to %s" % b_hostname],
|
|
)
|
|
if master is None:
|
|
entry['nsDS5ReplicatedAttributeList'] = [
|
|
'(objectclass=*) $ EXCLUDE %s' % " ".join(EXCLUDES)]
|
|
if isgssapi:
|
|
entry['nsds5replicatransportinfo'] = ['LDAP']
|
|
entry['nsds5replicabindmethod'] = ['SASL/GSSAPI']
|
|
else:
|
|
entry['nsds5replicabinddn'] = [repl_man_dn]
|
|
entry['nsds5replicacredentials'] = [repl_man_passwd]
|
|
entry['nsds5replicatransportinfo'] = ['TLS']
|
|
entry['nsds5replicabindmethod'] = ['simple']
|
|
|
|
if iswinsync:
|
|
self.setup_winsync_agmt(entry, win_subtree)
|
|
else:
|
|
entry['nsds5ReplicaStripAttrs'] = [" ".join(STRIP_ATTRS)]
|
|
|
|
a_conn.add_entry(entry)
|
|
|
|
try:
|
|
mod = [(ldap.MOD_ADD, 'nsDS5ReplicatedAttributeListTotal',
|
|
'(objectclass=*) $ EXCLUDE %s' % " ".join(TOTAL_EXCLUDES))]
|
|
a_conn.modify_s(dn, mod)
|
|
except ldap.LDAPError, e:
|
|
# Apparently there are problems set the total list
|
|
# Probably the master is an old 389-ds server, tell the caller
|
|
# that we will have to set the memberof fixup task
|
|
self.need_memberof_fixup = True
|
|
|
|
wait_for_entry(a_conn, entry)
|
|
|
|
def needs_memberof_fixup(self):
|
|
return self.need_memberof_fixup
|
|
|
|
def get_replica_principal_dns(self, a, b, retries):
|
|
"""
|
|
Get the DNs of the ldap principals we are going to convert
|
|
to using GSSAPI replication.
|
|
|
|
Arguments a and b are LDAP connections. retries is the number
|
|
of attempts that should be made to find the entries. It could
|
|
be that replication is slow.
|
|
|
|
If successful this returns a tuple (dn_a, dn_b).
|
|
|
|
If either of the DNs doesn't exist after the retries are
|
|
exhausted an exception is raised.
|
|
"""
|
|
filter_a = '(krbprincipalname=ldap/%s@%s)' % (a.host, self.realm)
|
|
filter_b = '(krbprincipalname=ldap/%s@%s)' % (b.host, self.realm)
|
|
|
|
a_entry = None
|
|
b_entry = None
|
|
error_message = ''
|
|
|
|
while (retries > 0 ):
|
|
root_logger.info('Getting ldap service principals for conversion: %s and %s' % (filter_a, filter_b))
|
|
try:
|
|
a_entry = b.get_entries(self.suffix, ldap.SCOPE_SUBTREE,
|
|
filter=filter_a)
|
|
except errors.NotFound:
|
|
pass
|
|
|
|
try:
|
|
b_entry = a.get_entries(self.suffix, ldap.SCOPE_SUBTREE,
|
|
filter=filter_b)
|
|
except errors.NotFound:
|
|
pass
|
|
|
|
if a_entry and b_entry:
|
|
root_logger.debug('Found both principals.')
|
|
break
|
|
|
|
# One or both is missing, force sync again
|
|
if not a_entry:
|
|
root_logger.debug('Unable to find entry for %s on %s'
|
|
% (filter_a, str(b)))
|
|
self.force_sync(a, b.host)
|
|
cn, dn = self.agreement_dn(b.host)
|
|
haserror, error_message = self.wait_for_repl_update(a, dn, 60)
|
|
|
|
if not b_entry:
|
|
root_logger.debug('Unable to find entry for %s on %s'
|
|
% (filter_b, str(a)))
|
|
self.force_sync(b, a.host)
|
|
cn, dn = self.agreement_dn(a.host)
|
|
haserror, error_message = self.wait_for_repl_update(b, dn, 60)
|
|
|
|
retries -= 1
|
|
|
|
if not a_entry or not b_entry:
|
|
error = 'One of the ldap service principals is missing. ' \
|
|
'Replication agreement cannot be converted.'
|
|
if error_message:
|
|
error += '\nReplication error message: %s' % error_message
|
|
raise RuntimeError(error)
|
|
|
|
return (a_entry[0].dn, b_entry[0].dn)
|
|
|
|
def setup_krb_princs_as_replica_binddns(self, a, b):
|
|
"""
|
|
Search the appropriate principal names so we can get
|
|
the correct DNs to store in the replication agreements.
|
|
Then modify the replica object to allow these DNs to act
|
|
as replication agents.
|
|
"""
|
|
|
|
rep_dn = self.replica_dn()
|
|
assert isinstance(rep_dn, DN)
|
|
(a_dn, b_dn) = self.get_replica_principal_dns(a, b, retries=100)
|
|
assert isinstance(a_dn, DN)
|
|
assert isinstance(b_dn, DN)
|
|
|
|
# Add kerberos principal DNs as valid bindDNs for replication
|
|
try:
|
|
mod = [(ldap.MOD_ADD, "nsds5replicabinddn", b_dn)]
|
|
a.modify_s(rep_dn, mod)
|
|
except ldap.TYPE_OR_VALUE_EXISTS:
|
|
pass
|
|
try:
|
|
mod = [(ldap.MOD_ADD, "nsds5replicabinddn", a_dn)]
|
|
b.modify_s(rep_dn, mod)
|
|
except ldap.TYPE_OR_VALUE_EXISTS:
|
|
pass
|
|
|
|
def gssapi_update_agreements(self, a, b):
|
|
|
|
self.setup_krb_princs_as_replica_binddns(a, b)
|
|
|
|
#change replication agreements to connect to other host using GSSAPI
|
|
mod = [(ldap.MOD_REPLACE, "nsds5replicatransportinfo", "LDAP"),
|
|
(ldap.MOD_REPLACE, "nsds5replicabindmethod", "SASL/GSSAPI"),
|
|
(ldap.MOD_DELETE, "nsds5replicabinddn", None),
|
|
(ldap.MOD_DELETE, "nsds5replicacredentials", None)]
|
|
|
|
cn, a_ag_dn = self.agreement_dn(b.host)
|
|
a.modify_s(a_ag_dn, mod)
|
|
|
|
cn, b_ag_dn = self.agreement_dn(a.host)
|
|
b.modify_s(b_ag_dn, mod)
|
|
|
|
# Finally remove the temporary replication manager user
|
|
try:
|
|
a.delete_entry(self.repl_man_dn)
|
|
except errors.NotFound:
|
|
pass
|
|
try:
|
|
b.delete_entry(self.repl_man_dn)
|
|
except errors.NotFound:
|
|
pass
|
|
|
|
def delete_agreement(self, hostname, dn=None):
|
|
"""
|
|
Delete a replication agreement.
|
|
|
|
@hostname: the hostname of the agreement to remove
|
|
@dn: optional dn of the agreement to remove
|
|
|
|
For IPA agreements we can easily calculate the DN of the agreement
|
|
to remove. Dogtag agreements are another matter, its agreement
|
|
names depend entirely on where it is created. In this case it is
|
|
better to pass the DN in directly.
|
|
"""
|
|
if dn is None:
|
|
cn, dn = self.agreement_dn(hostname)
|
|
return self.conn.delete_entry(dn)
|
|
|
|
def delete_referral(self, hostname):
|
|
dn = DN(('cn', self.suffix), ('cn', 'mapping tree'), ('cn', 'config'))
|
|
# TODO: should we detect proto/port somehow ?
|
|
mod = [(ldap.MOD_DELETE, 'nsslapd-referral',
|
|
'ldap://%s/%s' % (ipautil.format_netloc(hostname, 389), self.suffix))]
|
|
|
|
try:
|
|
self.conn.modify_s(dn, mod)
|
|
except Exception, e:
|
|
root_logger.debug("Failed to remove referral value: %s" % str(e))
|
|
|
|
def check_repl_init(self, conn, agmtdn, start):
|
|
done = False
|
|
hasError = 0
|
|
attrlist = ['cn', 'nsds5BeginReplicaRefresh',
|
|
'nsds5replicaUpdateInProgress',
|
|
'nsds5ReplicaLastInitStatus',
|
|
'nsds5ReplicaLastInitStart',
|
|
'nsds5ReplicaLastInitEnd']
|
|
entry = conn.get_entry(agmtdn, attrlist)
|
|
if not entry:
|
|
print "Error reading status from agreement", agmtdn
|
|
hasError = 1
|
|
else:
|
|
refresh = entry.single_value.get('nsds5BeginReplicaRefresh')
|
|
inprogress = entry.single_value.get('nsds5replicaUpdateInProgress')
|
|
status = entry.single_value.get('nsds5ReplicaLastInitStatus')
|
|
if not refresh: # done - check status
|
|
if not status:
|
|
print "No status yet"
|
|
elif status.find("replica busy") > -1:
|
|
print "[%s] reports: Replica Busy! Status: [%s]" % (conn.host, status)
|
|
done = True
|
|
hasError = 2
|
|
elif status.find("Total update succeeded") > -1:
|
|
print "\nUpdate succeeded"
|
|
done = True
|
|
elif inprogress.lower() == 'true':
|
|
print "\nUpdate in progress yet not in progress"
|
|
else:
|
|
print "\n[%s] reports: Update failed! Status: [%s]" % (conn.host, status)
|
|
hasError = 1
|
|
done = True
|
|
else:
|
|
now = datetime.datetime.now()
|
|
d = now - start
|
|
sys.stdout.write('\r')
|
|
sys.stdout.write("Update in progress, %d seconds elapsed" % int(d.total_seconds()))
|
|
sys.stdout.flush()
|
|
|
|
return done, hasError
|
|
|
|
def check_repl_update(self, conn, agmtdn):
|
|
done = False
|
|
hasError = 0
|
|
error_message = ''
|
|
attrlist = ['cn', 'nsds5replicaUpdateInProgress',
|
|
'nsds5ReplicaLastUpdateStatus', 'nsds5ReplicaLastUpdateStart',
|
|
'nsds5ReplicaLastUpdateEnd']
|
|
entry = conn.get_entry(agmtdn, attrlist)
|
|
if not entry:
|
|
print "Error reading status from agreement", agmtdn
|
|
hasError = 1
|
|
else:
|
|
inprogress = entry.single_value.get('nsds5replicaUpdateInProgress')
|
|
status = entry.single_value.get('nsds5ReplicaLastUpdateStatus')
|
|
try:
|
|
start = int(entry.single_value['nsds5ReplicaLastUpdateStart'])
|
|
except (ValueError, TypeError, KeyError):
|
|
start = 0
|
|
try:
|
|
end = int(entry.single_value['nsds5ReplicaLastUpdateEnd'])
|
|
except (ValueError, TypeError, KeyError):
|
|
end = 0
|
|
# incremental update is done if inprogress is false and end >= start
|
|
done = inprogress and inprogress.lower() == 'false' and start <= end
|
|
root_logger.info("Replication Update in progress: %s: status: %s: start: %d: end: %d" %
|
|
(inprogress, status, start, end))
|
|
if status: # always check for errors
|
|
# status will usually be a number followed by a string
|
|
# number != 0 means error
|
|
rc, msg = status.split(' ', 1)
|
|
if rc != '0':
|
|
hasError = 1
|
|
error_message = msg
|
|
done = True
|
|
|
|
return done, hasError, error_message
|
|
|
|
def wait_for_repl_init(self, conn, agmtdn):
|
|
done = False
|
|
haserror = 0
|
|
start = datetime.datetime.now()
|
|
while not done and not haserror:
|
|
time.sleep(1) # give it a few seconds to get going
|
|
done, haserror = self.check_repl_init(conn, agmtdn, start)
|
|
print ""
|
|
return haserror
|
|
|
|
def wait_for_repl_update(self, conn, agmtdn, maxtries=600):
|
|
done = False
|
|
haserror = 0
|
|
error_message = ''
|
|
while not done and not haserror and maxtries > 0:
|
|
time.sleep(1) # give it a few seconds to get going
|
|
done, haserror, error_message = self.check_repl_update(conn, agmtdn)
|
|
maxtries -= 1
|
|
if maxtries == 0: # too many tries
|
|
print "Error: timeout: could not determine agreement status: please check your directory server logs for possible errors"
|
|
haserror = 1
|
|
return haserror, error_message
|
|
|
|
def start_replication(self, conn, hostname=None, master=None):
|
|
print "Starting replication, please wait until this has completed."
|
|
if hostname == None:
|
|
hostname = self.conn.host
|
|
cn, dn = self.agreement_dn(hostname, master)
|
|
|
|
mod = [(ldap.MOD_ADD, 'nsds5BeginReplicaRefresh', 'start')]
|
|
conn.modify_s(dn, mod)
|
|
|
|
return self.wait_for_repl_init(conn, dn)
|
|
|
|
def basic_replication_setup(self, conn, replica_id, repldn, replpw):
|
|
assert isinstance(repldn, DN)
|
|
if replpw is not None:
|
|
self.add_replication_manager(conn, repldn, replpw)
|
|
self.replica_config(conn, replica_id, repldn)
|
|
self.setup_changelog(conn)
|
|
|
|
def setup_replication(self, r_hostname, r_port=389, r_sslport=636,
|
|
r_binddn=None, r_bindpw=None,
|
|
is_cs_replica=False, local_port=None):
|
|
assert isinstance(r_binddn, DN)
|
|
if local_port is None:
|
|
local_port = r_port
|
|
# note - there appears to be a bug in python-ldap - it does not
|
|
# allow connections using two different CA certs
|
|
r_conn = ipaldap.IPAdmin(
|
|
r_hostname, port=r_port, cacert=CACERT, protocol='ldap',
|
|
start_tls=True)
|
|
|
|
if r_bindpw:
|
|
r_conn.do_simple_bind(binddn=r_binddn, bindpw=r_bindpw)
|
|
else:
|
|
r_conn.do_sasl_gssapi_bind()
|
|
|
|
#Setup the first half
|
|
l_id = self._get_replica_id(self.conn, r_conn)
|
|
self.basic_replication_setup(self.conn, l_id,
|
|
self.repl_man_dn, self.repl_man_passwd)
|
|
|
|
# Now setup the other half
|
|
r_id = self._get_replica_id(r_conn, r_conn)
|
|
self.basic_replication_setup(r_conn, r_id,
|
|
self.repl_man_dn, self.repl_man_passwd)
|
|
|
|
if is_cs_replica:
|
|
self.setup_agreement(r_conn, self.conn.host, port=local_port,
|
|
repl_man_dn=self.repl_man_dn,
|
|
repl_man_passwd=self.repl_man_passwd,
|
|
master=False)
|
|
self.setup_agreement(self.conn, r_hostname, port=r_port,
|
|
repl_man_dn=self.repl_man_dn,
|
|
repl_man_passwd=self.repl_man_passwd,
|
|
master=True)
|
|
else:
|
|
self.setup_agreement(r_conn, self.conn.host, port=local_port,
|
|
repl_man_dn=self.repl_man_dn,
|
|
repl_man_passwd=self.repl_man_passwd)
|
|
self.setup_agreement(self.conn, r_hostname, port=r_port,
|
|
repl_man_dn=self.repl_man_dn,
|
|
repl_man_passwd=self.repl_man_passwd)
|
|
|
|
#Finally start replication
|
|
ret = self.start_replication(r_conn, master=False)
|
|
if ret != 0:
|
|
raise RuntimeError("Failed to start replication")
|
|
|
|
def setup_winsync_replication(self,
|
|
ad_dc_name, ad_binddn, ad_pwd,
|
|
passsync_pw, ad_subtree,
|
|
cacert=CACERT):
|
|
self.ad_suffix = ""
|
|
try:
|
|
# Validate AD connection
|
|
ad_conn = ldap.initialize('ldap://%s' % ipautil.format_netloc(ad_dc_name))
|
|
# the next one is to workaround bugs arounf opendalp libs+NSS db
|
|
# we need to first specify the OPT_X_TLS_CACERTFILE and _after_
|
|
# that initialize the context to prevent TLS connection errors:
|
|
# https://bugzilla.redhat.com/show_bug.cgi?id=800787
|
|
ad_conn.set_option(ldap.OPT_X_TLS_CACERTFILE, cacert)
|
|
ad_conn.set_option(ldap.OPT_X_TLS_NEWCTX, 0)
|
|
ad_conn.start_tls_s()
|
|
ad_conn.simple_bind_s(str(ad_binddn), ad_pwd)
|
|
res = ad_conn.search_s("", ldap.SCOPE_BASE, '(objectClass=*)',
|
|
['defaultNamingContext'])
|
|
for dn,entry in res:
|
|
if dn == "":
|
|
self.ad_suffix = entry['defaultNamingContext'][0]
|
|
root_logger.info("AD Suffix is: %s" % self.ad_suffix)
|
|
if self.ad_suffix == "":
|
|
raise RuntimeError("Failed to lookup AD's Ldap suffix")
|
|
ad_conn.unbind_s()
|
|
del ad_conn
|
|
except Exception, e:
|
|
root_logger.info("Failed to connect to AD server %s" % ad_dc_name)
|
|
root_logger.info("The error was: %s" % e)
|
|
raise RuntimeError("Failed to setup winsync replication")
|
|
|
|
# Setup the only half.
|
|
# there is no other side to get a replica ID from
|
|
# So we generate one locally
|
|
replica_id = self._get_replica_id(self.conn, self.conn)
|
|
self.basic_replication_setup(self.conn, replica_id,
|
|
self.repl_man_dn, self.repl_man_passwd)
|
|
|
|
#now add a passync user allowed to access the AD server
|
|
self.add_passsync_user(self.conn, passsync_pw)
|
|
self.setup_agreement(self.conn, ad_dc_name,
|
|
repl_man_dn=ad_binddn, repl_man_passwd=ad_pwd,
|
|
iswinsync=True, win_subtree=ad_subtree)
|
|
root_logger.info("Added new sync agreement, waiting for it to become ready . . .")
|
|
cn, dn = self.agreement_dn(ad_dc_name)
|
|
self.wait_for_repl_update(self.conn, dn, 300)
|
|
root_logger.info("Agreement is ready, starting replication . . .")
|
|
|
|
# Add winsync replica to the public DIT
|
|
dn = DN(('cn',ad_dc_name),('cn','replicas'),('cn','ipa'),('cn','etc'), self.suffix)
|
|
entry = self.conn.make_entry(
|
|
dn,
|
|
objectclass=["nsContainer", "ipaConfigObject"],
|
|
cn=[ad_dc_name],
|
|
ipaConfigString=["winsync:%s" % self.hostname],
|
|
)
|
|
|
|
try:
|
|
self.conn.add_entry(entry)
|
|
except Exception, e:
|
|
root_logger.info("Failed to create public entry for winsync replica")
|
|
|
|
#Finally start replication
|
|
ret = self.start_replication(self.conn, ad_dc_name)
|
|
if ret != 0:
|
|
raise RuntimeError("Failed to start replication")
|
|
|
|
def convert_to_gssapi_replication(self, r_hostname, r_binddn, r_bindpw):
|
|
r_conn = ipaldap.IPAdmin(r_hostname, port=PORT, cacert=CACERT)
|
|
if r_bindpw:
|
|
r_conn.do_simple_bind(binddn=r_binddn, bindpw=r_bindpw)
|
|
else:
|
|
r_conn.do_sasl_gssapi_bind()
|
|
|
|
# First off make sure servers are in sync so that both KDCs
|
|
# have all principals and their passwords and can release
|
|
# the right tickets. We do this by force pushing all our changes
|
|
self.force_sync(self.conn, r_hostname)
|
|
cn, dn = self.agreement_dn(r_hostname)
|
|
self.wait_for_repl_update(self.conn, dn, 300)
|
|
|
|
# now in the opposite direction
|
|
self.force_sync(r_conn, self.hostname)
|
|
cn, dn = self.agreement_dn(self.hostname)
|
|
self.wait_for_repl_update(r_conn, dn, 300)
|
|
|
|
# now that directories are in sync,
|
|
# change the agreements to use GSSAPI
|
|
self.gssapi_update_agreements(self.conn, r_conn)
|
|
|
|
def setup_gssapi_replication(self, r_hostname, r_binddn=None, r_bindpw=None):
|
|
"""
|
|
Directly sets up GSSAPI replication.
|
|
Only usable to connect 2 existing replicas (needs existing kerberos
|
|
principals)
|
|
"""
|
|
# note - there appears to be a bug in python-ldap - it does not
|
|
# allow connections using two different CA certs
|
|
r_conn = ipaldap.IPAdmin(r_hostname, port=PORT, cacert=CACERT)
|
|
if r_bindpw:
|
|
r_conn.do_simple_bind(binddn=r_binddn, bindpw=r_bindpw)
|
|
else:
|
|
r_conn.do_sasl_gssapi_bind()
|
|
|
|
# Allow krb principals to act as replicas
|
|
self.setup_krb_princs_as_replica_binddns(self.conn, r_conn)
|
|
|
|
# Create mutual replication agreementsausiung SASL/GSSAPI
|
|
self.setup_agreement(self.conn, r_hostname, isgssapi=True)
|
|
self.setup_agreement(r_conn, self.conn.host, isgssapi=True)
|
|
|
|
def initialize_replication(self, dn, conn):
|
|
mod = [(ldap.MOD_ADD, 'nsds5BeginReplicaRefresh', 'start'),
|
|
(ldap.MOD_REPLACE, 'nsds5ReplicaEnabled', 'on')]
|
|
try:
|
|
conn.modify_s(dn, mod)
|
|
except ldap.ALREADY_EXISTS:
|
|
return
|
|
|
|
def force_sync(self, conn, hostname):
|
|
|
|
newschedule = '2358-2359 0'
|
|
|
|
filter = self.get_agreement_filter(host=hostname)
|
|
try:
|
|
entries = conn.get_entries(
|
|
DN(('cn', 'config')), ldap.SCOPE_SUBTREE, filter)
|
|
except errors.NotFound:
|
|
root_logger.error("Unable to find replication agreement for %s" %
|
|
(hostname))
|
|
raise RuntimeError("Unable to proceed")
|
|
if len(entries) > 1:
|
|
root_logger.error("Found multiple agreements for %s" % hostname)
|
|
root_logger.error("Using the first one only (%s)" % entries[0].dn)
|
|
|
|
dn = entries[0].dn
|
|
schedule = entries[0].single_value.get('nsds5replicaupdateschedule')
|
|
|
|
# On the remote chance of a match. We force a synch to happen right
|
|
# now by setting the schedule to something and quickly removing it.
|
|
if schedule is not None:
|
|
if newschedule == schedule:
|
|
newschedule = '2358-2359 1'
|
|
root_logger.info("Setting agreement %s schedule to %s to force synch" %
|
|
(dn, newschedule))
|
|
mod = [(ldap.MOD_REPLACE, 'nsDS5ReplicaUpdateSchedule', [ newschedule ])]
|
|
conn.modify_s(dn, mod)
|
|
time.sleep(1)
|
|
root_logger.info("Deleting schedule %s from agreement %s" %
|
|
(newschedule, dn))
|
|
mod = [(ldap.MOD_DELETE, 'nsDS5ReplicaUpdateSchedule', None)]
|
|
conn.modify_s(dn, mod)
|
|
|
|
def get_agreement_type(self, hostname):
|
|
cn, dn = self.agreement_dn(hostname)
|
|
|
|
entry = self.conn.get_entry(dn)
|
|
|
|
objectclass = entry.get("objectclass")
|
|
|
|
for o in objectclass:
|
|
if o.lower() == "nsdswindowsreplicationagreement":
|
|
return WINSYNC
|
|
|
|
return IPA_REPLICA
|
|
|
|
def replica_cleanup(self, replica, realm, force=False):
|
|
"""
|
|
This function removes information about the replica in parts
|
|
of the shared tree that expose it, so clients stop trying to
|
|
use this replica.
|
|
"""
|
|
|
|
err = None
|
|
|
|
if replica == self.hostname:
|
|
raise RuntimeError("Can't cleanup self")
|
|
|
|
# delete master kerberos key and all its svc principals
|
|
try:
|
|
entries = self.conn.get_entries(
|
|
self.suffix, ldap.SCOPE_SUBTREE,
|
|
filter='(krbprincipalname=*/%s@%s)' % (replica, realm))
|
|
if entries:
|
|
entries.sort(key=len, reverse=True)
|
|
for entry in entries:
|
|
self.conn.delete_entry(entry)
|
|
except errors.NotFound:
|
|
pass
|
|
except Exception, e:
|
|
if not force:
|
|
raise e
|
|
else:
|
|
err = e
|
|
|
|
# remove replica memberPrincipal from s4u2proxy configuration
|
|
dn1 = DN(('cn', 'ipa-http-delegation'), api.env.container_s4u2proxy, self.suffix)
|
|
member_principal1 = "HTTP/%(fqdn)s@%(realm)s" % dict(fqdn=replica, realm=realm)
|
|
|
|
dn2 = DN(('cn', 'ipa-ldap-delegation-targets'), api.env.container_s4u2proxy, self.suffix)
|
|
member_principal2 = "ldap/%(fqdn)s@%(realm)s" % dict(fqdn=replica, realm=realm)
|
|
|
|
dn3 = DN(('cn', 'ipa-cifs-delegation-targets'), api.env.container_s4u2proxy, self.suffix)
|
|
member_principal3 = "cifs/%(fqdn)s@%(realm)s" % dict(fqdn=replica, realm=realm)
|
|
|
|
for (dn, member_principal) in ((dn1, member_principal1),
|
|
(dn2, member_principal2),
|
|
(dn3, member_principal3)):
|
|
try:
|
|
mod = [(ldap.MOD_DELETE, 'memberPrincipal', member_principal)]
|
|
self.conn.modify_s(dn, mod)
|
|
except (ldap.NO_SUCH_OBJECT, ldap.NO_SUCH_ATTRIBUTE):
|
|
root_logger.debug("Replica (%s) memberPrincipal (%s) not found in %s" % \
|
|
(replica, member_principal, dn))
|
|
except Exception, e:
|
|
if not force:
|
|
raise e
|
|
elif not err:
|
|
err = e
|
|
|
|
# delete master entry with all active services
|
|
try:
|
|
dn = DN(('cn', replica), ('cn', 'masters'), ('cn', 'ipa'),
|
|
('cn', 'etc'), self.suffix)
|
|
entries = self.conn.get_entries(dn, ldap.SCOPE_SUBTREE)
|
|
if entries:
|
|
entries.sort(key=len, reverse=True)
|
|
for entry in entries:
|
|
self.conn.delete_entry(entry)
|
|
except errors.NotFound:
|
|
pass
|
|
except Exception, e:
|
|
if not force:
|
|
raise e
|
|
elif not err:
|
|
err = e
|
|
|
|
try:
|
|
basedn = DN(('cn', 'etc'), self.suffix)
|
|
filter = '(dnaHostname=%s)' % replica
|
|
entries = self.conn.get_entries(
|
|
basedn, ldap.SCOPE_SUBTREE, filter=filter)
|
|
if len(entries) != 0:
|
|
for entry in entries:
|
|
self.conn.delete_entry(entry)
|
|
except errors.NotFound:
|
|
pass
|
|
except Exception, e:
|
|
if not force:
|
|
raise e
|
|
elif not err:
|
|
err = e
|
|
|
|
try:
|
|
dn = DN(('cn', 'default'), ('ou', 'profile'), self.suffix)
|
|
ret = self.conn.get_entry(dn)
|
|
srvlist = ret.single_value.get('defaultServerList', '')
|
|
srvlist = srvlist[0].split()
|
|
if replica in srvlist:
|
|
srvlist.remove(replica)
|
|
attr = ' '.join(srvlist)
|
|
mod = [(ldap.MOD_REPLACE, 'defaultServerList', attr)]
|
|
self.conn.modify_s(dn, mod)
|
|
except errors.NotFound:
|
|
pass
|
|
except ldap.NO_SUCH_ATTRIBUTE:
|
|
pass
|
|
except ldap.TYPE_OR_VALUE_EXISTS:
|
|
pass
|
|
except Exception, e:
|
|
if force and err:
|
|
raise err #pylint: disable=E0702
|
|
else:
|
|
raise e
|
|
|
|
if err:
|
|
raise err #pylint: disable=E0702
|
|
|
|
def set_readonly(self, readonly, critical=False):
|
|
"""
|
|
Set the database readonly status.
|
|
|
|
@readonly: boolean for read-only status
|
|
@critical: boolean to raise an exception on failure, default False.
|
|
"""
|
|
dn = DN(('cn', 'userRoot'), ('cn', 'ldbm database'),
|
|
('cn', 'plugins'), ('cn', 'config'))
|
|
|
|
mod = [(ldap.MOD_REPLACE, 'nsslapd-readonly', 'on' if readonly else 'off')]
|
|
try:
|
|
self.conn.modify_s(dn, mod)
|
|
except ldap.INSUFFICIENT_ACCESS, e:
|
|
# We can't modify the read-only status on the remote server.
|
|
# This usually isn't a show-stopper.
|
|
if critical:
|
|
raise e
|
|
root_logger.debug("No permission to modify replica read-only status, continuing anyway")
|
|
|
|
def cleanallruv(self, replicaId):
|
|
"""
|
|
Create a CLEANALLRUV task and monitor it until it has
|
|
completed.
|
|
"""
|
|
root_logger.debug("Creating CLEANALLRUV task for replica id %d" % replicaId)
|
|
|
|
dn = DN(('cn', 'clean %d' % replicaId), ('cn', 'cleanallruv'),('cn', 'tasks'), ('cn', 'config'))
|
|
e = self.conn.make_entry(
|
|
dn,
|
|
{
|
|
'objectclass': ['top', 'extensibleObject'],
|
|
'cn': ['clean %d' % replicaId],
|
|
'replica-base-dn': [api.env.basedn],
|
|
'replica-id': [replicaId],
|
|
}
|
|
)
|
|
try:
|
|
self.conn.add_entry(e)
|
|
except errors.DuplicateEntry:
|
|
print "CLEANALLRUV task for replica id %d already exists." % replicaId
|
|
else:
|
|
print "Background task created to clean replication data. This may take a while."
|
|
|
|
print "This may be safely interrupted with Ctrl+C"
|
|
|
|
wait_for_task(self.conn, dn)
|
|
|
|
def abortcleanallruv(self, replicaId):
|
|
"""
|
|
Create a task to abort a CLEANALLRUV operation.
|
|
"""
|
|
root_logger.debug("Creating task to abort a CLEANALLRUV operation for replica id %d" % replicaId)
|
|
|
|
dn = DN(('cn', 'abort %d' % replicaId), ('cn', 'abort cleanallruv'),('cn', 'tasks'), ('cn', 'config'))
|
|
e = self.conn.make_entry(
|
|
dn,
|
|
{
|
|
'replica-base-dn': [api.env.basedn],
|
|
'replica-id': [replicaId],
|
|
'objectclass': ['top', 'extensibleObject'],
|
|
'cn': ['abort %d' % replicaId],
|
|
}
|
|
)
|
|
try:
|
|
self.conn.add_entry(e)
|
|
except errors.DuplicateEntry:
|
|
print "An abort CLEANALLRUV task for replica id %d already exists." % replicaId
|
|
else:
|
|
print "Background task created. This may take a while."
|
|
|
|
print "This may be safely interrupted with Ctrl+C"
|
|
|
|
wait_for_task(self.conn, dn)
|
|
|
|
def get_DNA_range(self, hostname):
|
|
"""
|
|
Return the DNA range on this server as a tuple, (next, max), or
|
|
(None, None) if no range has been assigned yet.
|
|
|
|
Raises an exception on errors reading an entry.
|
|
"""
|
|
entry = self.conn.get_entry(DNA_DN)
|
|
|
|
nextvalue = int(entry.single_value.get("dnaNextValue", 0))
|
|
maxvalue = int(entry.single_value.get("dnaMaxValue", 0))
|
|
|
|
sharedcfgdn = entry.single_value.get("dnaSharedCfgDN")
|
|
if sharedcfgdn is not None:
|
|
sharedcfgdn = DN(sharedcfgdn)
|
|
|
|
shared_entry = self.conn.get_entry(sharedcfgdn)
|
|
remaining = int(shared_entry.single_value.get("dnaRemainingValues", 0))
|
|
else:
|
|
remaining = 0
|
|
|
|
if nextvalue == 0 and maxvalue == 0:
|
|
return (None, None)
|
|
|
|
# Check the magic values for an unconfigured DNA entry
|
|
if maxvalue == 1100 and nextvalue == 1101 and remaining == 0:
|
|
return (None, None)
|
|
else:
|
|
return (nextvalue, maxvalue)
|
|
|
|
def get_DNA_next_range(self, hostname):
|
|
"""
|
|
Return the DNA "on-deck" range on this server as a tuple, (next, max),
|
|
or
|
|
(None, None) if no range has been assigned yet.
|
|
|
|
Raises an exception on errors reading an entry.
|
|
"""
|
|
entry = self.conn.get_entry(DNA_DN)
|
|
|
|
range = entry.single_value.get("dnaNextRange")
|
|
|
|
if range is None:
|
|
return (None, None)
|
|
|
|
try:
|
|
(next, max) = range.split('-')
|
|
except ValueError:
|
|
# Should not happen, malformed entry, return nothing.
|
|
return (None, None)
|
|
|
|
return (int(next), int(max))
|
|
|
|
def save_DNA_next_range(self, next_start, next_max):
|
|
"""
|
|
Save a DNA range into the on-deck value.
|
|
|
|
This adds a dnaNextRange value to the DNA configuration. This
|
|
attribute takes the form of start-next.
|
|
|
|
Returns True on success.
|
|
Returns False if the range is already defined.
|
|
Raises an exception on failure.
|
|
"""
|
|
entry = self.conn.get_entry(DNA_DN)
|
|
|
|
range = entry.single_value.get("dnaNextRange")
|
|
|
|
if range is not None and next_start != 0 and next_max != 0:
|
|
return False
|
|
|
|
if next_start == 0 and next_max == 0:
|
|
entry["dnaNextRange"] = None
|
|
else:
|
|
entry["dnaNextRange"] = "%s-%s" % (next_start, next_max)
|
|
|
|
self.conn.update_entry(entry)
|
|
|
|
return True
|
|
|
|
def save_DNA_range(self, next_start, next_max):
|
|
"""
|
|
Save a DNA range.
|
|
|
|
This is potentially very dangerous.
|
|
|
|
Returns True on success. Raises an exception on failure.
|
|
"""
|
|
entry = self.conn.get_entry(DNA_DN)
|
|
|
|
entry["dnaNextValue"] = next_start
|
|
entry["dnaMaxValue"] = next_max
|
|
|
|
self.conn.update_entry(entry)
|
|
|
|
return True
|
|
|
|
def disable_agreement(self, hostname):
|
|
"""
|
|
Disable the replication agreement to hostname.
|
|
"""
|
|
cn, dn = self.agreement_dn(hostname)
|
|
|
|
entry = self.conn.get_entry(dn)
|
|
entry['nsds5ReplicaEnabled'] = 'off'
|
|
|
|
try:
|
|
self.conn.update_entry(entry)
|
|
except errors.EmptyModlist:
|
|
pass
|
|
|
|
def enable_agreement(self, hostname):
|
|
"""
|
|
Enable the replication agreement to hostname.
|
|
|
|
Note: for replication to work it needs to be enabled both ways.
|
|
"""
|
|
cn, dn = self.agreement_dn(hostname)
|
|
|
|
entry = self.conn.get_entry(dn)
|
|
entry['nsds5ReplicaEnabled'] = 'on'
|
|
|
|
try:
|
|
self.conn.update_entry(entry)
|
|
except errors.EmptyModlist:
|
|
pass
|
|
|
|
class CSReplicationManager(ReplicationManager):
|
|
"""ReplicationManager specific to CA agreements
|
|
|
|
Note that in most cases we don't know if we're connecting to an old-style
|
|
separate PKI DS, or to a host with a merged DB.
|
|
Use the get_cs_replication_manager function to determine this and return
|
|
an appropriate CSReplicationManager.
|
|
"""
|
|
|
|
def __init__(self, realm, hostname, dirman_passwd, port):
|
|
super(CSReplicationManager, self).__init__(
|
|
realm, hostname, dirman_passwd, port, starttls=True)
|
|
self.suffix = DN(('o', 'ipaca'))
|
|
self.hostnames = [] # set before calling or agreement_dn() will fail
|
|
|
|
def agreement_dn(self, hostname, master=None):
|
|
"""
|
|
Construct a dogtag replication agreement name. This needs to be much
|
|
more agressive than the IPA replication agreements because the name
|
|
is different on each side.
|
|
|
|
hostname is the local hostname, not the remote one, for both sides
|
|
NOTE: The agreement number is hardcoded in dogtag as well
|
|
|
|
TODO: configurable instance name
|
|
"""
|
|
dn = None
|
|
cn = None
|
|
if self.conn.port == 7389:
|
|
instance_name = 'pki-ca'
|
|
else:
|
|
instance_name = dogtag.configured_constants(api).PKI_INSTANCE_NAME
|
|
|
|
# if master is not None we know what dn to return:
|
|
if master is not None:
|
|
if master is True:
|
|
name = "master"
|
|
else:
|
|
name = "clone"
|
|
cn="%sAgreement1-%s-%s" % (name, hostname, instance_name)
|
|
dn = DN(('cn', cn), self.replica_dn())
|
|
return (cn, dn)
|
|
|
|
for host in self.hostnames:
|
|
for master in ["master", "clone"]:
|
|
try:
|
|
cn="%sAgreement1-%s-%s" % (master, host, instance_name)
|
|
dn = DN(('cn', cn), self.replica_dn())
|
|
self.conn.get_entry(dn)
|
|
return (cn, dn)
|
|
except errors.NotFound:
|
|
dn = None
|
|
cn = None
|
|
|
|
raise errors.NotFound(reason='No agreement found for %s' % hostname)
|
|
|
|
def delete_referral(self, hostname, port):
|
|
dn = DN(('cn', self.suffix), ('cn', 'mapping tree'), ('cn', 'config'))
|
|
entry = self.conn.get_entry(dn)
|
|
try:
|
|
# TODO: should we detect proto somehow ?
|
|
entry['nsslapd-referral'].remove('ldap://%s/%s' %
|
|
(ipautil.format_netloc(hostname, port), self.suffix))
|
|
self.conn.update_entry(entry)
|
|
except Exception, e:
|
|
root_logger.debug("Failed to remove referral value: %s" % e)
|
|
|
|
def has_ipaca(self):
|
|
try:
|
|
entry = self.conn.get_entry(self.suffix)
|
|
except errors.NotFound:
|
|
return False
|
|
else:
|
|
return True
|
|
|
|
def get_cs_replication_manager(realm, host, dirman_passwd):
|
|
"""Get a CSReplicationManager for a remote host
|
|
|
|
Detects if the host has a merged database, connects to appropriate port.
|
|
"""
|
|
|
|
# Try merged database port first. If it has the ipaca tree, return
|
|
# corresponding replication manager
|
|
# If we can't connect to it at all, we're not dealing with an IPA master
|
|
# anyway; let the exception propagate up
|
|
# Fall back to the old PKI-only DS port. Check that it has the ipaca tree
|
|
# (IPA with merged DB theoretically leaves port 7389 free for anyone).
|
|
# If it doesn't, raise exception.
|
|
ports = [
|
|
dogtag.Dogtag10Constants.DS_PORT,
|
|
dogtag.Dogtag9Constants.DS_PORT,
|
|
]
|
|
for port in ports:
|
|
root_logger.debug('Looking for PKI DS on %s:%s' % (host, port))
|
|
replication_manager = CSReplicationManager(
|
|
realm, host, dirman_passwd, port)
|
|
if replication_manager.has_ipaca():
|
|
root_logger.debug('PKI DS found on %s:%s' % (host, port))
|
|
return replication_manager
|
|
else:
|
|
root_logger.debug('PKI tree not found on %s:%s' % (host, port))
|
|
|
|
raise errors.NotFound(reason='Cannot reach PKI DS at %s on ports %s' % (host, ports))
|