mirror of
https://salsa.debian.org/freeipa-team/freeipa.git
synced 2025-02-25 18:55:28 -06:00
Handle races in replica config
When multiple replicas are installed in parallel, two replicas may try to create the cn=replica entry at the same time. This leads to a conflict on one of the replicas. replica_config() and ensure_replication_managers() now handle conflicts. ipaldap now maps TYPE_OR_VALUE_EXISTS to DuplicateEntry(). The type or value exists exception is raised, when an attribute value or type is already set. Fixes: https://pagure.io/freeipa/issue/7566 Signed-off-by: Christian Heimes <cheimes@redhat.com> Reviewed-By: Thierry Bordaz <tbordaz@redhat.com>
This commit is contained in:
@@ -1029,7 +1029,12 @@ class LDAPClient(object):
|
|||||||
except ldap.NO_SUCH_OBJECT:
|
except ldap.NO_SUCH_OBJECT:
|
||||||
raise errors.NotFound(reason=arg_desc or 'no such entry')
|
raise errors.NotFound(reason=arg_desc or 'no such entry')
|
||||||
except ldap.ALREADY_EXISTS:
|
except ldap.ALREADY_EXISTS:
|
||||||
|
# entry already exists
|
||||||
raise errors.DuplicateEntry()
|
raise errors.DuplicateEntry()
|
||||||
|
except ldap.TYPE_OR_VALUE_EXISTS:
|
||||||
|
# attribute type or attribute value already exists, usually only
|
||||||
|
# occurs, when two machines try to write at the same time.
|
||||||
|
raise errors.DuplicateEntry(message=desc)
|
||||||
except ldap.CONSTRAINT_VIOLATION:
|
except ldap.CONSTRAINT_VIOLATION:
|
||||||
# This error gets thrown by the uniqueness plugin
|
# This error gets thrown by the uniqueness plugin
|
||||||
_msg = 'Another entry with the same attribute value already exists'
|
_msg = 'Another entry with the same attribute value already exists'
|
||||||
|
@@ -34,7 +34,7 @@ import ldap
|
|||||||
from ipalib import api, errors
|
from ipalib import api, errors
|
||||||
from ipalib.cli import textui
|
from ipalib.cli import textui
|
||||||
from ipalib.text import _
|
from ipalib.text import _
|
||||||
from ipapython import ipautil, ipaldap, kerberos
|
from ipapython import ipautil, ipaldap
|
||||||
from ipapython.admintool import ScriptError
|
from ipapython.admintool import ScriptError
|
||||||
from ipapython.dn import DN
|
from ipapython.dn import DN
|
||||||
from ipapython.ipaldap import ldap_initialize
|
from ipapython.ipaldap import ldap_initialize
|
||||||
@@ -461,7 +461,7 @@ class ReplicationManager(object):
|
|||||||
return DN(('cn', 'replica'), ('cn', self.db_suffix),
|
return DN(('cn', 'replica'), ('cn', self.db_suffix),
|
||||||
('cn', 'mapping tree'), ('cn', 'config'))
|
('cn', 'mapping tree'), ('cn', 'config'))
|
||||||
|
|
||||||
def set_replica_binddngroup(self, r_conn, entry):
|
def _set_replica_binddngroup(self, r_conn, entry):
|
||||||
"""
|
"""
|
||||||
Set nsds5replicabinddngroup attribute on remote master's replica entry.
|
Set nsds5replicabinddngroup attribute on remote master's replica entry.
|
||||||
Older masters (ipa < 3.3) may not support setting this attribute. In
|
Older masters (ipa < 3.3) may not support setting this attribute. In
|
||||||
@@ -476,11 +476,6 @@ class ReplicationManager(object):
|
|||||||
mod.append((ldap.MOD_ADD, 'nsds5replicabinddngroup',
|
mod.append((ldap.MOD_ADD, 'nsds5replicabinddngroup',
|
||||||
self.repl_man_group_dn))
|
self.repl_man_group_dn))
|
||||||
|
|
||||||
if 'nsds5replicabinddngroupcheckinterval' not in entry:
|
|
||||||
mod.append(
|
|
||||||
(ldap.MOD_ADD,
|
|
||||||
'nsds5replicabinddngroupcheckinterval',
|
|
||||||
'60'))
|
|
||||||
if mod:
|
if mod:
|
||||||
try:
|
try:
|
||||||
r_conn.modify_s(entry.dn, mod)
|
r_conn.modify_s(entry.dn, mod)
|
||||||
@@ -488,49 +483,62 @@ class ReplicationManager(object):
|
|||||||
logger.debug(
|
logger.debug(
|
||||||
"nsds5replicabinddngroup attribute not supported on "
|
"nsds5replicabinddngroup attribute not supported on "
|
||||||
"remote master.")
|
"remote master.")
|
||||||
|
except (ldap.ALREADY_EXISTS, ldap.CONSTRAINT_VIOLATION):
|
||||||
|
logger.debug("No update to %s necessary", entry.dn)
|
||||||
|
|
||||||
def replica_config(self, conn, replica_id, replica_binddn):
|
def replica_config(self, conn, replica_id, replica_binddn):
|
||||||
assert isinstance(replica_binddn, DN)
|
assert isinstance(replica_binddn, DN)
|
||||||
dn = self.replica_dn()
|
dn = self.replica_dn()
|
||||||
assert isinstance(dn, DN)
|
assert isinstance(dn, DN)
|
||||||
|
|
||||||
|
logger.debug("Add or update replica config %s", dn)
|
||||||
try:
|
try:
|
||||||
entry = conn.get_entry(dn)
|
entry = conn.get_entry(dn)
|
||||||
except errors.NotFound:
|
except errors.NotFound:
|
||||||
pass
|
# no entry, create new one
|
||||||
else:
|
entry = conn.make_entry(
|
||||||
binddns = entry.setdefault('nsDS5ReplicaBindDN', [])
|
dn,
|
||||||
if replica_binddn not in {DN(m) for m in binddns}:
|
objectclass=["top", "nsds5replica", "extensibleobject"],
|
||||||
# Add the new replication manager
|
cn=["replica"],
|
||||||
binddns.append(replica_binddn)
|
nsds5replicaroot=[str(self.db_suffix)],
|
||||||
for key, value in REPLICA_CREATION_SETTINGS.items():
|
nsds5replicaid=[str(replica_id)],
|
||||||
entry[key] = value
|
nsds5replicatype=[self.get_replica_type()],
|
||||||
|
nsds5flags=["1"],
|
||||||
|
nsds5replicabinddn=[replica_binddn],
|
||||||
|
nsds5replicabinddngroup=[self.repl_man_group_dn],
|
||||||
|
nsds5replicalegacyconsumer=["off"],
|
||||||
|
**REPLICA_CREATION_SETTINGS
|
||||||
|
)
|
||||||
try:
|
try:
|
||||||
conn.update_entry(entry)
|
conn.add_entry(entry)
|
||||||
except errors.EmptyModlist:
|
except errors.DuplicateEntry:
|
||||||
pass
|
logger.debug("Lost race against another replica, updating")
|
||||||
|
# fetch entry that have been added by another replica
|
||||||
|
entry = conn.get_entry(dn)
|
||||||
|
else:
|
||||||
|
logger.debug("Added replica config %s", dn)
|
||||||
|
# added entry successfully
|
||||||
|
return entry
|
||||||
|
|
||||||
self.set_replica_binddngroup(conn, entry)
|
# either existing entry or lost race
|
||||||
|
binddns = entry.setdefault('nsDS5ReplicaBindDN', [])
|
||||||
|
if replica_binddn not in {DN(m) for m in binddns}:
|
||||||
|
# Add the new replication manager
|
||||||
|
binddns.append(replica_binddn)
|
||||||
|
|
||||||
# replication is already configured
|
for key, value in REPLICA_CREATION_SETTINGS.items():
|
||||||
return
|
entry[key] = value
|
||||||
|
|
||||||
replica_type = self.get_replica_type()
|
try:
|
||||||
|
conn.update_entry(entry)
|
||||||
|
except errors.EmptyModlist:
|
||||||
|
logger.debug("No update to %s necessary", entry.dn)
|
||||||
|
else:
|
||||||
|
logger.debug("Update replica config %s", entry.dn)
|
||||||
|
|
||||||
entry = conn.make_entry(
|
self._set_replica_binddngroup(conn, entry)
|
||||||
dn,
|
|
||||||
objectclass=["top", "nsds5replica", "extensibleobject"],
|
return entry
|
||||||
cn=["replica"],
|
|
||||||
nsds5replicaroot=[str(self.db_suffix)],
|
|
||||||
nsds5replicaid=[str(replica_id)],
|
|
||||||
nsds5replicatype=[replica_type],
|
|
||||||
nsds5flags=["1"],
|
|
||||||
nsds5replicabinddn=[replica_binddn],
|
|
||||||
nsds5replicabinddngroup=[self.repl_man_group_dn],
|
|
||||||
nsds5replicalegacyconsumer=["off"],
|
|
||||||
**REPLICA_CREATION_SETTINGS
|
|
||||||
)
|
|
||||||
conn.add_entry(entry)
|
|
||||||
|
|
||||||
def setup_changelog(self, conn):
|
def setup_changelog(self, conn):
|
||||||
ent = conn.get_entry(
|
ent = conn.get_entry(
|
||||||
@@ -690,7 +698,10 @@ class ReplicationManager(object):
|
|||||||
uid=["passsync"],
|
uid=["passsync"],
|
||||||
userPassword=[password],
|
userPassword=[password],
|
||||||
)
|
)
|
||||||
conn.add_entry(entry)
|
try:
|
||||||
|
conn.add_entry(entry)
|
||||||
|
except errors.DuplicateEntry:
|
||||||
|
pass
|
||||||
|
|
||||||
# Add the user to the list of users allowed to bypass password policy
|
# Add the user to the list of users allowed to bypass password policy
|
||||||
extop_dn = DN(('cn', 'ipa_pwd_extop'), ('cn', 'plugins'), ('cn', 'config'))
|
extop_dn = DN(('cn', 'ipa_pwd_extop'), ('cn', 'plugins'), ('cn', 'config'))
|
||||||
@@ -1658,7 +1669,10 @@ class ReplicationManager(object):
|
|||||||
objectclass=['top', 'groupofnames'],
|
objectclass=['top', 'groupofnames'],
|
||||||
cn=['replication managers']
|
cn=['replication managers']
|
||||||
)
|
)
|
||||||
conn.add_entry(entry)
|
try:
|
||||||
|
conn.add_entry(entry)
|
||||||
|
except errors.DuplicateEntry:
|
||||||
|
pass
|
||||||
|
|
||||||
def ensure_replication_managers(self, conn, r_hostname):
|
def ensure_replication_managers(self, conn, r_hostname):
|
||||||
"""
|
"""
|
||||||
@@ -1668,23 +1682,24 @@ class ReplicationManager(object):
|
|||||||
On FreeIPA 3.x masters lacking support for nsds5ReplicaBinddnGroup
|
On FreeIPA 3.x masters lacking support for nsds5ReplicaBinddnGroup
|
||||||
attribute, add replica bind DN directly into the replica entry.
|
attribute, add replica bind DN directly into the replica entry.
|
||||||
"""
|
"""
|
||||||
my_princ = kerberos.Principal((u'ldap', unicode(self.hostname)),
|
my_dn = DN(
|
||||||
realm=self.realm)
|
('krbprincipalname', u'ldap/%s@%s' % (self.hostname, self.realm)),
|
||||||
remote_princ = kerberos.Principal((u'ldap', unicode(r_hostname)),
|
api.env.container_service,
|
||||||
realm=self.realm)
|
api.env.basedn
|
||||||
services_dn = DN(api.env.container_service, api.env.basedn)
|
)
|
||||||
|
remote_dn = DN(
|
||||||
mydn, remote_dn = tuple(
|
('krbprincipalname', u'ldap/%s@%s' % (r_hostname, self.realm)),
|
||||||
DN(('krbprincipalname', unicode(p)), services_dn) for p in (
|
api.env.container_service,
|
||||||
my_princ, remote_princ))
|
api.env.basedn
|
||||||
|
)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
conn.get_entry(self.repl_man_group_dn)
|
conn.get_entry(self.repl_man_group_dn)
|
||||||
except errors.NotFound:
|
except errors.NotFound:
|
||||||
self._add_replica_bind_dn(conn, mydn)
|
self._add_replica_bind_dn(conn, my_dn)
|
||||||
self._add_replication_managers(conn)
|
self._add_replication_managers(conn)
|
||||||
|
|
||||||
self._add_dn_to_replication_managers(conn, mydn)
|
self._add_dn_to_replication_managers(conn, my_dn)
|
||||||
self._add_dn_to_replication_managers(conn, remote_dn)
|
self._add_dn_to_replication_managers(conn, remote_dn)
|
||||||
|
|
||||||
def add_temp_sasl_mapping(self, conn, r_hostname):
|
def add_temp_sasl_mapping(self, conn, r_hostname):
|
||||||
@@ -1704,7 +1719,10 @@ class ReplicationManager(object):
|
|||||||
|
|
||||||
entry = conn.get_entry(self.replica_dn())
|
entry = conn.get_entry(self.replica_dn())
|
||||||
entry['nsDS5ReplicaBindDN'].append(replica_binddn)
|
entry['nsDS5ReplicaBindDN'].append(replica_binddn)
|
||||||
conn.update_entry(entry)
|
try:
|
||||||
|
conn.update_entry(entry)
|
||||||
|
except errors.EmptyModlist:
|
||||||
|
pass
|
||||||
|
|
||||||
entry = conn.make_entry(
|
entry = conn.make_entry(
|
||||||
DN(('cn', 'Peer Master'), ('cn', 'mapping'), ('cn', 'sasl'),
|
DN(('cn', 'Peer Master'), ('cn', 'mapping'), ('cn', 'sasl'),
|
||||||
@@ -1716,7 +1734,10 @@ class ReplicationManager(object):
|
|||||||
nsSaslMapFilterTemplate=['(cn=&@%s)' % self.realm],
|
nsSaslMapFilterTemplate=['(cn=&@%s)' % self.realm],
|
||||||
nsSaslMapPriority=['1'],
|
nsSaslMapPriority=['1'],
|
||||||
)
|
)
|
||||||
conn.add_entry(entry)
|
try:
|
||||||
|
conn.add_entry(entry)
|
||||||
|
except errors.DuplicateEntry:
|
||||||
|
pass
|
||||||
|
|
||||||
def remove_temp_replication_user(self, conn, r_hostname):
|
def remove_temp_replication_user(self, conn, r_hostname):
|
||||||
"""
|
"""
|
||||||
|
@@ -419,7 +419,8 @@ class ldap2(CrudBackend, LDAPClient):
|
|||||||
modlist = [(a, b, self.encode(c))
|
modlist = [(a, b, self.encode(c))
|
||||||
for a, b, c in modlist]
|
for a, b, c in modlist]
|
||||||
self.conn.modify_s(str(group_dn), modlist)
|
self.conn.modify_s(str(group_dn), modlist)
|
||||||
except errors.DatabaseError:
|
except errors.DuplicateEntry:
|
||||||
|
# TYPE_OR_VALUE_EXISTS
|
||||||
raise errors.AlreadyGroupMember()
|
raise errors.AlreadyGroupMember()
|
||||||
|
|
||||||
def remove_entry_from_group(self, dn, group_dn, member_attr='member'):
|
def remove_entry_from_group(self, dn, group_dn, member_attr='member'):
|
||||||
|
Reference in New Issue
Block a user