Skip to main content

Cloudera Manager - Duplicate entry 'zookeeper' for key 'NAME'.

We had recently built a cluster using cloudera API’s and had all the services running on it with Kerberos enabled.
Next we had a requirement to add another kafka cluster to our already exsisting cluster in cloudera manager. Since it is a quick task to get the zookeeper and kafka up and running.
We decided to get this done using the cloudera manager instead of the API’s. But we faced the Duplicate entry 'zookeeper' for key 'NAME' issue as described in the bug below.
I have set up two clusters that share a Cloudera Manger.
The first I set up with the API and created the services with capital letter names, e.g., ZOOKEEPER, HDFS, HIVE.
Now, I add the second cluster using the Wizard.

Add Cluster->Select Hosts->Distribute Parcels->Select base HDFS Cluster install

On the next page i get SQL errros telling that the services i want to add already exist. 
I suspect that the check for existing service names does not include capitalized letters.
Note that renaming the services does not help, as it would only change the DisplayName in the database and not the Name column, 
which is unfortunately also a key column. 

Here the excerpt of the error message (attached full log).
javax.persistence.PersistenceException:org.hibernate.exception.ConstraintViolationException: could not perform addBatch

at AbstractEntityManagerImpl.java line 1387
in org.hibernate.ejb.AbstractEntityManagerImpl convert()
Caused by: java.sql.BatchUpdateException:Duplicate entry 'zookeeper' for key 'NAME'

at PreparedStatement.java line 2024
in com.mysql.jdbc.PreparedStatement executeBatchSerially()
Solutions:
For now solution was to use the API and change the names to other than what we had used earlier. 
Example we used `ZOOKEEPER` in out earlier build, so we changed it to `ZOOKEEPER001` in the new build. 

NOTE: Apparently Cloudera manager does not look for CAPITAL `ZOOKEEPER`, 
so if it did, then CM would generated a different name automatically.
And yes this bug was fixed recently and might take a couple of versions to see it on the stable release.
Issue is that when we deploy our services using API we named them as ZOOKEEPER (all caps) but cloudera manager check for all versions except ‘Capital’.
so it continues to build and fail with Duplicate error. If it detects then it would create a different name automatically.
Since this was not working, current workaround is to deploy the services using Cloudera API using a different name (Currently named as ZOOKEEPER001/KAFKA001) .
Another fix would be to change the API script to change the service name to SERVICE_NAME_, example: ‘ZOOKEEPER_HAD’, ‘ZOOKEEPER_KAF’ or a number ZOOKEEPER_1.

Comments

Post a Comment

Popular posts from this blog

Zabbix History Table Clean Up

Zabbix history table gets really big, and if you are in a situation where you want to clean it up. Then we can do so, using the below steps. Stop zabbix server. Take table backup - just in case. Create a temporary table. Update the temporary table with data required, upto a specific date using epoch . Move old table to a different table name. Move updated (new temporary) table to original table which needs to be cleaned-up. Drop the old table. (Optional) Restart Zabbix Since this is not offical procedure, but it has worked for me so use it at your own risk. Here is another post which will help is reducing the size of history tables - http://zabbixzone.com/zabbix/history-and-trends/ Zabbix Version : Zabbix v2.4 Make sure MySql 5.1 is set with InnoDB as innodb_file_per_table=ON Step 1 Stop the Zabbix server sudo service zabbix-server stop Script. echo "------------------------------------------" echo " 1. Stopping Zabbix Server ...

Installing Zabbix Version 2.4 Offline (Zabbix Server without Internet).

There might be situations where you have a remote/zabbix server which does not have internet connectivity, due to security or other reasons. So we create a custom repo on the remote/zabbix server so that we can install zabbix using rpms Here is how we are planning to do this. Download all the dependency rpms on a machine which has internet connection, using yum-downloadonly or repotrack . Transfer all the rpms to the remote server. Create a repo on the remote server. Update yum configuration. Install. NOTE: This method can be used to install any application, but here we have used zabbix as we had this requirement for a zabbix server. Download dependent rpms . On a machine which has internet connection install the package below. And download all the rpms . Make sure the system are similar (not required to be identical - At-least the OS should be of same version) mkdir /zabbix_rpms yum install yum-downloadonly Downloading all the rpms to location /zabbix_rpms/ ,...

Access Filter in SSSD `ldap_access_filter` [SSSD Access denied / Permission denied ]

Access Filter Setup with SSSD ldap_access_filter (string) If using access_provider = ldap , this option is mandatory. It specifies an LDAP search filter criteria that must be met for the user to be granted access on this host. If access_provider = ldap and this option is not set, it will result in all users being denied access. Use access_provider = allow to change this default behaviour. Example: access_provider = ldap ldap_access_filter = memberOf=cn=allowed_user_groups,ou=Groups,dc=example,dc=com Prerequisites yum install sssd Single LDAP Group Under domain/default in /etc/sssd/sssd.conf add: access_provider = ldap ldap_access_filter = memberOf=cn=Group Name,ou=Groups,dc=example,dc=com Multiple LDAP Groups Under domain/default in /etc/sssd/sssd.conf add: access_provider = ldap ldap_access_filter = (|(memberOf=cn=System Adminstrators,ou=Groups,dc=example,dc=com)(memberOf=cn=Database Users,ou=Groups,dc=example,dc=com)) ldap_access_filter accepts standa...