Before you upgrade a CDH cluster, you need to gather some information, review the limitations and release notes, and run some checks on the cluster. Fill in the My Environment form below to customize your CDH upgrade procedures.
Minimum Required Role: Cluster Administrator (also provided by Full Administrator)
Loading Filters ... 6.0.1 6.0.0 5.15 5.14 5.13 5.12 5.11 5.10 5.9 5.8 5.7 5.6 5.5 5.4 5.3 5.2 5.1 5.0 6.0.1 6.0.0 5.15 5.14 5.13 5.12 5.11 5.10 5.9 5.8 5.7 5.6 5.5 5.4 5.3 5.2 5.1 5.0
Limitations for Upgrading from CDH 5.x to CDH 6.0 or Higher
Preparing to Upgrade CDH
- You must have SSH access to the Cloudera Manager server hosts and be able to log in using the root account or an account that has password-less sudo permission to all the hosts.
- Review the CDH 5 and Cloudera Manager 5 Requirements and Supported Versions Cloudera Enterprise 6 Requirements and Supported Versions for the new versions you are upgrading to. If your hosts require an operating system upgrade, you must perform the upgrade before upgrading CDH. See Upgrading the Operating System.
- Ensure that Java 1.7 or 1.8 is installed across the cluster. Cloudera Manager 6.0.0 and higher and CDH 6.0.0 and higher require Java 1.8. For installation instructions and recommendations, see Upgrading the JDK.
- Review the following documents:
- Review the upgrade procedure and reserve a maintenance window with enough time allotted to perform all steps. For production clusters, Cloudera recommends allocating up to a full day maintenance window to perform the upgrade, depending on the number of hosts, the amount of experience you have with Hadoop and Linux, and the particular hardware you are using.
- If you are upgrading from CDH 5.1 or lower, and use Hive Date partition columns, you might need to update the date format. See Date Partition Columns.
- If the cluster uses Impala, check your SQL against the newest reserved words listed in incompatible changes. If upgrading across multiple versions, or in case of any problems, check against the full list of Impala keywords.
- Run the Security Inspector and fix any reported errors.
- Log in to any cluster node as the hdfs user, run the following commands, and correct any reported errors:
hdfs fsck / -includeSnapshots
hdfs dfsadmin -reportSee HDFS Commands Guide in the Apache Hadoop documentation.
- Log in to any DataNode as the hdfs user, run the following command, and correct any reported errors:
hbase hbckSee Checking Consistency in HBase Tables.
- If the cluster uses Kudu, log in to any cluster host and run the ksck command.
For the full syntax of this command, see Checking Cluster Health with ksck.
- If you have configured Hue to use TLS/SSL and you are upgrading from CDH 5.2 or lower to CDH 5.3 or higher, Hue validates CA certificates and requires a truststore. To create a truststore, follow the instructions in Hue as a TLS/SSL Client.
- If you are upgrading to CDH 6.0 or higher, and Hue is deployed in the cluster, and Hue is using PostgreSQL as its database, you must manually install psycopg2. See Installing Dependencies for Hue.
- If your cluster uses the Flume Kafka client, and you are upgrading to CDH 5.8.0 or CDH 5.8.1, perform the extra steps described in Upgrading to CDH 5.8.0 or CDH 5.8.1 When Using the Flume Kafka Client and then continue with the procedures in this topic.
- If your cluster uses Impala and Llama, this role has been deprecated as of CDH 5.9 and you must remove
the role from the Impala service before starting the upgrade. If you do not remove this role, the upgrade wizard will halt the upgrade.
To determine if Impala uses Llama:
To remove the Llama role:
- Go to the Impala service.
- Select the Instances tab.
- Examine the list of roles in the Role Type column. If Llama appears, the Impala service is using Llama.
- Go to the Impala service and select
The Disable YARN and Impala Integrated Resource Management wizard displays.
- Click Continue.
The Disable YARN and Impala Integrated Resource Management Command page displays the progress of the commands to disable the role.
- When the commands have completed, click Finish.
- If your cluster uses Sentry, and are upgrading from CDH 5.12 or lower, you might need to increase the Java heap memory for Sentry. See Performance Guidelines.
- The following services are no longer supported as of Enterprise 6.0.0:
- Sqoop 2
- MapReduce 1
- Spark 1.6
- Record Service
- Open the Cloudera Manager Admin console and collect the following information about your environment:
- The version of Cloudera Manager. Go to .
- The version of the JDK deployed. Go to .
- The version of CDH and whether the cluster was installed using parcels or packages. It is displayed next to the cluster name on the Home page.
- The services enabled in your cluster.
- Whether HDFS High Availability is enabled.
Go to Clusters click HDFS Service, click Actions menu. It is enabled if you see an menu item Disable High Availability.
- Back up Cloudera Manager before beginning the upgrade. See Backing Up Cloudera Manager.