Long term component architecture
As the main curator of open standards in Hadoop, Cloudera has a track record of bringing new open source solutions into its platform (such as Apache Spark, Apache HBase, and Apache Parquet) that are eventually adopted by the community at large. As standards, you can build longterm architecture on these components with confidence.
With the exception of DSSD support, Cloudera Enterprise 5.6.0 is identical to CDH 5.5.2/Cloudera Manager 5.5.3 If you do not need DSSD support, you do not need to upgrade if you are already using the latest 5.5.x release.
- System Requirements
- What's New
- Supported Operating Systems
- Supported Databases
- Supported JDK Versions
- Supported Internet Protocol
Supported Operating Systems
|Component||MySQL||SQLite||PostgreSQL||Oracle||Derby - see Note 4|
|Oozie||5.5, 5.6||-||8.4, 9.1, 9.2, 9.3
See Note 2
|Flume||-||-||-||-||Default (for the JDBC Channel only)|
See Note 1
|Default||8.4, 9.1, 9.2, 9.3
See Note 2
See Note 1
|-||8.4, 9.1, 9.2, 9.3
See Note 2
See Note 1
|-||8.4, 9.1, 9.2,, 9.3
See Note 2
|Sqoop 1||See Note 3||-||See Note 3||See Note 3||-|
|Sqoop 2||See Note 4||-||See Note 4||See Note 4||Default|
- MySQL 5.5 is supported on CDH 5.1. MySQL 5.6 is supported on CDH 5.1 and later.
- PostgreSQL 9.2 is supported on CDH 5.1 and later. PostgreSQL 9.3 is supported on CDH 5.2 and later.
- For the purposes of transferring data only, Sqoop 1 supports MySQL 5.0 and above, PostgreSQL 8.4 and above, Oracle 10.2 and above, Teradata 13.10 and above, and Netezza TwinFin 5.0 and above. The Sqoop metastore works only with HSQLDB (1.8.0 and higher 1.x versions; the metastore does not work with any HSQLDB 2.x versions).
- Sqoop 2 can transfer data to and from MySQL 5.0 and above, PostgreSQL 8.4 and above, Oracle 10.2 and above, and Microsoft SQL Server 2012 and above. The Sqoop 2 repository database is supported only on Derby.
- Derby is supported as shown in the table, but not always recommended. See the pages for individual components in the Cloudera Installation and Upgrade guide for recommendations.
Supported JDK Versions
CDH 5 is supported with the versions shown in the table that follows.
Table 1. Supported JDK Versions
|Latest Certified Version||Minimum Supported Version||Exceptions|
Supported Internet Protocol
What's New in CDH 5.2.1
This is a maintenance release that fixes the “POODLE” Vulnerability described below. All CDH 5.2.0 users should upgrade to 5.2.1 as soon as possible.
“POODLE” Vulnerability on SSL/TLS enabled ports
The POODLE (Padding Oracle On Downgraded Legacy Encryption) attack takes advantage of a cryptographic flaw in the obsolete SSLv3 protocol, after first forcing the use of that protocol. The only solution is to disable SSLv3 entirely. This requires changes across a wide variety of components of CDH and Cloudera Manager in 5.2.0 and all earlier versions. CDH 5.2.1 provides these changes for CDH 5.2.0 deployments.
CDH 5.2.1 also fixes the following issues:
- HADOOP-11243 - SSLFactory shouldn't allow SSLv3
- HADOOP-11217 - Disable SSLv3 in KMS
- HADOOP-11156 - DelegateToFileSystem should implement getFsStatus(final Path f).
- HADOOP-11176 - KMSClientProvider authentication fails when both currentUgi and loginUgi are a proxied user
- HDFS-7235 - DataNode#transferBlock should report blocks that don't exist using reportBadBlock
- HDFS-7274 - Disable SSLv3 in HttpFS
- HDFS-7391 - Reenable SSLv2Hello in HttpFS
- HDFS-6781 - Separate HDFS commands from CommandsManual.apt.vm
- HDFS-6831 - Inconsistency between 'hdfs dfsadmin' and 'hdfs dfsadmin -help'
- HDFS-7278 - Add a command that allows sysadmins to manually trigger full block reports from a DN
- YARN-2010 - Handle app-recovery failures gracefully
- YARN-2588 - Standby RM does not transitionToActive if previous transitionToActive is failed with ZK exception.
- YARN-2566 - DefaultContainerExecutor should pick a working directory randomly
- YARN-2641 - Decommission nodes on -refreshNodes instead of next NM-RM heartbeat
- MAPREDUCE-6147 - Support mapreduce.input.fileinputformat.split.maxsize
- HBASE-12376 - HBaseAdmin leaks ZK connections if failure starting watchers (ConnectionLossException)
- HBASE-12201 - Close the writers in the MOB sweep tool
- HBASE-12220 - Add hedgedReads and hedgedReadWins metrics
- HIVE-8693 - Separate out fair scheduler dependency from hadoop 0.23 shim
- HIVE-8634 - HiveServer2 fair scheduler queue mapping doesn't handle the secondary groups rules correctly
- HIVE-8675 - Increase thrift server protocol test coverage
- HIVE-8827 - Remove SSLv2Hello from list of disabled protocols protocols
- HIVE-8615 - beeline csv,tsv outputformat needs backward compatibility mode
- HIVE-8627 - Compute stats on a table from impala caused the table to be corrupted
- HIVE-7764 - Support all JDBC-HiveServer2 authentication modes on a secure cluster cluster
- HIVE-8182 - beeline fails when executing multiple-line queries with trailing spaces
- HUE-2438 - [core] Disable SSLv3 for Poodle vulnerability
- IMPALA-1361: FE Exceptions with BETWEEN predicates
- IMPALA-1397: free local expr allocations in scanner threads
- IMPALA-1400: Window function insert issue (LAG() + OVER)
- IMPALA-1401: raise MAX_PAGE_HEADER_SIZE and use scanner context to stitch together header buffer
- IMPALA-1410: accept "single character" character classes in regex functions
- IMPALA-1411: Create table as select produces incorrect results
- IMPALA-1416 - Queries fail with metastore exception after upgrade and compute stats
- OOZIE-2034 - Disable SSLv3 (POODLEbleed vulnerability)
- OOZIE-2063 - Cron syntax creates duplicate actions
- PARQUET-107 - Add option to disable summary metadata aggregation after MR jobs
- SPARK-3788 - Yarn dist cache code is not friendly to HDFS HA, Federation
- SPARK-3661 - spark.*.memory is ignored in cluster mode
- SPARK-3979 - Yarn backend's default file replication should match HDFS' default one
- SPARK-1720 - use LD_LIBRARY_PATH instead of -Djava.library.path
Want to Get Involved or Learn More?
Check out our other resources
Receive expert Hadoop training through Cloudera University, the industry's only truly dynamic Hadoop training curriculum that’s updated regularly to reflect the state of the art in big data.