Upgrading from CDH 4 to CDH 5 Parcels
Minimum Required Role: Full Administrator
This topic covers upgrading a CDH 4 cluster to a CDH 5 cluster using the upgrade wizard, which will install CDH 5 parcels. Your CDH 4 cluster can be using either parcels or packages; you can use the cluster upgrade wizard to upgrade using parcels in either case.
If you want to upgrade using CDH 5 packages, you can do so using a manual process. See Upgrading from CDH 4 Packages to CDH 5 Packages.
The steps to upgrade a CDH installation managed by Cloudera Manager using parcels are as follows.
- Before You Begin
- Stop All Services
- Perform Service-Specific Prerequisite Actions
- Remove CDH Packages
- Deactivate and Remove the GPL Extras Parcel
- Run the Upgrade Wizard
- Recover from Failed Steps
- Upgrade the GPL Extras Parcel
- Restart the Reports Manager Role
- Recompile JARs
- Finalize the HDFS Metadata Upgrade
- Upgrade Wizard Actions
Before You Begin
- Read the CDH 5 Release Notes.
- Read the Cloudera Manager 5 Release Notes.
- Upgrade to Cloudera Manager 5 before upgrading to CDH 5.
- Ensure Java 1.7 is installed across the cluster. For installation instructions and recommendations, see Upgrading to Oracle JDK 1.7 in a Cloudera Manager Deployment, and make sure you have read Known Issues and Workarounds in Cloudera Manager 5 before you proceed with the upgrade.
- Ensure that the Cloudera Manager minor version is equal to or greater than the CDH
minor version. For example:
Target CDH Version Minimum Cloudera Manager Version 5.0.5 5.0.x 5.1.4 5.1.x 5.4.1 5.4.x - Make sure there are no Oozie workflows in RUNNING or SUSPENDED status; otherwise the Oozie database upgrade will fail and you will have to reinstall CDH 4 to complete or kill those running workflows.
- Delete Symbolic Links in HDFS
If there are symbolic links in HDFS when you upgrade from CDH 4 to CDH 5, the upgrade will fail and you will have to downgrade to CDH 4, delete the symbolic links, and start over. To prevent this, proceed as follows.
- cd to the directory on the NameNode that contains the latest fsimage The location of this directory is specified as the value of dfs.namenode.name.dir (or dfs.name.dir) in hdfs-site.xml.
- Use a command such as the following to write out the path names in the fsimage:
$ hdfs oiv -i FSIMAGE -o /tmp/YYYY-MM-DD_FSIMAGE.txt
- Use a command such as the following to find the path names of any symbolic links listed in /tmp/YYYY-MM-DD_FSIMAGE.txt and write them out to the file /tmp/symlinks.txt:
$ grep -- "->" /tmp/YYYY-MM-DD_FSIMAGE.txt > /tmp/symlinks.txt
- Delete any symbolic links listed in /tmp/symlinks.txt.
- When upgrading from CDH 4 to CDH 5, Oozie upgrade can take a very long time. For upgrades from CDH 4.3 and
higher, you can reduce this time by reducing the amount of history Oozie retains. To reduce Oozie history:
- Go to the Oozie service.
- Click the Configuration tab.
- Click .
- In Oozie Server Advanced Configuration Snippet (Safety Valve) for
oozie-site.xml, enter the following
<property> <name>oozie.service.PurgeService.older.than</name> <value>7</value> </property> <property> <name>oozie.service.PurgeService.purge.limit</name> <value>1000</value> </property>
- For CDH lower than 5.2, enable DEBUG level logging:
- Click .
- Set Oozie Server Logging Threshold to DEBUG.
- Click Save Changes to commit the changes.
- Restart the Oozie Server role.
- Wait for the purge service to run and finish. By default, the service runs every hour. The purge service
emits the following messages in the Oozie server log:
STARTED Purge to purge Workflow Jobs older than [7] days, Coordinator Jobs older than [7] days, and Bundlejobs older than [7] days. ENDED Purge deleted [x] workflows, [y] coordinatorActions, [z] coordinators, [w] bundles
- Revert the purge service and log level settings to the default.
-
When upgrading from CDH 4 to CDH 5, Hue upgrade can take a very long time if the beeswax_queryhistory, beeswax_savedquery, and oozie_job tables are larger than 1000 records. You can reduce the upgrade time by running a script to reduce the size of the Hue database:
- Stop the Hue service.
- Back up the Hue database.
- Download the history cleanup script to the host running the Hue Server.
- Run the following as root:
- parcel installation
export HUE_CONF_DIR="/var/run/cloudera-scm-agent/process/`ls -1 /var/run/cloudera-scm-agent/process | grep HUE| sort -n | tail -1 `" /opt/cloudera/parcels/CDH/share/hue/build/env/bin/hue shell
- package installation
export HUE_CONF_DIR="/var/run/cloudera-scm-agent/process/`ls -1 /var/run/cloudera-scm-agent/process | grep HUE| sort -n | tail -1 `" /usr/share/hue/build/env/bin/hue shell
- parcel installation
- Run the downloaded script in the Hue shell.
-
If Using MySQL as Hue Backend: You may face issues after the upgrade if the default engine for MySQL does not match the engine used by the Hue tables. To confirm the match:
- Open the my.cnf file for MySQL, search for "default-storage-engine" and note its value.
- Connect to MySQL and run the following commands:
use hue; show create table auth_user;
- Search for the "ENGINE=" line and confirm that its value matches the one for the "default-storage-engine" above.
If the default engines do not match, Hue will display a warning on its start-up page (http://$HUE_HOST:$HUE_PORT/about). Work with your database administrator to convert the current Hue MySQL tables to the engine in use by MySQL, as noted by the "default-storage-engine" property.
- Whenever upgrading Impala, whether in CDH or a standalone parcel or package, check your SQL against the newest reserved words listed in incompatible changes. If upgrading across multiple versions or in case of any problems, check against the full list of Impala keywords.
- Run the Host Inspector and fix every issue.
- If using security, run the The Security Inspector.
- Run hdfs fsck / and hdfs dfsadmin -report and fix every issue.
- If using HBase:
- Run hbase hbck.
- Before you can upgrade HBase from CDH 4 to CDH 5, your HFiles must be upgraded from HFile
v1 format to HFile v2, because CDH 5 no longer supports HFile v1. The upgrade procedure itself is different if you are using Cloudera Manager or the command line, but has the same results. The first
step is to check for instances of HFile v1 in the HFiles and mark them to be upgraded to HFile v2, and to check for and report about corrupted files or files with unknown versions, which need to be
removed manually. The next step is to rewrite the HFiles during the next major compaction. After the HFiles are upgraded, you can continue the upgrade. After the upgrade is complete, you must
recompile custom coprocessors and JARs.To check and upgrade the files:
- In the Cloudera Admin Console, go to the HBase service and run .
- Check the output of the command in the stderr log.
Your output should be similar to the following:
Tables Processed: hdfs://localhost:41020/myHBase/.META. hdfs://localhost:41020/myHBase/usertable hdfs://localhost:41020/myHBase/TestTable hdfs://localhost:41020/myHBase/t Count of HFileV1: 2 HFileV1: hdfs://localhost:41020/myHBase/usertable /fa02dac1f38d03577bd0f7e666f12812/family/249450144068442524 hdfs://localhost:41020/myHBase/usertable /ecdd3eaee2d2fcf8184ac025555bb2af/family/249450144068442512 Count of corrupted files: 1 Corrupted Files: hdfs://localhost:41020/myHBase/usertable/fa02dac1f38d03577bd0f7e666f12812/family/1 Count of Regions with HFileV1: 2 Regions to Major Compact: hdfs://localhost:41020/myHBase/usertable/fa02dac1f38d03577bd0f7e666f12812 hdfs://localhost:41020/myHBase/usertable/ecdd3eaee2d2fcf8184ac025555bb2af
In the example above, you can see that the script has detected two HFile v1 files, one corrupt file and the regions to major compact. - Trigger a major compaction on each of the reported regions. This major compaction rewrites the files from
HFile v1 to HFile v2 format. To run the major compaction, start HBase Shell and issue the major_compact command.
$ /usr/lib/hbase/bin/hbase shell hbase> major_compact 'usertable'
You can also do this in a single step by using the echo shell built-in command.$ echo "major_compact 'usertable'" | /usr/lib/hbase/bin/hbase shell
- Review the upgrade procedure and reserve a maintenance window with enough time allotted to perform all steps. For production clusters, Cloudera recommends allocating up to a full day maintenance window to perform the upgrade, depending on the number of hosts, the amount of experience you have with Hadoop and Linux, and the particular hardware you are using.
- To avoid lots of alerts during the upgrade process, you can enable maintenance mode on your cluster before you start the upgrade. This will stop email alerts and SNMP traps from being sent, but will not stop checks and configuration validations from being made. Be sure to exit maintenance mode when you have finished the upgrade to re-enable Cloudera Manager alerts.
Stop All Services
- Stop the cluster.
- On the tab, click to the right of the cluster name and select Stop.
- Click Stop in the confirmation screen. The Command Details window shows the progress of stopping services.
When All services successfully stopped appears, the task is complete and you can close the Command Details window.
- Stop the Cloudera Management Service:
- Do one of the following:
-
- Select .
- Select .
-
- On the Cloudera Management Service and select Stop. tab, click to the right of
-
- Click Stop to confirm. The Command Details window shows the progress of stopping the roles.
- When Command completed with n/n successful subcommands appears, the task is complete. Click Close.
- Do one of the following:
Perform Service-Specific Prerequisite Actions
- Accumulo - if you have installed the Accumulo parcel, deactivate it following the instructions in Managing Parcels.
- HDFS - Back up HDFS metadata on the NameNode:
- Go to the HDFS service.
- Click the Configuration tab.
- In the Search field, search for "NameNode Data Directories" and note the value.
- On the active NameNode host, back up the directory listed in the NameNode Data Directories property. If
more than one is listed, make a backup of one directory, since each directory is a complete copy. For example, if the NameNode data directory is /data/dfs/nn, do the
following as root:
# cd /data/dfs/nn # tar -cvf /root/nn_backup_data.tar .
You should see output like this:
./ ./current/ ./current/fsimage ./current/fstime ./current/VERSION ./current/edits ./image/ ./image/fsimage
If there is a file with the extension lock in the NameNode data directory, the NameNode most likely is still running. Repeat the steps, starting by shutting down the NameNode role.
- Back up the Hive and Sqoop metastore databases.
- For each affected service:
- If not already stopped, stop the service.
- Back up the database. See Backing Up Databases.
- For each affected service:
Remove CDH Packages
- If your Hue service uses the embedded SQLite DB, back up /var/lib/hue/desktop.db to a location that is not /var/lib/hue as this directory is removed when the packages are removed.
- Uninstall the CDH packages. On each host:
Operating System Command RHEL $ sudo yum remove bigtop-jsvc bigtop-utils bigtop-tomcat hue-common sqoop2-client hbase-solr-doc solr-doc SLES $ sudo zypper remove bigtop-jsvc bigtop-utils bigtop-tomcat hue-common sqoop2-client hbase-solr-doc solr-doc Ubuntu or Debian $ sudo apt-get purge bigtop-jsvc bigtop-utils bigtop-tomcat hue-common sqoop2-client hbase-solr-doc solr-doc - Restart all the Cloudera Manager Agents to force an update of the installed binaries reported by the Agent. On each host:
$ sudo service cloudera-scm-agent restart
- Run the Host Inspector to verify that the packages have been removed:
- Click Hosts tab and then click the Host Inspector button.
- When the command completes, click Show Inspector Results.
- If your Hue service uses the embedded SQLite DB, restore the DB you backed up:
- Stop the Hue service.
- Copy the backup from the temporary location to the newly created Hue database directory, /var/lib/hue.
- Start the Hue service.
Deactivate and Remove the GPL Extras Parcel
If you are using LZO, deactivate and remove the CDH 4 GPL Extras parcel.
Run the Upgrade Wizard
- Log into the Cloudera Manager Admin console.
- From the tab, click next to the cluster name and select Upgrade Cluster. The Upgrade Wizard starts.
- If the option to pick between packages and parcels displays, select the Use Parcels option.
- In the Choose CDH Version (Parcels) field, select the CDH version. If there are no qualifying parcels, click the Modify the Remote Parcel Repository URLs link to go to the Parcel Configuration Settings page where you can add the locations of parcel repositories. Click Continue.
- Read the notices for steps you must complete before upgrading, click the Yes, I ... checkboxes after completing the steps, and click Continue.
- Cloudera Manager checks that hosts have the correct software installed. Click Continue.
- The selected parcels are downloaded and distributed. Click Continue.
- The Host Inspector runs and displays the CDH version on the hosts. Click Continue.
- Click Continue. Cloudera Manager performs all service upgrades and restarts the cluster.
- The wizard reports the result of the upgrade. Choose one of the following:
- Leave OK, set up YARN and import existing configuration from my MapReduce service
checked.
- Click Continue to proceed. Cloudera Manager stops the YARN service (if running) and its dependencies.
- Click Continue to proceed. The next page indicates some additional configuration required by YARN.
- Verify or modify the configurations and click Continue. The Switch Cluster to MR2 step proceeds.
- When all steps have completed, click Continue.
- Clear OK, set up YARN and import existing configuration from my MapReduce service.
- Leave OK, set up YARN and import existing configuration from my MapReduce service
checked.
- Click Finish to return to the tab.
- (Optional) Remove the MapReduce service.
- In the MapReduce row, right-click and select Delete. Click Delete to confirm.
Recover from Failed Steps
- If the converting configuration parameters step fails, Cloudera Manager rolls back all configurations to CDH 4. Fix any reported problems and retry the upgrade.
- If the upgrade command fails at any point after the convert configuration step, there is no retry support in Cloudera Manager. You must first correct the error, then manually re-run the individual commands. You can view the remaining commands in the Recent Commands page.
- If the HDFS upgrade metadata step fails, you cannot revert back to CDH 4 unless you restore a backup of Cloudera Manager.
Upgrade the GPL Extras Parcel
- Install the CDH 5 GPL Extras parcel. See Installing the GPL Extras Parcel.
- Reconfigure and restart services that use the parcel. See Configuring Services to Use the GPL Extras Parcel.
Restart the Reports Manager Role
- Do one of the following:
- Select .
- On the Cloudera Management Service table, click the Cloudera Management Service link. tab, in
- Click the Instances tab.
- Check the checkbox next to Reports Manager.
- Select and then Restart to confirm.
Recompile JARs
- MapReduce and YARN - Recompile JARs used in MapReduce applications. For further information, see For MapReduce Programmers: Writing and Running Jobs.
- HBase - Recompile coprocessor and custom JARs used by HBase applications.
Finalize the HDFS Metadata Upgrade
- Go to the HDFS service.
- Click the Instances tab.
- Click the NameNode instance.
- Select and click Finalize Metadata Upgrade to confirm.
Upgrade Wizard Actions
Upgrade HDFS Metadata
- Start the ZooKeeper service.
- Go to the HDFS service.
- Select Upgrade HDFS Metadata to confirm. and click
Upgrade HBase
- Go to the HBase service.
- Select Upgrade HBase to confirm. and click
Upgrade the Hive Metastore Database
- Go to the Hive service.
- Select Upgrade Hive Metastore Database Schema to confirm. and click
- If you have multiple instances of Hive, perform the upgrade on each metastore database.
Upgrade Oozie
- Go to the Oozie service.
- Select Upgrade Database to confirm. and click
- Start the Oozie service.
- Select Install Oozie ShareLib to confirm. and click
Upgrade Sqoop
- Go to the Sqoop service.
- Select Upgrade Sqoop to confirm. and click
Start Cluster Services
- On the tab, click to the right of the cluster name and select Start.
- Click Start that appears in the next screen to confirm. The Command Details window shows the progress of starting services.
When All services successfully started appears, the task is complete and you can close the Command Details window.
Deploy Client Configuration Files
- On the Home page, click to the right of the cluster name and select Deploy Client Configuration.
- Click the Deploy Client Configuration button in the confirmation pop-up that appears.
<< Upgrading CDH 4 to CDH 5 | ©2016 Cloudera, Inc. All rights reserved | Upgrading from CDH 4 Packages to CDH 5 Packages >> |
Terms and Conditions Privacy Policy |