X

Cloudera Tutorials

Optimize your time with detailed tutorials that clearly explain the best way to deploy, use, and manage Cloudera products. Login or register below to access all Cloudera tutorials.

Cloudera named a leader in 2022 Gartner® Magic Quadrant™ for Cloud Database Management Systems Get the report

Ready to Get Started?

 

NOTICE

 

As of January 31, 2021, this tutorial references legacy products that no longer represent Cloudera’s current product offerings.

Please visit recommended tutorials:

 

Introduction

In this tutorial, you will verify your sandbox IP is mapped to your desired hostname, your admin password is setup and the services that need to be on are activated.

Prerequisites

  • Map HDP Sandbox IP to hostname, if you need help, reference Learning the Ropes of the HDP Sandbox tutorial section ENVIRONMENT SETUP, go to Map Sandbox IP To Your Desired Hostname In The Hosts File in that tutorial

  • Map CDF Sandbox IP to hostname, if you need help, reference Learning the Ropes of the CDF Sandbox tutorial section ENVIRONMENT SETUP, go to Map Sandbox IP To Your Desired Hostname In The Hosts File in that tutorial

  • Set the Ambari admin password for HDP, if you need help, reference Learning the Ropes of the HDP Sandbox tutorial section Admin Password Reset

  • Set the Ambari admin password for CDF, if you need help, reference Learning the Ropes of the CDF Sandbox tutorial section Learning the Ropes of CDF Sandbox

  • Need to have data present in Druid, refer to Real-Time Event Processing In NiFi, SAM, Schema Registry and SuperSet tutorial to setup the SAM data pipeline to store data into Druid. All you need to do is step 1 through 3.

Outline

Started up all required services for HDF and HDP

If unsure, login to Ambari admin Dashboard

For example, to start Druid, you would do the following.

druid-service-start

After starting Druid and Superset, your Background Operations would look similar:

druid-superset-started.jpg

Setup Druid for Superset

We include reference images of what needs to be started in the data pipeline to get the data into Druid. You should have already done this step, which was pointed out in the prerequisites.

1. In the NiFi canvas http://sandbox-hdf.hortonworks.com:9090/nifi, start the NiFi DataFlow by pressing the green start button in the operate panel.

nifi-trucking-iot-flow.jpg

2. In the SAM canvas http://sandbox-hdf.hortonworks.com:7777/, start the SAM topology by pressing the green start button at the bottom right of the canvas.

sam-trucking-iot-topology

3. In the Superset UI http://sandbox-hdp.hortonworks.com:9089, login with credentials admin/admin, wait about 5 – 10 minutes for Kafka data to be consumed, then periodically, select the Sources dropdown and click on Refresh Druid Metadata. Eventually, the two Druid data sources will appear.

druid_metadata

Summary

Congratulations! Data is now in Druid. We can see the datasources in Superset. We are ready to start creating visualization representations of the data.

Further Reading



Your form submission has failed.

This may have been caused by one of the following:

  • Your request timed out
  • A plugin/browser extension blocked the submission. If you have an ad blocking plugin please disable it and close this message to reload the page.