Automobile manufacturer

Building a Hortonworks Big Data DEV Cluster –
Operation and onboarding of use cases

Starting point

  • Hadoop development clusters distributed across multiple business areas
  • Increased licensing and operational costs
  • Unclear and inappropriate operational processes for development
  • Time-to-market for each use case is lengthy
  • No multi-tenant cluster available
  • Confidential data must not be processed on the existing development clusters

Procedure

  • Sizing, procurement, and installation of a Hortonworks Hadoop cluster
  • Setting up Kerberos on the cluster
  • Planning and execution of tests for all Hadoop components
  • Development of a multi-tenant HDFS and Ranger architecture
  • Development of a unified service and operations model with internal billing
  • Transition and stabilisation phase for 3 use cases

Funktionen/Projektergebnis

  • Tested Hortonworks Hadoop DEV cluster with HDP 2.3.2
  • HDFS layout architecture for the multi-tenant data lake
  • Ranger access control model
  • Agreed service model with standardised processes & RACI responsibilities
  • Sizing sheet for use case evaluation
  • Standardised "Welcome Package" for fast onboarding of each use case on the platform
  • Three migrated use cases

Customer benefits

  • Professional operation of the Hadoop DEV cluster
  • Provision of comprehensive Big Data platform services, such as project support for onboarding, builds, operations, or changes
  • Ensuring high operational quality, even for complex issues, through the integration of data engineering expertise
  • Increased development efficiency through DevOps concepts
  • Cost reduction through the delegation of standard tasks to an operations team (including offshore involvement)
Marco Bruno | Senior Manager / Authorised Officer