Big Business Value from Big Data and Hadoop

Similar documents
Simplifying the Process of Uploading and Extracting Data from Apache Hadoop

Modernizing Your Data Warehouse with Azure

5th Annual. Cloudera, Inc. All rights reserved.

Angat Pinoy. Angat Negosyo. Angat Pilipinas.

Hadoop Course Content

Bringing the Power of SAS to Hadoop Title

Microsoft Big Data. Solution Brief

E-guide Hadoop Big Data Platforms Buyer s Guide part 1

Guide to Modernize Your Enterprise Data Warehouse How to Migrate to a Hadoop-based Big Data Lake

Introduction to Big Data(Hadoop) Eco-System The Modern Data Platform for Innovation and Business Transformation

BIG DATA AND HADOOP DEVELOPER

Intro to Big Data and Hadoop

Architecting an Open Data Lake for the Enterprise

Big Data The Big Story

Datametica. The Modern Data Platform Enterprise Data Hub Implementations. Why is workload moving to Cloud

Welcome! 2013 SAP AG or an SAP affiliate company. All rights reserved.

ABOUT THIS TRAINING: This Hadoop training will also prepare you for the Big Data Certification of Cloudera- CCP and CCA.

Analytics Platform System

Hadoop Roadmap 2012 A Hortonworks perspective

ORACLE DATA INTEGRATOR ENTERPRISE EDITION

Confidential

Common Customer Use Cases in FSI

NICE Customer Engagement Analytics - Architecture Whitepaper

GET MORE VALUE OUT OF BIG DATA

Hortonworks Data Platform

Hortonworks Connected Data Platforms

Big Data & Hadoop Advance

Cask Data Application Platform (CDAP)

MapR: Converged Data Pla3orm and Quick Start Solu;ons. Robin Fong Regional Director South East Asia

Analytics in Action transforming the way we use and consume information

Outline of Hadoop. Background, Core Services, and Components. David Schwab Synchronic Analytics Nov.

Architecture Optimization for the new Data Warehouse. Cloudera, Inc. All rights reserved.

Real-time Streaming Insight & Time Series Data Analytic For Smart Retail

Redefine Big Data: EMC Data Lake in Action. Andrea Prosperi Systems Engineer

WELCOME TO. Cloud Data Services: The Art of the Possible

AZURE HDINSIGHT. Azure Machine Learning Track Marek Chmel

ENABLING GLOBAL HADOOP WITH DELL EMC S ELASTIC CLOUD STORAGE (ECS)

Exploring Big Data and Data Analytics with Hadoop and IDOL. Brochure. You are experiencing transformational changes in the computing arena.


Spark and Hadoop Perfect Together

Cask Data Application Platform (CDAP) Extensions

Microsoft Azure Essentials

By: Shrikant Gawande (Cloudera Certified )

From Information to Insight: The Big Value of Big Data. Faire Ann Co Marketing Manager, Information Management Software, ASEAN

Modern Data Architecture with Apache Hadoop

ETL challenges on IOT projects. Pedro Martins Head of Implementation

Analytics in the Cloud, Cross Functional Teams, and Apache Hadoop is not a Thing Ryan Packer, Bank of New Zealand

Investor Presentation. Fourth Quarter 2015

Amsterdam. (technical) Updates & demonstration. Robert Voermans Governance architect

Aurélie Pericchi SSP APS Laurent Marzouk Data Insight & Cloud Architect

Hadoop Integration Deep Dive

MapR: Solution for Customer Production Success

Investor Presentation. Second Quarter 2016

Realising Value from Data

Hadoop Stories. Tim Marston. Director, Regional Alliances Page 1. Hortonworks Inc All Rights Reserved

Apache Hadoop in the Datacenter and Cloud

How Data Science is Changing the Way Companies Do Business Colin White

BIG DATA PROCESSING A DEEP DIVE IN HADOOP/SPARK & AZURE SQL DW

SIMPLIFYING BUSINESS ANALYTICS FOR COMPLEX DATA. Davidi Boyarski, Channel Manager

Business is being transformed by three trends

DataAdapt Active Insight

SAP Cloud Platform Big Data Services EXTERNAL. SAP Cloud Platform Big Data Services From Data to Insight

How In-Memory Computing can Maximize the Performance of Modern Payments

SAS and Hadoop Technology: Overview

Cloud Based Analytics for SAP

APAC Big Data & Cloud Summit 2013

StackIQ Enterprise Data Reference Architecture

Louis Bodine IBM STG WW BAO Tiger Team Leader

Hortonworks Powering the Future of Data

Spotlight Sessions. Nik Rouda. Director of Product Marketing Cloudera, Inc. All rights reserved. 1

Enterprise Analytics Accelerating Your Path to Value with an Open Analytics Platform

Table of Contents. Are You Ready for Digital Transformation? page 04. Take Advantage of This Big Data Opportunity with Cisco and Hortonworks page 06

Advancing Information Management and Analysis with Entity Resolution. Whitepaper ADVANCING INFORMATION MANAGEMENT AND ANALYSIS WITH ENTITY RESOLUTION

Building a Single Source of Truth across the Enterprise An Integrated Solution

Datametica DAMA. The Modern Data Platform Enterprise Data Hub Implementations. What is happening with Hadoop Why is workload moving to Cloud

The Intersection of Big Data and DB2

Five Advances in Analytics

Trifacta Data Wrangling for Hadoop: Accelerating Business Adoption While Ensuring Security & Governance

The Importance of good data management and Power BI

Optimal Infrastructure for Big Data

DLT AnalyticsStack. Powering big data, analytics and data science strategies for government agencies

Engaging in Big Data Transformation in the GCC

TechArch Day Digital Decoupling. Oscar Renalias. Accenture

Simplifying Your Modern Data Architecture Footprint

MapR Pentaho Business Solutions

THE CIO GUIDE TO BIG DATA ARCHIVING. How to pick the right product?

Charter Global. Digital Solutions and Consulting Services. Digital Solutions. QA Testing

ETL on Hadoop What is Required

Smarter Analytics for Big Data

Big Data. By Michael Covert. April 2012

Insights to HDInsight

IBM Big Data Summit 2012

Why Big Data Matters? Speaker: Paras Doshi

: Boosting Business Returns with Faster and Smarter Data Lakes

SAP Real-time Data Platform 9 th October Matteo Losi Head of Presales and Business Development Italy Italy EMEA

2013 PARTNER CONNECT

20775 Performing Data Engineering on Microsoft HD Insight

Accelerating Your Big Data Analytics. Jeff Healey, Director Product Marketing, HPE Vertica

20775A: Performing Data Engineering on Microsoft HD Insight

SOLUTION SHEET Hortonworks DataFlow (HDF ) End-to-end data flow management and streaming analytics platform

Transcription:

Big Business Value from Big Data and Hadoop Page 1

Topics The Big Data Explosion: Hype or Reality Introduction to Apache Hadoop The Business Case for Big Data Hortonworks Overview & Product Demo Page 2

Big Data: Hype or Reality? Page 3

What is Big Data? What is Big Data? Page 4

Big Data: Changing The Game for Organizations Petabytes BIG DATA Mobile Web Sentiment User Click Stream Transactions + Interactions + Observations = BIG DATA SMS/MMS Speech to Text Terabytes Gigabytes Megabytes WEB Social Interactions & Feeds Web logs Spatial & GPS Coordinates A/B testing Sensors / RFID / Devices Behavioral Targeting CRM Business Data Feeds Dynamic Pricing Segmentation External Demographics Search Marketing Customer Touches ERP User Generated Content Affiliate Networks Purchase detail Support Contacts HD Video, Audio, Images Dynamic Funnels Purchase record Payment record Offer details Offer history Product/Service Logs Increasing Data Variety and Complexity Page 5

Next Generation Data Platform Drivers Organizations will need to become more data driven to compete Business Drivers Enable new business models & drive faster growth (20%+) Find insights for competitive advantage & optimal returns Technical Drivers Data continues to grow exponentially Data is increasingly everywhere and in many formats Legacy solutions unfit for new requirements growth Financial Drivers Cost of data systems, as % of IT spend, continues to grow Cost advantages of commodity hardware & open source Page 6

What is a Data Driven Business? DEFINITION Better use of available data in the decision making process RULE Key metrics derived from data should be tied to goals PROVEN RESULTS Firms that adopt Data-Driven Decision Making have output and productivity that is 5-6% higher than what would be expected given their investments and usage of information technology* 110010100001010011101010100010010100100101001001000010010001001000001000100000100 100100100010000101110000100100010001010010010111101010010001001001010010100100111 1001010010100011111010001001010000010010001010010111101010011001001010010001000111 * Strength in Numbers: How Does Data-Driven Decisionmaking Affect Firm Performance? Brynjolfsson, Hitt and Kim (April 22, 2011) Page 7

Path to Becoming Big Data Driven 4 Simply put, because of big data, managers can measure, and hence know, radically more about their businesses, and directly translate that knowledge into improved decision making & performance. Key Considerations for a Data Driven Business 1. Large web properties were born this way, you may have to adapt a strategy 2. Start with a project tied to a key objective or KPI Don t OVER engineer 3. Make sure your Big Data strategy fits your organization and grow it over time 4. Don t do big data just to do big data you can get lost in all that data - Erik Brynjolfsson and Andrew McAfee Page 8

Wide Range of New Technologies NoSQL In Memory Storage MPP Cloud BigTable In Memory DB Hadoop HBase Pig MapReduce Apache YARN Scale-out Storage Page 9

A Few Definitions of New Technologies NoSQL A broad range of new database management technologies mostly focused on low latency access to large amounts of data. MPP or in database analytics Widely used approach in data warehouses that incorporates analytic algorithms with the data and then distribute processing Cloud The use of computing resources that are delivered as a service over a network Hadoop Open source data management software that combines massively parallel computing and highly-scalable distributed storage Page 10

Big Data: It s About Scale & Structure SCALE (storage & processing) STRUCTURE Traditional Database EDW MPP Analytics NoSQL Hadoop Distribution Standards and structured governance Loosely structured Limited, no data processing processing Processing coupled w/ data Structured data types Multi and unstructured A JDBC connector to Hive does not make you big data Page 11

Hadoop Market: Growing & Evolving Big data outranks virtualization as #1 trend driving spending initiatives Barclays CIO survey April 2012 Industry Applications, $12 Unstructured Data, $6 Hadoop, $14 NoSQL, $2 Overall market at $100B Enterprise Applications, $5 Hadoop 2 nd only to RDBMS in potential Storage for DB, $9 Estimates put market growth > 40% CAGR IDC expects Big Data tech and services market to grow to $16.9B in 2015 According to JPMC 50% of Big Data market will be influenced by Hadoop Server for DB, $12 Data Integration & Quality, $4 Business Intelligence, $13 RDBMS, $27 Source: Big Data II New Themes, New Opportunities, BofA Merrill Lynch, March 2012 Page 12

Hadoop: Poised for Rapid Growth relative % customers Orgs looking for use cases & ref arch Ecosystem evolving to create a pull market Enterprises endure 1-3 year adoption cycle Innovators, technology enthusiasts Early adopters, visionaries The CHASM Early majority, pragmatists Late majority, conservatives Laggards, Skeptics time Customers want technology & performance Customers want solutions & convenience Source: Geoffrey Moore - Crossing the Chasm Page 13

What s Needed to Drive Success? Enterprise tooling to become a complete data platform Open deployment & provisioning Higher quality data loading Monitoring and management APIs for easy integration Ecosystem needs support & development www.hortonworks.com/moore Existing infrastructure vendors need to continue to integrate Apps need to continue to be developed on this infrastructure Well defined use cases and solution architectures need to be promoted Market needs to rally around core Apache Hadoop To avoid splintering/market distraction To accelerate adoption Page 14

What is Hadoop? Page 15

What is Apache Hadoop? Open source data management software that combines massively parallel computing and highly-scalable distributed storage to provide high performance computation across large amounts of data Page 16

Big Data Driven Means a New Data Platform Facilitate data capture Collect data from all sources - structured and unstructured data At all speeds batch, asynchronous, streaming, real-time Comprehensive data processing Transform, refine, aggregate, analyze, report Simplified data exchange Interoperate with enterprise data systems for query and processing Share data with analytic applications Data between analyst Easy to operate platform Provision, monitor, diagnose, manage at scale Reliability, availability, affordability, scalability, interoperability Page 17

Enterprise Data Platform Requirements A data platform is an integrated set of components that allows you to capture, process and share data in any format, at scale Capture Store and integrate data sources in real time and batch Process Process and manage data of any size and includes tools to sort, filter, summarize and apply basic functions to the data Exchange Open and promote the exchange of data with new & existing external applications Operate Includes tools to manage and operate and gain insight into performance and assure availability Page 18

Apache Hadoop Open Source data management with scale-out storage & processing Processing Storage MapReduce Splits a task across processors near the data & assembles results Distributed across nodes HDFS Self-healing, high bandwidth clustered storage Natively redundant NameNode tracks locations Page 19

Apache Hadoop Characteristics Scalable Efficiently store and process petabytes of data Linear scale driven by additional processing and storage Reliable Redundant storage Failover across nodes and racks Flexible Store all types of data in any format Apply schema on analysis and sharing of the data Economical Use commodity hardware Open source software guards against vendor lock-in Page 20

Tale of the Tape Relational VS. Hadoop Required on write Reads are fast Standards and structured Limited, no data processing Structured schema speed governance processing data types Required on read Writes are fast Loosely structured Processing coupled with data Multi and unstructured Interactive OLAP Analytics Complex ACID Transactions Operational Data Store best fit use Data Discovery Processing unstructured data Massive Storage/Processing Page 21

What is a Hadoop Distribution A complimentary set of open source technologies that make up a complete data platform Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA Tested and pre-packaged to ease installation and usage Collects the right versions of the components that all have different release cycles and ensures they work together Page 22

Apache Hadoop Related Projects Process Capture Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA WebHDFS Operate Exchange REST API that supports the complete FileSystem interface for HDFS. Move data in and out and delete from HDFS Perform file and directory functions webhdfs://<host>:<http PORT>/PATH Standard and included in version 1.0 of Hadoop Page 23

Apache Hadoop Related Projects Process Capture Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA Apache Sqoop Operate Exchange Sqoop is a set of tools that allow non-hadoop data stores to interact with traditional relational databases and data warehouses. A series of connectors have been created to support explicit sources such as Oracle & Teradata It moves data in and out of Hadoop SQ-OOP: SQL to Hadoop Page 24

Apache Hadoop Related Projects Process Capture Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA Apache Flume Operate Exchange Distributed service for efficiently collecting, aggregating, and moving streams of log data into HDFS Streaming capability with many failover and recovery mechanisms Often used to move web log files directly into Hadoop Page 25

Apache Hadoop Related Projects Process Capture Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA Apache HBase Operate Exchange HBase is a non-relational database. It is columnar and provides fault-tolerant storage and quick access to large quantities of sparse data. It also adds transactional capabilities to Hadoop, allowing users to conduct updates, inserts and deletes. Page 26

Apache Hadoop Related Projects Process Capture Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA Apache Pig Operate Exchange Apache Pig allows you to write complex map reduce transformations using a simple scripting language. Pig latin (the language) defines a set of transformations on a data set such as aggregate, join and sort among others. Pig Latin is sometimes extended using UDF (User Defined Functions), which the user can write in Java and then call directly from the language. Page 27

Apache Hadoop Related Projects Process Capture Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA Apache Hive Operate Exchange Apache Hive is a data warehouse infrastructure built on top of Hadoop (originally by Facebook) for providing data summarization, ad-hoc query, and analysis of large datasets. It provides a mechanism to project structure onto this data and query the data using a SQL-like language called HiveQL (HQL). Page 28

Apache Hadoop Related Projects Process Capture Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA Apache HCatalog Operate Exchange HCatalog is a metadata management service for Apache Hadoop. It opens up the platform and allows interoperability across data processing tools such as Pig, Map Reduce and Hive. It also provides a table abstraction so that users need not be concerned with where or how their data is stored. Page 29

Apache Hadoop Related Projects Process Capture Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA Apache Ambari Operate Exchange Ambari is a monitoring, administration and lifecycle management project for Apache Hadoop clusters It provides a mechanism to provisions nodes Operationalizes Hadoop. Page 30

Apache Hadoop Related Projects Process Capture Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA Apache Oozie Exchange Oozie coordinates jobs written in multiple languages such as Map Reduce, Pig and Hive. It is a workflow system that links these jobs and allows specification of order and dependencies between them. Operate Page 31

Apache Hadoop Related Projects Process Capture Talend WebHDFS Sqoop Flume HCatalog Pig Hive HBase MapReduce HDFS Ambari Oozie ZooKeeper HA Apache ZooKeeper Exchange ZooKeeper is a centralized service for maintaining configuration information, naming, providing distributed synchronization, and providing group services. Operate Page 32

Using Hadoop out of the box 1. Provision your cluster w/ Ambari 2. Load your data into HDFS WebHDFS, distcp, Java 3. Express queries in Pig and Hive Shell or scripts 4. Write some UDFs and MR-jobs by hand 5. Graduate to using data scientists 6. Scale your cluster and your analytics Integrate to the rest of your enterprise data silos (TOPIC OF TODAY S DISCUSSION) Page 33

Hadoop in Enterprise Data Architectures Existing Business Infrastructure Web New Tech IDE & Dev Tools ODS & Datamarts Applications & Spreadsheets Visualization & Intelligence Web Applications Datameer Tableau Karmasphere Splunk Operations Discovery Tools EDW Low Latency/ NoSQL Custom Existing Templeton WebHDFS Sqoop HCatalog Pig Hive MapReduce Ambari Oozie ZooKeeper Flume HBase HDFS HA CRM ERP financials Social Media Exhaust Data logs files Big Data Sources (transactions, observations, interactions) Page 34

The Business Case for Hadoop Page 35

Apache Hadoop: Patterns of Use Big Data Transactions + Interactions + Observations Refine Explore Enrich $ Business Case $ Page 36

Operational Data Refinery Hadoop as platform for ETL modernization Refine Explore Enrich Unstructured Log files Capture and archive Parse & Cleanse Structure and join Upload Enterprise Data Warehouse DB data Refinery Capture Capture new unstructured data along with log files all alongside existing sources Retain inputs in raw form for audit and continuity purposes Process Parse the data & cleanse Apply structure and definition Join datasets together across disparate data sources Exchange Push to existing data warehouse for downstream consumption Feeds operational reporting and online systems Page 37

Big Bank Data Refinery Pipeline various upstream apps and feeds - EBCDIC - weblogs - relational feeds Empower scientists and systems - Best of breed exploration platforms - plus existing TD warehouse 10k files per day GPFS Veritca Aster Palantir Teradata transform to Hcatalog datatypes 1... 3-5 years retention.... Detect only changed records and upload to EDW... n HDFS (scalable-clustered storage) Page 38

Big Bank Data Refinery Key Benefits Capture Retain 3 5 years instead of 2 10 days Lower costs Improved compliance Process Turn upstream raw dumps into small list of new, update, delete customer records Convert fixed-width EBCDIC to UTF-8 (Java and DB compatible) Turn raw weblogs into sessions and behaviors Exchange Insert into Teradata for downstream as-is reporting and tools Insert into new exploration platform for scientists to play with Page 39

Big Data Exploration & Visualization Hadoop as agile, ad-hoc data mart Refine Explore Enrich Unstructured Log files DB data Optional EDW / Datamart Capture and archive upload Structure and join Categorize into tables JDBC / ODBC Explore Visualization Tools Capture Capture multi-structured data and retain inputs in raw form for iterative analysis Process Parse the data into queryable format Explore & analyze using Hive, Pig, Mahout and other tools to discover value Label data and type information for compatibility and later discovery Pre-compute stats, groupings, patterns in data to accelerate analysis Exchange Use visualization tools to facilitate exploration and find key insights Optionally move actionable insights into EDW or datamart Page 40

Hardware Manufacturer Unlocking Customer Survey Data product satisfaction and customer feedback analysis via Tableau Searchable text fields (Lucene index) Multiple choice survey data weblog traffic 1... Existing Customer Service DB....... n Hadoop cluster Page 41

Hardware Manufacturer Key Benefits Capture Store 10+ survey forms/year for > 3 years Capture text, audio, and systems data in one platform Process Unlock freeform text and audio data Un-anonymize customers Create HCatalog tables customer, survey, freeform text Exchange Visualize natural satisfaction levels and groups Tag customers as happy and report back to CRM database Page 42

Application Enrichment Deliver Hadoop analysis to online apps Refine Explore Enrich Unstructured Log files DB data Capture Capture data that was once too bulky and unmanageable Capture Parse Derive/Filter NoSQL, HBase Low Latency Enrich Scheduled & near real time Process Uncover aggregate characteristics across data Use Hive Pig and Map Reduce to identify patterns Filter useful data from mass streams (Pig) Micro or macro batch oriented schedules Online Applications Exchange Push results to HBase or other NoSQL alternative for real time delivery Use patterns to deliver right content/offer to the right person at the right time Page 43

Clothing Retailer Custom Homepages anonymous shopper cookied shopper Default Homepage Custom Homepage HBase custom homepage data 1... Retail website....... n Hadoop for "cluster analysis" Sales database weblogs Page 44

Clothing Retailer Key Benefits Capture Capture web logs together with sales order history, customer master Process Compute relationships between products over time people who buy shirts eventually need pants Score customer web behavior / sentiment Connect product recommendations to customer sentiment Exchange Load customer recommendations into HBase for rapid website service Page 45

Business Cases of Hadoop Vertical Refine Explore Enrich Social and Web MDM CRM Ad service models Paths Feature usefulness Friends and associations Content recommendations Ad service Retail Loyalty programs Cross-channel customer MDM CRM Referrers Brand and Sentiment Analysis Paths Taxonomic relationships Dynamic Pricing/Targeted Offer Intelligence Threat Identification Person of Interest Discovery Cross Jurisdiction Queries Finance Risk Modeling & Fraud Identification Trade Performance Analytics Surveillance and Fraud Detection Customer Risk Analysis Real-time upsell, cross sales marketing offers Energy Smart Grid: Production Optimization Grid Failure Prevention Smart Meters Individual Power Grid Manufacturing Supply Chain Optimization Customer Churn Analysis Dynamic Delivery Replacement parts Healthcare & Payer Electronic Medical Records (EMPI) Clinical Trials Analysis Insurance Premium Determination

Goal: Optimize Outcomes at Scale Media Intelligence Finance Advertising Fraud Retail / Wholesale Manufacturing Healthcare Education Government optimize optimize optimize optimize optimize optimize optimize optimize optimize optimize Content Detection Algorithms Performance Prevention Inventory turns Supply chains Patient outcomes Learning outcomes Citizen services Source: Geoffrey Moore. Hadoop Summit 2012 keynote presentation. Page 47

Customer: UC Irvine Medical Center Optimizing patient outcomes while lowering costs UC Irvine Medical Center is ranked among the nation's best hospitals by U.S. News & World Report for the 12th year More than 400 specialty and primary care physicians Opened in 1976 422-bed medical facility Current system, Epic holds 22 years of patient data, across admissions and clinical information Significant cost to maintain and run system Difficult to access, not-integrated into any systems, stand alone Apache Hadoop sunsets legacy system and augments new electronic medical records 1. Migrate all legacy Epic data to Apache Hadoop Replaced existing ETL and temporary databases with Hadoop resulting in faster more reliable transforms Captures all legacy data not just a subset. Exposes this data to EMR and other applications 2. Eliminate maintenance of legacy system and database licenses $500K in annual savings 3. Integrate data with EMR and clinical front-end Better service with complete patient history provided to admissions and doctors Enable improved research through complete information Page 48

Hortonworks Data Platform Page 49

Hortonworks Vision & Role We believe that by the end of 2015, more than half the world's data will be processed by Apache Hadoop. 1 2 3 4 5 Be diligent stewards of the open source core Be tireless innovators beyond the core Provide robust data platform services & open APIs Enable the ecosystem at each layer of the stack Make the platform enterprise-ready & easy to use Page 50

Enabling Hadoop as Enterprise Viable Platform Applications, Business Tools, Development Tools, Data Movement & Integration, Data Management Systems, Systems Management, Infrastructure Hortonworks Data Platform DEVELOPER Data Platform Services & Open APIs Installation & Configuration, Administration, Monitoring, High Availability, Replication, Multi-tenancy,.. Metadata, Indexing, Search, Security, Management, Data Extract & Load, APIs Page 51

Delivering on the Promise of Apache Hadoop Trusted Open Innovative Stewards of the core Original builders and operators of Hadoop 100+ years development experience Managed every viable Hadoop release HDP built on Hadoop 1.0 100% open platform No POS holdback Open to the community Open to the ecosystem Closely aligned to Apache Hadoop core Innovating current platform with HCatalog, Ambari, HA Innovating future platform with YARN, HA Complete vision for Hadoop-based platform Page 52

Hortonworks Apache Hadoop Expertise Hortonworks represent more than 90 years combined Hadoop development experience 8 of top 15 highest lines of code* contributors to Hadoop of all time and 3 of top 5 Hortonworkers are the builders, operators and core architects of Apache Hadoop Hortonworks engineers spend all of their time improving open source Apache Hadoop. Other Hadoop engineers are going to spend the bulk of their time working on their proprietary software We have noticed more activity over the last year from Hortonworks engineers on building out Apache Hadoop s more innovative features. These include YARN, Ambari and HCatalog.. Page 53

Hortonworks Data Platform Management & Monitoring Svcs (Ambari, Zookeeper) Workflow & Scheduling (Oozie) Non-Relational Database (HBase) Scripting (Pig) Metadata Services (HCatalog) 1 Query (Hive) Distributed Processing (MapReduce) Distributed Storage (HDFS) Hortonworks Data Platform Data Integration Services (HCatalog APIs, WebHDFS, Talend OS4BD, Sqoop, Flume) Delivers enterprise grade functionality on a proven Apache Hadoop distribution to ease management, simplify use and ease integration into the enterprise Simplify deployment to get started quickly and easily Monitor, manage any size cluster with familiar console and tools Only platform to include data integration services to interact with any data source Metadata services opens the platform for integration with existing applications Dependable high availability architecture The only 100% open source data platform for Apache Hadoop Page 54

Demo Page 55

Why Hortonworks Data Platform? ONLY Hortonworks Data Platform provides Tightly aligned to core Apache Hadoop development line - Reduces risk for customers who may add custom coding or projects Enterprise Integration - HCatalog provides scalable, extensible integration point to Hadoop data Most reliable Hadoop distribution - Full stack high availability on v1 delivers the strongest SLA guarantees Multi-tenant scheduling and resource management - Capacity and fair scheduling optimizes cluster resources Integration with operations, eases cluster management - Ambari is the most open/complete operations platform for Hadoop clusters Page 56

Hortonworks Growing Partner Ecosystem Page 57

Hortonworks Support Subscriptions Objective: help organizations to successfully develop and deploy solutions based upon Apache Hadoop Full-lifecycle technical support available Developer support for design, development and POCs Production support for staging and production environments Up to 24x7 with 1-hour response times Delivered by the Apache Hadoop experts Backed by development team that has released every major version of Apache Hadoop since 0.1 Forward-compatibility Hortonworks leadership role helps ensure bug fixes and patches can be included in future versions of Hadoop projects Page 58

Hortonworks Training The expert source for Apache Hadoop training & certification Role-based training for developers, administrators & data analysts Coursework built and maintained by the core Apache Hadoop development team. The right course, with the most extensive and realistic hands-on materials Provide an immersive experience into real-world Hadoop scenarios Public and Private courses available Comprehensive Apache Hadoop Certification Become a trusted and valuable Apache Hadoop expert Page 59

Next Steps? 1 Download Hortonworks Data Platform hortonworks.com/download 2 Use the getting started guide hortonworks.com/get-started 3 Learn more get support Hortonworks Support Expert role based training Course for admins, developers and operators Certification program Custom onsite options hortonworks.com/training Full lifecycle technical support across four service levels Delivered by Apache Hadoop Experts/Committers Forward-compatible hortonworks.com/support Page 60

Thank You! Questions & Answers Follow: @hortonworks Page 61

Q&A Page 62