draft nist big data interoperability framework: volume 5, architectures white paper...

53
NIST Special Publication 1500-5 DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper Survey NIST Big Data Public Working Group Reference Architecture Subgroup Draft Version 1 April 6, 2015 http://dx.doi.org/10.6028/NIST.SP.1500-5

Upload: others

Post on 30-Jul-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

NIST Special Publication 1500-5

DRAFT NIST Big Data Interoperability Framework:

Volume 5, Architectures White Paper Survey

NIST Big Data Public Working Group Reference Architecture Subgroup

Draft Version 1 April 6, 2015

http://dx.doi.org/10.6028/NIST.SP.1500-5

Page 2: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

NIST Special Publication 1500-5 Information Technology Laboratory

DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White

Paper Survey

Draft Version 1

NIST Big Data Public Working Group (NBD-PWG) Reference Architecture Subgroup

National Institute of Standards and Technology Gaithersburg, MD 20899

April 2015

U. S. Department of Commerce

Penny Pritzker, Secretary

National Institute of Standards and Technology Dr. Willie E. May, Under Secretary of Commerce for Standards and Technology and Director

Page 3: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

ii

National Institute of Standards and Technology Special Publication 1500-5 53 pages (April 6, 2015)

Certain commercial entities, equipment, or materials may be identified in this document in order to describe an experimental procedure or concept adequately. Such identification is not intended to imply recommendation or endorsement by NIST, nor is it intended to imply that the entities, materials, or equipment are necessarily the best available for the purpose.

There may be references in this publication to other publications currently under development by NIST in accordance with its assigned statutory responsibilities. The information in this publication, including concepts and methodologies, may be used by Federal agencies even before the completion of such companion publications. Thus, until each publication is completed, current requirements, guidelines, and procedures, where they exist, remain operative. For planning and transition purposes, Federal agencies may wish to closely follow the development of these new publications by NIST.

Organizations are encouraged to review all draft publications during public comment periods and provide feedback to NIST. All NIST Information Technology Laboratory publications, other than the ones noted above, are available at http://www.nist.gov/publication-portal.cfm.

Public comment period: April 6, 2015 through May 21, 2015

Comments on this publication may be submitted to Wo Chang

National Institute of Standards and Technology

Attn: Wo Chang, Information Technology Laboratory 100 Bureau Drive (Mail Stop 8900) Gaithersburg, MD 20899-8930

Email: [email protected]

Page 4: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

iii

Reports on Computer Systems Technology

The Information Technology Laboratory (ITL) at NIST promotes the U.S. economy and public welfare by providing technical leadership for the Nation’s measurement and standards infrastructure. ITL develops tests, test methods, reference data, proof of concept implementations, and technical analyses to advance the development and productive use of information technology. ITL’s responsibilities include the development of management, administrative, technical, and physical standards and guidelines for the cost-effective security and privacy of other than national security-related information in Federal information systems. This document reports on ITL’s research, guidance, and outreach efforts in Information Technology and its collaborative activities with industry, government, and academic organizations.

Abstract

Big Data is a term used to describe the deluge of data in our networked, digitized, sensor-laden, information-driven world. While great opportunities exist with Big Data, it can overwhelm traditional technical approaches and its growth is outpacing scientific and technological advances in data analytics. To advance progress in Big Data, the NIST Big Data Public Working Group (NBD-PWG) is working to develop consensus on important, fundamental questions related to Big Data. The results are reported in the NIST Big Data Interoperability Framework series of volumes. This volume, Volume 5, presents the results of the reference architecture survey. The reviewed reference architectures are described in detail, followed by a summary of the reference architecture comparison.

Keywords

Big Data, Big Data management, Big Data storage, Big Data analytics, application interfaces, Big Data infrastructure, reference architecture, architecture survey

Page 5: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

iv

Acknowledgements

This document reflects the contributions and discussions by the membership of the NBD-PWG, co-chaired by Wo Chang of the NIST Information Technology Laboratory, Robert Marcus of ET-Strategies, and Chaitanya Baru, University of California San Diego Supercomputer Center.

The document contains input from members of the NBD-PWG Reference Architecture Subgroup, led by Orit Levin (Microsoft), Don Krapohl (Augmented Intelligence), and James Ketner (AT&T).

NIST SP1500-5, Version 1 has been collaboratively authored by the NBD-PWG. As of the date of this publication, there are over six hundred NBD-PWG participants from industry, academia, and government. Federal agency participants include the National Archives and Records Administration (NARA), National Aeronautics and Space Administration (NASA), National Science Foundation (NSF), and the U.S. Departments of Agriculture, Commerce, Defense, Energy, Health and Human Services, Homeland Security, Transportation, Treasury, and Veterans Affairs.

NIST would like to acknowledge the specific contributions to this volume by the following NBD-PWG members:

Milind Bhandarkar EMC/Pivotal

Wo Chang National Institute of Standards and Technology

Yuri Demchenko University of Amsterdam

Barry Devlin 9sight Consulting

Harry Foxwell Oracle Press

James Kobielus IBM

Orit Levin Microsoft

Robert Marcus ET-Strategies

Tony Middleton LexisNexis

Sanjay Mishra Verizon

Sanjay Patil SAP

The editors for this document were Sanjay Mishra and Wo Chang.

Page 6: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

v

Notice to Readers

NIST is seeking feedback on the proposed working draft of the NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper Survey. Once public comments are received, compiled, and addressed by the NBD-PWG, and reviewed and approved by NIST internal editorial board, Version 1 of this volume will be published as final. Three versions are planned for this volume, with Versions 2 and 3 building on the first. Further explanation of the three planned versions and the information contained therein is included in Section 1.5 of this document.

Please be as specific as possible in any comments or edits to the text. Specific edits include, but are not limited to, changes in the current text, additional text further explaining a topic or explaining a new topic, additional references, or comments about the text, topics, or document organization. These specific edits can be recorded using one of the two following methods.

1. TRACK CHANGES: make edits to and comments on the text directly into this Word document using track changes

2. COMMENT TEMPLATE: capture specific edits using the Comment Template (http://bigdatawg.nist.gov/_uploadfiles/SP1500-1-to-7_comment_template.docx), which includes space for Section number, page number, comment, and text edits

Submit the edited file from either method 1 or 2 to [email protected] with the volume number in the subject line (e.g., Edits for Volume 5.)

Please contact Wo Chang ([email protected]) with any questions about the feedback submission process.

Big Data professionals continue to be welcome to join the NBD-PWG to help craft the work contained in the volumes of the NIST Big Data Interoperability Framework. Additional information about the NBD-PWG can be found at http://bigdatawg.nist.gov.

Page 7: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

vi

Table of Contents

EXECUTIVE SUMMARY .................................................................................................................................... VIII 

1  INTRODUCTION ......................................................................................................................................... 1 

1.1  BACKGROUND ................................................................................................................................................ 1 1.2  SCOPE AND OBJECTIVES OF THE REFERENCE ARCHITECTURE SUBGROUP ..................................................................... 2 1.3  REPORT PRODUCTION ...................................................................................................................................... 3 1.4  REPORT STRUCTURE ........................................................................................................................................ 3 1.5  FUTURE WORK ON THIS VOLUME ....................................................................................................................... 3 

2  BIG DATA ARCHITECTURE PROPOSALS RECEIVED ....................................................................................... 4 

2.1  BOB MARCUS ................................................................................................................................................ 4 2.1.1  General Architecture Description .................................................................................................. 4 2.1.2  Architecture Model ........................................................................................................................ 4 2.1.3  Key Components ............................................................................................................................ 5 

2.2  MICROSOFT ................................................................................................................................................... 7 2.2.1  General Architecture Description .................................................................................................. 7 2.2.2  Architecture Model ........................................................................................................................ 8 2.2.3  Key Components ............................................................................................................................ 8 

2.3  UNIVERSITY OF AMSTERDAM ............................................................................................................................. 9 2.3.1  General Architecture Description .................................................................................................. 9 2.3.2  Architecture Model ...................................................................................................................... 10 2.3.3  Key Components .......................................................................................................................... 11 

2.4  IBM ........................................................................................................................................................... 11 2.4.1  General Architecture Description ................................................................................................ 11 2.4.2  Architecture Model ...................................................................................................................... 12 2.4.3  Key Components .......................................................................................................................... 14 

2.5  ORACLE ....................................................................................................................................................... 14 2.5.1  General Architecture Description ................................................................................................ 14 2.5.2  Architecture Model ...................................................................................................................... 14 2.5.3  Key Components .......................................................................................................................... 15 

2.6  PIVOTAL ...................................................................................................................................................... 16 2.6.1  General Architecture Description ................................................................................................ 16 2.6.2  Architecture Model ...................................................................................................................... 16 2.6.3  Key Components .......................................................................................................................... 17 

2.7  SAP ........................................................................................................................................................... 19 2.7.1  General Architecture Description ................................................................................................ 19 2.7.2  Architecture Model ...................................................................................................................... 19 2.7.3  Key Components .......................................................................................................................... 20 

2.8  9SIGHT ....................................................................................................................................................... 21 2.8.1  General Architecture Description ................................................................................................ 21 2.8.2  Architecture Model ...................................................................................................................... 21 2.8.3  Key Components .......................................................................................................................... 21 

2.9  LEXISNEXIS .................................................................................................................................................. 23 2.9.1  General Architecture Description ................................................................................................ 23 2.9.2  Architecture Model ...................................................................................................................... 23 

Page 8: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

vii

2.9.3  Key Components .......................................................................................................................... 24 

3  SURVEY OF BIG DATA ARCHITECTURES ...................................................................................................... 26 

3.1  BOB MARCUS .............................................................................................................................................. 26 3.2  MICROSOFT ................................................................................................................................................. 28 3.3  UNIVERSITY OF AMSTERDAM ........................................................................................................................... 29 3.4  IBM ........................................................................................................................................................... 30 3.5  ORACLE ....................................................................................................................................................... 32 3.6  PIVOTAL ...................................................................................................................................................... 32 3.7  SAP ........................................................................................................................................................... 33 3.8  9SIGHT ....................................................................................................................................................... 34 3.9  LEXISNEXIS .................................................................................................................................................. 34 3.10  COMPARATIVE VIEW OF SURVEYED ARCHITECTURES ............................................................................................. 36 

4  CONCLUSIONS .......................................................................................................................................... 40 

APPENDIX A: ACRONYMS ................................................................................................................................ A‐1 

APPENDIX B: REFERENCES ............................................................................................................................... B‐1 

Figures FIGURE 1: COMPONENTS OF THE HIGH LEVEL REFERENCE MODEL ............................................................................................... 4 FIGURE 2: DESCRIPTION OF THE COMPONENTS OF THE LOW‐LEVEL REFERENCE MODEL ................................................................... 5 FIGURE 3: BIG DATA ECOSYSTEM REFERENCE ARCHITECTURE ...................................................................................................... 8 FIGURE 4: BIG DATA ARCHITECTURE FRAMEWORK .................................................................................................................. 10 FIGURE 5: IBM BIG DATA PLATFORM .................................................................................................................................. 12 FIGURE 6: HIGH LEVEL, CONCEPTUAL VIEW OF THE INFORMATION MANAGEMENT ECOSYSTEM ....................................................... 15 FIGURE 7: ORACLE BIG DATA REFERENCE ARCHITECTURE ......................................................................................................... 15 FIGURE 8: PIVOTAL ARCHITECTURE MODEL ........................................................................................................................... 16 FIGURE 9: PIVOTAL DATA FABRIC AND ANALYTICS .................................................................................................................. 17 FIGURE 10: SAP BIG DATA REFERENCE ARCHITECTURE ........................................................................................................... 20 FIGURE 11: 9SIGHT GENERAL ARCHITECTURE ........................................................................................................................ 21 FIGURE 12: 9SIGHT ARCHITECTURE MODEL .......................................................................................................................... 22 FIGURE 13: LEXIS NEXIS GENERAL ARCHITECTURE................................................................................................................... 23 FIGURE 14: LEXIS NEXIS HIGH PERFORMANCE COMPUTING CLUSTER ......................................................................................... 24 FIGURE 15: BIG DATA LAYERED ARCHITECTURE ...................................................................................................................... 28 FIGURE 16: DATA DISCOVERY AND EXPLORATION ................................................................................................................... 30 FIGURE 17(A): STACKED VIEW OF SURVEYED ARCHITECTURE .................................................................................................... 37 FIGURE 17(B): STACKED VIEW OF SURVEYED ARCHITECTURE (CONTINUED) ................................................................................. 38 FIGURE 17(C): STACKED VIEW OF SURVEYED ARCHITECTURE (CONTINUED) ................................................................................. 39 FIGURE 18: BIG DATA REFERENCE ARCHITECTURE .................................................................................................................. 41 

Tables TABLE 1: DATABASES AND INTERFACES IN THE LAYERED ARCHITECTURE FROM BOB MARCUS .......................................................... 26 TABLE 2: MICROSOFT DATA TRANSFORMATION STEPS ............................................................................................................. 28 

Page 9: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

viii

Executive Summary 1

This document, NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper 2 Survey, was prepared by the NIST Big Data Public Working Group (NBD-PWG) Reference Architecture 3 Subgroup to facilitate understanding of the operational intricacies in Big Data and to serve as a tool for 4 developing system-specific architectures using a common reference framework. The Subgroup surveyed 5 currently published Big Data platforms by leading companies or individuals supporting the Big Data 6 framework and analyzed the material. This effort revealed a remarkable consistency of Big Data 7 architecture. The most common themes occurring across the architectures surveyed are outlined below. 8

Big Data Management 9

Structured, semi-structured, and unstructured data 10 Velocity, variety, volume, and variability 11 SQL and NoSQL 12 Distributed file system 13

Big Data Analytics 14

Descriptive, predictive, and spatial 15 Real-time 16 Interactive 17 Batch analytics 18 Reporting 19 Dashboard 20

Big Data Infrastructure 21

In-memory data grids 22 Operational database 23 Analytic database 24 Relational database 25 Flat files 26 Content management system 27 Horizontal scalable architecture 28

The NIST Big Data Interoperability Framework consists of seven volumes, each of which addresses a 29 specific key topic, resulting from the work of the NBD-PWG. The seven volumes are as follows: 30

Volume 1, Definitions 31 Volume 2, Taxonomies 32 Volume 3, Use Cases and General Requirements 33 Volume 4, Security and Privacy 34 Volume 5, Architectures White Paper Survey 35 Volume 6, Reference Architecture 36 Volume 7, Standards Roadmap 37

38

Page 10: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

ix

The NIST Big Data Interoperability Framework will be released in three versions, which correspond to 39 the three stages of the NBD-PWG work. The three stages aim to achieve the following: 40

Stage 1: Identify the high-level NIST Big Data Reference Architecture (NBDRA) key components, 41 which are technology, infrastructure, and vendor agnostic 42

Stage 2: Define general interfaces between the NBDRA components 43 Stage 3: Validate the NBDRA by building Big Data general applications through the general interfaces 44

Potential areas of future work for the Subgroup during stage 2 are highlighted in Section 1.5 of this 45 volume. The current effort documented in this volume reflects concepts developed within the rapidly 46 evolving field of Big Data. 47

48

Page 11: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

1

1 INTRODUCTION 49

1.1 BACKGROUND 50

There is broad agreement among commercial, academic, and government leaders about the remarkable 51 potential of Big Data to spark innovation, fuel commerce, and drive progress. Big Data is the common 52 term used to describe the deluge of data in today’s networked, digitized, sensor-laden, and information-53 driven world. The availability of vast data resources carries the potential to answer questions previously 54 out of reach, including the following: 55

How can a potential pandemic reliably be detected early enough to intervene? 56 Can new materials with advanced properties be predicted before these materials have ever been 57

synthesized? 58 How can the current advantage of the attacker over the defender in guarding against cyber-59

security threats be reversed? 60

There is also broad agreement on the ability of Big Data to overwhelm traditional approaches. The growth 61 rates for data volumes, speeds, and complexity are outpacing scientific and technological advances in data 62 analytics, management, transport, and data user spheres. 63

Despite widespread agreement on the inherent opportunities and current limitations of Big Data, a lack of 64 consensus on some important, fundamental questions continues to confuse potential users and stymie 65 progress. These questions include the following: 66

What attributes define Big Data solutions? 67 How is Big Data different from traditional data environments and related applications? 68 What are the essential characteristics of Big Data environments? 69 How do these environments integrate with currently deployed architectures? 70 What are the central scientific, technological, and standardization challenges that need to be 71

addressed to accelerate the deployment of robust Big Data solutions? 72

Within this context, on March 29, 2012, the White House announced the Big Data Research and 73 Development Initiative.1 The initiative’s goals include helping to accelerate the pace of discovery in 74 science and engineering, strengthening national security, and transforming teaching and learning by 75 improving the ability to extract knowledge and insights from large and complex collections of digital 76 data. 77

Six federal departments and their agencies announced more than $200 million in commitments spread 78 across more than 80 projects, which aim to significantly improve the tools and techniques needed to 79 access, organize, and draw conclusions from huge volumes of digital data. The initiative also challenged 80 industry, research universities, and nonprofits to join with the federal government to make the most of the 81 opportunities created by Big Data. 82

Motivated by the White House initiative and public suggestions, the National Institute of Standards and 83 Technology (NIST) has accepted the challenge to stimulate collaboration among industry professionals to 84 further the secure and effective adoption of Big Data. As one result of NIST’s Cloud and Big Data Forum 85 held on January 15–17, 2013, there was strong encouragement for NIST to create a public working group 86 for the development of a Big Data Interoperability Framework. Forum participants noted that this 87 framework should define and prioritize Big Data requirements, including interoperability, portability, 88 reusability, extensibility, data usage, analytics, and technology infrastructure. In doing so, the framework 89 would accelerate the adoption of the most secure and effective Big Data techniques and technology. 90

Page 12: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

2

On June 19, 2013, the NIST Big Data Public Working Group (NBD-PWG) was launched with extensive 91 participation by industry, academia, and government from across the nation. The scope of the NBD-PWG 92 involves forming a community of interests from all sectors—including industry, academia, and 93 government—with the goal of developing consensus on definitions, taxonomies, secure reference 94 architectures, security and privacy, andfrom thesea standardsroadmap. Such a consensus would 95 create a vendor-neutral, technology- and infrastructure-independent framework that would enable Big 96 Data stakeholders to identify and use the best analytics tools for their processing and visualization 97 requirements on the most suitable computing platform and cluster, while also allowing value-added from 98 Big Data service providers. 99

The NIST Big Data Interoperability Framework consists of seven volumes, each of which addresses a 100 specific key topic, resulting from the work of the NBD-PWG. The seven volumes are as follows: 101

Volume 1, Definitions 102 Volume 2, Taxonomies 103 Volume 3, Use Cases and General Requirements 104 Volume 4, Security and Privacy 105 Volume 5, Architectures White Paper Survey 106 Volume 6, Reference Architecture 107 Volume 7, Standards Roadmap 108

The NIST Big Data Interoperability Framework will be released in three versions, which correspond to 109 the three stages of the NBD-PWG work. The three stages aim to achieve the following: 110

Stage 1: Identify the high-level NIST Big Data Reference Architecture (NBDRA) key components, 111 which are technology, infrastructure, and vendor agnostic 112

Stage 2: Define general interfaces between the NBDRA components 113 Stage 3: Validate the NBDRA by building Big Data general applications through the general interfaces 114

Potential areas of future work for the Subgroup during stage 2 are highlighted in Section 1.5 of this 115 volume. The current effort documented in this volume reflects concepts developed within the rapidly 116 evolving field of Big Data. 117

1.2 SCOPE AND OBJECTIVES OF THE REFERENCE ARCHITECTURE SUBGROUP 118

Reference architectures provide “an authoritative source of information about a specific subject area that 119 guides and constrains the instantiations of multiple architectures and solutions.”2 Reference architectures 120 generally serve as a reference foundation for solution architectures, and may also be used for comparison 121 and alignment purposes. This volume was prepared by the NBD-PWG Reference Architecture Subgroup. 122 The effort focused on developing an open reference Big Data architecture that achieves the following 123 objectives: 124

Provides a common language for the various stakeholders 125 Encourages adherence to common standards, specifications, and patterns 126 Provides consistent methods for implementation of technology to solve similar problem sets 127 Illustrates and improve understanding of the various Big Data components, processes, and 128

systems, in the context of vendor- and technology-agnostic Big Data conceptual model 129 Provides a technical reference for U.S. government departments, agencies, and other consumers 130

to understand, discuss, categorize, and compare Big Data solutions 131 Facilitates the analysis of candidate standards for interoperability, portability, reusability, and 132

extendibility 133

The NBDRA is intended to facilitate the understanding of the operational intricacies in Big Data. It does 134 not represent the system architecture of a specific Big Data system, but rather is a tool for describing, 135

Page 13: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

3

discussing, and developing system-specific architectures using a common framework. The reference 136 architecture achieves this by providing a generic, high-level conceptual model, which serves as an 137 effective tool for discussing the requirements, structures, and operations inherent to Big Data. The model 138 is not tied to any specific vendor products, services, or reference implementation, nor does it define 139 prescriptive solutions for advancing innovation. 140

The NBDRA does not address the following: 141

Detailed specifications for any organizations’ operational systems 142 Detailed specifications of information exchanges or services 143 Recommendations or standards for integration of infrastructure products 144

As a precursor to the development of the NBDRA, the NBD-PWG Reference Architecture Subgroup 145 surveyed the currently published Big Data platforms by leading companies supporting the Big Data 146 framework. All the reference architectures provided to the NBD-PWG are listed and the capabilities of 147 each surveyed platform are discussed in this document. 148

1.3 REPORT PRODUCTION 149

A wide spectrum of Big Data architectures were explored and developed from various industries, 150 academic, and government initiatives. The NBD-PWG Reference Architecture Subgroup produced this 151 report through the four steps outlined below. 152

1. Announced that the NBD-PWG Reference Architecture Subgroup is open to the public to attract 153 and solicit a wide array of subject matter experts and stakeholders in government, industry, and 154 academia 155

2. Gather publicly available Big Data architectures and materials representing various stakeholders, 156 different data types, and different use cases 157

3. Examine and analyze the Big Data material to better understand existing concepts, usage, goals, 158 objectives, characteristics, and key elements of the Big Data, and then document the findings 159 using NIST’s Big Data taxonomies model (presented in NIST Big Data Interoperability 160 Framework: Volume 2, Taxonomies) 161

4. Produce this report to document the findings and work of the NBD-PWG Reference Architecture 162 Subgroup 163

1.4 REPORT STRUCTURE 164

Following the introductory material presented in Section 1, the remainder of this document is organized 165 as follows: 166

Section 2 contains the reference architectures submitted to the NBD-PWG 167 Section 3 discusses the surveyed reference architectures and highlights key functionalities 168 Section 4 presents several conclusions gained from the evaluation of the reference architectures 169 Appendix A contains acronyms used in this document 170 Appendix B lists references provided throughout this document 171

1.5 FUTURE WORK ON THIS VOLUME 172

As presented in this document, information about existing reference architectures was collected and 173 analyzed by the NBD-PWG. The knowledge gained from the surveyed architectures was used in the 174 NBDRA development. Additional work is not anticipated for the reference architecture survey. 175

176

177

Page 14: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

4

2 BIG DATA ARCHITECTURE PROPOSALS RECEIVED 178

To begin the survey, NBD-PWG members were asked to provide information about Big Data 179 architectures. The architectures described in this section were discussed at length by the NBD-PWG. The 180 NBD-PWG Reference Architecture Subgroup sincerely appreciates contributions of the respective Big 181 Data architectures from the following individuals and organizations. This survey would not have been 182 possible without their contributions, 183

1. Robert Marcus, ET Strategies 184 2. Orit Levin, Microsoft 185 3. Yuri Demchenko, University of Amsterdam 186 4. James Kobielus, IBM 187 5. Harry Foxwell, Oracle 188 6. Milind Bhandarkar, EMC/Pivotal 189 7. Sanjay Patil, SAP 190 8. Barry Devlin, 9sight Consulting 191 9. Tony Middleton, LexisNexis 192

2.1 BOB MARCUS 193

2.1.1 GENERAL ARCHITECTURE DESCRIPTION 194 The High Level, Layered Reference Model and detailed Lower Level Reference Architecture in this 195 section are designed to support mappings from Big Data use cases, requirements, and technology gaps. 196 The Layered Reference Model is at a similar level to the NBDRA. The Lower Level Reference 197 Architecture (Section 2.1.3) is a detailed drill-down from the High Level, Layered Reference Model 198 (Section 2.1.2). 199

2.1.2 ARCHITECTURE MODEL 200 The High Level, Layered Reference Model in Figure 1 gives an overview of the key functions of Big 201 Data architectures. 202

203 Figure 1: Components of the High Level Reference Model 204

Page 15: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

5

A. External Data Sources and Sinks 205 This feature provides external data inputs and output to the internal Big Data components. 206

B. Stream and Extract, Transform, Load (ETL) Processing 207 This processing feature filters and transforms data flows between external data resources and internal Big 208 Data systems. 209

C. Highly Scalable Foundation 210 Horizontally scalable data stores and processing form the foundation of Big Data architectures. 211

D. Operational and Analytics Databases 212 Databases are integrated into the Big Data architecture. These can be horizontally scalable databases or 213 single platform databases with data extracted from the foundational data store. 214

E. Analytics and Database Interfaces 215 These are the interfaces to the data stores for queries, updates, and analytics. 216

F. Applications and User Interfaces 217 These are the applications (e.g. machine learning) and user interfaces (e.g. visualization) that are built on 218 Big Data components. 219

G. Supporting Services 220 These services include the components needed for the implementation and management of robust Big 221 Data systems. 222

2.1.3 KEY COMPONENTS 223 The Lower Level Reference Architecture in Figure 2 expands on some of the layers in the High Level 224 Layered Reference Model and shows some of the data flows. 225

 226 Figure 2: Description of the Components of the Low-Level Reference Model 227

Page 16: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

6

The Lower Level Reference Architecture (Figure 2) maps to the High Level Layered Reference Model 228 (Figure 1) as outlined below. 229

A. External Data Sources and Sinks 230 4. Data Sources and

Sinks These components clearly define interfaces to Big Data horizontally scalable internal data stores and applications.

231

B. Stream and ETL Processing 232 5. Scalable Stream

Processing This is processing of “data in movement” between data stores. It can be used for filtering, transforming, or routing data. For Big Data streams, the stream processing should be scalable to support distributed and/or pipelined processing.

233

C. Highly Scalable Foundation 234 1. Scalable

Infrastructure To support scalable Big Data stores and processing, the infrastructure should be able to support the easy addition of new resources. Possible platforms include public and/or private clouds.

2. Scalable Data Stores

This is the essence of Big Data architecture. Horizontal scalability using less expensive components can support the unlimited growth of data storage. However, there should be fault tolerance capabilities available to handle component failures.

3. Scalable Processing To take advantage of scalable distributed data stores, the scalable distributed parallel processing should have similar fault tolerance. In general, processing should be configured to minimize unnecessary data movement.

235

D. Operational and Analytics Databases 236 6. Analytics Databases Analytics databases are generally highly optimized for read-only interactions

(e.g. columnar storage, extensive indexing, and denormalization). It is often acceptable for database responses to have high latency (e.g. invoke scalable batch processing over large data sets).

7. Operational Databases

Operation databases generally support efficient write and read operations. NoSQL databases are often used in Big Data architectures in this capacity. Data can be later transformed and loaded into analytic databases to support analytic applications.

8. In Memory Data Grids

These high performance data caches and stores minimize writing to disk. They can be used for large-scale, real-time applications requiring transparent access to data.

237

Page 17: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

7

E. Analytics and Database Interfaces 238 9. Batch Analytics and

Interfaces These interfaces use batch scalable processing (e.g. Map-Reduce) to access data in scalable data stores (e.g., Hadoop File System). These interfaces can be SQL-like (e.g., Hive) or programmatic (e.g., Pig).

10. Interactive Analytics and Interfaces

These interfaces directly access data stores to provide interactive responses to end-users. The data stores can be horizontally scalable databases tuned for interactive responses (e.g., HBase) or query languages tuned to data models (e.g., Drill for nested data).

11. Real-Time Analytics and Interfaces

Some applications require real-time responses to events occurring within large data streams (e.g. algorithmic trading). This complex event processing uses machine-based analytics requiring very high performance data access to streams and data stores.

239

F. Applications and User Interfaces 240 12. Applications and

Visualization The key new capability available to Big Data analytic applications is the ability to avoid developing complex algorithms by utilizing vast amounts of distributed data (e.g. Google statistical language translation). However, taking advantage of the data available requires new distributed and parallel processing algorithms.

241

G. Supporting Services 242 13. Design, Develop,

and Deploy Tools High level tools are limited for the implementation of Big Data applications (e.g. Cascading). This should change to lower the skill levels needed by enterprise and government developers.

14. Security Current Big Data security and privacy controls are limited (e.g. only Kerberos authentication for Hadoop, Knox). They should be expanded in the future by commercial vendors (e.g. Cloudera Sentry) for enterprise and government applications.

15. Process Management

Commercial vendors are supplying process management tools to augment the initial open source implementations (e.g. Oozie).

16. Data Resource Management

Open Source data governance tools are still immature (e.g. Apache Falcon). These will be augmented in the future by commercial vendors.

17. System Management

Open source systems management tools are also immature (e.g. Ambari). Fortunately robust system management tools are commercially available for scalable infrastructure (e.g. cloud-based).

243

2.2 MICROSOFT 244

2.2.1 GENERAL ARCHITECTURE DESCRIPTION 245 This Big Data ecosystem reference architecture is a high level, data-centric diagram that depicts the Big 246 Data flow and possible data transformations from collection to usage. 247

Page 18: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

8

2.2.2 ARCHITECTURE MODEL 248 The Big Data ecosystem is comprised of four main components: Sources, Transformation, Infrastructure 249 and Usage, as shown in Figure 3. Security and Management are shown as examples of additional 250 supporting, crosscutting subsystems that provide backdrop services and functionality to the rest of the Big 251 Data ecosystem. 252

253 Figure 3: Big Data Ecosystem Reference Architecture 254

2.2.3 KEY COMPONENTS 255

Data Sources 256 Typically, the data behind Big Data is collected for a specific purpose, and is created in a form that 257 supports the known use at the data collection time. Once data is collected, it can be reused for a variety of 258 purposes, some potentially unknown at the collection time. Data sources are classified by four main 259 characteristics that define Big Data (i.e., volume, variety, velocity, and variability)3, and are independent 260 of the data content or context. 261

Data Transformation 262 As data circulates through the ecosystem, it is being processed and transformed in different ways to 263 extract value from the information. For the purpose of defining interoperability surfaces, it is important to 264 identify common transformations that are implemented by independent modules, systems, or deployed as 265 stand-alone services. The transformation functional blocks shown in Figure 3 can be performed by 266 separate systems or organizations, with data moving between those entities, such as the case with the 267 advertising ecosystem. Similar and additional transformational blocks are being used in enterprise data 268 warehouses, but typically they are closely integrated and rely on a common database to exchange the 269 information. 270

Each transformation function may have its specific pre-processing stage including registration and 271 metadata creation; may use different specialized data infrastructure best fitted for its requirements; and 272 may have its own privacy and other policy considerations. Common data transformations shown in Figure 273 3 are: 274

Page 19: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

9

Collection: Data can be collected in different types and forms. At the initial collection stage, sets 275 of data (e.g., data records) from similar sources and of similar structure are collected (and 276 combined) resulting in uniform security considerations, policies, etc. Initial metadata is created 277 (e.g., subjects with keys are identified) to facilitate subsequent aggregation or lookup method(s). 278

Aggregation: Sets of existing data collections with easily correlated metadata (e.g., identical 279 keys) are aggregated into a larger collection. As a result, the information about each object is 280 enriched or the number of objects in the collection grows. Security considerations and policies 281 concerning the resulting collection are typically similar to the original collections. 282

Matching: Sets of existing data collections with dissimilar metadata (e.g., keys) are aggregated 283 into a larger collection. For example, in the advertising industry, matching services correlate 284 HTTP cookies’ values with a person’s real name. As a result, the information about each object is 285 enriched. The security considerations and policies concerning the resulting collection are subject 286 to data exchange interfaces design. 287

Data Mining: According to DBTA4, “[d]ata mining can be defined as the process of extracting 288 data, analyzing it from many dimensions or perspectives, and then producing a summary of the 289 information in a useful form that identifies relationships within the data. There are two types of 290 data mining: descriptive, which gives information about existing data; and predictive, which 291 makes forecasts based on the data.” 292

Data Infrastructure: 293 Big Data infrastructure is a bundle of data storage or database software, servers, storage, and networking 294 used in support of the data transformation functions and for storage of data as needed. Data infrastructure 295 is placed to the right of the data transformation, to emphasize the natural role of data infrastructure in 296 support of data transformations. Note that the horizontal data retrieval and storage paths exist between the 297 two, which are different from the vertical data paths between them and data sources and data usage. 298

To achieve high efficiencies, data of different volume, variety, and velocity would typically be stored and 299 processed using computing and storage technologies tailored to those characteristics. The choice of 300 processing and storage technology is also dependent on the transformation itself. As a result, often the 301 same data can be transformed (either sequentially or in parallel) multiple times using independent data 302 infrastructure. 303

Examples of conditioning include de-identification, sampling, and fuzzing. 304

Examples of storage and retrieval include NoSQL and SQL Databases with various specialized types of 305 data load and queries. 306

Data Usage: 307 The results can be provided in different formats, (e.g., displayed or electronically encoded, with or 308 without metadata, at rest or streamed), different granularity (e.g., as individual records or aggregated), and 309 under different security considerations (e.g., public disclosure vs. internal use). 310

2.3 UNIVERSITY OF AMSTERDAM 311

2.3.1 GENERAL ARCHITECTURE DESCRIPTION 312 This Big Data Architecture Framework (BDAF) supports the extended Big Data definition presented in 313 the Systems and Network Engineering (SNE) technical report5 and reflects the main components and 314 processes in the Big Data Ecosystem (BDE). The BDAF, shown in Figure 4, is comprised of the 315 following five components that address different aspects of the SNE Big Data definition: 316

Data Models, Structures, and Types: The BDAF should support a variety of data types 317 produced by different data sources. These data must be stored and processed and will, to some 318 extent, define the Big Data infrastructure technologies and solutions. 319

Page 20: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

10

Big Data Management Infrastructure and Services: The BDAF should support Big Data 320 Lifecycle Management, provenance, curation, and archiving. Big Data Lifecycle Management 321 should support the major data transformations stages: collection, registration, filtering, 322 classification, analysis, modeling, prediction, delivery, presentation, and visualization. Big Data 323 Management capabilities can be partly addressed by defining scientific or business workflows 324 and using corresponding workflow management systems. 325

Big Data Analytics and Tools: These specifically address required data transformation 326 functionalities and related infrastructure components. 327

Big Data Infrastructure (BDI): This component includes storage, computing infrastructure, 328 network infrastructure, sensor networks, and target or actionable devices. 329

Big Data Security: Security should protect data in rest and in motion, ensure trusted processing 330 environments and reliable BDI operation, provide fine grained access control, and protect users’ 331 personal information. 332

2.3.2 ARCHITECTURE MODEL 333 Figure 4 illustrates the basic Big Data analytics capabilities as a part of the overall cloud-based BDI. 334

335 Figure 4: Big Data Architecture Framework 336

In addition to the general cloud-based infrastructure services (e.g., storage, computation, infrastructure or 337 VM management), the following specific applications and services will be required to support Big Data 338 and other data centric applications: 339

High-Performance Cluster systems 340 Hadoop related services and tools; distributed file systems 341 General analytics tools/systems: batch, real-time, interactive 342 Specialist data analytics tools: logs, events, data mining 343 Databases: operational and analytics; in-memory databases; streaming databases; SQL, NoSQL, 344

key-value storage, etc. 345 Streaming analytics and ETL processing 346

Page 21: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

11

Data reporting and visualization 347

Big Data analytics platforms should be vertically and horizontally scalable, which can be naturally 348 achieved when using cloud-based platforms and Intercloud integration models and architecture.6 349

2.3.3 KEY COMPONENTS 350 Big Data infrastructure, including the general infrastructure for general data management, is typically 351 cloud-based. Big Data analytics will use the High-Performance Computing (HPC) architectures and 352 technologies, as shown in Figure 4. General BDI includes the following capabilities, services and 353 components to support the whole Big Data lifecycle: 354

General cloud-based infrastructure, platform, services and applications to support creation, 355 deployment and operation of Big Data infrastructures and applications (using generic cloud 356 features of provisioning on-demand, scalability, measured services) 357

Big Data Management services and infrastructure, which includes data backup, replication, 358 curation, provenance 359

Registries, indexing/search, metadata, ontologies, namespaces 360 Security infrastructure (access control, policy enforcement, confidentiality, trust, availability, 361

accounting, identity management, privacy) 362 Collaborative environment infrastructure (groups management) and user-facing capabilities (user 363

portals, identity management/federation) 364

Big Data Infrastructure should be supported by broad network access and advanced network 365 infrastructure, which will play a key role in distributed heterogeneous BDI integration and reliable 366 operation: 367

Network infrastructure interconnects BDI components. These components are typically 368 distributed, increasingly multi-provider BDI components, and may include intra-cloud (intra-369 provider) and Intercloud network infrastructure. HPC clusters require high-speed network 370 infrastructure with low latency. Intercloud network infrastructure may require dedicated network 371 links and connectivity provisioned on demand. 372

FADI is presented in Figure 4 as a separate infrastructure/structural component to reflect its 373 importance, though it can be treated as a part of the general Intercloud infrastructure of the BDI. 374 FADI combines both Intercloud network infrastructure and corresponding federated security 375 infrastructure to support infrastructure components integration and users federation. 376

Heterogeneous multi-provider cloud services integration is addressed by the Intercloud Architecture 377 Framework (ICAF), and also, especially, the Intercloud Federation Framework (ICFF) being developed 378 by the authors.7 8 9 ICAF provides a common basis for building adaptive and on-demand provisioned 379 multi-provider cloud based services. 380

FADI is an important component of the overall cloud and Big Data infrastructure that interconnects all the 381 major components and domains in the multi-provider Intercloud infrastructure, including non-cloud and 382 legacy resources. Using a federation model for integrating multi-provider heterogeneous services and 383 resources reflects current practice in building and managing complex infrastructures and allows for inter-384 organizational resource sharing and identity federation. 385

2.4 IBM 386

2.4.1 GENERAL ARCHITECTURE DESCRIPTION 387 A Big Data platform must support all types of data and be able to run all necessary computations to drive 388 the analytics. 389

Page 22: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

12

The IBM reference architecture, as shown in Figure 5 outlines the architecture model. Section 2.4.2 390 provides additional details. 391

392 Figure 5: IBM Big Data Platform 393

2.4.2 ARCHITECTURE MODEL 394 To achieve these objectives, any Big Data platform should address six key imperatives: 395

Data Discovery and Exploration: The process of data analysis begins with understanding data sources, 396 figuring out what data is available within a particular source, and getting a sense of its quality and its 397 relationship to other data elements. This process, known as data discovery, enables data scientists to 398 create the right analytic model and computational strategy. Traditional approaches required data to be 399 physically moved to a central location before it could be discovered. With Big Data, this approach is too 400 expensive and impractical. To facilitate data discovery and unlock resident value within Big Data, the 401 platform should be able to discover data ‘in place.’ It should be able to support the indexing, searching, 402 and navigation of different sources of Big Data. It should be able to facilitate discovery of a diverse set of 403 data sources, such as databases, flat files, content management systems—pretty much any persistent data 404 store that contains structured, semi structured, or unstructured data. The security profile of the underlying 405 data systems should be strictly adhered to and preserved. These capabilities benefit analysts and data 406 scientists by helping them to quickly incorporate or discover new data sources in their analytic 407 applications. 408

Extreme Performance: Run Analytics Closer to the Data: Traditional architectures decoupled 409 analytical environments from data environments. Analytical software would run on its own infrastructure 410 and retrieve data from back-end data warehouses or other systems to perform complex analytics. The 411

Page 23: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

13

rationale behind this was that data environments were optimized for faster access to data, but not 412 necessarily for advanced mathematical computations. Hence, analytics were treated as a distinct workload 413 that had to be managed in a separate infrastructure. This architecture was expensive to manage and 414 operate, created data redundancy, and performed poorly with increasing data volumes. The analytic 415 architecture of the future should run both data processing and complex analytics on the same platform. It 416 should deliver petabyte scale performance throughput by seamlessly executing analytic models inside the 417 platform, against the entire data set, without replicating or sampling data. It should enable data scientists 418 to iterate through different models more quickly to facilitate discovery and experimentation with a “best 419 fit” yield. 420

Manage and Analyze Unstructured Data: For a long time, data has been classified on the basis of its 421 type—structured, semi structured, or unstructured. Existing infrastructures typically have barriers that 422 prevented the seamless correlation and holistic analysis of this data (e.g., independent systems to store 423 and manage these different data types.) Hybrid systems have also emerged, which have often not 424 performed as well as expected because of their inability to manage all data types. However, 425 organizational processes don’t distinguish between data types. To analyze customer support effectiveness, 426 structured information about a customer service representative (CSR) conversation (e.g., call duration, 427 call outcome, customer satisfaction, survey response) is as important as unstructured information gleaned 428 from that conversation (e.g., sentiment, customer feedback, verbally expressed concerns). Effective 429 analysis should factor in all components of an interaction, and analyze them within the same context, 430 regardless of whether the underlying data is structured or not. A game-changing analytics platform should 431 be able to manage, store, and retrieve both unstructured and structured data. It also should provide tools 432 for unstructured data exploration and analysis. 433

Analyze Data in Real Time: Performing analytics on activity as it unfolds presents a huge untapped 434 opportunity for the analytic enterprise. Historically, analytic models and computations ran on data that 435 was stored in databases. This worked well for transpired events from a few minutes, hours, or even days 436 back. These databases relied on disk drives to store and retrieve data. Even the best performing disk 437 drives had unacceptable latencies for reacting to certain events in real time. Enterprises that want to boost 438 their Big Data IQ need the capability to analyze data as it is being generated, and then to take appropriate 439 action. This allows insight to be derived before the data gets stored on physical disks. This type of data is 440 referred to as streaming data, and the resulting analysis is referred to as analytics of data in motion. 441 Depending on the time of day, or other contexts, the volume of the data stream can vary dramatically. For 442 example, consider a stream of data carrying stock trades in an exchange. Depending on trading activity, 443 that stream can quickly swell from 10 to 100 times its normal volume. This implies that a Big Data 444 platform should not only be able to support analytics of data in motion, but also should scale effectively 445 to manage increasing volumes of data streams. 446

A Rich Library of Analytical Functions and Tool Sets: One of the key goals of a Big Data platform 447 should be to reduce the analytic cycle time—the amount of time that it takes to discover and transform 448 data, develop and score models, and analyze and publish results. As noted earlier, when a platform 449 empowers the user to run extremely fast analytics, a foundation is provided on which to support multiple 450 analytic iterations and speed up model development (the snowball gets bigger and rotates faster). 451 Although this is the desired end goal, there should be a focus on improving developer productivity. By 452 making it easy to discover data, develop and deploy models, visualize results, and integrate with front-end 453 applications, the organization can enable practitioners, such as analysts and data scientists, to be more 454 effective in their respective jobs. This concept is referred to as the art of consumability. Most companies 455 do not have hundreds (if not thousands) of developers on hand, who are skilled in new age technologies. 456 Consumability is key to democratizing Big Data across the enterprise. The Big Data platform should be 457 able to flatten the time-to-analysis curve with a rich set of accelerators, libraries of analytic functions, and 458 a tool set that accelerates the development and visualization process. Because analytics is an emerging 459 discipline, it’s not uncommon to find data scientists who have their own preferred mechanisms for 460

Page 24: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

14

creating and visualizing models. They might use packaged applications, emerging open source libraries, 461 or build the models using procedural languages. Creating a restrictive development environment curtails 462 their productivity. A Big Data platform should support interaction with the most commonly available 463 analytic packages, with deep integration that facilitates pushing computationally intensive activities from 464 those packages, such as model scoring, into the platform. It should have a rich set of “parallelizable” 465 algorithms that have been developed and tested to run on Big Data. It has to have specific capabilities for 466 unstructured data analytics, such as text analytics routines and a framework for developing additional 467 algorithms. It must also provide the ability to visualize and publish results in an intuitive and easy-to-use 468 manner. 469

Integrate and Govern All Data Sources: Recently, the information management community has made 470 enormous progress in developing sound data management principles. These include policies, tools, and 471 technologies for data quality, security, governance, master data management, data integration, and 472 information lifecycle management. They establish veracity and trust in the data, and are extremely critical 473 to the success of any analytics program. 474

2.4.3 KEY COMPONENTS 475 The technological capabilities of the IBM framework that address these key strategic imperatives are as 476 follows: 477

Tools: These components support visualization, discovery, application development, and systems 478 management. 479

Accelerators: This component provides a rich library of analytical functions, schemas, tool sets, 480 and other artifacts for rapid development and delivery of value in Big Data projects. 481

Hadoop: This component supports managing and analyzing unstructured data. To support this 482 requirement, IBM InfoSphere BigInsights and PureData System for Hadoop support are required. 483

Stream Computing: This component supports analyzing in-motion data in real time. 484 Data Warehouse: This component supports business intelligence, advanced analytics, data 485

governance, and master data management on structured data. 486 Information Integration and Governance: This component supports integration and 487

governance of all data sources. Its capabilities include data integration, data quality, security, 488 lifecycle management, and master data management. 489

2.5 ORACLE 490

2.5.1 GENERAL ARCHITECTURE DESCRIPTION 491 Oracle’s Reference Architecture for Big Data provides a complete view of related technical capabilities, 492 how they fit together, and how they integrate into the larger information ecosystem. This reference 493 architecture helps to clarify Oracle’s Big Data strategy and to map specific products that support that 494 strategy. 495

Oracle offers an integrated solution to address enterprise Big Data requirements. Oracle’s Big Data 496 strategy is centered on extending current enterprise information architectures to incorporate Big Data. Big 497 Data technologies, such as Hadoop and Oracle NoSQL database, run alongside Oracle data warehouse 498 solutions to address Big Data requirements for acquiring, organizing, and analyzing data in support of 499 critical organizational decision-making. 500

2.5.2 ARCHITECTURE MODEL 501 Figure 6 shows a high level view of the Information Management ecosystem: 502

Page 25: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

15

503 Figure 6: High level, Conceptual View of the Information Management Ecosystem 504

2.5.3 KEY COMPONENTS 505 Figure 7 presents the Oracle Big Data reference architecture, detailing infrastructure services, data 506 sources, information provisioning, and information analysis components. 507

508 Figure 7: Oracle Big Data Reference Architecture 509

g g Consumers & Producers

Kn

ow

led

ge

Sec

uri

ty

Man

agem

en

t & M

on

itorin

g

Internal & External

Operational & Reporting Authoritative

Da

ta

Info

rmat

ion

Information Provisioning

Info

rma

tio

n

Mo

de

lin

g

Knowledge Workers

Information Stewards

Information Architects

Business Processes

Applications & Tools

Event Processors

Services

Decision Making Content System

<xxx>

Information Analytics

Information Delivery

Data Sources

Info

rmat

ion

Pro

vis

ion

ing

Distributed File Systems Faceted Unstructured Relational Data Streams

Data Warehouse

Big Data Processing

Massive Unstructured Data

& Stream Processing

Big Data Processing & Discovery

Information Discovery Data Conversion

Predictive Analytics (In-Database)

Operational Database

Data Mining Text Mining Statistical

Analysis

Semantic

Analysis Spatial In-DB

MapReduce Info

rmat

ion

A

na

lysi

s

Descriptive Analytics Reporting Dashboards

Infrastructure Services Connectivity

Hardware

Network Storage

Operating System Virtualization Security Management

NoSQL/Tag-Value Spatial/Relational

Bulk Data Movement

Page 26: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

16

2.6 PIVOTAL 510

2.6.1 GENERAL ARCHITECTURE DESCRIPTION 511 Pivotal, a spinoff from EMC and VMware, believes that Hadoop Distributed File System (HDFS) has 512 emerged as a standard interface between the data and the analytics layer. Marching on that path of 513 providing a unified platform, Pivotal integrated the Greenplum Database technology to work directly on 514 HDFS. It now has the first standards-compliant interactive SQL query engine, HAWQ, to store and query 515 both structured and semi-structured data on HDFS. Pivotal is also in the late stages of integrating real-516 time, in-memory analytics platforms with HDFS. Making HDFS a standard interface results in storage 517 optimization—only a single copy of the data is stored—and gives developers freedom to use an 518 appropriate analytics layer for generating insights for data applications. 519

2.6.2 ARCHITECTURE MODEL 520 Pivotal Big Data architecture is composed of three different layers: infrastructure, data ingestion and 521 analytics, and data-enabled applications. These layers, or fabrics, have to seamlessly integrate to provide a 522 frictionless environment for users of Big Data systems. Figure 8 implicitly corresponds to the three layers. 523 The infrastructure layer is the cloud (i.e., virtual) layer in the grey box. The black colored boxes represent 524 the data ingestion and analytics layer. The PaaS (Platform as a Service) is included as an example of an 525 application layer. 526

Figure 8: Pivotal Architecture Model 527

Infrastructure 528

Most of the early Big Data platforms instances assumed a networked bare metal infrastructure. 529 Speed to market drove the early adopters to offer this platform as a service on the bare metal 530 platforms. Monitoring and management capabilities were built around managing a bare metal 531 cluster at scale. The numbers of Big Data platform nodes at some of the early adopters are in the 532 many tens of thousands. Custom management tools were developed for implemented limited 533 elasticity by adding nodes, retiring nodes and monitoring and alerting in case of infrastructure 534 failures. 535

Since then, virtualization has been widely adopted in enterprises, since it offers enhanced 536 isolation, utilization, elasticity, and management capabilities inherent in most virtualization 537 technologies. 538

Page 27: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

17

Customers now expect a platform that will enable them to take advantage of their investments in 539 virtualized infrastructure, whether deployed on premise, such as with VMWare vSphere, or in 540 public clouds, such as Amazon AWS. 541

Data Storage and Analytics 542

Big Data platforms deal with the gravity of data (i.e., difficulty of moving large amounts of data 543 over existing network capabilities) by moving processing very close to the storage tier. As a 544 result, the analytics and data storage tiers were fused together in the Pivotal architecture. Every 545 attempt was made to move analytics closer to the data layer. Now, high-speed data center 546 interconnects (with 10+ Gbps bandwidth) are becoming more affordable, making it possible to 547 keep the processing within the same subnet and still deliver acceptable performance. 548

Virtualization technology is also innovating to add data locality awareness to the compute node 549 selection, thereby speeding up analytics workloads over large amounts of data, while offering the 550 benefits of resource isolation, better security, and improved utilization, compared to the 551 underlying bare metal platform. 552

Pivotal finds Data Lake is one of the most prevalent emerging usage patterns of the Big Data 553 platforms. Data Lake—a design pattern where streams of data are stored and available for 554 historical and operational analytics on the Big Data platform—is the starting point for most of the 555 enterprises. Once the data is available in one place, business users experiment with use cases and 556 solutions to business problems, resulting in second order benefits. 557

Big Data Applications 558

Pivotal considers extracting insight from a Big Data platform’s data as a Big Data application. 559 Excel spreadsheets, and PowerPoint presentations are used for early demonstrations of these 560 insights, and once the business cases are approved, applications need to be built to act on these 561 insights. Developers are looking for flexibility to integrate these new Big Data applications with 562 existing applications or build completely new applications. A good platform will make it easier 563 for the developers to do both at the same time. 564

In addition, the operations team for the application requires a platform to easily deploy, manage 565 and scale the application as the usage increases. These operability requirements are very critical 566 for enterprise customers. 567

2.6.3 Key Components 568

569 Figure 9: Pivotal Data Fabric and Analytics 570

Page 28: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

18

Analytics is all about distilling information (e.g., structured, unstructured across varying latencies) to 571 generate insights. Various stages in the distillation have distinctly different platform needs. Figure 9 572 illustrates Pivotal data and analytical products operating in the three layers of Pivotal architecture model. 573

Hypothesis: Historical structured and unstructured access pattern 574 Analytics problems typically start with a hypothesis that is tested on the historical data. This requires 575 analyzing massive amounts of historical data, and combining various genres of data, to extract features 576 relevant to the problem at hand. The ability to query and mix and match large amounts of data and 577 performance are the key requirements at this phase. Customers use the Hadoop interfaces when working 578 with ‘truly’ unstructured data (e.g., video, voice, images). For semi-unstructured data (e.g., machine 579 generated, text analytics, and transactional data) customers prefer structured interfaces, such as SQL, for 580 analytics. 581

Modeling: Model building requires interactive response times when accessing structured 582 data 583 Once features and relevant attributes are clearly defined and understood, models need to be built and 584 interactive response times become one of the most important requirements. Customers get tremendous 585 performance gains by using in-database analytics techniques, which optimize for reducing data 586 movements. HAWQ delivers interactive response time by parallelizing analytical computation. Most of 587 the enterprise customers have built internal teams of business analysts who are well-versed with SQL. By 588 leveraging the SQL interfaces, customers extend the life of their resource investment, continuing business 589 innovation at a much faster pace. 590

Model Operationalization: Run models in operational systems 591 Hypothesis validation and analytics finally result in analytical models that need to be operationalized. 592 Operationalizing is a process where the coefficients from the models are used in Big Data applications to 593 generate insights and scores for decision making. Some customers are building real-time applications that 594 generate insights and alerts based on the computed features. 595

Requirements from the customers across the three stages of analytics are as follows: 596

Interface compatibility across the three stages—SQL is a common standard for accessing data. 597 Pivotal Data Fabric is focused on providing a single consistent interface for all data on HDFS. 598

Quality of Service—Support varying response time for access to same data. Customers are 599 looking to get batch, interactive and real-time access to same datasets. The window of data access 600 drives the expected response times. 601

Ability to manage and monitor the models. Advanced customers are using the Application Fabric 602 technologies to build data applications. 603

Pivotal Data Fabric treats HDFS as the storage layer for all the data—low latency data, structured 604 interactive data, and unstructured batch data. This improves the storage efficiency and minimizes friction, 605 as the same data can be accessed via varying interfaces at varying latencies. Pivotal data fabric comprises 606 the following three core technologies: 607

1. Pivotal HD: for providing the HDFS capabilities along with the Hadoop interfaces to access 608 data (e.g., Pig, Hive, HBase and Map/Reduce). 609

2. HAWQ: for interactive analytics for the data stored on HDFS. 610 3. GemFire/SQLFire: for real-time access to the data streaming in and depositing the data on 611

HDFS. 612

The Data Fabric can also run on a bare metal platform or in the public/private clouds. The deployment 613 flexibility allows enterprise customers to select the ideal environment for managing their analytical 614 environments without worrying about changing the interfaces for their internal users. As the enterprise 615 needs change, the deployment infrastructure can be changed accordingly. 616

Page 29: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

19

2.7 SAP 617

2.7.1 GENERAL ARCHITECTURE DESCRIPTION 618 The data enterprises needed for managing data has grown exponentially in recent years. In addition to 619 traditional transactional data, businesses are finding it advantageous to accumulate all kinds of other data 620 such as weblogs, sensor data, and social media—generally referred to as Big Data—and leverage it to 621 enrich their business applications and gain new insights. 622

SAP Big Data architecture (Figure 10) provides businesses with a platform to handle the high volumes, 623 varieties and velocities of Big Data, as well as to rapidly extract and utilize business value from Big Data 624 in the context of business applications and analytics. 625

SAP HANA platform for Big Data helps relax the traditional constraints of creating business applications. 626

Traditional disk-based systems suffer from performance hits not just due to the necessity of moving data 627 to and from the disk, but also from the inherent limitations of disk-based architectures, such as the need to 628 create and maintain indices and materialized aggregates of data. SAP HANA platform largely eliminates 629 such drawbacks and enables the business to innovate freely and in real time. 630

Business applications are typically modeled to capture the rules and activities of real world business 631 processes. SAP HANA platform for Big Data makes a whole new set of data available for consumption 632 and relaxes the dependency on static rules in creating business applications. Applications can now 633 leverage real-time Big Data insights for decision making, which vastly enhances and expands their 634 functionality. 635

Traditionally, exploration and analytics of data is based on the knowledge of the structures and formats of 636 the data. With Big Data, it is often required to first investigate the data for what it contains, whether it is 637 relevant to the business, how it relates to the other data, and how it can be processed. SAP HANA 638 platform for Big Data supports tools and services for data scientists to conduct such research of Big Data 639 and enable its exploitation in the context of business applications. 640

2.7.2 ARCHITECTURE MODEL 641 SAP Big Data architecture provides a platform for business applications with features such as the ones 642 referenced above. The key principles of SAP Big Data architecture include: 643

An architecture that puts in-memory technology data at its core and maximizes computational 644 efficiencies by bringing the compute and data layers together. 645

Support for a variety of data processing engines (such as transactional, analytical, graph, and 646 spatial) operating directly on the same data set in memory. 647

Interoperability/integration of best of breed technologies such as Hadoop/Hive for the data layer, 648 including data storage and low level processing. 649

The ability to leverage these technologies to transform existing business applications and build 650 entirely new ones that were previously not practical. 651

Comprehensive native support for predictive analytics, as well as interoperability with popular 652 libraries such as R, for enabling roles such as data scientists to uncover and predict new business 653 potentialities. 654

In a nutshell, SAP Big Data architecture is not limited to handling large amounts of data, but instead is 655 meant to enable enterprises to identify and realize the business value of real-time Big Data. 656

Page 30: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

20

657 Figure 10: SAP Big Data Reference Architecture 658

2.7.3 KEY COMPONENTS 659 SAP Big Data architecture (Figure 10) enables an end-to-end platform and includes support for ingestion, 660 storage, processing and consumption of Big Data. 661

The ingestion of data includes acquisition of structured, semi-structured, and unstructured data from a 662 variety of sources to include traditional back end systems, sensors, social media, and event streams. 663 Managing data quality through stewardship and governance and maintaining a dependable metadata store 664 are key aspects of the data ingestion phase. 665

SAP Big Data architecture brings together transactional and analytical processing that directly operate on 666 the same copy of the enterprise data that is held entirely in memory. This architecture helps by 667 eliminating the latency between transactional and analytical applications, as there is no longer a need to 668 copy transactional data into separate systems for analytical purposes. 669

With in-memory computing, important application functionalities such as planning and simulation can be 670 executed in real time. SAP Big Data architecture includes a dedicated engine for planning and simulation 671 as a first class component, making it possible to iterate through various simulation and planning cycles in 672 real time. 673

SAP Big Data architecture includes a Graph engine. The elements of Big Data are typically loosely 674 structured. With the constant addition of new types of data, the structure and relationship between the data 675 is constantly evolving. In such environments, it is inefficient to impose an artificial structure (e.g. 676 relational) on the data. Modeling the data as graphs of complex and evolving interrelationships provides 677 the needed flexibility in capturing dynamic, multi-faceted data. 678

An ever increasing number of business applications are becoming location aware. For example, 679 businesses can now send promotions to the mobile device of a user walking into a retail store, which has a 680 much higher chance of capturing the user’s attention and generating a sale than traditional marketing. 681 Recognizing this trend, SAP Big Data architecture includes a spatial data processing engine to support 682 location aware business applications. For similar reasons, inherent capabilities for text, media, and social 683 data processing are also included. 684

Page 31: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

21

SAP Big Data architecture supports complex event processing throughout the entire stack. Event streams 685 (e.g., sensor data, update from capital markets) are not just an additional source of Big Data; they also 686 require sophisticated processing of events such as processing on the fly (e.g., ETL) and analytics on the 687 fly. 688

SAP Big Data architecture also enables customers to use low cost data storage and low level data 689 processing solutions such as Hadoop. By extending the SAP HANA platform for Big Data with Hadoop, 690 customers can bring the benefits of real-time processing to data in Hadoop systems. Scenarios for 691 extracting high value data from Hadoop, as well as federating data processing in SAP HANA with 692 Hadoop / Hive computational engines into a single SQL query, are fully supported. 693

SAP Big Data architecture enables applying common approaches, such as modeling, life cycle 694 management, landscape management, and security, across the platform. 695

In summary, SAP Big Data architecture takes full advantage of the SAP HANA platform, helping 696 businesses use Big Data in ways that will fundamentally transform their operations. 697

2.8 9SIGHT 698

2.8.1 GENERAL ARCHITECTURE DESCRIPTION 699 This simple picture (Figure 11) sets the overall scope for the 700 discussion and design between business and IT of systems 701 supporting modern business needs, which include Big Data and 702 real-time operation in a biz-tech ecosystem. Each layer is described 703 in terms of three axes or dimensions as follows: 704

Timeliness/Consistency: the balance between these two 705 demands commonly drives layering of data, e.g. in data 706 warehousing. 707

Structure/Context: an elaboration of 708 structured/unstructured descriptions that defines the 709 transformation of information to data. 710

Reliance/Usage: information trustworthiness based on its 711 sourcing and pre-processing. 712

The typical list of Big Data characteristics is subsumed in these 713 characteristic dimensions. 714

2.8.2 ARCHITECTURE MODEL 715 The REAL (Realistic, Extensible, Actionable, Labile) architecture supports building an IT environment 716 that can use Big Data in all business activities. The term “business” encompasses all social organizations 717 of people with the intention of pursuing a set of broadly related goals, including both for-profit and 718 nonprofit enterprises and governmental and nongovernmental stakeholders. The REAL architecture 719 covers all information and all processes that occur in such a business, but does not attempt to architect 720 people. The architecture model is shown in Figure 12. 721

2.8.3 KEY COMPONENTS 722 Business applications or workflows, whether operational, informational, or collaborative, with their 723 business focus and wide variety of goals and actions, are gathered together in a single component, 724 utilization. 725

Figure 11: 9Sight General Architecture

Page 32: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

22

Three information processing components are 726 identified. Instantiation is the means by 727 which measures, events, and messages from 728 the physical world are represented as, or 729 converted to, transactions or instances of 730 information within the enterprise 731 environment. Assimilation creates reconciled 732 and consistent information, using ETL tools 733 and data virtualization tools, before users 734 have access to it. Reification, which sits 735 between all utilization functions and the 736 information itself, provides a consistent, 737 cross-pillar view of information according to 738 an overarching model. Reification allows 739 access to the information in real time, and 740 corresponds to data virtualization for “online” 741 use. Modern data warehouse architectures use 742 such functions extensively, but the naming is 743 often overlapping and confusing, hence the 744 unusual function names used here. 745

The Service Oriented Architecture (SOA) 746 process- and services-based approach uses an 747 underlying choreography infrastructure, 748 which coordinates the actions of all 749 participating elements to produce desired outcomes. There are two subcomponents: adaptive workflow 750 management and an extensible message bus. These functions are well known in standard SOA work. 751

Finally, the organization component covers all design, management, and governance activities relating to 752 both processes and information. 753

Information/data is represented in pillars for three distinct classes as described below. 754

Human-sourced information: Information originates from people, because context comes only 755 from the human intellect. This information is the highly subjective record of human experiences 756 and is now almost entirely digitized and electronically stored everywhere from tweets to movies. 757 Loosely structured and often ungoverned, this information may not reliably represent for the 758 business what has happened in the real world. 759

Process-mediated data: Business processes record well-defined, legally binding business events. 760 This process-mediated data is highly structured and regulated, and includes transactions, 761 reference tables and relationships, and the metadata that sets its context. Process-mediated data 762 includes operational and BI systems and was the vast majority of what IT managed in the past. It 763 is amenable to information management and to storage and manipulation in relational database 764 systems. 765

Machine-generated data: Sourced from the sensors, computers, etc. used to record events and 766 measures in the physical world, such data is well-structured and usually reliable. As the Internet 767 of Things grows, well-structured machine-generated data is of growing importance to business. 768 Some claim that its size and speed is beyond traditional RDBMS, mandating NoSQL stores. 769 However, high-performance RDBMSs are also often used. 770

Context setting information (metadata) is an integral part of the information resource, spanning all pillars. 771

Figure 12: 9Sight Architecture Model

Page 33: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

23

2.9 LEXISNEXIS 772

2.9.1 GENERAL ARCHITECTURE DESCRIPTION 773

774 Figure 13: Lexis Nexis General Architecture 775

The High Performance Computing Cluster (HPCC) Systems platform is designed to handle massive, 776 multi-structured datasets ranging from hundreds of terabytes to tens of petabytes, serving as the backbone 777 for both LexisNexis online applications and programs within the U.S. Federal Government alike. The 778 technology has been in existence for over a decade, and was built from the ground up to address internal 779 company requirements pertaining to scalability, flexibility, agility and security. Prior to the technology 780 being released to the open source community in June 2011, the HPCC had been deployed to customer 781 premises as an appliance (software fused onto a preferred vendor’s hardware), but has since become 782 hardware-agnostic in an effort to meet the requirements of an expanding user base. The Lexis Nexis 783 general architecture is shown in Figure 13. 784

2.9.2 ARCHITECTURE MODEL 785 The HPCC is based on a distributed, shared-nothing architecture and contains two cluster types—one 786 optimized for “data refinery” activities (i.e., THOR cluster) and the other for “data delivery” (i.e., ROXIE 787 cluster). The nodes comprising both cluster types are homogenous, meaning all processing, memory and 788 disk components are the same and based on commercial-off-the-shelf (COTS) technology. 789

DATA CENTER

Data Factory Components

Data Build Cluster

(Data Refinery/Thor)

Landing Zone

Rapid Data Delivery Engine

Clusters(ROXIES)

Supercomputer infrastructure:LDAPDaliSashaAttribute RepositoryECL Server

Network Switch

Incoming Data

Physical Media

Monitoring Server

Query ESPServers

Operator Console(s)

FTPthroughfirewall

Query

Load Balancer

Load Balancer

Firewall

Query Deployment

Page 34: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

24

In addition to compute clusters, the HPCC environment also contains a number of system servers which 790 act as a gateway between the clusters and the outside world. The system servers are often referred to 791 collectively as the HPCC “middleware,” and include the following: 792

The Enterprise Control Language (ECL): compiler, executable code generator and job 793 server (ECL Server): Serves as the code generator and compiler that translate ECL code. 794

System data store (Dali): Used for environment configuration, message queue maintenance, and 795 enforcement of LDAP security restrictions. 796

Archiving server (Sasha): Serves as a companion ‘housekeeping’ server to Dali. 797 Distributed File Utility (DFU Server): Controls the spraying and despraying operations used to 798

move data onto and out of THOR. 799 The inter-component communication server (ESP Server): The inter-component 800

communication server that allows multiple services to be “plugged in” to provide various types of 801 functionality to client applications via multiple protocols. 802

2.9.3 KEY COMPONENTS 803 Core components of the HPCC include the THOR data refinery engine, ROXIE data delivery engine, and 804 an implicitly parallel, declarative programming language, ECL. Each component is outlined below in 805 further detail and shown in Figure 14. 806

THOR Data Refinery: THOR is a massively parallel ETL engine that can be used for 807 performing a variety of tasks such as massive: joins, merges, sorts, transformations, clustering, 808 and scaling. Essentially, THOR permits any problem with computational complexities O(n2) or 809 higher to become tractable. 810

ROXIE Data Delivery: ROXIE serves as a massively parallel, high throughput, structured query 811 response engine. It is suitable for performing volumes of structured queries and full text ranked 812 Boolean search, and can also operate in highly available (HA) environments due to its read-only 813 nature. ROXIE also provides real-time analytics capabilities, to address real-time classifications, 814 prediction, fraud detection and other problems that normally require handling processing and 815 analytics on data streams. 816

817 Figure 14: Lexis Nexis High Performance Computing Cluster 818

Page 35: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

25

The Enterprise Control Language (ECL): ECL is an open source, data-centric programming 819 language used by both THOR and ROXIE for large-scale data management and query processing. 820 ECL’s declarative nature enables users to solely focus on what they need to do with their data, 821 while leaving the exact steps for how this is accomplished within a massively parallel processing 822 architecture (MPP) to the ECL compiler. 823

As multi-structured data is ingested into the system and sprayed across the nodes of a THOR cluster, 824 users can begin to perform a multitude of ETL-like functions, including the following: 825

Mapping of source fields to common record layouts used in the data 826 Splitting or combining of source files, records, or fields to match the required layout 827 Standardization and cleaning of vital searchable fields, such as names, addresses, dates, etc. 828 Evaluation of current and historical timeframe of vital information for chronological 829

identification and location of subjects 830 Statistical and other direct analysis of the data for determining and maintaining quality as new 831

sources and updates are included 832 Mapping and translating source field data into common record layouts, depending on their 833

purpose 834 Applying duplication and combination rules to each source dataset and the common build 835

datasets, as required 836

THOR is capable of operating either independently or in tandem with ROXIE; when ROXIE is present it 837 hosts THOR results and makes them available to the end-user through a web service API. 838

839

Page 36: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

26

3 SURVEY OF BIG DATA ARCHITECTURES 840

Based on the Big Data platforms surveyed, a remarkable consistency in the core of the Big Data reference 841 platforms was observed. Each of the surveyed platforms minimally consists of the following components 842 in its architecture: 843

Data Management 844 Analytics 845 Infrastructure 846

In the subsections below, each of the nine surveyed architectures is discussed and key functionality for each 847 are highlighted. The main features of the submitted architectures are described in the sections below. 848

3.1 BOB MARCUS 849

Mr. Marcus, an individual contributor, presented a layered architecture model for Big Data. This model 850 proposes the following six components: 851

A. Data Sources and Sinks 852 This component provides external data inputs and output to the internal Big Data components. 853

B. Application and User Interfaces 854 These are the applications (e.g., machine learning) and user interfaces (e.g., visualization) built on Big 855 Data components. 856

C. Analytics Databases and Interfaces 857 The framework proposes integration of databases into the Big Data architecture. These can be 858 horizontally scalable databases or single platform databases, with data extracted from the foundational 859 data store. The framework outlines the following databases and interfaces (Table 1.) 860

Table 1: Databases and Interfaces in the Layered Architecture from Bob Marcus 861

Database Types Description

Databases Analytics Databases In general, these are highly optimized for read-only interactions and typically acceptable for database responses to have high latency (e.g., invoke scalable batch processing over large data sets).

Operational Databases

In general, these support efficient write and read operations. NoSQL databases are often used in Big Data architectures in this capacity. Data can later be transformed and loaded into analytic databases to support analytic applications.

In Memory Data Grids

These high performance data caches and stores minimize writing to disk. They can be used for large scale real-time applications requiring transparent access to data.

Page 37: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

27

Database Types Description

Analytics and Database Interfaces

Batch Analytics and Database Interfaces

These interfaces use batch scalable processing (e.g., Map-Reduce) to access data in scalable data stores (e.g., Hadoop File System). These interfaces can be SQL-like (e.g., Hive) or programmatic (e.g., Pig).

Interactive Analytics and Interfaces

These interfaces avoid direct access data stores to provide interactive responses to end users. The data stores can be horizontally scalable databases tuned for interactive responses (e.g., HBase) or query languages tuned to data models (e.g., Drill for nested data).

Real-Time Analytics and Interfaces

Some applications require real-time responses to events occurring within large data streams (e.g., algorithmic trading). This complex event processing uses machine-based analytics, which require very high performance data access to streams and data stores.

D. Scalable Stream and Data Processing 862 This component provides filters and transforms data flows between external data resources and internal 863 Big Data systems. 864

E. Scalable Infrastructure 865 The framework specifies scalable infrastructure that can support easy addition of new resources. Possible 866 platforms include public and/or private clouds and horizontal scaling data stores, as well as distributed 867 scalable data processing platforms. 868

F. Supporting Services 869 The framework specifies services needed for the implementation and management of robust Big Data 870 systems. The subcomponents specified within the supporting services are described below. 871

Design, Develop and Deploy Tools: The framework cautions that high level tools are limited for 872 the implementation of Big Data applications. This should change to lower the skill levels needed 873 by enterprise and government developers. 874

Security: The framework also notes lack of standardized and adequate support to address data 875 security and privacy. The framework cites that only Kerberos authentication for Hadoop, Knox 876 exists. Capabilities should be expanded in the future by commercial vendors for enterprise and 877 government applications. 878

Process Management: The framework notes that commercial vendors are supplying process 879 management tools to augment the initial open source implementations (e.g., Oozie). 880

Data Resource Management: The author notes that open source data governance tools are still 881 immature (e.g., Apache Falcon). These will be augmented in the future by commercial vendors. 882

System Management: The framework notes that open source systems management tools are also 883 immature (e.g., Ambari). However, robust system management tools are commercially available 884 for scalable infrastructure (e.g., cloud-based). 885

Figure 15 provides diagrammatic representation of the architecture. 886

Page 38: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

28

887 Figure 15: Big Data Layered Architecture 888

3.2 MICROSOFT 889

Microsoft defines a Big Data reference architecture that would have the four key functional capabilities, 890 as summarized below. 891

A. Data Sources 892

According to Microsoft, “data behind Big Data” is collected for a specific purpose, creating the data 893 objects in a form that supports the known use at the time of data collection. Once data is collected, it can 894 be reused for a variety of purposes, some potentially unknown at the collection time. Microsoft also 895 explains that data sources can be classified by four characteristics that are independent of the data content 896 and context: volume, velocity, variability, and variety. 897

B. Data Transformation 898 The second component of the Big Data reference architecture Microsoft describes is Data Transformation. 899 Microsoft defines this as the stage where data is processed and transformed in different ways to extract 900 value from the information. Each transformation function may have its specific pre-processing stage, 901 including registration and metadata creation; may use different specialized data infrastructure best fitted 902 for its requirements; and may have its own privacy and other policy considerations and interoperability. 903 Table 2 lists the common transformation functions defined by Microsoft. 904

Table 2: Microsoft Data Transformation Steps 905

Data Transformation Steps

Functional Description

Data Collection Data can be collected for different types and forms; similar sources and structure resulting in uniform security considerations, policies, and allows creation of an initial metadata

Aggregation This is defined by Microsoft as where sets of existing data is collected to form an easily correlated metadata (e.g., identical keys) and then aggregated into a larger collection thus enriching number of objects as the collection grows.

Page 39: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

29

Data Transformation Steps

Functional Description

Matching This is defined as where sets of existing data collections with dissimilar metadata (e.g., keys) are aggregated into a larger collection. Similar to aggregation this stage also enhances information about each object.

Data Mining Microsoft refers this as a process where data, analyzing it from many dimensions or perspectives, then producing a summary of the information in a useful form that identifies relationships within the data. There are two types of data mining: descriptive, which gives information about existing data; and predictive, which makes forecasts based on the data

C. Data Infrastructure 906 Microsoft defines Big Data infrastructure as a collection of data storage or database software, servers, 907 storage, and networking used in support of the data transformation functions and for storage of data as 908 needed. Furthermore, to achieve higher efficiencies, Microsoft defines infrastructure as a medium where 909 data of different volume, variety, variability, and velocity would typically be stored and processed using 910 computing and storage technologies tailored to those characteristics. The choice of processing and storage 911 technology is also dependent on the transformation itself. As a result, often the same data can be 912 transformed (either sequentially or in parallel) multiple times using independent data infrastructure. 913

D. Data Usage 914 The last component of the Big Data architecture framework is the data usage. After data has cycled 915 through a given infrastructure, the end-result can be provided in different formats, different granularity 916 and under different security considerations. 917

3.3 UNIVERSITY OF AMSTERDAM 918

University of Amsterdam (UVA) proposes a five-part Big Data framework as part of the overall cloud-919 based BDI. Each of the five subsections is explored below. 920

A. Data Models 921 The UVA Big Data architecture framework includes data models, structures and types that support a 922 variety of data types produced by different data sources and need to be stored and processed. 923

B. Big Data Analytics 924 In the UVA Big Data reference architecture, Big Data analytics is envisioned as a component that will use 925 HPC architectures and technologies. Additionally, in the proposed model, Big Data analytics is expected 926 to be scalable vertically and horizontally. This can be naturally achieved when using cloud-based 927 platform and Intercloud integration models/architecture. The architecture outlines the following analytics 928 capabilities supported by high-performance computing architectures and technologies: 929

Refinery, linking, and fusion 930 Real-time, interactive, batch, and streaming 931 Link analysis, cluster analysis, entity resolution, and complex analysis 932

C. Big Data Management 933 The UVA architecture describes Big Data management services with these components: 934

Data backup, replication, curation, provenance 935 Registries, indexing/search, metadata, ontologies, namespaces 936

Page 40: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

30

D. Big Data Infrastructure 937 The UVA architecture defines Big Data infrastructure that requires broad network access and advanced 938 network infrastructure to integrate distributed heterogeneous BDI integration offering reliable operation. 939 This includes: 940

General cloud-based infrastructure, platform, services, and applications to support creation, 941 deployment, and operation of Big Data infrastructures and applications (using generic cloud 942 features of provisioning on-demand, scalability, measured services. 943

Collaborative environment infrastructure (e.g., group management) and user-facing capabilities 944 (e.g., user portals, identity management/federation). 945

Network infrastructure that interconnects typically distributed and increasingly multi-provider 946 BDI components that may include intra-cloud (intra-provider) and Intercloud network 947 infrastructure. 948

FADI, which is to be treated as a part of the general Inter-cloud infrastructure of the BDI. FADI 949 combines both inter-cloud network infrastructure and corresponding federated security 950 infrastructure to support infrastructure components integration and users federation. FADI is an 951 important component of the overall cloud and Big Data infrastructure that interconnects all the 952 major components and domains in the multi-provider inter-cloud infrastructure including non-953 cloud and legacy resources. Using federation model for integrating multi-provider heterogeneous 954 services and resources reflects current practice in building and managing complex infrastructures 955 and allows for inter-organizational resource sharing and identity federation. 956

E. Big Data Security 957 The UVA Big Data reference architecture describes Big Data security that should protect data at rest and 958 in motion, ensure trusted processing environments and reliable BDI operation, provide fine-grained 959 access control, and protect users’ personal information 960

3.4 IBM 961

IBM’s Big Data reference model proposes a Big Data framework that can be summarized in four key 962 functional blocks, which are described below. 963

A. Data Discovery and Exploration (Data Source) 964 IBM explains data analysis as first understanding data sources, what is available in those data sources, the 965 quality of data, and its relationship to other data elements. IBM describes this process as data discovery, 966 which enables data scientists to create the right analytic model and computational strategy. Data 967 discovery also supports indexing, searching, and navigation. The discovery is independent of data sources 968 that include relational databases, flat files, and content management systems. The data store supports 969 structured, semi-structured, or unstructured data. 970

Data Discovery Functions

Indexing Searching Navigation

Str

uct

ure

d

Sem

i-st

ruct

ure

d

Un

stru

ctu

red

Relational Database

Flat Files

Content Management System

Figure 16: Data Discovery and Exploration 971

Page 41: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

31

B. Data Analytics 972 IBM’s Big Data platform architecture recommends running both data processing and complex analytics 973 on the same platform, as opposed to the traditional approach where analytics software runs on its own 974 infrastructure and retrieves data from back-end data warehouses or other systems. The rationale is that 975 data environments were customarily optimized for faster access to data, but not necessarily for advanced 976 mathematical computations. Therefore, analytics were treated as a distinct workload that had to be 977 managed in a separate infrastructure. 978

Within this context, IBM recommends the following: 979

Manage and analyze unstructured data: IBM notes that a game-changing analytics platform 980 must be able to manage, store, and retrieve both unstructured and structured data. It also has to 981 provide tools for unstructured data exploration and analysis. 982

Analyze data in real time: Performing analytics on activity as it unfolds presents a huge 983 untapped opportunity for the analytics enterprise. 984

In above context, IBM’s Big Data platform breaks down Big Data analytics capabilities as follows: 985

BI/Reporting 986 Exploration/Virtualization 987 Functional App 988 Industry App 989 Predictive Analytics 990 Content Analytics 991

C. Big Data Platform Infrastructure 992 IBM notes that one of the key goals of a Big Data platform should be to reduce the analytic cycle time 993 (i.e., the amount of time that it takes to discover and transform data, develop and score models, and 994 analyze and publish results.) IBM further emphasizes the importance of compute intensive infrastructure 995 with this statement, “A Big Data platform needs to support interaction with the most commonly available 996 analytic packages, with deep integration that facilitates pushing computationally intensive activities from 997 those packages, such as model scoring, into the platform. It needs to have a rich set of “parallelizable” 998 algorithms that have been developed and tested to run on Big Data. It has to have specific capabilities for 999 unstructured data analytics, such as text analytics routines and a framework for developing additional 1000 algorithms. It must also provide the ability to visualize and publish results in an intuitive and easy-to-use 1001 manner.” IBM outlines the tools that support Big Data infrastructure as follows: 1002

Hadoop: This component supports managing and analyzing unstructured data. To support this 1003 requirement, IBM InfoSphere, BigInsights, and PureData System for Hadoop support are 1004 required. 1005

Stream Computing: This component supports analyzing in-motion data in real time. 1006 Accelerators: This component provides a rich library of analytical functions, schemas, tool sets 1007

and other artifacts for rapid development and delivery of value in Big Data projects. 1008

D. Information Integration and Governance 1009 The last component of IBM’s Big Data framework is integration and governance of all data sources. This 1010 includes data integration, data quality, security, lifecycle management, and master data management. 1011

Page 42: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

32

3.5 ORACLE 1012

The architecture as provided by Oracle contains the following four components: 1013

A. Information Analytics 1014 The information analytics component has two major areas: (1) descriptive analytics, and (2) predictive 1015 analytics, with subcomponents supporting those two analytics. 1016

Descriptive Analytics 1017

o Reporting 1018 o Dashboard 1019

Predictive Analytics (In-Database) 1020

o Statistical Analysis 1021 o Semantic Analysis 1022 o Data Mining 1023 o Text Mining 1024 o In-DB Map/Reduce 1025 o Spatial 1026

B. Information Provisioning 1027 The information-provisioning component performs discovery, conversion, and processing of massive 1028 structured, unstructured, and streaming data. This is supported by both the operational database and data 1029 warehouse. 1030

C. Data Sources 1031 The Oracle Big Data reference architecture supports the following data types: 1032

Distributed file system 1033 Data streams 1034 NoSQL/Tag-value 1035 Relational 1036 Faceted unstructured 1037 Spatial/relational 1038

D. Infrastructure Services 1039 The following capabilities support the infrastructure services: 1040

Hardware 1041 Operating system 1042 Storage 1043 Security 1044 Network 1045 Connectivity 1046 Virtualization 1047 Management 1048

3.6 PIVOTAL 1049

Pivotal’s Big Data architecture is composed of three different layers: 1050

Infrastructure 1051 Data ingestion and analytics 1052 Data-enabled applications 1053

Page 43: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

33

These layers, or fabrics, have to seamlessly integrate to provide a frictionless environment for users of 1054 Big Data systems. Pivotal believes the ability to perform timely analytics largely depends on the 1055 proximity between the data store and where the analytics are being performed. This led Pivotal to fuse 1056 analytics and the data storage tier together. Additionally, Pivotal sees virtualization technology innovating 1057 to add data locality awareness to the compute node selection, thereby speeding up analytics workloads 1058 over large amounts of data, while offering the benefits of resource isolation, better security, and improved 1059 utilization, compared to the underlying bare metal platform. The following subsections further describe 1060 the capabilities: 1061

Pivotal Data Fabric and Analytics 1062 Pivotal uses its HAWQ analytics platform to support structured and unstructured data across varying 1063 latencies. The architecture recognizes the ability to query and mix and match large amounts of data and 1064 while supporting data performance. Pivotal supports Hadoop interfaces when working with “truly” 1065 unstructured data (e.g., video, voice, images). For semi-unstructured data (e.g., machine-generated, text 1066 analytics, and transactional data), it supports structured interfaces (e.g., SQL) for analytics. According to 1067 Pivotal, the data fabric is architected to run on bare metal platform or in the public/private clouds. 1068

Pivotal data fabric treats HDFS as the storage layer for all data—low latency data, structured interactive 1069 data, and unstructured batch data. Pivotal states that this improves the storage efficiency and minimizes 1070 friction as the same data can be accessed via the varying interfaces at varying latencies. Pivotal data fabric 1071 comprises the following three core technologies: 1072

Pivotal HD for providing the HDFS capabilities, through Greenplum Database technology, to 1073 work directly on HDFS along with the Hadoop interfaces to access data (e.g., Pig, Hive, HBase 1074 and Map/Reduce). 1075

HAWQ for interactive analytics for the data stored on HDFS—store and query both structured 1076 and semi-structured data on HDFS. 1077

GemFire/SQLFire for real-time access to the data streaming in and depositing the data on HDFS. 1078

3.7 SAP 1079

SAP’s Big Data reference architecture relies on three basic functions that support data ingestion, data 1080 storage, and data consumption. These three functions are then supported by vertical pillar functions that 1081 include data lifecycle management, infrastructure management, and data governance and security. The 1082 subsections below provide additional details to the three basic functions and the supporting pillar 1083 functions. 1084

A. Data Ingestion 1085 The data ingestion function provides the foundation for the Big Data platform. The SAP HANA platform 1086 supports several data types, including structured, semi-structured, and unstructured data, which can come 1087 from a variety of sources, such as traditional back-end systems, sensors, and social media and events 1088 streams. 1089

B. Data Storage and Processing 1090 In the SAP architecture, analytical, and transactional processing are closely tied to the data store that 1091 eliminates latency. Such architecture is seen as favorable when considering efficiency. Also, the 1092 architecture supports in-memory computing in real time. The SAP architecture also has features such as 1093 graph engine and spatial data processing. These components address complex and varying “multi-1094 faceted” data and location-aware data, respectively. The architecture also has built-in support for 1095 processing distributed file system via an integrated Hadoop platform. 1096

Page 44: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

34

C. Data Consumption 1097 The data consumption functional block supports both analytics and applications. The analytics provides 1098 various functions, including exploration, dashboards, reports, charting, and visualization. The Application 1099 component of this functional block supports machine learning and predictive and native HANA 1100 applications and services. 1101

D. Data lifecycle management 1102 The data lifecycle management provides common core functionality for the life of the data object that is 1103 managed by HANA platform. 1104

E. Infrastructure Management 1105 The Infrastructure management provides management of data objects and security for the data objects 1106 within the HANA platform. 1107

F. Data Security and Governance 1108 A host of functions vertically supports the above three functional blocks. These vertical services include 1109 data modeling, life cycle management, and data governance issues, including security, compliance, and 1110 audits. 1111

3.8 9SIGHT 1112

9Sight proposes a REAL architecture, aimed at IT, which supports building an IT environment capable of 1113 supporting Big Data in all business activities. The architecture model is described in the following six 1114 blocks: 1115

Utilization is a component that gathers operational, informational, or collaborative business 1116 applications/workflows with their business focus and wide variety of goals and actions. 1117

Instantiation is the process by which measures, events, and messages from the physical world 1118 are represented as, or converted to, transactions or instances of information within the enterprise 1119 environment. 1120

Assimilation is a process that creates reconciled and consistent information, using ETL and data 1121 virtualization tools, before users have access to it. 1122

Reification, which sits between all utilization functions and the information itself, provides a 1123 consistent, cross-pillar view of information according to an overarching model. Reification allows 1124 access to the information in real time, and corresponds to data virtualization for ‘online’ use. 1125

Choreography is a component that supports both the SOA process and services-based approach 1126 to delivering output. Choreography coordinates the actions of all participating elements to 1127 produce desired outcomes. There are two subcomponents: adaptive workflow management and an 1128 extensible message bus. These functions are well known in standard SOA work. 1129

Organization is a component that covers all design, management, and governance activities 1130 relating to both processes and information. 1131

3.9 LEXISNEXIS 1132

LexisNexis submitted a HPCC Systems platform that is designed to handle massive, multi-structured 1133 datasets ranging from hundreds of terabytes to tens of petabytes. This platform was built to address 1134 requirements pertaining to scalability, flexibility, agility, and security, and was built for LexisNexis’ 1135 internal company use and the U.S. federal government. Prior to the technology being released to the open 1136 source community in June 2011, the HPCC had been deployed to customer premises as an appliance 1137 (software fused onto a preferred vendor’s hardware), but has since become hardware-agnostic in an effort 1138 to meet the requirements of an expanding user base. 1139

Page 45: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

35

HPCC Architecture Model 1140 The HPCC is based on a distributed, shared-nothing architecture, and contains two cluster types: 1141

1. Data Refinery is a massively parallel ETL engine used for performing a variety of tasks such as 1142 massive: joins, merges, sorts, transformations, clustering, and scaling. The component permits 1143 any problem with computational complexities O(n2) or higher to become tractable. This 1144 component supports Big Data in structured, semi-structured, or unstructured data forms. 1145

2. Data Delivery component serves as a massively parallel, high throughput, structured query 1146 response engine. It is suitable for performing volumes of structured queries and full text ranked 1147 Boolean search, and can operate in HA environments due to its read-only nature. This component 1148 also provides real-time analytics capabilities to address real-time classifications, prediction, fraud 1149 detection, and other problems that normally require handling processing and analytics on data 1150 streams. 1151

The above two cluster types are supported by system servers that act as a gateway between the clusters 1152 and the outside world. The system servers are often referred to collectively as the HPCC “middleware” 1153 and include the following components. 1154

ECL compiler is an executable code generator and job server (ECL Server) that translates ECL 1155 code. The ECL is an open source, data-centric programming language that supports data refinery 1156 and data delivery components for large-scale data management and query processing. 1157

System Data Store is used for environment configuration, message queue maintenance, and 1158 enforcement of LDAP security restrictions. 1159

Archiving server is used for housekeeping purposes. 1160 Distributed File Utility moves data in/out of Data Refinery component. 1161 The inter-component communication server (ESP Server) allows multiple services to be 1162

“plugged in” to provide various types of functionality to client applications via multiple 1163 protocols. 1164

1165

Page 46: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

36

3.10 COMPARATIVE VIEW OF SURVEYED ARCHITECTURES 1166

Figure 17 was developed to visualize the commonalities and differences among the surveyed 1167 architectures. The components of each architecture were aligned horizontally. The architectures were 1168 stacked vertically to facilitate comparison of the components between the surveyed architectures. Three 1169 common components were observed in many of the surveyed architectures. These common components 1170 are as follows: 1171

Big Data Management and Storage 1172 Big Data Analytics and Application Interfaces 1173 Big Data Infrastructure 1174

This alignment of surveyed architecture helps in the development of a reference architecture along 1175 common areas. 1176

To facilitate visual comparison the following color scheme was used in Figure 17: 1177

Big Data Management and Storage

Big Data Analytics and Application Interfaces

Big Data Infrastructure

Other parts of the architecture 1178

1179

1180

Page 47: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

37

1181

Figure 17(a): Stacked View of Surveyed Architecture 1182

1183

Customer

Storage General Purpose

Sensor Experiment

High Log Data Behavioral

Industries /Businesses

Government (incl. Health

& Fin. Institution)

Application Development

Hadoop Stream Computing

Accelerators

Data Warehouse

Info Integration and Governance

Scalable Data Processing

Network Infrastructure Internet

Infrastrucutre Management/Monitoring

Security Infrastructure

Intercloud Multi-provider Heterogenous Infrastructure

Comute General Purpose

Storage Specialized

Collection

Minning

Aggregation Aggregation

Applications and User Interface

Analytics and Database Interfaces

Scalable Stream

Processing (filtering,

transforming, routing)

Volume

Network Operators /

Telecom

Data Management Supporting Services

Design, Develop & Deploy Tools

Security

Matching

Data Transformation Data Sources

Variety Velocity

Security

Data Objects

Management

System Management

Process Management

Data Resource

Management

Systems Management

Scalable Data File

StoreInteractive Analytics and Interfaces Batch Analytics and InterfacesReal-time Analytics

Functional App

Industry AppContent Analytics

Predictive Analytics

Security

Data Integration

Lifecycle Mgmt

Data Quality

Master Data Mgmt

University of Amsterdam

Microsoft

Academia

Storage & Retrieval

Conditioning

Fe

de

rate

d A

cce

ss a

nd

D

eliv

ery

Infr

ast

ruct

ure

Infrastrucutre

Analytics Applications: Link Analysis, Cluster Analysis, Entity Resolution, Complex Analysis

Data Management Data Source

Target Users, Processes,

objects, behaviors

Refinery, Linking, Fusion Real-time, Interactive, Batch, Streaming

Big Data Analytic/Tools

Bob Marcus (Individual Contribution)

IBM

Analytic Applications Big Data Platform

Exploration/ Visualization

Bi Reporting

Visualtions and Discovery

In Memory Data Grids

Operational Database

Analytics DatabaseData Sources & Sinks

Scalable Stream Processing

Data UsageData Infrastructure

Page 48: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

38

1184

Figure 17(b): Stacked View of Surveyed Architecture (continued) 1185

1186

Page 49: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

39

1187

Figure 17(c): Stacked View of Surveyed Architecture (continued) 1188

1189

Standardization and Cleaning of searchable

fields

Statistical and Direct Analysis

Massively parallel Extract, Transform and Load Engine Data Usage

Chronological

identification of location of

subjects

Applying duplications

and combination

rules to source data

Mapping and translating source field

data into common

record layout

Splitting or combining of source

files, records or

fields

PredictionFraud

Detection

Joins Merges Sorts Transformations Clustering ScalingReal-time

Classification

Data Delivery

Structured Query

Boolean Search

Instantiation

Messages

Transactional Data

Transactions

Measures Events

Ch

ore

og

rap

hy

Org

an

iza

tio

n

9sight

Reification

Human Sourced (information)Process Mediated (data)

Assimilation

Machine Generated (data)

LexisNexis

Utilization

Page 50: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

40

4 CONCLUSIONS 1190

Through the collection, review, and comparison of Big Data architecture implementations, numerous 1191 commonalities were identified. These commonalities between the surveyed architectures aided in the 1192 development of the NBDRA. Even though each Big Data system is tailored to the needs of the particular 1193 implementation, certain key components appear in most of the implementations. Three general 1194 components were observed in the surveyed architectures as outlined below with common features listed 1195 below the components. 1196

Big Data Management and Storage 1197

o Structured, semi-structured and unstructured data 1198 o Volume, variety, velocity, and variability 1199 o SQL and NoSQL 1200 o Distributed file system 1201

Big Data Analytics and Application Interfaces 1202

o Descriptive, predictive and spatial 1203 o Real-time 1204 o Interactive 1205 o Batch analytics 1206 o Reporting 1207 o Dashboard 1208

Big Data Infrastructure 1209

o In memory data grids 1210 o Operational database 1211 o Analytic database 1212 o Relational database 1213 o Flat files 1214 o Content management system 1215 o Horizontal scalable architecture 1216

Most of the surveyed architectures provide support for data users/consumers and orchestrators and 1217 capabilities such as systems management, data resource management, security, and data governance. The 1218 architectures also show a general lack of standardized and adequate support to address data security and 1219 privacy. Additional data security and privacy standardization would strengthen the Big Data platform. 1220

Figure 18 is a simplified view of a Big Data reference architecture using the common components and 1221 features observed in the surveyed architectures. The components in Figure 18 are as follows: 1222

Big Data Management and Storage: The data management component provides Big Data 1223 capabilities for structured and unstructured data. 1224

Big Data Analytics and Application Interfaces: This component includes data ETL, operational 1225 databases, data analytics, data visualization, and data governance. 1226

Big Data Infrastructure: The Big Data platform provides infrastructure and systems support 1227 with capabilities to integrate with internal and external components and to interact with Big Data 1228 platform users. 1229

Page 51: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

41

Together these components and features formed the basis for development of the NBDRA as developed 1230 and detailed in the NIST Big Data Interoperability Framework: Volume 6, Reference Architecture 1231 document. 1232

Figure 18: Big Data Reference Architecture 1233

1234

Big Data Management and Storage

Big Data Analytics and Application Interfaces

Big Data Infrastructure

Page 52: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

A-1

Appendix A: Acronyms 1235

BDAF Big Data Architecture Framework 1236

BDE Big Data Ecosystem 1237

BDI Big Data Infrastructure 1238

COTS commercial-off-the-shelf 1239

CSR customer service representative 1240

DFU Distributed File Utility 1241

ECL Enterprise Control Language 1242

ETL Extract, Transform, and Load 1243

ETL extract/transform/load 1244

FADI Federated Access and Delivery Infrastructure 1245

HDFS Hadoop Distributed File System 1246

HPCC High Performance Computing Cluster 1247

HA highly available 1248

HPC High-Performance Computing 1249

ITL Information Technology Laboratory 1250

ICAF Intercloud Architecture Framework 1251

ICFF Intercloud Federation Framework 1252

MPP massively parallel processing architecture 1253

NASA National Aeronautics and Space Administration 1254

NARA National Archives and Records Administration 1255

NIST National Institute of Standards and Technology 1256

NSF National Science Foundation 1257

NBD-PWG NIST Big Data Public Working Group 1258

NBDRA NIST Big Data Reference Architecture 1259

PaaS Platform as a Service 1260

REAL Realistic, Extensible, Actionable, and Labile 1261

SOA Service Oriented Architecture 1262

SNE Systems and Network Engineering 1263

UVA University of Amsterdam 1264

1265

Page 53: DRAFT NIST Big Data Interoperability Framework: Volume 5, Architectures White Paper …bigdatawg.nist.gov/_uploadfiles/M0396_v1_7656223932.pdf · 2015-04-27 · DRAFT NIST BIG DATA

DRAFT NIST BIG DATA INTEROPERABILITY FRAMEWORK: VOLUME 5, ARCHITECTURES WHITE PAPER SURVEY

B-1

Appendix B: References

DOCUMENT REFERENCES

1 The White House Office of Science and Technology Policy, “Big Data is a Big Deal,” OSTP Blog, accessed February 21, 2014, http://www.whitehouse.gov/blog/2012/03/29/big-data-big-deal.

2 Department of Defense Chief Information Officer, “Reference Architecture Description,” Department of Defense, June 2010, http://dodcio.defense.gov/Portals/0/Documents/DIEA/Ref_Archi_Description_Final_v1_18Jun10.pdf.

3 Gartner Press Release, “Gartner Says Solving ‘Big Data’ Challenge Involves More Than Just Managing Volumes of Data,” Gartner, Inc., June 27, 2011, http://www.gartner.com/it/page.jsp?id=1731916.

4 Lynn Greiner, “What is Data Analysis and Data Mining?,” Data Base Trends and Applications, January 7, 2011, http://www.dbta.com/Articles/Editorial/Trends-and-Applications/What-is-Data-Analysis-and-Data-Mining-73503.aspx.

5 Yuri Demchenko, Canh Ngo, and Peter Membrey, “Architecture Framework and Components for the Big Data Ecosystem Draft Version 0.2,” System and Network Engineering Group, UvA, September 12, 2013, http://www.uazone.org/demch/worksinprogress/sne-2013-02-techreport-bdaf-draft02.pdf. 

6 Y. Demchenko, M. Makkes, R. Strijkers, C. Ngo, and C. de Laat, “Intercloud Architecture Framework for Heterogeneous Multi-Provider Cloud based Infrastructure Services Provisioning,” The International Journal of Next-Generation Computing (IJNGC), Volume 4, Issue 2 (2013).

7 Y. Demchenko, M. Makkes, R. Strijkers, C. Ngo, and C. de Laat, “Intercloud Architecture Framework for Heterogeneous Multi-Provider Cloud based Infrastructure Services Provisioning,” The International Journal of Next-Generation Computing (IJNGC), Volume 4, Issue 2 (2013).

8 B. Khasnabish, J. Chu, S. Ma, N. So, P. Unbehagen, M. Morrow, M. Hasan, Y. Demchenko, and Y. Meng, “Cloud Reference Framework – Draft,” Internet Engineering Task Force, July 2, 2013, http://www.ietf.org/id/draft-khasnabish-cloud-reference-framework-05.txt.

9 Marc Makkes, Canh Ngo, Yuri Demchenko, Rudolf Strijkers, Robert Meijer, and Cees de Laat, “Defining Intercloud Federation Framework for Multi-provider Cloud Services Integration,” The Fourth International Conference on Cloud Computing, GRIDs, and Virtualization (CLOUD COMPUTING 2013), May 27 - June 1, 2013,Valencia, Spain.