Slides from my session “Big Data beyond Apache Hadoop – How to Integrate ALL your Data” at JavaOne 2013 in San Francisco are online.
Big data represents a significant paradigm shift in enterprise technology. Big data radically changes the nature of the data management profession as it introduces new concerns about the volume, velocity and variety of corporate data.
Apache Hadoop is the open source defacto standard for implementing big data solutions on the Java platform. Hadoop consists of its kernel, MapReduce, and the Hadoop Distributed Filesystem (HDFS). A challenging task is to send all data to Hadoop for processing and storage (and then get it back to your application later), because in practice data comes from many different applications (SAP, Salesforce, Siebel, etc.) and databases (File, SQL, NoSQL), uses different technologies and concepts for communication (e.g. HTTP, FTP, RMI, JMS), and consists of different data formats using CSV, XML, binary data, or other alternatives.
This session shows different open source frameworks and products (especially Apache Camel and Talend Open Studio for Big Data) to solve this challenging task. Learn how to use every thinkable data with Hadoop – without plenty of complex or redundant boilerplate code.
Click on the button to load the content from www.slideshare.net.
Real-time visibility has become essential in logistics. As supply chains grow more complex, providers must…
SAP Sapphire 2025 in Madrid brought together global SAP users, partners, and technology leaders to…
Agentic AI is emerging as a powerful pattern for building autonomous, intelligent, and collaborative systems.…
Fantasy sports has evolved into a data-driven, real-time digital industry with high stakes and massive…
Confluent, Databricks, and Snowflake are trusted by thousands of enterprises to power critical workloads—each with…
Enterprise data lives in complex ecosystems—SAP, Oracle, Salesforce, ServiceNow, IBM Mainframes, and more. This article…