Big Data Meets Microsoft Azure ! For Big Data & Cloud Community members this post on “Big Data, Meet Azure” is all about doing big on public cloud Azure. And sure, we no need definition for Big Data and Cloud Computing, but in a line; I would like to called both as Super Nova for […]
How to Ingest HDFS in JSON format using Apache Sqoop ? by NS Saravanan In current project use lambda architecture, so Data from sources system extracted in two ways, Real time streaming OR speed layer Batch process or Bach Layer Speed layer implemented using Attunity > Kafka > Spark streaming . The out of Spark stream […]
The top 79 beautiful lines for taking big data architecture from drawing board to production! Dear Data Community, Instead of titling this blog is “The top 79 beautiful lines for taking big data architecture from drawing board to production”, It would be very suitable if we call it as book talk, which is inspired by […]
Getting Started with Google Cloud Platform ! Last month got a chance to attend Bengaluru Google Cloud OnBoard, instructor led enablement event for Google Cloud Platform(Big Data). Big Data on GCP is simply superb, must try once. And presenting the prepared Getting Started with Google Cloud Platform artifact for our handy reference. Below are the quick […]
Big Data Stack 2.0 and Beyond! The Google File System (GFS), MapReduce, and Bigtable are Googles & data industries Big Data revolution, which constructs Big Data Stack 1.0. Dough Cutting actually integrated the above released concepts into a tool called Hadoop. GFS + MapReduce + Bigtable > HDFS + MapReduce + HBase; which is together […]
What is the best big data solution for working with all databases from Splunk ! The answer is Splunk DB Connect! In this blog we will see how the Splunk DB connect helps us to integrate all the databases from Splunk. Splunk DB Connect is the best solution for working with databases from Splunk. It […]
What is Beyond Classic Hadoop? Is it Spark and Flink? In this blog, we will explore the two new big data friends to Hadoop, and they are Apache Spark and Apache Flink. And if we take the Hadoop improvements with the parallel processing MapReduce; speed is very first focus. However, MapReduce is designed and developed for […]
The 7 Habits Of Successful Big Data and NoSQL Projects by Ben Lorica ! Let’s have firstname.lastname@example.org
Big Data Splunk’s Best & Better Practices ! Introduction to Splunk We see servers, devices, apps, logs, traffic, and clouds. We see data, big data, and fat data everywhere. Splunk offers the leading platform for Operational Intelligence. It enables the curious to look closely at what others ignore which is called machine data and find […]
Sqoop Use Cases Introduction: Sqoop was originally developed by Cloudera You can import data from relational Database to HDFS as well export it back to relational database from HDFS Sqoop supports many RDBMS and not limited to just MySQL etc. it also supports Legacy systems like Mainframes DB2. Sqoop Use cases: ELT: Extract Load Transform […]
A First Look at Big Data Apache Flink! There is abundance of interest in learning how to analyze streaming data in large-scale systems, partly because there are situations in which the time-value of data makes real-time analytics so eye-catching. But gathering in-the-moment insights made possible by very low latency applications is just one of the […]
Top 16 Hadoop Built-in Ingress and Egress Tools ! Hadoop has revolutionized data ingestion, data processing and enterprise data warehousing, but its explosive growth has come with a large amount of uncertainty, hype, and confusion. With this blog, enterprise decision makers will receive short quick insights on what all the 16 Hadoop build-in Ingress and […]
Big Data Meets Microsoft Azure ! For Big Data & Cloud...
How to Ingest HDFS in JSON format using Apache Sqoop ?...
The 4 Key Concepts in the Anatomy of an Apache Spark Job!...
The 1-2-3-4-5-6-7-8-9 of Cognitive Computing ! Dear Data...