Skip to content

Big Data

Analytics And More
  • Home
  • Spark
  • Design Patterns
  • streaming
  • Map Reduce
  • Hive
  • Hdfs & Yarn
  • Pig
  • Oozie
  • Hbase

Analytics && More

pig tutorial 2 – pig data types, relations, bags, tuples, fields and parameter substitution

July, 2017 adarsh

Relations, Bags, Tuples, Fields Pig Latin statements work with relations. A relation is a bag and a bag is a…

Continue Reading →

Posted in: Data Analytics, Pig, pig latin Filed under: pig latin, pig script

pig tutorial 1 – multiquery execution, store, dump, dependencies and replicated, skewed, merge joins

adarsh

A Pig Latin statement is an operator that takes a relation as input and produces another relation as output this…

Continue Reading →

Posted in: Data Analytics, Pig, pig latin Filed under: pig latin, pig script

input formats and output formats in hadoop and mapreduce

July, 2017 adarsh

There are many input and output formats supported in hadoop out of the box and we will explore the same…

Continue Reading →

Posted in: Data Analytics, hadoop input/output, Hdfs, Map Reduce Filed under: hadoop input output, hdfs, map reduce

default mappper, reducer, partitioner, multithreadedmapper and split size configuration in hadoop and mapreduce

adarsh

What will be the mapper,reducer and the partitioner that will be used in mapreduce program if we dont specify any…

Continue Reading →

Posted in: hadoop input/output, Hdfs, Map Reduce Filed under: hadoop input output, hdfs, map reduce

hadoop mapreduce reading the entire file content without splitting the file for example reading an xml file

adarsh 2d Comments

Some applications don’t want files to be split, as this allows a single mapper to process each input file in…

Continue Reading →

Posted in: Hdfs, Map Reduce Filed under: hdfs, hdfs filesystem, map reduce

handling failures in hadoop,mapreduce and yarn

July, 2017 adarsh 1 Comment

In the real world, user code is buggy, processes crash, and machines fail. One of the major benefits of using…

Continue Reading →

Posted in: Data Analytics, Hdfs, Map Reduce, yarn Filed under: hdfs, map reduce, yarn

life cycle of a mapreduce program – job submission,job initialization, task assignment, task execution, progress updates and job completion

adarsh

You can run a mapreduce job with a single method call submit() on a Job object or you can also…

Continue Reading →

Posted in: Data Analytics, Hdfs, Map Reduce, yarn Filed under: hdfs, map reduce, yarn

Post navigation

Page 27 of 33
← Previous 1 … 26 27 28 … 33 Next →

Recent Posts

  • Optimization for Using AWS Lambda to Send Messages to Amazon MSK
  • Rebalancing a Kafka Cluster in AWS MSK using CLI Commands
  • Using StsAssumeRoleCredentialsProvider with Glue Schema Registry Integration in Kafka Producer
  • Home
  • Contact Me
  • About Me
Copyright © 2017 Time Pass Techies
 

Loading Comments...