A Novel set of tools to “Extract, Rework, Batch” [ERB] Big Data with Apache Hadoop Environment

Journal Title: IOSR Journals (IOSR Journal of Computer Engineering) - Year 2014, Vol 16, Issue 5

Abstract

 Abstract: Over the previous couple of years, organizations across public and personal sectors have created a strategic call to show massive information into competitive advantage. The challenge of extracting price from huge information is comparable in many ways to the antique drawback of distilling business intelligence from transactional knowledge. At the center of this challenge is that the method costumed extract information from multiple sources, remodel it to suit your analytical wants, and load it into an information warehouse for consequent analysis, a process called “Extract, Rework , Batch” (ERB). The character of huge knowledge needs that the infrastructure for this method can scale cost-effectively. Apache Hadoop has emerged because the factual normal for managing huge data. This paper examines some of the platform hardware and software considerations in using Hadoop for (ERB).

Authors and Affiliations

Ketavarapu Nageswara Rao , Maddineni Karuna Sree , K. Narasimha Sastry , Duvvada Rajeswara Rao

Keywords

Related Articles

Collective Behavior of social Networking Sites

Now a days a huge data is generated by social media like Facebook, Twitter, Flickr, and YouTube This big data present opportunities and challenges to study collective behavior of data. In this work, we predict collective...

DELB: Differential Evolution BASED Load Balancing Technique for Load Balance in Cloud Computing

Cloud applications are often subject to unexpected events like flash crowds and hardware failures. Scholars addressed this matter on two various fronts: first, they presented replicas – application copies with same funct...

 Building a Diabetes Data Warehouse to Support Decision making in healthcare industry

 Abstract : data warehousing did not find its way easily and readily into healthcare and medicine, not like others financial institutions, Healthcare presents unique challenges for the architect of a data warehouse...

Optimizing Task Scheduling and Resource allocation in Cloud Data Center, using Enhanced Min-Min Algorithm

Abstracts: Cloud Computing provide the chance to use computing resources over the internet without owning the infrastructure. The main content of Cloud Computing is to manage Software application, data storage and proces...

I++Mapreduce: Incremental Mapreduce for Mining the Big Data

Abstract: Data mining is an interdisciplinary area of computer engineering. Incremental processing is a challenging approach to refreshing mining results and it uses saved states to avoid the cost of re computation from...

Download PDF file
  • EP ID EP116363
  • DOI 10.9790/0661-16586572
  • Views 107
  • Downloads 0

How To Cite

Ketavarapu Nageswara Rao, Maddineni Karuna Sree, K. Narasimha Sastry, Duvvada Rajeswara Rao (2014).  A Novel set of tools to “Extract, Rework, Batch” [ERB] Big Data with Apache Hadoop Environment. IOSR Journals (IOSR Journal of Computer Engineering), 16(5), 65-72. https://europub.co.uk/articles/-A-116363