• No results found

Resource Scalability for Efficient Parallel Processing in Cloud

N/A
N/A
Protected

Academic year: 2021

Share "Resource Scalability for Efficient Parallel Processing in Cloud"

Copied!
5
0
0

Loading.... (view fulltext now)

Full text

(1)

____________________________________________________

R S. Publication, rspublicationhouse@gmail.com Page 85

Resource Scalability for Efficient Parallel Processing in

Cloud

Govinda.K#1, Abirami.M#2, Divya Mercy Silva.J#3

#1 SCSE, VIT University #2 SITE , VIT University

#3 SITE , VIT University

ABSTRACT

In the era of digital world, the various technologies have been replaced by cloud computing. Where many of the recent applications are accessed using a single network called “cloud network”, having a great effect of impact with optimized usage to many purposes cloud computing is being adopted in a wide areas. Today evolving companies have to process a huge amount of data day-by-day for their customer's in a service come monetary value payment for its rendered service; the ad-hoc parallel data processing has become a hectic problem that arose in the recent trends of cloud computing for dynamically allocating resources. Today cloud computing environments have been assimilated frame works for parallel processing. These processing framework which currently used are designated for stagnate homogeneous constellation setups and it contempt’s. In this paper we are about to discuss a parallel processing framework, to dynamically allocate the resources. In such a situation tasks work much faster and wider through a Resource Assembler.

Key words: Scalability, Resource provisioning, Feedback processing, Resource assembler.

Corresponding Author: Govinda.K INTRODUCTION

In the growing trend huge amount of data are processed in a cost efficient manner in various fields. The various cloud operators such as Google, Yahoo, or Microsoft plays a vital role in dealing with vast amount of data day-by-day, which made traditional database system that are commonly very expensive and cost of maintaining it, has become a major problem. To solve such a situation, these operators have popularized with an architectural structure layout based on large commodity servers with Infrastructure-as-a-service cloud providers. Many of these operators have developed distributed application of architectures and also customized data processing frameworks. Some of the common examples are Google’s Map Reduce, Microsoft’s Dryad or Yahoo’s! Map-Reduce- Merge. They are classified generally such as High-Throughput Computing (HTC) or Many- Task Computing (MTC) where it all depends on the amount of data should be processed and the number of task involved in the computation. All these systems involved in the course of processing differs in design, size, speed, efficiency etc., but their major objective is to hiding the implementation of parallel programming, robustness, fault tolerance and execution optimization from the developers.

(2)

The developers are typically allowed to write sequential programs so that, processing framework is fragmented into small chunks of data that are distributed among the available nodes and then assembles each fragmented data and executes the fragment by assembling into the original data and then it is delivered to the client.

For companies like Google, Yahoo, Amazon, Microsoft, etc…have to process huge amount of data transformation so that there is no such constraint to have their own data center. Instead they allow their customers to rent with a cloud server for a payment-per usage basis for a limited amount of time. Operators of such are called (IaaS). The cloud operators offer to their customers to use the required amount of virtual machines that are instantiated to process a job and terminated when the job gets completed. These virtual machines are of different types, each type with their own configuration and cost defined by (IaaS) to its customers.

For companies like Google, Yahoo, Amazon, Microsoft, etc…have to process huge amount of data transformation so that there is no such constraint to have their own data center. Instead they allow their customers to rent with a cloud server for a payment-per usage basis for a limited amount of time. Operators of such are called (IaaS). The cloud operators offer to their customers to use the required amount of virtual machines that are instantiated to process a job and terminated when the job gets completed. These virtual machines are of different types, each type with their own configuration and cost defined by (IaaS) to its customers.

MATERIALS AND METHODS

In cloud computing technology Nephele’s architecture has a great impact with dynamically allocating resources. Recently regarding the parallel processing of a job in cloud network where dynamically resources are provisioned by the clients of a particular period of time the user own virtual machines and take the ownership for the rented time.

Table1. Comparison between Map Reduction Frameworks

Feature Dryad Hadoop

Monitoring Monitoring support for execution graphs

Monitoring support of HDFS Monitoring Map Reduce computations

Data Handling Shared directories/ Local disks HDFS Scheduling Data locality/ Network topology

based run time graph optimizations

Data locality/ Rack aware

Language Support

Programmable via C#c Dryad LINQ [24] provides LINQ programming API for Dryad

Implemented using Java Other languages are supported via Hadoop Streaming

Programming Model

DAG based execution flows MapReduce

____________________________________________________

R S. Publication, rspublicationhouse@gmail.com Page 86

(3)

A paper on ”Exploiting Dynamic Resource Allocation for Efficient Parallel Data Processing in the Cloud” [1] have established Nephele’s frameworks with a High-Throughput Computing (HTC) and (MTC),but as per their evaluation they cannot succeed in Resource overloading and underutilization which emerged as a major problem. In such a situation most of frameworks like MPI, Hadoop [7], Dryad [8],CGL- Map Reduce, Nimbus, Xen Hypervisor [6],Eucalyptus [5]

etc., most of the parallel applications have made a fine structure layout mechanism of frameworks in improving resources those are overloaded and they are also underutilized.In addition, the recent paper on “High Performance Parallel Computing with Clouds and Cloud Technologies”[9] they have made many researches on comparing various features of cloud technology but also they are not susceptible to latency on large data transfers which gradually performs lower communication, so overheads cannot be barred by Virtual Machines.

METHODS

Keeping Nephele’s architecture as the basis which serves as the classic master worker pattern which is directly connected to the public Network(internet) or private or (virtualized Network).

To illustrate the Nephele's architecture, in a Nephele's Job. The user must start a Virtual Machine in the public Network connected with the cloud called the Job Manager who is responsible for scheduling and coordinating the Job services rendered on behalf of the cloud operators and who has an authority in communicating with the interface called the “cloud controller”. The cloud controller’s main objective is to organize the Job Manager to allocate or deallocate the Virtual Machine according to the availability of the Resources. Each Virtual Machine in the cloud Networks are generally addressed as “Instances”. The instances are of different types with different hardware characteristics. The actual execution in the Nephele’s Job is carried out by a task manager who received a number of Jobs from the Job Manager at times. According to the Nephele’s architecture each Job is directly submitted to the Job Manager(public network) and then processed by the Task manager. It is an order to the Task manager to simultaneously give an update to the Job manager whether a Job is completed with a proper output or any possible errors has been raised. Unless an update is not made by the Task manager the Job Manager expects a set of Task manager to be empty. Once all necessary Task manager has successfully listed their completion or incompletion of the Job. The Job manager then triggers an execution of the scheduled Job. As a result a Nephele's Job is described with the DAG called directed acyclic graph (DAG) processing of a job is represented as a vertex consisting of task as it’s edges.

As our proposed methods relies on this basic components of Nephele’s architecture as basis job where each client are connected to the cloud network and each request of the client are directly submitted to the Job Manager by an interface called “cloud controller” who is an intermediate between them. The Job Manager then issues the job to the job scheduler the Job Scheduler directs the received job to the Task Manager. Similarly the Job Manager has a list that consists of the number of jobs arrived into the job queue. The Job Scheduler then splits the jobs into chunks, according to the capacity, capability, and the time requirements of virtual machines which are required to complete a particular task. Now these split jobs are distributed to the task manager, where it distributes the splitted job to the virtual machines according to jobs requirements and the availability of the virtual machines, another important feature of the Task

____________________________________________________

R S. Publication, rspublicationhouse@gmail.com Page 87

(4)

Manager is to maintain a scheduled list of the incoming job's current position whether it is in Running state, Ready state, Waiting state or in which mode does the job is processed is listed exactly by the Task Manger. Each Virtual Machine process the incoming fragmented job according to their specification and transfer them into the Task Manager meanwhile the Task Manager receives another set of splitted job from the Job Scheduler and waits for the Virtual Machines to complete the given set of jobs. Each completed job is then given to the Task Assembler.

The Task Assembler's main objective is to assemble the splitted job that is distributed to the various Virtual Machine and it reassemble the job into the original data. Though the Nephele's architecture has a Persistent storage all the processed job is completed and handed over to the Task Manager and the Task Manager has a bidirectional connection with the Resource Manager to intimate the number of Jobs arrived in the Job Queue. After a job has been processed the Resource Assembler provides a wide space for the incoming Jobs and at once the job is processed by the Virtual machines it gives another set of Jobs from the Job Queue to the available Virtual Machines. In such a way all the processed Jobs is directly delivered to the user by the Job Manager. Here the parallel processing occurs simultaneously between the Virtual Machines with the help of the Resource Manager.

Fig 1: Proposed architecture derived from nephele’s

____________________________________________________

R S. Publication, rspublicationhouse@gmail.com Page 88

(5)

CONCLUSION

In this paper we have studied the Nephele's basic architecture for parallel processing in cloud environment. The above framework/architecture fails to address resource underutilization and scalability of resources for user request. We have proposed a new architecture based on Nephele's architecture to provide resource scalability and to overcome resource underutilization over the cloud environment.

REFERENCE

[1] Daniel Warneke and Odej Kao, ]”Exploiting Dynamic Resource Allocation for Efficient Parallel Data Processing in the Cloud”, IEEE Transaction on parallel distributed systems, January 2011.

[2] AmazonWebServicesLLC.AmazonElasticComputeCloud. (AmazonEC2). http://aws.- Amazon.com /ec2/,2009.

[3] AmazonWebServicesLLC.AmazonElastic MapReduce http://aws.amazon.com/elasticmap- reduce / 2009.

[4] Elastic Hosts, http://www.elastichosts.com/.

[5] D. Nurmi, R. Wolski, C. Grzegorczyk, G. Obertelli, S. Soman, L.Youseff , and D.Zagorodnov,”The Eucalyptus Open-source Cloud-computing System,”CCGrid’09: The 9th IEEE International symposium on Cluster Computing and the grid ,shanghai,china,2009.

[6] P. Barham, B. Dragovic, K. Fraser, Hand, T. Harris, A. Ho, R. Neugebauer, I. Pratt, and A. Warfield, “Xen and the art of virtualization,”In Proceedings of the Nineteenth ACM Symposium on, Operating Systems Principles (Bolten Landing, NY, USA, October 19- 22, 2003).SOSP ’03. ACM, New York, NY, 164-177.DOI=

http://doi.acm.org/10.1145/945445.945462.

.

[7] Apache Hadoop, http://hadoop.apache.org/core/.

[8] M.Isard, M. Budiu, Y. Yu, A. Birrell, and D. Fetterly, “Dryad: Distributed data-parallel programs from sequential building blocks,” European Conference on Computer Systems, March 2007.

[9] Jaliya Ekanayake, ”High Performance parallel Computing with Clouds Technologies”

Geoffrey Fox Department of Computer Science, Indian University, Bloomington, IN 47404,USA .

____________________________________________________

R S. Publication, rspublicationhouse@gmail.com Page 89

References

Related documents

Offers and kidzee kalyan nagar is a east facing property also, such kind of tangy and ambience barbeque nation hospitality limited may also use the information.. Receiving a food

Based on this understanding, this work focuses on the effect of the inclusion of shallow donor dopant such as gallium into the normal cadmium chloride post-growth treatment as

Asset Portfolio Project Portfolio Development Portfolio •  Infrastructure, applications, data, people, processes •  Development projects and investments • 

Four eyes (9.1%), needed pars plana vitrectomy for the treatment of redetachment associated with proliferative vitreoretinopathy and scleral buckle revision surgery was

Three different short-term interest-rate series (the official discount rate, private banks’ average deposit rate, and the market rate of discount/money market rate) and

These include local authorities and large firms who are using the products as part of their internal networks as well as service providers and other operators who are using the

In this paper, we presented an empirical analysis of the effects of product market competition and ownership control on total factor productivity growth using a panel data set of

dielectric layer thickness, along with various growth parameters were studied and optimized using the various electrical characterization results from MOS, MIM and FET devices, to