UNIT-I
UNIT-I
UNIT-I
UNIT-I
Introduction
Evolutionary Trend in Distributed Computing
The chapter mainly asses the evolutionary change in machine architecture, operating system
platform, network connectivity and application workload.
The HPC & HTC system are both adapted by consumer and high end web scale
computing and information services.
In HTC system, Peer to Peer network are formed for distributed file sharing and
content delivery application.
A P2P systems built on many client system cloud computing and web services
platforms are focused on HTC application.
HPC systems emphasize the row speed performance. The speeds of HPC system are
improved by the demand from scientific, engineering and manufacturing communities. The
majority of computer users are using desktop computer on large severs when they conduct
internet searches and market drivers computing tasks.
HTC systems pays more attention to high flux computing .The main application of high flux
computing is in internet searches and web services. The throughput is defined as number of
tasks completed per unit of time .HTC not only improves bath processing speed ,but also
addresses problems of cost, energy, savings, security and reliability.
Computing paradigms
A) Centralized computing
B) Parallel computing
All processer are either tightly coupled with centralized shared memory (or) loosely
coupled with distributed memory. It is also accomplished through processing inter processor
communication is accomplished through shared memory via manger passing. A system
capable of parallel computing is known as parallel computer programs shows in parallel
computer are called parallel programs. The process of writing parallel programs is referred as
parallel programming.
C) Distributed computing
A Distributed system consists of multiple autonomous computer .each having its own
private memory communication through a computer network .Information exchange is
accomplished by message passing. A computer program that runs in distributed system is
known as distributed programs .The process of writing distributed programs is known a
distributed programming.
D) Cloud computing
An internet cloud of resources can either a centralized or distributed computing
systems .The cloud can be parallel or distributed computing .clouds can be built with physical
or virtualized resources ever large data centers.
E) Ubiquitous computing
It refers to computing with pervasive device at any place and time using wired (or)
wireless communication.
A) Efficiency
B) Dependability
It measures the reliability and self manager from clip to the system and application
levels.
The purpose is to provide high throughput services with quality of service (QOS)
even under failure state.
C) Adaption in Processing models
It measures the ability to support large jobs requests ever massive data sets and virtually
cloud resources under various workload and service models.
a) Bit level Parallelism (BLP): converts bit-serial processing to world level processing.
b) Instruction level parallelism (ILP): processor executes instructions simultaneously.
c) Data level parallelism (DLP): Through SIMD (simple instruction, multiple data).It
requires more hardware support and compiler assistance.
d)Task level parallelism (TLP): Due to an introduction of multi core processor and clip
multiprocessor (CMP).It is fair due to difficulty in programming and complication of code for
efficient execution on multi core CMPS.
e) Job level Parallelism (JLP): Due to move from parallel processing to distributed
processing. The cases grain parallelism is built on top of fine grain parallelism.
Utility computing
Internet of Things
The IOT refer to the networked interconnection of everyday object tools devices .These
things can be large or small and they vary respect to time and place The idea is to every
object using REID(or)sensor(or)electronic technology. The IOT needs to be designed to track
many static (or) morning objects simultaneously .It demands universal addressability of all
the objects. It reduces the complexity of identification search and storage and set the
threshold to filter out fine grain objects.
Three communication patterns exists
a) Human to Human (H2H)
b) Human to Thing (H2T)
c) Thing to Thing (T2T)
The concept behind their communication pattern is to connect things at any time and any
place intelligently .But still it is infancy stage of development.
It result of interaction between computational process and the physical world .A CPS
integrates cyber (heterogeneous, asynchronous) with physical (concurrent, information dense)
objects.
It merges 3c technologies.
a) Computation
b) Communication
c) Control -into an intelligent closed feedback system between the physical world and
information world.
It emphasizes the exploration of virtual Reality application in physical world.
ii) Multi-Threading
Five independent Threads of instruction to four pipelined data paths in five different
processors
a) Four- issue superscalar processor
b) Fine grain multi threaded processor
c) Coarse grain multi threaded processor
d) Dual are CMP
The super scalar processer is single threaded with four Functional units. Each of three
multiple threaded processer in four way multi threaded over four functional data paths .The
dual core processer assume two processing cores, each a single threaded two way superscalar
processors.
In superscalar processors same thread is executed .In fine grained switches the execution of
instruction from different thread per cycle. In coarse grained, executes instruction from
different threads simultaneously SMT allows simultaneously scheduling of instruction from
different threads in same cycle. The blank field indicates no instructions.
b) GPU Computing
i) Working of GPU
NVIDIA GPU has been upgraded to 128 cores on a single clip. Each core on a GPU
can handle eight threads of instruction. This translates to having up to 1024 threads executed
concurrently on a single GPU.
CPU is optimized for latency cache, while GPU is optimized to delivers high
throughput with explicit management of on-chip memory.
GPUs are designed to handle large no of floating point operations in parallel GPU
offloads the CPU from all data interactive calculation.
CPU is the conventional multi core processor with limited parallelism to exploit. GPU has
many cores architecture that has hundreds of simple processing cores organized as multi
processes.
Each core can have one or more threads. The CPU instructs the GPU to perform massive data
processing. CPU's floating point kernel computation role is largely offloaded to many cores
GPU
1. Memory Technology
The capacity increase of disk array will be greater. Faster processor speed and largely
memory capacity result in a wider gap between processors and memory. The memory will
may become even worse a problem limiting CPU performance.
The rapid growth of flash memory and Solid State Driver (SSD) also impacts the future of
HPC and HTC systems. Power increases linearly with respect to clock frequency. The clock
rate cannot be increased indefinitely. SSD's are expensive to replace stable disk arrays.
The nodes in small clusters are mostly interconnected by an Ethernet switch (or) LAN.LAN
is used to connect client hosts to big servers. SAN connects servers to network storage such
as disk array.NAS connects client hosts directly to disk arrays.
Hypervisor
Bare-metal Hosted
The user application running on its dedicated OS could be bundled together as a Virtual
appliance that can be ported to any hardware platform.
The massive systems are classified into four groups. They are
a) Clusters
b) P2P networks
c) Computing grids
d) Internet Clouds
These four system classes consists of millions of computers are participating nodes. These
machines work collectively, collaboratively at various levels.
a) Clusters
b) Computational Grids
Grids
Computational P2P
(OR) Data
P2P architecture offers a distributed model of networked systems. In P2P every node
acts as both client and server, providing part of system resources.
Peer machines are client connected over internet. All client machines act
autonomously to join or leave the system freely. There is no master slave relationship
and no control database.
Only the participating peers from the physical network at anytime. The physical
network is simply an adhoc network formed at various internet domain randomly
using TCP/IP protocols.
Files are distributed in participating peers. Based on communication peer IDs form an
overlay network at logical level. This overlay is a virtual network formed by mapping
each physical machine with ID.
When new peer joins the system, peer ID is added as node in overlay network. When
a peer removes the system, Peer ID is removed from overlay network
Unstructured Structured
ISSUES
d) Internet clouds
Cloud Computing
A cloud is a pool of virtualized computer resources. A cloud can host a variety of different
workloads, including batch style backend jobs and interactive and user facing applications.
Cloud computing applies a virtual platform with elastic resources on demand by provisions
hardware, software and data sets dynamically .Virtualize resources from data centers to form
an Internet cloud for paid users to run their application.
1. Infrastructure as a Service
Put together infrastructure demanded by users. The users can deploy and run on
multiple VM running guest OS on specific applications. The user does not manage or control
the underlying cloud infrastructure but can specify when to request and release the needed
resources.
2. Platform as a Service
This model enables the user to deploy user built applications onto a virtualized cloud
platform. It includes middleware, databases, development tools and runtime support. This
platform includes both hardware and software integrated with specific programming
interfaces. The user is freed from managing the cloud infrastructure.
3. Software as a Service
It is browser initiated application software over paid cloud customers. It applies to business
process, industry applications; ERP on customer side, there is no upfront investment in
servers. On provider side, costs are low, compared with hosting of user application.
Deployment modes
a) Private
b) Public
c) Managed &
d) Hybrid cloud
1. Desired location in areas with protected space and higher energy efficiency.
2. Sharing of load capacity among large pool of users.
3. Separation of infrastructure maintenance duties from domain specific application
development.
4. Cost reduction
5. Cloud computing programming and application development
6. Service and discovery
7. Privacy, security, copyright and reliability
8. Service agreements, business models and pricing.
In grid and web services, Java and CORBA architecture built on traditional OSI layers that
provide base networking abstraction. The base software environment such as .Net, Apache
Axis for web services, Java Virtual Machine for java.
The entity interfaces corresponds to WSDD. Java method and CORBA interface definition
language (IDF). These interfaces are linked with customized, high level communication
system. The communication system supports features including RPC, fault recovery and
specialized routing. These communications built on message oriented middleware
infrastructure such as web sphere (or) Java Message Service (JMS) which provides rich
functionality and support virtualization.
Security is a critical capability that either user or re implements the concepts of IP sec and
secure sockets. Entity communication is supported by higher level services.
Linked with
Entity Interface Communication System
Customized
Built on
The web service specifies all aspects of service and its environment. The specification
is carried out using SOAP communication message. The hosting environment then
becomes a universal distributed operating system with fully distributed capability
carried by SOAP messages.
The REST Approach adopts universal principle and delegates most of difficult
problems to application software. It has minimal information in header, message body
contains all information. It is appropriate for rapid technology and environments.
In CORBA and Java, the distributed entities are linked with RPC and the simplest
way to build composite applications, to view the entities as objects. For Java writing
java program with method calls replaced by (RMI) Remote Method Invocation, While
CORBA supports similar model with syntax reflecting c++ style of object interfaces.
A grid system applies static resources while cloud emphasizes elastic resources. Build a grid
out of multiple clouds. So Grid is better than cloud because it explicitly support negotiated
resource allocation.
• Resource sharing
• The grid can be thought of as a distributed system with non-interactive workloads that
involve a large number of files (data).
• Grids are often constructed with general-purpose grid middleware software libraries.
• Fabric Layer
• Connectivity Layer
• Resource services
• Collective Layer
• Application Layer
• OGSA
– defines requirements for framework for grid systems that support integration,
virtualization and management.
• OGSI
• OGSA-DAI
• Grid FTP