Workshop Agenda
Monday, June 30th
State of the Art and Future
scenarios of HPC and Grid |
||
|
Welcome Address |
|
|
J. Dongarra |
|
|
I. Foster |
“Towards an Open Analytics
Environment” |
|
D. Reed |
|
|
A. Gara |
“Present
and future challenges as we architect for the Exascale” |
|
Coffee Break |
|
|
A. Trefethen |
|
|
|
“The Evolution
of Research and Education Networks |
|
Concluding Remarks |
|
Emerging Computer Systems and Solutions |
||
|
F. Baetke |
|
|
S. Gillich |
“Intel -
Delivering Leadership HPC Technology Today and Tomorrow” |
|
T. Sasakura |
“NEC’s HPC Strategy - Consistency
and Innovation” |
|
Coffee Break |
|
|
T. Beckers |
“High
Performance Storage Solutions from DataDirect Networks” |
|
M. De Vries |
|
|
F. Magugliani |
|
|
Concluding Remarks |
Tuesday, July 1st
Advances in HPC Technology and
Systems 1 |
||
|
W. Hu |
“The
Godson-3 multi-core CPU and its application in High Performance Computers” |
|
R. Hetherington |
|
|
C. Jesshope |
“Managing
resources dynamically in SVP - from many-core to Grid” |
|
A. Shafarenko |
|
|
F. Cappello |
“Fault Tolerance
for PetaScale Systems: Current Knowledge, Challenges and Opportunities” |
|
Coffee Break |
|
|
P. Beckman |
“The Path to Exascale Computing” |
|
S. Matsuoka |
“Ultra
Low Power HPC --- scaling supercomputing by three orders of magnitude” |
|
J. Vetter |
“HPC
Interconnection Networks – The Key to Exascale Computing” |
|
Concluding Remarks |
|
Advances in HPC Technology and
Systems 2 |
||
|
J. Boisseau |
“Deployment
Experiences, Performance Observations, and Early Science Results on Ranger” |
|
R. Pennington |
“NCSA Blue
Waters: Preparing for the Sustained Petascale System” |
|
T. Lippert |
|
|
B. Miller |
|
|
Coffee Break |
|
|
PANEL DISCUSSION 1 Chairman: P. Messina Co-organizers: P. Beckman, P.
Messina Panelists: P. Beckman, A. Gara, D.
Reed, S. Matsuoka, R. Vetter |
Wednesday, July 2nd
Grid Technology and Systems 1 |
||
|
M. Livny |
“Old problems
never die – managing the multi-programming mix” |
|
D. Abramson |
“Active
Data: Blurring the distinction between data and computation” |
|
D. Talia |
“Using
Peer-to-Peer Dynamic Querying in Grid Information Services” |
|
Y. Robert |
“Algorithms and
scheduling techniques for clusters and grids” |
|
R. Sakellariou |
“Feedback
control for efficient autonomic solutions on the Grid” |
|
Coffee Break |
|
|
C. Catlett |
|
|
A. Algom |
“From
Grid Computing to Cloud Computing - The evolution of the Grid Marketplace” |
|
I. Llorente |
|
|
Concluding Remarks |
|
Grid Technology and Systems 2 |
||
|
M. Pasin |
“Network
resource reservation and virtualization for grid applications” |
|
A. Touzene |
“A Performance
Based Distribution Algorithm for Grid Computing |
|
C. Kesselman |
“Applications of Grid Technology
to Health Care Systems” |
|
Coffee Break |
|
|
PANEL DISCUSSION 2 “From Grids to
Cloud Services” Organizer: C. Catlett Panelists: Avner Algom, Pete
Beckman, Charlie Catlett, Ignacio Llorente, Satoshi Matsuoka |
Thursday, July 3rd
Infrastructures,
Instruments, Products, Solutions for High Performance Computing and Grids |
||
|
G. Fox |
|
|
A. Höing |
“Summary-based
Distributed Semantic Database for Resource and Service |
|
A. Streit |
|
|
W. Gentzsch |
|
|
M. Silberstein |
“Superlink-online
- delivering the power of GPUs, clusters and opportunistic grids to
geneticists” |
|
Coffee Break |
|
|
M. Bubak |
“Building
collaborative applications for system-level science” |
|
D. Simmel |
|
|
A. Congiusta |
|
|
Concluding Remarks |
|
National and International Grid
Infrastructures and Projects |
||
|
D. Abramson |
“e-Research
& Grid Computing in Australia: From Infrastructure to Research” |
|
K. Cho |
|
|
A. Gurtu |
|
|
Coffee Break |
|
|
A. Sachenko |
|
|
P. Öster |
|
|
M. Mazzucato |
“Italian Grid Infrastructure” |
|
Concluding Remarks |
Friday, July 4th
Challenging Applications of HPC
and Grids |
||
|
M. Athanasoulis |
“Building Shared High Performance Computing Infrastructure for
the Biomedical Sciences” |
|
P. Sloot |
“ViroLab:
Distributed Decision Support in a virtual laboratory |
|
U. Catalyurek |
“Processing of Large-Scale Biomedical Images on a Cluster of
Multi-Core CPUs and GPUs” |
|
T. David |
“A
Heterogeneous Computing Model for a Grand Challenge Problem” |
|
L. Grandinetti – P. Beraldi |
|
|
Coffee Break |
|
|
G. Aloisio – S. Fiore |
“Data Issues in
a challenging HPC application to Climate Change” |
|
G. Erbacci |
“An advanced
HPC infrastructure in Italy for
challenging scientific applications” |
|
K. Cho |
|
|
Concluding Remarks |
ABSTRACTS
Scheduling for Numerical Linear Algebra Library at Scale Jack
Dongarra Innovative
Computing Laboratory Computer
Science Dept. In this talk
we will look at some of the issues numerical library developers are facing
when using manycore systems with millions of threads of execution. |
Clouds and ManyCore: The Revolution Daniel A. Reed Microsoft Research Redmond, As Yogi
Berra famously noted, “It’s hard to make predictions, especially about the
future.” Without doubt, though, scientific discovery, business practice and
social interactions are moving rapidly from a world of homogeneous and local
systems to a world of distributed software, virtual organizations and cloud
computing infrastructure. In science,
a tsunami of new experimental and computational data and a suite of increasingly
ubiquitous sensors pose vexing problems in data analysis, transport,
visualization and collaboration. In society and business, software as a
service and cloud computing are empowering distributed groups. Let’s
step back and think about the longer term future. Where is the technology
going and what are the research implications?
What architectures are appropriate for 100-way or 1000-way multicore
designs? How do we build scalable
infrastructure? How do we develop and support software? What is the ecosystem of components in
which they will operate? How do we optimize performance, power and
reliability? Do we have ideas and
vision or are we constrained by ecosystem economics and research funding
parsimony? Biographical Sketch Daniel A. Reed is Microsoft’s Scalable and Multicore Computing
Strategist, responsible for re-envisioning the data center of the
future. Previously, he was the
Chancellor’s Eminent Professor at UNC Chapel Hill, as well as the Director of
the Renaissance Computing Institute (RENCI) and the Chancellor’s Senior Advisor
for Strategy and Innovation for UNC Chapel Hill. Dr. Reed is a member of President Bush’s
Council of Advisors on Science and Technology (PCAST) and a former member of
the President’s Information Technology Advisory Committee (PITAC). He recently chaired a review of the federal
networking and IT research portfolio, and he is chair of the board of
directors of the Computing Research Association. He was previously Head of the Department of Computer Science at the |
Present and future challenges as
we architect for the Exascale Alan Gara Dept. of Computer Science In this
presentation current trends toward achieving Petascale computing are
examined. These current trends will be contrasted with what is needed to
reach the Exascale. Possible directions and critical enabling technologies
will be discussed. |
Effective computing on heterogeneous platforms Anne
Trefethen We have entered an era where at every scale of computing
- desktop, high-performance and distributed - we need to deal with
heterogeneity. Systems are made up of
multicore chips and accelerators in an assortment of hardware architectures
and software environments. This has
created a complexity for scientific application developers and algorithm
developers alike. Our focus is on
effective algorithms and environments across these scales to support
efficient scientific application development. |
The Evolution of Research and Education Networks William E. Johnston Senior Scientist and Energy Sciences Network (ESnet) Department Head In the past
15 years there has been a remarkable increase in the volume of data that must
be analyzed in world-wide collaborations in order to accomplish the most
advanced science and a corresponding increase in network bandwidth,
deployment, and capabilities to meet these needs. Further, these changes have
touched all aspects of science including, in addition to data analysis,
remote conduct of experiments and multi-component distributed computational
simulation. Terabytes
of data from unique and very expensive instruments must be collaboratively
analyzed by the many science groups involved in the experiments. The highly
complex, long-running simulations needed to accurately represent macro-scale
phenomenon such as the climate, stellar formation, in-vivo cellular
functioning in complex organisms, etc., all involve building applications
that incorporate and use components that are located at the home institutions
of many different scientific groups. The
volume of traffic in research and education networks has increased
exponentially since about 1990. Virtually all of this increase – demonstrably
so in the past five years – is due to increased use of the network for moving
vast quantities of data among scientific instruments and widely distributed
analysis systems, and among supercomputers and remote analysis centers.
Further, this data movement is no longer optional for science: Increasingly
large-scale science is dependent on network-based data movement in order for
the science to be successful. Modern
science approaches require that networks provide not only high bandwidth, but
also advanced services. Scheduled and on-demand bandwidth enables connection
and simultaneous operation of instruments, local compute clusters,
supercomputers, and large storage systems. Low latency, high bandwidth,
secure circuits interconnect components of simulations running on systems
scattered around the country and internationally. Comprehensive, global
monitoring and reporting that allow distributed workflow systems to know
exactly how end-to-end paths that transit many different networks are
performing. At the same time, the network must provide a level of reliability
that is commensurate with the billion dollar instrument systems, scarce
supercomputers, and the hundreds of collaborating scientific groups being
interconnected that is typical of large-scale science. In this
talk I will look at how network architectures, technologies, and services
have evolved over the past 15 years to meet the needs of science that now
uses sophisticated distributed systems as an integral part of the process of
doing science. One result of this is that the R&E community has some
unique communications requirements and some of the most capable networks in
the world to satisfy those requirements. I will also look at the projected
requirements for science over the next 5 to 10 years and how the R&E
networks must further expand and evolve to meet these future requirements. |
Grids, Clouds and HPC: Opportunities and Challenges Dr.
Frank Baetke - Global HPC Technology Program Manager New trends in the HPC area can be derived
from increasing growth-rates at the lower end of the market, specifically at the
workgroup and departmental level, and from concepts which are based on the
original promises of computational grids. Those trends combined with the ever
increasing demand for even higher
component densities and higher energy efficiency generate additional
challenges: examples of new products will be shown which specifically address
those issues. |
Intel - Delivering Leadership HPC
Technology Today and Tomorrow Stephan Gillich Director HPC EMEA Intel We are
excited about the opportunity that lies in front of us as our |
High Performance
Storage Solutions from DataDirect Networks Toine Beckers DataDirect Networks
Inc., With the growing needs for High Performance
Computing clusters (from GFlops to TFlops and even PFlops systems) in many application
fields also the need for more and more data storage capacity increases as
well. This often leads to complex, difficult to manage storage solutions.
With the Silicon Storage Appliance products from DataDirect Networks an easy
to manage, scalable and high performance solution is provided which is
becoming widely accepted in the High Performance Computing Community. |
Next-Generation Cluster Management with ClusterVisionOS Martijn
De Vries Setting up and managing a large cluster can
be a challenging task without In this presentation, various aspects of the
ClusterVisionOS cluster |
Green Scalable High Performance Supercomputing Fabrizio
Magugliani EMEA
Business Development Director Sicortex As CPU speeds have reached a point where
simply increasing the clock |
The Godson-3 multi-core CPU and its application in High Performance
Computers Weiwu
Hu, Xianggao, Yunji Chen Institute
of Computing Technology, Godson-3
is a multi-core processor based on the 64-bit superscalar Godson-2 CPU core.
It takes a scalable CMP architecture in which processors and global addressed
L2 cache modules are connected in a distributed way and coherence of multiple
L1 copies of the same L2 block is maintained with a directory-based cache
coherence protocol. The
Godson-2 CPU core is a four-issue, out-of-order execution CPU which runs the
MIPS64 instruction set. The latest Godson Godson-3
adopts two-dimension mesh topology. Each node in the mesh include an 8*8
crossbar which connects four processor cores, four shared L2-cache banks and
four adjacent nodes in the East, South, West and North. A 2*2 mesh network
can connect a 16-core processor, and a 4*4 mesh network can connect a 64-core
processor. The distributed on-chip L2 cache modules are globally addressed.
Each cache block of L1 cache has a fixed L2 cache home node in which the
cache directory is maintained by directory-based cache coherence protocol.
Each node has one (or more) DDR2 memory controller. IO controllers are
connected through free crossbar ports of boundary nodes. Based on
the Godson-3 architecture, several product chips are defined and will be
physically implemented. The 4-core Godson-3 chip is designed and fabricated
based on 65nm STMicro CMOS technology. It includes one 4-core node, 4MB L2
cache, two DDR2/3 ports, two HT1.0 ports, two PCIE ports, one PCI port and
one LPC port. It will be taped out in first half of 2008. One
important application of Godson-3 is the low cost high performance computers
(HPC). Based on Godson-3, the design of one national PetaFLOPS HPC and one
personal TeraFLOPs HPC are planed. This presentation will introduces the HPC
plans based on the Godson-3 multi-core processor. |
Aggressively
Threaded Systems: A Wise Choice for HPC Rick Hetherington These
throughput workloads were not very computationally intensive but demanded
memory subsystems that provided high bandwidth and high capacity. The
second and third generations of The
result is a set of products that efficiently deliver high levels of
computational throughput. This talk
will discuss the UltraSparc T2 and T2+ processor designs as well as an
analysis of their behavior while executing 'technical' workloads. |
Managing resources dynamically in SVP – from many-core to Grid Chris
Jesshope Professor
of Computer Systems Architecture Our
computer systems are becoming pervasive and ubiquitous. It is now |
Nondeterministic Coordination
using S-Net Prof
Alex Shafarenko Department
of Computer Science Coordination languages have been used for many years
in order to separate computation and concurrency/communication, that is
coordination, concerns. Despite that, a typical coordination language
intrudes into the computational part of the code even though it provides
some abstract projection of those distributed computing realities. As a
result, units of an application program become barely readable in isolation,
without having the "big picture" in mind --- and that big picture
in turn is overburdened with interface details. We believe that the reason why coordination has
these problems is that true separation between computation and concurrency
concerns is only possible using a nondeterministic glue. Indeed deterministic
coordination abstracts application code as a state-transition system,
introducing synchonization over and above the inimum needed for correct
functioning of the application code. Nondeterministic coordination, which
we describe in this paper, leans towards loose, data-flow-style
composition using asynchronous computational structures ---
and synchronisers where necessary to ensure that the correct data
sets are worked on by fully encapsulated application code units. The paper will present a coordination language
S-Net, developed and implemented by the authors. The language is very compact, only using 4
combinators acting on user-defined boxes to create hierarchical networks of
asynchronously communicating components. The boxes are written in a
conventional language and use a conventional stream interface for output,
while the input comes as a standard parameter list. We expect ordinary engineers to be able to provide
these components. There is only one special box which the user cannot
create and which comes with the S-Net language: the synchrocell. The
significant expressive power of coordination in such a small language
is achieved by using a sophisticated type system with
subtyping, which influences the network "wiring" provided by
the combinators. The coordination program is thus a large algebraic formula
using the combinators, or several such formulae, and it is written by a
concurrency engineer who needs no detailed knowledge of the application
domain. Concurrency and self-adaptivity of S-Net is
helped by the fact that user-defined boxes are assumed to be without
persistent state, i.e. after the output stream has been flushed
and the box terminates, all local state is destroyed, so that the
next invocation of the box can take place at a different location
in the distributed system. Synchrocells retain their state between
invocations but they do not perform computations and consequently
consume no computing power. In conclusion, we will briefly dwell on the recent success
in applying S-Net to a signal processing problem in radar systems
industry at Thales Research & Technology, France. |
Fault Tolerance for PetaScale
Systems: Current Knowledge, Challenges and Opportunities Franck Cappello INRIA The
emergence of PetaScale systems reinvigorates the community interest about how
to manage failures in such systems and ensure that large applications
successfully complete. Existing results for several key mechanisms associated
with fault tolerance in HPC platforms will be presented during this talk. Most of
these key mechanisms come from the distributed system theory. Over the
last decade, they have received a lot of attention from the community and
there is probably little to gain by trying to optimize them again. We will
describe some of the latest findings in this domain. Unfortunately,
despite their high degree of optimization, existing approaches do not fit
well with the challenging evolutions of large scale systems. There is room
and even a need for new approaches. Opportunities may come from different
origins like adding hardware dedicated to fault tolerance or relaxing some of
the constraints inherited from the pure distributed system theory. We will sketch
some of these opportunities and their associated limitations. |
Ultra Low Power HPC --- scaling
supercomputing by three orders of Satoshi
Matsuoka Tokyo
Institute of Technology Low power
supercomputing as represented by various power efficient architectures such
as IBM BlueGene and power aware methods are starting to receive considerable
attention in the light of global agenda to reduce energy consumption and also
to alleviate increasing heat density problems. Our new project, Ultra
Low-Power HPC, greatly extend this horizon by taking the innovative
approaches to fundamentally slash energy consumption of supercomputing by up
to 3 orders of magnitude in 10 years. This is achieved by the comprehensive
use of new energy-efficient hardware devices and power-saving algorithms that
are modeled and optimized in a systemwide fashion. Early results from the
project are exhibiting good results in achieving 10-100 times energy efficiency,
mostly by the use of acceleration and new memory device technologies. |
HPC Interconnection Networks – The
Key to Exascale Computing Jeffrey Vetter Interconnection
networks play a critical role in the design of next generation HPC
architectures and the performance of important applications. Despite the
significance of interconnects, current trends in HPC interconnects do not appear
to fulfill the requirements for next generation multi-petaflop and exaflop
systems. Application requirements drive networks with high bandwidth, low
latency, and high message rate, while practical constraints, such as
signaling, packaging, and cost, limit improvements in hardware bandwidth and
latencies. To address these
challenges, Sandia and Oak Ridge National Laboratories have established the
Institute for Advanced Architectures and Algorithms (IAA). In this talk, I
will present some of the challenges and potential solutions for exa-scale
interconnection networks, which are being considered by IAA. |
Deployment Experiences,
Performance Observations, and Early Science Results on Ranger John (Jay) R. Boisseau, Ph.D. Director, The The Texas
Advanced Computing Center (TACC) at The University of |
NCSA Blue Waters: Preparing for
the Sustained Petascale System Robert Pennington, Urbana, IL, U.S.A. The NCSA
Blue Waters system will be installed at the |
The Impact of Petacomputing on
Models and Theories Thomas Lippert John von Neumann-Institute for Computing (NIC) FZ In 2008,
supercomputers have reached the Petaflop/s performance level. Machines likes
the IBM Blue Gene/P, the Los Alamos Roadrunner or the IBM Ranger at TACC
achieve their unprecedented power using O(100.000) cores. In my talk I will,
on the one hand, discuss the question if we have arrived at the limits of
scalability – I will present first scalability results from the Jülich Blue
Gene/P system with 64k cores –, and, on the other hand, argue how
Petacomputers with hundreds of thousands of processors might transform
science itself. |
Scalable Middleware for Large Scale Systems Barton P. Miller Computer Sciences Department I will discuss the problem of developing tools for large scale parallel
environments. We are especially interested in systems, both leadership class
parallel computers and clusters that have 10,000's or even millions of
processors. The infrastructure that we have developed to address this problem
is called MRNet, the Multicast/Reduction Network. MRNet's approach to scale
is to structure control and data flow in a tree-based overlay network (TBON)
that allows for efficient request distribution and flexible data reductions. The second part of this talk will present an overview of the MRNet
design, architecture, and computational model and then discuss several of the
applications of MRNet. The
applications include scalable automated performance analysis in Paradyn, a
vision clustering application and, most recently, an effort to develop our
first petascale tool, STAT, a scalable stack trace analyzer running currently
on 100,000's of processors. I will conclude with a brief description of a new fault tolerance
design that leverages natural redundancies in the tree structure to provide
recovery without checkpoints or message logging. |
Old problems never die – managing
the multi-programming mix Miron Livny Computer Sciences Department Old problems
never die; they just fade away as technologies and tradeoffs change. As the state of the art in hardware and
applications evolves further, they resurface.
When virtual memory was introduced almost 50 years ago, computer
systems had to find a way to prevent thrashing by controlling the number and
properties of the applications allowed to share their physical memory. The
recent proliferation of multi-core processors, usage of virtual machines and
deployment of complex I/O sub-systems require the development of similar
capabilities to control and manage at several scales the mix of applications
that share the compute and storage resources of today’s systems. |
Active Data: Blurring the distinction
between data and computation Tim Ho and David Abramson Clayton, The
amount of data being captured, generated, replicated and archived |
Using
Peer-to-Peer Dynamic Querying in Grid Information Services Domenico
Talia and Paolo Trunfio DEIS, Dynamic
querying (DQ) is a technique adopted in unstructured Peer-to-Peer (P2P) networks
to minimize the number of nodes that is necessary to visit to obtain the
desired number of results. In this talk we describe the use of the DQ
technique over a distributed hash table (DHT) to implement a scalable Grid
information service. The DQ-DHT (dynamic querying over a distributed hash
table) algorithm has been designed to perform DQ-like searches over DHT-based
networks. The aim of DQ-DHT is two-fold: allowing arbitrary queries to be
performed in structured P2P networks, and providing dynamic adaptation of
search according to the popularity of resources to be located. Through
the use of the DQ-DHT technique it is possible to implement a scalable Grid
information service supporting both structured search and execution of
arbitraries queries for searching Grid resources on the basis of complex
criteria or semantic features. |
Algorithms and scheduling techniques for
clusters and grids Yves
Robert Ecole
Normale Supérieure de Lyon, France In this talk we provide several
examples to |
Feedback
control for efficient autonomic solutions on the Grid Rizos Sakellariou This talk
will consider different approaches for |
Accidentally Using Grid Services Charlie
Catlett Maths
and Computer Science Division and Though
the term "grid" has fallen from the front page headlines, there is
an extremely active market of "grid services" - based on web
services and other standards - emerging. The web originally empowered
Internet users to create services and products with very little
infrastructure, and signs of success a decade ago included server meltdown
from high demand. Today one need not own any infrastructure at all to
launch a new service or product, and the combination of virtual and web
services offers not only near unlimited scaling but also reliability.
This talk will focus on a number of examples of new services,
illustrating that at least one measure of success is not only "ease of
use" but "accidental use" of transparent, but foundational,
services. |
From Grid Computing to Cloud Computing The evolution of the Grid Marketplace Avner
Algom The
Israeli Association of Grid Technologies Over the last few years we have seen grid
computing evolve from a niche technology associated with scientific and
technical computing, into a business-innovating technology that is driving
increased commercial adoption. Grid deployments accelerate application
performance, improve productivity and collaboration, and optimize the
resiliency of the IT infrastructure. Today, the maturity of the Virtualization
technologies, both at the VM and at the IT infrastructure levels, and the
convergence of the Grid, Virtualization and SOA concepts, enables the
business implementation of the Cloud Computing for utility and SaaS services. At last, the Grid Computing vision becomes a
reality: people that get electricity from their electrical outlet, on-demand,
can get applications, computing and storage services from the network,
on-demand. We can dynamically scale our computation and storage power, at no
time, and we pay only for what we use. This is going to change the marketplace as we know
it. |
Cloud
Computing for on-Demand Resource Provisioning Ignacio Llorente Distributed Systems
Architecture Group Universidad
Complutense de Madrid Madrid, Spain The aim of the presentation is to show the benefits
of the separation of resource provisioning from job execution management in
different deployment scenarios. Within an organization, the incorporation of
a new virtualization layer under existing Cluster and HPC middleware stacks
decouples the execution of the computing services from the physical
infrastructure. The dynamic execution of working nodes, on virtual resources
supported by virtual machine managers such as the OpenNEbula Virtual
Infrastructure Engine, provides multiple benefits, such as cluster
consolidation, cluster partitioning and heterogeneous workload execution.
When the computing platform is part of a Grid Infrastructure, this approach
additionally provides generic execution support, allowing Grid sites to
dynamically adapt to changing VO demands, so overcoming many of the obstacles
for Grid adoption. The previous scenario can
be modified so the computing services are executed on a remote virtual
infrastructure. This is the resource provision paradigm implemented by some
commercial and scientific infrastructure Cloud Computing solutions, such as
Globus VWS or Amazon EC2, which provide remote interfaces for control and
monitoring of virtual resources. In this way a computing platform could scale
out using resources provided on-demand by a provider, so supplementing local
physical computing services to satisfy peak or unusual demands. Cloud
interfaces can also provide support for the federation of virtualization
infrastructures, so allowing virtual machine managers to access resources from
remote resources providers or Cloud systems in order to meet fluctuating
demands. The OpenNEbula Virtual Infrastructure Engine is being enhanced to
access on-demand resources from EC2 and Globus-based clouds. This scenario is
being studied in the context of the RESERVOIR– Resources and Services
Virtualization without Barriers — EU-funded initiative. |
Network resource reservation and
virtualization for grid applications Marcelo Pasin INRIA, École Normale
Supérieure de Lyon Laboratoire de
l’informatique du parallélisme Lyon, France The
coordination of grid resource allocation often needs a service to
|
A Performance
Based Distribution Algorithm for Grid Computing Heterogeneous Tasks Abderezak
Touzene, Hussein AlMaqbali, Ahmed AlKindi, Khaled Day Department
of Recently
in [1] we proposed a performance based load-balancing algorithm for
independent tasks, which require similar computing need in the sense that the tasks are
almost identical. This paper extends the work and proposes a load
distribution algorithm for independent tasks with different computing
requirements including short and long tasks. We assume a preprocessing phase
of prediction of the number of instruction (TNI) needed for each task in the
grid. Our load distribution algorithm takes into account both the CPU speed
of the computing units and the TNI of different tasks. We design a simulation
model using steady-state, based on NS2 to study the performance of our load
distribution algorithm. Keywords: grid
computing, load-balancing, steady-state, resource management, performance
evaluation, simulation models. |
Parallel Data Mining from
Multicore to Cloudy Grids Geoffrey Fox We
describe a suite of data mining tools that cover clustering, Gaussian
modeling and dimensional reduction and embedding. These are applied to three
class of applications; Geographical information systems, cheminformatics and
bioinformatics. The data vary in
dimension from low (2), high (thousands) to undefined (sequences with
dissimilarities but not vectors defined). We use deterministic annealing to
provide more robust algorithms that are relatively insensitive to local
minima. We use embedding algorithms both to associate vectors with sequences
and to map high dimensional data to low dimensions for visualization. We
discuss the algorithm structure and their mapping to parallel architectures
of different types and look at the performance of the algorithms on three
classes of system; multicore, cluster and Grid using a MapReduce style
algorithm. Each approach is suitable in different application scenarios. |
Summary-based Distributed Semantic Database for Resource and Service
Discovery André Höing Electrical Engineering and
Computing Science Today's RDF
triple stores that are based on distributed hash tables (DHTs) distribute the
knowledge of all participating peers in the P2P network. They use hash values
of the subject, predicate, and object of each triple in order to identify
three nodes in the network that shall store a copy of the triple. Query
processors collect relevant triples by identifying responsible nodes using
the hash values of literals and constants occurring in the query. |
UNICORE 6 – A European Grid Technology Achim Streit Jülich Supercomputing
Centre (JSC) at Forschungszentrum The
development of UNICORE started back in 1997 with two projects funded by the German
ministry of education and research (BMBF). UNICORE is a vertically integrated
Grid middleware, which provides a seamless, secure, and intuitive access to
distributed resources and data and provides components on all levels of a
Grid architecture from an easy-to-use graphical client down to the interfaces
to the Grid resources. Furthermore, UNICORE has a strong support for
workflows while security is established through X.509 certificates. Since
2002 UNICORE is continuously improved to mature production ready quality and
enhanced with more functionalities in several European projects. Today
UNICORE is used in several national and international Grid infrastructures
like D-Grid and DEISA and is also providing access to the national
Supercomputer of the NIC in The talk
will give details about the new version of UNICORE 6, which is web-services
enabled, OGSA-based and standards-compliant. To begin with the underlying
design principles and concepts of UNICORE are presented. A detailed
architecture diagram shows the different components of UNICORE 6 and its
interdependencies with a special focus on workflows. This is followed by a
view on the adoption of common open standards in UNICORE 6, which allows
interoperability with other Grid technologies and a realisation of an open
and extensible architecture. The talk closes with some interesting use case
examples, where the UNICORE Grid technology is used. The
European UNICORE Grid Middleware is available as Open Source from http://www.unicore.eu. |
e-Science Applications on Grids -
The DEISA Success Story Wolfgang Gentzsch DEISA Distributed European
Infrastructure for Supercomputing Applications and We will
present selected compute and data intensive applications which Bio: Wolfgang Gentzsch DEISA, Duke University Wolfgang Gentzsch is Dissemination Advisor for the
DEISA Distributed European Initiative for Supercomputing Applications. He is
adjunct professor of computer science at |
Superlink-online
- delivering the power of GPUs, clusters and opportunistic grids to
geneticists M. Silberstein Technion-Israel Institute of Technology Haifa, Israel Genetic linkage analysis is a statistical tool used
by geneticists for mapping disease-susceptibility genes in the study of genetic
diseases. The analysis is based on the exact inference in very large
probabilistic (Bayesian) networks, which is often computationally hard
(ranging from seconds to years on a single CPU). We
present a distributed system for faster analysis of genetic data, called
Superlink-online. The system achieves high performance through parallel
execution of linkage analysis tasks over thousands of computational resources
residing in multiple opportunistic computing environments, aka Grids. It
utilizes the resources in many available grids, unifying thousands CPUs over
campus grids in the Technion and the Notably,
the system is available online, which allows geneticists to perform
computationally intensive analyses with no need for either While the grids potentially provide enormous amount
of computing power, we also explore an alternative approach of using Graphics
Processing Units (GPUs) to accelerate the genetic linkage computations. We
achieve up to two orders of magnitude speedups on average, and up to three
order of magnitude speedups on some particularly complex problem instances
versus the optimized application performance on a single CPU. The use of GPUs
is particularly appealing in the context of Community Grids,
considering the number of high performance GPUs available worldwide. |
Building Collaborative Applications for System-Level Science Marian Bubak 30-059 ACC CYFRONET AGH, Krakow, ul.
Nawojki 11, 30-950 A novel
approach to scientific investigations, besides analysis of individual
phenomena, integrates different, interdisciplinary sources of knowledge about
a complex system to obtain an understanding of the system as a whole. This
innovative way of research has recently been called system-level science [1].
Problem-solving
environments and virtual laboratories have been the subject of research and
development for many years [2]. Most of them are built on top of workflow
systems [3]. Their main drawbacks include limited expressiveness of the
programming model and lack of mechanisms for integration of computing
resources from grids, clusters and dedicated computers. The
ViroLab project [4] is developing a virtual laboratory [5] for research of
infectious diseases to facilitate medical knowledge discovery and provide
decision support for HIV drug resistance [6], and this virtual laboratory may
be useful in other
areas of system-level science. To
overcome the limitations of the programming methods, we have defined an
experiment plan notation based on a high-level scripting language - Ruby. For easy interfacing of
different technologies, we have introduced a grid object abstraction level
hierarchy [7]. Each grid object class is an abstract entity which defines the
operations that can be invoked from the script, each class may have multiple
implementations, representing the same functionality; and an implementation
may have multiple instances,running on different resources [8]. The
Experiment Planning Environment is an Eclipse-based tool supporting rapid
experiment plan development while Experiment Management Interface enables
loading and execution of experiments. The Experiment Repository stores
experiment plans prepared by developers and published for future usage, and
the laboratory database holds the obtained results.To enable high-level
programming, the virtual laboratory engine, called the GridSpace, includes
the Grid Operation Invoker which instantiates grid object representatives and
handles remote operation invocations. The GridSpace Application Optimizer is
responsible for optimal load balancing
on computational servers.The Data Access Service acquires data from remote
databases located in research institutions and hospitals. To meet the
specific requirements for exchanging biomedical information within such a
virtual environment, the solution introduced in DAS bases on existing Grid
technologies: Globus Toolkit, OGSA-DAI, and Shibboleth. The provenance
approach [9] in the ViroLab virtual laboratory brings together ontology-based
semantic modeling, monitoring of applications and the runtime infrastructure,
and database technologies, in order to collect rich information concerning
the execution of experiments, represent it in a meaningful way, and store it
in a scalable repository [10]. The
virtual laboratory has already been used to plan and execute a few
virological experiments, with various types of analysis of HIV virus
genotypes such as calculation of drug resistance based on virus genotype,
querying historical and provenance information about experiments, a drug
resistance system based on the Retrogram set of rules, data mining and
classification with Weka [5], and the molecular
dynamics NAMD application which has been installed on the CYFRONET EGEE site.
The
virtual laboratory provides an environment to collaboratively plan, develop
and use collaborative applications; it is dedicated for multi-expertise
task-oriented groups running complex computer simulations; its basic features
are: mechanisms for user-friendly experiment creation and execution,
possibility of reusing existing libraries,
tools etc., gathering and exposing provenance information, integration of
geographically-distributed data resources, access to WS, WSRF, MOCCA
components and jobs, secure access to data and applications. Acknowledgments
The
Virtual Laboratory is being developed at the Science
and CYFRONET AGH, Gridwise Technologies, Universiteit van Amsterdam, and HLRS
Stuttgart in the framework of the EU IST ViroLab and CoreGRID projects
as well as the related Polish SPUB-M and Foundation for Polish Science
grants. References [1] Exploration and IT Implications, IEEE Computer, vol. 39, no 11, 31-39,
2006 [2] K. Rycerz, M. Bubak, P.M.A. Sloot, V.
Getov: Problem Solving Environment for Distributed
Interactive Simulations in: Sergiei Gorlatch, Marian Bubak, and Thierry
Priol (Eds). Achievements in European Reseach on Grid Systems. CoreGRID
Integration Workshop 2006 (Selected Papers) ISBN-13: 978-0-387-72811-7; pp
55 - 66, Springer, 2008 [3] Y. Gil, E. Deelman, M. Ellisman, T. Fahringer, G. Fox, D. Gannon, C. Goble, M. Livny, L. Moreau, and
J. Myers. Examining the Challenges of Scientific Workflows. IEEE Computer vol 40, no 12 pp 24-32, 2007 [4] ViroLab - EU IST STREP Project 027446; www.virolab.org [5] ViroLab Virtual Laboratory, http://virolab.cyfronet.pl [6] P. M.A. Sloot, I. Altintas, M. Bubak, Ch.A. Boucher: From Molecule to Man: Decision
Support in Individualized E-Health, IEEE Computer vol. 39, no 11, 40-46,
2006 [7] T. Gubala, M. Bubak: GridSpace -
Semantic Programming Environment for the Grid, PPAM'2005, LNCS 3911,
172-179, 2006 [8] M. Malawski, M. Bubak, M. Placek, D.
Kurzyniec, V. Sunderam: Experiments with Distributed Component
Computing Across Grid Boundaries, Proc. HPC-GECO/CompFrame Workshop -
HPDC'2006, Paris, 2006 [9] D. de Roure, N.R. Jennings, N. Shadbolt,
The semantic grid: a future e-science infrastructure, Grid
Computing - Making the Global Infrastructure a
Reality, Wiley, 2003, pp. 437-470 [10] B.
Balis, M. Bubak, and J. Wach: User-Oriented Querying over Repositories of Data and
Provenance, In G. Fox, K. Chiu, and R. Buyya,
editors, Third IEEE International Conference on e-Science and Grid
Computing, e-Science 2007, IEEE Computer Society, 2007 |
DMOVER: Scheduled Data
Transfer for HPC Grid Workflows Derek Simmel Pittsburgh Supercomputing Center Pittsburgh, PA, U.S.A. TeraGrid
users have expressed a need for better tools to schedule and |
Grid Computing or the Internet of services? Opportunities and perspectives from research to business Antonio Congiusta NICE-ITALY, Cortanze,
Asti, Italy Experience has shown that solutions to
better enable organizations to take advantage of the benefits of Grid
computing, are based on clear identification of the requirements and the
application of the best available standardized and reliable technologies. Relevant examples of such principle with
related best practices can be extracted from some of the success stories that
recently have involved EnginFrame in the Oil & Gas industry, the Energy
and Automotive sectors, HPC support from collaboration facilities to
infrastructure provision and management, and also some fruitful cooperations
with strategical partners. In particular, beyond to well established
HPC activities within a primary European consortium for providing a
production quality infrastructure, a new trend has been undertaken towards
the integration of collaboration facilities to HPC environments. Quite
interesting are also the activities devoted to enable for workflow management
and distributed visualization, some of which are part of European-wide
research projects. From all such experiences we can
envision as future of the Grid an always strong evolution towards
interoperable key services, within a scenario in which comprehensive and
all-inclusive software is ever less important. In such a scenario, a key role
is played by integration technologies capable of homogenizing and enforcing
service interactions and access. |
e-Research
& Grid computing in David Abramson Clayton, Over the
past few years the Australian government has performed a major review of
its research infrastructure needs, from hard technological areas to the
social sciences. Along with this review, they have investigated the
electronic platforms required to support these various disciplines.
What has evolved is an grid computing strategy called "Platforms
for Collaboration" that addresses computation, networking and data
management. In addition to this, various computer science groups are
developing grid technologies that underpin this platform. In this talk I
will give an over of the Australian e-Research agenda and highlight a
few major research activities in grid computing. |
Kihyeon
Cho e-Science Division Korea Institute of Science and Technology Information Daejeon,
305-806, For Grid and e-Science in |
Atul Gurtu Tata Institute of Fundamental
Research Grid technology has changed the way advanced
research is being conducted today. In |
National Grid Initiative of Anatoly Sachenko Department of Information
Computing Systems and Control Uniting of
the existing Grid segments and supercomputer centers in scientific and educational
areas into joint Ukrainian National
Grid Initiative(UNGI) and the issues of UNGI integration into the European
Grid infrastructure are considered in this paper. The peculiarities of Grid segment at National Academy of
Science as well as the UGrid Project of Ministry of Education and Science are
described too. It’s stressed on the joint project UNGI for EGI and other
integration possibilities within INTAS, NATO and Frame 7 programs. Finally an
advanced approach for security strengthening
in Grid-systems is proposed. |
Per Öster CSC – Finnish IT Center for
Science The European Grid Initiative (EGI) has as goal to
ensure a long-term sustainability of grid infrastructures in |
Building Shared High Performance
Computing Infrastructure for the Biomedical Sciences Marcos Athanasoulis, Dr.PH, MPH In recent years high
performance computing has moved from the sidelines to the mainstream of
biomedical research. Increasingly researchers are employing
computational methods to facilitate their wet lab research. Some
emerging laboratories and approaches are based on a 100% computational
ramework. While there are many lessons to be learned from the
computational infrastructure put into place for the physical and mechanical
sciences, the character, nature and demands of biomedical computing differ
from the needs of the other sciences. Biomedical computational
problems, for example, tend to be less computationally intensive but more
“bursty” in their needs. This creates both an opportunity (it is easier
to meet capacity needs) and a challenge (job scheduling rules are more
complicated to accommodate the bursts). |
ViroLab: Distributed Decision Support in a virtual laboratory for
infectious diseases P. Sloot In future
years, genetic information is expected to become increasingly significant in
many areas of medicine. This expectation comes from the recent and
anticipated achievements in genomics, which provide an unparalleled
opportunity to advance the understanding of the role of genetic factors in
human health and disease, to allow more precise definition of the non-genetic
factors involved, and to apply this |
Processing of Large-Scale Biomedical Images on a Cluster of Multi-Core
CPUs and GPUs Umit Catalyurek Department of Biomedical
Informatics The As microprocessor
manufacturers strain to continue to increase performance, multi-core chips
are quickly becoming the norm. The demand in computer gaming industry also
brought us GPUs as an alternative fast, general purpose, streaming
co-processors. Commodity
GPUs and multi-core CPUs bring together an unprecedented combination
of high performance at low cost, and provide an ideal environment for
biomedical image analysis applications. In this
talk we will present our ongoing efforts on developing optimized biomedical
image analysis kernels for heterogeneous multi-core CPUs and GPUs. We will
also present how a cooperative cluster of multi-Core CPUs and GPUs can be
efficiently used for large scale biomedical image analysis. |
Grid
Computing for Financial Applications M. Al-Baali§,
P. Beraldi*, L. Grandinetti*, G. Aloisio^ I.
Epicoco^, A. Violi**, C. Figŕ Talamancaç § Dept. of Mathematics and
Statistics, * Department of Electronics,
Informatics and Systems, * * CESIC - ^ ç Innova spa In recent
years financial operators have shown an increasing interest in quantitative
tools able to efficiently measure, control and manage risk. Such an interest
is motivated by the necessity to operate in a very competitive and volatile
environment with a high level of
complexity increased by the
globalization of the economic activities and the continuous
introduction of innovative financial products. The complexity of the problems
to deal with and the necessity to operate in real time has highlighted the
serious computational constraints imposed by conventional numerical
platforms, prompting the need to take advantage of high performance computing
systems. In this
talk we present a prototypal system designed to support financial operators
in investment decisions concerning the strategic asset allocation
problem. The system has been designed
and tested within the European Project BEINGRID. At the
core of the system is the formulation of sophisticated optimization models
able to capture with an increasing level of realism with respect to
traditional approaches, the specific features of the applicative problem.
Moreover, the system is based on the integration of advanced scenario
generation procedures and efficient methods to solve the resulting huge sized
problems. The
system has been deployed on the SPACI grid infrastructure. In particular, an
user – friendly web grid environment
has been realized by using the GRB technology for the resource management and
the GRelC services for distributed data. |
Data Issues in a challenging HPC application
to Climate Change Giovanni Aloisio University of
Salento Lecce, Italy Earth
Science is strongly becoming a data intensive and oriented activity.
Petabytes of data, big collections, huge datasets are continuously produced,
managed and stored as well as accessed, transferred and analyzed by several
scientists and researchers at multiple sites. From the data grid perspective,
a key element to search, discover, manage and access huge amount of data
stored within distributed storages is the related data and metadata
framework. A new supercomputing centre, the Euro-Mediterranean Centre for
Climate Change (CMCC), was recently created by the Italian Government to
support research on Climate Change. The SPACI Consortium, one of the main
CMCC Associate Centres, provides know-how and expertise on High Performance
and Grid Computing. The GRelC Middleware (provided by SPACI Consortium) has
been recently adopted as part of the CMCC Data Grid framework in order to
provide a secure, transparent and scalable grid enabled metadata management
solution. We
present the CMCC initiative, the supercomputing facility as well as data grid
architectural and infrastructural issues concerning the adopted grid
data/metadata handling systems. |
Tim David Centre for “A
Heterogeneous Computing Model for a Grand Challenge Problem” |
The e-Science for High Energy Physics Kihyeon Cho, Ph.D. KISTI (Korea Institute of Science
and Technology Information) The
e-Science for High Energy Physics is to study High Energy Physics (HEP) any
time and anywhere even if we are not on-site of accelerator laboratories. The
components are 1) data production, 2) data processing and 3) data analysis
any time and anywhere. The data production is to do remote control and take
shifts remotely. The data processing is to run jobs anytime, anywhere using
Grid farms. The data analysis is to work together to publish papers using
collaborative environment. We apply this concept to LHC experiment at CERN
and Tevatron experiment at Fermilab. It this talk we will present the current
status and embodiment of the idea. |
A HPC infrastructure at the service of Scientific Research in Italy Giovanni
Erbacci CINECA -
System and Technology Deparment CINECA Inter-University Consortium,
Casalecchio di Reno, Italy State of the art HPC infrastructures are
fundamental to support scientific research and to advance science at European
level. Since many years, at Italian level, CINECA has been able to assure to
the scientific community a competitive
advantage by putting into timely production advanced HPC systems that have proven very wide applicability
and success. The CINECA HPC infrastructure de facto represents
the national facility for supercomputing and the CINECA HPC systems are part of the Italian research
Infrastructures system, integrated by means of the Italian academic and
research network facility (GARR). In this work we present the CINECA HPC
infrastructure, its evolution, and the
service model. Moreover, we outline
the CINECA role in the context of the
main HPC Infrastructure projects,
operating at the European level: DEISA, PRACE and HPC-Europa. DEISA is a consortium between the most advanced HPC centres in This infrastructure is mainly intended to support
challenge scientific applications by integrating and making easily accessible
supercomputers in different centres. PRACE is a feasibility project intended to build
the next generation of challenge HPC infrastructure and services at European level. The infrastructure will
consist of a limited number (3 to 5) of
PetaFlop/s class HPC systems integrated in a network of HPC systems on
a pyramidal model basis, with three different layers (European, National and Regional) in the European HPC eco-system. HPC-Europa supports the human network of
knowledge, experiences and expertise exchange, in the context of the
scientific research communities using
advanced HPC systems. HPC-Europa factively promotes such mission
supporting the mobility of the European researchers among the main research
institutions, and providing the access
to the computational resources offered by the main European HPC
infrastructures. |
PANELS