Tutorials

Three Tutorials will be held on September 3. All of these are free for registered participants.

TimeTutorialRoom
12:00-14:00Tutorial 1
Data Mining and Knowledge Discovery
by Ho To Bao
Conference room 6
14:15-16:15Tutorial 2
Grid Computing & OGSA
by Mr. K. Sugimoto and Ms. Y. Sawatani
Conference room 6
16:30-18:30Tutorial 3
Actuality and Trend of High Performance Computing and Parallelizing Compilations
by Minyi Guo
Conference room 6



Tutorial 1
"Data Mining and Knowledge Discovery"

by Professor Ho To Bao, JAIST

Abstract:

Knowledge discovery and data mining (KDD) has become an active and growing interdisciplinary area of information technology. It is not only of academic interest, but also of great practical significance. In the past few years, KDD attracted a large number of researchers and practitioners from many disciplines, e.g., machine learning, databases, AI, statistics, data visualization, high performance computing, etc. KDD is important simply because as long as one has a database, one can make use of this technology to uncover useful information.

This tutorial provides an intuitive introduction to basic concepts and techniques of KDD. It will cover the following main topics:

  • Basic concepts of KDD
  • The KDD process
  • Main KDD techniques: classification, association, clustering
  • Relation between KDD and high performance computing
  • KDD applications: case studies in medicine data mining
  • Challenges and some trends in KDD research
Tutorial 2
"Grid Computing & OGSA"

by Mr. K. Sugimoto and Ms. Y. Sawatani, IBM

Abstract:

Grid computing, that is distributed computing over the Internet using open standards, has recently attracted considerable attention. The capability is to form virtual, collaborative organizations that share applications and data in an open heterogeneous computing environment, and to aggregate large amounts of computing resources which are geographically dispersed in order to tackle large problems and workloads as if all the computing resources were located in a single site. Grid computing is already emerging in a number of key applications. In particular, the scientific and technical communities are using it to collaborate across institutions around the world in such application areas as high-energy physics, life sciences, and engineering design. In this tutorial, we first present an overview of grid computing that covers:

  • What is Grid computing?
  • Why is Grid computing important?
  • What are Grid applications in research and industry?

Then we focus on the Open Grid Services Architecture (OGSA), an evolving Globus architecture integrating Grid computing and Web services concepts and technologies. We cover:

  • What is the objective of OGSA?
  • What is the infrastructure focus area?
  • What is the OGSA service model?
  • What is the OGSA framework?
Additionally, we introduce our experience with some grid projects.



Tutorial 3
"Actuality and Trend of High Performance Computing and Parallelizing Compilations"

by Professor Minyi Guo, Aizu University

Abstract:

In recent years, high performance computing is more and more widely used in various aspects of our real life. There have been major efforts in developing approaches to parallelization of scientific applications. Parallelizing compilers play an important role by automatically customizing programs for complex processor architectures, improving portability and providing high performance to non-expert programmers.

In this tutorial, we first retrospect the techniques used in current parallelizing compilations. The parallelizing compilations can be classified as automatic parallelization and parallelizing compilers for parallel programming languages. For the automatic parallelization, we summarize various parallelism detection techniques such as data dependence, loop restructuring, data distribution, symbol analysis, inter-procedural analysis, etc. For the parallelizng compilers for parallel programming languages, we introduce the implementing techniques of HPF and OpenMP, which are most typical data parallel language for distributed memory multicomputers and task parallel language for shared memory multiprocessors, respectively.

Additionally, we give some experience for high performance programming with MPI, HPF, and OpenMP in the scientific computations. With our experience, we will prompt the better parallel solution for different kinds of computation problems.

We also prospect the trend of techniques of high performance computing and parallelizing compilations. We especially introduce the development of multigrain parallelism techniques, commutativity analysis of parallelizing compilers, and the development of HPF-OpenMP combination compilers.