Preview

High Performance Computing: Linux Cluster

Powerful Essays
Open Document
Open Document
4104 Words
Grammar
Grammar
Plagiarism
Plagiarism
Writing
Writing
Score
Score
High Performance Computing: Linux Cluster
HIGH PERFORMANCE COMPUTING: DESIGN, BUILD AND BENCHMARK A LINUX CLUSTER by GOH YOKE LOONG KEM 050024 Semester 1 Session 2008/09

Final year research project report submitted to the Faculty of Engineering University of Malaya in partial fulfillment of the requirement to be Bachelor’s degree of Engineering

DEPARTMENT OF MECHANICAL ENGINEERING ENGINEERING FACULTY UNIVERSITY OF MALAYA October 2008

ABSTRACT Nowadays almost every industry needs fast processing power especially in engineering field. Manipulating high-resolution interactive graphics in engineering, such as in aircraft engine design, has always been a challenge in terms of performance and scalability because of the sheer volume of data involved.
[1]

Linux clustering is

popular in many industries these days. With the advent of clustering technology and the growing acceptance of open source software, supercomputers can now be created for cheaper cost of traditional high-performance machines. Due to these conditions compromise, the number, variety, and specialized configurations of these machines are increasing dramatically with 32 – 128 node clusters being commonplace in science labs.
[2]

As a result, the intent of this research is to use the latest open source

software and computers available in the computer laboratory of Mechanical Engineering Department, University of Malaya to design and build a High Performance Linux Cluster. This paper will present the clustering fundamentals and details of how to setup the cluster. High performance cluster is mostly parallel programming. This paper shows how to run parallel programming with Message Passing Interface (MPI). High-quality implementations for MPI are freely available for FORTRAN, C, and C++ for Linux. In this research, MPICH 2 is used as MPI implementation. Extensive research will be carried out by benchmarking the performance of the cluster with the standard test codes. The results gained will use to compare with the existing

You May Also Find These Documents Helpful

  • Powerful Essays

    [4] Storage Conference. The Hadoop Distributed File System http://storageconference.org/ 2010/ Papers/ MSST/Shvachko.pdf [5] A Tutorial on Clustering Algorithms. K-Means Clustering http://home.dei.polimi.it/matteucc/ Clustering/ tutorial_html/kmeans.html [6] International Journal of Computer Science Issues. Setting up of an Open Source based Private Cloud http://ijcsi.org/papers/IJCSI-8-3-1-354-359.pdf [7] Eucalyptus. Modifying a prepackaged image http://open.eucalyptus.com/participate/wiki/modifyi ng-prepackaged-image [8] Michael G. Noll. Running Hadoop On Ubuntu Linux (Single-Node Cluster) http://www.michaelnoll.com/tutorials/running-hadoop-on-ubuntu-linuxsingle-node-cluster/ [9] 8K Miles Cloud Solutions. Hadoop: CDH3 – Cluster (Fully-Distributed) Setup http://cloudblog.8kmiles.com/2011/12/08/hadoopcdh3-cluster-fully-distributed-setup/ [10] Apache Mahout. Creating Vectors from Text https://cwiki.apache.org/MAHOUT/creatingvectors-from-text.html…

    • 3006 Words
    • 13 Pages
    Powerful Essays
  • Satisfactory Essays

    “This article describes programmability studies undertaken by Sun Microsystems in its HPCS participation.” FORTRAN has been the preferred language among the High Performance Computing (HPC) community, it is less likely that every programmer would have ended up preferring FORTRAN and more likely that the problems with the original source codes have more to do with reasons than the limitations of existing programming languages. The need for HPC benchmarks in the new languages could reduce source-code volume substantially, but rewriting these benchmarks even in FORTRAN could achieve similar source-code reductions and corresponding improvements in expressivity. This article mentions that the programmer rewrote many of HPC benchmarks and applications using modern FORTRAN taking into account the human costs of software development: programmability and associated characteristics such as readability, verifiability, and maintainability. “These are important considerations; although copy-and-paste is a fast way of writing lines of code, it degrades readability and increases maintenance costs.”…

    • 630 Words
    • 2 Pages
    Satisfactory Essays
  • Best Essays

    IBM SUPERCOMPUTER, WATSON

    • 2209 Words
    • 9 Pages

    The ability to coordinate all of these processors into one functioning logarithmic unit required a group of engineers from IBM to develop a specialized kernel-based virtual machine implementation with the ability to process eighty Tera-flops per seconds . The software that allowed all of this to occur is called Apache Hadoop. Hadoop is an open source framework software that is used to organize and manage grid computing environments. Since the theoretical limit of processors with current technology is set at a central processing unit (CPU) clock speed of three giga-hertz, a software model to enhance parallel processing for supercomputers had to be developed. With the use of Hadoop the programmers at IBM were able to more easily write applications for Watson that benefitted and took advantage of parallel processing to increase the speed at which problems could be solved and questions could be answered. The main reason why this makes things faster is the fact that one question can be researched in multiple paths at one time using parallel processing paths…

    • 2209 Words
    • 9 Pages
    Best Essays
  • Better Essays

    Systems with high performance server are used by scientists, engineers and analysts to analyse large quantity of data. Systems having high performance servers range from server farms to super computers. A summary of different industries and application are provided in Table I.…

    • 1229 Words
    • 5 Pages
    Better Essays
  • Good Essays

    WILEY SERIES ON PARALLEL AND DISTRIBUTED COMPUTING SERIES EDITOR: Albert Y. Zomaya Parallel & Distributed Simulation Systems / Richard Fujimoto Surviving the Design of Microprocessor and Multimicroprocessor Systems: Lessons Learned / Veljko Milutinovic Mobile Processing in Distributed and Open Environments / Peter Sapaty Introduction to Parallel Algorithms / C. Xavier and S.S. Iyengar Solutions to Parallel and Distributed Computing Problems: Lessons from Biological Sciences / Albert Y. Zomaya, Fikret Ercal, and Stephan Olariu (Editors) New Parallel Algorithms for Direct Solution of Linear Equations / C. Siva Ram Murthy, K.N. Balasubramanya Murthy, and Srinivas Aluru Practical PRAM Programming / Joerg Keller, Christoph Kessler, and Jesper Larsson Traeff Computational Collective Intelligence / Tadeusz M. Szuba Parallel & Distributed Computing: A Survey of Models, Paradigms, and Approaches / Claudia Leopold Fundamentals of Distributed Object Systems: A CORBA Perspective / Zahir Tari and Omran Bukhres Pipelined Processor Farms: Structured Design for Embedded Parallel Systems / Martin Fleury and Andrew Downton Handbook of Wireless Networks and Mobile Computing / Ivan Stojmenoviic (Editor) Internet-Based Workflow Management: Toward a Semantic Web / Dan C. Marinescu Parallel Computing on Heterogeneous Networks / Alexey L. Lastovetsky Tools and Environments for Parallel and Distributed Computing Tools / Salim Hariri and Manish Parashar Distributed Computing: Fundamentals, Simulations and Advanced Topics, Second Edition / Hagit Attiya and Jennifer Welch Smart Environments: Technology, Protocols and Applications / Diane J. Cook and Sajal K. Das (Editors) Fundamentals of Computer Organization…

    • 97856 Words
    • 392 Pages
    Good Essays
  • Powerful Essays

    It is now clear that silicon based processor chips are reaching their physical limits in processing speed, as they are constrained by the speed of electricity, light, and certain thermodynamic laws. A viable solution to overcome this limitation is to connect multiple processors working in coordination with each other to solve grand challenge problems. Hence, high performance computing requires the use of Massively Parallel Processing (MPP) systems containing thousands of powerful CPUs. A dominant representative computing system (hardware) built using MPP approach is C-DAC’s PARAM supercomputer. By the end of this century, all high performance systems will be parallel computer systems. High-end super computers will be the Massively Parallel Processing (MPP) systems having thousands of processors interconnected. To perform well, these parallel systems require an operating system radically different from current ones. Most researchers in the field of operating systems (including PARAS microkernel designers!) have found that these new operating systems will have to be much smaller than traditional ones to achieve the efficiency and flexibility needed. The solution appears to be to have a new kind of OS that is effectively a compromise between having no OS at all and having a large monolithic OS that does many things that are not needed. At the heart of this approach is a tiny operating system core called a microkernel. Dominant representative operating systems built using microkernel approach are Mach and C-DAC’s PARAS microkernel. This chapter presents an overview of parallel computing in general and correlates all those concepts to the PARAM and PARAS advented by the Centre for Development of Advanced Computing (CDAC). It starts with the discussion on need of parallel systems for High Performance Computing and Communication (HPCC). It also presents an overview of PARAM family of…

    • 3478 Words
    • 14 Pages
    Powerful Essays
  • Powerful Essays

    Pso Paralellization

    • 928 Words
    • 4 Pages

    S. Garcia Computer Science Department University of Oviedo Campus de Viesques Office 1.b.15 33206, Gijon Asturias Spain Tel.: +34-985-182487 Fax: +34-985-182156 E-mail: sgarcia@uniovi.es F. Reid Edinburgh Parallel Computing Centre Edinburgh Sotland Tel.: +44(0) 131-451-3410 Fax: +44(0)-131-650-6555 E-mail: fiona@epcc.ed.ac.uk D. Corne Heriot Watt University Edinburgh Sotland E-mail: dwcorne@macs.hw.ac.uk…

    • 928 Words
    • 4 Pages
    Powerful Essays
  • Powerful Essays

    Parallel Computing

    • 3147 Words
    • 13 Pages

    The highest level of parallel processing is conducted among multiple jobs or programs through multiprogramming, time-sharing, and multiprocessing. This presentation covers the basics of parallel computing. Beginning with a brief overview and some concepts and terminology associated with parallel computing, the topics of parallel memory architectures, Parallel computer architectures and Parallel programming models are then explored.…

    • 3147 Words
    • 13 Pages
    Powerful Essays
  • Better Essays

    In order to achieve these goals, the working set of the application should be optimal. The way to achieve an optimal working set is via data clustering. With good data clustering more data can be accessed in fewer pages; thus a high data density rate is obtained. A higher data density results in a smaller working set as well as a better chance of cache affinity. A smaller working set results in fewer page transfers. The following sections in this paper will explain several clustering patterns/techniques for achieving better performance via cache affinity, higher data density and a smaller…

    • 1188 Words
    • 5 Pages
    Better Essays
  • Good Essays

    The purpose of this post is to examine the introduction of parallel computing and the challenges of software development for Parallel execution environment. First I will introduce the idea of parallel computing and up next I will present and evaluate the challenges of parallel computing along with their solutions and finally some conclusion will be drawn.…

    • 988 Words
    • 4 Pages
    Good Essays
  • Good Essays

    Supercomputer

    • 311 Words
    • 2 Pages

    supercomputer is a computer that is at the frontline of current processing capacity, particularly speed of circulation. Supercomputers introduces in the 1960s were designed primarily by Seymour Cray at Control Data Corporation (CDC), and led the market into the 1970s until Cray left to form his own company, Cray Research. He then took over the supercomputer market with his new designs, holding the top spot in supercomputing for five years (1985-1990). In the 1980s a large number of smaller competitors entered the market, in parallel to the creation of the minicomputer market a decade earlier, but many of these disappeared in the mid-1990s “supercomputer market crash”. Today, supercomputer are typically one-of-a-kind custom designs produced by “traditional” companies such as Cray, IBM and Hewlett-Packard, who had purchased many of the 1980s companies to gain their experience. The IBM Roadrunner, located at Los Alamos National laboratory, is currently the fastest supercomputer in the world. The term supercomputer itself is rather fluid, and today’s supercomputer tends to become tomorrow’s ordinary computer. CDC’s early machines were simply very fast scalar processors, some ten times the speed of the fastest machines offered by other companies. In the 1970s most supercomputers were dedicated to running a vector processor, and many of the newer players developed their own such processor at a lower price to enter the market. The early and mid-1980s saw machines with a modest number of vector processors working in parallel to become the standard. Typical numbers of processors to massive parallel processing systems with thousand of “ordinary” CPUs, some being off the shelf units and others being custom design. Toda, parallel designs are based on “off the shelf” server-class microprocessors, such as the PowerPC, Opteron, or Xeon, and most modern supercomputers are now highly-tuned computer cluster using commodity processors, combined with custom interconnects.…

    • 311 Words
    • 2 Pages
    Good Essays
  • Good Essays

    Types of Computers

    • 418 Words
    • 2 Pages

    The types of computers range from the Hybrid to the Analog types. The computers you come across in the daily course of your day range from laptops, palmtops and towers, to desktop computers, to name a few. But the very word “computers” reminds one of the desktop computers used in offices or homes. Different categories of computes have been devised in keeping with our varied needs.…

    • 418 Words
    • 2 Pages
    Good Essays
  • Good Essays

    Virtual Computing

    • 573 Words
    • 2 Pages

    In the world of virtualization there are two kingpins; VMware and Microsoft. As I researched these two vendors it has become apparent that comparing these two giants would be like comparing apples and oranges; like apples and oranges are fruit, and VMware and Microsoft make virtualization software, they are different altogether. This happens to be the scenario that I have seen, not only between the two companies, but in blog posts and forums.…

    • 573 Words
    • 2 Pages
    Good Essays
  • Powerful Essays

    Generic Benchmarking

    • 6770 Words
    • 28 Pages

    M-Core is on the brink of creating one of the most innovative technological advances of the 21st century. The super processing chip will enhance the planet with the ability to have faster computers and decrease the energy used to make computers function. On a worldwide scale, this advancement is remarkable.…

    • 6770 Words
    • 28 Pages
    Powerful Essays
  • Satisfactory Essays

    Recent, revolutionary integrated circuit technology improvement coupled with computer architecture innovations have resulted to a speed of CPU performance development unprecedented since the middle of the 20th century. Computer architecture innovations has continuously improve Microprocessor-based machines at a speed range of 1.5 and 2 times per year for the last seven years. Improvement speed for minicomputers or mainframes computers are approximately 25% every year. Seemingly, recent…

    • 399 Words
    • 2 Pages
    Satisfactory Essays