Introduction to scientific computing

Lesson 1/75

Scientific computing

Scientific computing is a rapidly growing multidisciplinary field that uses advanced computing capabilities to understand and solve complex problems. It heavily utilizes HPC clusters, where clusters typically have a large number of computers (often called ‘nodes’) and, in general, most of these nodes would be configured identically.

The term cluster computing is used to denote nothing but two or more computers that are networked together to provide solutions as required. However, this idea should not be confused with a more general client-server model of computing as the idea behind clusters is quite unique.

A cluster of computers joins computational powers of the compute nodes to provide a more combined computational power. Therefore, as in the client-server model, rather than a simple client making requests of one or more servers, cluster computing utilize multiple machines to provide a more powerful computing environment perhaps through a single operating system.

In its simplest structure, HPC clusters are intended to utilize parallel computing to apply more processor force for the arrangement (solution) of a problem. HPC clusters typically have a large number of computers (often called ‘nodes’) and, in general, most of these nodes would be configured identically. Though from the out side the cluster may look like a single system, the internal workings to make this happen can be quite complex.

Computer clusters emerged as a result of convergence of a number of computing trends including the availability of low-cost microprocessors, high speed networks, and software for high-performance distributed computing. They have a wide range of applicability and deployment, ranging from small business clusters with a handful of nodes to some of the fastest supercomputers in the world!

Benefits

In scientific computing environments, coders derive their computational power by exploiting parallelism. Scientific programs for HPC systems must be splitted up into many smaller sub-programs which can be executed in parallel on different processors. Scientific computing systems can offer parallelism at a much larger scale, with 100’s or 1000’s, or even millions of tasks running concurrently.However, writing parallel software can be challenging, and many existing software packages do not already support parallelism & may require development.

Scientific computing on HPCs are userful when you have:

  • A program that can be recompiled or reconfigured to use optimized numerical libraries that are available on HPC systems but not on your own system;
  • You have a parallel problem, e.g. you have a single application that needs to be rerun many times with different parameters;
  • You have an application that has already been designed with parallelism;
  • To make use of the large memory available;
  • Your problem solutions require backups for future use. HPC facilities are reliable and regularly backed up.

However, it is not so good when:

  • You have a single threaded job which will only run one job at a time (typical of MatLab users);
  • You rely on DBMS/ databases;
  • You have a lot of data to transfer between your local machine and the HPC on a continuous basis (e.g. per job);
  • You need to have a GUI to interact with your program.

Class Lessons

Introduction Introduction to scientific comput    Virtual cluster    Cluster Computing Cluster computers Cluster computer components How to access? Data transfer Linux Foundation for Scientific C Introduction File and folder commands Display contents of a file    Wildcards    Redirection and pipes File system security access Processes and Jobs Useful UNIX commands    Compiling UNIX software packages Scientific Software Scientific software stack Software modules Scheduling jobs Portable Batch System Introduction to PBS PBS basic commads PBS job submission    PBS Statistics PBS jobs management PBS job states PBS variables PBS job arrays SLURM - Workload Manager Introduction to Slurm Slurm commands Slurm job submission    Slurm array job    C/C++ Foundation Variables, Literals and Constants Data Types Input and Output Operators Comments    Conditions Switch Loops Goto Functions Recursion Arrays Strings Structures Pointers C++ STL STL Vectors Lists Map Queue Set Stacks Pair C++ OpenMP OpenMP OpenMP- Hello world!    Open MP - reduction OpenMP - section C++ MPI (Message Passing Interfac MPI - Message Passing Interface MPI program structure MPI - Hello world! MPI send and receive C++ CUDA GPUs - graphics processing units CUDA programming overview CUDA Threads, Blocks and Grid CUDA memory model CUDA Kernel CUDA Hello world! CUDA square an array of numbers CUDA vector addition CUDA matrix multiplication C++ Boost Library Boost Boost Multiprecision C++ Scientific Computing MKL - Matrix Multiplication LAPACK - Linear Algebra Conclusion Conclusion and further readings

Quiz

GDPR

When you visit any of our websites, it may store or retrieve information on your browser, mostly in the form of cookies. This information might be about you, your preferences or your device and is mostly used to make the site work as you expect it to. The information does not usually directly identify you, but it can give you a more personalized web experience. Because we respect your right to privacy, you can choose not to allow some types of cookies. Click on the different category headings to find out more and manage your preferences. Please note, that blocking some types of cookies may impact your experience of the site and the services we are able to offer.