Folding@Clusters
This page contains notes about Folding@Clusters and about Earlham's involvement in Folding@Home (fah, folding at home, fahclient, etc.) more generally.
The Earlham CS Folding@Home team page contains our most recent data.
Contents
Specific resources
Folding@Home v8.4 (May 2025)
As of when these notes are created, most of our machines have F@H v7 installed on them and F@H just released v8.4, which includes a lot of changes on how users can manage F@H. I hope these notes make the switch from v7.x to v8.4 a bit easier for whoever is doing it.
- You will need to reinstall F@H v8.4 by copying the package link here and run
wget [PACKAGE_LINK]
. For v8.4, you only need to download 1 package. - After downloading, install the package:
dpkg -i --force-depends [PACKAGE]
. - Configure F@H at
/etc/fah-client/config.xml
. An example of the v8.4 config file that I made:
<config> <!-- User Information --> <user v='Seawave_Earlham'/> <team v='577'/> <account-token v=xxxxxxxxxxxxxxxx> <!-- Folding Slots --> <slot id='0' type='CPU'> <cpus v='2'/> </slot> </config>
Apply changes according to the machines that you are working on. Leave the account-token line commented out for now.
- Read and follow the Client Guide. You will need to create an account for headless machines like our cluster, so please follow the instruction and set up an account on the new web control browser. Passkey is not required when creating an account, but you can read up about it here if you decide to have one for our clusters.
- Once you have an account, go to Account Settings and copy the token. In the
/etc/fah-client/config.xml
file, add the token to the account-token field. - After the configuration with the account token, run
systemctl restart fah-client
orsystemctl start fah-client
. The token connects the machines to the account and the machines will show up on the web browser. - Now, you can choose FOLD ALL for the machines to start receiving works and folding. You can configure the machines straight from the web control.
Original notes from the cluster website about Folding@Clusters
Folding@Clusters
This software tool is designed to make high performance computational resources easily available to chemists and biologists for running simulations of large bio-molecules using open source molecular dynamics packages.
Josh Hursey, Josh McCoy, Charles Peck, and John Schaefer gave a presentation on this work at SuperComputing04, at the Purdue University Research area, in November, 2004. We also presented this work as a poster at SIAM's Computational Science and Engineering conference in February, 2005.
An article based on this work appears in the November, 2005 issue of Dr. Dobb's Journal.
The abstract for our SC04 submission follows:
Instead of traditional, tightly coupled massively parallel computing, current distributed computing projects such as SETI@home or Folding@Home use a client-server model to perform embarrassingly parallel computing, allowing for one to tap resources (hundreds of thousands of CPUs in PCs throughout the world) impossible to obtain by other means. However, certain algorithms could greatly benefit from a hybrid approach, combining the massive resources available to distributed computing with the tight coupling traditionally found only in supercomputers.
Towards this end, Folding@Clusters is an adaptive framework for harnessing tightly coupled cluster resources for protein folding research. It combines capability discovery, load balancing, process monitoring, and checkpoint/re-start services to provide a platform for molecular dynamics simulations on a range of grid-based parallel computing resources.
The raw computing power available for scientific inquiry continues to grow while the abstraction level of the tools available to scientists does not advance in a similar manner. Folding@Clusters provides chemists, investigating protein folding, with a high-level interface to a variety of parallel compute architectures, e.g. lab clusters, Beowulf clusters, large SMP machines, and clusters of SMP machines.
Folding@Clusters uses open source building blocks, such as the GROMACS molecular dynamics package and the LAM-MPI communications library, to provide the lowest-level functionality. Building on this foundation we construct a three-tier architecture: cluster, node, and science core, which provides a basis on which to abstract the process of performing a molecular dynamics simulation. This includes work unit preparation, distribution, and result aggregation, on a compute resource with arbitrary capabilities (CPU speed, CPU count, memory, etc.)
Other notes
- See config files /etc/fahclient/config.xml on layout or Whedon’s compute nodes, /var/lib/fahclient is working directory, tail the log file to see what’s going on - there’s a lot of waiting for a work unit
- Team stats for 577 - give resource name as part of the config, so Whedon and layout are tracked separately
- Init script: /etc/init.d/fahclient --help # Mostly it sits and runs, starts on reboot only on compute nodes