4 | | * [http://fedora.redhat.com Fedora 8] |
5 | | * [http://www.perceus.org/portal/project/warewulf Warewulf Cluster Toolkit] - Cluster administration and monitoring (A move to [http://www.perceus.org/portal/project/perceus Perceus] is planned) |
| 3 | One key design component of the new Limulus case is aoftware controlled power to the nodes. This will allow nodes to be powered-on only when needed. As an experiment, a simpel script was written that monitors the Grid Engine queue. If there are jobs waiting in the queue, nodes are powered-on. Once the work is done (i.e. nothing more in the queue) the nodes are powered off). |
| 4 | |
| 5 | As an example, an 8 core job was run on the Norbert cluster (in the Limulus case). The head node has 4 cores and each worker node has 2 cores for a total of 10 cores. An 8 node job was submitted via Grid Engine with only the head node powered-on. The script noticed the job waiting in the queue and turned on a single worker node to give 6 cores total, which were still not enough. Another node was powered-on and the total cores reached 8 and the job started to run. After completeion, the script noticed that there was nothing in the queue and shutdown the nodes. |
| 6 | |
| 7 | == Update January 2011 == |
| 8 | The following is a list of basic cluster RPMS that will be included in the software stack. The base distribution will be Scientific Linux. |
| 9 | |
| 10 | * [https://www.scientificlinux.org/ Scientific Linux V5.4] |
| 11 | * [http://www.perceus.org/ Perceus Cluster Toolkit] - Cluster administration |