Test Cases
From FarmShare
(Difference between revisions)
(→TC5: check maximum job numbers) |
|||
Line 70: | Line 70: | ||
That will submit 200 jobs, and you should see that only 100 of them will be accepted and an error like | That will submit 200 jobs, and you should see that only 100 of them will be accepted and an error like | ||
Unable to run job: job rejected: Only 100 jobs are allowed per user (current job count: 100). | Unable to run job: job rejected: Only 100 jobs are allowed per user (current job count: 100). | ||
+ | |||
+ | ==TC6: check disk throughput performance== | ||
+ | Use the bonnie++ executable and run it on local disk with a submit script like this: | ||
+ | <pre> | ||
+ | #!/bin/bash | ||
+ | |||
+ | #$ -m bes | ||
+ | #$ -M chekh@stanford.edu | ||
+ | #$ -cwd | ||
+ | |||
+ | BONNIE=/mnt/glusterfs/chekh/bonnie/bonnie++ | ||
+ | |||
+ | echo $TMPDIR | ||
+ | |||
+ | $BONNIE -d $TMPDIR | ||
+ | </pre> | ||
+ | |||
+ | Check that the performance numbers in the output roughly match these: |
Revision as of 14:21, 14 December 2011
This page should have some "test cases" that users or sysadmins can run to verify the functionality of the barleys.
Contents |
TC1: submit a job from a corn via /mnt/glusterfs
- cd /mnt/glusterfs/your_sunetid
- echo "hostname" | qsub -cwd
- qstat # check job status
- Check that the stderr output file is empty and the stdout output file contains the hostname of the machine that the job ran on
This test verifies that the shared filesystem is available and the job submission process works as expected.
TC2: submit a job from a corn with AUKS support
- kinit / klist -f # check that your ticket is forwardable
- aklog / tokens # check that you have an AFS token
- echo "hostname" | qsub
- qstat # check job status
- Check that the stderr output file is empty and the stdout output file contains the hostname of the machine that the job ran on
This test verifies that AUKS handles the Kerberos/AFS tickets/tokens correctly.
TC3: check memory tracking
R script:
$ cat R8GB.R x <- array(1:1073741824, dim=c(1024,1024,1024)) x <- gaussian(); Sys.sleep(60)
submit script:
$ cat r_test.script #!/bin/bash # use the current directory #$ -cwd # mail this address #$ -M chekh@stanford.edu # send mail on begin, end, suspend #$ -m bes # get rid of spurious messages about tty/terminal types #$ -S /bin/sh R --vanilla --no-save < R8GB.R
- submit this job with 'qsub r_test.script' (with AUKS or not)
- check that you get an e-mail
- check that the e-mail correctly reports ~8GB maxvmem (our current version does not, it's a bug, we need to upgrade)
TC4: check time tracking
submit a job like
echo "sleep 3600" | qsub -cwd -m bes -M chekh@stanford.edu
Check that the job completion mail says 1hr of walltime elapsed
submit a job like
echo "sleep 3600" | qsub -cwd -m bes -M chekh@stanford.edu -l h_rt=72:00:00
Check that the job went into long.q
TC5: check maximum job numbers
We currently have:
max_u_jobs 100 max_jobs 3000
for i in `seq -w 01 200`; do echo "sleep 600" | qsub -cwd ; done
That will submit 200 jobs, and you should see that only 100 of them will be accepted and an error like
Unable to run job: job rejected: Only 100 jobs are allowed per user (current job count: 100).
TC6: check disk throughput performance
Use the bonnie++ executable and run it on local disk with a submit script like this:
#!/bin/bash #$ -m bes #$ -M chekh@stanford.edu #$ -cwd BONNIE=/mnt/glusterfs/chekh/bonnie/bonnie++ echo $TMPDIR $BONNIE -d $TMPDIR
Check that the performance numbers in the output roughly match these: