Scratch
Default limit: 20GB per one core
The best practice to avoid problems with overload and slow response of NFS servers is: read all input files from the NFS server only once when your job starts, use /scratch at a local disk of an assigned worker node as your working directory, write all output files back to the NFS server when your job completes.
A local working directory /scratch is available at a local disk of each worker node, and is intended for temporary files. Upon a job start, the batch system creates a unique subdirectory /scratch/condor/dir_ID/, which is accessible via the $_CONDOR_SCRATCH_DIR variable. When the job finishes, the subdirectory is automatically deleted. This way, the local working directory /scratch is continuously cleaned.
You can appropriately adjust a default requirement for /scratch in a job_submission_file, e.g.
request_disk = 5G
request_disk = 30G
In case a job exceeds a required value for scratch, the job will go to HOLD state.