forked from nf-core/configs
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathsage.config
125 lines (113 loc) · 4.1 KB
/
sage.config
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
// Config profile metadata
params {
config_profile_description = 'The Sage Bionetworks Nextflow Config Profile'
config_profile_contact = 'Bruno Grande (@BrunoGrandePhD)'
config_profile_url = 'https://github.com/Sage-Bionetworks-Workflows'
}
// Leverage us-east-1 mirror of select human and mouse genomes
params {
igenomes_base = 's3://sage-igenomes/igenomes'
cpus = 4
max_cpus = 32
max_memory = 128.GB
max_time = 240.h
single_cpu_mem = 6.GB
}
// Enable retries globally for certain exit codes
process {
maxErrors = '-1'
maxRetries = 5
errorStrategy = { task.attempt <= 5 ? 'retry' : 'finish' }
}
// Increase time limit to allow file transfers to finish
// The default is 12 hours, which results in timeouts
threadPool.FileTransfer.maxAwait = '24 hour'
// Configure Nextflow to be more reliable on AWS
aws {
region = "us-east-1"
client {
uploadMaxThreads = 4
}
batch {
retryMode = 'built-in'
maxParallelTransfers = 1
maxTransferAttempts = 10
delayBetweenAttempts = '60 sec'
}
}
// Adjust default resource allocations (see `../docs/sage.md`)
process {
cpus = { check_max( 1 * factor(task, 2), 'cpus' ) }
memory = { check_max( 6.GB * factor(task, 1), 'memory' ) }
time = { check_max( 24.h * factor(task, 1), 'time' ) }
// Process-specific resource requirements
withLabel: 'process_single' {
cpus = { check_max( 1 * factor(task, 2), 'cpus' ) }
memory = { check_max( 6.GB * factor(task, 1), 'memory' ) }
time = { check_max( 24.h * factor(task, 1), 'time' ) }
}
withLabel: 'process_low' {
cpus = { check_max( 2 * factor(task, 2), 'cpus' ) }
memory = { check_max( 12.GB * factor(task, 1), 'memory' ) }
time = { check_max( 24.h * factor(task, 1), 'time' ) }
}
withLabel: 'process_medium' {
cpus = { check_max( 8 * factor(task, 2), 'cpus' ) }
memory = { check_max( 32.GB * factor(task, 1), 'memory' ) }
time = { check_max( 48.h * factor(task, 1), 'time' ) }
}
withLabel: 'process_high' {
cpus = { check_max( 16 * factor(task, 2), 'cpus' ) }
memory = { check_max( 64.GB * factor(task, 1), 'memory' ) }
time = { check_max( 96.h * factor(task, 1), 'time' ) }
}
withLabel: 'process_long' {
time = { check_max( 96.h * factor(task, 1), 'time' ) }
}
withLabel: 'process_high_memory|memory_max' {
memory = { check_max( 128.GB * factor(task, 1), 'memory' ) }
}
withLabel: 'cpus_max' {
cpus = { check_max( 32 * factor(task, 2), 'cpus' ) }
}
}
// Function to finely control the increase of the resource allocation
def factor(task, slow_factor = 1) {
if ( task.exitStatus in [143,137,104,134,139,247] ) {
return Math.ceil( task.attempt / slow_factor) as int
} else {
return 1 as int
}
}
// Function to ensure that resource requirements don't go
// beyond a maximum limit (copied here for Sarek v2)
def check_max(obj, type) {
if (type == 'memory') {
try {
if (obj.compareTo(params.max_memory as nextflow.util.MemoryUnit) == 1)
return params.max_memory as nextflow.util.MemoryUnit
else
return obj
} catch (all) {
println " ### ERROR ### Max memory '${params.max_memory}' is not valid! Using default value: $obj"
return obj
}
} else if (type == 'time') {
try {
if (obj.compareTo(params.max_time as nextflow.util.Duration) == 1)
return params.max_time as nextflow.util.Duration
else
return obj
} catch (all) {
println " ### ERROR ### Max time '${params.max_time}' is not valid! Using default value: $obj"
return obj
}
} else if (type == 'cpus') {
try {
return Math.min( obj, params.max_cpus as int )
} catch (all) {
println " ### ERROR ### Max cpus '${params.max_cpus}' is not valid! Using default value: $obj"
return obj
}
}
}