'extra']:
setattr(self, name, getattr(self.real, name))
self.cores = kwargs.pop('cores')
- self.scratch = self.disk
+ # libcloud disk sizes are in GB, Arvados/SLURM are in MB
+ # multiply by 1000 instead of 1024 to err on low side
+ self.scratch = self.disk * 1000
self.ram = int(self.ram * node_mem_scaling)
for name, override in kwargs.iteritems():
if not hasattr(self, name):
self.logger = logging.getLogger('arvnodeman.jobqueue')
self.logged_jobs = set()
+ self.logger.info("Using cloud node sizes:")
+ for s in self.cloud_sizes:
+ self.logger.info(str(s.__dict__))
+
@staticmethod
def coerce_int(x, fallback):
try:
{'min_ram_mb_per_node': 256},
{'min_nodes': 6},
{'min_nodes': 12},
- {'min_scratch_mb_per_node': 200})
+ {'min_scratch_mb_per_node': 300000})
self.assertEqual(6, len(servlist))
def test_ignore_too_expensive_jobs(self):