diff --git a/src/slurm_plugin/instance_manager.py b/src/slurm_plugin/instance_manager.py index e8704c8c4..42b33446e 100644 --- a/src/slurm_plugin/instance_manager.py +++ b/src/slurm_plugin/instance_manager.py @@ -79,7 +79,6 @@ def get_manager( run_instances_overrides: dict = None, create_fleet_overrides: dict = None, job_level_scaling: bool = False, - temp_jls_for_node_sharing: bool = False, ): if job_level_scaling: return JobLevelScalingInstanceManager( @@ -95,7 +94,6 @@ def get_manager( fleet_config=fleet_config, run_instances_overrides=run_instances_overrides, create_fleet_overrides=create_fleet_overrides, - temp_jls_for_node_sharing=temp_jls_for_node_sharing, ) else: return NodeListScalingInstanceManager( @@ -493,7 +491,6 @@ def __init__( fleet_config: Dict[str, any] = None, run_instances_overrides: dict = None, create_fleet_overrides: dict = None, - temp_jls_for_node_sharing: bool = False, ): super().__init__( region=region, @@ -510,7 +507,6 @@ def __init__( create_fleet_overrides=create_fleet_overrides, ) self.unused_launched_instances = {} - self.temp_jls_for_node_sharing = temp_jls_for_node_sharing def _clear_unused_launched_instances(self): """Clear and reset unused launched instances list.""" diff --git a/src/slurm_plugin/resume.py b/src/slurm_plugin/resume.py index 6c90225a6..8f77a76be 100644 --- a/src/slurm_plugin/resume.py +++ b/src/slurm_plugin/resume.py @@ -47,7 +47,6 @@ class SlurmResumeConfig: "fleet_config_file": "/etc/parallelcluster/slurm_plugin/fleet-config.json", "all_or_nothing_batch": True, "job_level_scaling": True, - "temp_jls_for_node_sharing": False, } def __init__(self, config_file_path): @@ -96,9 +95,6 @@ def _get_config(self, config_file_path): self.job_level_scaling = config.getboolean( "slurm_resume", "job_level_scaling", fallback=self.DEFAULTS.get("job_level_scaling") ) - self.temp_jls_for_node_sharing = config.getboolean( - "slurm_resume", "temp_jls_for_node_sharing", fallback=self.DEFAULTS.get("temp_jls_for_node_sharing") - ) fleet_config_file = config.get( "slurm_resume", "fleet_config_file", fallback=self.DEFAULTS.get("fleet_config_file") ) @@ -209,7 +205,6 @@ def _resume(arg_nodes, resume_config, slurm_resume): run_instances_overrides=resume_config.run_instances_overrides, create_fleet_overrides=resume_config.create_fleet_overrides, job_level_scaling=resume_config.job_level_scaling, - temp_jls_for_node_sharing=resume_config.temp_jls_for_node_sharing, ) instance_manager.add_instances( slurm_resume=slurm_resume, diff --git a/tests/slurm_plugin/test_resume.py b/tests/slurm_plugin/test_resume.py index f03faad09..4004fba71 100644 --- a/tests/slurm_plugin/test_resume.py +++ b/tests/slurm_plugin/test_resume.py @@ -798,7 +798,6 @@ def test_resume_launch( job_level_scaling=job_level_scaling, assign_node_max_batch_size=500, terminate_max_batch_size=1000, - temp_jls_for_node_sharing=False, ) mocker.patch("slurm_plugin.resume.is_clustermgtd_heartbeat_valid", autospec=True, return_value=is_heartbeat_valid) mock_handle_failed_nodes = mocker.patch("slurm_plugin.resume._handle_failed_nodes", autospec=True)