Skip to content

Commit

Permalink
Remove temporary resume setting
Browse files Browse the repository at this point in the history
Remove temporary resume setting used during development of the node-sharing job-level scaling feature

Signed-off-by: Luca Carrogu <[email protected]>
  • Loading branch information
lukeseawalker committed Sep 26, 2023
1 parent 018d55a commit 5bb429d
Show file tree
Hide file tree
Showing 3 changed files with 0 additions and 10 deletions.
4 changes: 0 additions & 4 deletions src/slurm_plugin/instance_manager.py
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,6 @@ def get_manager(
run_instances_overrides: dict = None,
create_fleet_overrides: dict = None,
job_level_scaling: bool = False,
temp_jls_for_node_sharing: bool = False,
):
if job_level_scaling:
return JobLevelScalingInstanceManager(
Expand All @@ -95,7 +94,6 @@ def get_manager(
fleet_config=fleet_config,
run_instances_overrides=run_instances_overrides,
create_fleet_overrides=create_fleet_overrides,
temp_jls_for_node_sharing=temp_jls_for_node_sharing,
)
else:
return NodeListScalingInstanceManager(
Expand Down Expand Up @@ -493,7 +491,6 @@ def __init__(
fleet_config: Dict[str, any] = None,
run_instances_overrides: dict = None,
create_fleet_overrides: dict = None,
temp_jls_for_node_sharing: bool = False,
):
super().__init__(
region=region,
Expand All @@ -510,7 +507,6 @@ def __init__(
create_fleet_overrides=create_fleet_overrides,
)
self.unused_launched_instances = {}
self.temp_jls_for_node_sharing = temp_jls_for_node_sharing

def _clear_unused_launched_instances(self):
"""Clear and reset unused launched instances list."""
Expand Down
5 changes: 0 additions & 5 deletions src/slurm_plugin/resume.py
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,6 @@ class SlurmResumeConfig:
"fleet_config_file": "/etc/parallelcluster/slurm_plugin/fleet-config.json",
"all_or_nothing_batch": True,
"job_level_scaling": True,
"temp_jls_for_node_sharing": False,
}

def __init__(self, config_file_path):
Expand Down Expand Up @@ -96,9 +95,6 @@ def _get_config(self, config_file_path):
self.job_level_scaling = config.getboolean(
"slurm_resume", "job_level_scaling", fallback=self.DEFAULTS.get("job_level_scaling")
)
self.temp_jls_for_node_sharing = config.getboolean(
"slurm_resume", "temp_jls_for_node_sharing", fallback=self.DEFAULTS.get("temp_jls_for_node_sharing")
)
fleet_config_file = config.get(
"slurm_resume", "fleet_config_file", fallback=self.DEFAULTS.get("fleet_config_file")
)
Expand Down Expand Up @@ -209,7 +205,6 @@ def _resume(arg_nodes, resume_config, slurm_resume):
run_instances_overrides=resume_config.run_instances_overrides,
create_fleet_overrides=resume_config.create_fleet_overrides,
job_level_scaling=resume_config.job_level_scaling,
temp_jls_for_node_sharing=resume_config.temp_jls_for_node_sharing,
)
instance_manager.add_instances(
slurm_resume=slurm_resume,
Expand Down
1 change: 0 additions & 1 deletion tests/slurm_plugin/test_resume.py
Original file line number Diff line number Diff line change
Expand Up @@ -798,7 +798,6 @@ def test_resume_launch(
job_level_scaling=job_level_scaling,
assign_node_max_batch_size=500,
terminate_max_batch_size=1000,
temp_jls_for_node_sharing=False,
)
mocker.patch("slurm_plugin.resume.is_clustermgtd_heartbeat_valid", autospec=True, return_value=is_heartbeat_valid)
mock_handle_failed_nodes = mocker.patch("slurm_plugin.resume._handle_failed_nodes", autospec=True)
Expand Down

0 comments on commit 5bb429d

Please sign in to comment.