Skip to content

Commit

Permalink
Remove unused cluster util audit keys
Browse files Browse the repository at this point in the history
  • Loading branch information
Kevin Yan committed Nov 17, 2023
1 parent 531432a commit b3d9d18
Show file tree
Hide file tree
Showing 2 changed files with 0 additions and 58 deletions.
1 change: 0 additions & 1 deletion deltacat/compute/compactor/compaction_session.py
Original file line number Diff line number Diff line change
Expand Up @@ -293,7 +293,6 @@ def _execute_compaction_round(
f"{node_resource_keys}"
)

compaction_audit.set_cluster_cpu_max(cluster_cpus)
# create a remote options provider to round-robin tasks across all nodes or allocated bundles
logger.info(f"Setting round robin scheduling with node id:{node_resource_keys}")
round_robin_opt_provider = functools.partial(
Expand Down
57 changes: 0 additions & 57 deletions deltacat/compute/compactor/model/compaction_session_audit_info.py
Original file line number Diff line number Diff line change
Expand Up @@ -98,14 +98,6 @@ def hash_bucket_count(self) -> int:
"""
return self.get("hashBucketCount")

@property
def cluster_cpu_max(self) -> float:
"""
Total cluster cpu allocated for the compaction job. If it is autoscaling cluster,
max cpu at any time will be reported.
"""
return self.get("clusterCpuMax")

@property
def compaction_time_in_seconds(self) -> float:
"""
Expand Down Expand Up @@ -423,35 +415,6 @@ def hash_bucket_processed_size_bytes(self) -> int:
"""
return self.get("hashBucketProcessedSizeBytes")

@property
def total_cpu_seconds(self) -> float:
"""
Total number of vCPUs provisioned in the cluster weighted over time.
"""
return self.get("totalCPUSeconds")

@property
def used_cpu_seconds(self) -> float:
"""
Total used vCPU in the cluster weighted over time.
"""
return self.get("usedCPUSeconds")

@property
def used_memory_gb_seconds(self) -> float:
"""
The used memory in the cluster weighted over time. This
determines opportunities for better memory estimation.
"""
return self.get("usedMemoryGBSeconds")

@property
def total_memory_gb_seconds(self) -> float:
"""
Total memory in the cluster weighted over time in GB.
"""
return self.get("totalMemoryGBSeconds")

@property
def pyarrow_version(self) -> str:
"""
Expand Down Expand Up @@ -510,10 +473,6 @@ def set_hash_bucket_count(
self["hashBucketCount"] = hash_bucket_count
return self

def set_cluster_cpu_max(self, cluster_cpu_max: float) -> CompactionSessionAuditInfo:
self["clusterCpuMax"] = cluster_cpu_max
return self

def set_compaction_time_in_seconds(
self, compaction_time_in_seconds: float
) -> CompactionSessionAuditInfo:
Expand Down Expand Up @@ -778,22 +737,6 @@ def set_hash_bucket_processed_size_bytes(
self["hashBucketProcessedSizeBytes"] = size
return self

def set_total_cpu_seconds(self, value: float) -> CompactionSessionAuditInfo:
self["totalCPUSeconds"] = value
return self

def set_used_cpu_seconds(self, value: float) -> CompactionSessionAuditInfo:
self["usedCPUSeconds"] = value
return self

def set_used_memory_gb_seconds(self, value: float) -> CompactionSessionAuditInfo:
self["usedMemoryGBSeconds"] = value
return self

def set_total_memory_gb_seconds(self, value: float) -> CompactionSessionAuditInfo:
self["totalMemoryGBSeconds"] = value
return self

def set_pyarrow_version(self, value: str) -> CompactionSessionAuditInfo:
self["pyarrowVersion"] = value
return self
Expand Down

0 comments on commit b3d9d18

Please sign in to comment.