generated from amazon-archives/__template_Apache-2.0
-
Notifications
You must be signed in to change notification settings - Fork 234
Expand file tree
/
Copy pathsmart_benchmark_sweeper.py
More file actions
162 lines (129 loc) · 6.41 KB
/
smart_benchmark_sweeper.py
File metadata and controls
162 lines (129 loc) · 6.41 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
from dataclasses import dataclass
import itertools
import logging
from pathlib import Path
from typing import Any, Dict, List, Optional
from hydra.types import HydraContext
from hydra.core.config_store import ConfigStore
from hydra.core.override_parser.overrides_parser import OverridesParser
from hydra.core.override_parser.types import Override
from hydra.core.plugins import Plugins
from hydra.core.utils import JobStatus
from hydra.plugins.launcher import Launcher
from hydra.plugins.sweeper import Sweeper
from hydra.types import TaskFunction
from omegaconf import DictConfig, OmegaConf
log = logging.getLogger(__name__)
@dataclass
class SmartBenchmarkSweeperConf:
"""Configuration for SmartBenchmarkSweeper.
Attributes:
max_batch_size: Maximum number of jobs to run in a single batch (currently unused)
params: Base parameters to apply to all benchmark configurations
fail_fast: If True, stops execution immediately after first benchmark failure.
Use True for quick validation during development/debugging.
Use False (default) to run all benchmarks and collect all results.
"""
_target_: str = "hydra_plugins.smart_sweeper.smart_benchmark_sweeper.SmartBenchmarkSweeper"
max_batch_size: Optional[int] = None
params: Optional[Dict[str, str]] = None
fail_fast: bool = False
ConfigStore.instance().store(group="hydra/sweeper", name="smart_benchmark", node=SmartBenchmarkSweeperConf)
class SmartBenchmarkSweeper(Sweeper):
def __init__(
self, max_batch_size: Optional[int] = None, params: Optional[Dict[str, str]] = None, fail_fast: bool = False
):
self.max_batch_size = max_batch_size
self.params = params or {}
self.fail_fast = fail_fast
self.config: Optional[DictConfig] = None
self.launcher: Optional[Launcher] = None
self.hydra_context: Optional[HydraContext] = None
def setup(self, *, hydra_context: HydraContext, task_function: TaskFunction, config: DictConfig) -> None:
self.config = config
self.launcher = Plugins.instance().instantiate_launcher(
hydra_context=hydra_context, task_function=task_function, config=config
)
self.hydra_context = hydra_context
def _load_benchmark_params(self, benchmark_type: str) -> List[str]:
try:
config_path = Path("conf") / "hydra" / "sweeper" / f"{benchmark_type}.yaml"
if config_path.exists():
benchmark_config = OmegaConf.load(config_path)
params = benchmark_config.get("params", {})
return [f"{key}={value}" for key, value in params.items()]
return []
except Exception as e:
log.error(f"Failed to load config for {benchmark_type}: {e}")
return []
def sweep(self, arguments: List[str]) -> Any:
benchmark_types = self._extract_benchmark_types(arguments)
log.info(f"Running benchmark types: {benchmark_types}")
# Save sweep config
sweep_dir = Path(self.config.hydra.sweep.dir)
sweep_dir.mkdir(parents=True, exist_ok=True)
OmegaConf.save(self.config, sweep_dir / "multirun.yaml")
base_params_conf = []
for k, v in self.params.items():
base_params_conf.append(f"{k}={v}")
base_params_conf.extend(arguments)
all_combinations = []
# For a given benchmark type, this will load parameters defined in
# only the base and benchmark_type config files.
for benchmark_type in benchmark_types:
benchmark_params = self._load_benchmark_params(benchmark_type)
params_conf = base_params_conf + benchmark_params
parser = OverridesParser.create()
parsed = parser.parse_overrides(params_conf)
type_combinations = self._generate_combinations_for_type(benchmark_type, parsed)
all_combinations.extend(type_combinations)
log.info(f"Generated {len(all_combinations)} total combinations")
returns = []
initial_job_idx = 0
if all_combinations:
self.validate_batch_is_legal(all_combinations)
returns = self._execute_batches(all_combinations, initial_job_idx)
return returns
def _execute_batches(self, all_combinations: List[List[str]], initial_job_idx: int) -> List[Any]:
"""
Execute benchmark combinations in batches.
When fail_fast=False: Launches all combinations in one batch
When fail_fast=True: Launches one combination at a time, stopping on first failure
Args:
all_combinations: List of parameter combinations to execute
initial_job_idx: Starting job index for the launcher
Returns:
List of results from launcher.launch() calls
"""
returns = []
batch_size = 1 if self.fail_fast else len(all_combinations)
for i in range(0, len(all_combinations), batch_size):
batch = all_combinations[i : i + batch_size]
results = self.launcher.launch(batch, initial_job_idx=i)
# Check results immediately if fail_fast enabled
if self.fail_fast:
for r in results:
if r.status == JobStatus.FAILED:
raise r._return_value
returns.append(results)
return returns
def _extract_benchmark_types(self, arguments: List[str]) -> List[str]:
for arg in arguments:
if arg.startswith("benchmark_type="):
benchmark_type_str = arg.split("=", 1)[1]
return [bt.strip() for bt in benchmark_type_str.split(",")]
return ["fio"]
def _generate_combinations_for_type(self, benchmark_type: str, parsed_overrides: List[Override]) -> List[List[str]]:
param_lists = [[f"benchmark_type={benchmark_type}"]]
for param_override in parsed_overrides:
param_key = param_override.get_key_element()
if param_key == "benchmark_type":
continue
if param_override.is_sweep_override():
sweep = [f"{param_key}={val}" for val in param_override.sweep_string_iterator()]
param_lists.append(sweep)
else:
value = param_override.get_value_element_as_str()
param_lists.append([f"{param_key}={value}"])
combinations = list(itertools.product(*param_lists))
return [list(combination) for combination in combinations]