| |
| |
| |
| |
|
|
|
|
| import glob |
| import os |
|
|
| from tensorboard.backend.event_processing import event_accumulator |
|
|
| from isaacsim.benchmark.services import BaseIsaacBenchmark |
| from isaacsim.benchmark.services.metrics.measurements import DictMeasurement, ListMeasurement, SingleMeasurement |
|
|
|
|
| def parse_tf_logs(log_dir: str): |
| """Search for the latest tfevents file in log_dir folder and returns |
| the tensorboard logs in a dictionary. |
| |
| Args: |
| log_dir: directory used to search for tfevents files |
| """ |
|
|
| |
| list_of_files = glob.glob(f"{log_dir}/events*") |
| latest_file = max(list_of_files, key=os.path.getctime) |
|
|
| log_data = {} |
| ea = event_accumulator.EventAccumulator(latest_file) |
| ea.Reload() |
| tags = ea.Tags()["scalars"] |
| for tag in tags: |
| log_data[tag] = [] |
| for event in ea.Scalars(tag): |
| log_data[tag].append(event.value) |
|
|
| return log_data |
|
|
|
|
| |
| |
| |
|
|
|
|
| def log_min_max_mean_stats(benchmark: BaseIsaacBenchmark, values: dict): |
| for k, v in values.items(): |
| measurement = SingleMeasurement(name=f"Min {k}", value=min(v), unit="ms") |
| benchmark.store_custom_measurement("runtime", measurement) |
| measurement = SingleMeasurement(name=f"Max {k}", value=max(v), unit="ms") |
| benchmark.store_custom_measurement("runtime", measurement) |
| measurement = SingleMeasurement(name=f"Mean {k}", value=sum(v) / len(v), unit="ms") |
| benchmark.store_custom_measurement("runtime", measurement) |
|
|
|
|
| def log_app_start_time(benchmark: BaseIsaacBenchmark, value: float): |
| measurement = SingleMeasurement(name="App Launch Time", value=value, unit="ms") |
| benchmark.store_custom_measurement("startup", measurement) |
|
|
|
|
| def log_python_imports_time(benchmark: BaseIsaacBenchmark, value: float): |
| measurement = SingleMeasurement(name="Python Imports Time", value=value, unit="ms") |
| benchmark.store_custom_measurement("startup", measurement) |
|
|
|
|
| def log_task_start_time(benchmark: BaseIsaacBenchmark, value: float): |
| measurement = SingleMeasurement(name="Task Creation and Start Time", value=value, unit="ms") |
| benchmark.store_custom_measurement("startup", measurement) |
|
|
|
|
| def log_scene_creation_time(benchmark: BaseIsaacBenchmark, value: float): |
| measurement = SingleMeasurement(name="Scene Creation Time", value=value, unit="ms") |
| benchmark.store_custom_measurement("startup", measurement) |
|
|
|
|
| def log_simulation_start_time(benchmark: BaseIsaacBenchmark, value: float): |
| measurement = SingleMeasurement(name="Simulation Start Time", value=value, unit="ms") |
| benchmark.store_custom_measurement("startup", measurement) |
|
|
|
|
| def log_total_start_time(benchmark: BaseIsaacBenchmark, value: float): |
| measurement = SingleMeasurement(name="Total Start Time (Launch to Train)", value=value, unit="ms") |
| benchmark.store_custom_measurement("startup", measurement) |
|
|
|
|
| def log_runtime_step_times(benchmark: BaseIsaacBenchmark, value: dict, compute_stats=True): |
| measurement = DictMeasurement(name="Step Frametimes", value=value) |
| benchmark.store_custom_measurement("runtime", measurement) |
| if compute_stats: |
| log_min_max_mean_stats(benchmark, value) |
|
|
|
|
| def log_rl_policy_rewards(benchmark: BaseIsaacBenchmark, value: list): |
| measurement = ListMeasurement(name="Rewards", value=value) |
| benchmark.store_custom_measurement("train", measurement) |
| |
| measurement = SingleMeasurement(name="Max Rewards", value=max(value), unit="float") |
| benchmark.store_custom_measurement("train", measurement) |
|
|
|
|
| def log_rl_policy_episode_lengths(benchmark: BaseIsaacBenchmark, value: list): |
| measurement = ListMeasurement(name="Episode Lengths", value=value) |
| benchmark.store_custom_measurement("train", measurement) |
| |
| measurement = SingleMeasurement(name="Max Episode Lengths", value=max(value), unit="float") |
| benchmark.store_custom_measurement("train", measurement) |
|
|