Spaces:
Running
Running
| import requests | |
| BASE_URL = "http://localhost:7860" # Default HF space port | |
| def evaluate_baseline(task_id): | |
| requests.post(f"{BASE_URL}/reset?task_id={task_id}") | |
| done = False | |
| while not done: | |
| state = requests.get(f"{BASE_URL}/state").json()["observation"] | |
| # Simple policy: If queue is larger than active GPUs, provision more. | |
| gpus_needed = state["queue_size"] - state["active_gpus"] | |
| action = {"gpus_to_provision": max(-1, min(2, gpus_needed))} # Throttle scaling | |
| step_res = requests.post(f"{BASE_URL}/step", json=action).json() | |
| done = step_res["done"] | |
| score = requests.get(f"{BASE_URL}/grader").json()["score"] | |
| print(f"Task: {task_id} | Final Score: {score:.3f}") | |
| if __name__ == "__main__": | |
| for task in ["easy", "medium", "hard"]: | |
| evaluate_baseline(task) | |