samrat-rm commited on
Commit
f0681d9
Β·
1 Parent(s): ff8ce5f

docs: updating readme with state changes and test

Browse files
Files changed (1) hide show
  1. README.md +14 -1
README.md CHANGED
@@ -308,6 +308,18 @@ Stdout will stream `[START]` / `[STEP]` / `[END]` lines per episode. Internal lo
308
 
309
  ---
310
 
 
 
 
 
 
 
 
 
 
 
 
 
311
  ### Local Agent β€” No API Key Required
312
 
313
  To smoke test the full pipeline without calling an external LLM, you can run inference with a local model via [Ollama](https://ollama.com/).
@@ -346,8 +358,9 @@ USE_LOCAL=true uv run python inference.py
346
  ```
347
  WhyDidItFail/
348
  β”œβ”€β”€ inference.py # Baseline inference script
 
349
  β”œβ”€β”€ client.py # WhyDidItFailEnv client (WebSocket)
350
- β”œβ”€β”€ models.py # Action and Observation Pydantic models
351
  β”œβ”€β”€ openenv.yaml # OpenEnv manifest
352
  β”œβ”€β”€ Dockerfile # Container image
353
  └── server/
 
308
 
309
  ---
310
 
311
+ ### 6. Verify environment state (optional)
312
+
313
+ To confirm the state endpoint is working correctly β€” episode tracking, inspection order, required sources:
314
+
315
+ ```bash
316
+ uv run python test_state.py
317
+ ```
318
+
319
+ Runs a single hard episode and calls `state()` after every action. Prints `OK` or `FAIL` per checkpoint.
320
+
321
+ ---
322
+
323
  ### Local Agent β€” No API Key Required
324
 
325
  To smoke test the full pipeline without calling an external LLM, you can run inference with a local model via [Ollama](https://ollama.com/).
 
358
  ```
359
  WhyDidItFail/
360
  β”œβ”€β”€ inference.py # Baseline inference script
361
+ β”œβ”€β”€ test_state.py # State endpoint verification script
362
  β”œβ”€β”€ client.py # WhyDidItFailEnv client (WebSocket)
363
+ β”œβ”€β”€ models.py # Action, Observation, and State Pydantic models
364
  β”œβ”€β”€ openenv.yaml # OpenEnv manifest
365
  β”œβ”€β”€ Dockerfile # Container image
366
  └── server/