nanigock commited on
Commit
9f2fc5c
·
verified ·
1 Parent(s): 80892e8

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +6 -0
  2. oracle_context_cache_v3/0xricksanchez__like-dbg.json +0 -0
  3. oracle_context_cache_v3/15r10nk__inline-snapshot.json +0 -0
  4. oracle_context_cache_v3/AlignmentResearch__tuned-lens.json +0 -0
  5. oracle_context_cache_v3/AndrewAnnex__SpiceyPy.json +1 -0
  6. oracle_context_cache_v3/AnonymouX47__term-image.json +0 -0
  7. oracle_context_cache_v3/Azure-Samples__rag-postgres-openai-python.json +1 -0
  8. oracle_context_cache_v3/BayesWitnesses__m2cgen.json +1 -0
  9. oracle_context_cache_v3/BoboTiG__python-mss.json +0 -0
  10. oracle_context_cache_v3/BrainBlend-AI__atomic-agents.json +0 -0
  11. oracle_context_cache_v3/CalebBell__fluids.json +0 -0
  12. oracle_context_cache_v3/Chen-zexi__vllm-cli.json +0 -0
  13. oracle_context_cache_v3/Cloxl__xhshow.json +0 -0
  14. oracle_context_cache_v3/Cranot__roam-code.json +0 -0
  15. oracle_context_cache_v3/CursorTouch__Windows-MCP.json +0 -0
  16. oracle_context_cache_v3/DHI__terracotta.json +0 -0
  17. oracle_context_cache_v3/DLR-RM__stable-baselines3.json +0 -0
  18. oracle_context_cache_v3/DebarghaG__proofofthought.json +0 -0
  19. oracle_context_cache_v3/DeepLcom__deepl-python.json +1 -0
  20. oracle_context_cache_v3/Delgan__loguru.json +0 -0
  21. oracle_context_cache_v3/DenisCarriere__geocoder.json +0 -0
  22. oracle_context_cache_v3/DisnakeDev__disnake.json +0 -0
  23. oracle_context_cache_v3/DonDebonair__slack-machine.json +0 -0
  24. oracle_context_cache_v3/Donkie__Spoolman.json +1 -0
  25. oracle_context_cache_v3/EbodShojaei__bake.json +0 -0
  26. oracle_context_cache_v3/Filimoa__open-parse.json +0 -0
  27. oracle_context_cache_v3/FinanceData__FinanceDataReader.json +1 -0
  28. oracle_context_cache_v3/Forethought-Technologies__AutoChain.json +0 -0
  29. oracle_context_cache_v3/GitGuardian__ggshield.json +0 -0
  30. oracle_context_cache_v3/IDSIA__sacred.json +0 -0
  31. oracle_context_cache_v3/JWock82__Pynite.json +0 -0
  32. oracle_context_cache_v3/JoshuaC215__agent-service-toolkit.json +0 -0
  33. oracle_context_cache_v3/JuanBindez__pytubefix.json +0 -0
  34. oracle_context_cache_v3/Kludex__mangum.json +0 -0
  35. oracle_context_cache_v3/Lancetnik__Propan.json +0 -0
  36. oracle_context_cache_v3/LonamiWebs__Telethon.json +0 -0
  37. oracle_context_cache_v3/LuteOrg__lute-v3.json +0 -0
  38. oracle_context_cache_v3/Lux-Luna__LunaVox.json +0 -0
  39. oracle_context_cache_v3/MarshalX__atproto.json +0 -0
  40. oracle_context_cache_v3/MartenBE__mkslides.json +1 -0
  41. oracle_context_cache_v3/MasoniteFramework__masonite.json +0 -0
  42. oracle_context_cache_v3/MaxHalford__prince.json +1 -0
  43. oracle_context_cache_v3/Mayitzin__ahrs.json +1 -0
  44. oracle_context_cache_v3/MerrimanInd__drawpyo.json +0 -0
  45. oracle_context_cache_v3/MiniMax-AI__Mini-Agent.json +0 -0
  46. oracle_context_cache_v3/MinishLab__model2vec.json +0 -0
  47. oracle_context_cache_v3/MinishLab__semhash.json +0 -0
  48. oracle_context_cache_v3/MolecularAI__aizynthfinder.json +0 -0
  49. oracle_context_cache_v3/MongoEngine__mongoengine.json +0 -0
  50. oracle_context_cache_v3/MrPowers__chispa.json +0 -0
.gitattributes CHANGED
@@ -80,3 +80,9 @@ oracle_context_cache/dbos-inc__dbos-transact-py.json filter=lfs diff=lfs merge=l
80
  oracle_context_cache/hdwallet-io__python-hdwallet.json filter=lfs diff=lfs merge=lfs -text
81
  oracle_context_cache/oraios__serena.json filter=lfs diff=lfs merge=lfs -text
82
  oracle_context_cache/sammchardy__python-binance.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
80
  oracle_context_cache/hdwallet-io__python-hdwallet.json filter=lfs diff=lfs merge=lfs -text
81
  oracle_context_cache/oraios__serena.json filter=lfs diff=lfs merge=lfs -text
82
  oracle_context_cache/sammchardy__python-binance.json filter=lfs diff=lfs merge=lfs -text
83
+ oracle_context_cache_v3/aiogram__aiogram.json filter=lfs diff=lfs merge=lfs -text
84
+ oracle_context_cache_v3/benavlabs__fastcrud.json filter=lfs diff=lfs merge=lfs -text
85
+ oracle_context_cache_v3/dbos-inc__dbos-transact-py.json filter=lfs diff=lfs merge=lfs -text
86
+ oracle_context_cache_v3/hdwallet-io__python-hdwallet.json filter=lfs diff=lfs merge=lfs -text
87
+ oracle_context_cache_v3/oraios__serena.json filter=lfs diff=lfs merge=lfs -text
88
+ oracle_context_cache_v3/sammchardy__python-binance.json filter=lfs diff=lfs merge=lfs -text
oracle_context_cache_v3/0xricksanchez__like-dbg.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/15r10nk__inline-snapshot.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/AlignmentResearch__tuned-lens.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/AndrewAnnex__SpiceyPy.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"repo": "AndrewAnnex/SpiceyPy", "n_pairs": 42, "version": "v3_compressed", "max_tokens": 6000, "contexts": {"src/spiceypy/benchmarks/test_cyice.py::618": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["cyice", "pytest"], "enclosing_function": "test_getelm", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/tests/gettestkernels.py::202": {"resolved_imports": [], "used_names": ["error", "hashlib", "request", "time", "urllib"], "enclosing_function": "attempt_download", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 0, "n_chars_extracted": 0}, "src/spiceypy/tests/test_context_manager.py::113": {"resolved_imports": ["src/spiceypy/__init__.py"], "used_names": ["CoreKernels", "pytest"], "enclosing_function": "test_unload_if_error", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "src/spiceypy/tests/test_context_manager.py::71": {"resolved_imports": ["src/spiceypy/__init__.py"], "used_names": ["CoreKernels", "pytest"], "enclosing_function": "test_side_effect", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "src/spiceypy/tests/test_spiceerrors.py::150": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py", "src/spiceypy/found_catcher.py"], "used_names": [], "enclosing_function": "test_error_to_str", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/tests/test_spiceerrors.py::41": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py", "src/spiceypy/found_catcher.py"], "used_names": [], "enclosing_function": "test_geterror", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/tests/test_wrapper.py::332": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/utils/callbacks.py", "src/spiceypy/utils/support_types.py"], "used_names": ["CoreKernels", "testing"], "enclosing_function": "test_bodvcd", "extracted_code": "", "n_imports_parsed": 11, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/tests/test_context_manager.py::32": {"resolved_imports": ["src/spiceypy/__init__.py"], "used_names": ["CoreKernels", "pytest"], "enclosing_function": "test_input_types", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "src/spiceypy/tests/test_support_types.py::183": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/utils/support_types.py"], "used_names": ["array", "pytest"], "enclosing_function": "test_to_double_matrix", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 2, "n_chars_extracted": 0}, "src/spiceypy/benchmarks/test_cyice.py::398": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["cyice", "pytest", "time"], "enclosing_function": "test_et2lst", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/benchmarks/test_cyice.py::549": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["cyice", "pytest"], "enclosing_function": "test_fovray_v", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/tests/test_support_types.py::84": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/utils/support_types.py"], "used_names": [], "enclosing_function": "test_spicecell_len0", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 2, "n_chars_extracted": 0}, "src/spiceypy/tests/test_spiceerrors.py::120": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py", "src/spiceypy/found_catcher.py"], "used_names": ["CoreKernels", "ExtraKernels", "cyice", "pytest", "spiceypy"], "enclosing_function": "test_no_loaded_files_exception", "extracted_code": "# Source: src/spiceypy/__init__.py\n__version__ = \"8.0.2\"\n\nfrom .spiceypy import *\nfrom .utils import support_types\nfrom .utils import exceptions\n\n# Default setting for error reporting so that programs don't just exit out!\nerract(\"set\", 10, \"return\")\nerrdev(\"set\", 10, \"null\")\n\n\n# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 344, "extracted_code_full": "# Source: src/spiceypy/__init__.py\n__version__ = \"8.0.2\"\n\nfrom .spiceypy import *\nfrom .utils import support_types\nfrom .utils import exceptions\n\n# Default setting for error reporting so that programs don't just exit out!\nerract(\"set\", 10, \"return\")\nerrdev(\"set\", 10, \"null\")\n\n\n# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 344, "compression_ratio": 1.0}, "src/spiceypy/benchmarks/test_cyice.py::400": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["cyice", "pytest", "time"], "enclosing_function": "test_et2lst", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/benchmarks/test_cyice.py::981": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["cyice", "pytest"], "enclosing_function": "test_occult", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/benchmarks/test_cyice.py::901": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["ExtraKernels", "cyice", "pytest"], "enclosing_function": "test_limbpt", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/tests/test_support_types.py::102": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/utils/support_types.py"], "used_names": [], "enclosing_function": "test_spicecell_equality", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 2, "n_chars_extracted": 0}, "src/spiceypy/tests/test_spiceerrors.py::56": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py", "src/spiceypy/found_catcher.py"], "used_names": ["cwd", "os", "pytest"], "enclosing_function": "test_get_spiceypy_exceptions", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/tests/test_support_types.py::65": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/utils/support_types.py"], "used_names": ["pytest"], "enclosing_function": "test_SpiceCell", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 2, "n_chars_extracted": 0}, "src/spiceypy/tests/test_context_manager.py::109": {"resolved_imports": ["src/spiceypy/__init__.py"], "used_names": ["CoreKernels", "pytest"], "enclosing_function": "test_unload_if_error", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "src/spiceypy/tests/test_support_types.py::103": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/utils/support_types.py"], "used_names": [], "enclosing_function": "test_spicecell_equality", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 2, "n_chars_extracted": 0}, "src/spiceypy/tests/test_context_manager.py::125": {"resolved_imports": ["src/spiceypy/__init__.py"], "used_names": ["CoreKernels"], "enclosing_function": "test_actually_works", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "src/spiceypy/benchmarks/test_cyice.py::512": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["ExtraKernels", "cyice", "pytest"], "enclosing_function": "test_evsgp4_v", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/benchmarks/test_cyice.py::419": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["cyice", "pytest", "time"], "enclosing_function": "test_et2lst_v", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/benchmarks/test_cyice.py::1982": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["cyice", "pytest"], "enclosing_function": "test_spkpvn_v", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/tests/test_wrapper.py::405": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/utils/callbacks.py", "src/spiceypy/utils/support_types.py"], "used_names": [], "enclosing_function": "test_bsrchd", "extracted_code": "", "n_imports_parsed": 11, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/benchmarks/test_cyice.py::480": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["ExtraKernels", "cyice", "pytest"], "enclosing_function": "test_evsgp4", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/tests/test_spiceerrors.py::34": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py", "src/spiceypy/found_catcher.py"], "used_names": [], "enclosing_function": "test_tkversion", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/tests/test_spiceerrors.py::101": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py", "src/spiceypy/found_catcher.py"], "used_names": ["CoreKernels", "ExtraKernels", "cyice", "pytest", "spiceypy"], "enclosing_function": "test_no_loaded_files_exception", "extracted_code": "# Source: src/spiceypy/__init__.py\n__version__ = \"8.0.2\"\n\nfrom .spiceypy import *\nfrom .utils import support_types\nfrom .utils import exceptions\n\n# Default setting for error reporting so that programs don't just exit out!\nerract(\"set\", 10, \"return\")\nerrdev(\"set\", 10, \"null\")\n\n\n# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 344, "extracted_code_full": "# Source: src/spiceypy/__init__.py\n__version__ = \"8.0.2\"\n\nfrom .spiceypy import *\nfrom .utils import support_types\nfrom .utils import exceptions\n\n# Default setting for error reporting so that programs don't just exit out!\nerract(\"set\", 10, \"return\")\nerrdev(\"set\", 10, \"null\")\n\n\n# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 344, "compression_ratio": 1.0}, "src/spiceypy/tests/test_wrapper.py::364": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/utils/callbacks.py", "src/spiceypy/utils/support_types.py"], "used_names": [], "enclosing_function": "test_bschoc", "extracted_code": "", "n_imports_parsed": 11, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/tests/test_wrapper.py::353": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/utils/callbacks.py", "src/spiceypy/utils/support_types.py"], "used_names": [], "enclosing_function": "test_brckti", "extracted_code": "", "n_imports_parsed": 11, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/tests/test_wrapper.py::365": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/utils/callbacks.py", "src/spiceypy/utils/support_types.py"], "used_names": [], "enclosing_function": "test_bschoc", "extracted_code": "", "n_imports_parsed": 11, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/tests/test_spiceerrors.py::135": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py", "src/spiceypy/found_catcher.py"], "used_names": ["pytest"], "enclosing_function": "test_found_error_checker", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/tests/test_context_manager.py::69": {"resolved_imports": ["src/spiceypy/__init__.py"], "used_names": ["CoreKernels", "pytest"], "enclosing_function": "test_side_effect", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "src/spiceypy/tests/test_logging.py::51": {"resolved_imports": [], "used_names": ["subprocess", "sys"], "enclosing_function": "test_import_default_does_not_emit_info", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "src/spiceypy/tests/test_context_manager.py::85": {"resolved_imports": ["src/spiceypy/__init__.py"], "used_names": ["CoreKernels", "pytest"], "enclosing_function": "test_invalid_input", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "src/spiceypy/tests/test_spiceerrors.py::91": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py", "src/spiceypy/found_catcher.py"], "used_names": ["CoreKernels", "ExtraKernels", "cyice", "pytest", "spiceypy"], "enclosing_function": "test_no_loaded_files_exception", "extracted_code": "# Source: src/spiceypy/__init__.py\n__version__ = \"8.0.2\"\n\nfrom .spiceypy import *\nfrom .utils import support_types\nfrom .utils import exceptions\n\n# Default setting for error reporting so that programs don't just exit out!\nerract(\"set\", 10, \"return\")\nerrdev(\"set\", 10, \"null\")\n\n\n# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 344, "extracted_code_full": "# Source: src/spiceypy/__init__.py\n__version__ = \"8.0.2\"\n\nfrom .spiceypy import *\nfrom .utils import support_types\nfrom .utils import exceptions\n\n# Default setting for error reporting so that programs don't just exit out!\nerract(\"set\", 10, \"return\")\nerrdev(\"set\", 10, \"null\")\n\n\n# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 344, "compression_ratio": 1.0}, "src/spiceypy/tests/test_context_manager.py::30": {"resolved_imports": ["src/spiceypy/__init__.py"], "used_names": ["CoreKernels", "pytest"], "enclosing_function": "test_input_types", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "src/spiceypy/tests/test_gettestkernels.py::34": {"resolved_imports": [], "used_names": ["CoreKernels", "attempt_download", "pytest"], "enclosing_function": "test_gettestkernels", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "src/spiceypy/benchmarks/test_cyice.py::1120": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["cyice", "pytest"], "enclosing_function": "test_pi", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}, "src/spiceypy/tests/test_spiceerrors.py::42": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py", "src/spiceypy/found_catcher.py"], "used_names": [], "enclosing_function": "test_geterror", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 0}, "src/spiceypy/benchmarks/test_cyice.py::399": {"resolved_imports": ["src/spiceypy/__init__.py", "src/spiceypy/cyice/__init__.py"], "used_names": ["cyice", "pytest", "time"], "enclosing_function": "test_et2lst", "extracted_code": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_imports_parsed": 10, "n_files_resolved": 2, "n_chars_extracted": 66, "extracted_code_full": "# Source: src/spiceypy/cyice/__init__.py\n\"\"\"\n\nfrom .cyice import *", "n_chars_compressed": 66, "compression_ratio": 1.0}}}
oracle_context_cache_v3/AnonymouX47__term-image.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/Azure-Samples__rag-postgres-openai-python.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"repo": "Azure-Samples/rag-postgres-openai-python", "n_pairs": 51, "version": "v3_compressed", "max_tokens": 6000, "contexts": {"tests/test_openai_clients.py::65": {"resolved_imports": [], "used_names": ["common_parameters", "pytest"], "enclosing_function": "test_github_models_with_custom_values", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_frontend_routes.py::31": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_favicon", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_openai_clients.py::66": {"resolved_imports": [], "used_names": ["common_parameters", "pytest"], "enclosing_function": "test_github_models_with_custom_values", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_openai_clients.py::39": {"resolved_imports": [], "used_names": ["common_parameters", "create_openai_chat_client", "create_openai_embed_client", "pytest"], "enclosing_function": "test_github_models_configuration", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_api_routes.py::66": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_similar_handler_422", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_openai_clients.py::40": {"resolved_imports": [], "used_names": ["common_parameters", "create_openai_chat_client", "create_openai_embed_client", "pytest"], "enclosing_function": "test_github_models_configuration", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_postgres_engine.py::24": {"resolved_imports": [], "used_names": ["create_postgres_engine", "os", "pytest"], "enclosing_function": "test_create_postgres_engine", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_api_routes.py::213": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_chat_non_json_422", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_openai_clients.py::49": {"resolved_imports": [], "used_names": ["common_parameters", "create_openai_chat_client", "create_openai_embed_client", "pytest"], "enclosing_function": "test_github_models_configuration", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_postgres_searcher.py::8": {"resolved_imports": [], "used_names": [], "enclosing_function": "test_postgres_build_filter_clause_without_filters", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/e2e.py::111": {"resolved_imports": [], "used_names": ["Route"], "enclosing_function": "handle", "extracted_code": "", "n_imports_parsed": 10, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_postgres_searcher.py::36": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_postgres_searcher_search_empty_text_search", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/e2e.py::114": {"resolved_imports": [], "used_names": ["Route"], "enclosing_function": "handle", "extracted_code": "", "n_imports_parsed": 10, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_postgres_engine.py::26": {"resolved_imports": [], "used_names": ["create_postgres_engine", "os", "pytest"], "enclosing_function": "test_create_postgres_engine", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_postgres_engine.py::27": {"resolved_imports": [], "used_names": ["create_postgres_engine", "os", "pytest"], "enclosing_function": "test_create_postgres_engine", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_frontend_routes.py::40": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_assets_non_existent_404", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_api_routes.py::14": {"resolved_imports": [], "used_names": ["json", "pytest", "test_data"], "enclosing_function": "test_item_handler", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_postgres_engine.py::23": {"resolved_imports": [], "used_names": ["create_postgres_engine", "os", "pytest"], "enclosing_function": "test_create_postgres_engine", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/e2e.py::65": {"resolved_imports": [], "used_names": ["Route"], "enclosing_function": "handle", "extracted_code": "", "n_imports_parsed": 10, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_openai_clients.py::15": {"resolved_imports": [], "used_names": ["create_openai_embed_client", "pytest", "test_data"], "enclosing_function": "test_create_openai_embed_client", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_postgres_searcher.py::24": {"resolved_imports": [], "used_names": ["Filter"], "enclosing_function": "test_postgres_build_filter_clause_with_filters_numeric", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_frontend_routes.py::41": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_assets_non_existent_404", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/e2e.py::113": {"resolved_imports": [], "used_names": ["Route"], "enclosing_function": "handle", "extracted_code": "", "n_imports_parsed": 10, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_api_routes.py::105": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_search_handler_422", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_api_routes.py::64": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_similar_handler_422", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_frontend_routes.py::16": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_index", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_openai_clients.py::25": {"resolved_imports": [], "used_names": ["create_openai_chat_client", "pytest"], "enclosing_function": "test_create_openai_chat_client", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_frontend_routes.py::58": {"resolved_imports": [], "used_names": ["os", "pytest"], "enclosing_function": "test_assets", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_postgres_searcher.py::13": {"resolved_imports": [], "used_names": ["Filter"], "enclosing_function": "test_postgres_build_filter_clause_with_filters", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_frontend_routes.py::15": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_index", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_dependencies.py::9": {"resolved_imports": [], "used_names": ["common_parameters", "pytest"], "enclosing_function": "test_get_common_parameters", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_dependencies.py::20": {"resolved_imports": [], "used_names": ["common_parameters", "pytest"], "enclosing_function": "test_get_common_parameters_ollama", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_api_routes.py::16": {"resolved_imports": [], "used_names": ["json", "pytest", "test_data"], "enclosing_function": "test_item_handler", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_postgres_engine.py::25": {"resolved_imports": [], "used_names": ["create_postgres_engine", "os", "pytest"], "enclosing_function": "test_create_postgres_engine", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_api_routes.py::30": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_item_handler_404", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_dependencies.py::11": {"resolved_imports": [], "used_names": ["common_parameters", "pytest"], "enclosing_function": "test_get_common_parameters", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_dependencies.py::40": {"resolved_imports": [], "used_names": ["get_azure_credential", "pytest"], "enclosing_function": "test_get_azure_credential", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_dependencies.py::19": {"resolved_imports": [], "used_names": ["common_parameters", "pytest"], "enclosing_function": "test_get_common_parameters_ollama", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_postgres_searcher.py::41": {"resolved_imports": [], "used_names": ["ItemPublic", "pytest", "test_data"], "enclosing_function": "test_postgres_searcher_search", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_openai_clients.py::11": {"resolved_imports": [], "used_names": ["create_openai_embed_client", "pytest", "test_data"], "enclosing_function": "test_create_openai_embed_client", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_frontend_routes.py::17": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_index", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_dependencies.py::21": {"resolved_imports": [], "used_names": ["common_parameters", "pytest"], "enclosing_function": "test_get_common_parameters_ollama", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_embeddings.py::17": {"resolved_imports": [], "used_names": ["compute_text_embedding", "create_openai_embed_client", "pytest", "test_data"], "enclosing_function": "test_compute_text_embedding", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_openai_clients.py::50": {"resolved_imports": [], "used_names": ["common_parameters", "create_openai_chat_client", "create_openai_embed_client", "pytest"], "enclosing_function": "test_github_models_configuration", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_dependencies.py::29": {"resolved_imports": [], "used_names": ["common_parameters", "pytest"], "enclosing_function": "test_get_common_parameters_openai", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_frontend_routes.py::42": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_assets_non_existent_404", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_dependencies.py::41": {"resolved_imports": [], "used_names": ["get_azure_credential", "pytest"], "enclosing_function": "test_get_azure_credential", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_dependencies.py::10": {"resolved_imports": [], "used_names": ["common_parameters", "pytest"], "enclosing_function": "test_get_common_parameters", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_api_routes.py::32": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_item_handler_404", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_frontend_routes.py::18": {"resolved_imports": [], "used_names": ["pytest"], "enclosing_function": "test_index", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/e2e.py::112": {"resolved_imports": [], "used_names": ["Route"], "enclosing_function": "handle", "extracted_code": "", "n_imports_parsed": 10, "n_files_resolved": 0, "n_chars_extracted": 0}}}
oracle_context_cache_v3/BayesWitnesses__m2cgen.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"repo": "BayesWitnesses/m2cgen", "n_pairs": 58, "version": "v3_compressed", "max_tokens": 6000, "contexts": {"tests/test_ast.py::10": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast", "pytest"], "enclosing_function": "test_count_exprs", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::14": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast", "pytest"], "enclosing_function": "test_count_exprs", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::18": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast", "pytest"], "enclosing_function": "test_count_exprs", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::25": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast", "pytest"], "enclosing_function": "test_count_exprs", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::33": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast", "pytest"], "enclosing_function": "test_count_exprs", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::40": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast"], "enclosing_function": "test_count_exprs_exclude_list", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::91": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast"], "enclosing_function": "test_count_all_exprs_types", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::129": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast"], "enclosing_function": "test_num_val", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::128": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast"], "enclosing_function": "test_num_val", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::96": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["deepcopy"], "enclosing_function": "test_exprs_equality", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::125": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast"], "enclosing_function": "test_num_val", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::126": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["ast"], "enclosing_function": "test_num_val", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_ast.py::101": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py"], "used_names": ["deepcopy"], "enclosing_function": "test_exprs_hash", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_cli.py::72": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/cli.py"], "used_names": ["cli", "io", "mock", "sys"], "enclosing_function": "test_language_is_required", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_cli.py::45": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/cli.py"], "used_names": ["cli", "io"], "enclosing_function": "test_file_as_input", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_cli.py::43": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/cli.py"], "used_names": ["cli", "io"], "enclosing_function": "test_file_as_input", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_cli.py::61": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/cli.py"], "used_names": ["capture", "cli"], "enclosing_function": "test_stdin_as_input", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_cli.py::96": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/cli.py"], "used_names": ["cli"], "enclosing_function": "test_function_name_csharp_default", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_cli.py::153": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/cli.py"], "used_names": ["__version__", "cli", "io", "mock", "sys"], "enclosing_function": "test_version", "extracted_code": "# Source: m2cgen/__init__.py\n__version__ = (Path(__file__).absolute().parent / \"VERSION.txt\").read_text(encoding=\"utf-8\").strip()", "n_imports_parsed": 6, "n_files_resolved": 2, "n_chars_extracted": 129, "extracted_code_full": "# Source: m2cgen/__init__.py\n__version__ = (Path(__file__).absolute().parent / \"VERSION.txt\").read_text(encoding=\"utf-8\").strip()", "n_chars_compressed": 129, "compression_ratio": 1.0}, "tests/test_cli.py::71": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/cli.py"], "used_names": ["cli", "io", "mock", "sys"], "enclosing_function": "test_language_is_required", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_fallback_expressions.py::45": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["CInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_abs_fallback_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c.interpreter import CInterpreter\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\n\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,", "n_imports_parsed": 4, "n_files_resolved": 3, "n_chars_extracted": 981, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c.interpreter import CInterpreter\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\n\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,", "n_chars_compressed": 981, "compression_ratio": 1.0}, "tests/test_fallback_expressions.py::17": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["PythonInterpreter", "ast", "pytest"], "enclosing_function": "test_required_funs_without_fallbacks", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,", "n_imports_parsed": 4, "n_files_resolved": 3, "n_chars_extracted": 852, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,", "n_chars_compressed": 852, "compression_ratio": 1.0}, "tests/test_fallback_expressions.py::21": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["PythonInterpreter", "ast", "pytest"], "enclosing_function": "test_required_funs_without_fallbacks", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,", "n_imports_parsed": 4, "n_files_resolved": 3, "n_chars_extracted": 852, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,", "n_chars_compressed": 852, "compression_ratio": 1.0}, "tests/utils.py::216": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/assemblers/__init__.py", "m2cgen/interpreters/utils.py"], "used_names": [], "enclosing_function": "assert_code_equal", "extracted_code": "", "n_imports_parsed": 25, "n_files_resolved": 4, "n_chars_extracted": 0}, "tests/utils.py::188": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/assemblers/__init__.py", "m2cgen/interpreters/utils.py"], "used_names": ["ast"], "enclosing_function": "cmp_exprs", "extracted_code": "", "n_imports_parsed": 25, "n_files_resolved": 4, "n_chars_extracted": 0}, "tests/assemblers/test_boosting_lightgbm.py::276": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/assemblers/__init__.py"], "used_names": ["LightGBMModelAssembler", "pytest"], "enclosing_function": "test_unknown_output_transform", "extracted_code": "# Source: m2cgen/assemblers/__init__.py\nfrom m2cgen.assemblers.boosting import (\n LightGBMModelAssembler,\n XGBoostLinearModelAssembler,\n XGBoostModelAssemblerSelector,\n XGBoostTreeModelAssembler\n)\nfrom m2cgen.assemblers.ensemble import RandomForestModelAssembler\nfrom m2cgen.assemblers.linear import (\n ProcessMLEModelAssembler,\n SklearnGLMModelAssembler,\n SklearnLinearModelAssembler,\n\n XGBoostTreeModelAssembler,\n XGBoostLinearModelAssembler,\n LightGBMModelAssembler,\n SklearnSVMModelAssembler,\n LightningSVMModelAssembler,\n StatsmodelsGLMModelAssembler,\n StatsmodelsModelAssemblerSelector,\n SklearnGLMModelAssembler,\n]\n\n\nSUPPORTED_MODELS = {\n\nSUPPORTED_MODELS = {\n # LightGBM\n \"lightgbm_LGBMClassifier\": LightGBMModelAssembler,\n \"lightgbm_LGBMRegressor\": LightGBMModelAssembler,\n\n # XGBoost\n \"xgboost_XGBClassifier\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRFClassifier\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRegressor\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRFRegressor\": XGBoostModelAssemblerSelector,\n\n # Sklearn SVM\n\n # LightGBM\n \"lightgbm_LGBMClassifier\": LightGBMModelAssembler,\n \"lightgbm_LGBMRegressor\": LightGBMModelAssembler,\n\n # XGBoost\n \"xgboost_XGBClassifier\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRFClassifier\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRegressor\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRFRegressor\": XGBoostModelAssemblerSelector,\n\n # Sklearn SVM\n \"sklearn_LinearSVC\": SklearnLinearModelAssembler,", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 1566, "extracted_code_full": "# Source: m2cgen/assemblers/__init__.py\nfrom m2cgen.assemblers.boosting import (\n LightGBMModelAssembler,\n XGBoostLinearModelAssembler,\n XGBoostModelAssemblerSelector,\n XGBoostTreeModelAssembler\n)\nfrom m2cgen.assemblers.ensemble import RandomForestModelAssembler\nfrom m2cgen.assemblers.linear import (\n ProcessMLEModelAssembler,\n SklearnGLMModelAssembler,\n SklearnLinearModelAssembler,\n\n XGBoostTreeModelAssembler,\n XGBoostLinearModelAssembler,\n LightGBMModelAssembler,\n SklearnSVMModelAssembler,\n LightningSVMModelAssembler,\n StatsmodelsGLMModelAssembler,\n StatsmodelsModelAssemblerSelector,\n SklearnGLMModelAssembler,\n]\n\n\nSUPPORTED_MODELS = {\n\nSUPPORTED_MODELS = {\n # LightGBM\n \"lightgbm_LGBMClassifier\": LightGBMModelAssembler,\n \"lightgbm_LGBMRegressor\": LightGBMModelAssembler,\n\n # XGBoost\n \"xgboost_XGBClassifier\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRFClassifier\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRegressor\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRFRegressor\": XGBoostModelAssemblerSelector,\n\n # Sklearn SVM\n\n # LightGBM\n \"lightgbm_LGBMClassifier\": LightGBMModelAssembler,\n \"lightgbm_LGBMRegressor\": LightGBMModelAssembler,\n\n # XGBoost\n \"xgboost_XGBClassifier\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRFClassifier\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRegressor\": XGBoostModelAssemblerSelector,\n \"xgboost_XGBRFRegressor\": XGBoostModelAssemblerSelector,\n\n # Sklearn SVM\n \"sklearn_LinearSVC\": SklearnLinearModelAssembler,", "n_chars_compressed": 1566, "compression_ratio": 1.0}, "tests/assemblers/test_linear_sklearn.py::178": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/assemblers/__init__.py", "m2cgen/ast.py"], "used_names": ["assemblers", "linear_model", "pytest"], "enclosing_function": "test_glm_unknown_link_func", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 0}, "tests/assemblers/test_linear_statsmodels.py::159": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/assemblers/__init__.py", "m2cgen/ast.py"], "used_names": ["assemblers", "pytest", "utils"], "enclosing_function": "test_unknown_constant_position", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 0}, "tests/assemblers/test_linear_statsmodels.py::171": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/assemblers/__init__.py", "m2cgen/ast.py"], "used_names": ["assemblers", "pytest", "utils"], "enclosing_function": "test_unknown_model", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 0}, "tests/assemblers/test_linear_statsmodels.py::593": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/assemblers/__init__.py", "m2cgen/ast.py"], "used_names": ["assemblers", "pytest", "utils"], "enclosing_function": "test_glm_unknown_link_func", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 0}, "tests/assemblers/test_meta.py::38": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/assemblers/__init__.py"], "used_names": ["DummyRegressor", "RANSACModelAssembler", "RANSACRegressor", "pytest"], "enclosing_function": "test_ransac_unknown_base_estimator", "extracted_code": "# Source: m2cgen/assemblers/__init__.py\nStatsmodelsModelAssemblerSelector\n)\nfrom m2cgen.assemblers.meta import RANSACModelAssembler\nfrom m2cgen.assemblers.svm import LightningSVMModelAssembler, SklearnSVMModelAssembler\nfrom m2cgen.assemblers.tree import TreeModelAssembler\n\n__all__ = [\n SklearnLinearModelAssembler,\n StatsmodelsLinearModelAssembler,\n ProcessMLEModelAssembler,\n RANSACModelAssembler,\n TreeModelAssembler,\n\n StatsmodelsLinearModelAssembler,\n ProcessMLEModelAssembler,\n RANSACModelAssembler,\n TreeModelAssembler,\n RandomForestModelAssembler,\n XGBoostModelAssemblerSelector,\n XGBoostTreeModelAssembler,\n XGBoostLinearModelAssembler,\n LightGBMModelAssembler,\n SklearnSVMModelAssembler,\n LightningSVMModelAssembler,\n StatsmodelsGLMModelAssembler,\n\n \"sklearn_PassiveAggressiveRegressor\": SklearnLinearModelAssembler,\n \"sklearn_PoissonRegressor\": SklearnGLMModelAssembler,\n \"sklearn_RANSACRegressor\": RANSACModelAssembler,\n \"sklearn_Ridge\": SklearnLinearModelAssembler,\n \"sklearn_RidgeCV\": SklearnLinearModelAssembler,\n \"sklearn_SGDRegressor\": SklearnLinearModelAssembler,\n \"sklearn_TheilSenRegressor\": SklearnLinearModelAssembler,\n \"sklearn_TweedieRegressor\": SklearnGLMModelAssembler,\n\n # Statsmodels Linear Regressors\n \"statsmodels_GLMResultsWrapper\": StatsmodelsGLMModelAssembler,\n \"statsmodels_ProcessMLEResults\": ProcessMLEModelAssembler,", "n_imports_parsed": 7, "n_files_resolved": 3, "n_chars_extracted": 1441, "extracted_code_full": "# Source: m2cgen/assemblers/__init__.py\n StatsmodelsModelAssemblerSelector\n)\nfrom m2cgen.assemblers.meta import RANSACModelAssembler\nfrom m2cgen.assemblers.svm import LightningSVMModelAssembler, SklearnSVMModelAssembler\nfrom m2cgen.assemblers.tree import TreeModelAssembler\n\n__all__ = [\n SklearnLinearModelAssembler,\n StatsmodelsLinearModelAssembler,\n ProcessMLEModelAssembler,\n RANSACModelAssembler,\n TreeModelAssembler,\n\n StatsmodelsLinearModelAssembler,\n ProcessMLEModelAssembler,\n RANSACModelAssembler,\n TreeModelAssembler,\n RandomForestModelAssembler,\n XGBoostModelAssemblerSelector,\n XGBoostTreeModelAssembler,\n XGBoostLinearModelAssembler,\n LightGBMModelAssembler,\n SklearnSVMModelAssembler,\n LightningSVMModelAssembler,\n StatsmodelsGLMModelAssembler,\n\n \"sklearn_PassiveAggressiveRegressor\": SklearnLinearModelAssembler,\n \"sklearn_PoissonRegressor\": SklearnGLMModelAssembler,\n \"sklearn_RANSACRegressor\": RANSACModelAssembler,\n \"sklearn_Ridge\": SklearnLinearModelAssembler,\n \"sklearn_RidgeCV\": SklearnLinearModelAssembler,\n \"sklearn_SGDRegressor\": SklearnLinearModelAssembler,\n \"sklearn_TheilSenRegressor\": SklearnLinearModelAssembler,\n \"sklearn_TweedieRegressor\": SklearnGLMModelAssembler,\n\n # Statsmodels Linear Regressors\n \"statsmodels_GLMResultsWrapper\": StatsmodelsGLMModelAssembler,\n \"statsmodels_ProcessMLEResults\": ProcessMLEModelAssembler,", "n_chars_compressed": 1437, "compression_ratio": 0.9972241498959056}, "tests/assemblers/test_svm_lightning.py::30": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/assemblers/__init__.py"], "used_names": ["KernelSVC", "LightningSVMModelAssembler", "cmp_exprs", "utils"], "enclosing_function": "test_norm_in_cosine_kernel", "extracted_code": "# Source: m2cgen/assemblers/__init__.py\n)\nfrom m2cgen.assemblers.meta import RANSACModelAssembler\nfrom m2cgen.assemblers.svm import LightningSVMModelAssembler, SklearnSVMModelAssembler\nfrom m2cgen.assemblers.tree import TreeModelAssembler\n\n__all__ = [\n SklearnLinearModelAssembler,\n StatsmodelsLinearModelAssembler,\n ProcessMLEModelAssembler,\n RANSACModelAssembler,\n TreeModelAssembler,\n RandomForestModelAssembler,\n\n LightGBMModelAssembler,\n SklearnSVMModelAssembler,\n LightningSVMModelAssembler,\n StatsmodelsGLMModelAssembler,\n StatsmodelsModelAssemblerSelector,\n SklearnGLMModelAssembler,\n]\n\n\nSUPPORTED_MODELS = {\n # LightGBM\n \"lightgbm_LGBMClassifier\": LightGBMModelAssembler,\n\n\n# Lightning SVM\n \"lightning_KernelSVC\": LightningSVMModelAssembler,\n \"lightning_LinearSVC\": SklearnLinearModelAssembler,\n \"lightning_LinearSVR\": SklearnLinearModelAssembler,\n\n # Sklearn Linear Regressors\n \"sklearn_ARDRegression\": SklearnLinearModelAssembler,\n \"sklearn_BayesianRidge\": SklearnLinearModelAssembler,\n \"sklearn_ElasticNet\": SklearnLinearModelAssembler,\n \"sklearn_ElasticNetCV\": SklearnLinearModelAssembler,\n \"sklearn_GammaRegressor\": SklearnGLMModelAssembler,", "n_imports_parsed": 6, "n_files_resolved": 3, "n_chars_extracted": 1228, "extracted_code_full": "# Source: m2cgen/assemblers/__init__.py\n)\nfrom m2cgen.assemblers.meta import RANSACModelAssembler\nfrom m2cgen.assemblers.svm import LightningSVMModelAssembler, SklearnSVMModelAssembler\nfrom m2cgen.assemblers.tree import TreeModelAssembler\n\n__all__ = [\n SklearnLinearModelAssembler,\n StatsmodelsLinearModelAssembler,\n ProcessMLEModelAssembler,\n RANSACModelAssembler,\n TreeModelAssembler,\n RandomForestModelAssembler,\n\n LightGBMModelAssembler,\n SklearnSVMModelAssembler,\n LightningSVMModelAssembler,\n StatsmodelsGLMModelAssembler,\n StatsmodelsModelAssemblerSelector,\n SklearnGLMModelAssembler,\n]\n\n\nSUPPORTED_MODELS = {\n # LightGBM\n \"lightgbm_LGBMClassifier\": LightGBMModelAssembler,\n\n\n # Lightning SVM\n \"lightning_KernelSVC\": LightningSVMModelAssembler,\n \"lightning_LinearSVC\": SklearnLinearModelAssembler,\n \"lightning_LinearSVR\": SklearnLinearModelAssembler,\n\n # Sklearn Linear Regressors\n \"sklearn_ARDRegression\": SklearnLinearModelAssembler,\n \"sklearn_BayesianRidge\": SklearnLinearModelAssembler,\n \"sklearn_ElasticNet\": SklearnLinearModelAssembler,\n \"sklearn_ElasticNetCV\": SklearnLinearModelAssembler,\n \"sklearn_GammaRegressor\": SklearnGLMModelAssembler,", "n_chars_compressed": 1224, "compression_ratio": 0.996742671009772}, "tests/assemblers/test_svm_sklearn.py::109": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/assemblers/__init__.py"], "used_names": ["SVC", "SklearnSVMModelAssembler", "pytest"], "enclosing_function": "test_unknown_kernel", "extracted_code": "# Source: m2cgen/assemblers/__init__.py\n)\nfrom m2cgen.assemblers.meta import RANSACModelAssembler\nfrom m2cgen.assemblers.svm import LightningSVMModelAssembler, SklearnSVMModelAssembler\nfrom m2cgen.assemblers.tree import TreeModelAssembler\n\n__all__ = [\n SklearnLinearModelAssembler,\n StatsmodelsLinearModelAssembler,\n ProcessMLEModelAssembler,\n RANSACModelAssembler,\n TreeModelAssembler,\n RandomForestModelAssembler,\n\n XGBoostLinearModelAssembler,\n LightGBMModelAssembler,\n SklearnSVMModelAssembler,\n LightningSVMModelAssembler,\n StatsmodelsGLMModelAssembler,\n StatsmodelsModelAssemblerSelector,\n SklearnGLMModelAssembler,\n]\n\n\nSUPPORTED_MODELS = {\n # LightGBM\n\n \"sklearn_LinearSVC\": SklearnLinearModelAssembler,\n \"sklearn_LinearSVR\": SklearnLinearModelAssembler,\n \"sklearn_NuSVC\": SklearnSVMModelAssembler,\n \"sklearn_NuSVR\": SklearnSVMModelAssembler,\n \"sklearn_OneClassSVM\": SklearnSVMModelAssembler,\n \"sklearn_SVC\": SklearnSVMModelAssembler,\n \"sklearn_SVR\": SklearnSVMModelAssembler,\n\n # Lightning SVM\n \"lightning_KernelSVC\": LightningSVMModelAssembler,\n \"lightning_LinearSVC\": SklearnLinearModelAssembler,\n \"lightning_LinearSVR\": SklearnLinearModelAssembler,\n\n \"sklearn_LinearSVR\": SklearnLinearModelAssembler,\n \"sklearn_NuSVC\": SklearnSVMModelAssembler,\n \"sklearn_NuSVR\": SklearnSVMModelAssembler,\n \"sklearn_OneClassSVM\": SklearnSVMModelAssembler,\n \"sklearn_SVC\": SklearnSVMModelAssembler,\n \"sklearn_SVR\": SklearnSVMModelAssembler,\n\n # Lightning SVM\n \"lightning_KernelSVC\": LightningSVMModelAssembler,\n \"lightning_LinearSVC\": SklearnLinearModelAssembler,\n \"lightning_LinearSVR\": SklearnLinearModelAssembler,\n\n\n\"sklearn_NuSVC\": SklearnSVMModelAssembler,\n \"sklearn_NuSVR\": SklearnSVMModelAssembler,\n \"sklearn_OneClassSVM\": SklearnSVMModelAssembler,\n \"sklearn_SVC\": SklearnSVMModelAssembler,\n \"sklearn_SVR\": SklearnSVMModelAssembler,\n\n # Lightning SVM\n \"lightning_KernelSVC\": LightningSVMModelAssembler,\n \"lightning_LinearSVC\": SklearnLinearModelAssembler,\n \"lightning_LinearSVR\": SklearnLinearModelAssembler,\n\n # Sklearn Linear Regressors", "n_imports_parsed": 7, "n_files_resolved": 3, "n_chars_extracted": 2174, "extracted_code_full": "# Source: m2cgen/assemblers/__init__.py\n)\nfrom m2cgen.assemblers.meta import RANSACModelAssembler\nfrom m2cgen.assemblers.svm import LightningSVMModelAssembler, SklearnSVMModelAssembler\nfrom m2cgen.assemblers.tree import TreeModelAssembler\n\n__all__ = [\n SklearnLinearModelAssembler,\n StatsmodelsLinearModelAssembler,\n ProcessMLEModelAssembler,\n RANSACModelAssembler,\n TreeModelAssembler,\n RandomForestModelAssembler,\n\n XGBoostLinearModelAssembler,\n LightGBMModelAssembler,\n SklearnSVMModelAssembler,\n LightningSVMModelAssembler,\n StatsmodelsGLMModelAssembler,\n StatsmodelsModelAssemblerSelector,\n SklearnGLMModelAssembler,\n]\n\n\nSUPPORTED_MODELS = {\n # LightGBM\n\n \"sklearn_LinearSVC\": SklearnLinearModelAssembler,\n \"sklearn_LinearSVR\": SklearnLinearModelAssembler,\n \"sklearn_NuSVC\": SklearnSVMModelAssembler,\n \"sklearn_NuSVR\": SklearnSVMModelAssembler,\n \"sklearn_OneClassSVM\": SklearnSVMModelAssembler,\n \"sklearn_SVC\": SklearnSVMModelAssembler,\n \"sklearn_SVR\": SklearnSVMModelAssembler,\n\n # Lightning SVM\n \"lightning_KernelSVC\": LightningSVMModelAssembler,\n \"lightning_LinearSVC\": SklearnLinearModelAssembler,\n \"lightning_LinearSVR\": SklearnLinearModelAssembler,\n\n \"sklearn_LinearSVR\": SklearnLinearModelAssembler,\n \"sklearn_NuSVC\": SklearnSVMModelAssembler,\n \"sklearn_NuSVR\": SklearnSVMModelAssembler,\n \"sklearn_OneClassSVM\": SklearnSVMModelAssembler,\n \"sklearn_SVC\": SklearnSVMModelAssembler,\n \"sklearn_SVR\": SklearnSVMModelAssembler,\n\n # Lightning SVM\n \"lightning_KernelSVC\": LightningSVMModelAssembler,\n \"lightning_LinearSVC\": SklearnLinearModelAssembler,\n \"lightning_LinearSVR\": SklearnLinearModelAssembler,\n\n\n \"sklearn_NuSVC\": SklearnSVMModelAssembler,\n \"sklearn_NuSVR\": SklearnSVMModelAssembler,\n \"sklearn_OneClassSVM\": SklearnSVMModelAssembler,\n \"sklearn_SVC\": SklearnSVMModelAssembler,\n \"sklearn_SVR\": SklearnSVMModelAssembler,\n\n # Lightning SVM\n \"lightning_KernelSVC\": LightningSVMModelAssembler,\n \"lightning_LinearSVC\": SklearnLinearModelAssembler,\n \"lightning_LinearSVR\": SklearnLinearModelAssembler,\n\n # Sklearn Linear Regressors", "n_chars_compressed": 2170, "compression_ratio": 0.9981600735970562}, "tests/interpreters/test_c.py::30": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["CInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c.interpreter import CInterpreter\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\n\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 981, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c.interpreter import CInterpreter\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\n\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,", "n_chars_compressed": 981, "compression_ratio": 1.0}, "tests/interpreters/test_c.py::90": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["CInterpreter", "assert_code_equal", "ast", "product", "pytest"], "enclosing_function": "test_associativity_in_bin_num_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c.interpreter import CInterpreter\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\n\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 981, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c.interpreter import CInterpreter\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\n\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,", "n_chars_compressed": 981, "compression_ratio": 1.0}, "tests/interpreters/test_c_sharp.py::34": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["CSharpInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c.interpreter import CInterpreter\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\n\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1065, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c.interpreter import CInterpreter\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\n\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,", "n_chars_compressed": 1065, "compression_ratio": 1.0}, "tests/interpreters/test_dart.py::30": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["DartInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c.interpreter import CInterpreter\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\n\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1031, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c.interpreter import CInterpreter\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\n\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 1031, "compression_ratio": 1.0}, "tests/interpreters/test_elixir.py::42": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["ElixirInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\n\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 927, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.c_sharp.interpreter import CSharpInterpreter\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\n\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 927, "compression_ratio": 1.0}, "tests/interpreters/test_f_sharp.py::28": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["FSharpInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\n\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 967, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.dart.interpreter import DartInterpreter\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\n\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 967, "compression_ratio": 1.0}, "tests/interpreters/test_go.py::30": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["GoInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\n\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1122, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\n\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,", "n_chars_compressed": 1122, "compression_ratio": 1.0}, "tests/interpreters/test_go.py::90": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["GoInterpreter", "assert_code_equal", "ast", "product", "pytest"], "enclosing_function": "test_associativity_in_bin_num_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\n\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1122, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.elixir.interpreter import ElixirInterpreter\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\n\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,", "n_chars_compressed": 1122, "compression_ratio": 1.0}, "tests/interpreters/test_haskell.py::31": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["HaskellInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1019, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 1019, "compression_ratio": 1.0}, "tests/interpreters/test_haskell.py::94": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["HaskellInterpreter", "assert_code_equal", "ast", "product", "pytest"], "enclosing_function": "test_associativity_in_bin_num_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1019, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.f_sharp.interpreter import FSharpInterpreter\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 1019, "compression_ratio": 1.0}, "tests/interpreters/test_java.py::32": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["JavaInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1032, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.go.interpreter import GoInterpreter\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,", "n_chars_compressed": 1031, "compression_ratio": 0.999031007751938}, "tests/interpreters/test_javascript.py::30": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["JavascriptInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1015, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,", "n_chars_compressed": 1015, "compression_ratio": 1.0}, "tests/interpreters/test_javascript.py::90": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["JavascriptInterpreter", "assert_code_equal", "ast", "product", "pytest"], "enclosing_function": "test_associativity_in_bin_num_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1015, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.haskell.interpreter import HaskellInterpreter\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,", "n_chars_compressed": 1015, "compression_ratio": 1.0}, "tests/interpreters/test_php.py::31": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["PhpInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 896, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 896, "compression_ratio": 1.0}, "tests/interpreters/test_php.py::94": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["PhpInterpreter", "assert_code_equal", "ast", "product", "pytest"], "enclosing_function": "test_associativity_in_bin_num_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 896, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.java.interpreter import JavaInterpreter\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 896, "compression_ratio": 1.0}, "tests/interpreters/test_powershell.py::30": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["PowershellInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 905, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 905, "compression_ratio": 1.0}, "tests/interpreters/test_powershell.py::90": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["PowershellInterpreter", "assert_code_equal", "ast", "product", "pytest"], "enclosing_function": "test_associativity_in_bin_num_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 905, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.javascript.interpreter import JavascriptInterpreter\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 905, "compression_ratio": 1.0}, "tests/interpreters/test_python.py::382": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["PythonInterpreter", "ast"], "enclosing_function": "test_deep_expression", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 852, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,", "n_chars_compressed": 852, "compression_ratio": 1.0}, "tests/interpreters/test_python.py::27": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["PythonInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 852, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.php.interpreter import PhpInterpreter\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,", "n_chars_compressed": 852, "compression_ratio": 1.0}, "tests/interpreters/test_r.py::29": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["RInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 774, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.powershell.interpreter import PowershellInterpreter\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 774, "compression_ratio": 1.0}, "tests/interpreters/test_ruby.py::29": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["RubyInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 636, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 636, "compression_ratio": 1.0}, "tests/interpreters/test_ruby.py::103": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["RubyInterpreter", "assert_code_equal", "ast", "product", "pytest"], "enclosing_function": "test_associativity_in_bin_num_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 636, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.python.interpreter import PythonInterpreter\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 636, "compression_ratio": 1.0}, "tests/interpreters/test_rust.py::30": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["RustInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 550, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 550, "compression_ratio": 1.0}, "tests/interpreters/test_rust.py::90": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["RustInterpreter", "assert_code_equal", "ast", "product", "pytest"], "enclosing_function": "test_associativity_in_bin_num_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 550, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.r.interpreter import RInterpreter\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,\n ElixirInterpreter,\n]", "n_chars_compressed": 550, "compression_ratio": 1.0}, "tests/interpreters/test_visual_basic.py::32": {"resolved_imports": ["m2cgen/__init__.py", "m2cgen/ast.py", "m2cgen/interpreters/__init__.py"], "used_names": ["VisualBasicInterpreter", "assert_code_equal", "ast"], "enclosing_function": "test_if_expr", "extracted_code": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 696, "extracted_code_full": "# Source: m2cgen/interpreters/__init__.py\nfrom m2cgen.interpreters.ruby.interpreter import RubyInterpreter\nfrom m2cgen.interpreters.rust.interpreter import RustInterpreter\nfrom m2cgen.interpreters.visual_basic.interpreter import VisualBasicInterpreter\n\n__all__ = [\n JavaInterpreter,\n PythonInterpreter,\n CInterpreter,\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n\n GoInterpreter,\n JavascriptInterpreter,\n VisualBasicInterpreter,\n CSharpInterpreter,\n PowershellInterpreter,\n RInterpreter,\n PhpInterpreter,\n DartInterpreter,\n HaskellInterpreter,\n RubyInterpreter,\n FSharpInterpreter,\n RustInterpreter,", "n_chars_compressed": 696, "compression_ratio": 1.0}}}
oracle_context_cache_v3/BoboTiG__python-mss.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/BrainBlend-AI__atomic-agents.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/CalebBell__fluids.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/Chen-zexi__vllm-cli.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/Cloxl__xhshow.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/Cranot__roam-code.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/CursorTouch__Windows-MCP.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/DHI__terracotta.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/DLR-RM__stable-baselines3.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/DebarghaG__proofofthought.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/DeepLcom__deepl-python.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"repo": "DeepLcom/deepl-python", "n_pairs": 60, "version": "v3_compressed", "max_tokens": 6000, "contexts": {"tests/test_cli.py::31": {"resolved_imports": ["deepl/__init__.py", "deepl/__main__.py"], "used_names": [], "enclosing_function": "test_help", "extracted_code": "", "n_imports_parsed": 8, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_cli.py::60": {"resolved_imports": ["deepl/__init__.py", "deepl/__main__.py"], "used_names": [], "enclosing_function": "test_no_auth", "extracted_code": "", "n_imports_parsed": 8, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_cli.py::106": {"resolved_imports": ["deepl/__init__.py", "deepl/__main__.py"], "used_names": [], "enclosing_function": "test_no_command", "extracted_code": "", "n_imports_parsed": 8, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_general.py::113": {"resolved_imports": ["deepl/__init__.py"], "used_names": [], "enclosing_function": "test_glossary_languages", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_general.py::102": {"resolved_imports": ["deepl/__init__.py"], "used_names": [], "enclosing_function": "test_language", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_general.py::132": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "example_text", "os", "patch"], "enclosing_function": "test_user_agent", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_glossary.py::27": {"resolved_imports": ["deepl/__init__.py"], "used_names": [], "enclosing_function": "test_glossary_create", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_glossary.py::130": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "pytest"], "enclosing_function": "test_glossary_get_entries", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_glossary.py::202": {"resolved_imports": ["deepl/__init__.py"], "used_names": [], "enclosing_function": "test_glossary_translate_text_basic", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_glossary.py::208": {"resolved_imports": ["deepl/__init__.py"], "used_names": [], "enclosing_function": "test_glossary_translate_text_basic", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_glossary.py::23": {"resolved_imports": ["deepl/__init__.py"], "used_names": [], "enclosing_function": "test_glossary_create", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_glossary.py::24": {"resolved_imports": ["deepl/__init__.py"], "used_names": [], "enclosing_function": "test_glossary_create", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_glossary.py::178": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["needs_real_server"], "enclosing_function": "test_glossary_translate_text_sentence", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_glossary.py::69": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "pytest"], "enclosing_function": "test_glossary_create_invalid", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_glossary.py::91": {"resolved_imports": ["deepl/__init__.py"], "used_names": [], "enclosing_function": "test_glossary_create_large", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_glossary.py::22": {"resolved_imports": ["deepl/__init__.py"], "used_names": [], "enclosing_function": "test_glossary_create", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_glossary.py::30": {"resolved_imports": ["deepl/__init__.py"], "used_names": [], "enclosing_function": "test_glossary_create", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_multilingual_glossary.py::201": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["util"], "enclosing_function": "test_glossary_create_csv", "extracted_code": "# Source: deepl/__init__.py\n)\n\nfrom .util import ( # noqa\n auth_key_is_free_account,\n convert_tsv_to_dict,\n convert_dict_to_tsv,\n validate_glossary_term,\n)\n\n__all__ = [\n \"__version__\",\n \"__author__\",", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 218, "extracted_code_full": "# Source: deepl/__init__.py\n)\n\nfrom .util import ( # noqa\n auth_key_is_free_account,\n convert_tsv_to_dict,\n convert_dict_to_tsv,\n validate_glossary_term,\n)\n\n__all__ = [\n \"__version__\",\n \"__author__\",", "n_chars_compressed": 218, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::107": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["MultilingualGlossaryDictionaryEntries"], "enclosing_function": "test_glossary_dictionary_update", "extracted_code": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1441, "extracted_code_full": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_chars_compressed": 1441, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::108": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["MultilingualGlossaryDictionaryEntries"], "enclosing_function": "test_glossary_dictionary_update", "extracted_code": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1441, "extracted_code_full": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_chars_compressed": 1441, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::314": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["MultilingualGlossaryDictionaryEntries", "deepl", "pytest", "util"], "enclosing_function": "test_glossary_get_entries", "extracted_code": "# Source: deepl/__init__.py\n)\n\nfrom .util import ( # noqa\n auth_key_is_free_account,\n convert_tsv_to_dict,\n convert_dict_to_tsv,\n validate_glossary_term,\n)\n\n__all__ = [\n \"__version__\",\n \"__author__\",\n\n\n# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1662, "extracted_code_full": "# Source: deepl/__init__.py\n)\n\nfrom .util import ( # noqa\n auth_key_is_free_account,\n convert_tsv_to_dict,\n convert_dict_to_tsv,\n validate_glossary_term,\n)\n\n__all__ = [\n \"__version__\",\n \"__author__\",\n\n\n# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_chars_compressed": 1662, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::132": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["MultilingualGlossaryDictionaryEntries", "pytest"], "enclosing_function": "test_glossary_name_update", "extracted_code": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1441, "extracted_code_full": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_chars_compressed": 1441, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::439": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["MultilingualGlossaryDictionaryEntries"], "enclosing_function": "test_glossary_translate_text_basic", "extracted_code": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1441, "extracted_code_full": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_chars_compressed": 1441, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::140": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["MultilingualGlossaryDictionaryEntries", "pytest"], "enclosing_function": "test_glossary_name_update", "extracted_code": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1441, "extracted_code_full": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_chars_compressed": 1441, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::205": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["util"], "enclosing_function": "test_glossary_create_csv", "extracted_code": "# Source: deepl/__init__.py\n)\n\nfrom .util import ( # noqa\n auth_key_is_free_account,\n convert_tsv_to_dict,\n convert_dict_to_tsv,\n validate_glossary_term,\n)\n\n__all__ = [\n \"__version__\",\n \"__author__\",", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 218, "extracted_code_full": "# Source: deepl/__init__.py\n)\n\nfrom .util import ( # noqa\n auth_key_is_free_account,\n convert_tsv_to_dict,\n convert_dict_to_tsv,\n validate_glossary_term,\n)\n\n__all__ = [\n \"__version__\",\n \"__author__\",", "n_chars_compressed": 218, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::206": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["util"], "enclosing_function": "test_glossary_create_csv", "extracted_code": "# Source: deepl/__init__.py\n)\n\nfrom .util import ( # noqa\n auth_key_is_free_account,\n convert_tsv_to_dict,\n convert_dict_to_tsv,\n validate_glossary_term,\n)\n\n__all__ = [\n \"__version__\",\n \"__author__\",", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 218, "extracted_code_full": "# Source: deepl/__init__.py\n)\n\nfrom .util import ( # noqa\n auth_key_is_free_account,\n convert_tsv_to_dict,\n convert_dict_to_tsv,\n validate_glossary_term,\n)\n\n__all__ = [\n \"__version__\",\n \"__author__\",", "n_chars_compressed": 218, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::416": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["MultilingualGlossaryDictionaryEntries", "needs_real_server"], "enclosing_function": "test_glossary_translate_text_sentence", "extracted_code": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1441, "extracted_code_full": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_chars_compressed": 1441, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::171": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["MultilingualGlossaryDictionaryEntries", "util"], "enclosing_function": "test_glossary_dictionary_replace", "extracted_code": "# Source: deepl/__init__.py\n)\n\nfrom .util import ( # noqa\n auth_key_is_free_account,\n convert_tsv_to_dict,\n convert_dict_to_tsv,\n validate_glossary_term,\n)\n\n__all__ = [\n \"__version__\",\n \"__author__\",\n\n\n# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1662, "extracted_code_full": "# Source: deepl/__init__.py\n)\n\nfrom .util import ( # noqa\n auth_key_is_free_account,\n convert_tsv_to_dict,\n convert_dict_to_tsv,\n validate_glossary_term,\n)\n\n__all__ = [\n \"__version__\",\n \"__author__\",\n\n\n# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_chars_compressed": 1662, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::33": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["MultilingualGlossaryDictionaryEntries"], "enclosing_function": "test_glossary_create", "extracted_code": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1441, "extracted_code_full": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_chars_compressed": 1441, "compression_ratio": 1.0}, "tests/test_multilingual_glossary.py::51": {"resolved_imports": ["deepl/__init__.py", "deepl/util.py", "deepl/api_data.py"], "used_names": ["MultilingualGlossaryDictionaryEntries"], "enclosing_function": "test_glossary_create", "extracted_code": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_imports_parsed": 5, "n_files_resolved": 3, "n_chars_extracted": 1441, "extracted_code_full": "# Source: deepl/api_data.py\nclass MultilingualGlossaryDictionaryEntries:\n def __init__(\n self,\n source_lang: str,\n target_lang: str,\n entries: Dict[str, str],\n ):\n self._source_lang = source_lang\n self._target_lang = target_lang\n self._entries = entries\n\n def __str__(self) -> str:\n return (\n \"MultilingualGlossaryDictionaryEntries: Source Language \"\n f\"{self._source_lang}, Target Language {self._target_lang} \"\n f\"Contents: {self._entries}\"\n )\n\n @staticmethod\n def from_json(json) -> \"MultilingualGlossaryDictionaryEntries\":\n \"\"\"Create MultilingualGlossaryDictionaryEntries from the given\n API JSON object.\n \"\"\"\n return MultilingualGlossaryDictionaryEntries(\n str(json[\"source_lang\"]),\n str(json[\"target_lang\"]),\n json[\"entries\"],\n )\n\n def to_json(self):\n \"\"\"Create API JSON object from\n MultilingualGlossaryDictionaryEntries\n \"\"\"\n return {\n \"source_lang\": self._source_lang,\n \"target_lang\": self._target_lang,\n \"entries\": self._entries,\n }\n\n @property\n def source_lang(self) -> str:\n return self._source_lang\n\n @property\n def target_lang(self) -> str:\n return self._target_lang\n\n @property\n def entries(self) -> Dict[str, str]:\n return self._entries", "n_chars_compressed": 1441, "compression_ratio": 1.0}, "tests/test_rephrase_text.py::31": {"resolved_imports": ["deepl/api_data.py"], "used_names": ["WriteResult"], "enclosing_function": "_check_sanity_of_improvements", "extracted_code": "# Source: deepl/api_data.py\nclass WriteResult:\n \"\"\"Holds the result of a text improvement request.\"\"\"\n\n def __init__(\n self, text: str, detected_source_language: str, target_language: str\n ):\n self.text = text\n self.detected_source_language = detected_source_language\n self.target_language = target_language\n\n def __str__(self):\n return self.text", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 393, "extracted_code_full": "# Source: deepl/api_data.py\nclass WriteResult:\n \"\"\"Holds the result of a text improvement request.\"\"\"\n\n def __init__(\n self, text: str, detected_source_language: str, target_language: str\n ):\n self.text = text\n self.detected_source_language = detected_source_language\n self.target_language = target_language\n\n def __str__(self):\n return self.text", "n_chars_compressed": 393, "compression_ratio": 1.0}, "tests/test_rephrase_text.py::34": {"resolved_imports": ["deepl/api_data.py"], "used_names": ["WriteResult"], "enclosing_function": "_check_sanity_of_improvements", "extracted_code": "# Source: deepl/api_data.py\nclass WriteResult:\n \"\"\"Holds the result of a text improvement request.\"\"\"\n\n def __init__(\n self, text: str, detected_source_language: str, target_language: str\n ):\n self.text = text\n self.detected_source_language = detected_source_language\n self.target_language = target_language\n\n def __str__(self):\n return self.text", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 393, "extracted_code_full": "# Source: deepl/api_data.py\nclass WriteResult:\n \"\"\"Holds the result of a text improvement request.\"\"\"\n\n def __init__(\n self, text: str, detected_source_language: str, target_language: str\n ):\n self.text = text\n self.detected_source_language = detected_source_language\n self.target_language = target_language\n\n def __str__(self):\n return self.text", "n_chars_compressed": 393, "compression_ratio": 1.0}, "tests/test_style_rules.py::15": {"resolved_imports": [], "used_names": ["needs_mock_server"], "enclosing_function": "test_get_all_style_rules", "extracted_code": "", "n_imports_parsed": 1, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_style_rules.py::18": {"resolved_imports": [], "used_names": ["needs_mock_server"], "enclosing_function": "test_get_all_style_rules", "extracted_code": "", "n_imports_parsed": 1, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_style_rules.py::20": {"resolved_imports": [], "used_names": ["needs_mock_server"], "enclosing_function": "test_get_all_style_rules", "extracted_code": "", "n_imports_parsed": 1, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/test_translate_document.py::318": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "example_text", "needs_mock_server", "time"], "enclosing_function": "test_translate_document_request_fields", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_document.py::204": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "example_text", "pytest", "re"], "enclosing_function": "test_document_failure_during_translation", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_document.py::186": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["example_text", "needs_mock_server"], "enclosing_function": "test_document_output_format", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_translate_document.py::203": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "example_text", "pytest", "re"], "enclosing_function": "test_document_failure_during_translation", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_document.py::285": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["example_text", "time"], "enclosing_function": "test_translate_document_string", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_translate_document.py::28": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "example_text"], "enclosing_function": "test_translate_document_from_filepath", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_document.py::261": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "needs_mock_server", "time"], "enclosing_function": "test_translate_document_low_level", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_document.py::128": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["io", "needs_mock_server"], "enclosing_function": "test_translate_large_document", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_translate_document.py::27": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "example_text"], "enclosing_function": "test_translate_document_from_filepath", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_document.py::29": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "example_text"], "enclosing_function": "test_translate_document_from_filepath", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 7, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_text.py::432": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "needs_real_server"], "enclosing_function": "test_custom_instructions", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_text.py::153": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["example_text", "needs_mock_server", "time"], "enclosing_function": "test_translate_with_retries", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_translate_text.py::28": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "example_text", "pytest"], "enclosing_function": "test_model_type", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_text.py::433": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "needs_real_server"], "enclosing_function": "test_custom_instructions", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_text.py::411": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "example_text"], "enclosing_function": "test_extra_body_params", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_translate_text.py::141": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_invalid_text", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_translate_text.py::14": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["example_text"], "enclosing_function": "test_single_text", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_translate_text.py::16": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["example_text"], "enclosing_function": "test_single_text", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_util.py::27": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "pytest"], "enclosing_function": "test_convert_dict_to_tsv", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_util.py::11": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "pytest"], "enclosing_function": "test_convert_tsv_to_dict", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_util.py::29": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "pytest"], "enclosing_function": "test_convert_dict_to_tsv", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_util.py::32": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "pytest"], "enclosing_function": "test_convert_dict_to_tsv", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_util.py::19": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "pytest"], "enclosing_function": "test_convert_tsv_to_dict", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_util.py::13": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "pytest"], "enclosing_function": "test_convert_tsv_to_dict", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}, "tests/test_util.py::16": {"resolved_imports": ["deepl/__init__.py"], "used_names": ["deepl", "pytest"], "enclosing_function": "test_convert_tsv_to_dict", "extracted_code": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 949, "extracted_code_full": "# Source: deepl/__init__.py\n# Copyright 2022 DeepL SE (https://www.deepl.com)\n# Use of this source code is governed by an MIT\n# license that can be found in the LICENSE file.\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\n\nfrom .version import VERSION as __version__ # noqa\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n\n__author__ = \"DeepL SE <python-api@deepl.com>\"\n\nfrom .deepl_client import DeepLClient\n\nfrom .exceptions import ( # noqa\n AuthorizationException,\n ConnectionException,\n DeepLException,\n DocumentNotReadyException,\n DocumentTranslationException,\n GlossaryNotFoundException,\n TooManyRequestsException,", "n_chars_compressed": 949, "compression_ratio": 1.0}}}
oracle_context_cache_v3/Delgan__loguru.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/DenisCarriere__geocoder.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/DisnakeDev__disnake.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/DonDebonair__slack-machine.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/Donkie__Spoolman.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"repo": "Donkie/Spoolman", "n_pairs": 115, "version": "v3_compressed", "max_tokens": 6000, "contexts": {"tests_integration/tests/conftest.py::295": {"resolved_imports": [], "used_names": ["Any", "Iterable", "pytest"], "enclosing_function": "assert_lists_compatible", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/conftest.py::280": {"resolved_imports": [], "used_names": ["Any"], "enclosing_function": "assert_dicts_compatible", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/fields/test_create.py::199": {"resolved_imports": [], "used_names": ["URL", "httpx", "json"], "enclosing_function": "test_add_text_field_invalid_data", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/fields/test_create.py::230": {"resolved_imports": [], "used_names": ["URL", "httpx", "json"], "enclosing_function": "test_add_choice_field_invalid_choices", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/fields/test_create.py::317": {"resolved_imports": [], "used_names": ["URL", "assert_httpx_success", "httpx", "json"], "enclosing_function": "test_update_field_change_field_type", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/fields/test_delete.py::29": {"resolved_imports": [], "used_names": ["URL", "assert_httpx_success", "httpx", "json"], "enclosing_function": "test_delete_field", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/fields/test_delete.py::20": {"resolved_imports": [], "used_names": ["URL", "assert_httpx_success", "httpx", "json"], "enclosing_function": "test_delete_field", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/fields/test_get.py::20": {"resolved_imports": [], "used_names": ["URL", "assert_httpx_success", "assert_lists_compatible", "httpx", "json"], "enclosing_function": "test_get_field", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/fields/test_utilize.py::59": {"resolved_imports": [], "used_names": ["URL", "assert_httpx_code", "httpx"], "enclosing_function": "test_add_vendor_with_invalid_extra_field", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/fields/test_utilize.py::20": {"resolved_imports": [], "used_names": ["URL", "assert_httpx_success", "httpx", "json"], "enclosing_function": "test_add_vendor_with_extra_field", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_add.py::74": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "datetime", "httpx", "timezone"], "enclosing_function": "test_add_filament", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_add.py::174": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_add_filament_multi_color_errors", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_add.py::127": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_add_filament_color_hex_alpha", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_add.py::152": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_add_filament_multi_color", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_add.py::153": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_add_filament_multi_color", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_add.py::96": {"resolved_imports": [], "used_names": ["URL", "assert_dicts_compatible", "httpx"], "enclosing_function": "test_add_filament_required", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_add.py::50": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "datetime", "httpx", "timezone"], "enclosing_function": "test_add_filament", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_delete.py::40": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx"], "enclosing_function": "test_delete_filament", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_delete.py::51": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_delete_filament_not_found", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_find.py::169": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_filaments_limit_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_find.py::217": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_filaments_limit_asc_offset_outside_range", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_find.py::167": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_filaments_limit_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_find.py::119": {"resolved_imports": [], "used_names": ["URL", "assert_lists_compatible", "httpx"], "enclosing_function": "test_find_all_filaments", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_find.py::130": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_filaments_sort_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_find.py::432": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_filaments_by_article_number", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_find.py::129": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_filaments_sort_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_get.py::87": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_filament_not_found", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_get.py::89": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_filament_not_found", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_get.py::54": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "httpx"], "enclosing_function": "test_get_filament", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_update.py::173": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx"], "enclosing_function": "test_update_filament_cant_set_none", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_update.py::188": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_update_filament_not_found", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_update.py::190": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_update_filament_not_found", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_update.py::127": {"resolved_imports": [], "used_names": ["URL", "assert_dicts_compatible", "httpx"], "enclosing_function": "test_update_filament_multi_color", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/filament/test_update.py::72": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "httpx"], "enclosing_function": "test_update_filament", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/setting/test_get.py::27": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_unknown", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/setting/test_get.py::38": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_all", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/setting/test_get.py::16": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_currency", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/setting/test_set.py::75": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_set_unknown", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/setting/test_set.py::85": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_set_currency_wrong_type", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/setting/test_set.py::21": {"resolved_imports": [], "used_names": ["URL", "httpx", "json"], "enclosing_function": "test_set_currency", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/setting/test_set.py::61": {"resolved_imports": [], "used_names": ["URL", "httpx", "json"], "enclosing_function": "test_unset_currency", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/setting/test_set.py::104": {"resolved_imports": [], "used_names": ["URL", "httpx", "json"], "enclosing_function": "test_set_big_value", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_add.py::73": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "datetime", "httpx", "length_from_weight", "pytest", "timezone"], "enclosing_function": "test_add_spool_remaining_weight", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_add.py::197": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx"], "enclosing_function": "test_add_spool_both_used_and_remaining_weight", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_add.py::60": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "datetime", "httpx", "length_from_weight", "pytest", "timezone"], "enclosing_function": "test_add_spool_remaining_weight", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_add.py::62": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "datetime", "httpx", "length_from_weight", "pytest", "timezone"], "enclosing_function": "test_add_spool_remaining_weight", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_add.py::321": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "datetime", "httpx", "length_from_weight", "pytest", "timezone"], "enclosing_function": "test_add_spool_spool_weight", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_add.py::249": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "datetime", "httpx", "length_from_weight", "pytest", "timezone"], "enclosing_function": "test_add_spool_initial_weight", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_add.py::59": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "datetime", "httpx", "length_from_weight", "pytest", "timezone"], "enclosing_function": "test_add_spool_remaining_weight", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_add.py::61": {"resolved_imports": [], "used_names": ["Any", "URL", "assert_dicts_compatible", "datetime", "httpx", "length_from_weight", "pytest", "timezone"], "enclosing_function": "test_add_spool_remaining_weight", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_delete.py::30": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx"], "enclosing_function": "test_delete_spool", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_delete.py::41": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_delete_spool_not_found", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_find.py::165": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_spools_limit_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_find.py::213": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_spools_limit_asc_offset_outside_range", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_find.py::405": {"resolved_imports": [], "used_names": ["URL", "assert_lists_compatible", "httpx"], "enclosing_function": "test_find_spools_by_multiple_vendor_ids", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_find.py::163": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_spools_limit_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_find.py::187": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_spools_limit_asc_offset", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_find.py::132": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_spools_sort_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_find.py::131": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_spools_sort_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_find.py::372": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_spools_by_empty_filament_vendor_name", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_find.py::422": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_spools_by_empty_filament_vendor_id", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_find.py::435": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_spools_by_location", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_get.py::144": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_spool_not_found", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_get.py::44": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx"], "enclosing_function": "test_get_spool", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_get.py::146": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_spool_not_found", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_get.py::86": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx", "pytest"], "enclosing_function": "test_get_spool_default_weights", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_get.py::87": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx", "pytest"], "enclosing_function": "test_get_spool_default_weights", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_measure.py::49": {"resolved_imports": [], "used_names": ["Any", "URL", "datetime", "httpx", "pytest", "timezone"], "enclosing_function": "test_measure_spool", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_measure.py::44": {"resolved_imports": [], "used_names": ["Any", "URL", "datetime", "httpx", "pytest", "timezone"], "enclosing_function": "test_measure_spool", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_measure.py::46": {"resolved_imports": [], "used_names": ["Any", "URL", "datetime", "httpx", "pytest", "timezone"], "enclosing_function": "test_measure_spool", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_measure.py::98": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx", "pytest"], "enclosing_function": "test_measure_spool_higher_initial", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_measure.py::97": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx", "pytest"], "enclosing_function": "test_measure_spool_higher_initial", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_update.py::175": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx"], "enclosing_function": "test_update_spool_both_used_and_remaining_weight", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_update.py::188": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx"], "enclosing_function": "test_update_spool_not_found", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_update.py::74": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx", "length_from_weight", "pytest"], "enclosing_function": "test_update_spool", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_update.py::71": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx", "length_from_weight", "pytest"], "enclosing_function": "test_update_spool", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_update.py::72": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx", "length_from_weight", "pytest"], "enclosing_function": "test_update_spool", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_update.py::190": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx"], "enclosing_function": "test_update_spool_not_found", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_update.py::70": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx", "length_from_weight", "pytest"], "enclosing_function": "test_update_spool", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_update.py::73": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx", "length_from_weight", "pytest"], "enclosing_function": "test_update_spool", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_use.py::47": {"resolved_imports": [], "used_names": ["Any", "URL", "datetime", "httpx", "pytest", "timezone"], "enclosing_function": "test_use_spool_weight", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_use.py::121": {"resolved_imports": [], "used_names": ["Any", "URL", "httpx"], "enclosing_function": "test_use_spool_weight_and_length", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_use.py::134": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_use_spool_not_found", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_use.py::136": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_use_spool_not_found", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_use.py::42": {"resolved_imports": [], "used_names": ["Any", "URL", "datetime", "httpx", "pytest", "timezone"], "enclosing_function": "test_use_spool_weight", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_use.py::93": {"resolved_imports": [], "used_names": ["Any", "URL", "datetime", "httpx", "math", "pytest", "timezone"], "enclosing_function": "test_use_spool_length", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_use.py::177": {"resolved_imports": [], "used_names": ["Any", "URL", "asyncio", "httpx", "pytest"], "enclosing_function": "test_use_spool_concurrent", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/spool/test_use.py::43": {"resolved_imports": [], "used_names": ["Any", "URL", "datetime", "httpx", "pytest", "timezone"], "enclosing_function": "test_use_spool_weight", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_add.py::41": {"resolved_imports": [], "used_names": ["URL", "assert_dicts_compatible", "datetime", "httpx", "timezone"], "enclosing_function": "test_add_vendor", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_add.py::59": {"resolved_imports": [], "used_names": ["URL", "assert_dicts_compatible", "httpx"], "enclosing_function": "test_add_vendor_required", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_add.py::28": {"resolved_imports": [], "used_names": ["URL", "assert_dicts_compatible", "datetime", "httpx", "timezone"], "enclosing_function": "test_add_vendor", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_delete.py::29": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_delete_vendor", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_delete.py::40": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_delete_vendor_not_found", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_find.py::93": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_vendors_limit_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_find.py::141": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_vendors_limit_asc_offset_outside_range", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_find.py::91": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_vendors_limit_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_find.py::60": {"resolved_imports": [], "used_names": ["URL", "assert_lists_compatible", "httpx"], "enclosing_function": "test_find_all_vendors", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_find.py::71": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_vendors_sort_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_find.py::70": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_vendors_sort_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_find.py::174": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_vendors_by_name", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_find.py::187": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_vendors_by_empty_name", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_find.py::200": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_vendors_by_external_id", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_find.py::94": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_find_all_vendors_limit_asc", "extracted_code": "", "n_imports_parsed": 6, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_get.py::43": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_vendor_not_found", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_get.py::28": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_vendor", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_get.py::29": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_vendor", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_get.py::45": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_vendor_not_found", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_get.py::30": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_vendor", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_get.py::31": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_get_vendor", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_update.py::46": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_update_vendor_not_found", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_update.py::48": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_update_vendor_not_found", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_update.py::31": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_update_vendor", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_update.py::32": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_update_vendor", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_update.py::33": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_update_vendor", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests_integration/tests/vendor/test_update.py::34": {"resolved_imports": [], "used_names": ["URL", "httpx"], "enclosing_function": "test_update_vendor", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}}}
oracle_context_cache_v3/EbodShojaei__bake.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/Filimoa__open-parse.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/FinanceData__FinanceDataReader.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"repo": "FinanceData/FinanceDataReader", "n_pairs": 30, "version": "v3_compressed", "max_tokens": 6000, "contexts": {"tests/test_us.py::28": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_us_data_reader_yahoo", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_snap.py::23": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_snap_naver_finstate", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_basic.py::68": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_index", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_basic.py::170": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_stocklistings", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_krx.py::73": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_stock_data_reader", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_us.py::21": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_us_sp500_listing", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_snap.py::17": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_snap_krx_index_stocks", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_krx.py::22": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_stock_listing_kosdaq", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_krx.py::29": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_delisting_listing", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_snap.py::11": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_snap_krx_indices", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_economy.py::8": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_fred_data_reader", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_krx.py::36": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_administrative_listing", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_us.py::10": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_us_stock_listing_nasdaq", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_us.py::11": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_us_stock_listing_nasdaq", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_basic.py::13": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_daily", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_snap.py::30": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_snap_ecos_keystat", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_international.py::26": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_market_listing_hose", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_us.py::16": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_us_stock_listing_nyse", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_krx.py::13": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_stock_listing", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_krx.py::21": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_stock_listing_kosdaq", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_international.py::32": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_data_reader_international", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_international.py::20": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_market_listing_tse", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_krx.py::14": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_stock_listing", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_international.py::8": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_market_listing_sse", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_investing.py::13": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_investing_data_reader", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_economy.py::9": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_fred_data_reader", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_international.py::14": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_market_listing_hkex", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_basic.py::42": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_major_index", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_krx.py::42": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_krx_data_reader_basics", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_basic.py::164": {"resolved_imports": ["src/FinanceDataReader/__init__.py"], "used_names": ["pytest"], "enclosing_function": "test_stocklistings", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}}}
oracle_context_cache_v3/Forethought-Technologies__AutoChain.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/GitGuardian__ggshield.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/IDSIA__sacred.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/JWock82__Pynite.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/JoshuaC215__agent-service-toolkit.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/JuanBindez__pytubefix.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/Kludex__mangum.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/Lancetnik__Propan.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/LonamiWebs__Telethon.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/LuteOrg__lute-v3.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/Lux-Luna__LunaVox.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/MarshalX__atproto.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/MartenBE__mkslides.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"repo": "MartenBE/mkslides", "n_pairs": 67, "version": "v3_compressed", "max_tokens": 6000, "contexts": {"tests/utils.py::41": {"resolved_imports": [], "used_names": ["Path", "subprocess"], "enclosing_function": "__run_build_generic", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/utils.py::76": {"resolved_imports": [], "used_names": ["Path"], "enclosing_function": "assert_html_contains", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/baseline/test_baseline.py::31": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_process_directory_without_config", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/baseline/test_baseline.py::35": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_process_directory_without_config", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/baseline/test_baseline.py::32": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_process_directory_without_config", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/baseline/test_baseline.py::37": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_process_directory_without_config", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/baseline/test_baseline.py::36": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_process_directory_without_config", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/favicons/test_favicons.py::67": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_local_slideshow_favicon_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/favicons/test_favicons.py::45": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_local_index_favicon_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/favicons/test_favicons.py::62": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_local_slideshow_favicon_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/favicons/test_favicons.py::72": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_local_slideshow_favicon_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/favicons/test_favicons.py::16": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_absolute_url_index_favicon_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/frontmatter/test_frontmatter.py::162": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_frontmatter_paths_are_relative_to_mdfile", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/frontmatter/test_frontmatter.py::136": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_frontmatter_paths_are_relative_to_mdfile", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/frontmatter/test_frontmatter.py::166": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_frontmatter_paths_are_relative_to_mdfile", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/frontmatter/test_frontmatter.py::149": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_frontmatter_paths_are_relative_to_mdfile", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/frontmatter/test_frontmatter.py::21": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_frontmatter_overrides_default", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/frontmatter/test_frontmatter.py::140": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_frontmatter_paths_are_relative_to_mdfile", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/frontmatter/test_frontmatter.py::29": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_frontmatter_overrides_default", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/frontmatter/test_frontmatter.py::26": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_frontmatter_overrides_default", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/frontmatter/test_frontmatter.py::35": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_frontmatter_overrides_default", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/frontmatter/test_frontmatter.py::31": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_frontmatter_overrides_default", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/images/test_images.py::21": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "run_build_strict"], "enclosing_function": "test_process_directory_without_config", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/images/test_images.py::18": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "run_build_strict"], "enclosing_function": "test_process_directory_without_config", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/images/test_images.py::19": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "run_build_strict"], "enclosing_function": "test_process_directory_without_config", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/images/test_images.py::20": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "run_build_strict"], "enclosing_function": "test_process_directory_without_config", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/index/test_index.py::21": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_index_title", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/index/test_index.py::20": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_index_title", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/index/test_index.py::32": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_index_banner", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/markdown_data_options/test_markdown_data_options.py::19": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_revealjs_markdown_data_options", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/navtree/test_navtree.py::91": {"resolved_imports": ["src/mkslides/config.py", "src/mkslides/markupgenerator.py", "src/mkslides/navtree.py"], "used_names": ["Any", "DeepDiff", "MarkupGenerator", "NavTree", "get_config", "json"], "enclosing_function": "test_navtree_from_md_files", "extracted_code": "# Source: src/mkslides/config.py\ndef get_config(config_file: Path | None = None) -> DictConfig:\n config = OmegaConf.structured(Config)\n\n if not config_file and DEFAULT_CONFIG_LOCATION.exists():\n config_file = DEFAULT_CONFIG_LOCATION.resolve(strict=True).absolute()\n\n config.internal.config_path = config_file\n if config_file:\n try:\n loaded_config = OmegaConf.load(config_file)\n config = OmegaConf.merge(config, loaded_config)\n\n logger.info(f\"Loaded config from '{config_file}'\")\n except Exception:\n logger.exception(f\"Failed to load config from {config_file}\")\n raise\n\n assert OmegaConf.is_dict(config)\n\n logger.debug(f\"Used config:\\n\\n{OmegaConf.to_yaml(config, resolve=True)}\")\n\n return config\n\n\n# Source: src/mkslides/markupgenerator.py\nclass MarkupGenerator:\n def __init__(\n self,\n global_config: DictConfig,\n md_root_path: Path,\n output_directory_path: Path,\n strict: bool,\n ) -> None:\n self.global_config = global_config\n self.md_root_path = md_root_path.resolve(strict=True)\n self.output_directory_path = output_directory_path.resolve(strict=False)\n logger.info(\n f\"Output directory: '{self.output_directory_path.absolute()}'\",\n )\n\n self.output_assets_path = self.output_directory_path / OUTPUT_ASSETS_DIRNAME\n self.output_revealjs_path = self.output_assets_path / \"reveal-js\"\n self.output_highlightjs_themes_path = (\n self.output_assets_path / \"highlight-js-themes\"\n )\n\n self.strict = strict\n\n def process_markdown(self) -> None:\n \"\"\"Process the markdown files and generate HTML slideshows.\"\"\"\n logger.debug(\"Processing markdown\")\n start_time = time.perf_counter()\n\n self.__create_or_clear_output_directory()\n\n if self.md_root_path.is_file():\n assert self.md_root_path.suffix == \".md\", (\n \"md_root_path must be a markdown file\"\n )\n self.__process_markdown_file()\n else:\n self.__process_markdown_directory()\n\n end_time = time.perf_counter()\n logger.info(\n f\"Finished processing markdown in {end_time - start_time:.2f} seconds\",\n )\n\n def __create_or_clear_output_directory(self) -> None:\n \"\"\"Clear or create the output directory and copy reveal.js.\"\"\"\n if self.output_directory_path.exists():\n shutil.rmtree(self.output_directory_path)\n logger.debug(\"Output directory already exists, deleted\")\n\n self.output_directory_path.mkdir(parents=True, exist_ok=True)\n logger.debug(\"Output directory created\")\n\n with resources.as_file(REVEALJS_RESOURCE) as revealjs_path:\n self.__copy(revealjs_path, self.output_revealjs_path)\n\n with resources.as_file(HIGHLIGHTJS_THEMES_RESOURCE) as highlightjs_themes_path:\n self.__copy(highlightjs_themes_path, self.output_highlightjs_themes_path)\n\n def scan_files(self) -> tuple[list[MdFileToProcess], list[Path]]:\n \"\"\"Scan the markdown directory for markdown files and other files.\"\"\"\n md_files: list[MdFileToProcess] = []\n non_md_files: list[Path] = []\n\n for file in self.md_root_path.rglob(\"*\"):\n if file.is_file():\n resolved_file = file.resolve(strict=True)\n if resolved_file.suffix.lower() == \".md\":\n destination_path = (\n self.output_directory_path\n / resolved_file.relative_to(self.md_root_path).with_suffix(\n \".html\",\n )\n )\n\n md_files.append(\n self.__create_md_file_to_process(\n resolved_file,\n destination_path,\n ),\n )\n\n else:\n non_md_files.append(resolved_file)\n\n return md_files, non_md_files\n\n def __create_md_file_to_process(\n self,\n source_path: Path,\n destination_path: Path,\n ) -> MdFileToProcess:\n \"\"\"Create an MdFileToProcess instance from a markdown file.\"\"\"\n content = source_path.read_text(encoding=\"utf-8-sig\")\n frontmatter_metadata, markdown_content = frontmatter.parse(content)\n\n slide_config = self.__generate_slide_config(\n source_path,\n destination_path,\n frontmatter_metadata,\n )\n assert slide_config\n\n markdown_content = emojize(markdown_content, language=\"alias\")\n\n if preprocess_script := slide_config.slides.preprocess_script:\n preprocess_function = load_preprocessing_function(preprocess_script)\n if not preprocess_function:\n msg = (\n f\"Preprocessing function '{preprocess_script}' could not be loaded\"\n )\n raise ImportError(msg)\n markdown_content = preprocess_function(markdown_content)\n logger.debug(\n f\"Applied preprocessing function '{preprocess_script}' to markdown content of '{source_path}'\",\n )\n\n return MdFileToProcess(\n source_path=source_path,\n destination_path=destination_path,\n slide_config=slide_config,\n markdown_content=markdown_content,\n )\n\n def __process_markdown_file(self) -> None:\n \"\"\"Process the detected markdown file.\"\"\"\n absolute_input_path = self.md_root_path.absolute()\n logger.debug(f\"Processing markdown file at '{absolute_input_path}'\")\n logger.warning(\n f\"When you use a single file like '{absolute_input_path}' as `PATH`, only default static assets will be copied to the output folder. If you want to include images or other files, create a folder instead and pass that as `PATH`. Using a file as `PATH` is more meant for a quick slideshow in a pinch using only text.\",\n )\n\n destination_path = self.output_directory_path / \"index.html\"\n md_file_data = self.__create_md_file_to_process(\n self.md_root_path,\n destination_path,\n )\n\n self.__process_detected_markdown_files([md_file_data])\n\n def __process_markdown_directory(self) -> None:\n \"\"\"Process the detected markdown files in a directory.\"\"\"\n logger.debug(\n f\"Processing markdown directory at '{self.md_root_path.absolute()}'\",\n )\n\n md_files, non_md_files = self.scan_files()\n\n self.__process_detected_markdown_files(md_files, non_md_files)\n\n def __process_detected_markdown_files(\n self,\n md_files: list,\n non_md_files: list | None = None,\n ) -> None:\n \"\"\"Process the detected markdown files and copy non-markdown files.\"\"\"\n if non_md_files:\n for file in non_md_files:\n destination_path = self.output_directory_path / file.relative_to(\n self.md_root_path,\n )\n self.__copy(file, destination_path)\n\n self.__handle_relative_links(md_files)\n\n templates = self.__load_templates(md_files)\n\n if len(md_files) == 1:\n md_files[0].destination_path = self.output_directory_path / \"index.html\"\n else:\n self.__generate_index(md_files)\n\n self.__render_slideshows(md_files, templates)\n\n def __render_slideshows(\n self,\n md_files: list[MdFileToProcess],\n templates: dict[str, Template],\n ) -> None:\n \"\"\"Render all markdown files to HTML slideshows.\"\"\"\n for md_file_data in md_files:\n slide_config = md_file_data.slide_config\n\n slideshow_template = None\n if template_config := slide_config.slides.template:\n slideshow_template = templates[template_config]\n else:\n slideshow_template = DEFAULT_SLIDESHOW_TEMPLATE\n\n revealjs_path = self.output_revealjs_path.relative_to(\n md_file_data.destination_path.parent,\n walk_up=True,\n )\n\n # https://revealjs.com/markdown/#external-markdown\n markdown_data_options = {\n key: value\n for key, value in {\n \"data-separator\": slide_config.slides.separator,\n \"data-separator-vertical\": slide_config.slides.separator_vertical,\n \"data-separator-notes\": slide_config.slides.separator_notes,\n \"data-charset\": slide_config.slides.charset,\n }.items()\n if value\n }\n\n markup = slideshow_template.render(\n favicon=slide_config.slides.favicon,\n theme=slide_config.slides.theme,\n highlight_theme=slide_config.slides.highlight_theme,\n revealjs_path=revealjs_path,\n markdown_data_options=markdown_data_options,\n markdown=md_file_data.markdown_content,\n revealjs_config=OmegaConf.to_container(slide_config.revealjs),\n plugins=slide_config.plugins,\n )\n\n self.__create_or_overwrite_file(\n md_file_data.destination_path,\n markup,\n )\n\n def __load_templates(\n self,\n md_files: list[MdFileToProcess],\n ) -> dict[str, Template]:\n \"\"\"Load Jinja2 templates from the markdown files.\"\"\"\n templates: dict[str, Template] = {}\n\n for md_file_data in md_files:\n template = md_file_data.slide_config.slides.template\n if template and template not in templates:\n templates[template] = LOCAL_JINJA2_ENVIRONMENT.get_template(template)\n logger.debug(f\"Loaded custom template '{template}'\")\n\n return templates\n\n def __generate_theme_url(\n self,\n destination_path: Path,\n slide_config: DictConfig,\n frontmatter_metadata: dict[str, object],\n ) -> str | None:\n \"\"\"Generate the reveal.js theme URL.\"\"\"\n theme = slide_config.slides.theme\n\n if theme is None:\n return None\n\n if theme in REVEALJS_THEMES_LIST:\n return str(\n (\n self.output_revealjs_path / \"dist\" / \"theme\" / f\"{theme}.css\"\n ).relative_to(destination_path.parent, walk_up=True),\n )\n\n if get_url_type(theme) != URLType.RELATIVE or (\n \"slides\" in frontmatter_metadata\n and isinstance(frontmatter_metadata[\"slides\"], dict)\n and frontmatter_metadata[\"slides\"].get(\"theme\")\n ):\n return theme\n\n return str(\n (self.output_directory_path / theme).relative_to(\n destination_path.parent,\n walk_up=True,\n ),\n )\n\n def __generate_highlight_theme_url(\n self,\n destination_path: Path,\n slide_config: DictConfig,\n frontmatter_metadata: dict[str, object],\n ) -> str | None:\n \"\"\"Generate the highlight.js theme URL.\"\"\"\n highlight_theme = slide_config.slides.highlight_theme\n\n if highlight_theme is None:\n return None\n\n if highlight_theme in HIGHLIGHTJS_THEMES_LIST:\n return str(\n (\n self.output_highlightjs_themes_path / f\"{highlight_theme}.css\"\n ).relative_to(destination_path.parent, walk_up=True),\n )\n\n if get_url_type(highlight_theme) != URLType.RELATIVE or (\n \"slides\" in frontmatter_metadata\n and isinstance(frontmatter_metadata[\"slides\"], dict)\n and frontmatter_metadata[\"slides\"].get(\"highlight_theme\")\n ):\n return highlight_theme\n\n return str(\n (self.output_directory_path / highlight_theme).relative_to(\n destination_path.parent,\n walk_up=True,\n ),\n )\n\n def __generate_favicon_url(\n self,\n destination_path: Path,\n slide_config: DictConfig,\n frontmatter_metadata: dict[str, object],\n ) -> str | None:\n favicon = slide_config.slides.favicon\n\n if favicon is None:\n return None\n\n if get_url_type(favicon) != URLType.RELATIVE or (\n \"slides\" in frontmatter_metadata\n and isinstance(frontmatter_metadata[\"slides\"], dict)\n and frontmatter_metadata[\"slides\"].get(\"favicon\")\n ):\n return favicon\n\n return str(\n (self.output_directory_path / favicon).relative_to(\n destination_path.parent,\n walk_up=True,\n ),\n )\n\n def __generate_preprocess_script_absolute_path(\n self,\n source_path: Path,\n slide_config: DictConfig,\n frontmatter_metadata: dict[str, object],\n ) -> str | None:\n \"\"\"Generate the absolute path for the preprocess script if it is a relative URL.\"\"\"\n preprocess_script = slide_config.slides.preprocess_script\n\n if slide_config.slides.preprocess_script is None:\n return None\n\n if get_url_type(preprocess_script) != URLType.RELATIVE:\n return preprocess_script\n\n if (\n \"slides\" in frontmatter_metadata\n and isinstance(frontmatter_metadata[\"slides\"], dict)\n and frontmatter_metadata[\"slides\"].get(\"preprocess_script\")\n ):\n return str(\n (source_path.parent / preprocess_script).resolve(strict=True),\n )\n\n return str(\n (\n self.global_config.internal.config_path.parent / preprocess_script\n ).resolve(strict=True),\n )\n\n def __generate_slide_config(\n self,\n source_path: Path,\n destination_path: Path,\n frontmatter_metadata: dict[str, object],\n ) -> DictConfig:\n \"\"\"Generate the slide configuration by merging the metadata retrieved from the frontmatter of the markdown and the global configuration.\"\"\"\n slide_config: DictConfig = deepcopy(self.global_config)\n\n if frontmatter_metadata:\n for key in FRONTMATTER_ALLOWED_KEYS:\n if key in frontmatter_metadata:\n OmegaConf.update(slide_config, key, frontmatter_metadata[key])\n\n slide_config.slides.theme = self.__generate_theme_url(\n destination_path,\n slide_config,\n frontmatter_metadata,\n )\n\n slide_config.slides.highlight_theme = self.__generate_highlight_theme_url(\n destination_path,\n slide_config,\n frontmatter_metadata,\n )\n\n slide_config.slides.favicon = self.__generate_favicon_url(\n destination_path,\n slide_config,\n frontmatter_metadata,\n )\n\n slide_config.slides.preprocess_script = (\n self.__generate_preprocess_script_absolute_path(\n source_path,\n slide_config,\n frontmatter_metadata,\n )\n )\n\n return slide_config\n\n def __generate_index(self, md_files: list[MdFileToProcess]) -> None:\n \"\"\"Generate an index.html file in the output directory.\"\"\"\n logger.debug(\"Generating index\")\n\n navtree = NavTree(self.md_root_path, self.output_directory_path)\n if self.global_config.index.nav:\n nav_from_config = OmegaConf.to_container(self.global_config.index.nav)\n assert isinstance(nav_from_config, list), \"nav must be a list\"\n logger.debug(\"Generating navigation tree from config\")\n navtree.from_config_json(nav_from_config)\n navtree.validate_with_md_files(md_files, strict=self.strict)\n else:\n logger.debug(\"Generating navigation tree from markdown files\")\n navtree.from_md_files(md_files)\n\n logger.debug(\n f\"Generated navigation tree with input root path {navtree.input_root_path.absolute()} and output root path {navtree.output_root_path.absolute()}\",\n )\n\n if logger.isEnabledFor(logging.DEBUG):\n navtree_json = json.dumps(json.loads(navtree.to_json()), indent=4)\n logger.debug(f\"Navigation tree:\\n\\n{navtree_json}\\n\")\n\n # Refresh the templates here, so they have effect when live reloading\n index_template = None\n if template_config := self.global_config.index.template:\n index_template = LOCAL_JINJA2_ENVIRONMENT.get_template(template_config)\n else:\n index_template = DEFAULT_INDEX_TEMPLATE\n\n content = index_template.render(\n favicon=self.global_config.index.favicon,\n title=self.global_config.index.title,\n theme=self.global_config.index.theme,\n navtree=navtree,\n build_datetime=datetime.datetime.now(tz=datetime.UTC),\n enable_footer=self.global_config.index.enable_footer,\n )\n self.__create_or_overwrite_file(\n self.output_directory_path / \"index.html\",\n content,\n )\n\n def __create_or_overwrite_file(self, destination_path: Path, content: Any) -> None:\n \"\"\"Create or overwrite a file with the given content.\"\"\"\n is_overwrite = destination_path.exists()\n\n destination_path.parent.mkdir(parents=True, exist_ok=True)\n destination_path.write_text(content, encoding=\"utf-8\")\n\n action = \"Overwritten\" if is_overwrite else \"Created\"\n logger.debug(f\"{action} file '{destination_path}'\")\n\n def __copy(self, source_path: Path, destination_path: Path) -> None:\n \"\"\"Copy a file or directory from the source path to the destination path.\"\"\"\n is_overwrite = destination_path.exists()\n is_directory = source_path.is_dir()\n\n destination_path.parent.mkdir(parents=True, exist_ok=True)\n\n if is_directory:\n shutil.copytree(source_path, destination_path, dirs_exist_ok=True)\n else:\n shutil.copy(source_path, destination_path)\n\n action = \"Overwritten\" if is_overwrite else \"Copied\"\n file_or_directory = \"directory\" if is_directory else \"file\"\n logger.debug(\n f\"{action} {file_or_directory} '{source_path.absolute()}' to '{destination_path.absolute()}'\",\n )\n\n def __handle_relative_links(\n self,\n md_file_data: list[MdFileToProcess],\n ) -> None:\n \"\"\"Check if all relative link targets are present and normalize .md links.\"\"\"\n for md_file in md_file_data:\n content = md_file.markdown_content\n\n for link in self.__find_all_relative_links(content):\n link_path = md_file.source_path.parent / link\n relative_source_path = md_file.source_path.relative_to(\n self.md_root_path,\n )\n\n if not link_path.exists():\n msg = f\"File '{relative_source_path}' contains a link '{link}', but the target is not found among slide files.\"\n if self.strict:\n raise FileNotFoundError(msg)\n logger.warning(msg)\n elif link.lower().endswith(\".md\"):\n content = self.__replace_md_link_target(content, link)\n\n md_file.markdown_content = content\n\n def __find_all_relative_links(self, markdown_content: str) -> set[str]:\n \"\"\"Find all relative links in the given markdown content.\"\"\"\n html_content = markdown.markdown(markdown_content, extensions=[\"extra\"])\n soup = BeautifulSoup(html_content, \"html.parser\")\n\n found_links = set()\n\n for link in soup.find_all(\"a\", href=True):\n if not link.find_parents([\"code\", \"pre\"]):\n found_links.add(link[\"href\"])\n\n for link in soup.find_all(\"img\", src=True):\n if not link.find_parents([\"code\", \"pre\"]):\n found_links.add(link[\"src\"])\n\n for link in soup.find_all(\"source\", src=True):\n if not link.find_parents([\"code\", \"pre\"]):\n found_links.add(link[\"src\"])\n\n for comment in soup.find_all(string=lambda text: isinstance(text, Comment)):\n if match := HTML_BACKGROUND_IMAGE_REGEX.search(comment):\n found_links.add(match.group(\"location\"))\n\n relative_links = {\n link for link in found_links if get_url_type(link) == URLType.RELATIVE\n }\n\n return relative_links\n\n def __replace_md_link_target(self, content: str, link: str) -> str:\n \"\"\"Replace a specific .md link target with .html in markdown and HTML links.\"\"\"\n\n def _replacer(match: re.Match, *, link: str) -> str:\n matched_location = match.group(\"location\")\n\n # Only touch matches that correspond exactly to this link\n if matched_location != link:\n return match.group(0)\n\n new_location = MD_EXTENSION_REGEX.sub(\".html\", matched_location)\n return match.group(0).replace(matched_location, new_location)\n\n for regex in (MD_RELATIVE_LINK_REGEX, HTML_RELATIVE_LINK_REGEX):\n bound_replacer = partial(_replacer, link=link)\n content = regex.sub(bound_replacer, content)\n\n return content\n\n\n# Source: src/mkslides/navtree.py\nclass NavTree:\n def __init__(self, input_root_path: Path, output_root_path: Path) -> None:\n self.input_root_path = input_root_path\n self.output_root_path = output_root_path\n\n # Relative path as str is the index, title as str the data.\n self.tree = Tree()\n self.tree.create_node(identifier=\"root\")\n\n def from_md_files(self, md_files: list[MdFileToProcess]) -> None:\n for md_file in md_files:\n relative_source_path = md_file.source_path.relative_to(\n self.input_root_path,\n )\n parts = relative_source_path.parts\n\n current_relative_source_path = Path()\n parent_node_id = str(self.tree.root)\n for part in parts:\n current_relative_source_path /= part\n\n node_id = None\n if (self.input_root_path / current_relative_source_path).is_dir():\n node_id = str(current_relative_source_path)\n else:\n node_id = str(current_relative_source_path.with_suffix(\".html\"))\n\n node_data = None\n if md_file.slide_config.slides.title:\n node_data = md_file.slide_config.slides.title\n else:\n node_data = current_relative_source_path.stem\n\n if node_id not in self.tree:\n self.tree.create_node(\n identifier=node_id,\n parent=parent_node_id,\n data=node_data,\n )\n\n parent_node_id = node_id\n\n def from_config_json(self, json_data: list) -> None:\n assert isinstance(json_data, list), \"json data must be a list\"\n\n for item in json_data:\n self.__node_from_config_json(\n item,\n self.output_root_path,\n str(self.tree.root),\n )\n\n def __node_from_config_json(\n self,\n json_data: dict | str,\n current_path: Path,\n parent_node_id: str,\n ) -> None:\n # leaf node\n #\n # - filename.md\n #\n if isinstance(json_data, str):\n destination_path = (current_path / json_data).with_suffix(\".html\")\n node_id = str(destination_path.relative_to(self.output_root_path))\n node_data = destination_path.stem\n\n self.tree.create_node(\n identifier=node_id,\n parent=parent_node_id,\n data=node_data,\n )\n\n # category or leaf node with custom file name\n elif isinstance(json_data, dict):\n assert len(json_data.keys()) == 1, \"json dict must have one key\"\n\n title, content = next(iter(json_data.items()))\n\n # leaf node with custom name\n #\n # - custom-file-name: filename.md\n #\n if isinstance(content, str):\n destination_path = (current_path / content).with_suffix(\".html\")\n node_id = str(destination_path.relative_to(self.output_root_path))\n node_data = title\n\n self.tree.create_node(\n identifier=node_id,\n parent=parent_node_id,\n data=node_data,\n )\n\n # category node\n #\n # - category:\n # - ...\n #\n elif isinstance(content, list):\n destination_path = current_path / title\n node_id = str(f\"{destination_path.relative_to(self.output_root_path)}\")\n node_data = title\n\n self.tree.create_node(\n identifier=node_id,\n parent=parent_node_id,\n data=node_data,\n )\n\n for item in content:\n self.__node_from_config_json(item, destination_path, node_id)\n\n else:\n msg = f\"json dict must have a string or list as value, but value is of {type(content)}\"\n raise TypeError(msg)\n\n else:\n msg = (\n f\"json data must be a string or dict, but is of type {type(json_data)}\"\n )\n\n raise TypeError(msg)\n\n def is_node_leaf(self, node_id: str) -> bool:\n return self.tree[node_id].is_leaf(self.tree.identifier)\n\n def get_node_children(self, node_id: str) -> list:\n return sorted(self.tree.children(node_id), key=lambda n: n.identifier)\n\n def to_json(self) -> str:\n if not self.tree:\n return \"{}\"\n\n return self.tree.to_json(with_data=True)\n\n def validate_with_md_files(\n self,\n md_files: list[MdFileToProcess],\n strict: bool,\n ) -> None:\n md_file_relative_destination_paths = [\n str(md_file.destination_path.relative_to(self.output_root_path))\n for md_file in md_files\n ]\n\n files_not_in_navtree = []\n for md_file_relative_destination_path in md_file_relative_destination_paths:\n if md_file_relative_destination_path not in self.tree:\n source_file_name = str(\n Path(md_file_relative_destination_path).with_suffix(\".md\"),\n )\n files_not_in_navtree.append(source_file_name)\n\n if files_not_in_navtree:\n logger.info(\n \"The following pages exist in the slides directory, but are not included in the 'nav' configuration:\",\n )\n\n for file_name in files_not_in_navtree:\n logger.info(f\"\\t- {file_name}\")\n\n for node_id in self.tree.expand_tree():\n node = self.tree.get_node(node_id)\n assert node\n if (\n node.is_leaf(self.tree.identifier)\n and node.identifier not in md_file_relative_destination_paths\n ):\n source_file_name = Path(node.identifier).with_suffix(\".md\").name\n msg = f\"A reference to '{source_file_name}' is included in the 'nav' configuration, which is not found in the slideshow files.\"\n if strict:\n raise FileNotFoundError(msg)\n logger.warning(msg)", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 27622, "extracted_code_full": "# Source: src/mkslides/config.py\ndef get_config(config_file: Path | None = None) -> DictConfig:\n config = OmegaConf.structured(Config)\n\n if not config_file and DEFAULT_CONFIG_LOCATION.exists():\n config_file = DEFAULT_CONFIG_LOCATION.resolve(strict=True).absolute()\n\n config.internal.config_path = config_file\n if config_file:\n try:\n loaded_config = OmegaConf.load(config_file)\n config = OmegaConf.merge(config, loaded_config)\n\n logger.info(f\"Loaded config from '{config_file}'\")\n except Exception:\n logger.exception(f\"Failed to load config from {config_file}\")\n raise\n\n assert OmegaConf.is_dict(config)\n\n logger.debug(f\"Used config:\\n\\n{OmegaConf.to_yaml(config, resolve=True)}\")\n\n return config\n\n\n# Source: src/mkslides/markupgenerator.py\nclass MarkupGenerator:\n def __init__(\n self,\n global_config: DictConfig,\n md_root_path: Path,\n output_directory_path: Path,\n strict: bool,\n ) -> None:\n self.global_config = global_config\n self.md_root_path = md_root_path.resolve(strict=True)\n self.output_directory_path = output_directory_path.resolve(strict=False)\n logger.info(\n f\"Output directory: '{self.output_directory_path.absolute()}'\",\n )\n\n self.output_assets_path = self.output_directory_path / OUTPUT_ASSETS_DIRNAME\n self.output_revealjs_path = self.output_assets_path / \"reveal-js\"\n self.output_highlightjs_themes_path = (\n self.output_assets_path / \"highlight-js-themes\"\n )\n\n self.strict = strict\n\n def process_markdown(self) -> None:\n \"\"\"Process the markdown files and generate HTML slideshows.\"\"\"\n logger.debug(\"Processing markdown\")\n start_time = time.perf_counter()\n\n self.__create_or_clear_output_directory()\n\n if self.md_root_path.is_file():\n assert self.md_root_path.suffix == \".md\", (\n \"md_root_path must be a markdown file\"\n )\n self.__process_markdown_file()\n else:\n self.__process_markdown_directory()\n\n end_time = time.perf_counter()\n logger.info(\n f\"Finished processing markdown in {end_time - start_time:.2f} seconds\",\n )\n\n def __create_or_clear_output_directory(self) -> None:\n \"\"\"Clear or create the output directory and copy reveal.js.\"\"\"\n if self.output_directory_path.exists():\n shutil.rmtree(self.output_directory_path)\n logger.debug(\"Output directory already exists, deleted\")\n\n self.output_directory_path.mkdir(parents=True, exist_ok=True)\n logger.debug(\"Output directory created\")\n\n with resources.as_file(REVEALJS_RESOURCE) as revealjs_path:\n self.__copy(revealjs_path, self.output_revealjs_path)\n\n with resources.as_file(HIGHLIGHTJS_THEMES_RESOURCE) as highlightjs_themes_path:\n self.__copy(highlightjs_themes_path, self.output_highlightjs_themes_path)\n\n def scan_files(self) -> tuple[list[MdFileToProcess], list[Path]]:\n \"\"\"Scan the markdown directory for markdown files and other files.\"\"\"\n md_files: list[MdFileToProcess] = []\n non_md_files: list[Path] = []\n\n for file in self.md_root_path.rglob(\"*\"):\n if file.is_file():\n resolved_file = file.resolve(strict=True)\n if resolved_file.suffix.lower() == \".md\":\n destination_path = (\n self.output_directory_path\n / resolved_file.relative_to(self.md_root_path).with_suffix(\n \".html\",\n )\n )\n\n md_files.append(\n self.__create_md_file_to_process(\n resolved_file,\n destination_path,\n ),\n )\n\n else:\n non_md_files.append(resolved_file)\n\n return md_files, non_md_files\n\n def __create_md_file_to_process(\n self,\n source_path: Path,\n destination_path: Path,\n ) -> MdFileToProcess:\n \"\"\"Create an MdFileToProcess instance from a markdown file.\"\"\"\n content = source_path.read_text(encoding=\"utf-8-sig\")\n frontmatter_metadata, markdown_content = frontmatter.parse(content)\n\n slide_config = self.__generate_slide_config(\n source_path,\n destination_path,\n frontmatter_metadata,\n )\n assert slide_config\n\n markdown_content = emojize(markdown_content, language=\"alias\")\n\n if preprocess_script := slide_config.slides.preprocess_script:\n preprocess_function = load_preprocessing_function(preprocess_script)\n if not preprocess_function:\n msg = (\n f\"Preprocessing function '{preprocess_script}' could not be loaded\"\n )\n raise ImportError(msg)\n markdown_content = preprocess_function(markdown_content)\n logger.debug(\n f\"Applied preprocessing function '{preprocess_script}' to markdown content of '{source_path}'\",\n )\n\n return MdFileToProcess(\n source_path=source_path,\n destination_path=destination_path,\n slide_config=slide_config,\n markdown_content=markdown_content,\n )\n\n def __process_markdown_file(self) -> None:\n \"\"\"Process the detected markdown file.\"\"\"\n absolute_input_path = self.md_root_path.absolute()\n logger.debug(f\"Processing markdown file at '{absolute_input_path}'\")\n logger.warning(\n f\"When you use a single file like '{absolute_input_path}' as `PATH`, only default static assets will be copied to the output folder. If you want to include images or other files, create a folder instead and pass that as `PATH`. Using a file as `PATH` is more meant for a quick slideshow in a pinch using only text.\",\n )\n\n destination_path = self.output_directory_path / \"index.html\"\n md_file_data = self.__create_md_file_to_process(\n self.md_root_path,\n destination_path,\n )\n\n self.__process_detected_markdown_files([md_file_data])\n\n def __process_markdown_directory(self) -> None:\n \"\"\"Process the detected markdown files in a directory.\"\"\"\n logger.debug(\n f\"Processing markdown directory at '{self.md_root_path.absolute()}'\",\n )\n\n md_files, non_md_files = self.scan_files()\n\n self.__process_detected_markdown_files(md_files, non_md_files)\n\n def __process_detected_markdown_files(\n self,\n md_files: list,\n non_md_files: list | None = None,\n ) -> None:\n \"\"\"Process the detected markdown files and copy non-markdown files.\"\"\"\n if non_md_files:\n for file in non_md_files:\n destination_path = self.output_directory_path / file.relative_to(\n self.md_root_path,\n )\n self.__copy(file, destination_path)\n\n self.__handle_relative_links(md_files)\n\n templates = self.__load_templates(md_files)\n\n if len(md_files) == 1:\n md_files[0].destination_path = self.output_directory_path / \"index.html\"\n else:\n self.__generate_index(md_files)\n\n self.__render_slideshows(md_files, templates)\n\n def __render_slideshows(\n self,\n md_files: list[MdFileToProcess],\n templates: dict[str, Template],\n ) -> None:\n \"\"\"Render all markdown files to HTML slideshows.\"\"\"\n for md_file_data in md_files:\n slide_config = md_file_data.slide_config\n\n slideshow_template = None\n if template_config := slide_config.slides.template:\n slideshow_template = templates[template_config]\n else:\n slideshow_template = DEFAULT_SLIDESHOW_TEMPLATE\n\n revealjs_path = self.output_revealjs_path.relative_to(\n md_file_data.destination_path.parent,\n walk_up=True,\n )\n\n # https://revealjs.com/markdown/#external-markdown\n markdown_data_options = {\n key: value\n for key, value in {\n \"data-separator\": slide_config.slides.separator,\n \"data-separator-vertical\": slide_config.slides.separator_vertical,\n \"data-separator-notes\": slide_config.slides.separator_notes,\n \"data-charset\": slide_config.slides.charset,\n }.items()\n if value\n }\n\n markup = slideshow_template.render(\n favicon=slide_config.slides.favicon,\n theme=slide_config.slides.theme,\n highlight_theme=slide_config.slides.highlight_theme,\n revealjs_path=revealjs_path,\n markdown_data_options=markdown_data_options,\n markdown=md_file_data.markdown_content,\n revealjs_config=OmegaConf.to_container(slide_config.revealjs),\n plugins=slide_config.plugins,\n )\n\n self.__create_or_overwrite_file(\n md_file_data.destination_path,\n markup,\n )\n\n def __load_templates(\n self,\n md_files: list[MdFileToProcess],\n ) -> dict[str, Template]:\n \"\"\"Load Jinja2 templates from the markdown files.\"\"\"\n templates: dict[str, Template] = {}\n\n for md_file_data in md_files:\n template = md_file_data.slide_config.slides.template\n if template and template not in templates:\n templates[template] = LOCAL_JINJA2_ENVIRONMENT.get_template(template)\n logger.debug(f\"Loaded custom template '{template}'\")\n\n return templates\n\n def __generate_theme_url(\n self,\n destination_path: Path,\n slide_config: DictConfig,\n frontmatter_metadata: dict[str, object],\n ) -> str | None:\n \"\"\"Generate the reveal.js theme URL.\"\"\"\n theme = slide_config.slides.theme\n\n if theme is None:\n return None\n\n if theme in REVEALJS_THEMES_LIST:\n return str(\n (\n self.output_revealjs_path / \"dist\" / \"theme\" / f\"{theme}.css\"\n ).relative_to(destination_path.parent, walk_up=True),\n )\n\n if get_url_type(theme) != URLType.RELATIVE or (\n \"slides\" in frontmatter_metadata\n and isinstance(frontmatter_metadata[\"slides\"], dict)\n and frontmatter_metadata[\"slides\"].get(\"theme\")\n ):\n return theme\n\n return str(\n (self.output_directory_path / theme).relative_to(\n destination_path.parent,\n walk_up=True,\n ),\n )\n\n def __generate_highlight_theme_url(\n self,\n destination_path: Path,\n slide_config: DictConfig,\n frontmatter_metadata: dict[str, object],\n ) -> str | None:\n \"\"\"Generate the highlight.js theme URL.\"\"\"\n highlight_theme = slide_config.slides.highlight_theme\n\n if highlight_theme is None:\n return None\n\n if highlight_theme in HIGHLIGHTJS_THEMES_LIST:\n return str(\n (\n self.output_highlightjs_themes_path / f\"{highlight_theme}.css\"\n ).relative_to(destination_path.parent, walk_up=True),\n )\n\n if get_url_type(highlight_theme) != URLType.RELATIVE or (\n \"slides\" in frontmatter_metadata\n and isinstance(frontmatter_metadata[\"slides\"], dict)\n and frontmatter_metadata[\"slides\"].get(\"highlight_theme\")\n ):\n return highlight_theme\n\n return str(\n (self.output_directory_path / highlight_theme).relative_to(\n destination_path.parent,\n walk_up=True,\n ),\n )\n\n def __generate_favicon_url(\n self,\n destination_path: Path,\n slide_config: DictConfig,\n frontmatter_metadata: dict[str, object],\n ) -> str | None:\n favicon = slide_config.slides.favicon\n\n if favicon is None:\n return None\n\n if get_url_type(favicon) != URLType.RELATIVE or (\n \"slides\" in frontmatter_metadata\n and isinstance(frontmatter_metadata[\"slides\"], dict)\n and frontmatter_metadata[\"slides\"].get(\"favicon\")\n ):\n return favicon\n\n return str(\n (self.output_directory_path / favicon).relative_to(\n destination_path.parent,\n walk_up=True,\n ),\n )\n\n def __generate_preprocess_script_absolute_path(\n self,\n source_path: Path,\n slide_config: DictConfig,\n frontmatter_metadata: dict[str, object],\n ) -> str | None:\n \"\"\"Generate the absolute path for the preprocess script if it is a relative URL.\"\"\"\n preprocess_script = slide_config.slides.preprocess_script\n\n if slide_config.slides.preprocess_script is None:\n return None\n\n if get_url_type(preprocess_script) != URLType.RELATIVE:\n return preprocess_script\n\n if (\n \"slides\" in frontmatter_metadata\n and isinstance(frontmatter_metadata[\"slides\"], dict)\n and frontmatter_metadata[\"slides\"].get(\"preprocess_script\")\n ):\n return str(\n (source_path.parent / preprocess_script).resolve(strict=True),\n )\n\n return str(\n (\n self.global_config.internal.config_path.parent / preprocess_script\n ).resolve(strict=True),\n )\n\n def __generate_slide_config(\n self,\n source_path: Path,\n destination_path: Path,\n frontmatter_metadata: dict[str, object],\n ) -> DictConfig:\n \"\"\"Generate the slide configuration by merging the metadata retrieved from the frontmatter of the markdown and the global configuration.\"\"\"\n slide_config: DictConfig = deepcopy(self.global_config)\n\n if frontmatter_metadata:\n for key in FRONTMATTER_ALLOWED_KEYS:\n if key in frontmatter_metadata:\n OmegaConf.update(slide_config, key, frontmatter_metadata[key])\n\n slide_config.slides.theme = self.__generate_theme_url(\n destination_path,\n slide_config,\n frontmatter_metadata,\n )\n\n slide_config.slides.highlight_theme = self.__generate_highlight_theme_url(\n destination_path,\n slide_config,\n frontmatter_metadata,\n )\n\n slide_config.slides.favicon = self.__generate_favicon_url(\n destination_path,\n slide_config,\n frontmatter_metadata,\n )\n\n slide_config.slides.preprocess_script = (\n self.__generate_preprocess_script_absolute_path(\n source_path,\n slide_config,\n frontmatter_metadata,\n )\n )\n\n return slide_config\n\n def __generate_index(self, md_files: list[MdFileToProcess]) -> None:\n \"\"\"Generate an index.html file in the output directory.\"\"\"\n logger.debug(\"Generating index\")\n\n navtree = NavTree(self.md_root_path, self.output_directory_path)\n if self.global_config.index.nav:\n nav_from_config = OmegaConf.to_container(self.global_config.index.nav)\n assert isinstance(nav_from_config, list), \"nav must be a list\"\n logger.debug(\"Generating navigation tree from config\")\n navtree.from_config_json(nav_from_config)\n navtree.validate_with_md_files(md_files, strict=self.strict)\n else:\n logger.debug(\"Generating navigation tree from markdown files\")\n navtree.from_md_files(md_files)\n\n logger.debug(\n f\"Generated navigation tree with input root path {navtree.input_root_path.absolute()} and output root path {navtree.output_root_path.absolute()}\",\n )\n\n if logger.isEnabledFor(logging.DEBUG):\n navtree_json = json.dumps(json.loads(navtree.to_json()), indent=4)\n logger.debug(f\"Navigation tree:\\n\\n{navtree_json}\\n\")\n\n # Refresh the templates here, so they have effect when live reloading\n index_template = None\n if template_config := self.global_config.index.template:\n index_template = LOCAL_JINJA2_ENVIRONMENT.get_template(template_config)\n else:\n index_template = DEFAULT_INDEX_TEMPLATE\n\n content = index_template.render(\n favicon=self.global_config.index.favicon,\n title=self.global_config.index.title,\n theme=self.global_config.index.theme,\n navtree=navtree,\n build_datetime=datetime.datetime.now(tz=datetime.UTC),\n enable_footer=self.global_config.index.enable_footer,\n )\n self.__create_or_overwrite_file(\n self.output_directory_path / \"index.html\",\n content,\n )\n\n def __create_or_overwrite_file(self, destination_path: Path, content: Any) -> None:\n \"\"\"Create or overwrite a file with the given content.\"\"\"\n is_overwrite = destination_path.exists()\n\n destination_path.parent.mkdir(parents=True, exist_ok=True)\n destination_path.write_text(content, encoding=\"utf-8\")\n\n action = \"Overwritten\" if is_overwrite else \"Created\"\n logger.debug(f\"{action} file '{destination_path}'\")\n\n def __copy(self, source_path: Path, destination_path: Path) -> None:\n \"\"\"Copy a file or directory from the source path to the destination path.\"\"\"\n is_overwrite = destination_path.exists()\n is_directory = source_path.is_dir()\n\n destination_path.parent.mkdir(parents=True, exist_ok=True)\n\n if is_directory:\n shutil.copytree(source_path, destination_path, dirs_exist_ok=True)\n else:\n shutil.copy(source_path, destination_path)\n\n action = \"Overwritten\" if is_overwrite else \"Copied\"\n file_or_directory = \"directory\" if is_directory else \"file\"\n logger.debug(\n f\"{action} {file_or_directory} '{source_path.absolute()}' to '{destination_path.absolute()}'\",\n )\n\n def __handle_relative_links(\n self,\n md_file_data: list[MdFileToProcess],\n ) -> None:\n \"\"\"Check if all relative link targets are present and normalize .md links.\"\"\"\n for md_file in md_file_data:\n content = md_file.markdown_content\n\n for link in self.__find_all_relative_links(content):\n link_path = md_file.source_path.parent / link\n relative_source_path = md_file.source_path.relative_to(\n self.md_root_path,\n )\n\n if not link_path.exists():\n msg = f\"File '{relative_source_path}' contains a link '{link}', but the target is not found among slide files.\"\n if self.strict:\n raise FileNotFoundError(msg)\n logger.warning(msg)\n elif link.lower().endswith(\".md\"):\n content = self.__replace_md_link_target(content, link)\n\n md_file.markdown_content = content\n\n def __find_all_relative_links(self, markdown_content: str) -> set[str]:\n \"\"\"Find all relative links in the given markdown content.\"\"\"\n html_content = markdown.markdown(markdown_content, extensions=[\"extra\"])\n soup = BeautifulSoup(html_content, \"html.parser\")\n\n found_links = set()\n\n for link in soup.find_all(\"a\", href=True):\n if not link.find_parents([\"code\", \"pre\"]):\n found_links.add(link[\"href\"])\n\n for link in soup.find_all(\"img\", src=True):\n if not link.find_parents([\"code\", \"pre\"]):\n found_links.add(link[\"src\"])\n\n for link in soup.find_all(\"source\", src=True):\n if not link.find_parents([\"code\", \"pre\"]):\n found_links.add(link[\"src\"])\n\n for comment in soup.find_all(string=lambda text: isinstance(text, Comment)):\n if match := HTML_BACKGROUND_IMAGE_REGEX.search(comment):\n found_links.add(match.group(\"location\"))\n\n relative_links = {\n link for link in found_links if get_url_type(link) == URLType.RELATIVE\n }\n\n return relative_links\n\n def __replace_md_link_target(self, content: str, link: str) -> str:\n \"\"\"Replace a specific .md link target with .html in markdown and HTML links.\"\"\"\n\n def _replacer(match: re.Match, *, link: str) -> str:\n matched_location = match.group(\"location\")\n\n # Only touch matches that correspond exactly to this link\n if matched_location != link:\n return match.group(0)\n\n new_location = MD_EXTENSION_REGEX.sub(\".html\", matched_location)\n return match.group(0).replace(matched_location, new_location)\n\n for regex in (MD_RELATIVE_LINK_REGEX, HTML_RELATIVE_LINK_REGEX):\n bound_replacer = partial(_replacer, link=link)\n content = regex.sub(bound_replacer, content)\n\n return content\n\n\n# Source: src/mkslides/navtree.py\nclass NavTree:\n def __init__(self, input_root_path: Path, output_root_path: Path) -> None:\n self.input_root_path = input_root_path\n self.output_root_path = output_root_path\n\n # Relative path as str is the index, title as str the data.\n self.tree = Tree()\n self.tree.create_node(identifier=\"root\")\n\n def from_md_files(self, md_files: list[MdFileToProcess]) -> None:\n for md_file in md_files:\n relative_source_path = md_file.source_path.relative_to(\n self.input_root_path,\n )\n parts = relative_source_path.parts\n\n current_relative_source_path = Path()\n parent_node_id = str(self.tree.root)\n for part in parts:\n current_relative_source_path /= part\n\n node_id = None\n if (self.input_root_path / current_relative_source_path).is_dir():\n node_id = str(current_relative_source_path)\n else:\n node_id = str(current_relative_source_path.with_suffix(\".html\"))\n\n node_data = None\n if md_file.slide_config.slides.title:\n node_data = md_file.slide_config.slides.title\n else:\n node_data = current_relative_source_path.stem\n\n if node_id not in self.tree:\n self.tree.create_node(\n identifier=node_id,\n parent=parent_node_id,\n data=node_data,\n )\n\n parent_node_id = node_id\n\n def from_config_json(self, json_data: list) -> None:\n assert isinstance(json_data, list), \"json data must be a list\"\n\n for item in json_data:\n self.__node_from_config_json(\n item,\n self.output_root_path,\n str(self.tree.root),\n )\n\n def __node_from_config_json(\n self,\n json_data: dict | str,\n current_path: Path,\n parent_node_id: str,\n ) -> None:\n # leaf node\n #\n # - filename.md\n #\n if isinstance(json_data, str):\n destination_path = (current_path / json_data).with_suffix(\".html\")\n node_id = str(destination_path.relative_to(self.output_root_path))\n node_data = destination_path.stem\n\n self.tree.create_node(\n identifier=node_id,\n parent=parent_node_id,\n data=node_data,\n )\n\n # category or leaf node with custom file name\n elif isinstance(json_data, dict):\n assert len(json_data.keys()) == 1, \"json dict must have one key\"\n\n title, content = next(iter(json_data.items()))\n\n # leaf node with custom name\n #\n # - custom-file-name: filename.md\n #\n if isinstance(content, str):\n destination_path = (current_path / content).with_suffix(\".html\")\n node_id = str(destination_path.relative_to(self.output_root_path))\n node_data = title\n\n self.tree.create_node(\n identifier=node_id,\n parent=parent_node_id,\n data=node_data,\n )\n\n # category node\n #\n # - category:\n # - ...\n #\n elif isinstance(content, list):\n destination_path = current_path / title\n node_id = str(f\"{destination_path.relative_to(self.output_root_path)}\")\n node_data = title\n\n self.tree.create_node(\n identifier=node_id,\n parent=parent_node_id,\n data=node_data,\n )\n\n for item in content:\n self.__node_from_config_json(item, destination_path, node_id)\n\n else:\n msg = f\"json dict must have a string or list as value, but value is of {type(content)}\"\n raise TypeError(msg)\n\n else:\n msg = (\n f\"json data must be a string or dict, but is of type {type(json_data)}\"\n )\n\n raise TypeError(msg)\n\n def is_node_leaf(self, node_id: str) -> bool:\n return self.tree[node_id].is_leaf(self.tree.identifier)\n\n def get_node_children(self, node_id: str) -> list:\n return sorted(self.tree.children(node_id), key=lambda n: n.identifier)\n\n def to_json(self) -> str:\n if not self.tree:\n return \"{}\"\n\n return self.tree.to_json(with_data=True)\n\n def validate_with_md_files(\n self,\n md_files: list[MdFileToProcess],\n strict: bool,\n ) -> None:\n md_file_relative_destination_paths = [\n str(md_file.destination_path.relative_to(self.output_root_path))\n for md_file in md_files\n ]\n\n files_not_in_navtree = []\n for md_file_relative_destination_path in md_file_relative_destination_paths:\n if md_file_relative_destination_path not in self.tree:\n source_file_name = str(\n Path(md_file_relative_destination_path).with_suffix(\".md\"),\n )\n files_not_in_navtree.append(source_file_name)\n\n if files_not_in_navtree:\n logger.info(\n \"The following pages exist in the slides directory, but are not included in the 'nav' configuration:\",\n )\n\n for file_name in files_not_in_navtree:\n logger.info(f\"\\t- {file_name}\")\n\n for node_id in self.tree.expand_tree():\n node = self.tree.get_node(node_id)\n assert node\n if (\n node.is_leaf(self.tree.identifier)\n and node.identifier not in md_file_relative_destination_paths\n ):\n source_file_name = Path(node.identifier).with_suffix(\".md\").name\n msg = f\"A reference to '{source_file_name}' is included in the 'nav' configuration, which is not found in the slideshow files.\"\n if strict:\n raise FileNotFoundError(msg)\n logger.warning(msg)", "n_chars_compressed": 27622, "compression_ratio": 1.0}, "tests/navtree/test_navtree.py::194": {"resolved_imports": ["src/mkslides/config.py", "src/mkslides/markupgenerator.py", "src/mkslides/navtree.py"], "used_names": ["Any", "re", "subprocess"], "enclosing_function": "test_files_not_in_folder_without_strict", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 3, "n_chars_extracted": 0}, "tests/plugins/test_plugins.py::25": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_plugins", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/plugins/test_plugins.py::29": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_plugins", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/plugins/test_plugins.py::21": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_plugins", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/plugins/test_plugins.py::37": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_plugins", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/plugins/test_plugins.py::33": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_plugins", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/preprocessing/test_preprocessing.py::17": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_preprocessing", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/preprocessing/test_preprocessing.py::31": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_preprocessing", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/relative_links/test_relative_links.py::43": {"resolved_imports": [], "used_names": ["Any", "re", "subprocess"], "enclosing_function": "test_non_existing_relative_links_without_strict", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/relative_slideshow_links/test_relative_slideshow_links.py::94": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_relative_slideshow_links", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/relative_slideshow_links/test_relative_slideshow_links.py::18": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_relative_slideshow_links", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/relative_slideshow_links/test_relative_slideshow_links.py::19": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_relative_slideshow_links", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/relative_slideshow_links/test_relative_slideshow_links.py::98": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_relative_slideshow_links", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/relative_slideshow_links/test_relative_slideshow_links.py::20": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_relative_slideshow_links", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/revealjs_options/test_revealjs_options.py::21": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_revealjs_default_options", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/revealjs_options/test_revealjs_options.py::53": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_revealjs_string_options", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/revealjs_options/test_revealjs_options.py::30": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains_regexp", "re", "run_build_strict"], "enclosing_function": "test_revealjs_integer_options", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/single_md_in_folder/test_single_md_in_folder.py::33": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_multiple_md_files_in_folder", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/single_md_in_folder/test_single_md_in_folder.py::20": {"resolved_imports": [], "used_names": ["Any", "assert_file_does_not_exist", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_single_md_files_in_folder", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/single_md_in_folder/test_single_md_in_folder.py::21": {"resolved_imports": [], "used_names": ["Any", "assert_file_does_not_exist", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_single_md_files_in_folder", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/single_md_in_folder/test_single_md_in_folder.py::23": {"resolved_imports": [], "used_names": ["Any", "assert_file_does_not_exist", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_single_md_files_in_folder", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/single_md_in_folder/test_single_md_in_folder.py::24": {"resolved_imports": [], "used_names": ["Any", "assert_file_does_not_exist", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_single_md_files_in_folder", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/text/test_emojis.py::23": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_emojize", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/text/test_emojis.py::22": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_emojize", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::46": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_local_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::20": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_local_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::33": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_local_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::50": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_local_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::59": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_local_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::108": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_builtin_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::79": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_absolute_url_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::83": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_absolute_url_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::75": {"resolved_imports": [], "used_names": ["Any", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_absolute_url_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::118": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_builtin_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::123": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_builtin_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}, "tests/themes/test_themes.py::132": {"resolved_imports": [], "used_names": ["Any", "assert_file_exist", "assert_html_contains", "run_build_strict"], "enclosing_function": "test_builtin_slideshow_theme_path", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 0, "n_chars_extracted": 0}}}
oracle_context_cache_v3/MasoniteFramework__masonite.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/MaxHalford__prince.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"repo": "MaxHalford/prince", "n_pairs": 42, "version": "v3_compressed", "max_tokens": 6000, "contexts": {"tests/test_ca.py::95": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 13, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_famd.py::73": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_gpa.py::47": {"resolved_imports": ["prince/__init__.py"], "used_names": ["prince"], "enclosing_function": "test_fit_bad_init", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_gpa.py::34": {"resolved_imports": ["prince/__init__.py"], "used_names": ["prince"], "enclosing_function": "test_fit", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_mca.py::83": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_col_cos2", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_mca.py::70": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_col_coords", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_pca.py::137": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_gpa.py::82": {"resolved_imports": ["prince/__init__.py"], "used_names": ["prince"], "enclosing_function": "test_copy", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_pca.py::128": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_ca.py::113": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_row_contrib", "extracted_code": "", "n_imports_parsed": 13, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_famd.py::66": {"resolved_imports": ["prince/__init__.py"], "used_names": [], "enclosing_function": "test_cat_cols", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_pca.py::165": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_row_contrib", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_mfa.py::76": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_ca.py::94": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 13, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_ca.py::108": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_row_coords", "extracted_code": "", "n_imports_parsed": 13, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_ca.py::120": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_row_cosine_similarities", "extracted_code": "", "n_imports_parsed": 13, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_pca.py::127": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_pca.py::160": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_row_cosine_similarities", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_pca.py::149": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R", "pytest"], "enclosing_function": "test_row_coords", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_gpa.py::68": {"resolved_imports": ["prince/__init__.py"], "used_names": ["prince"], "enclosing_function": "test_fit_transform_equal", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_pca.py::129": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_gpa.py::75": {"resolved_imports": ["prince/__init__.py"], "used_names": ["prince"], "enclosing_function": "test_fit_transform_single", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_famd.py::72": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_ca.py::96": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 13, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_famd.py::71": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_svd.py::64": {"resolved_imports": ["prince/__init__.py", "prince/svd.py"], "used_names": ["load_df_from_R", "svd"], "enclosing_function": "test_U", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_famd.py::82": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R", "pytest"], "enclosing_function": "test_row_coords", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_mfa.py::77": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_mfa.py::75": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_svd.py::71": {"resolved_imports": ["prince/__init__.py", "prince/svd.py"], "used_names": ["robjects", "svd"], "enclosing_function": "test_s", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_svd.py::68": {"resolved_imports": ["prince/__init__.py", "prince/svd.py"], "used_names": ["load_df_from_R", "svd"], "enclosing_function": "test_U", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_svd.py::78": {"resolved_imports": ["prince/__init__.py", "prince/svd.py"], "used_names": ["load_df_from_R", "svd"], "enclosing_function": "test_V", "extracted_code": "", "n_imports_parsed": 7, "n_files_resolved": 2, "n_chars_extracted": 0}, "tests/test_pca.py::136": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_eigenvalues", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_gpa.py::59": {"resolved_imports": ["prince/__init__.py"], "used_names": ["prince"], "enclosing_function": "test_transform", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_mca.py::199": {"resolved_imports": ["prince/__init__.py"], "used_names": ["prince"], "enclosing_function": "test_type_doesnt_matter", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_gpa.py::86": {"resolved_imports": ["prince/__init__.py"], "used_names": ["prince"], "enclosing_function": "test_copy", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_mfa.py::106": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R", "pytest"], "enclosing_function": "test_row_coords", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_famd.py::58": {"resolved_imports": ["prince/__init__.py"], "used_names": [], "enclosing_function": "test_num_cols", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_gpa.py::60": {"resolved_imports": ["prince/__init__.py"], "used_names": ["prince"], "enclosing_function": "test_transform", "extracted_code": "", "n_imports_parsed": 5, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_famd.py::87": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_row_contrib", "extracted_code": "", "n_imports_parsed": 9, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_mfa.py::111": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R"], "enclosing_function": "test_row_contrib", "extracted_code": "", "n_imports_parsed": 12, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_ca.py::79": {"resolved_imports": ["prince/__init__.py"], "used_names": ["load_df_from_R", "sparse"], "enclosing_function": "test_svd_U", "extracted_code": "", "n_imports_parsed": 13, "n_files_resolved": 1, "n_chars_extracted": 0}}}
oracle_context_cache_v3/Mayitzin__ahrs.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"repo": "Mayitzin/ahrs", "n_pairs": 45, "version": "v3_compressed", "max_tokens": 6000, "contexts": {"tests/test_core_functions.py::11": {"resolved_imports": ["ahrs/utils/core.py"], "used_names": [], "enclosing_function": "test_assert_numerical_iterable", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_estimators.py::353": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_adaptive_gain", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_dcm.py::40": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_rotation_matrix_from_euler_angles", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_common_tools.py::14": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_geometry_circle_default", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::22": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_normal_gravity", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_quaternions.py::27": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_identity_quaternion", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_common_tools.py::23": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_geometry_circle_custom", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::21": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_normal_gravity", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::7": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_correct_values", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::23": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_normal_gravity", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_common_tools.py::16": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_geometry_circle_default", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_dcm.py::60": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_wrong_input_matrix", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_estimators.py::24": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_wrong_frame", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::13": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_correct_values", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wmm.py::84": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["os"], "enclosing_function": "test_wmm2015", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::28": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_values", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wmm.py::85": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["os"], "enclosing_function": "test_wmm2015", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_common_tools.py::12": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_geometry_circle_default", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_quaternions.py::338": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_conjugate", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::27": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_values", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_quaternions.py::287": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_identity_quaternion", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::12": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_correct_values", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::24": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_normal_gravity", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_metrics.py::8": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_correct_values", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_dcm.py::39": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_rotation_matrix_from_euler_angles", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_core_functions.py::8": {"resolved_imports": ["ahrs/utils/core.py"], "used_names": [], "enclosing_function": "test_assert_same_shapes", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_dcm.py::38": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_rotation_matrix_from_euler_angles", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wmm.py::88": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["os"], "enclosing_function": "test_wmm2015", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_core_functions.py::19": {"resolved_imports": ["ahrs/utils/core.py"], "used_names": [], "enclosing_function": "test_get_nan_intervals", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wmm.py::86": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["os"], "enclosing_function": "test_wmm2015", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_quaternions.py::108": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_wrong_input_array", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_estimators.py::21": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_multiple_values", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wmm.py::87": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["os"], "enclosing_function": "test_wmm2015", "extracted_code": "", "n_imports_parsed": 4, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::8": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_correct_values", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::31": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_values", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_estimators.py::358": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_adaptive_gain", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_quaternions.py::126": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_random_attitudes", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_wgs84.py::14": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_correct_values", "extracted_code": "", "n_imports_parsed": 2, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_metrics.py::30": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_guard_clauses", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_quaternions.py::396": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_one_random_attitude", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_estimators.py::350": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_adaptive_gain", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_quaternions.py::28": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_identity_quaternion", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_dcm.py::20": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_rotation_matrix_in_SO3", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_quaternions.py::35": {"resolved_imports": ["ahrs/__init__.py"], "used_names": [], "enclosing_function": "test_conjugate", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}, "tests/test_common_tools.py::21": {"resolved_imports": ["ahrs/__init__.py"], "used_names": ["ahrs"], "enclosing_function": "test_geometry_circle_custom", "extracted_code": "", "n_imports_parsed": 3, "n_files_resolved": 1, "n_chars_extracted": 0}}}
oracle_context_cache_v3/MerrimanInd__drawpyo.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/MiniMax-AI__Mini-Agent.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/MinishLab__model2vec.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/MinishLab__semhash.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/MolecularAI__aizynthfinder.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/MongoEngine__mongoengine.json ADDED
The diff for this file is too large to render. See raw diff
 
oracle_context_cache_v3/MrPowers__chispa.json ADDED
The diff for this file is too large to render. See raw diff