Datasets:
Update README.md
Browse files
README.md
CHANGED
|
@@ -4,17 +4,18 @@ language:
|
|
| 4 |
- en
|
| 5 |
tags:
|
| 6 |
- climate
|
|
|
|
| 7 |
---
|
| 8 |
|
| 9 |
Argovis Argo Ocean Profiles
|
| 10 |
|
| 11 |
-
## Dataset summary
|
| 12 |
|
| 13 |
This dataset contains ocean profile data collected by the international Argo float program and accessed via the Argovis API. Each record corresponds to a single profile measured by an autonomous drifting float, including time, location, basin, and associated profile metadata fields that can be joined to the underlying temperature and salinity data structures. The goal of this dataset is to provide a ready-to-use subset of Argo profiles for machine learning, geospatial analysis, and educational use.[1][2][3][4]
|
| 14 |
|
| 15 |
The files were programmatically fetched using the Argovis API in Python, converted to pandas DataFrames, and exported as CSV. This makes it easy to load the data in common data science environments (Python, R, Julia) without needing to write custom API integration code.[5][6][1]
|
| 16 |
|
| 17 |
-
## Source and provenance
|
| 18 |
|
| 19 |
- Original data source: Argo Global Data Assembly Centers (GDACs), accessed through the Argovis platform.[2][4]
|
| 20 |
- Access method: Argovis API (https://argovis-api.colorado.edu/argo) with query filters on time and optional geographic constraints.[1][5]
|
|
@@ -25,7 +26,7 @@ The files were programmatically fetched using the Argovis API in Python, convert
|
|
| 25 |
|
| 26 |
These data are a derivative, convenience-formatted view of the original Argo profile data; they do not modify scientific content, only representation (JSON → tabular CSV).
|
| 27 |
|
| 28 |
-
## Files included
|
| 29 |
|
| 30 |
Depending on how you upload, you might have some or all of:
|
| 31 |
|
|
@@ -41,7 +42,7 @@ Depending on how you upload, you might have some or all of:
|
|
| 41 |
|
| 42 |
If additional files are added later (e.g., separate files for metadata, variables, or different time windows), the filenames should clearly reflect their content and time coverage.
|
| 43 |
|
| 44 |
-
## Data fields (high-level)
|
| 45 |
|
| 46 |
Typical columns in the main CSV include:[8][1]
|
| 47 |
|
|
@@ -59,7 +60,7 @@ Typical columns in the main CSV include:[8][1]
|
|
| 59 |
|
| 60 |
Users are encouraged to consult Argovis and Argo documentation for full definitions of scientific variables, QC flags, and conventions.[4][3][1]
|
| 61 |
|
| 62 |
-
## Intended uses
|
| 63 |
|
| 64 |
This dataset is useful for:
|
| 65 |
|
|
@@ -75,7 +76,7 @@ This dataset is useful for:
|
|
| 75 |
|
| 76 |
Because this dataset is derived from the operational Argo network, it reflects real measurement noise, missing data patterns, and QC flags that are valuable for realistic ML pipelines.[4][3]
|
| 77 |
|
| 78 |
-
## How to load the data
|
| 79 |
|
| 80 |
Example in Python with pandas:
|
| 81 |
|
|
@@ -101,7 +102,7 @@ df = dataset["train"].to_pandas()
|
|
| 101 |
|
| 102 |
Adjust split name and file mapping depending on how the dataset is configured on the Hub.[9][10]
|
| 103 |
|
| 104 |
-
## Limitations and caveats
|
| 105 |
|
| 106 |
- This is a subset in time (and optionally space), not the full Argo archive.[2][4]
|
| 107 |
- Deep scientific interpretation (e.g., water mass analysis) requires careful handling of:
|
|
@@ -112,7 +113,7 @@ Adjust split name and file mapping depending on how the dataset is configured on
|
|
| 112 |
|
| 113 |
Users should always refer to the official Argo documentation and Argovis API documentation for authoritative descriptions of variables and processing.[11][1][3]
|
| 114 |
|
| 115 |
-
## License and attribution
|
| 116 |
|
| 117 |
The underlying Argo data are described as freely available without restriction and are treated similarly to open data in many catalogs. However, proper acknowledgment of the Argo program is required in any work using these data.[12][2]
|
| 118 |
|
|
@@ -122,7 +123,7 @@ If using this dataset, please include an acknowledgment along the lines of:
|
|
| 122 |
|
| 123 |
Also cite this Hugging Face dataset if it is used as a curated, preprocessed source in your workflows.
|
| 124 |
|
| 125 |
-
## Contact and contributions
|
| 126 |
|
| 127 |
If you find issues in the CSV export, want additional time ranges or regions, or would like to contribute parsing scripts or example notebooks (e.g., for plotting sections or training ML models), feel free to:
|
| 128 |
|
|
|
|
| 4 |
- en
|
| 5 |
tags:
|
| 6 |
- climate
|
| 7 |
+
pretty_name: 'ARGO_Profiles:'
|
| 8 |
---
|
| 9 |
|
| 10 |
Argovis Argo Ocean Profiles
|
| 11 |
|
| 12 |
+
## Dataset summary:
|
| 13 |
|
| 14 |
This dataset contains ocean profile data collected by the international Argo float program and accessed via the Argovis API. Each record corresponds to a single profile measured by an autonomous drifting float, including time, location, basin, and associated profile metadata fields that can be joined to the underlying temperature and salinity data structures. The goal of this dataset is to provide a ready-to-use subset of Argo profiles for machine learning, geospatial analysis, and educational use.[1][2][3][4]
|
| 15 |
|
| 16 |
The files were programmatically fetched using the Argovis API in Python, converted to pandas DataFrames, and exported as CSV. This makes it easy to load the data in common data science environments (Python, R, Julia) without needing to write custom API integration code.[5][6][1]
|
| 17 |
|
| 18 |
+
## Source and provenance:
|
| 19 |
|
| 20 |
- Original data source: Argo Global Data Assembly Centers (GDACs), accessed through the Argovis platform.[2][4]
|
| 21 |
- Access method: Argovis API (https://argovis-api.colorado.edu/argo) with query filters on time and optional geographic constraints.[1][5]
|
|
|
|
| 26 |
|
| 27 |
These data are a derivative, convenience-formatted view of the original Argo profile data; they do not modify scientific content, only representation (JSON → tabular CSV).
|
| 28 |
|
| 29 |
+
## Files included:
|
| 30 |
|
| 31 |
Depending on how you upload, you might have some or all of:
|
| 32 |
|
|
|
|
| 42 |
|
| 43 |
If additional files are added later (e.g., separate files for metadata, variables, or different time windows), the filenames should clearly reflect their content and time coverage.
|
| 44 |
|
| 45 |
+
## Data fields (high-level):
|
| 46 |
|
| 47 |
Typical columns in the main CSV include:[8][1]
|
| 48 |
|
|
|
|
| 60 |
|
| 61 |
Users are encouraged to consult Argovis and Argo documentation for full definitions of scientific variables, QC flags, and conventions.[4][3][1]
|
| 62 |
|
| 63 |
+
## Intended uses:
|
| 64 |
|
| 65 |
This dataset is useful for:
|
| 66 |
|
|
|
|
| 76 |
|
| 77 |
Because this dataset is derived from the operational Argo network, it reflects real measurement noise, missing data patterns, and QC flags that are valuable for realistic ML pipelines.[4][3]
|
| 78 |
|
| 79 |
+
## How to load the data:
|
| 80 |
|
| 81 |
Example in Python with pandas:
|
| 82 |
|
|
|
|
| 102 |
|
| 103 |
Adjust split name and file mapping depending on how the dataset is configured on the Hub.[9][10]
|
| 104 |
|
| 105 |
+
## Limitations and caveats:
|
| 106 |
|
| 107 |
- This is a subset in time (and optionally space), not the full Argo archive.[2][4]
|
| 108 |
- Deep scientific interpretation (e.g., water mass analysis) requires careful handling of:
|
|
|
|
| 113 |
|
| 114 |
Users should always refer to the official Argo documentation and Argovis API documentation for authoritative descriptions of variables and processing.[11][1][3]
|
| 115 |
|
| 116 |
+
## License and attribution:
|
| 117 |
|
| 118 |
The underlying Argo data are described as freely available without restriction and are treated similarly to open data in many catalogs. However, proper acknowledgment of the Argo program is required in any work using these data.[12][2]
|
| 119 |
|
|
|
|
| 123 |
|
| 124 |
Also cite this Hugging Face dataset if it is used as a curated, preprocessed source in your workflows.
|
| 125 |
|
| 126 |
+
## Contact and contributions:
|
| 127 |
|
| 128 |
If you find issues in the CSV export, want additional time ranges or regions, or would like to contribute parsing scripts or example notebooks (e.g., for plotting sections or training ML models), feel free to:
|
| 129 |
|