Datasets:
Languages:
Greek
Size:
10K - 100K
Tags:
speech_corpus
greek_language
text-to-speech
neural_TTS
speech_synthesis
under_resourced_languages
License:
Update README.md
Browse files
README.md
CHANGED
|
@@ -55,7 +55,7 @@ You can load this dataset directly in Python using the Hugging Face `datasets` l
|
|
| 55 |
from datasets import load_dataset
|
| 56 |
|
| 57 |
# Download and Load the dataset
|
| 58 |
-
dataset = load_dataset("aangelakis/STOMA",
|
| 59 |
|
| 60 |
# Inspect the first example
|
| 61 |
sample = dataset[0]
|
|
@@ -63,6 +63,10 @@ sample = dataset[0]
|
|
| 63 |
# Print the Greek text
|
| 64 |
print("Text:", sample["text"])
|
| 65 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 66 |
# The 'audio' column is automatically decoded into a dictionary with 'array' and 'sampling_rate'
|
| 67 |
audio_data = sample["audio"]["array"]
|
| 68 |
sampling_rate = sample["audio"]["sampling_rate"]
|
|
|
|
| 55 |
from datasets import load_dataset
|
| 56 |
|
| 57 |
# Download and Load the dataset
|
| 58 |
+
dataset = load_dataset("aangelakis/STOMA", split="train")
|
| 59 |
|
| 60 |
# Inspect the first example
|
| 61 |
sample = dataset[0]
|
|
|
|
| 63 |
# Print the Greek text
|
| 64 |
print("Text:", sample["text"])
|
| 65 |
|
| 66 |
+
# Print metadata
|
| 67 |
+
print("Speaker:", sample["speaker_id"])
|
| 68 |
+
print("Corpus Section:", sample["section"])
|
| 69 |
+
|
| 70 |
# The 'audio' column is automatically decoded into a dictionary with 'array' and 'sampling_rate'
|
| 71 |
audio_data = sample["audio"]["array"]
|
| 72 |
sampling_rate = sample["audio"]["sampling_rate"]
|