| # Real Human Conversations Worldwide — Multilingual Dataset |
|
|
| A massive, unified dataset of **real human conversations** from across the world, collected from publicly available sources. This is NOT synthetic customer-service data — it contains genuine human dialogue from Reddit, Discord, Twitch, Usenet, Telegram, therapy sessions, and more. |
|
|
| ## 📊 Stats |
|
|
| - **Total rows**: 1,609,769 |
| - **Total size**: ~470 MB (zstd-compressed Parquet) |
| - **Format**: Parquet |
| - **Languages**: English, Russian, Italian, Japanese, Korean |
|
|
| ## 🌍 Languages |
|
|
| | Language | Rows | Size | Sources | |
| |----------|------|------|---------| |
| | English | 1,084,282 | 98.3 MB | Discord, Reddit, Twitch, therapy, YouTube mix | |
| | Russian | 200,000 | 34.1 MB | Telegram chats | |
| | Italian | 198,508 | 318.3 MB | Usenet newsgroups | |
| | Japanese | 100,000 | 10.4 MB | Text conversations | |
| | Korean | 26,979 | 9.1 MB | Everyday chat | |
|
|
| ## 📁 Dataset Format |
|
|
| All data unified into a consistent schema: |
|
|
| | Column | Type | Description | |
| |--------|------|-------------| |
| | `text` | string | Conversation / dialogue / transcript text | |
| | `source` | string | Origin dataset name | |
| | `language` | string | ISO language code | |
| | `domain` | string | Conversation type | |
| | `turns` | int | Number of dialogue turns | |
| | `metadata` | string | JSON with extra info | |
|
|
| ## 🚀 Usage |
|
|
| ```python |
| from datasets import load_dataset |
| |
| # Load everything |
| ds = load_dataset("asdf98/human-chats-worldwide", split="train") |
| |
| # By language |
| en = ds.filter(lambda x: x["language"] == "en") |
| ru = ds.filter(lambda x: x["language"] == "ru") |
| it = ds.filter(lambda x: x["language"] == "it") |
| ja = ds.filter(lambda x: x["language"] == "ja") |
| ko = ds.filter(lambda x: x["language"] == "ko") |
| ``` |
|
|
| ## 🔗 Sources |
|
|
| All data from publicly available Hugging Face datasets: |
|
|
| | File | Source | Type | Language | |
| |------|--------|------|----------| |
| | `discord_dialogues.parquet` | mookiezi/Discord-Dialogues | Real Discord chats | en | |
| | `reddit_comments.parquet` | HuggingFaceGECLM/REDDIT_comments | Reddit threads | en | |
| | `reddit_confessions.parquet` | SocialGrep/one-million-reddit-confessions | Real confessions | en | |
| | `reddit_questions.parquet` | SocialGrep/one-million-reddit-questions | AskReddit | en | |
| | `reddit_youtube_mix.parquet` | fsteig/conversations-30gb | Reddit+YouTube mix | en | |
| | `russian_dialogues.parquet` | Den4ikAI/russian_dialogues_2 | Telegram chats | ru | |
| | `italian_usenet.parquet` | mii-community/UsenetArchiveIT-conversations | Usenet forums | it | |
| | `twitch_chat.parquet` | lparkourer10/twitch_chat | Live stream chat | en | |
| | `mental_health.parquet` | Amod/mental_health_counseling_conversations | Therapy sessions | en | |
| | `japanese_text.parquet` | izumi-lab/llm-japanese-dataset | Japanese conversations | ja | |
| | `korean_conversations.parquet` | jojo0217/korean_safe_conversation | Everyday Korean chat | ko | |
|
|
| ## ⚠️ Notes |
|
|
| - All data is **publicly posted** by real humans on public platforms |
| - No private messages or non-consensual data |
| - Filtered to remove `[deleted]`, `[removed]`, and very short posts |
| - Some content may be informal, colloquial, or contain strong language |
|
|