# Real Human Conversations Worldwide — Multilingual Dataset A massive, unified dataset of **real human conversations** from across the world, collected from publicly available sources. This is NOT synthetic customer-service data — it contains genuine human dialogue from Reddit, Discord, Twitch, Usenet, Telegram, therapy sessions, and more. ## 📊 Stats - **Total rows**: 1,609,769 - **Total size**: ~470 MB (zstd-compressed Parquet) - **Format**: Parquet - **Languages**: English, Russian, Italian, Japanese, Korean ## 🌍 Languages | Language | Rows | Size | Sources | |----------|------|------|---------| | English | 1,084,282 | 98.3 MB | Discord, Reddit, Twitch, therapy, YouTube mix | | Russian | 200,000 | 34.1 MB | Telegram chats | | Italian | 198,508 | 318.3 MB | Usenet newsgroups | | Japanese | 100,000 | 10.4 MB | Text conversations | | Korean | 26,979 | 9.1 MB | Everyday chat | ## 📁 Dataset Format All data unified into a consistent schema: | Column | Type | Description | |--------|------|-------------| | `text` | string | Conversation / dialogue / transcript text | | `source` | string | Origin dataset name | | `language` | string | ISO language code | | `domain` | string | Conversation type | | `turns` | int | Number of dialogue turns | | `metadata` | string | JSON with extra info | ## 🚀 Usage ```python from datasets import load_dataset # Load everything ds = load_dataset("asdf98/human-chats-worldwide", split="train") # By language en = ds.filter(lambda x: x["language"] == "en") ru = ds.filter(lambda x: x["language"] == "ru") it = ds.filter(lambda x: x["language"] == "it") ja = ds.filter(lambda x: x["language"] == "ja") ko = ds.filter(lambda x: x["language"] == "ko") ``` ## 🔗 Sources All data from publicly available Hugging Face datasets: | File | Source | Type | Language | |------|--------|------|----------| | `discord_dialogues.parquet` | mookiezi/Discord-Dialogues | Real Discord chats | en | | `reddit_comments.parquet` | HuggingFaceGECLM/REDDIT_comments | Reddit threads | en | | `reddit_confessions.parquet` | SocialGrep/one-million-reddit-confessions | Real confessions | en | | `reddit_questions.parquet` | SocialGrep/one-million-reddit-questions | AskReddit | en | | `reddit_youtube_mix.parquet` | fsteig/conversations-30gb | Reddit+YouTube mix | en | | `russian_dialogues.parquet` | Den4ikAI/russian_dialogues_2 | Telegram chats | ru | | `italian_usenet.parquet` | mii-community/UsenetArchiveIT-conversations | Usenet forums | it | | `twitch_chat.parquet` | lparkourer10/twitch_chat | Live stream chat | en | | `mental_health.parquet` | Amod/mental_health_counseling_conversations | Therapy sessions | en | | `japanese_text.parquet` | izumi-lab/llm-japanese-dataset | Japanese conversations | ja | | `korean_conversations.parquet` | jojo0217/korean_safe_conversation | Everyday Korean chat | ko | ## ⚠️ Notes - All data is **publicly posted** by real humans on public platforms - No private messages or non-consensual data - Filtered to remove `[deleted]`, `[removed]`, and very short posts - Some content may be informal, colloquial, or contain strong language