word stringlengths 1 64 | c int64 10 200M |
|---|---|
the | 199,512,185 |
of | 102,620,319 |
in | 87,602,893 |
and | 82,981,549 |
a | 59,625,532 |
to | 55,793,585 |
was | 33,272,070 |
is | 24,651,348 |
for | 24,412,023 |
on | 24,102,693 |
as | 23,833,908 |
by | 21,368,319 |
with | 20,226,361 |
from | 17,116,492 |
he | 16,727,862 |
at | 16,544,587 |
that | 15,134,112 |
his | 13,154,847 |
it | 12,322,375 |
an | 10,829,152 |
were | 8,668,220 |
also | 8,100,516 |
which | 7,859,030 |
are | 7,504,614 |
first | 6,753,639 |
this | 6,680,512 |
be | 6,455,235 |
s | 6,446,021 |
new | 6,444,250 |
had | 6,364,061 |
or | 6,246,620 |
has | 5,841,720 |
references | 5,637,367 |
one | 5,606,668 |
after | 5,491,902 |
their | 5,428,865 |
she | 5,212,502 |
its | 5,209,126 |
her | 5,205,042 |
who | 5,178,898 |
but | 4,840,072 |
american | 4,827,178 |
not | 4,778,107 |
two | 4,768,153 |
th | 4,687,898 |
they | 4,566,994 |
people | 4,487,606 |
have | 4,195,976 |
been | 3,906,109 |
all | 3,806,161 |
other | 3,787,325 |
time | 3,631,352 |
during | 3,604,452 |
when | 3,472,234 |
university | 3,428,342 |
united | 3,397,434 |
school | 3,295,855 |
may | 3,292,442 |
into | 3,242,772 |
national | 3,238,032 |
year | 3,165,652 |
world | 3,079,173 |
state | 3,028,807 |
players | 3,005,979 |
there | 3,004,965 |
born | 2,996,674 |
i | 2,918,389 |
external | 2,896,780 |
links | 2,884,827 |
states | 2,859,826 |
up | 2,839,486 |
city | 2,827,438 |
century | 2,808,654 |
more | 2,784,250 |
years | 2,766,682 |
over | 2,760,447 |
film | 2,746,120 |
de | 2,708,653 |
would | 2,696,577 |
south | 2,669,235 |
three | 2,650,424 |
later | 2,647,896 |
season | 2,639,467 |
only | 2,635,230 |
between | 2,608,914 |
where | 2,558,101 |
no | 2,551,935 |
about | 2,531,321 |
st | 2,482,645 |
most | 2,479,804 |
team | 2,463,352 |
out | 2,428,066 |
e | 2,424,431 |
county | 2,402,544 |
war | 2,394,606 |
under | 2,391,749 |
series | 2,383,512 |
second | 2,366,612 |
d | 2,346,636 |
history | 2,344,265 |
End of preview. Expand in Data Studio
Wordcounts for the English Wikipedia dump (2023-11-01), including words that occur at least 10 times in the corpus. Created using the following script:
import re
import duckdb
from collections import Counter
from datasets import load_dataset
from tqdm.auto import tqdm
conn = duckdb.connect(":memory:")
def ensure_db(conn: duckdb.DuckDBPyConnection):
conn.execute("""
CREATE TABLE IF NOT EXISTS wc (
word TEXT PRIMARY KEY,
c BIGINT
);
""")
ensure_db(conn)
def merge_batch(conn: duckdb.DuckDBPyConnection, counts: Counter):
if not counts:
return
df = pd.DataFrame({"word": list(counts.keys()), "c": list(map(int, counts.values()))})
# Register the batch dataframe as a view, then MERGE (UPSERT)
conn.register("batch_df", df)
conn.execute("""
MERGE INTO wc AS t
USING batch_df AS s
ON t.word = s.word
WHEN MATCHED THEN UPDATE SET c = t.c + s.c
WHEN NOT MATCHED THEN INSERT (word, c) VALUES (s.word, s.c);
""")
conn.unregister("batch_df")
TOKEN_RE = re.compile(r"[a-z]+(?:'[a-z]+)?") # keep internal apostrophes
def tokenize_en_lower(text: str):
if not text:
return []
return TOKEN_RE.findall(text.lower())
batch_size = 500
limit = 0
ds_iter = load_dataset("wikimedia/wikipedia", "20231101.en", split="train", streaming=True)
buf = Counter()
n = 0
pbar = tqdm(desc="Processing (streaming)", unit="art")
for ex in ds_iter:
buf.update(tokenize_en_lower(ex.get("text", "")))
n += 1
if n % batch_size == 0:
merge_batch(conn, buf); buf.clear()
pbar.update(batch_size)
if limit and n >= limit:
break
if buf:
merge_batch(conn, buf); pbar.update(n % batch_size)
pbar.close()
- Downloads last month
- 15