is_exist bool 2
classes | url stringlengths 28 74 | created_at stringlengths 20 20 | description stringlengths 5 348 | pdf_text stringlengths 98 67.1k | readme_text stringlengths 0 81.3k | nlp_taxonomy_classifier_labels sequencelengths 0 11 | awesome_japanese_nlp_labels sequencelengths 0 2 |
|---|---|---|---|---|---|---|---|
true | https://github.com/takapy0210/nlplot | 2020-05-06T15:09:24Z | Visualization Module for Natural Language Processing | takapy0210 / nlplot
Public
Branches
Tags
Go to file
Go to file
Code
.github/workflows
docs
nlplot
tests
.gitignore
LICENSE
MANIFEST.in
README.md
requirements-dev.txt
requirements.txt
setup.py
nlplot: Analysis and visualization module for Natural Language Processing 📈
Facilitates the visualization of natural language p... | # Carp
<img src="resources/logo/carp_logo_300_c.png" alt="Logo" align="right"/>
[](https://github.com/carp-lang/Carp/actions?query=workflow%3A%22Linux+CI%22)
[](htt... | [
"Natural Language Interfaces",
"Structured Data in NLP",
"Syntactic Text Processing",
"Visual Data in NLP"
] | [] |
true | https://github.com/chezou/Mykytea-python | 2011-07-15T08:34:12Z | Python wrapper for KyTea | chezou / Mykytea-python
Public
Branches
Tags
Go to file
Go to file
Code
.github/workfl…
lib
.gitattribute
.gitignore
LICENSE
MANIFEST.in
Makefile
README.md
pyproject.toml
setup.cfg
setup.py
Sponsor
Sponsor
Mykytea-python is a Python wrapper module for KyTea, a general text analysis toolkit. KyTea is developed by
KyTe... | # KyTea wrapper for Python
[](https://badge.fury.io/py/kytea)
[](https://github.com/sponsors/chezou)
Mykytea-python is a Python wrapper module for KyTea, a general text analysis toolkit.
KyTea i... | [
"Morphology",
"Robustness in NLP",
"Syntactic Text Processing",
"Tagging",
"Text Segmentation"
] | [] |
true | https://github.com/nicolas-raoul/kakasi-java | 2012-01-18T08:30:56Z | Kanji transliteration to hiragana/katakana/romaji, in Java | nicolas-raoul / kakasi-java
Public
Branches
Tags
Go to file
Go to file
Code
dict
docs
src/co…
.gitignore
AUTH…
COPYI…
READ…
build.xml
UPDATE: I just created Jakaroma, its kanji
transliteration is much more accurate than Kakasi-java
so please use it instead, thanks! Also open source.
Kakasi-java Convert Japanese kanji i... | UPDATE: I just created [Jakaroma](https://github.com/nicolas-raoul/jakaroma), its kanji transliteration is much more accurate than Kakasi-java so please use it instead, thanks! Also open source.
Kakasi-java
Convert Japanese kanji into romaji
See also http://kakasi.namazu.org
Originally written by Tomoyuki Kawao
Forke... | [
"Syntactic Text Processing",
"Text Normalization"
] | [] |
true | https://github.com/miurahr/pykakasi | 2012-08-14T14:48:14Z | Lightweight converter from Japanese Kana-kanji sentences into Kana-Roman. | miurahr / pykakasi
Public archive
6 Branches
52 Tags
Go to file
Go to file
Code
miurahr README: fix link
51fe14d · 2 years ago
.github
Actions: instal…
3 years ago
bin
Introduce dep…
3 years ago
docs
PEP8: black f…
3 years ago
src
Support latin1…
2 years ago
tests
Support latin1…
2 years ago
utils
Merge pull re…
3 year... | ========
Pykakasi
========
Overview
========
.. image:: https://readthedocs.org/projects/pykakasi/badge/?version=latest
:target: https://pykakasi.readthedocs.io/en/latest/?badge=latest
:alt: Documentation Status
.. image:: https://badge.fury.io/py/pykakasi.png
:target: http://badge.fury.io/py/Pykakasi
:a... | [
"Syntactic Text Processing",
"Text Normalization"
] | [] |
true | https://github.com/yohokuno/jsc | 2012-08-23T23:39:40Z | Joint source channel model for Japanese Kana Kanji conversion, Chinese pinyin input and CJE mixed input. | yohokuno / jsc
Public
Branches
Tags
Go to file
Go to file
Code
data
src
tools
README.md
TODO
waf
wscript
JSC is an implementation of joint source channel or joint n-gram model with monotonic decoder.
It can be used for machine transliteration, Japanese kana-kanji conversion, Chinese pinyin input, English word segmentat... | JSC: Joint Souce Channel Model and Decoder
===
**JSC** is an implementation of joint source channel or joint n-gram model with monotonic decoder.
A Joint Source-Channel Model for Machine Transliteration, Li Haizhou, Zhang Min, Su Jian.
http://acl.ldc.upenn.edu/acl2004/main/pdf/121_pdf_2-col.pdf
It can be use... | [
"Language Models",
"Multilinguality",
"Syntactic Text Processing"
] | [] |
true | https://github.com/lovell/hepburn | 2013-06-28T10:06:51Z | Node.js module for converting Japanese Hiragana and Katakana script to, and from, Romaji using Hepburn romanisation | lovell / hepburn
Public
Branches
Tags
Go to file
Go to file
Code
.github…
lib
tests
.gitignore
LICEN…
READ…
packag…
Node.js module for converting Japanese Hiragana and Katakana
script to, and from, Romaji using Hepburn romanisation.
Based partly on Takaaki Komura's kana2hepburn.
About
Node.js module for converting
Japa... | # Hepburn
Node.js module for converting Japanese Hiragana and Katakana script to, and from, Romaji using [Hepburn romanisation](http://en.wikipedia.org/wiki/Hepburn_romanization).
Based partly on Takaaki Komura's [kana2hepburn](https://github.com/emon/kana2hepburn).
## Install
npm install hepburn
## Usage
```jav... | [
"Phonology",
"Syntactic Text Processing",
"Text Normalization"
] | [] |
true | https://github.com/jeresig/node-romaji-name | 2013-08-24T17:50:11Z | Normalize and fix common issues with Romaji-based Japanese names. | jeresig / node-romaji-name
Public
Branches
Tags
Go to file
Go to file
Code
.gitignore
LICEN…
READ…
packag…
packag…
romaji-…
setting…
test-pe…
test.js
This is a utility primarily designed for consuming, parsing, and correcting Japanese
names written in rōmaji using proper Hepburn romanization form.
Beyond fixing common ... | romaji-name
================
This is a utility primarily designed for consuming, parsing, and correcting Japanese names written in [rōmaji](https://en.wikipedia.org/wiki/Romanization_of_Japanese) using proper [Hepburn romanization](https://en.wikipedia.org/wiki/Hepburn_romanization) form.
Beyond fixing common problem... | [
"Syntactic Text Processing",
"Text Error Correction",
"Text Normalization"
] | [] |
true | https://github.com/WaniKani/WanaKana | 2013-08-27T19:57:41Z | Javascript library for detecting and transliterating Hiragana <--> Katakana <--> Romaji | WaniKani / WanaKana
Public
Branches
Tags
Go to file
Go to file
Code
.github
cypress
gh-pages
scripts
src
test
.browserslis…
.editorconfig
.eslintrc
.gitignore
.prettierrc
.travis.yml
CHANGEL…
CONTRIBU…
LICENSE
README.md
VERSION
babel.confi…
cypress.json
jsdoc.json
package.json
rollup.confi…
tsconfig.json
yarn.lock
Abou... | <div align="center">
<!-- Npm Version -->
<a href="https://www.npmjs.com/package/wanakana">
<img src="https://img.shields.io/npm/v/wanakana.svg" alt="NPM package" />
</a>
<!-- Build Status -->
<a href="https://travis-ci.org/WaniKani/WanaKana">
<img src="https://img.shields.io/travis/WaniKani/WanaKana.... | [
"Syntactic Text Processing",
"Text Normalization"
] | [] |
true | https://github.com/gojp/nihongo | 2013-09-02T15:17:52Z | Japanese Dictionary | gojp / nihongo
Public
Branches
Tags
Go to file
Go to file
Code
.github…
data
edict2
lib
static
templa…
.gitignore
LICEN…
READ…
go.mod
go.sum
main.go
go report
go report A+
A+
Open source Japanese Dictionary written in Go:
https://nihongo.io
1. git clone
https://github.com/gojp/nihongo.git
About
Japanese Dictionary
# g... | nihongo.io
=========
[](https://goreportcard.com/report/github.com/gojp/nihongo)
Open source Japanese Dictionary written in Go: [https://nihongo.io](https://nihongo.io)
### How to run:
1. `git clone https://github.com/gojp/nihongo.git`
2. Run th... | [] | [
"Vocabulary, Dictionary, and Language Input Method"
] |
true | https://github.com/studio-ousia/mojimoji | 2013-11-02T16:23:06Z | A fast converter between Japanese hankaku and zenkaku characters | studio-ousia / mojimoji
Public
Branches
Tags
Go to file
Go to file
Code
.github…
mojimoji
CONT…
LICEN…
MANIF…
READ…
mojimo…
pyproj…
require…
setup.cfg
setup.py
test_m…
Test
Test
passing
passing pypi
pypi v0.0.13
v0.0.13 pip downloads
pip downloads 7M
7M
A Cython-based fast converter between Japanese hankaku and zenka... | mojimoji
========
.. image:: https://github.com/studio-ousia/mojimoji/actions/workflows/test.yml/badge.svg
:target: https://github.com/studio-ousia/mojimoji/actions/workflows/test.yml
.. image:: https://img.shields.io/pypi/v/mojimoji.svg
:target: https://pypi.org/project/mojimoji/
.. image:: https://static.p... | [
"Syntactic Text Processing",
"Text Normalization"
] | [] |
true | https://github.com/cihai/cihai | 2013-12-03T17:42:52Z | Python library for CJK (Chinese, Japanese, and Korean) language dictionary | Access to this site has been restricted.
If you believe this is an error, please contact Support.
GitHub Status — @githubstatus
| # cihai · [](https://pypi.org/project/cihai/) [](https://github.com/cihai/cihai/blob/master/LICENSE) [](https... | [
"Multilinguality",
"Syntactic Text Processing"
] | [
"Annotation and Dataset Development"
] |
true | https://github.com/SamuraiT/mecab-python3 | 2014-05-31T08:47:04Z | mecab-python. mecab-python. you can find original version here:http://taku910.github.io/mecab/ | SamuraiT / mecab-python3
Public
Branches
Tags
Go to file
Go to file
Code
.github
debian
src/Me…
test
.gitattri…
.gitignore
AUTH…
BSD
COPYI…
Docker…
GPL
LGPL
MANIF…
READ…
setup.py
tox.ini
test-manylinux
test-manylinux
passing
passing downloads
downloads 745k/month
745k/month
platforms
platforms linux macosx windows
l... | [](https://pypi.org/project/mecab-python3/)

[](https://pypi.org/project/mecab-python3/)
![Supp... | [
"Morphology",
"Syntactic Text Processing",
"Tagging",
"Text Segmentation"
] | [] |
true | https://github.com/hakatashi/kyujitai.js | 2014-09-06T08:05:01Z | Utility collections for making Japanese text old-fashioned | hakatashi / kyujitai.js
Public
7 Branches
8 Tags
Go to file
Go to file
Code
hakatashi 1.3.0
8b627c7 · 4 years ago
calibrate
Remove BO…
10 years ago
data
Add a charact…
4 years ago
dist
Update dist
10 years ago
lib
Change pack…
9 years ago
test
Change test t…
10 years ago
.gitattri…
Add .gitattribu…
4 years ago
.gitigno... | # kyujitai.js
[](https://travis-ci.org/hakatashi/kyujitai.js)
[](https://greenkeeper.io/)
Utility collections for making Japanese text old-fashioned.
## install
npm instal... | [
"Low-Resource NLP",
"Syntactic Text Processing"
] | [] |
true | https://github.com/ikegami-yukino/rakutenma-python | 2015-01-01T21:40:43Z | Rakuten MA (Python version) | ikegami-yukino / rakutenma-python
Public
Branches
Tags
Go to file
Go to file
Code
bin
rakutenma
tests
.gitignore
.travis.yml
CHANGES.rst
LICENSE.txt
MANIFEST.in
README.rst
setup.py
travis-ci.org
python
python 2.6 | 2.7 | 3.3 | 3.4 | 3.5 | 3.6
2.6 | 2.7 | 3.3 | 3.4 | 3.5 | 3.6 pypi
pypi v0.3.3
v0.3.3
Code Health li... | Rakuten MA Python
===================
|travis| |coveralls| |pyversion| |version| |landscape| |license|
Rakuten MA Python (morphological analyzer) is a Python version of Rakuten MA (word segmentor + PoS Tagger) for Chinese and Japanese.
For details about Rakuten MA, See https://github.com/rakuten-nlp/rakutenma
See... | [
"Morphology",
"Syntactic Text Processing",
"Tagging",
"Text Segmentation"
] | [] |
true | https://github.com/google/budoux | 2015-03-18T18:22:31Z | Standalone. Small. Language-neutral. BudouX is the successor to Budou, the machine learning powered line break organizer tool. | google / budoux
Public
Branches
Tags
Go to file
Go to file
Code
.github
budoux
data/finetuning/ja
demo
java
javascript
scripts
tests
.gitignore
.markdownlint.yaml
CONTRIBUTING.md
LICENSE
MANIFEST.in
README.md
bump_version.py
example.png
pyproject.toml
setup.cfg
setup.py
pypi
pypi v0.6.3
v0.6.3 npm
npm v0.6.3
v0.6.3 m... | <!-- markdownlint-disable MD014 -->
# BudouX
[](https://pypi.org/project/budoux/) [](https://www.npmjs.com/package/budoux) [](... | [
"Chunking",
"Syntactic Text Processing",
"Text Segmentation"
] | [] |
true | https://github.com/scriptin/topokanji | 2015-05-28T17:52:28Z | Topologically ordered lists of kanji for effective learning | scriptin / topokanji
Public
Branches
Tags
Go to file
Go to file
Code
data
depen…
lib
lists
.gitignore
.jshintrc
READ…
build.js
packag…
30 seconds explanation for people who want to
learn kanji:
It is best to learn kanji starting from simple
characters and then learning complex ones as
compositions of "parts", which are... | # TopoKanji
> **30 seconds explanation for people who want to learn kanji:**
>
> It is best to learn kanji starting from simple characters and then learning complex ones as compositions of "parts", which are called "radicals" or "components". For example:
>
> - 一 → 二 → 三
> - 丨 → 凵 → 山 → 出
> - 言 → 五 → 口 → 語
>
> It is a... | [] | [
"Annotation and Dataset Development"
] |
true | https://github.com/Kensuke-Mitsuzawa/JapaneseTokenizers | 2015-09-01T10:24:45Z | A set of metrics for feature selection from text data | Kensuke-Mitsuzawa / JapaneseTokenizers
Public
Branches
Tags
Go to file
Go to file
Code
Japan…
exampl…
test
.gitignore
.travis.…
LICEN…
MANIF…
Makefile
READ…
install_…
setup.py
travis-…
license
license MIT
MIT
Build Status
This is simple python-wrapper for Japanese Tokenizers(A.K.A Tokenizer)
This project aims to call t... | [](LICENSE)[](https://travis-ci.org/Kensuke-Mitsuzawa/JapaneseTokenizers)
# What's this?
This is simple python-wrapper for Japanese Tokenizers(A.K.A ... | [
"Morphology",
"Responsible & Trustworthy NLP",
"Robustness in NLP",
"Syntactic Text Processing",
"Tagging",
"Text Segmentation"
] | [] |
true | https://github.com/tokuhirom/akaza | 2015-10-14T01:17:00Z | Yet another Japanese IME for IBus/Linux | akaza-im / akaza
Public
Branches
Tags
Go to file
Go to file
Code
.github…
.run
akaza-…
akaza-…
akaza-…
develo…
docs
ibus-ak…
ibus-sys
keymap
libakaza
marisa…
romkan
.gitattri…
.gitignore
.gitmo…
CONT…
Cargo.l…
Cargo.…
Chang…
LICEN…
About
Yet another Japanese IME for
IBus/Linux
# nlp # rust # ime # ibus
Readme
MIT ... | # ibus-akaza
Yet another kana-kanji-converter on IBus, written in Rust.
統計的かな漢字変換による日本語IMEです。
Rust で書いています。
**現在、開発途中のプロダクトです。非互換の変更が予告なくはいります**
## モチベーション
いじりやすくて **ある程度** UIが使いやすいかな漢字変換があったら面白いなと思ったので作ってみています。
「いじりやすくて」というのはつまり、Hack-able であるという意味です。
モデルデータを自分で生成できて、特定の企業に依存しない自由なかな漢字変換エンジンを作りたい。
## 特徴
* UI/Lo... | [
"Language Models",
"Syntactic Text Processing"
] | [
"Vocabulary, Dictionary, and Language Input Method"
] |
true | https://github.com/hexenq/kuroshiro | 2016-01-03T09:16:40Z | Japanese language library for converting Japanese sentence to Hiragana, Katakana or Romaji with furigana and okurigana modes supported. | hexenq / kuroshiro
Public
7 Branches
18 Tags
Go to file
Go to file
Code
src
test
.babelrc
.eslintrc
.gitignore
.npmignore
.travis.yml
CHANGELOG.md
CONTRIBUTING.md
LICENSE
README.eo-eo.md
README.jp.md
README.md
README.zh-cn.md
README.zh-tw.md
package.json
Build Status coverage
coverage
96%
96%
gitter
gitter
join chat... | 
# kuroshiro
[](https://travis-ci.org/hexenq/kuroshiro)
[](https://coveralls.io/r/hexenq/kuroshiro)
[
- Output: Multi-label classification of NLP research fields
Problem Setting of the Dataset:
- Training Data: GitHub repositories from before 2022
- Test Data: GitHub repositories from 2023
- Objective: To predict multi-labels of research field labels in the NLP domain
- We used the positive examples from awesome-japanese-nlp-classification-dataset.
Dataset Features:
- A multimodal dataset including GitHub summaries, README files, PDF files, and screenshot images.
- The annotation labels were assigned with reference to Exploring-NLP-Research.
If you need image data and PDF files, please submit a request to the community. This dataset uses text extracted from PDF files and does not include image data.
Based on GitHub's terms of service, please use this dataset for research purposes only.
How to use this dataset
Install the datasets library.
pip install datasets
How to load in Python.
from datasets import load_dataset
dataset = load_dataset("taishi-i/awesome-japanese-nlp-multilabel-dataset")
# To access the dataset sample
print(dataset["train"][0])
Here is a sample of the dataset.
{
"is_exist": True,
"url": "https://github.com/scriptin/jmdict-simplified",
"created_at": "2016-02-07T16:34:32Z",
"description": "JMdict and JMnedict in JSON format",
"pdf_text": "scriptin / jmdict-simplified\nPublic\nBranches\nTags\n",
"readme_text": "# jmdict-simplified\n\n**[JMdict][], [JMnedict][], [Kanjidic][], and [Kradfile/Radkfile][Kradfile] in JSON format**<br>\n",
"nlp_taxonomy_classifier_labels": [
"Multilinguality"
],
"awesome_japanese_nlp_labels": [
"Annotation and Dataset Development",
"Vocabulary, Dictionary, and Language Input Method"
]
}
is_exist: A boolean value indicating whether the GitHub repository exists (Trueif it exists) at the time of retrieval.url: The URL of the GitHub repository.created_at: The timestamp indicating when the repository was created (in ISO 8601 format).description: A short description of the repository provided on GitHub.pdf_text: Extracted text from a PDF file that contains a saved version of the GitHub repository's top page.readme_text: Extracted text from the repository's README file.nlp_taxonomy_classifier_labels: Manually annotated multi-label annotations of NLP research field labels.awesome_japanese_nlp_labels: Manually annotated multi-label annotations of research field labels specifically for Japanese natural language processing.
Details of the dataset.
DatasetDict({
train: Dataset({
features: ['is_exist', 'url', 'created_at', 'description', 'pdf_text', 'readme_text', 'nlp_taxonomy_classifier_labels', 'awesome_japanese_nlp_labels'],
num_rows: 407
})
validation: Dataset({
features: ['is_exist', 'url', 'created_at', 'description', 'pdf_text', 'readme_text', 'nlp_taxonomy_classifier_labels', 'awesome_japanese_nlp_labels'],
num_rows: 17
})
test: Dataset({
features: ['is_exist', 'url', 'created_at', 'description', 'pdf_text', 'readme_text', 'nlp_taxonomy_classifier_labels', 'awesome_japanese_nlp_labels'],
num_rows: 60
})
})
Baseline
The baseline model was used for TimSchopf/nlp_taxonomy_classifier.
The fine-tuned model was trained using this dataset to fine-tune the baseline model.
| Classification Method | Model | Description | Dev Prec. | Dev Rec. | Dev F1 | Eval Prec. | Eval Rec. | Eval F1 |
|---|---|---|---|---|---|---|---|---|
| Random Prediction | - | - | 0.034 | 0.455 | 0.064 | 0.042 | 0.513 | 0.078 |
| Baseline | TimSchopf/nlp_taxonomy_classifier | ✓ | 0.538 | 0.382 | 0.447 | 0.360 | 0.354 | 0.349 |
| Fine-Tuning | TimSchopf/nlp_taxonomy_classifier | ✓ | 0.538 | 0.509 | 0.523 | 0.436 | 0.484 | 0.521 |
| Zero-Shot | gpt-4o-2024-08-06 | ✓ | 0.560 | 0.255 | 0.350 | 0.476 | 0.184 | 0.265 |
License
We collect and publish this dataset under GitHub Acceptable Use Policies - 7. Information Usage Restrictions and GitHub Terms of Service - H. API Terms for research purposes. This dataset should be used solely for research verification purposes. Adhering to GitHub's regulations is mandatory.
- Downloads last month
- 18