Dataset Viewer
Auto-converted to Parquet Duplicate
is_exist
bool
2 classes
url
stringlengths
28
74
created_at
stringlengths
20
20
description
stringlengths
5
348
pdf_text
stringlengths
98
67.1k
readme_text
stringlengths
0
81.3k
nlp_taxonomy_classifier_labels
sequencelengths
0
11
awesome_japanese_nlp_labels
sequencelengths
0
2
true
https://github.com/takapy0210/nlplot
2020-05-06T15:09:24Z
Visualization Module for Natural Language Processing
takapy0210 / nlplot Public Branches Tags Go to file Go to file Code .github/workflows docs nlplot tests .gitignore LICENSE MANIFEST.in README.md requirements-dev.txt requirements.txt setup.py nlplot: Analysis and visualization module for Natural Language Processing 📈 Facilitates the visualization of natural language p...
# Carp <img src="resources/logo/carp_logo_300_c.png" alt="Logo" align="right"/> [![Linux CI](https://github.com/carp-lang/Carp/workflows/Linux%20CI/badge.svg)](https://github.com/carp-lang/Carp/actions?query=workflow%3A%22Linux+CI%22) [![MacOS CI](https://github.com/carp-lang/Carp/workflows/MacOS%20CI/badge.svg)](htt...
[ "Natural Language Interfaces", "Structured Data in NLP", "Syntactic Text Processing", "Visual Data in NLP" ]
[]
true
https://github.com/chezou/Mykytea-python
2011-07-15T08:34:12Z
Python wrapper for KyTea
chezou / Mykytea-python Public Branches Tags Go to file Go to file Code .github/workfl… lib .gitattribute .gitignore LICENSE MANIFEST.in Makefile README.md pyproject.toml setup.cfg setup.py Sponsor Sponsor Mykytea-python is a Python wrapper module for KyTea, a general text analysis toolkit. KyTea is developed by KyTe...
# KyTea wrapper for Python [![PyPI version](https://badge.fury.io/py/kytea.svg)](https://badge.fury.io/py/kytea) [![](https://img.shields.io/badge/-Sponsor-fafbfc?logo=GitHub%20Sponsors )](https://github.com/sponsors/chezou) Mykytea-python is a Python wrapper module for KyTea, a general text analysis toolkit. KyTea i...
[ "Morphology", "Robustness in NLP", "Syntactic Text Processing", "Tagging", "Text Segmentation" ]
[]
true
https://github.com/nicolas-raoul/kakasi-java
2012-01-18T08:30:56Z
Kanji transliteration to hiragana/katakana/romaji, in Java
nicolas-raoul / kakasi-java Public Branches Tags Go to file Go to file Code dict docs src/co… .gitignore AUTH… COPYI… READ… build.xml UPDATE: I just created Jakaroma, its kanji transliteration is much more accurate than Kakasi-java so please use it instead, thanks! Also open source. Kakasi-java Convert Japanese kanji i...
UPDATE: I just created [Jakaroma](https://github.com/nicolas-raoul/jakaroma), its kanji transliteration is much more accurate than Kakasi-java so please use it instead, thanks! Also open source. Kakasi-java Convert Japanese kanji into romaji See also http://kakasi.namazu.org Originally written by Tomoyuki Kawao Forke...
[ "Syntactic Text Processing", "Text Normalization" ]
[]
true
https://github.com/miurahr/pykakasi
2012-08-14T14:48:14Z
Lightweight converter from Japanese Kana-kanji sentences into Kana-Roman.
miurahr / pykakasi Public archive 6 Branches 52 Tags Go to file Go to file Code miurahr README: fix link 51fe14d · 2 years ago .github Actions: instal… 3 years ago bin Introduce dep… 3 years ago docs PEP8: black f… 3 years ago src Support latin1… 2 years ago tests Support latin1… 2 years ago utils Merge pull re… 3 year...
======== Pykakasi ======== Overview ======== .. image:: https://readthedocs.org/projects/pykakasi/badge/?version=latest :target: https://pykakasi.readthedocs.io/en/latest/?badge=latest :alt: Documentation Status .. image:: https://badge.fury.io/py/pykakasi.png :target: http://badge.fury.io/py/Pykakasi :a...
[ "Syntactic Text Processing", "Text Normalization" ]
[]
true
https://github.com/yohokuno/jsc
2012-08-23T23:39:40Z
Joint source channel model for Japanese Kana Kanji conversion, Chinese pinyin input and CJE mixed input.
yohokuno / jsc Public Branches Tags Go to file Go to file Code data src tools README.md TODO waf wscript JSC is an implementation of joint source channel or joint n-gram model with monotonic decoder. It can be used for machine transliteration, Japanese kana-kanji conversion, Chinese pinyin input, English word segmentat...
JSC: Joint Souce Channel Model and Decoder === **JSC** is an implementation of joint source channel or joint n-gram model with monotonic decoder. A Joint Source-Channel Model for Machine Transliteration, Li Haizhou, Zhang Min, Su Jian. http://acl.ldc.upenn.edu/acl2004/main/pdf/121_pdf_2-col.pdf It can be use...
[ "Language Models", "Multilinguality", "Syntactic Text Processing" ]
[]
true
https://github.com/lovell/hepburn
2013-06-28T10:06:51Z
Node.js module for converting Japanese Hiragana and Katakana script to, and from, Romaji using Hepburn romanisation
lovell / hepburn Public Branches Tags Go to file Go to file Code .github… lib tests .gitignore LICEN… READ… packag… Node.js module for converting Japanese Hiragana and Katakana script to, and from, Romaji using Hepburn romanisation. Based partly on Takaaki Komura's kana2hepburn. About Node.js module for converting Japa...
# Hepburn Node.js module for converting Japanese Hiragana and Katakana script to, and from, Romaji using [Hepburn romanisation](http://en.wikipedia.org/wiki/Hepburn_romanization). Based partly on Takaaki Komura's [kana2hepburn](https://github.com/emon/kana2hepburn). ## Install npm install hepburn ## Usage ```jav...
[ "Phonology", "Syntactic Text Processing", "Text Normalization" ]
[]
true
https://github.com/jeresig/node-romaji-name
2013-08-24T17:50:11Z
Normalize and fix common issues with Romaji-based Japanese names.
jeresig / node-romaji-name Public Branches Tags Go to file Go to file Code .gitignore LICEN… READ… packag… packag… romaji-… setting… test-pe… test.js This is a utility primarily designed for consuming, parsing, and correcting Japanese names written in rōmaji using proper Hepburn romanization form. Beyond fixing common ...
romaji-name ================ This is a utility primarily designed for consuming, parsing, and correcting Japanese names written in [rōmaji](https://en.wikipedia.org/wiki/Romanization_of_Japanese) using proper [Hepburn romanization](https://en.wikipedia.org/wiki/Hepburn_romanization) form. Beyond fixing common problem...
[ "Syntactic Text Processing", "Text Error Correction", "Text Normalization" ]
[]
true
https://github.com/WaniKani/WanaKana
2013-08-27T19:57:41Z
Javascript library for detecting and transliterating Hiragana <--> Katakana <--> Romaji
WaniKani / WanaKana Public Branches Tags Go to file Go to file Code .github cypress gh-pages scripts src test .browserslis… .editorconfig .eslintrc .gitignore .prettierrc .travis.yml CHANGEL… CONTRIBU… LICENSE README.md VERSION babel.confi… cypress.json jsdoc.json package.json rollup.confi… tsconfig.json yarn.lock Abou...
<div align="center"> <!-- Npm Version --> <a href="https://www.npmjs.com/package/wanakana"> <img src="https://img.shields.io/npm/v/wanakana.svg" alt="NPM package" /> </a> <!-- Build Status --> <a href="https://travis-ci.org/WaniKani/WanaKana"> <img src="https://img.shields.io/travis/WaniKani/WanaKana....
[ "Syntactic Text Processing", "Text Normalization" ]
[]
true
https://github.com/gojp/nihongo
2013-09-02T15:17:52Z
Japanese Dictionary
gojp / nihongo Public Branches Tags Go to file Go to file Code .github… data edict2 lib static templa… .gitignore LICEN… READ… go.mod go.sum main.go go report go report A+ A+ Open source Japanese Dictionary written in Go: https://nihongo.io 1. git clone https://github.com/gojp/nihongo.git About Japanese Dictionary # g...
nihongo.io ========= [![Go Report Card](https://goreportcard.com/badge/github.com/gojp/nihongo)](https://goreportcard.com/report/github.com/gojp/nihongo) Open source Japanese Dictionary written in Go: [https://nihongo.io](https://nihongo.io) ### How to run: 1. `git clone https://github.com/gojp/nihongo.git` 2. Run th...
[]
[ "Vocabulary, Dictionary, and Language Input Method" ]
true
https://github.com/studio-ousia/mojimoji
2013-11-02T16:23:06Z
A fast converter between Japanese hankaku and zenkaku characters
studio-ousia / mojimoji Public Branches Tags Go to file Go to file Code .github… mojimoji CONT… LICEN… MANIF… READ… mojimo… pyproj… require… setup.cfg setup.py test_m… Test Test passing passing pypi pypi v0.0.13 v0.0.13 pip downloads pip downloads 7M 7M A Cython-based fast converter between Japanese hankaku and zenka...
mojimoji ======== .. image:: https://github.com/studio-ousia/mojimoji/actions/workflows/test.yml/badge.svg :target: https://github.com/studio-ousia/mojimoji/actions/workflows/test.yml .. image:: https://img.shields.io/pypi/v/mojimoji.svg :target: https://pypi.org/project/mojimoji/ .. image:: https://static.p...
[ "Syntactic Text Processing", "Text Normalization" ]
[]
true
https://github.com/cihai/cihai
2013-12-03T17:42:52Z
Python library for CJK (Chinese, Japanese, and Korean) language dictionary
Access to this site has been restricted. If you believe this is an error, please contact Support. GitHub Status — @githubstatus
# cihai &middot; [![Python Package](https://img.shields.io/pypi/v/cihai.svg)](https://pypi.org/project/cihai/) [![License](https://img.shields.io/github/license/cihai/cihai.svg)](https://github.com/cihai/cihai/blob/master/LICENSE) [![Code Coverage](https://codecov.io/gh/cihai/cihai/branch/master/graph/badge.svg)](https...
[ "Multilinguality", "Syntactic Text Processing" ]
[ "Annotation and Dataset Development" ]
true
https://github.com/SamuraiT/mecab-python3
2014-05-31T08:47:04Z
mecab-python. mecab-python. you can find original version here:http://taku910.github.io/mecab/
SamuraiT / mecab-python3 Public Branches Tags Go to file Go to file Code .github debian src/Me… test .gitattri… .gitignore AUTH… BSD COPYI… Docker… GPL LGPL MANIF… READ… setup.py tox.ini test-manylinux test-manylinux passing passing downloads downloads 745k/month 745k/month platforms platforms linux macosx windows l...
[![Current PyPI packages](https://badge.fury.io/py/mecab-python3.svg)](https://pypi.org/project/mecab-python3/) ![Test Status](https://github.com/SamuraiT/mecab-python3/workflows/test-manylinux/badge.svg) [![PyPI - Downloads](https://img.shields.io/pypi/dm/mecab-python3)](https://pypi.org/project/mecab-python3/) ![Supp...
[ "Morphology", "Syntactic Text Processing", "Tagging", "Text Segmentation" ]
[]
true
https://github.com/hakatashi/kyujitai.js
2014-09-06T08:05:01Z
Utility collections for making Japanese text old-fashioned
hakatashi / kyujitai.js Public 7 Branches 8 Tags Go to file Go to file Code hakatashi 1.3.0 8b627c7 · 4 years ago calibrate Remove BO… 10 years ago data Add a charact… 4 years ago dist Update dist 10 years ago lib Change pack… 9 years ago test Change test t… 10 years ago .gitattri… Add .gitattribu… 4 years ago .gitigno...
# kyujitai.js [![Build Status](https://travis-ci.org/hakatashi/kyujitai.js.svg)](https://travis-ci.org/hakatashi/kyujitai.js) [![Greenkeeper badge](https://badges.greenkeeper.io/hakatashi/kyujitai.js.svg)](https://greenkeeper.io/) Utility collections for making Japanese text old-fashioned. ## install npm instal...
[ "Low-Resource NLP", "Syntactic Text Processing" ]
[]
true
https://github.com/ikegami-yukino/rakutenma-python
2015-01-01T21:40:43Z
Rakuten MA (Python version)
ikegami-yukino / rakutenma-python Public Branches Tags Go to file Go to file Code bin rakutenma tests .gitignore .travis.yml CHANGES.rst LICENSE.txt MANIFEST.in README.rst setup.py travis-ci.org python python 2.6 | 2.7 | 3.3 | 3.4 | 3.5 | 3.6 2.6 | 2.7 | 3.3 | 3.4 | 3.5 | 3.6 pypi pypi v0.3.3 v0.3.3 Code Health li...
Rakuten MA Python =================== |travis| |coveralls| |pyversion| |version| |landscape| |license| Rakuten MA Python (morphological analyzer) is a Python version of Rakuten MA (word segmentor + PoS Tagger) for Chinese and Japanese. For details about Rakuten MA, See https://github.com/rakuten-nlp/rakutenma See...
[ "Morphology", "Syntactic Text Processing", "Tagging", "Text Segmentation" ]
[]
true
https://github.com/google/budoux
2015-03-18T18:22:31Z
Standalone. Small. Language-neutral. BudouX is the successor to Budou, the machine learning powered line break organizer tool.
google / budoux Public Branches Tags Go to file Go to file Code .github budoux data/finetuning/ja demo java javascript scripts tests .gitignore .markdownlint.yaml CONTRIBUTING.md LICENSE MANIFEST.in README.md bump_version.py example.png pyproject.toml setup.cfg setup.py pypi pypi v0.6.3 v0.6.3 npm npm v0.6.3 v0.6.3 m...
<!-- markdownlint-disable MD014 --> # BudouX [![PyPI](https://img.shields.io/pypi/v/budoux?color=blue)](https://pypi.org/project/budoux/) [![npm](https://img.shields.io/npm/v/budoux?color=yellow)](https://www.npmjs.com/package/budoux) [![Maven Central](https://img.shields.io/maven-central/v/com.google.budoux/budoux)](...
[ "Chunking", "Syntactic Text Processing", "Text Segmentation" ]
[]
true
https://github.com/scriptin/topokanji
2015-05-28T17:52:28Z
Topologically ordered lists of kanji for effective learning
scriptin / topokanji Public Branches Tags Go to file Go to file Code data depen… lib lists .gitignore .jshintrc READ… build.js packag… 30 seconds explanation for people who want to learn kanji: It is best to learn kanji starting from simple characters and then learning complex ones as compositions of "parts", which are...
# TopoKanji > **30 seconds explanation for people who want to learn kanji:** > > It is best to learn kanji starting from simple characters and then learning complex ones as compositions of "parts", which are called "radicals" or "components". For example: > > - 一 → 二 → 三 > - 丨 → 凵 → 山 → 出 > - 言 → 五 → 口 → 語 > > It is a...
[]
[ "Annotation and Dataset Development" ]
true
https://github.com/Kensuke-Mitsuzawa/JapaneseTokenizers
2015-09-01T10:24:45Z
A set of metrics for feature selection from text data
Kensuke-Mitsuzawa / JapaneseTokenizers Public Branches Tags Go to file Go to file Code Japan… exampl… test .gitignore .travis.… LICEN… MANIF… Makefile READ… install_… setup.py travis-… license license MIT MIT Build Status This is simple python-wrapper for Japanese Tokenizers(A.K.A Tokenizer) This project aims to call t...
[![MIT License](http://img.shields.io/badge/license-MIT-blue.svg?style=flat)](LICENSE)[![Build Status](https://travis-ci.org/Kensuke-Mitsuzawa/JapaneseTokenizers.svg?branch=master)](https://travis-ci.org/Kensuke-Mitsuzawa/JapaneseTokenizers) # What's this? This is simple python-wrapper for Japanese Tokenizers(A.K.A ...
[ "Morphology", "Responsible & Trustworthy NLP", "Robustness in NLP", "Syntactic Text Processing", "Tagging", "Text Segmentation" ]
[]
true
https://github.com/tokuhirom/akaza
2015-10-14T01:17:00Z
Yet another Japanese IME for IBus/Linux
akaza-im / akaza Public Branches Tags Go to file Go to file Code .github… .run akaza-… akaza-… akaza-… develo… docs ibus-ak… ibus-sys keymap libakaza marisa… romkan .gitattri… .gitignore .gitmo… CONT… Cargo.l… Cargo.… Chang… LICEN… About Yet another Japanese IME for IBus/Linux # nlp # rust # ime # ibus Readme MIT ...
# ibus-akaza Yet another kana-kanji-converter on IBus, written in Rust. 統計的かな漢字変換による日本語IMEです。 Rust で書いています。 **現在、開発途中のプロダクトです。非互換の変更が予告なくはいります** ## モチベーション いじりやすくて **ある程度** UIが使いやすいかな漢字変換があったら面白いなと思ったので作ってみています。 「いじりやすくて」というのはつまり、Hack-able であるという意味です。 モデルデータを自分で生成できて、特定の企業に依存しない自由なかな漢字変換エンジンを作りたい。 ## 特徴 * UI/Lo...
[ "Language Models", "Syntactic Text Processing" ]
[ "Vocabulary, Dictionary, and Language Input Method" ]
true
https://github.com/hexenq/kuroshiro
2016-01-03T09:16:40Z
Japanese language library for converting Japanese sentence to Hiragana, Katakana or Romaji with furigana and okurigana modes supported.
hexenq / kuroshiro Public 7 Branches 18 Tags Go to file Go to file Code src test .babelrc .eslintrc .gitignore .npmignore .travis.yml CHANGELOG.md CONTRIBUTING.md LICENSE README.eo-eo.md README.jp.md README.md README.zh-cn.md README.zh-tw.md package.json Build Status coverage coverage 96% 96% gitter gitter join chat...
![kuroshiro](https://kuroshiro.org/kuroshiro.png) # kuroshiro [![Build Status](https://travis-ci.org/hexenq/kuroshiro.svg?branch=master)](https://travis-ci.org/hexenq/kuroshiro) [![Coverage Status](https://coveralls.io/repos/hexenq/kuroshiro/badge.svg)](https://coveralls.io/r/hexenq/kuroshiro) [![npm version](https:/...
[ "Syntactic Text Processing", "Text Normalization" ]
[]
true
https://github.com/scriptin/kanji-frequency
2016-01-24T01:51:10Z
Kanji usage frequency data collected from various sources
scriptin / kanji-frequency Public Branches Tags Go to file Go to file Code .github… .vscode data data2015 public scripts src .editor… .gitignore .prettie… .prettie… CONT… LICEN… READ… astro.c… packag… packag… tailwin… tsconfi… About Kanji usage frequency data collected from various sources scriptin.github.io/kanji-freq...
# Kanji usage frequency Datasets built from various Japanese language corpora <https://scriptin.github.io/kanji-frequency/> - see this website for the dataset description. This readme describes only technical aspects. You can download the datasets here: <https://github.com/scriptin/kanji-frequency/tree/master/data> ...
[ "Information Extraction & Text Mining", "Structured Data in NLP", "Term Extraction" ]
[ "Annotation and Dataset Development" ]
End of preview. Expand in Data Studio

Dataset overview

This is a dataset for Japanese natural language processing with multi-label annotations of research field labels for GitHub repositories in the NLP domain. Please refer to this paper for the specific method of constructing the dataset. It is written in Japanese.

Input and Output

  • Input: Information from GitHub repositories (description, README text, PDF text, screenshot images)
  • Output: Multi-label classification of NLP research fields

Problem Setting of the Dataset:

  • Training Data: GitHub repositories from before 2022
  • Test Data: GitHub repositories from 2023
  • Objective: To predict multi-labels of research field labels in the NLP domain
  • We used the positive examples from awesome-japanese-nlp-classification-dataset.

Dataset Features:

  • A multimodal dataset including GitHub summaries, README files, PDF files, and screenshot images.
  • The annotation labels were assigned with reference to Exploring-NLP-Research.

If you need image data and PDF files, please submit a request to the community. This dataset uses text extracted from PDF files and does not include image data.

Based on GitHub's terms of service, please use this dataset for research purposes only.

How to use this dataset

Install the datasets library.

pip install datasets

How to load in Python.

from datasets import load_dataset

dataset = load_dataset("taishi-i/awesome-japanese-nlp-multilabel-dataset")

# To access the dataset sample
print(dataset["train"][0])

Here is a sample of the dataset.

{
  "is_exist": True,
  "url": "https://github.com/scriptin/jmdict-simplified",
  "created_at": "2016-02-07T16:34:32Z",
  "description": "JMdict and JMnedict in JSON format",
  "pdf_text": "scriptin / jmdict-simplified\nPublic\nBranches\nTags\n",
  "readme_text": "# jmdict-simplified\n\n**[JMdict][], [JMnedict][], [Kanjidic][], and [Kradfile/Radkfile][Kradfile] in JSON format**<br>\n",
  "nlp_taxonomy_classifier_labels": [
    "Multilinguality"
  ],
  "awesome_japanese_nlp_labels": [
    "Annotation and Dataset Development",
    "Vocabulary, Dictionary, and Language Input Method"
  ]
}
  • is_exist: A boolean value indicating whether the GitHub repository exists (True if it exists) at the time of retrieval.
  • url: The URL of the GitHub repository.
  • created_at: The timestamp indicating when the repository was created (in ISO 8601 format).
  • description: A short description of the repository provided on GitHub.
  • pdf_text: Extracted text from a PDF file that contains a saved version of the GitHub repository's top page.
  • readme_text: Extracted text from the repository's README file.
  • nlp_taxonomy_classifier_labels: Manually annotated multi-label annotations of NLP research field labels.
  • awesome_japanese_nlp_labels: Manually annotated multi-label annotations of research field labels specifically for Japanese natural language processing.

Details of the dataset.

DatasetDict({
    train: Dataset({
        features: ['is_exist', 'url', 'created_at', 'description', 'pdf_text', 'readme_text', 'nlp_taxonomy_classifier_labels', 'awesome_japanese_nlp_labels'],
        num_rows: 407
    })
    validation: Dataset({
        features: ['is_exist', 'url', 'created_at', 'description', 'pdf_text', 'readme_text', 'nlp_taxonomy_classifier_labels', 'awesome_japanese_nlp_labels'],
        num_rows: 17
    })
    test: Dataset({
        features: ['is_exist', 'url', 'created_at', 'description', 'pdf_text', 'readme_text', 'nlp_taxonomy_classifier_labels', 'awesome_japanese_nlp_labels'],
        num_rows: 60
    })
})

Baseline

The baseline model was used for TimSchopf/nlp_taxonomy_classifier.

The fine-tuned model was trained using this dataset to fine-tune the baseline model.

Classification Method Model Description Dev Prec. Dev Rec. Dev F1 Eval Prec. Eval Rec. Eval F1
Random Prediction - - 0.034 0.455 0.064 0.042 0.513 0.078
Baseline TimSchopf/nlp_taxonomy_classifier 0.538 0.382 0.447 0.360 0.354 0.349
Fine-Tuning TimSchopf/nlp_taxonomy_classifier 0.538 0.509 0.523 0.436 0.484 0.521
Zero-Shot gpt-4o-2024-08-06 0.560 0.255 0.350 0.476 0.184 0.265

License

We collect and publish this dataset under GitHub Acceptable Use Policies - 7. Information Usage Restrictions and GitHub Terms of Service - H. API Terms for research purposes. This dataset should be used solely for research verification purposes. Adhering to GitHub's regulations is mandatory.

Downloads last month
18