| import gradio as gr |
| import rebiber |
| import os |
| import uuid |
| |
|
|
| |
| filepath = os.path.abspath(rebiber.__file__).replace("__init__.py","") |
| bib_list_path = os.path.join(filepath, "bib_list.txt") |
| abbr_tsv_path = "abbr.tsv" |
|
|
| bib_db = rebiber.construct_bib_db(bib_list_path, start_dir=filepath) |
|
|
| abbr_dict = rebiber.normalize.load_abbr_tsv(abbr_tsv_path) |
|
|
|
|
| def process(input_bib, shorten, remove_keys, deduplicate, sort): |
| if "@" not in input_bib: |
| return "N/A" |
| global abbr_dict |
| |
| random_id = uuid.uuid4().hex |
| with open(f"input_{random_id}.bib", "w") as f: |
| f.write(input_bib.replace("\t", " ")) |
| all_bib_entries = rebiber.load_bib_file(f"input_{random_id}.bib") |
| print("# Input Bib Entries:", len(all_bib_entries)) |
| abbr_dict_pass = [] |
| if shorten: |
| abbr_dict_pass = abbr_dict |
| rebiber.normalize_bib(bib_db, all_bib_entries, f"output_{random_id}.bib", |
| abbr_dict=abbr_dict_pass, |
| deduplicate=deduplicate, |
| sort=sort, |
| removed_value_names=remove_keys) |
| with open(f"output_{random_id}.bib") as f: |
| output_bib = f.read().replace("\n ", "\n ") |
| |
| |
| return output_bib, random_id, gr.update(visible=True) |
|
|
|
|
| example_input = """ |
| @article{lin2020birds, |
| title={Birds have four legs?! NumerSense: Probing Numerical Commonsense Knowledge of Pre-trained Language Models}, |
| author={Lin, Bill Yuchen and Lee, Seyeon and Khanna, Rahul and Ren, Xiang}, |
| journal={arXiv preprint arXiv:2005.00683}, |
| year={2020} |
| } |
| @inproceedings{Lin2020CommonGenAC, |
| title={CommonGen: A Constrained Text Generation Challenge for Generative Commonsense Reasoning}, |
| author={Bill Yuchen Lin and Minghan Shen and Wangchunshu Zhou and Pei Zhou and Chandra Bhagavatula and Yejin Choi and Xiang Ren}, |
| booktitle={Findings}, |
| year={2020} |
| } |
| """ |
|
|
| examples = [[example_input]] |
|
|
|
|
| |
| |
| |
| |
| |
| |
|
|
|
|
|
|
|
|
|
|
| with gr.Blocks() as demo: |
| |
| gr.Markdown( |
| '''# Rebiber: A tool for normalizing bibtex with official info. |
| <table> |
| <tr> |
| <td> |
| <a href="https://yuchenlin.xyz/"> |
| <img src="https://img.shields.io/badge/Yuchen%20Lin-🐼-blue?style=social"> |
| </a> |
| </td> |
| <td> |
| <a href="https://github.com/yuchenlin/rebiber"> |
| <img src="https://img.shields.io/badge/Github--blue?style=social&logo=github"> |
| </a> |
| </td> |
| <td> |
| <a href="https://twitter.com/billyuchenlin/status/1353850378438070272?s=20"> |
| <img src="https://img.shields.io/badge/Tweet--blue?style=social&logo=twitter"> |
| </a> |
| </td> |
| </tr> |
| </table> |
| <span style="font-size:13pt"> |
| |
| We often cite papers using their arXiv versions without noting that they are already __PUBLISHED__ in some conferences. These unofficial bib entries might violate rules about submissions or camera-ready versions for some conferences. |
| We introduce __Rebiber__, a simple tool in Python to fix them automatically. It is based on the official conference information from the [DBLP](https://dblp.org/) or [the ACL anthology](https://www.aclweb.org/anthology/) (for NLP conferences)! |
| Apart from handling outdated arXiv citations, __Rebiber__ also normalizes citations in a unified way (DBLP-style), supporting abbreviation and value selection. |
| |
| </span> |
| ''' |
| ) |
| |
| with gr.Row(): |
| with gr.Column(scale=3): |
| input_bib = gr.Textbox(lines=15, label="Input BIB", value=example_input, interactive=True) |
| removekeys = gr.CheckboxGroup(["url", "biburl", "address", "publisher", "pages", "doi", "volume", "bibsource"], |
| value=[False, False, False, False, False, False, False, False], |
| label="Remove Keys", info="Which keys to remove?") |
| shorten = gr.Checkbox(label="Abbreviation", info="Shorten the conference/journal names (e.g., `Proceedings of the 2020 International Conference of ...` --> `Proc. of ICML')", value=False) |
| dedup = gr.Checkbox(label="Deduplicate entries.", value=False) |
| sort = gr.Checkbox(label="Sort alphabetically by ID.", value=False) |
| with gr.Row(): |
| clr_button = gr.Button("Clear") |
| button = gr.Button("Submit") |
| ex_uuid = gr.Text(label="UUID") |
| ex_uuid.visible = False |
| with gr.Column(scale=3): |
| output=gr.Textbox(label="Output BIB (Note that you can copy the output bib file by clicking the top-right button.)").style(show_copy_button=True, interactive=False) |
| download_btn = gr.Button("Generate Bib File") |
| download_btn.visible = False |
| download_content = gr.outputs.File() |
| download_content.visible = False |
| def download_file(ex_uuid): |
| global download_content |
| |
| file_path = f"output_{ex_uuid}.bib" |
| download_content.update(visible=False) |
| return file_path, gr.update(visible=True) |
| download_btn.click(download_file, inputs=ex_uuid, outputs=[download_content,download_content]) |
| button.click(process, inputs=[input_bib, shorten, removekeys, dedup, sort], outputs=[output, ex_uuid, download_btn], api_name = "process") |
| def clean(text): |
| return "" |
| clr_button.click(clean, input_bib, input_bib) |
| |
| |
| |
| |
| |
| |
| |
|
|
|
|
| if __name__ == "__main__": |
| demo.launch() |
|
|
|
|
| """ |
| @article{lin2020birds, |
| title={Birds have four legs?! NumerSense: Probing Numerical Commonsense Knowledge of Pre-trained Language Models}, |
| author={Lin, Bill Yuchen and Lee, Seyeon and Khanna, Rahul and Ren, Xiang}, |
| journal={arXiv preprint arXiv:2005.00683}, |
| year={2020} |
| } |
| |
| @inproceedings{lin2020birds, |
| address = {Online}, |
| author = {Lin, Bill Yuchen and |
| Lee, Seyeon and |
| Khanna, Rahul and |
| Ren, Xiang}, |
| booktitle = {Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)}, |
| doi = {10.18653/v1/2020.emnlp-main.557}, |
| pages = {6862--6868}, |
| publisher = {Association for Computational Linguistics}, |
| title = {{B}irds have four legs?! {N}umer{S}ense: {P}robing {N}umerical {C}ommonsense {K}nowledge of {P}re-{T}rained {L}anguage {M}odels}, |
| url = {https://aclanthology.org/2020.emnlp-main.557}, |
| year = {2020} |
| } |
| """ |