| { |
| "1706.03762": { |
| "arxivId": "1706.03762", |
| "title": "Attention is All you Need" |
| }, |
| "1810.04805": { |
| "arxivId": "1810.04805", |
| "title": "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding" |
| }, |
| "1405.0312": { |
| "arxivId": "1405.0312", |
| "title": "Microsoft COCO: Common Objects in Context" |
| }, |
| "2005.14165": { |
| "arxivId": "2005.14165", |
| "title": "Language Models are Few-Shot Learners" |
| }, |
| "1910.13461": { |
| "arxivId": "1910.13461", |
| "title": "BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension" |
| }, |
| "2203.02155": { |
| "arxivId": "2203.02155", |
| "title": "Training language models to follow instructions with human feedback" |
| }, |
| "2302.13971": { |
| "arxivId": "2302.13971", |
| "title": "LLaMA: Open and Efficient Foundation Language Models" |
| }, |
| "1804.07461": { |
| "arxivId": "1804.07461", |
| "title": "GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding" |
| }, |
| "1706.04599": { |
| "arxivId": "1706.04599", |
| "title": "On Calibration of Modern Neural Networks" |
| }, |
| "2204.02311": { |
| "arxivId": "2204.02311", |
| "title": "PaLM: Scaling Language Modeling with Pathways" |
| }, |
| "1904.09675": { |
| "arxivId": "1904.09675", |
| "title": "BERTScore: Evaluating Text Generation with BERT" |
| }, |
| "1610.02413": { |
| "arxivId": "1610.02413", |
| "title": "Equality of Opportunity in Supervised Learning" |
| }, |
| "2107.03374": { |
| "arxivId": "2107.03374", |
| "title": "Evaluating Large Language Models Trained on Code" |
| }, |
| "2108.07258": { |
| "arxivId": "2108.07258", |
| "title": "On the Opportunities and Risks of Foundation Models" |
| }, |
| "2205.01068": { |
| "arxivId": "2205.01068", |
| "title": "OPT: Open Pre-trained Transformer Language Models" |
| }, |
| "2009.03300": { |
| "arxivId": "2009.03300", |
| "title": "Measuring Massive Multitask Language Understanding" |
| }, |
| "1706.03741": { |
| "arxivId": "1706.03741", |
| "title": "Deep Reinforcement Learning from Human Preferences" |
| }, |
| "2303.12712": { |
| "arxivId": "2303.12712", |
| "title": "Sparks of Artificial General Intelligence: Early experiments with GPT-4" |
| }, |
| "2306.05685": { |
| "arxivId": "2306.05685", |
| "title": "Judging LLM-as-a-judge with MT-Bench and Chatbot Arena" |
| }, |
| "1905.00537": { |
| "arxivId": "1905.00537", |
| "title": "SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems" |
| }, |
| "1705.03551": { |
| "arxivId": "1705.03551", |
| "title": "TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension" |
| }, |
| "2206.07682": { |
| "arxivId": "2206.07682", |
| "title": "Emergent Abilities of Large Language Models" |
| }, |
| "2303.18223": { |
| "arxivId": "2303.18223", |
| "title": "A Survey of Large Language Models" |
| }, |
| "2012.15723": { |
| "arxivId": "2012.15723", |
| "title": "Making Pre-trained Language Models Better Few-shot Learners" |
| }, |
| "2212.13138": { |
| "arxivId": "2212.13138", |
| "title": "Large language models encode clinical knowledge" |
| }, |
| "2206.04615": { |
| "arxivId": "2206.04615", |
| "title": "Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models" |
| }, |
| "2201.08239": { |
| "arxivId": "2201.08239", |
| "title": "LaMDA: Language Models for Dialog Applications" |
| }, |
| "1909.08593": { |
| "arxivId": "1909.08593", |
| "title": "Fine-Tuning Language Models from Human Preferences" |
| }, |
| "2109.07958": { |
| "arxivId": "2109.07958", |
| "title": "TruthfulQA: Measuring How Models Mimic Human Falsehoods" |
| }, |
| "1708.08559": { |
| "arxivId": "1708.08559", |
| "title": "DeepTest: Automated Testing of Deep-Neural-Network-Driven Autonomous Cars" |
| }, |
| "2302.04761": { |
| "arxivId": "2302.04761", |
| "title": "Toolformer: Language Models Can Teach Themselves to Use Tools" |
| }, |
| "2109.00859": { |
| "arxivId": "2109.00859", |
| "title": "CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation" |
| }, |
| "2302.04023": { |
| "arxivId": "2302.04023", |
| "title": "A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and Interactivity" |
| }, |
| "2103.03874": { |
| "arxivId": "2103.03874", |
| "title": "Measuring Mathematical Problem Solving With the MATH Dataset" |
| }, |
| "2210.02414": { |
| "arxivId": "2210.02414", |
| "title": "GLM-130B: An Open Bilingual Pre-trained Model" |
| }, |
| "2005.04118": { |
| "arxivId": "2005.04118", |
| "title": "Beyond Accuracy: Behavioral Testing of NLP Models with CheckList" |
| }, |
| "2009.11462": { |
| "arxivId": "2009.11462", |
| "title": "RealToxicityPrompts: Evaluating Neural Toxic Degeneration in Language Models" |
| }, |
| "2103.03097": { |
| "arxivId": "2103.03097", |
| "title": "Generalizing to Unseen Domains: A Survey on Domain Generalization" |
| }, |
| "1910.14599": { |
| "arxivId": "1910.14599", |
| "title": "Adversarial NLI: A New Benchmark for Natural Language Understanding" |
| }, |
| "2203.13474": { |
| "arxivId": "2203.13474", |
| "title": "CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis" |
| }, |
| "2104.08663": { |
| "arxivId": "2104.08663", |
| "title": "BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models" |
| }, |
| "2302.11382": { |
| "arxivId": "2302.11382", |
| "title": "A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPT" |
| }, |
| "2201.11990": { |
| "arxivId": "2201.11990", |
| "title": "Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model" |
| }, |
| "2303.17580": { |
| "arxivId": "2303.17580", |
| "title": "HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in Hugging Face" |
| }, |
| "2211.01910": { |
| "arxivId": "2211.01910", |
| "title": "Large Language Models Are Human-Level Prompt Engineers" |
| }, |
| "2112.00861": { |
| "arxivId": "2112.00861", |
| "title": "A General Language Assistant as a Laboratory for Alignment" |
| }, |
| "2302.06476": { |
| "arxivId": "2302.06476", |
| "title": "Is ChatGPT a General-Purpose Natural Language Processing Task Solver?" |
| }, |
| "2207.05221": { |
| "arxivId": "2207.05221", |
| "title": "Language Models (Mostly) Know What They Know" |
| }, |
| "2307.06281": { |
| "arxivId": "2307.06281", |
| "title": "MMBench: Is Your Multi-modal Model an All-around Player?" |
| }, |
| "1705.08500": { |
| "arxivId": "1705.08500", |
| "title": "Selective Classification for Deep Neural Networks" |
| }, |
| "2105.09938": { |
| "arxivId": "2105.09938", |
| "title": "Measuring Coding Challenge Competence With APPS" |
| }, |
| "2306.13394": { |
| "arxivId": "2306.13394", |
| "title": "MME: A Comprehensive Evaluation Benchmark for Multimodal Large Language Models" |
| }, |
| "2305.01210": { |
| "arxivId": "2305.01210", |
| "title": "Is Your Code Generated by ChatGPT Really Correct? Rigorous Evaluation of Large Language Models for Code Generation" |
| }, |
| "2306.08302": { |
| "arxivId": "2306.08302", |
| "title": "Unifying Large Language Models and Knowledge Graphs: A Roadmap" |
| }, |
| "2302.14045": { |
| "arxivId": "2302.14045", |
| "title": "Language Is Not All You Need: Aligning Perception with Language Models" |
| }, |
| "1707.06875": { |
| "arxivId": "1707.06875", |
| "title": "Why We Need New Evaluation Metrics for NLG" |
| }, |
| "2008.02275": { |
| "arxivId": "2008.02275", |
| "title": "Aligning AI With Shared Human Values" |
| }, |
| "2307.16789": { |
| "arxivId": "2307.16789", |
| "title": "ToolLLM: Facilitating Large Language Models to Master 16000+ Real-world APIs" |
| }, |
| "2305.14387": { |
| "arxivId": "2305.14387", |
| "title": "AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback" |
| }, |
| "2305.10355": { |
| "arxivId": "2305.10355", |
| "title": "Evaluating Object Hallucination in Large Vision-Language Models" |
| }, |
| "2305.14251": { |
| "arxivId": "2305.14251", |
| "title": "FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long Form Text Generation" |
| }, |
| "2305.17926": { |
| "arxivId": "2305.17926", |
| "title": "Large Language Models are not Fair Evaluators" |
| }, |
| "2305.08322": { |
| "arxivId": "2305.08322", |
| "title": "C-Eval: A Multi-Level Multi-Discipline Chinese Evaluation Suite for Foundation Models" |
| }, |
| "2308.02490": { |
| "arxivId": "2308.02490", |
| "title": "MM-Vet: Evaluating Large Multimodal Models for Integrated Capabilities" |
| }, |
| "2304.06364": { |
| "arxivId": "2304.06364", |
| "title": "AGIEval: A Human-Centric Benchmark for Evaluating Foundation Models" |
| }, |
| "2303.04048": { |
| "arxivId": "2303.04048", |
| "title": "Is ChatGPT a Good NLG Evaluator? A Preliminary Study" |
| }, |
| "2309.01219": { |
| "arxivId": "2309.01219", |
| "title": "Siren's Song in the AI Ocean: A Survey on Hallucination in Large Language Models" |
| }, |
| "2104.14337": { |
| "arxivId": "2104.14337", |
| "title": "Dynabench: Rethinking Benchmarking in NLP" |
| }, |
| "2307.16125": { |
| "arxivId": "2307.16125", |
| "title": "SEED-Bench: Benchmarking Multimodal LLMs with Generative Comprehension" |
| }, |
| "2301.13867": { |
| "arxivId": "2301.13867", |
| "title": "Mathematical Capabilities of ChatGPT" |
| }, |
| "2101.11718": { |
| "arxivId": "2101.11718", |
| "title": "BOLD: Dataset and Metrics for Measuring Biases in Open-Ended Language Generation" |
| }, |
| "2306.11698": { |
| "arxivId": "2306.11698", |
| "title": "DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models" |
| }, |
| "2303.08896": { |
| "arxivId": "2303.08896", |
| "title": "SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models" |
| }, |
| "2307.04657": { |
| "arxivId": "2307.04657", |
| "title": "BeaverTails: Towards Improved Safety Alignment of LLM via a Human-Preference Dataset" |
| }, |
| "2304.05335": { |
| "arxivId": "2304.05335", |
| "title": "Toxicity in ChatGPT: Analyzing Persona-assigned Language Models" |
| }, |
| "2110.08193": { |
| "arxivId": "2110.08193", |
| "title": "BBQ: A hand-built bias benchmark for question answering" |
| }, |
| "2309.05922": { |
| "arxivId": "2309.05922", |
| "title": "A Survey of Hallucination in Large Foundation Models" |
| }, |
| "2207.08143": { |
| "arxivId": "2207.08143", |
| "title": "Can large language models reason about medical questions?" |
| }, |
| "2204.04991": { |
| "arxivId": "2204.04991", |
| "title": "TRUE: Re-evaluating Factual Consistency Evaluation" |
| }, |
| "2304.01852": { |
| "arxivId": "2304.01852", |
| "title": "Summary of ChatGPT/GPT-4 Research and Perspective Towards the Future of Large Language Models" |
| }, |
| "2304.05613": { |
| "arxivId": "2304.05613", |
| "title": "ChatGPT Beyond English: Towards a Comprehensive Evaluation of Large Language Models in Multilingual Learning" |
| }, |
| "2301.12867": { |
| "arxivId": "2301.12867", |
| "title": "Exploring AI Ethics of ChatGPT: A Diagnostic Analysis" |
| }, |
| "2309.12284": { |
| "arxivId": "2309.12284", |
| "title": "MetaMath: Bootstrap Your Own Mathematical Questions for Large Language Models" |
| }, |
| "2304.03439": { |
| "arxivId": "2304.03439", |
| "title": "Evaluating the Logical Reasoning Ability of ChatGPT and GPT-4" |
| }, |
| "2305.14975": { |
| "arxivId": "2305.14975", |
| "title": "Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback" |
| }, |
| "2304.09542": { |
| "arxivId": "2304.09542", |
| "title": "Is ChatGPT Good at Search? Investigating Large Language Models as Re-Ranking Agent" |
| }, |
| "2303.12528": { |
| "arxivId": "2303.12528", |
| "title": "MEGA: Multilingual Evaluation of Generative AI" |
| }, |
| "2210.07197": { |
| "arxivId": "2210.07197", |
| "title": "Towards a Unified Multi-Dimensional Evaluator for Text Generation" |
| }, |
| "2302.12095": { |
| "arxivId": "2302.12095", |
| "title": "On the Robustness of ChatGPT: An Adversarial and Out-of-distribution Perspective" |
| }, |
| "2304.03738": { |
| "arxivId": "2304.03738", |
| "title": "Should ChatGPT be Biased? Challenges and Risks of Bias in Large Language Models" |
| }, |
| "2307.02046": { |
| "arxivId": "2307.02046", |
| "title": "Recommender Systems in the Era of Large Language Models (LLMs)" |
| }, |
| "2305.09645": { |
| "arxivId": "2305.09645", |
| "title": "StructGPT: A General Framework for Large Language Model to Reason over Structured Data" |
| }, |
| "2111.02840": { |
| "arxivId": "2111.02840", |
| "title": "Adversarial GLUE: A Multi-Task Benchmark for Robustness Evaluation of Language Models" |
| }, |
| "2105.04054": { |
| "arxivId": "2105.04054", |
| "title": "Societal Biases in Language Generation: Progress and Challenges" |
| }, |
| "2304.08354": { |
| "arxivId": "2304.08354", |
| "title": "Tool Learning with Foundation Models" |
| }, |
| "2305.03514": { |
| "arxivId": "2305.03514", |
| "title": "Can Large Language Models Transform Computational Social Science?" |
| }, |
| "2203.03580": { |
| "arxivId": "2203.03580", |
| "title": "The Unsurprising Effectiveness of Pre-Trained Vision Models for Control" |
| }, |
| "2306.09212": { |
| "arxivId": "2306.09212", |
| "title": "CMMLU: Measuring massive multitask language understanding in Chinese" |
| }, |
| "2306.05087": { |
| "arxivId": "2306.05087", |
| "title": "PandaLM: An Automatic Evaluation Benchmark for LLM Instruction Tuning Optimization" |
| }, |
| "2305.15005": { |
| "arxivId": "2305.15005", |
| "title": "Sentiment Analysis in the Era of Large Language Models: A Reality Check" |
| }, |
| "2305.15771": { |
| "arxivId": "2305.15771", |
| "title": "On the Planning Abilities of Large Language Models - A Critical Investigation" |
| }, |
| "2303.09038": { |
| "arxivId": "2303.09038", |
| "title": "Translating radiology reports into plain language using ChatGPT and GPT-4 with prompt learning: results, limitations, and potential" |
| }, |
| "2305.18486": { |
| "arxivId": "2305.18486", |
| "title": "A Systematic Study and Comprehensive Evaluation of ChatGPT on Benchmark Datasets" |
| }, |
| "2301.01768": { |
| "arxivId": "2301.01768", |
| "title": "The political ideology of conversational AI: Converging evidence on ChatGPT's pro-environmental, left-libertarian orientation" |
| }, |
| "2103.06268": { |
| "arxivId": "2103.06268", |
| "title": "CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review" |
| }, |
| "2305.02182": { |
| "arxivId": "2305.02182", |
| "title": "Uncovering ChatGPT\u2019s Capabilities in Recommender Systems" |
| }, |
| "2306.14565": { |
| "arxivId": "2306.14565", |
| "title": "Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning" |
| }, |
| "2306.04528": { |
| "arxivId": "2306.04528", |
| "title": "PromptRobust: Towards Evaluating the Robustness of Large Language Models on Adversarial Prompts" |
| }, |
| "2307.02477": { |
| "arxivId": "2307.02477", |
| "title": "Reasoning or Reciting? Exploring the Capabilities and Limitations of Language Models Through Counterfactual Tasks" |
| }, |
| "2303.17466": { |
| "arxivId": "2303.17466", |
| "title": "Assessing Cross-Cultural Alignment between ChatGPT and Human Societies: An Empirical Study" |
| }, |
| "2305.07609": { |
| "arxivId": "2305.07609", |
| "title": "Is ChatGPT Fair for Recommendation? Evaluating Fairness in Large Language Model Recommendation" |
| }, |
| "2205.12615": { |
| "arxivId": "2205.12615", |
| "title": "Autoformalization with Large Language Models" |
| }, |
| "2205.12255": { |
| "arxivId": "2205.12255", |
| "title": "TALM: Tool Augmented Language Models" |
| }, |
| "2306.06687": { |
| "arxivId": "2306.06687", |
| "title": "LAMM: Language-Assisted Multi-Modal Instruction-Tuning Dataset, Framework, and Benchmark" |
| }, |
| "2304.04339": { |
| "arxivId": "2304.04339", |
| "title": "Is ChatGPT a Good Sentiment Analyzer? A Preliminary Study" |
| }, |
| "2310.03214": { |
| "arxivId": "2310.03214", |
| "title": "FreshLLMs: Refreshing Large Language Models with Search Engine Augmentation" |
| }, |
| "2306.09265": { |
| "arxivId": "2306.09265", |
| "title": "LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models" |
| }, |
| "2304.07619": { |
| "arxivId": "2304.07619", |
| "title": "Can ChatGPT Forecast Stock Price Movements? Return Predictability and Large Language Models" |
| }, |
| "2303.13835": { |
| "arxivId": "2303.13835", |
| "title": "Where to Go Next for Recommender Systems? ID- vs. Modality-based Recommender Models Revisited" |
| }, |
| "2309.07915": { |
| "arxivId": "2309.07915", |
| "title": "MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning" |
| }, |
| "2309.10691": { |
| "arxivId": "2309.10691", |
| "title": "MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback" |
| }, |
| "1804.02667": { |
| "arxivId": "1804.02667", |
| "title": "J-PLUS: The Javalambre Photometric Local Universe Survey" |
| }, |
| "2304.02015": { |
| "arxivId": "2304.02015", |
| "title": "How well do Large Language Models perform in Arithmetic tasks?" |
| }, |
| "2305.16934": { |
| "arxivId": "2305.16934", |
| "title": "On Evaluating Adversarial Robustness of Large Vision-Language Models" |
| }, |
| "2309.11998": { |
| "arxivId": "2309.11998", |
| "title": "LMSYS-Chat-1M: A Large-Scale Real-World LLM Conversation Dataset" |
| }, |
| "2304.01938": { |
| "arxivId": "2304.01938", |
| "title": "Evaluating large language models on a highly-specialized topic, radiation oncology physics" |
| }, |
| "2305.17306": { |
| "arxivId": "2305.17306", |
| "title": "Chain-of-Thought Hub: A Continuous Effort to Measure Large Language Models' Reasoning Performance" |
| }, |
| "2306.04181": { |
| "arxivId": "2306.04181", |
| "title": "Benchmarking Foundation Models with Language-Model-as-an-Examiner" |
| }, |
| "2307.00184": { |
| "arxivId": "2307.00184", |
| "title": "Personality Traits in Large Language Models" |
| }, |
| "2304.02210": { |
| "arxivId": "2304.02210", |
| "title": "Document-Level Machine Translation with Large Language Models" |
| }, |
| "2306.05715": { |
| "arxivId": "2306.05715", |
| "title": "Exploring the Responses of Large Language Models to Beginner Programmers\u2019 Help Requests" |
| }, |
| "2305.13711": { |
| "arxivId": "2305.13711", |
| "title": "LLM-Eval: Unified Multi-Dimensional Automatic Evaluation for Open-Domain Conversations with Large Language Models" |
| }, |
| "2304.07333": { |
| "arxivId": "2304.07333", |
| "title": "The Self-Perception and Political Biases of ChatGPT" |
| }, |
| "2303.16421": { |
| "arxivId": "2303.16421", |
| "title": "ChatGPT Is a Knowledgeable but Inexperienced Solver: An Investigation of Commonsense Problem in Large Language Models" |
| }, |
| "2308.01862": { |
| "arxivId": "2308.01862", |
| "title": "Wider and Deeper LLM Networks are Fairer LLM Evaluators" |
| }, |
| "2211.08073": { |
| "arxivId": "2211.08073", |
| "title": "GLUE-X: Evaluating Natural Language Understanding Models from an Out-of-distribution Generalization Perspective" |
| }, |
| "2305.12474": { |
| "arxivId": "2305.12474", |
| "title": "Evaluating the Performance of Large Language Models on GAOKAO Benchmark" |
| }, |
| "2307.09705": { |
| "arxivId": "2307.09705", |
| "title": "CValues: Measuring the Values of Chinese Large Language Models from Safety to Responsibility" |
| }, |
| "2306.05179": { |
| "arxivId": "2306.05179", |
| "title": "M3Exam: A Multilingual, Multimodal, Multilevel Benchmark for Examining Large Language Models" |
| }, |
| "2302.06706": { |
| "arxivId": "2302.06706", |
| "title": "On the Planning Abilities of Large Language Models (A Critical Investigation with a Proposed Benchmark)" |
| }, |
| "2205.00445": { |
| "arxivId": "2205.00445", |
| "title": "MRKL Systems: A modular, neuro-symbolic architecture that combines large language models, external knowledge sources and discrete reasoning" |
| }, |
| "2305.15269": { |
| "arxivId": "2305.15269", |
| "title": "Testing the General Deductive Reasoning Capacity of Large Language Models Using OOD Examples" |
| }, |
| "2305.11171": { |
| "arxivId": "2305.11171", |
| "title": "TrueTeacher: Learning Factual Consistency Evaluation with Large Language Models" |
| }, |
| "2304.00723": { |
| "arxivId": "2304.00723", |
| "title": "Exploring the Use of Large Language Models for Reference-Free Text Quality Evaluation: A Preliminary Empirical Study" |
| }, |
| "2106.06052": { |
| "arxivId": "2106.06052", |
| "title": "Dynaboard: An Evaluation-As-A-Service Platform for Holistic Next-Generation Benchmarking" |
| }, |
| "2309.07045": { |
| "arxivId": "2309.07045", |
| "title": "SafetyBench: Evaluating the Safety of Large Language Models with Multiple Choice Questions" |
| }, |
| "2308.08833": { |
| "arxivId": "2308.08833", |
| "title": "CMB: A Comprehensive Medical Benchmark in Chinese" |
| }, |
| "2305.14938": { |
| "arxivId": "2305.14938", |
| "title": "Do LLMs Understand Social Knowledge? Evaluating the Sociability of Large Language Models with SocKET Benchmark" |
| }, |
| "2306.07799": { |
| "arxivId": "2306.07799", |
| "title": "ChatGPT vs Human-authored Text: Insights into Controllable Text Summarization and Sentence Style Transfer" |
| }, |
| "2306.09296": { |
| "arxivId": "2306.09296", |
| "title": "KoLA: Carefully Benchmarking World Knowledge of Large Language Models" |
| }, |
| "2306.04757": { |
| "arxivId": "2306.04757", |
| "title": "InstructEval: Towards Holistic Evaluation of Instruction-Tuned Large Language Models" |
| }, |
| "2306.09841": { |
| "arxivId": "2306.09841", |
| "title": "Are Large Language Models Really Good Logical Reasoners? A Comprehensive Evaluation From Deductive, Inductive and Abductive Views" |
| }, |
| "2307.09042": { |
| "arxivId": "2307.09042", |
| "title": "Emotional intelligence of Large Language Models" |
| }, |
| "2306.01248": { |
| "arxivId": "2306.01248", |
| "title": "How Ready are Pre-trained Abstractive Models and LLMs for Legal Case Judgement Summarization?" |
| }, |
| "2306.05783": { |
| "arxivId": "2306.05783", |
| "title": "Xiezhi: An Ever-Updating Benchmark for Holistic Domain Knowledge Evaluation" |
| }, |
| "2306.03090": { |
| "arxivId": "2306.03090", |
| "title": "Is ChatGPT a Good Teacher Coach? Measuring Zero-Shot Performance For Scoring and Providing Actionable Insights on Classroom Instruction" |
| }, |
| "2301.12868": { |
| "arxivId": "2301.12868", |
| "title": "On Robustness of Prompt-based Semantic Parsing with Large Pre-trained Language Model: An Empirical Study on Codex" |
| }, |
| "2209.12106": { |
| "arxivId": "2209.12106", |
| "title": "Moral Mimicry: Large Language Models Produce Moral Rationalizations Tailored to Political Identity" |
| }, |
| "2306.04618": { |
| "arxivId": "2306.04618", |
| "title": "Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations" |
| }, |
| "2306.01337": { |
| "arxivId": "2306.01337", |
| "title": "MathChat: Converse to Tackle Challenging Math Problems with LLM Agents" |
| }, |
| "2306.07075": { |
| "arxivId": "2306.07075", |
| "title": "Large language models as tax attorneys: a case study in legal capabilities emergence" |
| }, |
| "2305.11700": { |
| "arxivId": "2305.11700", |
| "title": "Exploring the Upper Limits of Text-Based Collaborative Filtering Using Large Language Models: Discoveries and Insights" |
| }, |
| "2305.18365": { |
| "arxivId": "2305.18365", |
| "title": "What indeed can GPT models do in chemistry? A comprehensive benchmark on eight tasks" |
| }, |
| "2305.15074": { |
| "arxivId": "2305.15074", |
| "title": "Have LLMs Advanced Enough? A Challenging Problem Solving Benchmark For Large Language Models" |
| }, |
| "2303.02155": { |
| "arxivId": "2303.02155", |
| "title": "ChatGPT and Other Large Language Models as Evolutionary Engines for Online Interactive Collaborative Game Design" |
| }, |
| "2301.11596": { |
| "arxivId": "2301.11596", |
| "title": "ThoughtSource: A central hub for large language model reasoning data" |
| }, |
| "2306.11507": { |
| "arxivId": "2306.11507", |
| "title": "TrustGPT: A Benchmark for Trustworthy and Responsible Large Language Models" |
| }, |
| "2305.16151": { |
| "arxivId": "2305.16151", |
| "title": "Understanding the Capabilities of Large Language Models for Automated Planning" |
| }, |
| "2308.03656": { |
| "arxivId": "2308.03656", |
| "title": "Emotionally Numb or Empathetic? Evaluating How LLMs Feel Using EmotionBench" |
| }, |
| "2306.01694": { |
| "arxivId": "2306.01694", |
| "title": "Evaluating Language Models for Mathematics through Interactions" |
| }, |
| "2309.09150": { |
| "arxivId": "2309.09150", |
| "title": "Can Large Language Models Understand Real-World Complex Instructions?" |
| }, |
| "2303.07142": { |
| "arxivId": "2303.07142", |
| "title": "Large Language Models in the Workplace: A Case Study on Prompt Engineering for Job Type Classification" |
| }, |
| "2205.09148": { |
| "arxivId": "2205.09148", |
| "title": "DDXPlus: A New Dataset For Automatic Medical Diagnosis" |
| }, |
| "2307.13692": { |
| "arxivId": "2307.13692", |
| "title": "ARB: Advanced Reasoning Benchmark for Large Language Models" |
| }, |
| "2305.16837": { |
| "arxivId": "2305.16837", |
| "title": "ChatGPT: A Study on its Utility for Ubiquitous Software Engineering Tasks" |
| }, |
| "2306.10512": { |
| "arxivId": "2306.10512", |
| "title": "From Static Benchmarks to Adaptive Testing: Psychometrics in AI Evaluation" |
| }, |
| "2301.12307": { |
| "arxivId": "2301.12307", |
| "title": "MQAG: Multiple-choice Question Answering and Generation for Assessing Information Consistency in Summarization" |
| }, |
| "2212.02774": { |
| "arxivId": "2212.02774", |
| "title": "Adaptive Testing of Computer Vision Models" |
| }, |
| "2306.02408": { |
| "arxivId": "2306.02408", |
| "title": "Evaluating and Improving Tool-Augmented Computation-Intensive Math Reasoning" |
| }, |
| "2307.01135": { |
| "arxivId": "2307.01135", |
| "title": "ChatGPT vs. Google: A Comparative Study of Search Performance and User Experience" |
| }, |
| "2304.00228": { |
| "arxivId": "2304.00228", |
| "title": "Accuracy and Political Bias of News Source Credibility Ratings by Large Language Models" |
| }, |
| "2306.16636": { |
| "arxivId": "2306.16636", |
| "title": "CMATH: Can Your Language Model Pass Chinese Elementary School Math Test?" |
| }, |
| "2306.04504": { |
| "arxivId": "2306.04504", |
| "title": "Evaluation of ChatGPT on Biomedical Tasks: A Zero-Shot Comparison with Fine-Tuned Generative Transformers" |
| }, |
| "2306.04308": { |
| "arxivId": "2306.04308", |
| "title": "Personality testing of GPT-3: Limited temporal reliability, but highlighted social desirability of GPT-3's personality instruments results" |
| }, |
| "2305.10263": { |
| "arxivId": "2305.10263", |
| "title": "M3KE: A Massive Multi-Level Multi-Subject Knowledge Evaluation Benchmark for Chinese Large Language Models" |
| }, |
| "2305.01181": { |
| "arxivId": "2305.01181", |
| "title": "A Paradigm Shift: The Future of Machine Translation Lies with Large Language Models" |
| }, |
| "2306.13651": { |
| "arxivId": "2306.13651", |
| "title": "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models" |
| }, |
| "2306.06331": { |
| "arxivId": "2306.06331", |
| "title": "Investigating the Effectiveness of ChatGPT in Mathematical Reasoning and Problem Solving: Evidence from the Vietnamese National High School Graduation Examination" |
| }, |
| "2305.11792": { |
| "arxivId": "2305.11792", |
| "title": "Chain-of-thought prompting for responding to in-depth dialogue questions with LLM" |
| }, |
| "2306.04563": { |
| "arxivId": "2306.04563", |
| "title": "ChatGPT is fun, but it is not funny! Humor is still challenging Large Language Models" |
| }, |
| "2306.07622": { |
| "arxivId": "2306.07622", |
| "title": "Human-Like Intuitive Behavior and Reasoning Biases Emerged in Language Models - and Disappeared in GPT-4" |
| }, |
| "2303.12057": { |
| "arxivId": "2303.12057", |
| "title": "Large Language Models Can Be Used to Estimate the Ideologies of Politicians in a Zero-Shot Learning Setting" |
| }, |
| "2304.01457": { |
| "arxivId": "2304.01457", |
| "title": "Exploring Vision-Language Models for Imbalanced Learning" |
| }, |
| "2307.00112": { |
| "arxivId": "2307.00112", |
| "title": "Performance of ChatGPT on USMLE: Unlocking the Potential of Large Language Models for AI-Assisted Medical Education" |
| }, |
| "2306.08997": { |
| "arxivId": "2306.08997", |
| "title": "Exploring the MIT Mathematics and EECS Curriculum Using Large Language Models" |
| }, |
| "2305.14693": { |
| "arxivId": "2305.14693", |
| "title": "Have Large Language Models Developed a Personality?: Applicability of Self-Assessment Tests in Measuring Personality in LLMs" |
| }, |
| "2311.15296": { |
| "arxivId": "2311.15296", |
| "title": "UHGEval: Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation" |
| }, |
| "2305.11262": { |
| "arxivId": "2305.11262", |
| "title": "CHBias: Bias Evaluation and Mitigation of Chinese Conversational Language Models" |
| }, |
| "2306.01590": { |
| "arxivId": "2306.01590", |
| "title": "An Evaluation of Log Parsing with ChatGPT" |
| }, |
| "2306.15261": { |
| "arxivId": "2306.15261", |
| "title": "A Survey on Out-of-Distribution Evaluation of Neural NLP Models" |
| }, |
| "2306.06264": { |
| "arxivId": "2306.06264", |
| "title": "Measuring and Modifying Factual Knowledge in Large Language Models" |
| }, |
| "2306.02864": { |
| "arxivId": "2306.02864", |
| "title": "Leveraging Large Language Models for Topic Classification in the Domain of Public Affairs" |
| }, |
| "2302.12297": { |
| "arxivId": "2302.12297", |
| "title": "Dynamic Benchmarking of Masked Language Models on Temporal Concept Drift with Multiple Views" |
| }, |
| "2306.02549": { |
| "arxivId": "2306.02549", |
| "title": "Evaluation of AI Chatbots for Patient-Specific EHR Questions" |
| }, |
| "2204.01906": { |
| "arxivId": "2204.01906", |
| "title": "Dynatask: A Framework for Creating Dynamic AI Benchmark Tasks" |
| }, |
| "2306.01499": { |
| "arxivId": "2306.01499", |
| "title": "Can LLMs like GPT-4 outperform traditional AI tools in dementia diagnosis? Maybe, but not today" |
| }, |
| "2304.07849": { |
| "arxivId": "2304.07849", |
| "title": "ChatPLUG: Open-Domain Generative Dialogue System with Internet-Augmented Instruction Tuning for Digital Human" |
| }, |
| "2111.08181": { |
| "arxivId": "2111.08181", |
| "title": "Adversarially Constructed Evaluation Sets Are More Challenging, but May Not Be Fair" |
| }, |
| "2305.12421": { |
| "arxivId": "2305.12421", |
| "title": "Evaluating Open Question Answering Evaluation" |
| }, |
| "2310.02174": { |
| "arxivId": "2310.02174", |
| "title": "Ask Again, Then Fail: Large Language Models' Vacillations in Judgement" |
| }, |
| "2309.11737": { |
| "arxivId": "2309.11737", |
| "title": "Choice-75: A Dataset on Decision Branching in Script Learning" |
| }, |
| "2305.15268": { |
| "arxivId": "2305.15268", |
| "title": "EvEval: A Comprehensive Evaluation of Event Semantics for Large Language Models" |
| }, |
| "2306.04926": { |
| "arxivId": "2306.04926", |
| "title": "covLLM: Large Language Models for COVID-19 Biomedical Literature" |
| }, |
| "2306.04610": { |
| "arxivId": "2306.04610", |
| "title": "The Two Word Test: A Semantic Benchmark for Large Language Models" |
| } |
| } |