id stringlengths 40 40 | pid stringlengths 42 42 | input stringlengths 8.37k 169k | output stringlengths 1 1.63k |
|---|---|---|---|
6548db45fc28e8a8b51f114635bad14a13eaec5b | 6548db45fc28e8a8b51f114635bad14a13eaec5b_0 | Q: Which GAN do they use?
Text: Introduction
Generative adversarial nets (GAN) (Goodfellow et al., 2014) belong to a class of generative models which are trainable and can generate artificial data examples similar to the existing ones. In a GAN model, there are two sub-models simultaneously trained: a generative model ... | We construct a GAN model which combines different sets of word embeddings INLINEFORM4 , INLINEFORM5 , into a single set of word embeddings INLINEFORM6 . |
6548db45fc28e8a8b51f114635bad14a13eaec5b | 6548db45fc28e8a8b51f114635bad14a13eaec5b_1 | Q: Which GAN do they use?
Text: Introduction
Generative adversarial nets (GAN) (Goodfellow et al., 2014) belong to a class of generative models which are trainable and can generate artificial data examples similar to the existing ones. In a GAN model, there are two sub-models simultaneously trained: a generative model ... | weGAN, deGAN |
4c4f76837d1329835df88b0921f4fe8bda26606f | 4c4f76837d1329835df88b0921f4fe8bda26606f_0 | Q: Do they evaluate grammaticality of generated text?
Text: Introduction
Generative adversarial nets (GAN) (Goodfellow et al., 2014) belong to a class of generative models which are trainable and can generate artificial data examples similar to the existing ones. In a GAN model, there are two sub-models simultaneously ... | No |
819d2e97f54afcc7cdb3d894a072bcadfba9b747 | 819d2e97f54afcc7cdb3d894a072bcadfba9b747_0 | Q: Which corpora do they use?
Text: Introduction
Generative adversarial nets (GAN) (Goodfellow et al., 2014) belong to a class of generative models which are trainable and can generate artificial data examples similar to the existing ones. In a GAN model, there are two sub-models simultaneously trained: a generative mo... | CNN, TIME, 20 Newsgroups, and Reuters-21578 |
637aa32a34b20b4b0f1b5dfa08ef4e0e5ed33d52 | 637aa32a34b20b4b0f1b5dfa08ef4e0e5ed33d52_0 | Q: Do they report results only on English datasets?
Text: Introduction
Understanding a user's intent and sentiment is of utmost importance for current intelligent chatbots to respond appropriately to human requests. However, current systems are not able to perform to their best capacity when presented with incomplete d... | Yes |
4b8257cdd9a60087fa901da1f4250e7d910896df | 4b8257cdd9a60087fa901da1f4250e7d910896df_0 | Q: How do the authors define or exemplify 'incorrect words'?
Text: Introduction
Understanding a user's intent and sentiment is of utmost importance for current intelligent chatbots to respond appropriately to human requests. However, current systems are not able to perform to their best capacity when presented with inc... | typos in spellings or ungrammatical words |
7e161d9facd100544fa339b06f656eb2fc64ed28 | 7e161d9facd100544fa339b06f656eb2fc64ed28_0 | Q: How many vanilla transformers do they use after applying an embedding layer?
Text: Introduction
Understanding a user's intent and sentiment is of utmost importance for current intelligent chatbots to respond appropriately to human requests. However, current systems are not able to perform to their best capacity when... | Unanswerable |
abc5836c54fc2ac8465aee5a83b9c0f86c6fd6f5 | abc5836c54fc2ac8465aee5a83b9c0f86c6fd6f5_0 | Q: Do they test their approach on a dataset without incomplete data?
Text: Introduction
Understanding a user's intent and sentiment is of utmost importance for current intelligent chatbots to respond appropriately to human requests. However, current systems are not able to perform to their best capacity when presented ... | No |
abc5836c54fc2ac8465aee5a83b9c0f86c6fd6f5 | abc5836c54fc2ac8465aee5a83b9c0f86c6fd6f5_1 | Q: Do they test their approach on a dataset without incomplete data?
Text: Introduction
Understanding a user's intent and sentiment is of utmost importance for current intelligent chatbots to respond appropriately to human requests. However, current systems are not able to perform to their best capacity when presented ... | No |
4debd7926941f1a02266b1a7be2df8ba6e79311a | 4debd7926941f1a02266b1a7be2df8ba6e79311a_0 | Q: Should their approach be applied only when dealing with incomplete data?
Text: Introduction
Understanding a user's intent and sentiment is of utmost importance for current intelligent chatbots to respond appropriately to human requests. However, current systems are not able to perform to their best capacity when pre... | No |
4debd7926941f1a02266b1a7be2df8ba6e79311a | 4debd7926941f1a02266b1a7be2df8ba6e79311a_1 | Q: Should their approach be applied only when dealing with incomplete data?
Text: Introduction
Understanding a user's intent and sentiment is of utmost importance for current intelligent chatbots to respond appropriately to human requests. However, current systems are not able to perform to their best capacity when pre... | No |
3b745f086fb5849e7ce7ce2c02ccbde7cfdedda5 | 3b745f086fb5849e7ce7ce2c02ccbde7cfdedda5_0 | Q: By how much do they outperform other models in the sentiment in intent classification tasks?
Text: Introduction
Understanding a user's intent and sentiment is of utmost importance for current intelligent chatbots to respond appropriately to human requests. However, current systems are not able to perform to their be... | In the sentiment classification task by 6% to 8% and in the intent classification task by 0.94% on average |
44c7c1fbac80eaea736622913d65fe6453d72828 | 44c7c1fbac80eaea736622913d65fe6453d72828_0 | Q: What is the sample size of people used to measure user satisfaction?
Text: Introduction
Amazon Alexa Prize BIBREF0 provides a platform to collect real human-machine conversation data and evaluate performance on speech-based social conversational systems. Our system, Gunrock BIBREF1 addresses several limitations of p... | 34,432 user conversations |
44c7c1fbac80eaea736622913d65fe6453d72828 | 44c7c1fbac80eaea736622913d65fe6453d72828_1 | Q: What is the sample size of people used to measure user satisfaction?
Text: Introduction
Amazon Alexa Prize BIBREF0 provides a platform to collect real human-machine conversation data and evaluate performance on speech-based social conversational systems. Our system, Gunrock BIBREF1 addresses several limitations of p... | 34,432 |
3e0c9469821cb01a75e1818f2acb668d071fcf40 | 3e0c9469821cb01a75e1818f2acb668d071fcf40_0 | Q: What are all the metrics to measure user engagement?
Text: Introduction
Amazon Alexa Prize BIBREF0 provides a platform to collect real human-machine conversation data and evaluate performance on speech-based social conversational systems. Our system, Gunrock BIBREF1 addresses several limitations of prior chatbots BI... | overall rating, mean number of turns |
3e0c9469821cb01a75e1818f2acb668d071fcf40 | 3e0c9469821cb01a75e1818f2acb668d071fcf40_1 | Q: What are all the metrics to measure user engagement?
Text: Introduction
Amazon Alexa Prize BIBREF0 provides a platform to collect real human-machine conversation data and evaluate performance on speech-based social conversational systems. Our system, Gunrock BIBREF1 addresses several limitations of prior chatbots BI... | overall rating, mean number of turns |
a725246bac4625e6fe99ea236a96ccb21b5f30c6 | a725246bac4625e6fe99ea236a96ccb21b5f30c6_0 | Q: What the system designs introduced?
Text: Introduction
Amazon Alexa Prize BIBREF0 provides a platform to collect real human-machine conversation data and evaluate performance on speech-based social conversational systems. Our system, Gunrock BIBREF1 addresses several limitations of prior chatbots BIBREF2, BIBREF3, B... | Amazon Conversational Bot Toolkit, natural language understanding (NLU) (nlu) module, dialog manager, knowledge bases, natural language generation (NLG) (nlg) module, text to speech (TTS) (tts) |
516626825e51ca1e8a3e0ac896c538c9d8a747c8 | 516626825e51ca1e8a3e0ac896c538c9d8a747c8_0 | Q: Do they specify the model they use for Gunrock?
Text: Introduction
Amazon Alexa Prize BIBREF0 provides a platform to collect real human-machine conversation data and evaluate performance on speech-based social conversational systems. Our system, Gunrock BIBREF1 addresses several limitations of prior chatbots BIBREF2... | No |
77af93200138f46bb178c02f710944a01ed86481 | 77af93200138f46bb178c02f710944a01ed86481_0 | Q: Do they gather explicit user satisfaction data on Gunrock?
Text: Introduction
Amazon Alexa Prize BIBREF0 provides a platform to collect real human-machine conversation data and evaluate performance on speech-based social conversational systems. Our system, Gunrock BIBREF1 addresses several limitations of prior chatb... | Yes |
71538776757a32eee930d297f6667cd0ec2e9231 | 71538776757a32eee930d297f6667cd0ec2e9231_0 | Q: How do they correlate user backstory queries to user satisfaction?
Text: Introduction
Amazon Alexa Prize BIBREF0 provides a platform to collect real human-machine conversation data and evaluate performance on speech-based social conversational systems. Our system, Gunrock BIBREF1 addresses several limitations of pri... | modeled the relationship between word count and the two metrics of user engagement (overall rating, mean number of turns) in separate linear regressions |
7aa8375cdf4690fc3b9b1799b0f5a9ec1c1736ed | 7aa8375cdf4690fc3b9b1799b0f5a9ec1c1736ed_0 | Q: Is ROUGE their only baseline?
Text: Introduction
Producing sentences which are perceived as natural by a human addressee—a property which we will denote as fluency throughout this paper —is a crucial goal of all natural language generation (NLG) systems: it makes interactions more natural, avoids misunderstandings a... | No |
7aa8375cdf4690fc3b9b1799b0f5a9ec1c1736ed | 7aa8375cdf4690fc3b9b1799b0f5a9ec1c1736ed_1 | Q: Is ROUGE their only baseline?
Text: Introduction
Producing sentences which are perceived as natural by a human addressee—a property which we will denote as fluency throughout this paper —is a crucial goal of all natural language generation (NLG) systems: it makes interactions more natural, avoids misunderstandings a... | No, other baseline metrics they use besides ROUGE-L are n-gram overlap, negative cross-entropy, perplexity, and BLEU. |
3ac30bd7476d759ea5d9a5abf696d4dfc480175b | 3ac30bd7476d759ea5d9a5abf696d4dfc480175b_0 | Q: what language models do they use?
Text: Introduction
Producing sentences which are perceived as natural by a human addressee—a property which we will denote as fluency throughout this paper —is a crucial goal of all natural language generation (NLG) systems: it makes interactions more natural, avoids misunderstandin... | LSTM LMs |
0e57a0983b4731eba9470ba964d131045c8c7ea7 | 0e57a0983b4731eba9470ba964d131045c8c7ea7_0 | Q: what questions do they ask human judges?
Text: Introduction
Producing sentences which are perceived as natural by a human addressee—a property which we will denote as fluency throughout this paper —is a crucial goal of all natural language generation (NLG) systems: it makes interactions more natural, avoids misunder... | Unanswerable |
f0317e48dafe117829e88e54ed2edab24b86edb1 | f0317e48dafe117829e88e54ed2edab24b86edb1_0 | Q: What misbehavior is identified?
Text: Introduction
In machine translation, neural networks have attracted a lot of research attention. Recently, the attention-based encoder-decoder framework BIBREF0 , BIBREF1 has been largely adopted. In this approach, Recurrent Neural Networks (RNNs) map source sequences of words t... | if the attention loose track of the objects in the picture and "gets lost", the model still takes it into account and somehow overrides the information brought by the text-based annotations |
f0317e48dafe117829e88e54ed2edab24b86edb1 | f0317e48dafe117829e88e54ed2edab24b86edb1_1 | Q: What misbehavior is identified?
Text: Introduction
In machine translation, neural networks have attracted a lot of research attention. Recently, the attention-based encoder-decoder framework BIBREF0 , BIBREF1 has been largely adopted. In this approach, Recurrent Neural Networks (RNNs) map source sequences of words t... | if the attention loose track of the objects in the picture and "gets lost", the model still takes it into account and somehow overrides the information brought by the text-based annotations |
ec91b87c3f45df050e4e16018d2bf5b62e4ca298 | ec91b87c3f45df050e4e16018d2bf5b62e4ca298_0 | Q: What is the baseline used?
Text: Introduction
In machine translation, neural networks have attracted a lot of research attention. Recently, the attention-based encoder-decoder framework BIBREF0 , BIBREF1 has been largely adopted. In this approach, Recurrent Neural Networks (RNNs) map source sequences of words to tar... | Unanswerable |
f129c97a81d81d32633c94111018880a7ffe16d1 | f129c97a81d81d32633c94111018880a7ffe16d1_0 | Q: Which attention mechanisms do they compare?
Text: Introduction
In machine translation, neural networks have attracted a lot of research attention. Recently, the attention-based encoder-decoder framework BIBREF0 , BIBREF1 has been largely adopted. In this approach, Recurrent Neural Networks (RNNs) map source sequence... | Soft attention, Hard Stochastic attention, Local Attention |
100cf8b72d46da39fedfe77ec939fb44f25de77f | 100cf8b72d46da39fedfe77ec939fb44f25de77f_0 | Q: Which paired corpora did they use in the other experiment?
Text: Introduction
Making article comments is a fundamental ability for an intelligent machine to understand the article and interact with humans. It provides more challenges because commenting requires the abilities of comprehending the article, summarizing... | dataset that contains article-comment parallel contents INLINEFORM0 , and an unpaired dataset that contains the documents (articles or comments) INLINEFORM1 |
100cf8b72d46da39fedfe77ec939fb44f25de77f | 100cf8b72d46da39fedfe77ec939fb44f25de77f_1 | Q: Which paired corpora did they use in the other experiment?
Text: Introduction
Making article comments is a fundamental ability for an intelligent machine to understand the article and interact with humans. It provides more challenges because commenting requires the abilities of comprehending the article, summarizing... | Chinese dataset BIBREF0 |
8cc56fc44136498471754186cfa04056017b4e54 | 8cc56fc44136498471754186cfa04056017b4e54_0 | Q: By how much does their system outperform the lexicon-based models?
Text: Introduction
Making article comments is a fundamental ability for an intelligent machine to understand the article and interact with humans. It provides more challenges because commenting requires the abilities of comprehending the article, sum... | Under the retrieval evaluation setting, their proposed model + IR2 had better MRR than NVDM by 0.3769, better MR by 4.6, and better Recall@10 by 20 .
Under the generative evaluation setting the proposed model + IR2 had better BLEU by 0.044 , better CIDEr by 0.033, better ROUGE by 0.032, and better METEOR by 0.029 |
8cc56fc44136498471754186cfa04056017b4e54 | 8cc56fc44136498471754186cfa04056017b4e54_1 | Q: By how much does their system outperform the lexicon-based models?
Text: Introduction
Making article comments is a fundamental ability for an intelligent machine to understand the article and interact with humans. It provides more challenges because commenting requires the abilities of comprehending the article, sum... | Proposed model is better than both lexical based models by significan margin in all metrics: BLEU 0.261 vs 0.250, ROUGLE 0.162 vs 0.155 etc. |
5fa431b14732b3c47ab6eec373f51f2bca04f614 | 5fa431b14732b3c47ab6eec373f51f2bca04f614_0 | Q: Which lexicon-based models did they compare with?
Text: Introduction
Making article comments is a fundamental ability for an intelligent machine to understand the article and interact with humans. It provides more challenges because commenting requires the abilities of comprehending the article, summarizing the main... | TF-IDF, NVDM |
33ccbc401b224a48fba4b167e86019ffad1787fb | 33ccbc401b224a48fba4b167e86019ffad1787fb_0 | Q: How many comments were used?
Text: Introduction
Making article comments is a fundamental ability for an intelligent machine to understand the article and interact with humans. It provides more challenges because commenting requires the abilities of comprehending the article, summarizing the main ideas, mining the op... | from 50K to 4.8M |
cca74448ab0c518edd5fc53454affd67ac1a201c | cca74448ab0c518edd5fc53454affd67ac1a201c_0 | Q: How many articles did they have?
Text: Introduction
Making article comments is a fundamental ability for an intelligent machine to understand the article and interact with humans. It provides more challenges because commenting requires the abilities of comprehending the article, summarizing the main ideas, mining th... | 198,112 |
b69ffec1c607bfe5aa4d39254e0770a3433a191b | b69ffec1c607bfe5aa4d39254e0770a3433a191b_0 | Q: What news comment dataset was used?
Text: Introduction
Making article comments is a fundamental ability for an intelligent machine to understand the article and interact with humans. It provides more challenges because commenting requires the abilities of comprehending the article, summarizing the main ideas, mining... | Chinese dataset BIBREF0 |
f5cf8738e8d211095bb89350ed05ee7f9997eb19 | f5cf8738e8d211095bb89350ed05ee7f9997eb19_0 | Q: By how much do they outperform standard BERT?
Text: Introduction
With ever-increasing amounts of data available, there is an increase in the need to offer tooling to speed up processing, and eventually making sense of this data. Because fully-automated tools to extract meaning from any given input to any desired lev... | up to four percentage points in accuracy |
bed527bcb0dd5424e69563fba4ae7e6ea1fca26a | bed527bcb0dd5424e69563fba4ae7e6ea1fca26a_0 | Q: What dataset do they use?
Text: Introduction
With ever-increasing amounts of data available, there is an increase in the need to offer tooling to speed up processing, and eventually making sense of this data. Because fully-automated tools to extract meaning from any given input to any desired level of detail have ye... | 2019 GermEval shared task on hierarchical text classification |
bed527bcb0dd5424e69563fba4ae7e6ea1fca26a | bed527bcb0dd5424e69563fba4ae7e6ea1fca26a_1 | Q: What dataset do they use?
Text: Introduction
With ever-increasing amounts of data available, there is an increase in the need to offer tooling to speed up processing, and eventually making sense of this data. Because fully-automated tools to extract meaning from any given input to any desired level of detail have ye... | GermEval 2019 shared task |
aeab5797b541850e692f11e79167928db80de1ea | aeab5797b541850e692f11e79167928db80de1ea_0 | Q: How do they combine text representations with the knowledge graph embeddings?
Text: Introduction
With ever-increasing amounts of data available, there is an increase in the need to offer tooling to speed up processing, and eventually making sense of this data. Because fully-automated tools to extract meaning from an... | all three representations are concatenated and passed into a MLP |
cda4612b4bda3538d19f4b43dde7bc30c1eda4e5 | cda4612b4bda3538d19f4b43dde7bc30c1eda4e5_0 | Q: What are the traditional methods to identifying important attributes?
Text: The problem we solve in this paper
Knowledge graph(KG) has been proposed for several years and its most prominent application is in web search, for example, Google search triggers a certain entity card when a user's query matches or mentions... | automated attribute-value extraction, score the attributes using the Bayes model, evaluate their importance with several different frequency metrics, aggregate the weights from different sources into one consistent typicality score using a Ranking SVM model, OntoRank algorithm |
cda4612b4bda3538d19f4b43dde7bc30c1eda4e5 | cda4612b4bda3538d19f4b43dde7bc30c1eda4e5_1 | Q: What are the traditional methods to identifying important attributes?
Text: The problem we solve in this paper
Knowledge graph(KG) has been proposed for several years and its most prominent application is in web search, for example, Google search triggers a certain entity card when a user's query matches or mentions... | TextRank, Word2vec BIBREF19, GloVe BIBREF20 |
e12674f0466f8c0da109b6076d9939b30952c7da | e12674f0466f8c0da109b6076d9939b30952c7da_0 | Q: What do you use to calculate word/sub-word embeddings
Text: The problem we solve in this paper
Knowledge graph(KG) has been proposed for several years and its most prominent application is in web search, for example, Google search triggers a certain entity card when a user's query matches or mentions an entity based... | FastText |
9fe6339c7027a1a0caffa613adabe8b5bb6a7d4a | 9fe6339c7027a1a0caffa613adabe8b5bb6a7d4a_0 | Q: What user generated text data do you use?
Text: The problem we solve in this paper
Knowledge graph(KG) has been proposed for several years and its most prominent application is in web search, for example, Google search triggers a certain entity card when a user's query matches or mentions an entity based on some sta... | Unanswerable |
b5c3787ab3784214fc35f230ac4926fe184d86ba | b5c3787ab3784214fc35f230ac4926fe184d86ba_0 | Q: Did they propose other metrics?
Text: Introduction
Characteristic metrics are a set of unsupervised measures that quantitatively describe or summarize the properties of a data collection. These metrics generally do not use ground-truth labels and only measure the intrinsic characteristics of data. The most prominent... | Yes |
9174aded45bc36915f2e2adb6f352f3c7d9ada8b | 9174aded45bc36915f2e2adb6f352f3c7d9ada8b_0 | Q: Which real-world datasets did they use?
Text: Introduction
Characteristic metrics are a set of unsupervised measures that quantitatively describe or summarize the properties of a data collection. These metrics generally do not use ground-truth labels and only measure the intrinsic characteristics of data. The most p... | SST-2 (Stanford Sentiment Treebank, version 2), Snips |
9174aded45bc36915f2e2adb6f352f3c7d9ada8b | 9174aded45bc36915f2e2adb6f352f3c7d9ada8b_1 | Q: Which real-world datasets did they use?
Text: Introduction
Characteristic metrics are a set of unsupervised measures that quantitatively describe or summarize the properties of a data collection. These metrics generally do not use ground-truth labels and only measure the intrinsic characteristics of data. The most p... | SST-2, Snips |
a8f1029f6766bffee38a627477f61457b2d6ed5c | a8f1029f6766bffee38a627477f61457b2d6ed5c_0 | Q: How did they obtain human intuitions?
Text: Introduction
Characteristic metrics are a set of unsupervised measures that quantitatively describe or summarize the properties of a data collection. These metrics generally do not use ground-truth labels and only measure the intrinsic characteristics of data. The most pro... | Unanswerable |
a2103e7fe613549a9db5e65008f33cf2ee0403bd | a2103e7fe613549a9db5e65008f33cf2ee0403bd_0 | Q: What are the country-specific drivers of international development rhetoric?
Text: Introduction
Decisions made in international organisations are fundamental to international development efforts and initiatives. It is in these global governance arenas that the rules of the global economic system, which have a huge i... | wealth , democracy , population, levels of ODA, conflict |
13b36644357870008d70e5601f394ec3c6c07048 | 13b36644357870008d70e5601f394ec3c6c07048_0 | Q: Is the dataset multilingual?
Text: Introduction
Decisions made in international organisations are fundamental to international development efforts and initiatives. It is in these global governance arenas that the rules of the global economic system, which have a huge impact on development outcomes are agreed on; dec... | No |
13b36644357870008d70e5601f394ec3c6c07048 | 13b36644357870008d70e5601f394ec3c6c07048_1 | Q: Is the dataset multilingual?
Text: Introduction
Decisions made in international organisations are fundamental to international development efforts and initiatives. It is in these global governance arenas that the rules of the global economic system, which have a huge impact on development outcomes are agreed on; dec... | No |
e4a19b91b57c006a9086ae07f2d6d6471a8cf0ce | e4a19b91b57c006a9086ae07f2d6d6471a8cf0ce_0 | Q: How are the main international development topics that states raise identified?
Text: Introduction
Decisions made in international organisations are fundamental to international development efforts and initiatives. It is in these global governance arenas that the rules of the global economic system, which have a hug... | They focus on exclusivity and semantic coherence measures: Highly frequent words in a given topic that do not appear very often in other topics are viewed as making that topic exclusive. They select select the 16-topic model, which has the largest positive residual in the regression fit, and provides higher exclusivit... |
fd0ef5a7b6f62d07776bf672579a99c67e61a568 | fd0ef5a7b6f62d07776bf672579a99c67e61a568_0 | Q: What experiments do the authors present to validate their system?
Text: Introduction
QnAMaker aims to simplify the process of bot creation by extracting Question-Answer (QA) pairs from data given by users into a Knowledge Base (KB) and providing a conversational layer over it. KB here refers to one instance of azure... | we measure our system's performance for datasets across various domains, evaluations are done by managed judges who understands the knowledge base and then judge user queries relevance to the QA pairs |
071bcb4b054215054f17db64bfd21f17fd9e1a80 | 071bcb4b054215054f17db64bfd21f17fd9e1a80_0 | Q: How does the conversation layer work?
Text: Introduction
QnAMaker aims to simplify the process of bot creation by extracting Question-Answer (QA) pairs from data given by users into a Knowledge Base (KB) and providing a conversational layer over it. KB here refers to one instance of azure search index, where the ext... | Unanswerable |
f399d5a8dbeec777a858f81dc4dd33a83ba341a2 | f399d5a8dbeec777a858f81dc4dd33a83ba341a2_0 | Q: What components is the QnAMaker composed of?
Text: Introduction
QnAMaker aims to simplify the process of bot creation by extracting Question-Answer (QA) pairs from data given by users into a Knowledge Base (KB) and providing a conversational layer over it. KB here refers to one instance of azure search index, where ... | QnAMaker Portal, QnaMaker Management APIs, Azure Search Index, QnaMaker WebApp, Bot |
f399d5a8dbeec777a858f81dc4dd33a83ba341a2 | f399d5a8dbeec777a858f81dc4dd33a83ba341a2_1 | Q: What components is the QnAMaker composed of?
Text: Introduction
QnAMaker aims to simplify the process of bot creation by extracting Question-Answer (QA) pairs from data given by users into a Knowledge Base (KB) and providing a conversational layer over it. KB here refers to one instance of azure search index, where ... | QnAMaker Portal, QnaMaker Management APIs, Azure Search Index, QnaMaker WebApp, Bot |
d28260b5565d9246831e8dbe594d4f6211b60237 | d28260b5565d9246831e8dbe594d4f6211b60237_0 | Q: How they measure robustness in experiments?
Text: Introduction
Since Och BIBREF0 proposed minimum error rate training (MERT) to exactly optimize objective evaluation measures, MERT has become a standard model tuning technique in statistical machine translation (SMT). Though MERT performs better by improving its sear... | We empirically provide a formula to measure the richness in the scenario of machine translation. |
d28260b5565d9246831e8dbe594d4f6211b60237 | d28260b5565d9246831e8dbe594d4f6211b60237_1 | Q: How they measure robustness in experiments?
Text: Introduction
Since Och BIBREF0 proposed minimum error rate training (MERT) to exactly optimize objective evaluation measures, MERT has become a standard model tuning technique in statistical machine translation (SMT). Though MERT performs better by improving its sear... | boost the training BLEU very greatly, the over-fitting problem of the Plackett-Luce models PL($k$) is alleviated with moderately large $k$ |
8670989ca39214eda6c1d1d272457a3f3a92818b | 8670989ca39214eda6c1d1d272457a3f3a92818b_0 | Q: Is new method inferior in terms of robustness to MIRAs in experiments?
Text: Introduction
Since Och BIBREF0 proposed minimum error rate training (MERT) to exactly optimize objective evaluation measures, MERT has become a standard model tuning technique in statistical machine translation (SMT). Though MERT performs b... | Unanswerable |
923b12c0a50b0ee22237929559fad0903a098b7b | 923b12c0a50b0ee22237929559fad0903a098b7b_0 | Q: What experiments with large-scale features are performed?
Text: Introduction
Since Och BIBREF0 proposed minimum error rate training (MERT) to exactly optimize objective evaluation measures, MERT has become a standard model tuning technique in statistical machine translation (SMT). Though MERT performs better by impr... | Plackett-Luce Model for SMT Reranking |
67131c15aceeb51ae1d3b2b8241c8750a19cca8e | 67131c15aceeb51ae1d3b2b8241c8750a19cca8e_0 | Q: Which ASR system(s) is used in this work?
Text: Introduction
Currently, voice-controlled smart devices are widely used in multiple areas to fulfill various tasks, e.g. playing music, acquiring weather information and booking tickets. The SLU system employs several modules to enable the understanding of the semantics... | Oracle |
579a0603ec56fc2b4aa8566810041dbb0cd7b5e7 | 579a0603ec56fc2b4aa8566810041dbb0cd7b5e7_0 | Q: What are the series of simple models?
Text: Introduction
Currently, voice-controlled smart devices are widely used in multiple areas to fulfill various tasks, e.g. playing music, acquiring weather information and booking tickets. The SLU system employs several modules to enable the understanding of the semantics of ... | perform experiments to utilize ASR $n$-best hypotheses during evaluation |
c9c85eee41556c6993f40e428fa607af4abe80a9 | c9c85eee41556c6993f40e428fa607af4abe80a9_0 | Q: Over which datasets/corpora is this work evaluated?
Text: Introduction
Currently, voice-controlled smart devices are widely used in multiple areas to fulfill various tasks, e.g. playing music, acquiring weather information and booking tickets. The SLU system employs several modules to enable the understanding of the... | $\sim $ 8.7M annotated anonymised user utterances |
c9c85eee41556c6993f40e428fa607af4abe80a9 | c9c85eee41556c6993f40e428fa607af4abe80a9_1 | Q: Over which datasets/corpora is this work evaluated?
Text: Introduction
Currently, voice-controlled smart devices are widely used in multiple areas to fulfill various tasks, e.g. playing music, acquiring weather information and booking tickets. The SLU system employs several modules to enable the understanding of the... | on $\sim $ 8.7M annotated anonymised user utterances |
982979cb3c71770d8d7d2d1be8f92b66223dec85 | 982979cb3c71770d8d7d2d1be8f92b66223dec85_0 | Q: What new metrics are suggested to track progress?
Text: Introduction
Word embeddings have great practical importance since they can be used as pre-computed high-density features to ML models, significantly reducing the amount of training data required in a variety of NLP tasks. However, there are several inter-relat... | For example, one metric could consist in checking whether for any given word, all words that are known to belong to the same class are closer than any words belonging to different classes, independently of the actual cosine |
5ba6f7f235d0f5d1d01fd97dd5e4d5b0544fd212 | 5ba6f7f235d0f5d1d01fd97dd5e4d5b0544fd212_0 | Q: What intrinsic evaluation metrics are used?
Text: Introduction
Word embeddings have great practical importance since they can be used as pre-computed high-density features to ML models, significantly reducing the amount of training data required in a variety of NLP tasks. However, there are several inter-related cha... | Class Membership Tests, Class Distinction Test, Word Equivalence Test |
5ba6f7f235d0f5d1d01fd97dd5e4d5b0544fd212 | 5ba6f7f235d0f5d1d01fd97dd5e4d5b0544fd212_1 | Q: What intrinsic evaluation metrics are used?
Text: Introduction
Word embeddings have great practical importance since they can be used as pre-computed high-density features to ML models, significantly reducing the amount of training data required in a variety of NLP tasks. However, there are several inter-related cha... | coverage metric, being distinct (cosine INLINEFORM0 0.7 or 0.8), belonging to the same class (cosine INLINEFORM1 0.7 or 0.8), being equivalent (cosine INLINEFORM2 0.85 or 0.95) |
7ce7edd06925a943e32b59f3e7b5159ccb7acaf6 | 7ce7edd06925a943e32b59f3e7b5159ccb7acaf6_0 | Q: What experimental results suggest that using less than 50% of the available training examples might result in overfitting?
Text: Introduction
Word embeddings have great practical importance since they can be used as pre-computed high-density features to ML models, significantly reducing the amount of training data r... | consistent increase in the validation loss after about 15 epochs |
a883bb41449794e0a63b716d9766faea034eb359 | a883bb41449794e0a63b716d9766faea034eb359_0 | Q: What multimodality is available in the dataset?
Text: Introduction
A great deal of commonsense knowledge about the world we live is procedural in nature and involves steps that show ways to achieve specific goals. Understanding and reasoning about procedural texts (e.g. cooking recipes, how-to guides, scientific pro... | context is a procedural text, the question and the multiple choice answers are composed of images |
a883bb41449794e0a63b716d9766faea034eb359 | a883bb41449794e0a63b716d9766faea034eb359_1 | Q: What multimodality is available in the dataset?
Text: Introduction
A great deal of commonsense knowledge about the world we live is procedural in nature and involves steps that show ways to achieve specific goals. Understanding and reasoning about procedural texts (e.g. cooking recipes, how-to guides, scientific pro... | images and text |
5d83b073635f5fd8cd1bdb1895d3f13406583fbd | 5d83b073635f5fd8cd1bdb1895d3f13406583fbd_0 | Q: What are previously reported models?
Text: Introduction
A great deal of commonsense knowledge about the world we live is procedural in nature and involves steps that show ways to achieve specific goals. Understanding and reasoning about procedural texts (e.g. cooking recipes, how-to guides, scientific processes) are... | Hasty Student, Impatient Reader, BiDAF, BiDAF w/ static memory |
171ebfdc9b3a98e4cdee8f8715003285caeb2f39 | 171ebfdc9b3a98e4cdee8f8715003285caeb2f39_0 | Q: How better is accuracy of new model compared to previously reported models?
Text: Introduction
A great deal of commonsense knowledge about the world we live is procedural in nature and involves steps that show ways to achieve specific goals. Understanding and reasoning about procedural texts (e.g. cooking recipes, h... | Average accuracy of proposed model vs best prevous result:
Single-task Training: 57.57 vs 55.06
Multi-task Training: 50.17 vs 50.59 |
3c3cb51093b5fd163e87a773a857496a4ae71f03 | 3c3cb51093b5fd163e87a773a857496a4ae71f03_0 | Q: How does the scoring model work?
Text: Introduction
Electronic health records (EHRs) systematically collect patients' clinical information, such as health profiles, histories of present illness, past medical histories, examination results and treatment plans BIBREF0 . By analyzing EHRs, many useful information, clos... | First, mapping the segmented sentence to a sequence of candidate word embeddings. Then, the scoring model takes the word embedding sequence as input, scoring over each individual candidate word |
3c3cb51093b5fd163e87a773a857496a4ae71f03 | 3c3cb51093b5fd163e87a773a857496a4ae71f03_1 | Q: How does the scoring model work?
Text: Introduction
Electronic health records (EHRs) systematically collect patients' clinical information, such as health profiles, histories of present illness, past medical histories, examination results and treatment plans BIBREF0 . By analyzing EHRs, many useful information, clos... | the scoring model takes the word embedding sequence as input, scoring over each individual candidate word from two perspectives: (1) the possibility that the candidate word itself can be regarded as a legal word; (2) the rationality of the link that the candidate word directly follows previous segmentation history |
53a0763eff99a8148585ac642705637874be69d4 | 53a0763eff99a8148585ac642705637874be69d4_0 | Q: How does the active learning model work?
Text: Introduction
Electronic health records (EHRs) systematically collect patients' clinical information, such as health profiles, histories of present illness, past medical histories, examination results and treatment plans BIBREF0 . By analyzing EHRs, many useful informati... | Active learning methods has a learning engine (mainly used for training of classification problems) and the selection engine (which chooses samples that need to be relabeled by annotators from unlabeled data). Then, relabeled samples are added to training set for classifier to re-train, thus continuously improving the ... |
0bfed6f9cfe93617c5195c848583e3945f2002ff | 0bfed6f9cfe93617c5195c848583e3945f2002ff_0 | Q: Which neural network architectures are employed?
Text: Introduction
Electronic health records (EHRs) systematically collect patients' clinical information, such as health profiles, histories of present illness, past medical histories, examination results and treatment plans BIBREF0 . By analyzing EHRs, many useful i... | gated neural network |
352c081c93800df9654315e13a880d6387b91919 | 352c081c93800df9654315e13a880d6387b91919_0 | Q: What are the key points in the role of script knowledge that can be studied?
Text: Motivation
A script is “a standardized sequence of events that describes some stereotypical human activity such as going to a restaurant or visiting a doctor” BIBREF0 . Script events describe an action/activity along with the involved... | Unanswerable |
18fbf9c08075e3b696237d22473c463237d153f5 | 18fbf9c08075e3b696237d22473c463237d153f5_0 | Q: Did the annotators agreed and how much?
Text: Motivation
A script is “a standardized sequence of events that describes some stereotypical human activity such as going to a restaurant or visiting a doctor” BIBREF0 . Script events describe an action/activity along with the involved participants. For example, in the sc... | For event types and participant types, there was a moderate to substantial level of agreement using the Fleiss' Kappa. For coreference chain annotation, there was average agreement of 90.5%. |
18fbf9c08075e3b696237d22473c463237d153f5 | 18fbf9c08075e3b696237d22473c463237d153f5_1 | Q: Did the annotators agreed and how much?
Text: Motivation
A script is “a standardized sequence of events that describes some stereotypical human activity such as going to a restaurant or visiting a doctor” BIBREF0 . Script events describe an action/activity along with the involved participants. For example, in the sc... | Moderate agreement of 0.64-0.68 Fleiss’ Kappa over event type labels, 0.77 Fleiss’ Kappa over participant labels, and good agreement of 90.5% over coreference information. |
a37ef83ab6bcc6faff3c70a481f26174ccd40489 | a37ef83ab6bcc6faff3c70a481f26174ccd40489_0 | Q: How many subjects have been used to create the annotations?
Text: Motivation
A script is “a standardized sequence of events that describes some stereotypical human activity such as going to a restaurant or visiting a doctor” BIBREF0 . Script events describe an action/activity along with the involved participants. Fo... | four different annotators |
bc9c31b3ce8126d1d148b1025c66f270581fde10 | bc9c31b3ce8126d1d148b1025c66f270581fde10_0 | Q: What datasets are used to evaluate this approach?
Text: Introduction
Knowledge graphs (KG) play a critical role in many real-world applications such as search, structured data management, recommendations, and question answering. Since KGs often suffer from incompleteness and noise in their facts (links), a number of... | Kinship and Nations knowledge graphs, YAGO3-10 and WN18KGs knowledge graphs |
bc9c31b3ce8126d1d148b1025c66f270581fde10 | bc9c31b3ce8126d1d148b1025c66f270581fde10_1 | Q: What datasets are used to evaluate this approach?
Text: Introduction
Knowledge graphs (KG) play a critical role in many real-world applications such as search, structured data management, recommendations, and question answering. Since KGs often suffer from incompleteness and noise in their facts (links), a number of... | WN18 and YAGO3-10 |
185841e979373808d99dccdade5272af02b98774 | 185841e979373808d99dccdade5272af02b98774_0 | Q: How is this approach used to detect incorrect facts?
Text: Introduction
Knowledge graphs (KG) play a critical role in many real-world applications such as search, structured data management, recommendations, and question answering. Since KGs often suffer from incompleteness and noise in their facts (links), a number... | if there is an error in the graph, the triple is likely to be inconsistent with its neighborhood, and thus the model should put least trust on this triple. In other words, the error triple should have the least influence on the model's prediction of the training data. |
d427e3d41c4c9391192e249493be23926fc5d2e9 | d427e3d41c4c9391192e249493be23926fc5d2e9_0 | Q: Can this adversarial approach be used to directly improve model accuracy?
Text: Introduction
Knowledge graphs (KG) play a critical role in many real-world applications such as search, structured data management, recommendations, and question answering. Since KGs often suffer from incompleteness and noise in their fa... | Yes |
330f2cdeab689670b68583fc4125f5c0b26615a8 | 330f2cdeab689670b68583fc4125f5c0b26615a8_0 | Q: what are the advantages of the proposed model?
Text: Introduction
Topic models, such as latent Dirichlet allocation (LDA), allow us to analyze large collections of documents by revealing their underlying themes, or topics, and how each document exhibits them BIBREF0 . Therefore, it is not surprising that topic model... | he proposed model outperforms all the baselines, being the svi version the one that performs best., the svi version converges much faster to higher values of the log marginal likelihood when compared to the batch version, which reflects the efficiency of the svi algorithm. |
c87b2dd5c439d5e68841a705dd81323ec0d64c97 | c87b2dd5c439d5e68841a705dd81323ec0d64c97_0 | Q: what are the state of the art approaches?
Text: Introduction
Topic models, such as latent Dirichlet allocation (LDA), allow us to analyze large collections of documents by revealing their underlying themes, or topics, and how each document exhibits them BIBREF0 . Therefore, it is not surprising that topic models hav... | Bosch 2006 (mv), LDA + LogReg (mv), LDA + Raykar, LDA + Rodrigues, Blei 2003 (mv), sLDA (mv) |
f7789313a804e41fcbca906a4e5cf69039eeef9f | f7789313a804e41fcbca906a4e5cf69039eeef9f_0 | Q: what datasets were used?
Text: Introduction
Topic models, such as latent Dirichlet allocation (LDA), allow us to analyze large collections of documents by revealing their underlying themes, or topics, and how each document exhibits them BIBREF0 . Therefore, it is not surprising that topic models have become a standa... | Reuters-21578 BIBREF30, LabelMe BIBREF31, 20-Newsgroups benchmark corpus BIBREF29 |
f7789313a804e41fcbca906a4e5cf69039eeef9f | f7789313a804e41fcbca906a4e5cf69039eeef9f_1 | Q: what datasets were used?
Text: Introduction
Topic models, such as latent Dirichlet allocation (LDA), allow us to analyze large collections of documents by revealing their underlying themes, or topics, and how each document exhibits them BIBREF0 . Therefore, it is not surprising that topic models have become a standa... | 20-Newsgroups benchmark corpus , Reuters-21578, LabelMe |
2376c170c343e2305dac08ba5f5bda47c370357f | 2376c170c343e2305dac08ba5f5bda47c370357f_0 | Q: How was the dataset collected?
Text: Introduction
Recently, there have been a variety of task-oriented dialogue models thanks to the prosperity of neural architectures BIBREF0, BIBREF1, BIBREF2, BIBREF3, BIBREF4, BIBREF5. However, the research is still largely limited by the availability of large-scale high-quality ... | Database Construction: we crawled travel information in Beijing from the Web, including Hotel, Attraction, and Restaurant domains (hereafter we name the three domains as HAR domains). Then, we used the metro information of entities in HAR domains to build the metro database. , Goal Generation: a multi-domain goal gener... |
2376c170c343e2305dac08ba5f5bda47c370357f | 2376c170c343e2305dac08ba5f5bda47c370357f_1 | Q: How was the dataset collected?
Text: Introduction
Recently, there have been a variety of task-oriented dialogue models thanks to the prosperity of neural architectures BIBREF0, BIBREF1, BIBREF2, BIBREF3, BIBREF4, BIBREF5. However, the research is still largely limited by the availability of large-scale high-quality ... | They crawled travel information from the Web to build a database, created a multi-domain goal generator from the database, collected dialogue between workers an automatically annotated dialogue acts. |
0137ecebd84a03b224eb5ca51d189283abb5f6d9 | 0137ecebd84a03b224eb5ca51d189283abb5f6d9_0 | Q: What are the benchmark models?
Text: Introduction
Recently, there have been a variety of task-oriented dialogue models thanks to the prosperity of neural architectures BIBREF0, BIBREF1, BIBREF2, BIBREF3, BIBREF4, BIBREF5. However, the research is still largely limited by the availability of large-scale high-quality ... | BERTNLU from ConvLab-2, a rule-based model (RuleDST) , TRADE (Transferable Dialogue State Generator) , a vanilla policy trained in a supervised fashion from ConvLab-2 (SL policy) |
5f6fbd57cce47f20a0fda27d954543c00c4344c2 | 5f6fbd57cce47f20a0fda27d954543c00c4344c2_0 | Q: How was the corpus annotated?
Text: Introduction
Recently, there have been a variety of task-oriented dialogue models thanks to the prosperity of neural architectures BIBREF0, BIBREF1, BIBREF2, BIBREF3, BIBREF4, BIBREF5. However, the research is still largely limited by the availability of large-scale high-quality d... | The workers were also asked to annotate both user states and system states, we used some rules to automatically annotate dialogue acts according to user states, system states, and dialogue histories |
d6e2b276390bdc957dfa7e878de80cee1f41fbca | d6e2b276390bdc957dfa7e878de80cee1f41fbca_0 | Q: What models other than standalone BERT is new model compared to?
Text: Introduction
As traditional word embedding algorithms BIBREF1 are known to struggle with rare words, several techniques for improving their representations have been proposed over the last few years. These approaches exploit either the contexts i... | Only Bert base and Bert large are compared to proposed approach. |
32537fdf0d4f76f641086944b413b2f756097e5e | 32537fdf0d4f76f641086944b413b2f756097e5e_0 | Q: How much is representaton improved for rare/medum frequency words compared to standalone BERT and previous work?
Text: Introduction
As traditional word embedding algorithms BIBREF1 are known to struggle with rare words, several techniques for improving their representations have been proposed over the last few years... | improving the score for WNLaMPro-medium by 50% compared to BERT$_\text{base}$ and 31% compared to Attentive Mimicking |
ef081d78be17ef2af792e7e919d15a235b8d7275 | ef081d78be17ef2af792e7e919d15a235b8d7275_0 | Q: What are three downstream task datasets?
Text: Introduction
As traditional word embedding algorithms BIBREF1 are known to struggle with rare words, several techniques for improving their representations have been proposed over the last few years. These approaches exploit either the contexts in which rare words occur... | MNLI BIBREF21, AG's News BIBREF22, DBPedia BIBREF23 |
ef081d78be17ef2af792e7e919d15a235b8d7275 | ef081d78be17ef2af792e7e919d15a235b8d7275_1 | Q: What are three downstream task datasets?
Text: Introduction
As traditional word embedding algorithms BIBREF1 are known to struggle with rare words, several techniques for improving their representations have been proposed over the last few years. These approaches exploit either the contexts in which rare words occur... | MNLI, AG's News, DBPedia |
537b2d7799124d633892a1ef1a485b3b071b303d | 537b2d7799124d633892a1ef1a485b3b071b303d_0 | Q: What is dataset for word probing task?
Text: Introduction
As traditional word embedding algorithms BIBREF1 are known to struggle with rare words, several techniques for improving their representations have been proposed over the last few years. These approaches exploit either the contexts in which rare words occur B... | WNLaMPro dataset |
9aca4b89e18ce659c905eccc78eda76af9f0072a | 9aca4b89e18ce659c905eccc78eda76af9f0072a_0 | Q: How fast is the model compared to baselines?
Text: Introduction
Entity Linking (EL), which is also called Entity Disambiguation (ED), is the task of mapping mentions in text to corresponding entities in a given knowledge Base (KB). This task is an important and challenging stage in text understanding because mention... | Unanswerable |
b0376a7f67f1568a7926eff8ff557a93f434a253 | b0376a7f67f1568a7926eff8ff557a93f434a253_0 | Q: How big is the performance difference between this method and the baseline?
Text: Introduction
Entity Linking (EL), which is also called Entity Disambiguation (ED), is the task of mapping mentions in text to corresponding entities in a given knowledge Base (KB). This task is an important and challenging stage in tex... | Comparing with the highest performing baseline: 1.3 points on ACE2004 dataset, 0.6 points on CWEB dataset, and 0.86 points in the average of all scores. |
dad8cc543a87534751f9f9e308787e1af06f0627 | dad8cc543a87534751f9f9e308787e1af06f0627_0 | Q: What datasets used for evaluation?
Text: Introduction
Entity Linking (EL), which is also called Entity Disambiguation (ED), is the task of mapping mentions in text to corresponding entities in a given knowledge Base (KB). This task is an important and challenging stage in text understanding because mentions are usua... | AIDA-B, ACE2004, MSNBC, AQUAINT, WNED-CWEB, WNED-WIKI |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.