id stringlengths 40 40 | pid stringlengths 42 42 | input stringlengths 8.37k 169k | output stringlengths 1 1.63k |
|---|---|---|---|
dad8cc543a87534751f9f9e308787e1af06f0627 | dad8cc543a87534751f9f9e308787e1af06f0627_1 | Q: What datasets used for evaluation?
Text: Introduction
Entity Linking (EL), which is also called Entity Disambiguation (ED), is the task of mapping mentions in text to corresponding entities in a given knowledge Base (KB). This task is an important and challenging stage in text understanding because mentions are usua... | AIDA-CoNLL, ACE2004, MSNBC, AQUAINT, WNED-CWEB, WNED-WIKI, OURSELF-WIKI |
0481a8edf795768d062c156875d20b8fb656432c | 0481a8edf795768d062c156875d20b8fb656432c_0 | Q: what are the mentioned cues?
Text: Introduction
Entity Linking (EL), which is also called Entity Disambiguation (ED), is the task of mapping mentions in text to corresponding entities in a given knowledge Base (KB). This task is an important and challenging stage in text understanding because mentions are usually am... | output of global LSTM network at time $V_{m_i}^t$5 , which encodes the mention context and target entity information from $V_{m_i}^t$6 to $V_{m_i}^t$7 |
b6a4ab009e6f213f011320155a7ce96e713c11cf | b6a4ab009e6f213f011320155a7ce96e713c11cf_0 | Q: How did the author's work rank among other submissions on the challenge?
Text: Introduction
The BioASQ Challenge includes a question answering task (Phase B, part B) where the aim is to find the βideal answerβ β that is, an answer that would normally be given by a person BIBREF0. This is in contrast with most other ... | Unanswerable |
cfffc94518d64cb3c8789395707e4336676e0345 | cfffc94518d64cb3c8789395707e4336676e0345_0 | Q: What approaches without reinforcement learning have been tried?
Text: Introduction
The BioASQ Challenge includes a question answering task (Phase B, part B) where the aim is to find the βideal answerβ β that is, an answer that would normally be given by a person BIBREF0. This is in contrast with most other question ... | classification, regression, neural methods |
cfffc94518d64cb3c8789395707e4336676e0345 | cfffc94518d64cb3c8789395707e4336676e0345_1 | Q: What approaches without reinforcement learning have been tried?
Text: Introduction
The BioASQ Challenge includes a question answering task (Phase B, part B) where the aim is to find the βideal answerβ β that is, an answer that would normally be given by a person BIBREF0. This is in contrast with most other question ... | Support Vector Regression (SVR) and Support Vector Classification (SVC), deep learning regression models of BIBREF2 to convert them to classification models |
f60629c01f99de3f68365833ee115b95a3388699 | f60629c01f99de3f68365833ee115b95a3388699_0 | Q: What classification approaches were experimented for this task?
Text: Introduction
The BioASQ Challenge includes a question answering task (Phase B, part B) where the aim is to find the βideal answerβ β that is, an answer that would normally be given by a person BIBREF0. This is in contrast with most other question ... | NNC SU4 F1, NNC top 5, Support Vector Classification (SVC) |
a7cb4f8e29fd2f3d1787df64cd981a6318b65896 | a7cb4f8e29fd2f3d1787df64cd981a6318b65896_0 | Q: Did classification models perform better than previous regression one?
Text: Introduction
The BioASQ Challenge includes a question answering task (Phase B, part B) where the aim is to find the βideal answerβ β that is, an answer that would normally be given by a person BIBREF0. This is in contrast with most other qu... | Yes |
642c4704a71fd01b922a0ef003f234dcc7b223cd | 642c4704a71fd01b922a0ef003f234dcc7b223cd_0 | Q: What are the main sources of recall errors in the mapping?
Text: Introduction
The two largest standardized, cross-lingual datasets for morphological annotation are provided by the Universal Dependencies BIBREF1 and Universal Morphology BIBREF2 , BIBREF3 projects. Each project's data are annotated according to its ow... | irremediable annotation discrepancies, differences in choice of attributes to annotate, The resources themselves would need updating to encode the relevant morphosyntactic information. Some languages had a very low number of overlapping forms, and no tag matches or near-matches between them, the two annotations encode ... |
e477e494fe15a978ff9c0a5f1c88712cdaec0c5c | e477e494fe15a978ff9c0a5f1c88712cdaec0c5c_0 | Q: Do they look for inconsistencies between different languages' annotations in UniMorph?
Text: Introduction
The two largest standardized, cross-lingual datasets for morphological annotation are provided by the Universal Dependencies BIBREF1 and Universal Morphology BIBREF2 , BIBREF3 projects. Each project's data are a... | Yes |
04495845251b387335bf2e77e2c423130f43c7d9 | 04495845251b387335bf2e77e2c423130f43c7d9_0 | Q: Do they look for inconsistencies between different UD treebanks?
Text: Introduction
The two largest standardized, cross-lingual datasets for morphological annotation are provided by the Universal Dependencies BIBREF1 and Universal Morphology BIBREF2 , BIBREF3 projects. Each project's data are annotated according to ... | Yes |
564dcaf8d0bcc274ab64c784e4c0f50d7a2c17ee | 564dcaf8d0bcc274ab64c784e4c0f50d7a2c17ee_0 | Q: Which languages do they validate on?
Text: Introduction
The two largest standardized, cross-lingual datasets for morphological annotation are provided by the Universal Dependencies BIBREF1 and Universal Morphology BIBREF2 , BIBREF3 projects. Each project's data are annotated according to its own cross-lingual schema... | Ar, Bg, Ca, Cs, Da, De, En, Es, Eu, Fa, Fi, Fr, Ga, He, Hi, Hu, It, La, Lt, Lv, Nb, Nl, Nn, PL, Pt, Ro, Ru, Sl, Sv, Tr, Uk, Ur |
564dcaf8d0bcc274ab64c784e4c0f50d7a2c17ee | 564dcaf8d0bcc274ab64c784e4c0f50d7a2c17ee_1 | Q: Which languages do they validate on?
Text: Introduction
The two largest standardized, cross-lingual datasets for morphological annotation are provided by the Universal Dependencies BIBREF1 and Universal Morphology BIBREF2 , BIBREF3 projects. Each project's data are annotated according to its own cross-lingual schema... | We apply this conversion to the 31 languages, Arabic, Hindi, Lithuanian, Persian, and Russian. , Dutch, Spanish |
f3d0e6452b8d24b7f9db1fd898d1fbe6cd23f166 | f3d0e6452b8d24b7f9db1fd898d1fbe6cd23f166_0 | Q: Does the paper evaluate any adjustment to improve the predicion accuracy of face and audio features?
Text: Introduction
Automatic emotion recognition is commonly understood as the task of assigning an emotion to a predefined instance, for example an utterance (as audio signal), an image (for instance with a depicted... | No |
9b1d789398f1f1a603e4741a5eee63ccaf0d4a4f | 9b1d789398f1f1a603e4741a5eee63ccaf0d4a4f_0 | Q: How is face and audio data analysis evaluated?
Text: Introduction
Automatic emotion recognition is commonly understood as the task of assigning an emotion to a predefined instance, for example an utterance (as audio signal), an image (for instance with a depicted face), or a textual unit (e.g., a transcribed utteran... | confusion matrices, $\text{F}_1$ score |
00bcdffff7e055f99aaf1b05cf41c98e2748e948 | 00bcdffff7e055f99aaf1b05cf41c98e2748e948_0 | Q: What is the baseline method for the task?
Text: Introduction
Automatic emotion recognition is commonly understood as the task of assigning an emotion to a predefined instance, for example an utterance (as audio signal), an image (for instance with a depicted face), or a textual unit (e.g., a transcribed utterance, a... | For the emotion recognition from text they use described neural network as baseline.
For audio and face there is no baseline. |
f92ee3c5fce819db540bded3cfcc191e21799cb1 | f92ee3c5fce819db540bded3cfcc191e21799cb1_0 | Q: What are the emotion detection tools used for audio and face input?
Text: Introduction
Automatic emotion recognition is commonly understood as the task of assigning an emotion to a predefined instance, for example an utterance (as audio signal), an image (for instance with a depicted face), or a textual unit (e.g., ... | We apply an off-the-shelf tool for emotion recognition (the manufacturer cannot be disclosed due to licensing restrictions) |
f92ee3c5fce819db540bded3cfcc191e21799cb1 | f92ee3c5fce819db540bded3cfcc191e21799cb1_1 | Q: What are the emotion detection tools used for audio and face input?
Text: Introduction
Automatic emotion recognition is commonly understood as the task of assigning an emotion to a predefined instance, for example an utterance (as audio signal), an image (for instance with a depicted face), or a textual unit (e.g., ... | cannot be disclosed due to licensing restrictions |
4547818a3bbb727c4bb4a76554b5a5a7b5c5fedb | 4547818a3bbb727c4bb4a76554b5a5a7b5c5fedb_0 | Q: what amounts of size were used on german-english?
Text: Introduction
While neural machine translation (NMT) has achieved impressive performance in high-resource data conditions, becoming dominant in the field BIBREF0 , BIBREF1 , BIBREF2 , recent research has argued that these models are highly data-inefficient, and ... | Training data with 159000, 80000, 40000, 20000, 10000 and 5000 sentences, and 7584 sentences for development |
4547818a3bbb727c4bb4a76554b5a5a7b5c5fedb | 4547818a3bbb727c4bb4a76554b5a5a7b5c5fedb_1 | Q: what amounts of size were used on german-english?
Text: Introduction
While neural machine translation (NMT) has achieved impressive performance in high-resource data conditions, becoming dominant in the field BIBREF0 , BIBREF1 , BIBREF2 , recent research has argued that these models are highly data-inefficient, and ... | ultra-low data condition (100k words of training data) and the full IWSLT 14 training corpus (3.2M words) |
07d7652ad4a0ec92e6b44847a17c378b0d9f57f5 | 07d7652ad4a0ec92e6b44847a17c378b0d9f57f5_0 | Q: what were their experimental results in the low-resource dataset?
Text: Introduction
While neural machine translation (NMT) has achieved impressive performance in high-resource data conditions, becoming dominant in the field BIBREF0 , BIBREF1 , BIBREF2 , recent research has argued that these models are highly data-i... | 10.37 BLEU |
9f3444c9fb2e144465d63abf58520cddd4165a01 | 9f3444c9fb2e144465d63abf58520cddd4165a01_0 | Q: what are the methods they compare with in the korean-english dataset?
Text: Introduction
While neural machine translation (NMT) has achieved impressive performance in high-resource data conditions, becoming dominant in the field BIBREF0 , BIBREF1 , BIBREF2 , recent research has argued that these models are highly da... | gu-EtAl:2018:EMNLP1 |
2348d68e065443f701d8052018c18daa4ecc120e | 2348d68e065443f701d8052018c18daa4ecc120e_0 | Q: what pitfalls are mentioned in the paper?
Text: Introduction
While neural machine translation (NMT) has achieved impressive performance in high-resource data conditions, becoming dominant in the field BIBREF0 , BIBREF1 , BIBREF2 , recent research has argued that these models are highly data-inefficient, and underper... | highly data-inefficient, underperform phrase-based statistical machine translation |
5679fabeadf680e35a4f7b092d39e8638dca6b4d | 5679fabeadf680e35a4f7b092d39e8638dca6b4d_0 | Q: Does the paper report the results of previous models applied to the same tasks?
Text: Introduction ::: Background
Over the past two decades, the rise of social media and the digitization of news and discussion platforms have radically transformed how individuals and groups create, process and share news and informat... | Yes |
5679fabeadf680e35a4f7b092d39e8638dca6b4d | 5679fabeadf680e35a4f7b092d39e8638dca6b4d_1 | Q: Does the paper report the results of previous models applied to the same tasks?
Text: Introduction ::: Background
Over the past two decades, the rise of social media and the digitization of news and discussion platforms have radically transformed how individuals and groups create, process and share news and informat... | No |
a939a53cabb4893b2fd82996f3dbe8688fdb7bbb | a939a53cabb4893b2fd82996f3dbe8688fdb7bbb_0 | Q: How is the quality of the discussion evaluated?
Text: Introduction ::: Background
Over the past two decades, the rise of social media and the digitization of news and discussion platforms have radically transformed how individuals and groups create, process and share news and information. As Alan Rusbridger, former-... | Unanswerable |
8b99767620fd4efe51428b68841cc3ec06699280 | 8b99767620fd4efe51428b68841cc3ec06699280_0 | Q: What is the technique used for text analysis and mining?
Text: Introduction ::: Background
Over the past two decades, the rise of social media and the digitization of news and discussion platforms have radically transformed how individuals and groups create, process and share news and information. As Alan Rusbridger... | Unanswerable |
312417675b3dc431eb7e7b16a917b7fed98d4376 | 312417675b3dc431eb7e7b16a917b7fed98d4376_0 | Q: What are the causal mapping methods employed?
Text: Introduction ::: Background
Over the past two decades, the rise of social media and the digitization of news and discussion platforms have radically transformed how individuals and groups create, process and share news and information. As Alan Rusbridger, former-ed... | Axelrod's causal mapping method |
792d7b579cbf7bfad8fe125b0d66c2059a174cf9 | 792d7b579cbf7bfad8fe125b0d66c2059a174cf9_0 | Q: What is the previous work's model?
Text: Introduction
Hinglish is a linguistic blend of Hindi (very widely spoken language in India) and English (an associate language of urban areas) and is spoken by upwards of 350 million people in India. While the name is based on the Hindi language, it does not refer exclusively... | Ternary Trans-CNN |
44a2a8e187f8adbd7d63a51cd2f9d2d324d0c98d | 44a2a8e187f8adbd7d63a51cd2f9d2d324d0c98d_0 | Q: What dataset is used?
Text: Introduction
Hinglish is a linguistic blend of Hindi (very widely spoken language in India) and English (an associate language of urban areas) and is spoken by upwards of 350 million people in India. While the name is based on the Hindi language, it does not refer exclusively to Hindi, bu... | HEOT , A labelled dataset for a corresponding english tweets |
44a2a8e187f8adbd7d63a51cd2f9d2d324d0c98d | 44a2a8e187f8adbd7d63a51cd2f9d2d324d0c98d_1 | Q: What dataset is used?
Text: Introduction
Hinglish is a linguistic blend of Hindi (very widely spoken language in India) and English (an associate language of urban areas) and is spoken by upwards of 350 million people in India. While the name is based on the Hindi language, it does not refer exclusively to Hindi, bu... | HEOT |
5908d7fb6c48f975c5dfc5b19bb0765581df2b25 | 5908d7fb6c48f975c5dfc5b19bb0765581df2b25_0 | Q: How big is the dataset?
Text: Introduction
Hinglish is a linguistic blend of Hindi (very widely spoken language in India) and English (an associate language of urban areas) and is spoken by upwards of 350 million people in India. While the name is based on the Hindi language, it does not refer exclusively to Hindi, ... | 3189 rows of text messages |
5908d7fb6c48f975c5dfc5b19bb0765581df2b25 | 5908d7fb6c48f975c5dfc5b19bb0765581df2b25_1 | Q: How big is the dataset?
Text: Introduction
Hinglish is a linguistic blend of Hindi (very widely spoken language in India) and English (an associate language of urban areas) and is spoken by upwards of 350 million people in India. While the name is based on the Hindi language, it does not refer exclusively to Hindi, ... | Resulting dataset was 7934 messages for train and 700 messages for test. |
cca3301f20db16f82b5d65a102436bebc88a2026 | cca3301f20db16f82b5d65a102436bebc88a2026_0 | Q: How is the dataset collected?
Text: Introduction
Hinglish is a linguistic blend of Hindi (very widely spoken language in India) and English (an associate language of urban areas) and is spoken by upwards of 350 million people in India. While the name is based on the Hindi language, it does not refer exclusively to H... | A labelled dataset for a corresponding english tweets were also obtained from a study conducted by Davidson et al, HEOT obtained from one of the past studies done by Mathur et al |
cfd67b9eeb10e5ad028097d192475d21d0b6845b | cfd67b9eeb10e5ad028097d192475d21d0b6845b_0 | Q: Was each text augmentation technique experimented individually?
Text: Introduction
Hinglish is a linguistic blend of Hindi (very widely spoken language in India) and English (an associate language of urban areas) and is spoken by upwards of 350 million people in India. While the name is based on the Hindi language, ... | No |
e1c681280b5667671c7f78b1579d0069cba72b0e | e1c681280b5667671c7f78b1579d0069cba72b0e_0 | Q: What models do previous work use?
Text: Introduction
Hinglish is a linguistic blend of Hindi (very widely spoken language in India) and English (an associate language of urban areas) and is spoken by upwards of 350 million people in India. While the name is based on the Hindi language, it does not refer exclusively ... | Ternary Trans-CNN , Hybrid multi-channel CNN and LSTM |
58d50567df71fa6c3792a0964160af390556757d | 58d50567df71fa6c3792a0964160af390556757d_0 | Q: Does the dataset contain content from various social media platforms?
Text: Introduction
Hinglish is a linguistic blend of Hindi (very widely spoken language in India) and English (an associate language of urban areas) and is spoken by upwards of 350 million people in India. While the name is based on the Hindi lang... | No |
07c79edd4c29635dbc1c2c32b8df68193b7701c6 | 07c79edd4c29635dbc1c2c32b8df68193b7701c6_0 | Q: What dataset is used?
Text: Introduction
Hinglish is a linguistic blend of Hindi (very widely spoken language in India) and English (an associate language of urban areas) and is spoken by upwards of 350 million people in India. While the name is based on the Hindi language, it does not refer exclusively to Hindi, bu... | HEOT , A labelled dataset for a corresponding english tweets |
66125cfdf11d3bf8e59728428e02021177142c3a | 66125cfdf11d3bf8e59728428e02021177142c3a_0 | Q: How they demonstrate that language-neutral component is sufficiently general in terms of modeling semantics to allow high-accuracy word-alignment?
Text: Introduction
Multilingual BERT (mBERT; BIBREF0) is gaining popularity as a contextual representation for various multilingual tasks, such as dependency parsing BIBR... | Table TABREF15 shows that word-alignment based on mBERT representations surpasses the outputs of the standard FastAlign tool even if it was provided large parallel corpus. This suggests that word-level semantics are well captured by mBERT contextual embeddings. For this task, learning an explicit projection had a negli... |
66125cfdf11d3bf8e59728428e02021177142c3a | 66125cfdf11d3bf8e59728428e02021177142c3a_1 | Q: How they demonstrate that language-neutral component is sufficiently general in terms of modeling semantics to allow high-accuracy word-alignment?
Text: Introduction
Multilingual BERT (mBERT; BIBREF0) is gaining popularity as a contextual representation for various multilingual tasks, such as dependency parsing BIBR... | explicit projection had a negligible effect on the performance |
222b2469eede9a0448e0226c6c742e8c91522af3 | 222b2469eede9a0448e0226c6c742e8c91522af3_0 | Q: Are language-specific and language-neutral components disjunctive?
Text: Introduction
Multilingual BERT (mBERT; BIBREF0) is gaining popularity as a contextual representation for various multilingual tasks, such as dependency parsing BIBREF1, BIBREF2, cross-lingual natural language inference (XNLI) or named-entity re... | No |
6f8386ad64dce3a20bc75165c5c7591df8f419cf | 6f8386ad64dce3a20bc75165c5c7591df8f419cf_0 | Q: How they show that mBERT representations can be split into a language-specific component and a language-neutral component?
Text: Introduction
Multilingual BERT (mBERT; BIBREF0) is gaining popularity as a contextual representation for various multilingual tasks, such as dependency parsing BIBREF1, BIBREF2, cross-ling... | We thus try to remove the language-specific information from the representations by centering the representations of sentences in each language so that their average lies at the origin of the vector space. |
81dc39ee6cdacf90d5f0f62134bf390a29146c65 | 81dc39ee6cdacf90d5f0f62134bf390a29146c65_0 | Q: What challenges this work presents that must be solved to build better language-neutral representations?
Text: Introduction
Multilingual BERT (mBERT; BIBREF0) is gaining popularity as a contextual representation for various multilingual tasks, such as dependency parsing BIBREF1, BIBREF2, cross-lingual natural langua... | contextual embeddings do not represent similar semantic phenomena similarly and therefore they are not directly usable for zero-shot cross-lingual tasks |
eeaceee98ef1f6c971dac7b0b8930ee8060d71c2 | eeaceee98ef1f6c971dac7b0b8930ee8060d71c2_0 | Q: What approaches they propose?
Text: Introduction
Fueled by recent advances in deep-learning and language processing, NLP systems are increasingly being used for prediction and decision-making in many fields BIBREF0, including sensitive ones such as health, commerce and law BIBREF1. Unfortunately, these highly flexib... | Across models and tasks: The degree (as grayscale) of faithfulness at the level of specific models and tasks., Across input space: The degree of faithfulness at the level of subspaces of the input space, such as neighborhoods of similar inputs, or singular inputs themselves. |
3371d586a3a81de1552d90459709c57c0b1a2594 | 3371d586a3a81de1552d90459709c57c0b1a2594_0 | Q: What faithfulness criteria does they propose?
Text: Introduction
Fueled by recent advances in deep-learning and language processing, NLP systems are increasingly being used for prediction and decision-making in many fields BIBREF0, including sensitive ones such as health, commerce and law BIBREF1. Unfortunately, the... | Across models and tasks: The degree (as grayscale) of faithfulness at the level of specific models and tasks., Across input space: The degree of faithfulness at the level of subspaces of the input space, such as neighborhoods of similar inputs, or singular inputs themselves. |
d4b9cdb4b2dfda1e0d96ab6c3b5e2157fd52685e | d4b9cdb4b2dfda1e0d96ab6c3b5e2157fd52685e_0 | Q: Which are three assumptions in current approaches for defining faithfulness?
Text: Introduction
Fueled by recent advances in deep-learning and language processing, NLP systems are increasingly being used for prediction and decision-making in many fields BIBREF0, including sensitive ones such as health, commerce and ... | Two models will make the same predictions if and only if they use the same reasoning process., On similar inputs, the model makes similar decisions if and only if its reasoning is similar., Certain parts of the input are more important to the model reasoning than others. Moreover, the contributions of different parts o... |
d4b9cdb4b2dfda1e0d96ab6c3b5e2157fd52685e | d4b9cdb4b2dfda1e0d96ab6c3b5e2157fd52685e_1 | Q: Which are three assumptions in current approaches for defining faithfulness?
Text: Introduction
Fueled by recent advances in deep-learning and language processing, NLP systems are increasingly being used for prediction and decision-making in many fields BIBREF0, including sensitive ones such as health, commerce and ... | Two models will make the same predictions if and only if they use the same reasoning process., On similar inputs, the model makes similar decisions if and only if its reasoning is similar., Certain parts of the input are more important to the model reasoning than others. Moreover, the contributions of different parts o... |
2a859e80d8647923181cb2d8f9a2c67b1c3f4608 | 2a859e80d8647923181cb2d8f9a2c67b1c3f4608_0 | Q: Which are key points in guidelines for faithfulness evaluation?
Text: Introduction
Fueled by recent advances in deep-learning and language processing, NLP systems are increasingly being used for prediction and decision-making in many fields BIBREF0, including sensitive ones such as health, commerce and law BIBREF1. ... | Be explicit in what you evaluate., Faithfulness evaluation should not involve human-judgement on the quality of interpretation., Faithfulness evaluation should not involve human-provided gold labels., Do not trust βinherent interpretabilityβ claims., Faithfulness evaluation of IUI systems should not rely on user perfor... |
aceac4ad16ffe1af0f01b465919b1d4422941a6b | aceac4ad16ffe1af0f01b465919b1d4422941a6b_0 | Q: Did they use the state-of-the-art model to analyze the attention?
Text: Introduction
Deep learning has achieved tremendous success for many NLP tasks. However, unlike traditional methods that provide optimized weights for human understandable features, the behavior of deep learning models is much harder to interpret... | we provide an extensive analysis of the state-of-the-art model |
f7070b2e258beac9b09514be2bfcc5a528cc3a0e | f7070b2e258beac9b09514be2bfcc5a528cc3a0e_0 | Q: What is the performance of their model?
Text: Introduction
Deep learning has achieved tremendous success for many NLP tasks. However, unlike traditional methods that provide optimized weights for human understandable features, the behavior of deep learning models is much harder to interpret. Due to the high dimensio... | Unanswerable |
f7070b2e258beac9b09514be2bfcc5a528cc3a0e | f7070b2e258beac9b09514be2bfcc5a528cc3a0e_1 | Q: What is the performance of their model?
Text: Introduction
Deep learning has achieved tremendous success for many NLP tasks. However, unlike traditional methods that provide optimized weights for human understandable features, the behavior of deep learning models is much harder to interpret. Due to the high dimensio... | Unanswerable |
2efdcebebeb970021233553104553205ce5d6567 | 2efdcebebeb970021233553104553205ce5d6567_0 | Q: How many layers are there in their model?
Text: Introduction
Deep learning has achieved tremendous success for many NLP tasks. However, unlike traditional methods that provide optimized weights for human understandable features, the behavior of deep learning models is much harder to interpret. Due to the high dimens... | two LSTM layers |
4fa851d91388f0803e33f6cfae519548598cd37c | 4fa851d91388f0803e33f6cfae519548598cd37c_0 | Q: Did they compare with gradient-based methods?
Text: Introduction
Deep learning has achieved tremendous success for many NLP tasks. However, unlike traditional methods that provide optimized weights for human understandable features, the behavior of deep learning models is much harder to interpret. Due to the high di... | Unanswerable |
a891039441e008f1fd0a227dbed003f76c140737 | a891039441e008f1fd0a227dbed003f76c140737_0 | Q: What MC abbreviate for?
Text: Introduction
Enabling computers to understand given documents and answer questions about their content has recently attracted intensive interest, including but not limited to the efforts as in BIBREF0 , BIBREF1 , BIBREF2 , BIBREF3 , BIBREF4 , BIBREF5 . Many specific problems such as mac... | machine comprehension |
73738e42d488b32c9db89ac8adefc75403fa2653 | 73738e42d488b32c9db89ac8adefc75403fa2653_0 | Q: how much of improvement the adaptation model can get?
Text: Introduction
Enabling computers to understand given documents and answer questions about their content has recently attracted intensive interest, including but not limited to the efforts as in BIBREF0 , BIBREF1 , BIBREF2 , BIBREF3 , BIBREF4 , BIBREF5 . Many... | 69.10%/78.38% |
6c8bd7fa1cfb1b2bbeb011cc9c712dceac0c8f06 | 6c8bd7fa1cfb1b2bbeb011cc9c712dceac0c8f06_0 | Q: what is the architecture of the baseline model?
Text: Introduction
Enabling computers to understand given documents and answer questions about their content has recently attracted intensive interest, including but not limited to the efforts as in BIBREF0 , BIBREF1 , BIBREF2 , BIBREF3 , BIBREF4 , BIBREF5 . Many speci... | word embedding, input encoder, alignment, aggregation, and prediction. |
6c8bd7fa1cfb1b2bbeb011cc9c712dceac0c8f06 | 6c8bd7fa1cfb1b2bbeb011cc9c712dceac0c8f06_1 | Q: what is the architecture of the baseline model?
Text: Introduction
Enabling computers to understand given documents and answer questions about their content has recently attracted intensive interest, including but not limited to the efforts as in BIBREF0 , BIBREF1 , BIBREF2 , BIBREF3 , BIBREF4 , BIBREF5 . Many speci... | Our baseline model is composed of the following typical components: word embedding, input encoder, alignment, aggregation, and prediction. |
fa218b297d9cdcae238cef71096752ce27ca8f4a | fa218b297d9cdcae238cef71096752ce27ca8f4a_0 | Q: What is the exact performance on SQUAD?
Text: Introduction
Enabling computers to understand given documents and answer questions about their content has recently attracted intensive interest, including but not limited to the efforts as in BIBREF0 , BIBREF1 , BIBREF2 , BIBREF3 , BIBREF4 , BIBREF5 . Many specific prob... | Our model achieves a 68.73% EM score and 77.39% F1 score |
ff28d34d1aaa57e7ad553dba09fc924dc21dd728 | ff28d34d1aaa57e7ad553dba09fc924dc21dd728_0 | Q: What are their correlation results?
Text: Introduction
Quality Estimation (QE) is a term used in machine translation (MT) to refer to methods that measure the quality of automatically translated text without relying on human references BIBREF0, BIBREF1. In this study, we address QE for summarization. Our proposed mo... | High correlation results range from 0.472 to 0.936 |
ae8354e67978b7c333094c36bf9d561ca0c2d286 | ae8354e67978b7c333094c36bf9d561ca0c2d286_0 | Q: What dataset do they use?
Text: Introduction
Quality Estimation (QE) is a term used in machine translation (MT) to refer to methods that measure the quality of automatically translated text without relying on human references BIBREF0, BIBREF1. In this study, we address QE for summarization. Our proposed model, Sum-Q... | datasets from the NIST DUC-05, DUC-06 and DUC-07 shared tasks |
02348ab62957cb82067c589769c14d798b1ceec7 | 02348ab62957cb82067c589769c14d798b1ceec7_0 | Q: What simpler models do they look at?
Text: Introduction
Quality Estimation (QE) is a term used in machine translation (MT) to refer to methods that measure the quality of automatically translated text without relying on human references BIBREF0, BIBREF1. In this study, we address QE for summarization. Our proposed m... | BiGRU s with attention, ROUGE, Language model (LM), Next sentence prediction |
02348ab62957cb82067c589769c14d798b1ceec7 | 02348ab62957cb82067c589769c14d798b1ceec7_1 | Q: What simpler models do they look at?
Text: Introduction
Quality Estimation (QE) is a term used in machine translation (MT) to refer to methods that measure the quality of automatically translated text without relying on human references BIBREF0, BIBREF1. In this study, we address QE for summarization. Our proposed m... | BiGRUs with attention, ROUGE, Language model, and next sentence prediction |
3748787379b3a7d222c3a6254def3f5bfb93a60e | 3748787379b3a7d222c3a6254def3f5bfb93a60e_0 | Q: What linguistic quality aspects are addressed?
Text: Introduction
Quality Estimation (QE) is a term used in machine translation (MT) to refer to methods that measure the quality of automatically translated text without relying on human references BIBREF0, BIBREF1. In this study, we address QE for summarization. Our ... | Grammaticality, non-redundancy, referential clarity, focus, structure & coherence |
6852217163ea678f2009d4726cb6bd03cf6a8f78 | 6852217163ea678f2009d4726cb6bd03cf6a8f78_0 | Q: What benchmark datasets are used for the link prediction task?
Text: Introduction
Knowledge graphs are usually collections of factual triplesβ(head entity, relation, tail entity), which represent human knowledge in a structured way. In the past few years, we have witnessed the great achievement of knowledge graphs i... | WN18RR, FB15k-237, YAGO3-10 |
6852217163ea678f2009d4726cb6bd03cf6a8f78 | 6852217163ea678f2009d4726cb6bd03cf6a8f78_1 | Q: What benchmark datasets are used for the link prediction task?
Text: Introduction
Knowledge graphs are usually collections of factual triplesβ(head entity, relation, tail entity), which represent human knowledge in a structured way. In the past few years, we have witnessed the great achievement of knowledge graphs i... | WN18RR BIBREF26, FB15k-237 BIBREF18, YAGO3-10 BIBREF27 |
cd1ad7e18d8eef8f67224ce47f3feec02718ea1a | cd1ad7e18d8eef8f67224ce47f3feec02718ea1a_0 | Q: What are state-of-the art models for this task?
Text: Introduction
Knowledge graphs are usually collections of factual triplesβ(head entity, relation, tail entity), which represent human knowledge in a structured way. In the past few years, we have witnessed the great achievement of knowledge graphs in many areas, s... | TransE, DistMult, ComplEx, ConvE, RotatE |
9c9e90ceaba33242342a5ae7568e89fe660270d5 | 9c9e90ceaba33242342a5ae7568e89fe660270d5_0 | Q: How better does HAKE model peform than state-of-the-art methods?
Text: Introduction
Knowledge graphs are usually collections of factual triplesβ(head entity, relation, tail entity), which represent human knowledge in a structured way. In the past few years, we have witnessed the great achievement of knowledge graphs... | 0.021 higher MRR, a 2.4% higher H@1, and a 2.4% higher H@3 against RotatE, respectively, doesn't outperform the previous state-of-the-art as much as that of WN18RR and YAGO3-10, HAKE gains a 0.050 higher MRR, 6.0% higher H@1 and 4.6% higher H@3 than RotatE, respectively |
2a058f8f6bd6f8e80e8452e1dba9f8db5e3c7de8 | 2a058f8f6bd6f8e80e8452e1dba9f8db5e3c7de8_0 | Q: How are entities mapped onto polar coordinate system?
Text: Introduction
Knowledge graphs are usually collections of factual triplesβ(head entity, relation, tail entity), which represent human knowledge in a structured way. In the past few years, we have witnessed the great achievement of knowledge graphs in many ar... | radial coordinate and the angular coordinates correspond to the modulus part and the phase part, respectively |
db9021ddd4593f6fadf172710468e2fdcea99674 | db9021ddd4593f6fadf172710468e2fdcea99674_0 | Q: What additional techniques are incorporated?
Text: Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of traditional programming language. On na... | Unanswerable |
db9021ddd4593f6fadf172710468e2fdcea99674 | db9021ddd4593f6fadf172710468e2fdcea99674_1 | Q: What additional techniques are incorporated?
Text: Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of traditional programming language. On na... | incorporating coding syntax tree model |
8ea4bd4c1d8a466da386d16e4844ea932c44a412 | 8ea4bd4c1d8a466da386d16e4844ea932c44a412_0 | Q: What dataset do they use?
Text: Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of traditional programming language. On naturalness of comput... | A parallel corpus where the source is an English expression of code and the target is Python code. |
8ea4bd4c1d8a466da386d16e4844ea932c44a412 | 8ea4bd4c1d8a466da386d16e4844ea932c44a412_1 | Q: What dataset do they use?
Text: Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of traditional programming language. On naturalness of comput... | text-code parallel corpus |
92240eeab107a4f636705b88f00cefc4f0782846 | 92240eeab107a4f636705b88f00cefc4f0782846_0 | Q: Do they compare to other models?
Text: Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of traditional programming language. On naturalness of... | No |
4196d329061f5a9d147e1e77aeed6a6bd9b35d18 | 4196d329061f5a9d147e1e77aeed6a6bd9b35d18_0 | Q: What is the architecture of the system?
Text: Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of traditional programming language. On natural... | seq2seq translation |
a37e4a21ba98b0259c36deca0d298194fa611d2f | a37e4a21ba98b0259c36deca0d298194fa611d2f_0 | Q: How long are expressions in layman's language?
Text: Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of traditional programming language. On ... | Unanswerable |
321429282557e79061fe2fe02a9467f3d0118cdd | 321429282557e79061fe2fe02a9467f3d0118cdd_0 | Q: What additional techniques could be incorporated to further improve accuracy?
Text: Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of tradit... | phrase-based word embedding, Abstract Syntax Tree(AST) |
891cab2e41d6ba962778bda297592c916b432226 | 891cab2e41d6ba962778bda297592c916b432226_0 | Q: What programming language is target language?
Text: Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of traditional programming language. On n... | Python |
1eeabfde99594b8d9c6a007f50b97f7f527b0a17 | 1eeabfde99594b8d9c6a007f50b97f7f527b0a17_0 | Q: What dataset is used to measure accuracy?
Text: Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of traditional programming language. On natur... | validation data |
e96adf8466e67bd19f345578d5a6dc68fd0279a1 | e96adf8466e67bd19f345578d5a6dc68fd0279a1_0 | Q: Is text-to-image synthesis trained is suppervized or unsuppervized manner?
Text: Introduction
β (GANs), and the variations that are now being proposed is the most interesting idea in the last 10 years in ML, in my opinion.β (2016)
β Yann LeCun
A picture is worth a thousand words! While written text provide efficient... | unsupervised |
e96adf8466e67bd19f345578d5a6dc68fd0279a1 | e96adf8466e67bd19f345578d5a6dc68fd0279a1_1 | Q: Is text-to-image synthesis trained is suppervized or unsuppervized manner?
Text: Introduction
β (GANs), and the variations that are now being proposed is the most interesting idea in the last 10 years in ML, in my opinion.β (2016)
β Yann LeCun
A picture is worth a thousand words! While written text provide efficient... | Even though natural language and image synthesis were part of several contributions on the supervised side of deep learning, unsupervised learning saw recently a tremendous rise in input from the research community specially on two subproblems: text-based natural language and image synthesis |
c1477a6c86bd1670dd17407590948000c9a6b7c6 | c1477a6c86bd1670dd17407590948000c9a6b7c6_0 | Q: What challenges remain unresolved?
Text: Introduction
β (GANs), and the variations that are now being proposed is the most interesting idea in the last 10 years in ML, in my opinion.β (2016)
β Yann LeCun
A picture is worth a thousand words! While written text provide efficient, effective, and concise ways for commun... | give more independence to the several learning methods (e.g. less human intervention) involved in the studies, increasing the size of the output images |
e020677261d739c35c6f075cde6937d0098ace7f | e020677261d739c35c6f075cde6937d0098ace7f_0 | Q: What is the conclusion of comparison of proposed solution?
Text: Introduction
β (GANs), and the variations that are now being proposed is the most interesting idea in the last 10 years in ML, in my opinion.β (2016)
β Yann LeCun
A picture is worth a thousand words! While written text provide efficient, effective, and... | HDGAN produced relatively better visual results on the CUB and Oxford datasets while AttnGAN produced far more impressive results than the rest on the more complex COCO dataset, In terms of inception score (IS), which is the metric that was applied to majority models except DC-GAN, the results in Table TABREF48 show th... |
6389d5a152151fb05aae00b53b521c117d7b5e54 | 6389d5a152151fb05aae00b53b521c117d7b5e54_0 | Q: What is typical GAN architecture for each text-to-image synhesis group?
Text: Introduction
β (GANs), and the variations that are now being proposed is the most interesting idea in the last 10 years in ML, in my opinion.β (2016)
β Yann LeCun
A picture is worth a thousand words! While written text provide efficient, e... | Semantic Enhancement GANs: DC-GANs, MC-GAN
Resolution Enhancement GANs: StackGANs, AttnGAN, HDGAN
Diversity Enhancement GANs: AC-GAN, TAC-GAN etc.
Motion Enhancement GAGs: T2S, T2V, StoryGAN |
7fe48939ce341212c1d801095517dc552b98e7b3 | 7fe48939ce341212c1d801095517dc552b98e7b3_0 | Q: Where do they employ feature-wise sigmoid gating?
Text: Introduction
Incorporating sub-word structures like substrings, morphemes and characters to the creation of word representations significantly increases their quality as reflected both by intrinsic metrics and performance in a wide range of downstream tasks BIB... | gating mechanism acts upon each dimension of the word and character-level vectors |
65ad17f614b7345f0077424c04c94971c831585b | 65ad17f614b7345f0077424c04c94971c831585b_0 | Q: Which model architecture do they use to obtain representations?
Text: Introduction
Incorporating sub-word structures like substrings, morphemes and characters to the creation of word representations significantly increases their quality as reflected both by intrinsic metrics and performance in a wide range of downst... | BiLSTM with max pooling |
323e100a6c92d3fe503f7a93b96d821408f92109 | 323e100a6c92d3fe503f7a93b96d821408f92109_0 | Q: Which downstream sentence-level tasks do they evaluate on?
Text: Introduction
Incorporating sub-word structures like substrings, morphemes and characters to the creation of word representations significantly increases their quality as reflected both by intrinsic metrics and performance in a wide range of downstream ... | BIBREF13 , BIBREF18 |
9f89bff89cea722debc991363f0826de945bc582 | 9f89bff89cea722debc991363f0826de945bc582_0 | Q: Which similarity datasets do they use?
Text: Introduction
Incorporating sub-word structures like substrings, morphemes and characters to the creation of word representations significantly increases their quality as reflected both by intrinsic metrics and performance in a wide range of downstream tasks BIBREF0 , BIBR... | MEN, MTurk287, MTurk771, RG, RW, SimLex999, SimVerb3500, WS353, WS353R, WS353S |
9f89bff89cea722debc991363f0826de945bc582 | 9f89bff89cea722debc991363f0826de945bc582_1 | Q: Which similarity datasets do they use?
Text: Introduction
Incorporating sub-word structures like substrings, morphemes and characters to the creation of word representations significantly increases their quality as reflected both by intrinsic metrics and performance in a wide range of downstream tasks BIBREF0 , BIBR... | WS353S, SimLex999, SimVerb3500 |
735f58e28d84ee92024a36bc348cfac2ee114409 | 735f58e28d84ee92024a36bc348cfac2ee114409_0 | Q: Are there datasets with relation tuples annotated, how big are datasets available?
Text: Introduction
Distantly-supervised information extraction systems extract relation tuples with a set of pre-defined relations from text. Traditionally, researchers BIBREF0, BIBREF1, BIBREF2 use pipeline approaches where a named e... | Yes |
710fa8b3e74ee63d2acc20af19f95f7702b7ce5e | 710fa8b3e74ee63d2acc20af19f95f7702b7ce5e_0 | Q: Which one of two proposed approaches performed better in experiments?
Text: Introduction
Distantly-supervised information extraction systems extract relation tuples with a set of pre-defined relations from text. Traditionally, researchers BIBREF0, BIBREF1, BIBREF2 use pipeline approaches where a named entity recogni... | WordDecoding (WDec) model |
56123dd42cf5c77fc9a88fc311ed2e1eb672126e | 56123dd42cf5c77fc9a88fc311ed2e1eb672126e_0 | Q: What is previous work authors reffer to?
Text: Introduction
Distantly-supervised information extraction systems extract relation tuples with a set of pre-defined relations from text. Traditionally, researchers BIBREF0, BIBREF1, BIBREF2 use pipeline approaches where a named entity recognition (NER) system is used to ... | SPTree, Tagging, CopyR, HRL, GraphR, N-gram Attention |
1898f999626f9a6da637bd8b4857e5eddf2fc729 | 1898f999626f9a6da637bd8b4857e5eddf2fc729_0 | Q: How higher are F1 scores compared to previous work?
Text: Introduction
Distantly-supervised information extraction systems extract relation tuples with a set of pre-defined relations from text. Traditionally, researchers BIBREF0, BIBREF1, BIBREF2 use pipeline approaches where a named entity recognition (NER) system ... | WordDecoding (WDec) model achieves F1 scores that are $3.9\%$ and $4.1\%$ higher than HRL on the NYT29 and NYT24 datasets respectively, PtrNetDecoding (PNDec) model achieves F1 scores that are $3.0\%$ and $1.3\%$ higher than HRL on the NYT29 and NYT24 datasets respectively |
1898f999626f9a6da637bd8b4857e5eddf2fc729 | 1898f999626f9a6da637bd8b4857e5eddf2fc729_1 | Q: How higher are F1 scores compared to previous work?
Text: Introduction
Distantly-supervised information extraction systems extract relation tuples with a set of pre-defined relations from text. Traditionally, researchers BIBREF0, BIBREF1, BIBREF2 use pipeline approaches where a named entity recognition (NER) system ... | Our WordDecoding (WDec) model achieves F1 scores that are $3.9\%$ and $4.1\%$ higher than HRL on the NYT29 and NYT24 datasets respectively, In the ensemble scenario, compared to HRL, WDec achieves $4.2\%$ and $3.5\%$ higher F1 scores |
d32b6ac003cfe6277f8c2eebc7540605a60a3904 | d32b6ac003cfe6277f8c2eebc7540605a60a3904_0 | Q: what were the baselines?
Text: [block]I.1em
[block]i.1em
Learning to Rank Scientific Documents from the CrowdLearning to Rank Scientific Documents from the Crowd
-4
[1]1
Introduction
The number of biomedical research papers published has increased dramatically in recent years. As of October, 2016, PubMed houses over... | Rank by the number of times a citation is mentioned in the document, Rank by the number of times the citation is cited in the literature (citation impact). , Rank using Google Scholar Related Articles., Rank by the TF*IDF weighted cosine similarity. , ank using a learning-to-rank model trained on text similarity ranki... |
d32b6ac003cfe6277f8c2eebc7540605a60a3904 | d32b6ac003cfe6277f8c2eebc7540605a60a3904_1 | Q: what were the baselines?
Text: [block]I.1em
[block]i.1em
Learning to Rank Scientific Documents from the CrowdLearning to Rank Scientific Documents from the Crowd
-4
[1]1
Introduction
The number of biomedical research papers published has increased dramatically in recent years. As of October, 2016, PubMed houses over... | (1) Rank by the number of times a citation is mentioned in the document., (2) Rank by the number of times the citation is cited in the literature (citation impact)., (3) Rank using Google Scholar Related Articles., (4) Rank by the TF*IDF weighted cosine similarity., (5) Rank using a learning-to-rank model trained on te... |
c10f38ee97ed80484c1a70b8ebba9b1fb149bc91 | c10f38ee97ed80484c1a70b8ebba9b1fb149bc91_0 | Q: what is the supervised model they developed?
Text: [block]I.1em
[block]i.1em
Learning to Rank Scientific Documents from the CrowdLearning to Rank Scientific Documents from the Crowd
-4
[1]1
Introduction
The number of biomedical research papers published has increased dramatically in recent years. As of October, 2016... | SVMRank |
340501f23ddc0abe344a239193abbaaab938cc3a | 340501f23ddc0abe344a239193abbaaab938cc3a_0 | Q: what is the size of this built corpus?
Text: [block]I.1em
[block]i.1em
Learning to Rank Scientific Documents from the CrowdLearning to Rank Scientific Documents from the Crowd
-4
[1]1
Introduction
The number of biomedical research papers published has increased dramatically in recent years. As of October, 2016, PubM... | 90 annotated documents with 5 citations each ranked 1 to 5, where 1 is least relevant and 5 is most relevant for a total of 450 annotated citations |
fbb85cbd41de6d2818e77e8f8d4b91e431931faa | fbb85cbd41de6d2818e77e8f8d4b91e431931faa_0 | Q: what crowdsourcing platform is used?
Text: [block]I.1em
[block]i.1em
Learning to Rank Scientific Documents from the CrowdLearning to Rank Scientific Documents from the Crowd
-4
[1]1
Introduction
The number of biomedical research papers published has increased dramatically in recent years. As of October, 2016, PubMed... | asked the authors to rank by closeness five citations we selected from their paper |
1951cde612751410355610074c3c69cec94824c2 | 1951cde612751410355610074c3c69cec94824c2_0 | Q: Which deep learning model performed better?
Text: Introduction
In recent years, social media, forums, blogs and other forms of online communication tools have radically affected everyday life, especially how people express their opinions and comments. The extraction of useful information (such as people's opinion ab... | autoencoders |
1951cde612751410355610074c3c69cec94824c2 | 1951cde612751410355610074c3c69cec94824c2_1 | Q: Which deep learning model performed better?
Text: Introduction
In recent years, social media, forums, blogs and other forms of online communication tools have radically affected everyday life, especially how people express their opinions and comments. The extraction of useful information (such as people's opinion ab... | CNN |
4140d8b5a78aea985546aa1e323de12f63d24add | 4140d8b5a78aea985546aa1e323de12f63d24add_0 | Q: By how much did the results improve?
Text: Introduction
In recent years, social media, forums, blogs and other forms of online communication tools have radically affected everyday life, especially how people express their opinions and comments. The extraction of useful information (such as people's opinion about com... | Unanswerable |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.