uid int64 4 318k | paper_url stringlengths 39 81 | arxiv_id stringlengths 9 16 ⌀ | title stringlengths 6 365 | abstract stringlengths 0 7.27k | url_abs stringlengths 17 601 | url_pdf stringlengths 21 819 | proceeding stringlengths 7 1.03k ⌀ | authors sequence | tasks sequence | date float64 422B 1,672B ⌀ | methods list | __index_level_0__ int64 1 197k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
172,339 | https://paperswithcode.com/paper/maximum-a-posteriori-signal-recovery-for | 2010.15682 | Maximum a posteriori signal recovery for optical coherence tomography angiography image generation and denoising | Optical coherence tomography angiography (OCTA) is a novel and clinically promising imaging modality to image retinal and sub-retinal vasculature. Based on repeated optical coherence tomography (OCT) scans, intensity changes are observed over time and used to compute OCTA image data. OCTA data are prone to noise and ar... | https://arxiv.org/abs/2010.15682v1 | https://arxiv.org/pdf/2010.15682v1.pdf | null | [
"Lennart Husvogt",
"Stefan B. Ploner",
"Siyu Chen",
"Daniel Stromer",
"Julia Schottenhamml",
"A. Yasin Alibhai",
"Eric Moult",
"Nadia K. Waheed",
"James G. Fujimoto",
"Andreas Maier"
] | [
"Denoising",
"Image Generation"
] | 1,603,929,600,000 | [] | 25,324 |
3,841 | https://paperswithcode.com/paper/code-completion-with-neural-attention-and | 1711.09573 | Code Completion with Neural Attention and Pointer Networks | Intelligent code completion has become an essential research task to
accelerate modern software development. To facilitate effective code completion
for dynamically-typed programming languages, we apply neural language models by
learning from large codebases, and develop a tailored attention mechanism for
code completi... | http://arxiv.org/abs/1711.09573v2 | http://arxiv.org/pdf/1711.09573v2.pdf | null | [
"Jian Li",
"Yue Wang",
"Michael R. Lyu",
"Irwin King"
] | [
"Code Completion"
] | 1,511,740,800,000 | [] | 140,067 |
151,672 | https://paperswithcode.com/paper/naist-s-machine-translation-systems-for-iwslt | null | NAIST's Machine Translation Systems for IWSLT 2020 Conversational Speech Translation Task | This paper describes NAIST{'}s NMT system submitted to the IWSLT 2020 conversational speech translation task. We focus on the translation disfluent speech transcripts that include ASR errors and non-grammatical utterances. We tried a domain adaptation method by transferring the styles of out-of-domain data (United Nati... | https://aclanthology.org/2020.iwslt-1.21 | https://aclanthology.org/2020.iwslt-1.21.pdf | WS 2020 7 | [
"Ryo Fukuda",
"Katsuhito Sudoh",
"Satoshi Nakamura"
] | [
"Domain Adaptation",
"Machine Translation",
"Style Transfer"
] | 1,593,561,600,000 | [] | 124,264 |
124,349 | https://paperswithcode.com/paper/influence-aware-memory-for-deep-reinforcement-1 | 1911.07643 | Influence-aware Memory Architectures for Deep Reinforcement Learning | Due to its perceptual limitations, an agent may have too little information about the state of the environment to act optimally. In such cases, it is important to keep track of the observation history to uncover hidden state. Recent deep reinforcement learning methods use recurrent neural networks (RNN) to memorize pas... | https://arxiv.org/abs/1911.07643v4 | https://arxiv.org/pdf/1911.07643v4.pdf | null | [
"Miguel Suau",
"Jinke He",
"Elena Congeduti",
"Rolf A. N. Starre",
"Aleksander Czechowski",
"Frans A. Oliehoek"
] | [
"reinforcement-learning"
] | 1,574,035,200,000 | [] | 166,238 |
101,001 | https://paperswithcode.com/paper/deep-unified-multimodal-embeddings-for | 1905.07075 | Deep Unified Multimodal Embeddings for Understanding both Content and Users in Social Media Networks | There has been an explosion of multimodal content generated on social media networks in the last few years, which has necessitated a deeper understanding of social media content and user behavior. We present a novel content-independent content-user-reaction model for social multimedia content analysis. Compared to prio... | https://arxiv.org/abs/1905.07075v3 | https://arxiv.org/pdf/1905.07075v3.pdf | null | [
"Karan Sikka",
"Lucas Van Bramer",
"Ajay Divakaran"
] | [
"Cross-Modal Retrieval"
] | 1,558,051,200,000 | [] | 108,730 |
105,815 | https://paperswithcode.com/paper/few-shot-learning-with-per-sample-rich | 1906.03859 | Few-Shot Learning with Per-Sample Rich Supervision | Learning with few samples is a major challenge for parameter-rich models like deep networks. In contrast, people learn complex new concepts even from very few examples, suggesting that the sample complexity of learning can often be reduced. Many approaches to few-shot learning build on transferring a representation fro... | https://arxiv.org/abs/1906.03859v1 | https://arxiv.org/pdf/1906.03859v1.pdf | null | [
"Roman Visotsky",
"Yuval Atzmon",
"Gal Chechik"
] | [
"Few-Shot Learning",
"Classification",
"Meta-Learning",
"Scene Classification"
] | 1,560,124,800,000 | [] | 81,212 |
9,528 | https://paperswithcode.com/paper/constrained-image-generation-using-binarized | 1802.08795 | Constrained Image Generation Using Binarized Neural Networks with Decision Procedures | We consider the problem of binary image generation with given properties.
This problem arises in a number of practical applications, including generation
of artificial porous medium for an electrode of lithium-ion batteries, for
composed materials, etc. A generated image represents a porous medium and, as
such, it is s... | http://arxiv.org/abs/1802.08795v1 | http://arxiv.org/pdf/1802.08795v1.pdf | null | [
"Svyatoslav Korneev",
"Nina Narodytska",
"Luca Pulina",
"Armando Tacchella",
"Nikolaj Bjorner",
"Mooly Sagiv"
] | [
"Image Generation"
] | 1,519,430,400,000 | [] | 121,913 |
63,916 | https://paperswithcode.com/paper/learning-to-predict-denotational | null | Learning to Predict Denotational Probabilities For Modeling Entailment | We propose a framework that captures the denotational probabilities of words and phrases by embedding them in a vector space, and present a method to induce such an embedding from a dataset of denotational probabilities. We show that our model successfully predicts denotational probabilities for unseen phrases, and tha... | https://aclanthology.org/E17-1068 | https://aclanthology.org/E17-1068.pdf | EACL 2017 4 | [
"Alice Lai",
"Julia Hockenmaier"
] | [
"Coreference Resolution",
"Natural Language Inference"
] | 1,491,004,800,000 | [] | 74,007 |
201,003 | https://paperswithcode.com/paper/adversarially-guided-actor-critic-1 | 2102.04376 | Adversarially Guided Actor-Critic | Despite definite success in deep reinforcement learning problems, actor-critic algorithms are still confronted with sample inefficiency in complex environments, particularly in tasks where efficient exploration is a bottleneck. These methods consider a policy (the actor) and a value function (the critic) whose respecti... | https://arxiv.org/abs/2102.04376v1 | https://arxiv.org/pdf/2102.04376v1.pdf | ICLR 2021 1 | [
"Yannis Flet-Berliac",
"Johan Ferret",
"Olivier Pietquin",
"Philippe Preux",
"Matthieu Geist"
] | [
"Efficient Exploration"
] | 1,612,742,400,000 | [] | 50,348 |
75,241 | https://paperswithcode.com/paper/generative-entity-networks-disentangling | null | Generative Entity Networks: Disentangling Entitites and Attributes in Visual Scenes using Partial Natural Language Descriptions | Generative image models have made significant progress in the last few years, and are now able to generate low-resolution images which sometimes look realistic. However the state-of-the-art models utilize fully entangled latent representations where small changes to a single neuron can effect every output pixel in rela... | https://openreview.net/forum?id=BJInMmWC- | https://openreview.net/pdf?id=BJInMmWC- | ICLR 2018 1 | [
"Charlie Nash",
"Sebastian Nowozin",
"Nate Kushman"
] | [
"Question Answering"
] | 1,514,764,800,000 | [
{
"code_snippet_url": "https://github.com/L1aoXingyu/pytorch-beginner/blob/9c86be785c7c318a09cf29112dd1f1a58613239b/08-AutoEncoder/simple_autoencoder.py#L38",
"description": "An **Autoencoder** is a bottleneck architecture that turns a high-dimensional input into a latent low-dimensional code (encoder), and... | 5,299 |
298,219 | https://paperswithcode.com/paper/where-are-my-neighbors-exploiting-patches | 2206.00481 | Where are my Neighbors? Exploiting Patches Relations in Self-Supervised Vision Transformer | Vision Transformers (ViTs) enabled the use of transformer architecture on vision tasks showing impressive performances when trained on big datasets. However, on relatively small datasets, ViTs are less accurate given their lack of inductive bias. To this end, we propose a simple but still effective self-supervised lear... | https://arxiv.org/abs/2206.00481v1 | https://arxiv.org/pdf/2206.00481v1.pdf | null | [
"Guglielmo Camporese",
"Elena Izzo",
"Lamberto Ballan"
] | [
"Inductive Bias",
"Self-Supervised Learning"
] | 1,654,041,600,000 | [] | 192,503 |
197,581 | https://paperswithcode.com/paper/fakebuster-a-deepfakes-detection-tool-for | 2101.03321 | FakeBuster: A DeepFakes Detection Tool for Video Conferencing Scenarios | This paper proposes a new DeepFake detector FakeBuster for detecting impostors during video conferencing and manipulated faces on social media. FakeBuster is a standalone deep learning based solution, which enables a user to detect if another person's video is manipulated or spoofed during a video conferencing based me... | https://arxiv.org/abs/2101.03321v1 | https://arxiv.org/pdf/2101.03321v1.pdf | null | [
"Vineet Mehta",
"Parul Gupta",
"Ramanathan Subramanian",
"Abhinav Dhall"
] | [
"Face Swapping"
] | 1,610,150,400,000 | [] | 5,388 |
168,778 | https://paperswithcode.com/paper/a-deep-learning-based-interactive-sketching | 2010.04413 | A deep learning based interactive sketching system for fashion images design | In this work, we propose an interactive system to design diverse high-quality garment images from fashion sketches and the texture information. The major challenge behind this system is to generate high-quality and detailed texture according to the user-provided texture information. Prior works mainly use the texture p... | https://arxiv.org/abs/2010.04413v1 | https://arxiv.org/pdf/2010.04413v1.pdf | null | [
"Yao Li",
"Xianggang Yu",
"Xiaoguang Han",
"Nianjuan Jiang",
"Kui Jia",
"Jiangbo Lu"
] | [
"Intrinsic Image Decomposition",
"Texture Synthesis"
] | 1,602,201,600,000 | [] | 17,119 |
227,557 | https://paperswithcode.com/paper/reinforcement-learning-based-dialogue-guided | 2106.12384 | Reinforcement Learning-based Dialogue Guided Event Extraction to Exploit Argument Relations | Event extraction is a fundamental task for natural language processing. Finding the roles of event arguments like event participants is essential for event extraction. However, doing so for real-life event descriptions is challenging because an argument's role often varies in different contexts. While the relationship ... | https://arxiv.org/abs/2106.12384v2 | https://arxiv.org/pdf/2106.12384v2.pdf | null | [
"Qian Li",
"Hao Peng",
"JianXin Li",
"Jia Wu",
"Yuanxing Ning",
"Lihong Wang",
"Philip S. Yu",
"Zheng Wang"
] | [
"Event Extraction",
"Incremental Learning",
"reinforcement-learning"
] | 1,624,406,400,000 | [] | 134,800 |
26,039 | https://paperswithcode.com/paper/adversarial-examples-for-generative-models | 1702.06832 | Adversarial examples for generative models | We explore methods of producing adversarial examples on deep generative
models such as the variational autoencoder (VAE) and the VAE-GAN. Deep learning
architectures are known to be vulnerable to adversarial examples, but previous
work has focused on the application of adversarial examples to classification
tasks. Deep... | http://arxiv.org/abs/1702.06832v1 | http://arxiv.org/pdf/1702.06832v1.pdf | null | [
"Jernej Kos",
"Ian Fischer",
"Dawn Song"
] | [
"Classification",
"Classification"
] | 1,487,721,600,000 | [
{
"code_snippet_url": "https://github.com/L1aoXingyu/pytorch-beginner/blob/9c86be785c7c318a09cf29112dd1f1a58613239b/08-AutoEncoder/simple_autoencoder.py#L38",
"description": "An **Autoencoder** is a bottleneck architecture that turns a high-dimensional input into a latent low-dimensional code (encoder), and... | 153,759 |
279,975 | https://paperswithcode.com/paper/cake-a-scalable-commonsense-aware-framework | 2202.13785 | CAKE: A Scalable Commonsense-Aware Framework For Multi-View Knowledge Graph Completion | Knowledge graphs store a large number of factual triples while they are still incomplete, inevitably. The previous knowledge graph completion (KGC) models predict missing links between entities merely relying on fact-view data, ignoring the valuable commonsense knowledge. The previous knowledge graph embedding (KGE) te... | https://arxiv.org/abs/2202.13785v3 | https://arxiv.org/pdf/2202.13785v3.pdf | ACL 2022 5 | [
"Guanglin Niu",
"Bo Li",
"Yongfei Zhang",
"ShiLiang Pu"
] | [
"Graph Embedding",
"Knowledge Graph Completion",
"Knowledge Graph Embedding",
"Knowledge Graphs",
"Link Prediction"
] | 1,645,747,200,000 | [] | 53,744 |
184,651 | https://paperswithcode.com/paper/mufold-betaturn-a-deep-dense-inception | 1808.04322 | MUFold-BetaTurn: A Deep Dense Inception Network for Protein Beta-Turn Prediction | Beta-turn prediction is useful in protein function studies and experimental
design. Although recent approaches using machine-learning techniques such as
SVM, neural networks, and K-NN have achieved good results for beta-turn
pre-diction, there is still significant room for improvement. As previous
predictors utilized f... | http://arxiv.org/abs/1808.04322v1 | http://arxiv.org/pdf/1808.04322v1.pdf | null | [] | [
"Experimental Design",
"Feature Engineering"
] | 1,534,118,400,000 | [] | 97,061 |
137,241 | https://paperswithcode.com/paper/pool-based-unsupervised-active-learning-for | 2003.07658 | Pool-Based Unsupervised Active Learning for Regression Using Iterative Representativeness-Diversity Maximization (iRDM) | Active learning (AL) selects the most beneficial unlabeled samples to label, and hence a better machine learning model can be trained from the same number of labeled samples. Most existing active learning for regression (ALR) approaches are supervised, which means the sampling process must use some label information, o... | https://arxiv.org/abs/2003.07658v2 | https://arxiv.org/pdf/2003.07658v2.pdf | null | [
"Ziang Liu",
"Xue Jiang",
"Hanbin Luo",
"Weili Fang",
"Jiajing Liu",
"Dongrui Wu"
] | [
"Active Learning"
] | 1,584,403,200,000 | [
{
"code_snippet_url": null,
"description": "**Linear Regression** is a method for modelling a relationship between a dependent variable and independent variables. These models can be fit with numerous approaches. The most common is *least squares*, where we minimize the mean square error between the predict... | 120,211 |
293,867 | https://paperswithcode.com/paper/cross-modal-cloze-task-a-new-task-to-brain-to | null | Cross-Modal Cloze Task: A New Task to Brain-to-Word Decoding | Decoding language from non-invasive brain activity has attracted increasing attention from both researchers in neuroscience and natural language processing. Due to the noisy nature of brain recordings, existing work has simplified brain-to-word decoding as a binary classification task which is to discriminate a brain s... | https://aclanthology.org/2022.findings-acl.54 | https://aclanthology.org/2022.findings-acl.54.pdf | Findings (ACL) 2022 5 | [
"Shuxian Zou",
"Shaonan Wang",
"Jiajun Zhang",
"Chengqing Zong"
] | [
"Language Modelling"
] | 1,651,363,200,000 | [] | 154,832 |
227,847 | https://paperswithcode.com/paper/bayesian-inference-in-high-dimensional-time-1 | 2106.13379 | Bayesian Inference in High-Dimensional Time-Serieswith the Orthogonal Stochastic Linear Mixing Model | Many modern time-series datasets contain large numbers of output response variables sampled for prolonged periods of time. For example, in neuroscience, the activities of 100s-1000's of neurons are recorded during behaviors and in response to sensory stimuli. Multi-output Gaussian process models leverage the nonparamet... | https://arxiv.org/abs/2106.13379v2 | https://arxiv.org/pdf/2106.13379v2.pdf | null | [
"Rui Meng",
"Kristofer Bouchard"
] | [
"Bayesian Inference",
"Gaussian Processes",
"Time Series"
] | 1,624,579,200,000 | [
{
"code_snippet_url": null,
"description": "**Gaussian Processes** are non-parametric models for approximating functions. They rely upon a measure of similarity between points (the kernel function) to predict the value for an unseen point from training data. The models are fully probabilistic so uncertainty... | 102,352 |
236,184 | https://paperswithcode.com/paper/modulating-language-models-with-emotions | 2108.07886 | Modulating Language Models with Emotions | Generating context-aware language that embodies diverse emotions is an important step towards building empathetic NLP systems. In this paper, we propose a formulation of modulated layer normalization -- a technique inspired by computer vision -- that allows us to use large-scale language models for emotional response g... | https://arxiv.org/abs/2108.07886v1 | https://arxiv.org/pdf/2108.07886v1.pdf | Findings (ACL) 2021 8 | [
"Ruibo Liu",
"Jason Wei",
"Chenyan Jia",
"Soroush Vosoughi"
] | [
"Response Generation"
] | 1,629,158,400,000 | [
{
"code_snippet_url": "https://github.com/CyberZHG/torch-layer-normalization/blob/89f405b60f53f85da6f03fe685c190ef394ce50c/torch_layer_normalization/layer_normalization.py#L8",
"description": "Unlike [batch normalization](https://paperswithcode.com/method/batch-normalization), **Layer Normalization** direct... | 97,900 |
290,977 | https://paperswithcode.com/paper/defending-against-person-hiding-adversarial | 2204.13004 | Defending Against Person Hiding Adversarial Patch Attack with a Universal White Frame | Object detection has attracted great attention in the computer vision area and has emerged as an indispensable component in many vision systems. In the era of deep learning, many high-performance object detection networks have been proposed. Although these detection networks show high performance, they are vulnerable t... | https://arxiv.org/abs/2204.13004v1 | https://arxiv.org/pdf/2204.13004v1.pdf | null | [
"Youngjoon Yu",
"Hong Joo Lee",
"Hakmin Lee",
"Yong Man Ro"
] | [
"Autonomous Driving",
"Object Detection",
"Object Detection"
] | 1,651,017,600,000 | [] | 191,602 |
290,047 | https://paperswithcode.com/paper/towards-fewer-labels-support-pair-active | 2204.10008 | Towards Fewer Labels: Support Pair Active Learning for Person Re-identification | Supervised-learning based person re-identification (re-id) require a large amount of manual labeled data, which is not applicable in practical re-id deployment. In this work, we propose a Support Pair Active Learning (SPAL) framework to lower the manual labeling cost for large-scale person reidentification. The support... | https://arxiv.org/abs/2204.10008v1 | https://arxiv.org/pdf/2204.10008v1.pdf | null | [
"Dapeng Jin",
"Minxian Li"
] | [
"Active Learning",
"Person Re-Identification"
] | 1,650,499,200,000 | [] | 22,530 |
822 | https://paperswithcode.com/paper/addition-of-code-mixed-features-to-enhance | 1806.03821 | Addition of Code Mixed Features to Enhance the Sentiment Prediction of Song Lyrics | Sentiment analysis, also called opinion mining, is the field of study that
analyzes people's opinions,sentiments, attitudes and emotions. Songs are
important to sentiment analysis since the songs and mood are mutually dependent
on each other. Based on the selected song it becomes easy to find the mood of
the listener, ... | http://arxiv.org/abs/1806.03821v1 | http://arxiv.org/pdf/1806.03821v1.pdf | null | [
"Gangula Rama Rohit Reddy",
"Radhika Mamidi"
] | [
"Language Identification",
"Opinion Mining",
"Sentiment Analysis"
] | 1,528,675,200,000 | [] | 174,454 |
6,803 | https://paperswithcode.com/paper/multi-lingual-neural-title-generation-for-e | 1804.01041 | Multi-lingual neural title generation for e-Commerce browse pages | To provide better access of the inventory to buyers and better search engine
optimization, e-Commerce websites are automatically generating millions of
easily searchable browse pages. A browse page consists of a set of slot
name/value pairs within a given category, grouping multiple items which share
some characteristi... | http://arxiv.org/abs/1804.01041v1 | http://arxiv.org/pdf/1804.01041v1.pdf | NAACL 2018 6 | [
"Prashant Mathur",
"Nicola Ueffing",
"Gregor Leusch"
] | [
"Transfer Learning"
] | 1,522,713,600,000 | [] | 185,413 |
193,153 | https://paperswithcode.com/paper/understanding-interpretability-by-generalized | 2012.03089 | Understanding Interpretability by generalized distillation in Supervised Classification | The ability to interpret decisions taken by Machine Learning (ML) models is fundamental to encourage trust and reliability in different practical applications. Recent interpretation strategies focus on human understanding of the underlying decision mechanisms of the complex ML models. However, these strategies are rest... | https://arxiv.org/abs/2012.03089v1 | https://arxiv.org/pdf/2012.03089v1.pdf | null | [
"Adit Agarwal",
"Dr. K. K. Shukla",
"Arjan Kuijper",
"Anirban Mukhopadhyay"
] | [
"Classification",
"Classification"
] | 1,607,126,400,000 | [
{
"code_snippet_url": null,
"description": "Please enter a description about the method here",
"full_name": "Interpretability",
"introduced_year": 2000,
"main_collection": {
"area": "Computer Vision",
"description": "**Image Models** are methods that build representations of images f... | 60,649 |
313,207 | https://paperswithcode.com/paper/improving-multilayer-perceptron-mlp-based | 2208.09711 | Improving Multilayer-Perceptron(MLP)-based Network Anomaly Detection with Birch Clustering on CICIDS-2017 Dataset | Machine learning algorithms have been widely used in intrusion detection systems, including Multi-layer Perceptron (MLP). In this study, we proposed a two-stage model that combines the Birch clustering algorithm and MLP classifier to improve the performance of network anomaly multi-classification. In our proposed metho... | https://arxiv.org/abs/2208.09711v1 | https://arxiv.org/pdf/2208.09711v1.pdf | null | [
"Yuhua Yin",
"Julian Jang-Jaccard",
"Fariza Sabrina",
"Jin Kwak"
] | [
"Anomaly Detection",
"Intrusion Detection",
"pseudo label"
] | 1,660,953,600,000 | [
{
"code_snippet_url": "https://cryptoabout.info",
"description": "**k-Means Clustering** is a clustering algorithm that divides a training set into $k$ different clusters of examples that are near each other. It works by initializing $k$ different centroids {$\\mu\\left(1\\right),\\ldots,\\mu\\left(k\\right... | 92,023 |
52,195 | https://paperswithcode.com/paper/twitter-sentiment-analysis-via-bi-sense-emoji | 1807.07961 | Twitter Sentiment Analysis via Bi-sense Emoji Embedding and Attention-based LSTM | Sentiment analysis on large-scale social media data is important to bridge
the gaps between social media contents and real world activities including
political election prediction, individual and public emotional status
monitoring and analysis, and so on. Although textual sentiment analysis has
been well studied based ... | http://arxiv.org/abs/1807.07961v2 | http://arxiv.org/pdf/1807.07961v2.pdf | null | [
"Yuxiao Chen",
"Jianbo Yuan",
"Quanzeng You",
"Jiebo Luo"
] | [
"Sentiment Analysis",
"Twitter Sentiment Analysis"
] | 1,532,044,800,000 | [
{
"code_snippet_url": "https://github.com/aykutaaykut/Memory-Networks",
"description": "A **Memory Network** provides a memory component that can be read from and written to with the inference capabilities of a neural network model. The motivation is that many neural networks lack a long-term memory compone... | 87,823 |
164,737 | https://paperswithcode.com/paper/an-incentive-mechanism-for-federated-learning | 2009.10269 | An Incentive Mechanism for Federated Learning in Wireless Cellular network: An Auction Approach | Federated Learning (FL) is a distributed learning framework that can deal with the distributed issue in machine learning and still guarantee high learning performance. However, it is impractical that all users will sacrifice their resources to join the FL algorithm. This motivates us to study the incentive mechanism de... | https://arxiv.org/abs/2009.10269v1 | https://arxiv.org/pdf/2009.10269v1.pdf | null | [
"Tra Huong Thi Le",
"Nguyen H. Tran",
"Yan Kyaw Tun",
"Minh N. H. Nguyen",
"Shashi Raj Pandey",
"Zhu Han",
"Choong Seon Hong"
] | [
"Federated Learning"
] | 1,600,732,800,000 | [] | 25,683 |
314,754 | https://paperswithcode.com/paper/spoofing-aware-attention-based-asv-back-end | 2209.00423 | Spoofing-Aware Attention based ASV Back-end with Multiple Enrollment Utterances and a Sampling Strategy for the SASV Challenge 2022 | Current state-of-the-art automatic speaker verification (ASV) systems are vulnerable to presentation attacks, and several countermeasures (CMs), which distinguish bona fide trials from spoofing ones, have been explored to protect ASV. However, ASV systems and CMs are generally developed and optimized independently with... | https://arxiv.org/abs/2209.00423v1 | https://arxiv.org/pdf/2209.00423v1.pdf | null | [
"Chang Zeng",
"Lin Zhang",
"Meng Liu",
"Junichi Yamagishi"
] | [
"Speaker Verification"
] | 1,661,990,400,000 | [] | 186,256 |
256,745 | https://paperswithcode.com/paper/parbleu-augmenting-metrics-with-automatic | null | ParBLEU: Augmenting Metrics with Automatic Paraphrases for the WMT’20 Metrics Shared Task | We describe parBLEU, parCHRF++, and parESIM, which augment baseline metrics with automatically generated paraphrases produced by PRISM (Thompson and Post, 2020a), a multilingual neural machine translation system. We build on recent work studying how to improve BLEU by using diverse automatically paraphrased references ... | https://aclanthology.org/2020.wmt-1.98 | https://aclanthology.org/2020.wmt-1.98.pdf | WMT (EMNLP) 2020 11 | [
"Rachel Bawden",
"Biao Zhang",
"Andre Tättar",
"Matt Post"
] | [
"Machine Translation"
] | 1,604,188,800,000 | [] | 32,834 |
207,192 | https://paperswithcode.com/paper/learning-to-simulate-on-sparse-trajectory | 2103.11845 | Learning to Simulate on Sparse Trajectory Data | Simulation of the real-world traffic can be used to help validate the transportation policies. A good simulator means the simulated traffic is similar to real-world traffic, which often requires dense traffic trajectories (i.e., with a high sampling rate) to cover dynamic situations in the real world. However, in most ... | https://arxiv.org/abs/2103.11845v1 | https://arxiv.org/pdf/2103.11845v1.pdf | null | [
"Hua Wei",
"Chacha Chen",
"Chang Liu",
"Guanjie Zheng",
"Zhenhui Li"
] | [
"Imitation Learning"
] | 1,616,371,200,000 | [] | 148,197 |
13,588 | https://paperswithcode.com/paper/a-variational-approach-to-shape-from-shading | 1709.10354 | A Variational Approach to Shape-from-shading Under Natural Illumination | A numerical solution to shape-from-shading under natural illumination is
presented. It builds upon an augmented Lagrangian approach for solving a
generic PDE-based shape-from-shading model which handles directional or
spherical harmonic lighting, orthographic or perspective projection, and
greylevel or multi-channel im... | http://arxiv.org/abs/1709.10354v2 | http://arxiv.org/pdf/1709.10354v2.pdf | null | [
"Yvain Quéau",
"Jean Mélou",
"Fabien Castan",
"Daniel Cremers",
"Jean-Denis Durou"
] | [
"Denoising"
] | 1,506,643,200,000 | [] | 131,612 |
212,741 | https://paperswithcode.com/paper/unsupervised-learning-of-explainable-parse | 2104.04998 | Unsupervised Learning of Explainable Parse Trees for Improved Generalisation | Recursive neural networks (RvNN) have been shown useful for learning sentence representations and helped achieve competitive performance on several natural language inference tasks. However, recent RvNN-based models fail to learn simple grammar and meaningful semantics in their intermediate tree representation. In this... | https://arxiv.org/abs/2104.04998v1 | https://arxiv.org/pdf/2104.04998v1.pdf | null | [
"Atul Sahay",
"Ayush Maheshwari",
"Ritesh Kumar",
"Ganesh Ramakrishnan",
"Manjesh Kumar Hanawal",
"Kavi Arya"
] | [
"Natural Language Inference",
"Sentiment Analysis"
] | 1,618,099,200,000 | [] | 137,812 |
277,335 | https://paperswithcode.com/paper/towards-weakly-supervised-text-spotting-using | 2202.05508 | Towards Weakly-Supervised Text Spotting using a Multi-Task Transformer | Text spotting end-to-end methods have recently gained attention in the literature due to the benefits of jointly optimizing the text detection and recognition components. Existing methods usually have a distinct separation between the detection and recognition branches, requiring exact annotations for the two tasks. We... | https://arxiv.org/abs/2202.05508v2 | https://arxiv.org/pdf/2202.05508v2.pdf | CVPR 2022 1 | [
"Yair Kittenplon",
"Inbal Lavi",
"Sharon Fogel",
"Yarin Bar",
"R. Manmatha",
"Pietro Perona"
] | [
"Text Spotting"
] | 1,644,537,600,000 | [] | 6,532 |
168,919 | https://paperswithcode.com/paper/a-novel-strategy-for-covid-19-classification | 2010.05690 | COVID-19 Classification Using Staked Ensembles: A Comprehensive Analysis | The issue of COVID-19, increasing with a massive mortality rate. This led to the WHO declaring it as a pandemic. In this situation, it is crucial to perform efficient and fast diagnosis. The reverse transcript polymerase chain reaction (RTPCR) test is conducted to detect the presence of SARS-CoV-2. This test is time-co... | https://arxiv.org/abs/2010.05690v3 | https://arxiv.org/pdf/2010.05690v3.pdf | null | [
"Lalith Bharadwaj B",
"Rohit Boddeda",
"Sai Vardhan K",
"Madhu G"
] | [
"Classification"
] | 1,602,028,800,000 | [] | 2,990 |
264,422 | https://paperswithcode.com/paper/multilingual-pre-training-with-language-and | null | Multilingual pre-training with Language and Task Adaptation for Multilingual Text Style Transfer | We exploit the pre-trained seq2seq model mBART for multilingual text style transfer. Using machine translated data as well as gold aligned English sentences yields state-of-the-art results in the three target languages we consider. Besides, in view of the general scarcity of parallel data, we propose a modular approac... | https://openreview.net/forum?id=rWPLdCIiY6g | https://openreview.net/pdf?id=rWPLdCIiY6g | ACL ARR November 2021 11 | [
"Anonymous"
] | [
"Style Transfer",
"Text Style Transfer"
] | 1,637,020,800,000 | [
{
"code_snippet_url": "https://github.com/pytorch/pytorch/blob/96aaa311c0251d24decb9dc5da4957b7c590af6f/torch/nn/modules/activation.py#L329",
"description": "**Tanh Activation** is an activation function used for neural networks:\r\n\r\n$$f\\left(x\\right) = \\frac{e^{x} - e^{-x}}{e^{x} + e^{-x}}$$\r\n\r\nH... | 2,148 |
215,525 | https://paperswithcode.com/paper/discovering-an-aid-policy-to-minimize-student | 2104.10258 | Discovering an Aid Policy to Minimize Student Evasion Using Offline Reinforcement Learning | High dropout rates in tertiary education expose a lack of efficiency that causes frustration of expectations and financial waste. Predicting students at risk is not enough to avoid student dropout. Usually, an appropriate aid action must be discovered and applied in the proper time for each student. To tackle this sequ... | https://arxiv.org/abs/2104.10258v1 | https://arxiv.org/pdf/2104.10258v1.pdf | null | [
"Leandro M. de Lima",
"Renato A. Krohling"
] | [
"reinforcement-learning"
] | 1,618,876,800,000 | [
{
"code_snippet_url": "https://github.com/google/jax/blob/7f3078b70d0ed9bea6228efa420879c56f72ef69/jax/experimental/stax.py#L271-L275",
"description": "**Dropout** is a regularization technique for neural networks that drops a unit (along with connections) at training time with a specified probability $p$ (... | 77,804 |
8,616 | https://paperswithcode.com/paper/learning-approximate-inference-networks-for | 1803.03376 | Learning Approximate Inference Networks for Structured Prediction | Structured prediction energy networks (SPENs; Belanger & McCallum 2016) use
neural network architectures to define energy functions that can capture
arbitrary dependencies among parts of structured outputs. Prior work used
gradient descent for inference, relaxing the structured output to a set of
continuous variables a... | http://arxiv.org/abs/1803.03376v1 | http://arxiv.org/pdf/1803.03376v1.pdf | ICLR 2018 1 | [
"Lifu Tu",
"Kevin Gimpel"
] | [
"Language Modelling",
"Multi-Label Classification",
"Part-Of-Speech Tagging",
"Structured Prediction"
] | 1,520,553,600,000 | [] | 56,649 |
221,481 | https://paperswithcode.com/paper/stytr-2-unbiased-image-style-transfer-with | 2105.14576 | StyTr$^2$: Image Style Transfer with Transformers | The goal of image style transfer is to render an image with artistic features guided by a style reference while maintaining the original content. Owing to the locality in convolutional neural networks (CNNs), extracting and maintaining the global information of input images is difficult. Therefore, traditional neural s... | https://arxiv.org/abs/2105.14576v3 | https://arxiv.org/pdf/2105.14576v3.pdf | null | [
"Yingying Deng",
"Fan Tang",
"WeiMing Dong",
"Chongyang Ma",
"Xingjia Pan",
"Lei Wang",
"Changsheng Xu"
] | [
"Style Transfer"
] | 1,622,332,800,000 | [] | 130,489 |
206,830 | https://paperswithcode.com/paper/consistency-based-active-learning-for-object | 2103.10374 | Consistency-based Active Learning for Object Detection | Active learning aims to improve the performance of task model by selecting the most informative samples with a limited budget. Unlike most recent works that focused on applying active learning for image classification, we propose an effective Consistency-based Active Learning method for object Detection (CALD), which f... | https://arxiv.org/abs/2103.10374v3 | https://arxiv.org/pdf/2103.10374v3.pdf | null | [
"Weiping Yu",
"Sijie Zhu",
"Taojiannan Yang",
"Chen Chen"
] | [
"Active Learning",
"Classification",
"Classification",
"Image Classification",
"Object Detection",
"Object Detection"
] | 1,616,025,600,000 | [
{
"code_snippet_url": "https://github.com/pytorch/vision/blob/5e9ebe8dadc0ea2841a46cfcd82a93b4ce0d4519/torchvision/ops/roi_pool.py#L10",
"description": "**Region of Interest Pooling**, or **RoIPool**, is an operation for extracting a small feature map (e.g., $7×7$) from each RoI in detection and segmentatio... | 62,718 |
52,784 | https://paperswithcode.com/paper/news-session-based-recommendations-using-deep | 1808.00076 | News Session-Based Recommendations using Deep Neural Networks | News recommender systems are aimed to personalize users experiences and help
them to discover relevant articles from a large and dynamic search space.
Therefore, news domain is a challenging scenario for recommendations, due to
its sparse user profiling, fast growing number of items, accelerated item's
value decay, and... | http://arxiv.org/abs/1808.00076v3 | http://arxiv.org/pdf/1808.00076v3.pdf | null | [
"Gabriel de Souza P. Moreira",
"Felipe Ferreira",
"Adilson Marques da Cunha"
] | [
"News Recommendation",
"Recommendation Systems",
"Session-Based Recommendations"
] | 1,532,995,200,000 | [] | 166,734 |
254,403 | https://paperswithcode.com/paper/are-factuality-checkers-reliable-adversarial | null | Are Factuality Checkers Reliable? Adversarial Meta-evaluation of Factuality in Summarization | With the continuous upgrading of the summarization systems driven by deep neural networks, researchers have higher requirements on the quality of the generated summaries, which should be not only fluent and informative but also factually correct. As a result, the field of factual evaluation has developed rapidly recent... | https://aclanthology.org/2021.findings-emnlp.179 | https://aclanthology.org/2021.findings-emnlp.179.pdf | Findings (EMNLP) 2021 11 | [
"Yiran Chen",
"PengFei Liu",
"Xipeng Qiu"
] | [
"Data Augmentation"
] | 1,635,724,800,000 | [] | 110,904 |
169,201 | https://paperswithcode.com/paper/block-term-tensor-neural-networks | 2010.04963 | Block-term Tensor Neural Networks | Deep neural networks (DNNs) have achieved outstanding performance in a wide range of applications, e.g., image classification, natural language processing, etc. Despite the good performance, the huge number of parameters in DNNs brings challenges to efficient training of DNNs and also their deployment in low-end device... | https://arxiv.org/abs/2010.04963v2 | https://arxiv.org/pdf/2010.04963v2.pdf | null | [
"Jinmian Ye",
"Guangxi Li",
"Di Chen",
"Haiqin Yang",
"Shandian Zhe",
"Zenglin Xu"
] | [
"Image Classification"
] | 1,602,288,000,000 | [] | 150,066 |
244,768 | https://paperswithcode.com/paper/aggregation-with-feature-detection | null | Aggregation With Feature Detection | Aggregating features from different depths of a network is widely adopted to improve the network capability. Lots of modern architectures are equipped with skip connections, which actually makes the feature aggregation happen in all these networks. Since different features tell different semantic meanings, there a... | http://openaccess.thecvf.com//content/ICCV2021/html/Sun_Aggregation_With_Feature_Detection_ICCV_2021_paper.html | http://openaccess.thecvf.com//content/ICCV2021/papers/Sun_Aggregation_With_Feature_Detection_ICCV_2021_paper.pdf | ICCV 2021 10 | [
"Shuyang Sun",
"Xiaoyu Yue",
"Xiaojuan Qi",
"Wanli Ouyang",
"Victor Adrian Prisacariu",
"Philip H.S. Torr"
] | [
"Instance Segmentation",
"Object Detection",
"Object Detection",
"Semantic Segmentation"
] | 1,609,459,200,000 | [
{
"code_snippet_url": "",
"description": "**Average Pooling** is a pooling operation that calculates the average value for patches of a feature map, and uses it to create a downsampled (pooled) feature map. It is usually used after a convolutional layer. It adds a small amount of translation invariance - me... | 39,332 |
186,426 | https://paperswithcode.com/paper/towards-adversarial-learning-of-speaker | 1903.09606 | Towards adversarial learning of speaker-invariant representation for speech emotion recognition | Speech emotion recognition (SER) has attracted great attention in recent
years due to the high demand for emotionally intelligent speech interfaces.
Deriving speaker-invariant representations for speech emotion recognition is
crucial. In this paper, we propose to apply adversarial training to SER to
learn speaker-invar... | http://arxiv.org/abs/1903.09606v1 | http://arxiv.org/pdf/1903.09606v1.pdf | null | [] | [
"Classification",
"Emotion Classification",
"Emotion Recognition",
"Representation Learning",
"Speech Emotion Recognition"
] | 1,553,212,800,000 | [] | 91,257 |
110,612 | https://paperswithcode.com/paper/chinese-relation-extraction-with-multi | null | Chinese Relation Extraction with Multi-Grained Information and External Linguistic Knowledge | Chinese relation extraction is conducted using neural networks with either character-based or word-based inputs, and most existing methods typically suffer from segmentation errors and ambiguity of polysemy. To address the issues, we propose a multi-grained lattice framework (MG lattice) for Chinese relation extraction... | https://aclanthology.org/P19-1430 | https://aclanthology.org/P19-1430.pdf | ACL 2019 7 | [
"Ziran Li",
"Ning Ding",
"Zhiyuan Liu",
"Hai-Tao Zheng",
"Ying Shen"
] | [
"Relation Extraction"
] | 1,561,939,200,000 | [] | 122,862 |
98,124 | https://paperswithcode.com/paper/transformable-bottleneck-networks | 1904.06458 | Transformable Bottleneck Networks | We propose a novel approach to performing fine-grained 3D manipulation of image content via a convolutional neural network, which we call the Transformable Bottleneck Network (TBN). It applies given spatial transformations directly to a volumetric bottleneck within our encoder-bottleneck-decoder architecture. Multi-vie... | https://arxiv.org/abs/1904.06458v5 | https://arxiv.org/pdf/1904.06458v5.pdf | ICCV 2019 10 | [
"Kyle Olszewski",
"Sergey Tulyakov",
"Oliver Woodford",
"Hao Li",
"Linjie Luo"
] | [
"3D Reconstruction",
"Novel View Synthesis"
] | 1,555,113,600,000 | [] | 120,802 |
107,961 | https://paperswithcode.com/paper/volmap-a-real-time-model-for-semantic | 1906.11873 | VolMap: A Real-time Model for Semantic Segmentation of a LiDAR surrounding view | This paper introduces VolMap, a real-time approach for the semantic segmentation of a 3D LiDAR surrounding view system in autonomous vehicles. We designed an optimized deep convolution neural network that can accurately segment the point cloud produced by a 360\degree{} LiDAR setup, where the input consists of a volume... | https://arxiv.org/abs/1906.11873v1 | https://arxiv.org/pdf/1906.11873v1.pdf | null | [
"Hager Radi",
"Waleed Ali"
] | [
"3D Object Detection",
"Autonomous Vehicles",
"Object Detection",
"Object Detection",
"Semantic Segmentation"
] | 1,560,297,600,000 | [
{
"code_snippet_url": null,
"description": "A **convolution** is a type of matrix operation, consisting of a kernel, a small matrix of weights, that slides over input data performing element-wise multiplication with the part of the input it is on, then summing the results into an output.\r\n\r\nIntuitively,... | 71,378 |
123,059 | https://paperswithcode.com/paper/using-dynamic-embeddings-to-improve-static | 1911.02929 | How Can BERT Help Lexical Semantics Tasks? | Contextualized embeddings such as BERT can serve as strong input representations to NLP tasks, outperforming their static embeddings counterparts such as skip-gram, CBOW and GloVe. However, such embeddings are dynamic, calculated according to a sentence-level context, which limits their use in lexical semantics tasks. ... | https://arxiv.org/abs/1911.02929v2 | https://arxiv.org/pdf/1911.02929v2.pdf | null | [
"Yile Wang",
"Leyang Cui",
"Yue Zhang"
] | [
"Word Embeddings"
] | 1,573,084,800,000 | [
{
"code_snippet_url": "",
"description": "**GloVe Embeddings** are a type of word embedding that encode the co-occurrence probability ratio between two words as vector differences. GloVe uses a weighted least squares objective $J$ that minimizes the difference between the dot product of the vectors of two w... | 135,696 |
307,643 | https://paperswithcode.com/paper/funqg-molecular-representation-learning-via | 2207.08597 | FunQG: Molecular Representation Learning Via Quotient Graphs | Learning expressive molecular representations is crucial to facilitate the accurate prediction of molecular properties. Despite the significant advancement of graph neural networks (GNNs) in molecular representation learning, they generally face limitations such as neighbors-explosion, under-reaching, over-smoothing, a... | https://arxiv.org/abs/2207.08597v1 | https://arxiv.org/pdf/2207.08597v1.pdf | null | [
"Hossein Hajiabolhassan",
"Zahra Taheri",
"Ali Hojatnia",
"Yavar Taheri Yeganeh"
] | [
"Molecular Property Prediction",
"Representation Learning"
] | 1,658,102,400,000 | [] | 54,202 |
182,790 | https://paperswithcode.com/paper/mosaicked-multispectral-image-compression | 1801.03577 | Mosaicked multispectral image compression based on inter- and intra-band correlation | Multispectral imaging has been utilized in many fields, but the cost of
capturing and storing image data is still high. Single-sensor cameras with
multispectral filter arrays can reduce the cost of capturing images at the
expense of slightly lower image quality. When multispectral filter arrays are
used, conventional m... | http://arxiv.org/abs/1801.03577v1 | http://arxiv.org/pdf/1801.03577v1.pdf | null | [] | [
"Image Compression"
] | 1,515,542,400,000 | [] | 149,774 |
98,226 | https://paperswithcode.com/paper/swtvm-exploring-the-automated-compilation-for | 1904.07404 | swTVM: Towards Optimized Tensor Code Generation for Deep Learning on Sunway Many-Core Processor | The flourish of deep learning frameworks and hardware platforms has been demanding an efficient compiler that can shield the diversity in both software and hardware in order to provide application portability. Among the existing deep learning compilers, TVM is well known for its efficiency in code generation and optimi... | https://arxiv.org/abs/1904.07404v3 | https://arxiv.org/pdf/1904.07404v3.pdf | null | [
"Mingzhen Li",
"Changxi Liu",
"Jianjin Liao",
"Xuegui Zheng",
"Hailong Yang",
"Rujun Sun",
"Jun Xu",
"Lin Gan",
"Guangwen Yang",
"Zhongzhi Luan",
"Depei Qian"
] | [
"Code Generation"
] | 1,555,372,800,000 | [
{
"code_snippet_url": "https://www.healthnutra.org/es/maxup/",
"description": "A **1 x 1 Convolution** is a [convolution](https://paperswithcode.com/method/convolution) with some special properties in that it can be used for dimensionality reduction, efficient low dimensional embeddings, and applying non-li... | 62,359 |
197,959 | https://paperswithcode.com/paper/instantaneous-psd-estimation-for-speech | 2007.00542 | Instantaneous PSD Estimation for Speech Enhancement based on Generalized Principal Components | Power spectral density (PSD) estimates of various microphone signal components are essential to many speech enhancement procedures. As speech is highly non-nonstationary, performance improvements may be gained by maintaining time-variations in PSD estimates. In this paper, we propose an instantaneous PSD estimation app... | https://arxiv.org/abs/2007.00542v1 | https://arxiv.org/pdf/2007.00542v1.pdf | null | [] | [
"Speech Enhancement"
] | 1,593,561,600,000 | [] | 166,124 |
300,148 | https://paperswithcode.com/paper/transformer-based-urdu-handwritten-text | 2206.04575 | Transformer based Urdu Handwritten Text Optical Character Reader | Extracting Handwritten text is one of the most important components of digitizing information and making it available for large scale setting. Handwriting Optical Character Reader (OCR) is a research problem in computer vision and natural language processing computing, and a lot of work has been done for English, but u... | https://arxiv.org/abs/2206.04575v1 | https://arxiv.org/pdf/2206.04575v1.pdf | null | [
"Mohammad Daniyal Shaiq",
"Musa Dildar Ahmed Cheema",
"Ali Kamal"
] | [
"Natural Language Understanding",
"Optical Character Recognition"
] | 1,654,732,800,000 | [] | 884 |
End of preview. Expand in Data Studio
- Downloads last month
- 15