1 |
Universal Conditional Masked Language Pre-training for Neural Machine Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
2 |
Compilable Neural Code Generation with Compiler Feedback ...
|
|
Wang, Xin; Wang, Yasheng; Wan, Yao; Mi, Fei; Li, Yitong; Zhou, Pingyi; Liu, Jin; Wu, Hao; Jiang, Xin; Liu, Qun. - : arXiv, 2022
|
|
Abstract:
Automatically generating compilable programs with (or without) natural language descriptions has always been a touchstone problem for computational linguistics and automated software engineering. Existing deep-learning approaches model code generation as text generation, either constrained by grammar structures in decoder, or driven by pre-trained language models on large-scale code corpus (e.g., CodeGPT, PLBART, and CodeT5). However, few of them account for compilability of the generated programs. To improve compilability of the generated programs, this paper proposes COMPCODER, a three-stage pipeline utilizing compiler feedback for compilable code generation, including language model fine-tuning, compilability reinforcement, and compilability discrimination. Comprehensive experiments on two code generation tasks demonstrate the effectiveness of our proposed approach, improving the success rate of compilation from 44.18 to 89.18 in code completion on average and from 70.3 to 96.2 in text-to-code generation, ... : Accepted by ACL 2022 ...
|
|
Keyword:
Artificial Intelligence cs.AI; Computation and Language cs.CL; FOS Computer and information sciences; Programming Languages cs.PL
|
|
URL: https://arxiv.org/abs/2203.05132 https://dx.doi.org/10.48550/arxiv.2203.05132
|
|
BASE
|
|
Hide details
|
|
3 |
Sub-Character Tokenization for Chinese Pretrained Language Models ...
|
|
|
|
BASE
|
|
Show details
|
|
4 |
Training Multilingual Pre-trained Language Model with Byte-level Subwords ...
|
|
|
|
BASE
|
|
Show details
|
|
5 |
Multilingual Speech Translation with Unified Transformer: Huawei Noah's Ark Lab at IWSLT 2021 ...
|
|
|
|
BASE
|
|
Show details
|
|
7 |
Learning Multilingual Representation for Natural Language Understanding with Enhanced Cross-Lingual Supervision ...
|
|
|
|
BASE
|
|
Show details
|
|
8 |
LightMBERT: A Simple Yet Effective Method for Multilingual BERT Distillation ...
|
|
|
|
BASE
|
|
Show details
|
|
9 |
Uncertainty-Aware Balancing for Multilingual and Multi-Domain Neural Machine Translation Training ...
|
|
|
|
BASE
|
|
Show details
|
|
10 |
Improving Unsupervised Question Answering via Summarization-Informed Question Generation ...
|
|
|
|
BASE
|
|
Show details
|
|
11 |
CCA-MDD: A Coupled Cross-Attention based Framework for Streaming Mispronunciation detection and diagnosis ...
|
|
|
|
BASE
|
|
Show details
|
|
12 |
A Mutual Information Maximization Approach for the Spurious Solution Problem in Weakly Supervised Question Answering ...
|
|
|
|
BASE
|
|
Show details
|
|
13 |
HyKnow: End-to-End Task-Oriented Dialog Modeling with Hybrid Knowledge Management ...
|
|
|
|
BASE
|
|
Show details
|
|
14 |
AutoTinyBERT: Automatic Hyper-parameter Optimization for Efficient Pre-trained Language Models ...
|
|
|
|
BASE
|
|
Show details
|
|
15 |
TGEA: An Error-Annotated Dataset and Benchmark Tasks for TextGeneration from Pretrained Language Models ...
|
|
|
|
BASE
|
|
Show details
|
|
16 |
Two Parents, One Child: {D}ual Transfer for Low-Resource Neural Machine Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
17 |
RealTranS: End-to-End Simultaneous Speech Translation with Convolutional Weighted-Shrinking Transformer ...
|
|
|
|
BASE
|
|
Show details
|
|
18 |
Uncertainty-Aware Balancing for Multilingual and Multi-Domain Neural Machine Translation Training ...
|
|
|
|
BASE
|
|
Show details
|
|
19 |
DyLex: Incorporating Dynamic Lexicons into BERT for Sequence Labeling ...
|
|
|
|
BASE
|
|
Show details
|
|
|
|