You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[["ling2016latent", "Latent Predictor Networks for Code Generation"], ["karpathy2015visualizing", "Visualizing and Understanding Recurrent Networks"], ["yin2017syntactic", "A Syntactic Neural Model for General-Purpose Code Generation"], ["xu2020incorporating", "Incorporating External Knowledge through Pre-training for Natural Language to Code Generation"]]
1
+
[["yin2022natural", "Natural Language to Code Generation in Interactive Data Science Notebooks"], ["ling2016latent", "Latent Predictor Networks for Code Generation"], ["karpathy2015visualizing", "Visualizing and Understanding Recurrent Networks"], ["yin2017syntactic", "A Syntactic Neural Model for General-Purpose Code Generation"]]
[["xu2022systematic", "A Systematic Evaluation of Large Language Models of Code"], ["svyatkovskiy2020fast", "Fast and Memory-Efficient Neural Code Completion"], ["zhang2023repocoder", "RepoCoder: Repository-Level Code Completion Through Iterative Retrieval and Generation"], ["wei2023typet5", "TypeT5: Seq2seq Type Inference using Static Analysis"]]
1
+
[["xu2022systematic", "A Systematic Evaluation of Large Language Models of Code"], ["shrivastava2023repofusion", "RepoFusion: Training Code Models to Understand Your Repository"], ["ding2023static", "A Static Evaluation of Code Completion by Large Language Models"], ["svyatkovskiy2020fast", "Fast and Memory-Efficient Neural Code Completion"]]
[["bhatia2016automated", "Automated Correction for Syntax Errors in Programming Assignments using Recurrent Neural Networks"], ["hellendoorn2018deep", "Deep Learning Type Inference"], ["wu2021prototransformer", "ProtoTransformer: A Meta-Learning Approach to Providing Student Feedback"], ["santos2018syntax", "Syntax and Sensibility: Using language models to detect and correct syntax errors"]]
1
+
[["bhatia2016automated", "Automated Correction for Syntax Errors in Programming Assignments using Recurrent Neural Networks"], ["wu2021prototransformer", "ProtoTransformer: A Meta-Learning Approach to Providing Student Feedback"], ["hellendoorn2018deep", "Deep Learning Type Inference"], ["santos2018syntax", "Syntax and Sensibility: Using language models to detect and correct syntax errors"]]
[["feng2020codebert", "CodeBERT: A Pre-Trained Model for Programming and Natural Languages"], ["hellendoorn2015will", "Will they like this? Evaluating Code Contributions With Language Models"], ["raychev2016learning", "Learning Programs from Noisy Data"], ["saberi2023model", "Model-Agnostic Syntactical Information for Pre-Trained Programming Language Models"]]
1
+
[["feng2020codebert", "CodeBERT: A Pre-Trained Model for Programming and Natural Languages"], ["hellendoorn2015will", "Will they like this? Evaluating Code Contributions With Language Models"], ["saberi2023model", "Model-Agnostic Syntactical Information for Pre-Trained Programming Language Models"], ["raychev2016learning", "Learning Programs from Noisy Data"]]
[["garg2022deepperf", "DeepPERF: A Deep Learning-Based Approach For Improving Software Performance"], ["wang2023codet5", "CodeT5+: Open Code Large Language Models for Code Understanding and Generation"], ["gupta2023grace", "Grace: Language Models Meet Code Edits"], ["li2023hitchhiker", "The Hitchhiker's Guide to Program Analysis: A Journey with Large Language Models"]]
[["shrivastava2020repository", "Repository-Level Prompt Generation for Large Language Models of Code"], ["ye2020leveraging", "Leveraging Code Generation to Improve Code Retrieval and Summarization via Dual Learning"], ["lomshakov2023fine", "Fine-Tuning Large Language Models for Answering Programming Questions with Code Snippets"], ["hindle2012naturalness", "On the Naturalness of Software"]]
1
+
[["ye2020leveraging", "Leveraging Code Generation to Improve Code Retrieval and Summarization via Dual Learning"], ["shrivastava2020repository", "Repository-Level Prompt Generation for Large Language Models of Code"], ["hindle2012naturalness", "On the Naturalness of Software"], ["bui2021efficient", "Self-Supervised Contrastive Learning for Code Retrieval and Summarization via Semantic-Preserving Transformations"]]
[["fried2022incoder", "InCoder: A Generative Model for Code Infilling and Synthesis"], ["kocetkov2022stack", "The Stack: 3TB of permissively licensed source code"], ["mir2021manytypes4py", "ManyTypes4Py: A Benchmark Python Dataset for Machine Learning-based Type Inference"], ["bavarian2022efficient", "Efficient Training of Language Models to Fill in the Middle"]]
1
+
[["li2023starcoder", "StarCoder: may the source be with you!"], ["fried2022incoder", "InCoder: A Generative Model for Code Infilling and Synthesis"], ["kocetkov2022stack", "The Stack: 3TB of permissively licensed source code"], ["mir2021manytypes4py", "ManyTypes4Py: A Benchmark Python Dataset for Machine Learning-based Type Inference"]]
0 commit comments