GitHub. 2023. Github copilot.
https://github.
com/features/copilot
. Accessed: 2024-02-
14.
Gemini Google, Rohan Anil, Sebastian Borgeaud,
Yonghui Wu, Jean-Baptiste Alayrac, Jiahui Yu,
Radu Soricut, Johan Schalkwyk, Andrew M Dai,
Anja Hauth, et al. 2023. Gemini: a family of
highly capable multimodal models. ArXiv preprint,
abs/2312.11805.
Daya Guo, Qihao Zhu, Dejian Yang, Zhenda Xie, Kai
Dong, Wentao Zhang, Guanting Chen, Xiao Bi,
Y Wu, YK Li, et al. 2024. Deepseek-coder: When the
large language model meets programming–the rise of
code intelligence. ArXiv preprint, abs/2401.14196.
Albert Q Jiang, Alexandre Sablayrolles, Arthur Men-
sch, Chris Bamford, Devendra Singh Chaplot, Diego
de las Casas, Florian Bressand, Gianna Lengyel, Guil-
laume Lample, Lucile Saulnier, et al. 2023. Mistral
7b. ArXiv preprint, abs/2310.06825.
Hung Le, Yue Wang, Akhilesh Deepak Gotmare, Silvio
Savarese, and Steven Chu Hong Hoi. 2022. Coderl:
Mastering code generation through pretrained models
and deep reinforcement learning. Advances in Neural
Information Processing Systems, 35:21314–21328.
Raymond Li, Loubna Ben Allal, Yangtian Zi, Niklas
Muennighoff, Denis Kocetkov, Chenghao Mou, Marc
Marone, Christopher Akiki, Jia Li, Jenny Chim, et al.
2023a. Starcoder: may the source be with you!
ArXiv preprint, abs/2305.06161.
Rongao Li, Jie Fu, Bo-Wen Zhang, Tao Huang, Zhihong
Sun, Chen Lyu, Guang Liu, Zhi Jin, and Ge Li. 2023b.
Taco: Topics in algorithmic code generation dataset.
ArXiv preprint, abs/2312.14852.
Yujia Li, David Choi, Junyoung Chung, Nate Kushman,
Julian Schrittwieser, Rémi Leblond, Tom Eccles,
James Keeling, Felix Gimeno, Agustin Dal Lago,
et al. 2022. Competition-level code generation with
alphacode. Science, 378(6624):1092–1097.
Jiawei Liu, Chunqiu Steven Xia, Yuyao Wang, and
LINGMING ZHANG. 2023. Is your code gener-
ated by chatGPT really correct? rigorous evalua-
tion of large language models for code generation.
In Thirty-seventh Conference on Neural Information
Processing Systems.
Ziyang Luo, Can Xu, Pu Zhao, Qingfeng Sun, Xi-
ubo Geng, Wenxiang Hu, Chongyang Tao, Jing Ma,
Qingwei Lin, and Daxin Jiang. 2023. Wizardcoder:
Empowering code large language models with evol-
instruct. ArXiv preprint, abs/2306.08568.
Aman Madaan, Niket Tandon, Peter Clark, and Yim-
ing Yang. 2022. Memory-assisted prompt editing
to improve GPT-3 after deployment. In Proceed-
ings of the 2022 Conference on Empirical Methods
in Natural Language Processing, pages 2833–2861,
Abu Dhabi, United Arab Emirates. Association for
Computational Linguistics.
Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler
Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon,
Nouha Dziri, Shrimai Prabhumoye, Yiming Yang,
et al. 2023. Self-refine: Iterative refinement with
self-feedback. ArXiv preprint, abs/2303.17651.
Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan
Wang, Yingbo Zhou, Silvio Savarese, and Caiming
Xiong. 2022. Codegen: An open large language
model for code with multi-turn program synthesis.
ArXiv preprint, abs/2203.13474.
OpenAI. 2022. ChatGPT: Optimizing Language
Models for Dialogue.
https://openai.com/
blog/chatgpt/. Accessed on 14 Feb. 2024.
OpenAI. 2023. Gpt-4 technical report.
Phind. 2023. Phind/phind-codellama-34b-v2.
Jack W Rae, Sebastian Borgeaud, Trevor Cai, Katie
Millican, Jordan Hoffmann, Francis Song, John
Aslanides, Sarah Henderson, Roman Ring, Susan-
nah Young, et al. 2021. Scaling language models:
Methods, analysis & insights from training gopher.
ArXiv preprint, abs/2112.11446.
Baptiste Roziere, Jonas Gehring, Fabian Gloeckle, Sten
Sootla, Itai Gat, Xiaoqing Ellen Tan, Yossi Adi,
Jingyu Liu, Tal Remez, Jérémy Rapin, et al. 2023.
Code llama: Open foundation models for code.
ArXiv preprint, abs/2308.12950.
William Saunders, Catherine Yeh, Jeff Wu, Steven Bills,
Long Ouyang, Jonathan Ward, and Jan Leike. 2022.
Self-critiquing models for assisting human evaluators.
ArXiv preprint, abs/2206.05802.
Jérémy Scheurer, Jon Ander Campos, Jun Shern Chan,
Angelica Chen, Kyunghyun Cho, and Ethan Perez.
2022. Training language models with natural lan-
guage feedback. ArXiv preprint, abs/2204.14146.
Weijia Shi, Sewon Min, Maria Lomeli, Chunting Zhou,
Margaret Li, Victoria Lin, Noah A Smith, Luke
Zettlemoyer, Scott Yih, and Mike Lewis. 2023. In-
context pretraining: Language modeling beyond doc-
ument boundaries. ArXiv preprint, abs/2310.10638.
Parshin Shojaee, Aneesh Jain, Sindhu Tipirneni, and
Chandan K Reddy. 2023. Execution-based code gen-
eration using deep reinforcement learning. ArXiv
preprint, abs/2301.13816.
speechless. 2023. speechless-codellama-34b-v2.0.
Niket Tandon, Aman Madaan, Peter Clark, Keisuke
Sakaguchi, and Yiming Yang. 2021. Interscript: A
dataset for interactive learning of scripts through er-
ror feedback. ArXiv preprint, abs/2112.07867.
Ross Taylor, Marcin Kardas, Guillem Cucurull, Thomas
Scialom, Anthony Hartshorn, Elvis Saravia, Andrew
Poulton, Viktor Kerkez, and Robert Stojnic. 2022.
Galactica: A large language model for science. ArXiv
preprint, abs/2211.09085.