The following papers were recommended by the Semantic Scholar API
\n- \n
- Magicoder: Source Code Is All You Need (2023) \n
- DeepSeek LLM: Scaling Open-Source Language Models with Longtermism (2024) \n
- WaveCoder: Widespread And Versatile Enhanced Instruction Tuning with Refined Data Generation (2023) \n
- On the Effectiveness of Large Language Models in Domain-Specific Code Generation (2023) \n
- AST-T5: Structure-Aware Pretraining for Code Generation and Understanding (2024) \n
Please give a thumbs up to this comment if you found it helpful!
\nIf you want recommendations for any Paper on Hugging Face checkout this Space
\n","updatedAt":"2024-01-27T01:21:40.786Z","author":{"_id":"63d3e0e8ff1384ce6c5dd17d","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1674830754237-63d3e0e8ff1384ce6c5dd17d.jpeg","fullname":"Librarian Bot (Bot)","name":"librarian-bot","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":264}},"numEdits":0,"identifiedLanguage":{"language":"en","probability":0.7159170508384705},"editors":["librarian-bot"],"editorAvatarUrls":["https://cdn-avatars.huggingface.co/v1/production/uploads/1674830754237-63d3e0e8ff1384ce6c5dd17d.jpeg"],"reactions":[],"isReport":false}},{"id":"65b6ba2e791c522128f8ebe6","author":{"_id":"60bdd16a54fdf5333724c90c","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1623052635576-noauth.jpeg","fullname":"Sangmin Ahn","name":"Sangmin","type":"user","isPro":true,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":8},"createdAt":"2024-01-28T20:33:50.000Z","type":"comment","data":{"edited":true,"hidden":true,"hiddenBy":"","latest":{"raw":"This comment has been hidden","html":"This comment has been hidden","updatedAt":"2024-01-28T20:35:50.954Z","author":{"_id":"60bdd16a54fdf5333724c90c","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1623052635576-noauth.jpeg","fullname":"Sangmin Ahn","name":"Sangmin","type":"user","isPro":true,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":8}},"numEdits":3,"editors":[],"editorAvatarUrls":[],"reactions":[]}},{"id":"6799b94148e36d6187b6bf78","author":{"_id":"6799afa8da59b0f6aedaf101","avatarUrl":"/avatars/5cd95bbb03f762e21d35b8b392714d4c.svg","fullname":"Abcd","name":"Afan-75748","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false},"createdAt":"2025-01-29T05:14:41.000Z","type":"comment","data":{"edited":true,"hidden":true,"hiddenBy":"","latest":{"raw":"This comment has been hidden","html":"This comment has been hidden","updatedAt":"2025-01-29T05:14:52.949Z","author":{"_id":"6799afa8da59b0f6aedaf101","avatarUrl":"/avatars/5cd95bbb03f762e21d35b8b392714d4c.svg","fullname":"Abcd","name":"Afan-75748","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false}},"numEdits":0,"editors":[],"editorAvatarUrls":[],"reactions":[]}},{"id":"67d0620166d901d1fa2d9fe6","author":{"_id":"67862789f5b24159e9ab1510","avatarUrl":"/avatars/9e909c0abe15760cd1976f0dc9b8f148.svg","fullname":"Neon","name":"REDFECT23","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false},"createdAt":"2025-03-11T16:17:05.000Z","type":"comment","data":{"edited":true,"hidden":true,"hiddenBy":"","latest":{"raw":"This comment has been hidden","html":"This comment has been hidden","updatedAt":"2025-03-11T16:17:15.355Z","author":{"_id":"67862789f5b24159e9ab1510","avatarUrl":"/avatars/9e909c0abe15760cd1976f0dc9b8f148.svg","fullname":"Neon","name":"REDFECT23","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false}},"numEdits":0,"editors":[],"editorAvatarUrls":[],"reactions":[]}}],"primaryEmailConfirmed":false,"paper":{"id":"2401.14196","authors":[{"_id":"65b32177b0a5a381b60c4e11","user":{"_id":"653df20eaa1f487614da4db1","avatarUrl":"/avatars/12b27ce2c59f53b7e464039deab36a5d.svg","isPro":false,"fullname":"Daya Guo","user":"guoday","type":"user"},"name":"Daya Guo","status":"admin_assigned","statusLastChangedAt":"2024-01-26T09:38:34.569Z","hidden":false},{"_id":"65b32177b0a5a381b60c4e12","user":{"_id":"63cd76b4374057a338e8e703","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/63cd76b4374057a338e8e703/i4Qk5-0aYx3oRhC8b50aJ.jpeg","isPro":false,"fullname":"zhuqihao","user":"zqh11","type":"user"},"name":"Qihao Zhu","status":"admin_assigned","statusLastChangedAt":"2024-01-26T09:38:47.601Z","hidden":false},{"_id":"65b32177b0a5a381b60c4e13","user":{"_id":"6225bb44c6e650de3a65dbaa","avatarUrl":"/avatars/99c99ced2461978df572c27c1b3a4904.svg","isPro":false,"fullname":"DejianYang","user":"DejianYang","type":"user"},"name":"Dejian Yang","status":"admin_assigned","statusLastChangedAt":"2024-01-26T09:39:04.053Z","hidden":false},{"_id":"65b32177b0a5a381b60c4e14","user":{"_id":"6539f6ea26df26ecd1393c37","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/6539f6ea26df26ecd1393c37/9VJusLLAiLhUxfBAFggpF.jpeg","isPro":false,"fullname":"Zhenda Xie","user":"zdaxie","type":"user"},"name":"Zhenda Xie","status":"admin_assigned","statusLastChangedAt":"2024-01-26T09:39:10.898Z","hidden":false},{"_id":"65b32177b0a5a381b60c4e15","name":"Kai Dong","hidden":false},{"_id":"65b32177b0a5a381b60c4e16","name":"Wentao Zhang","hidden":false},{"_id":"65b32177b0a5a381b60c4e17","name":"Guanting Chen","hidden":false},{"_id":"65b32177b0a5a381b60c4e18","name":"Xiao Bi","hidden":false},{"_id":"65b32177b0a5a381b60c4e19","user":{"_id":"64d8aa52bcab729cb4e7973f","avatarUrl":"/avatars/d9f8118839e8ad3ec86ecd40cf59772f.svg","isPro":false,"fullname":"Y. WU","user":"y-wu-pkutech","type":"user"},"name":"Y. Wu","status":"admin_assigned","statusLastChangedAt":"2024-01-26T09:40:38.084Z","hidden":false},{"_id":"65b32177b0a5a381b60c4e1a","name":"Y. K. Li","hidden":false},{"_id":"65b32177b0a5a381b60c4e1b","user":{"_id":"6538815d1bdb3c40db94fbfa","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/6538815d1bdb3c40db94fbfa/id7aSY8JUgKK2agKWLERt.jpeg","isPro":false,"fullname":"Fuli Luo","user":"luofuli","type":"user"},"name":"Fuli Luo","status":"admin_assigned","statusLastChangedAt":"2024-01-26T09:40:53.712Z","hidden":false},{"_id":"65b32177b0a5a381b60c4e1c","name":"Yingfei Xiong","hidden":false},{"_id":"65b32177b0a5a381b60c4e1d","name":"Wenfeng Liang","hidden":false}],"publishedAt":"2024-01-25T14:17:53.000Z","submittedOnDailyAt":"2024-01-26T00:35:28.508Z","title":"DeepSeek-Coder: When the Large Language Model Meets Programming -- The\n Rise of Code Intelligence","submittedOnDailyBy":{"_id":"60f1abe7544c2adfd699860c","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1674929746905-60f1abe7544c2adfd699860c.jpeg","isPro":false,"fullname":"AK","user":"akhaliq","type":"user"},"summary":"The rapid development of large language models has revolutionized code\nintelligence in software development. However, the predominance of\nclosed-source models has restricted extensive research and development. To\naddress this, we introduce the DeepSeek-Coder series, a range of open-source\ncode models with sizes from 1.3B to 33B, trained from scratch on 2 trillion\ntokens. These models are pre-trained on a high-quality project-level code\ncorpus and employ a fill-in-the-blank task with a 16K window to enhance code\ngeneration and infilling. Our extensive evaluations demonstrate that\nDeepSeek-Coder not only achieves state-of-the-art performance among open-source\ncode models across multiple benchmarks but also surpasses existing\nclosed-source models like Codex and GPT-3.5. Furthermore, DeepSeek-Coder models\nare under a permissive license that allows for both research and unrestricted\ncommercial use.","upvotes":66,"discussionId":"65b32178b0a5a381b60c4e68","ai_summary":"DeepSeek-Coder, a series of open-source code models trained from scratch, outperforms both open-source and closed-source models across multiple benchmarks.","ai_keywords":["DeepSeek-Coder","large language models","code generation","infilling","pre-trained","fill-in-the-blank task","project-level code corpus"]},"canReadDatabase":false,"canManagePapers":false,"canSubmit":false,"hasHfLevelAccess":false,"upvoted":false,"upvoters":[{"_id":"6083902e1e36b13a64497d91","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/6083902e1e36b13a64497d91/h4rGHMn2c6z5GesF0F6VU.png","isPro":false,"fullname":"cheng","user":"zhoujun","type":"user"},{"_id":"62bdcb9ce06c2f0ff85a6e08","avatarUrl":"/avatars/7b442b8d3022fcb04ed3ea7c1a8e0fb4.svg","isPro":false,"fullname":"Amrit","user":"abaveja313","type":"user"},{"_id":"63a3b90e7abdaa25a81b723a","avatarUrl":"/avatars/62b6a9c9e82285c1f03d3767b02609d3.svg","isPro":false,"fullname":"Iordanis","user":"iordanissh","type":"user"},{"_id":"655e1c11accde1bbc8c4034b","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/655e1c11accde1bbc8c4034b/UZYLm8-LZupx8nJbFIth9.jpeg","isPro":false,"fullname":"Mahesh Sathiamoorthy","user":"madiator","type":"user"},{"_id":"620783f24e28382272337ba4","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/620783f24e28382272337ba4/zkUveQPNiDfYjgGhuFErj.jpeg","isPro":false,"fullname":"GuoLiangTang","user":"Tommy930","type":"user"},{"_id":"64747f7e33192631bacd8831","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/64747f7e33192631bacd8831/dstkZJ4sHJSeqLesV5cOC.jpeg","isPro":false,"fullname":"Taufiq Dwi Purnomo","user":"taufiqdp","type":"user"},{"_id":"63feebe688b9695964c3cc6b","avatarUrl":"/avatars/7ba8b420c64d753cc81352cf5ab51b92.svg","isPro":false,"fullname":"Anthony W Figueroa","user":"THEFIG","type":"user"},{"_id":"6539f6ea26df26ecd1393c37","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/6539f6ea26df26ecd1393c37/9VJusLLAiLhUxfBAFggpF.jpeg","isPro":false,"fullname":"Zhenda Xie","user":"zdaxie","type":"user"},{"_id":"62f207564cc9fe83448ad97b","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1660028717717-noauth.png","isPro":false,"fullname":"Shen Zhuoran","user":"CMSFlash","type":"user"},{"_id":"644f30e733542c9c4b77eff8","avatarUrl":"/avatars/66b9eb77894785c40431f0377d78092a.svg","isPro":false,"fullname":"Bayo Falade","user":"Mobayode","type":"user"},{"_id":"6538119803519fddb4a17e10","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/6538119803519fddb4a17e10/ffJMkdx-rM7VvLTCM6ri_.jpeg","isPro":false,"fullname":"samusenps","user":"samusenps","type":"user"},{"_id":"648eb1eb59c4e5c87dc116e0","avatarUrl":"/avatars/c636cea39c2c0937f01398c94ead5dad.svg","isPro":false,"fullname":"fdsqefsgergd","user":"T-representer","type":"user"}],"acceptLanguages":["*"],"dailyPaperRank":2}">DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence
Abstract
DeepSeek-Coder, a series of open-source code models trained from scratch, outperforms both open-source and closed-source models across multiple benchmarks.
The rapid development of large language models has revolutionized code intelligence in software development. However, the predominance of closed-source models has restricted extensive research and development. To address this, we introduce the DeepSeek-Coder series, a range of open-source code models with sizes from 1.3B to 33B, trained from scratch on 2 trillion tokens. These models are pre-trained on a high-quality project-level code corpus and employ a fill-in-the-blank task with a 16K window to enhance code generation and infilling. Our extensive evaluations demonstrate that DeepSeek-Coder not only achieves state-of-the-art performance among open-source code models across multiple benchmarks but also surpasses existing closed-source models like Codex and GPT-3.5. Furthermore, DeepSeek-Coder models are under a permissive license that allows for both research and unrestricted commercial use.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Magicoder: Source Code Is All You Need (2023)
- DeepSeek LLM: Scaling Open-Source Language Models with Longtermism (2024)
- WaveCoder: Widespread And Versatile Enhanced Instruction Tuning with Refined Data Generation (2023)
- On the Effectiveness of Large Language Models in Domain-Specific Code Generation (2023)
- AST-T5: Structure-Aware Pretraining for Code Generation and Understanding (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space