Highlights
- Pro
Lists (10)
Sort Name ascending (A-Z)
Benchmark
Code_and_Text
code2code,code2text,text2code,text2text(Code) LLMs
New (Code) LLMs Evaluation of LLMs New methods for better use of LLMsCollections or Surveys
Collections of certain topics Surveys for certain topicsDefault
I do not know how to classify them(Large) Code Models
Code Models among LLMsLearning List
A list for me to learn something newResearch Track
A collection of replication packages from conference papersStars
This is the repository for the paper titled "ThinkRepair: Self-Directed Automated Program Repair" accepted by ISSTA'24.
Reverse Engineering: Decompiling Binary Code with Large Language Models
The leaderboard website of REval benchmark (ICSE 2025 Paper "Reasoning Runtime Behavior of a Program with LLM: How Far Are We?")
Enhancing AI Software Engineering with Repository-level Code Graph
🚀🚀 「大模型」2小时完全从0训练26M的小参数GPT!🌏 Train a 26M-parameter GPT from scratch in just 2h!
A dataset of LLM-generated chain-of-thought steps annotated with mistake location.
Interaction2Code: How Far Are We From Automatic Interactive Webpage Generation?
《开源大模型食用指南》针对中国宝宝量身打造的基于Linux环境快速微调(全参数/Lora)、部署国内外开源大模型(LLM)/多模态大模型(MLLM)教程
中文nlp解决方案(大模型、数据、模型、训练、推理)
A high-throughput and memory-efficient inference and serving engine for LLMs
Open-source repository for the OOPSLA'24 paper "CYCLE: Learning to Self-Refine Code Generation"
A Systematic Literature Review on Large Language Models for Automated Program Repair
CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)
TeCo: an ML+Execution model for test completion
The Open Cookbook for Top-Tier Code Large Language Model
Fast lexical search implementing BM25 in Python using Numpy, Numba and Scipy
A modular graph-based Retrieval-Augmented Generation (RAG) system
A Pair Programming Framework for Code Generation via Multi-Plan Exploration and Feedback-Driven Refinement, ASE 2024 (Distinguished Paper Award)
Code and data for XLCoST: A Benchmark Dataset for Cross-lingual Code Intelligence
A distributed, extensible, secure solution for evaluating machine generated code with unit tests in multiple programming languages.
[LREC-COLING'24] HumanEval-XL: A Multilingual Code Generation Benchmark for Cross-lingual Natural Language Generalization