|
1 | 1 | # aiXcoder-7B Code Large Language Model |
2 | 2 |
|
3 | 3 | <p align="center"> |
4 | | - 🏠 <a href="https://www.aixcoder.com/" target="_blank">Official website</a>|🛠 <a href="https://marketplace.visualstudio.com/items?itemName=aixcoder-plugin.aixcoder" target="_blank">VS Code Plugin</a>|🛠 <a href="https://plugins.jetbrains.com/plugin/13574-aixcoder-code-completer" target="_blank">Jetbrains Plugin</a>|🤗 <a href="https://huggingface.co/aiXcoder/aixcoder-7b-base" target="_blank">Model Weights</a>|<a href="./assets/wechat_1.jpg" target="_blank">WeChat</a>|<a href="./assets/wechat_2.jpg" target="_blank">WeChat Official Account</a> |
| 4 | + 🏠 <a href="https://www.aixcoder.com/" target="_blank">Official website</a>|📄 <a href="https://arxiv.org/pdf/2410.13187" target="_blank">Paper</a> | 🛠 <a href="https://marketplace.visualstudio.com/items?itemName=aixcoder-plugin.aixcoder" target="_blank">VS Code Plugin</a>|🛠 <a href="https://plugins.jetbrains.com/plugin/13574-aixcoder-code-completer" target="_blank">Jetbrains Plugin</a>|🤗 <a href="https://huggingface.co/aiXcoder/aixcoder-7b-base" target="_blank">Model Weights</a>|<a href="./assets/wechat_1.jpg" target="_blank">WeChat</a>|<a href="./assets/wechat_2.jpg" target="_blank">WeChat Official Account</a> |
5 | 5 | </p> |
6 | 6 |
|
7 | 7 | Welcome to the official repository of aiXcoder-7B Code Large Language Model. This model is designed to understand and generate code across multiple programming languages, offering state-of-the-art performance in code completion, comprehension, generation, and more tasks about programming languages. |
8 | 8 |
|
9 | 9 | Table of Contents |
10 | 10 |
|
11 | | -1. [Model Introduction](#model-introduction) |
12 | | -2. [Quickstart](#quickstart) |
| 11 | +- [aiXcoder-7B Code Large Language Model](#aixcoder-7b-code-large-language-model) |
| 12 | + - [Model Introduction](#model-introduction) |
| 13 | + - [Quickstart](#quickstart) |
13 | 14 | - [Environment Requirements](#environment-requirements) |
| 15 | + - [Option 1: Build Env](#option-1-build-env) |
| 16 | + - [Option 2: Docker](#option-2-docker) |
14 | 17 | - [Model Weights](#model-weights) |
15 | 18 | - [Inference Example](#inference-example) |
| 19 | + - [Command Line Execution](#command-line-execution) |
| 20 | + - [Python Script Execution](#python-script-execution) |
16 | 21 | - [Quantized through bitsandbytes](#quantized-through-bitsandbytes) |
17 | 22 | - [Fine-tuning example](#fine-tuning-example) |
18 | | -3. [Data for aiXcoder 7B](#data-for-aixcoder-7b) |
19 | | -4. [Training](#training) |
| 23 | + - [Data for aiXcoder 7B](#data-for-aixcoder-7b) |
| 24 | + - [Training](#training) |
20 | 25 | - [Training Hyperparameters](#training-hyperparameters) |
21 | 26 | - [Batch processing method](#batch-processing-method) |
22 | 27 | - [Pre-training Tasks](#pre-training-tasks) |
23 | | -5. [Details of Experimental Results](#details-of-experimental-results) |
| 28 | + - [Details of Experimental Results](#details-of-experimental-results) |
24 | 29 | - [NL2Code Benchmarks](#nl2code-benchmarks) |
25 | 30 | - [Code Completion (Fill in the Middle)](#code-completion-fill-in-the-middle) |
26 | 31 | - [Cross-file Code Evaluation](#cross-file-code-evaluation) |
27 | | -6. [License](#license) |
28 | | -7. [Acknowledgments](#acknowledgments) |
| 32 | + - [License](#license) |
| 33 | + - [Acknowledgments](#acknowledgments) |
| 34 | + - [Citation](#citation) |
29 | 35 |
|
30 | 36 |
|
31 | 37 |
|
@@ -493,3 +499,34 @@ The model weights are licensed under the [Model License](./MODEL_LICENSE) for ac |
493 | 499 | We would like to thank all contributors to the open-source projects and datasets that made this work possible. |
494 | 500 |
|
495 | 501 | Thank you for your interest in our Code Large Language Model. We look forward to your contributions and feedback! |
| 502 | + |
| 503 | +## Citation |
| 504 | + |
| 505 | +If you use the code or aiXcoder-7B in your work, please cite the following paper: |
| 506 | +``` |
| 507 | +@article{aiXcoder-7B, |
| 508 | + author = {Siyuan Jiang and |
| 509 | + Jia Li and |
| 510 | + He Zong and |
| 511 | + Huanyu Liu and |
| 512 | + Hao Zhu and |
| 513 | + Shukai Hu and |
| 514 | + Erlu Li and |
| 515 | + Jiazheng Ding and |
| 516 | + Yu Han and |
| 517 | + Wei Ning and |
| 518 | + Gen Wang and |
| 519 | + Yihong Dong and |
| 520 | + Kechi Zhang and |
| 521 | + Ge Li}, |
| 522 | + title = {aiXcoder-7B: {A} Lightweight and Effective Large Language Model for |
| 523 | + Code Processing}, |
| 524 | + journal = {CoRR}, |
| 525 | + volume = {abs/2410.13187}, |
| 526 | + year = {2024}, |
| 527 | + url = {https://doi.org/10.48550/arXiv.2410.13187}, |
| 528 | + doi = {10.48550/ARXIV.2410.13187}, |
| 529 | + eprinttype = {arXiv}, |
| 530 | + eprint = {2410.13187} |
| 531 | +} |
| 532 | +``` |
0 commit comments