Skip to content

Commit 0c58a98

Browse files
authored
Merge pull request #180 from FlagAI-Open/LongContext-ICL-Annotation
Add LongContext-ICL annotations
2 parents 3c5184d + 916c84d commit 0c58a98

File tree

3 files changed

+76
-119
lines changed

3 files changed

+76
-119
lines changed
Lines changed: 73 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,73 @@
1+
# 超长长上下文场景中LLM自动数据标注挑战赛
2+
3+
---
4+
5+
## 消息
6+
<!-- BEGIN NEWS -->
7+
- **[2026-01-20] `发布`** 赛事信息已在 **Kaggle** 正式上线。详情见:[FlagOS Open Computing Global Challenge](https://www.kaggle.com/competitions/flag-os-open-computing-global-challenge).
8+
- **[2026-01-06] `发布`****众智 FlagOS 社区****北京智源人工智能研究院(BAAI)****CCF ODTC** 联合主办的综合性大赛 **FlagOS 开放计算全球挑战赛** 正式发布。详情见:
9+
[FlagOS开放计算全球挑战赛- AI赛事通 | 数据算法赛](https://www.competehub.dev/zh/competitions/modelscope180)
10+
<!-- END NEWS -->
11+
12+
---
13+
14+
15+
## 快速开始
16+
### 1. 环境
17+
18+
```bash
19+
openai
20+
torch
21+
flagScale
22+
```
23+
24+
推荐在NVIDIA平台使用 `cd src && bash create_env_nvidia.sh` 创建环境。
25+
26+
### 2. 下载模型权重
27+
```bash
28+
hf download Qwen/Qwen3-4B --local-dir Qwen3-4B
29+
# or
30+
modelscope download --model Qwen/Qwen3-4B
31+
```
32+
### 3. 长文本配置
33+
`Qwen3-4B/config.json`将原有配置替换为:
34+
```json
35+
"rope_scaling": {
36+
"rope_type": "yarn",
37+
"factor": 4.0,
38+
"original_max_position_embeddings": 32768
39+
}
40+
```
41+
### 4. 模型部署
42+
43+
请根据实际需求,配置 `llm_config.yaml` 文件。启动配置
44+
45+
```bash
46+
cd FlagScale
47+
python run.py --config-path .. --config-name llm_config action=run
48+
```
49+
50+
在模型服务启动后,可通过以下方式测试本地 API:
51+
52+
```bash
53+
python api_test.py
54+
```
55+
56+
如需停止服务,请执行:
57+
58+
```bash
59+
python run.py --config-path .. --config-name llm_config action=stop
60+
```
61+
62+
### 5. 运行/改进基线方法(Baseline)
63+
64+
启动如下命令开始模型标注
65+
```bash
66+
python main.py
67+
```
68+
69+
实现新的标注方法,请修改`method.py`文件。你可以在该文件中:
70+
* 定义新的指令模板、
71+
* 定义新的上下文示例选择策略
72+
* 定义新的模型推理、标注方案
73+
* 添加自定义后处理逻辑

openseek/competition/LongContext-ICL-Annotation/READMD_zh_cn.md

Lines changed: 0 additions & 90 deletions
This file was deleted.

openseek/competition/LongContext-ICL-Annotation/README.md

Lines changed: 3 additions & 29 deletions
Original file line numberDiff line numberDiff line change
@@ -2,42 +2,16 @@
22

33
Large Language Models Automatic Data Annotation under Long-Context Scenarios.
44

5+
---
6+
57
## News
68
<!-- BEGIN NEWS -->
79
- **[2026-01-20] `Release`:** The competition is now officially live on **Kaggle**. See details: [FlagOS Open Computing Global Challenge](https://www.kaggle.com/competitions/flag-os-open-computing-global-challenge).
810
- **[2026-01-06] `Release`:** The comprehensive competition **FlagOS Open Computing Global Challenge** was officially announced, co-hosted by the **FlagOS Community**, the **Beijing Academy of Artificial Intelligence (BAAI)**, and **CCF ODTC**. See details:
911
[FlagOS开放计算全球挑战赛- AI赛事通 | 数据算法赛](https://www.competehub.dev/zh/competitions/modelscope180)
1012
<!-- END NEWS -->
1113

12-
## Introduction
13-
14-
The LongContext-ICL-Annotation Challenge focuses on automatic data annotation under long-context settings using Large Language Models (LLMs). The competition is built upon the Qwen3-4B model and adopts the In-context Learning (ICL) paradigm to investigate scalable and high-quality automated annotation methods.
15-
16-
Participating teams are required to use the officially provided datasets and design effective ICL-based annotation solutions tailored for ultra-long context scenarios. All submissions will be evaluated on a unified benchmark dataset. The Organizing Committee will conduct standardized evaluations and determine the final rankings based on the official evaluation results.
17-
18-
## Objectives
19-
20-
This challenge takes Large Language Models (LLMs) as the core technical foundation and targets automated data annotation under ultra-long context constraints, aiming to explore novel paradigms that balance annotation efficiency and annotation accuracy. The competition focuses on the following key scientific and engineering challenges:
21-
22-
- 1. Instruction and Prompt Design:
23-
24-
How can effective model instructions and prompt strategies be designed in ultra-long context scenarios to guide LLMs toward stable and high-quality data annotation?
25-
- 2. Ultra-Long Context Construction:
26-
27-
When the number of available annotation examples significantly exceeds the model’s context capacity, how can information-dense and structurally coherent ultra-long context inputs be constructed for target data annotation?
28-
- 3. Multi-Turn and Continuous Annotation:
29-
30-
In automated multi-round dialogue or continuous interaction settings, how can ultra-long contexts be efficiently leveraged to achieve both consistency and scalability in data annotation?
31-
32-
## Challenge Details
33-
34-
- Participating teams are expected to independently design a complete LLM-based automatic data annotation pipeline and validate their approach under a unified dataset and evaluation protocol. Evaluation scores and rankings will be published on a standardized leaderboard.
35-
36-
- In addition to prediction results, teams must submit a technical report and fully reproducible source code in accordance with the competition requirements. The Organizing Committee will reproduce submitted solutions and review the technical design. The final score will be calculated as a weighted combination of prediction performance and technical solution evaluation, with detailed rules specified by the competition.
37-
38-
- Teams are required to submit their technical reports and complete source code to the official OpenSeek GitHub repository designated by the competition.
39-
40-
- For additional details, please refer to [FlagOS platform](https://flagos.io/RaceDetail?id=296fmsd8&lang=en). All competition-related information is subject to the announcements published on the official platform.
14+
---
4115

4216
## Quick Start
4317

0 commit comments

Comments
 (0)