Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
73 changes: 73 additions & 0 deletions openseek/competition/LongContext-ICL-Annotation/READMD_cn.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,73 @@
# 超长长上下文场景中LLM自动数据标注挑战赛

---

## 消息
<!-- BEGIN NEWS -->
- **[2026-01-20] `发布`:** 赛事信息已在 **Kaggle** 正式上线。详情见:[FlagOS Open Computing Global Challenge](https://www.kaggle.com/competitions/flag-os-open-computing-global-challenge).
- **[2026-01-06] `发布`:** 由 **众智 FlagOS 社区**、**北京智源人工智能研究院(BAAI)** 与 **CCF ODTC** 联合主办的综合性大赛 **FlagOS 开放计算全球挑战赛** 正式发布。详情见:
[FlagOS开放计算全球挑战赛- AI赛事通 | 数据算法赛](https://www.competehub.dev/zh/competitions/modelscope180)
<!-- END NEWS -->

---


## 快速开始
### 1. 环境

```bash
openai
torch
flagScale
```

推荐在NVIDIA平台使用 `cd src && bash create_env_nvidia.sh` 创建环境。

### 2. 下载模型权重
```bash
hf download Qwen/Qwen3-4B --local-dir Qwen3-4B
# or
modelscope download --model Qwen/Qwen3-4B
```
### 3. 长文本配置
在`Qwen3-4B/config.json`将原有配置替换为:
```json
"rope_scaling": {
"rope_type": "yarn",
"factor": 4.0,
"original_max_position_embeddings": 32768
}
```
### 4. 模型部署

请根据实际需求,配置 `llm_config.yaml` 文件。启动配置

```bash
cd FlagScale
python run.py --config-path .. --config-name llm_config action=run
```

在模型服务启动后,可通过以下方式测试本地 API:

```bash
python api_test.py
```

如需停止服务,请执行:

```bash
python run.py --config-path .. --config-name llm_config action=stop
```

### 5. 运行/改进基线方法(Baseline)

启动如下命令开始模型标注
```bash
python main.py
```

实现新的标注方法,请修改`method.py`文件。你可以在该文件中:
* 定义新的指令模板、
* 定义新的上下文示例选择策略
* 定义新的模型推理、标注方案
* 添加自定义后处理逻辑
90 changes: 0 additions & 90 deletions openseek/competition/LongContext-ICL-Annotation/READMD_zh_cn.md

This file was deleted.

32 changes: 3 additions & 29 deletions openseek/competition/LongContext-ICL-Annotation/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,42 +2,16 @@

Large Language Models Automatic Data Annotation under Long-Context Scenarios.

---

## News
<!-- BEGIN NEWS -->
- **[2026-01-20] `Release`:** The competition is now officially live on **Kaggle**. See details: [FlagOS Open Computing Global Challenge](https://www.kaggle.com/competitions/flag-os-open-computing-global-challenge).
- **[2026-01-06] `Release`:** The comprehensive competition **FlagOS Open Computing Global Challenge** was officially announced, co-hosted by the **FlagOS Community**, the **Beijing Academy of Artificial Intelligence (BAAI)**, and **CCF ODTC**. See details:
[FlagOS开放计算全球挑战赛- AI赛事通 | 数据算法赛](https://www.competehub.dev/zh/competitions/modelscope180)
<!-- END NEWS -->

## Introduction

The LongContext-ICL-Annotation Challenge focuses on automatic data annotation under long-context settings using Large Language Models (LLMs). The competition is built upon the Qwen3-4B model and adopts the In-context Learning (ICL) paradigm to investigate scalable and high-quality automated annotation methods.

Participating teams are required to use the officially provided datasets and design effective ICL-based annotation solutions tailored for ultra-long context scenarios. All submissions will be evaluated on a unified benchmark dataset. The Organizing Committee will conduct standardized evaluations and determine the final rankings based on the official evaluation results.

## Objectives

This challenge takes Large Language Models (LLMs) as the core technical foundation and targets automated data annotation under ultra-long context constraints, aiming to explore novel paradigms that balance annotation efficiency and annotation accuracy. The competition focuses on the following key scientific and engineering challenges:

- 1. Instruction and Prompt Design:

How can effective model instructions and prompt strategies be designed in ultra-long context scenarios to guide LLMs toward stable and high-quality data annotation?
- 2. Ultra-Long Context Construction:

When the number of available annotation examples significantly exceeds the model’s context capacity, how can information-dense and structurally coherent ultra-long context inputs be constructed for target data annotation?
- 3. Multi-Turn and Continuous Annotation:

In automated multi-round dialogue or continuous interaction settings, how can ultra-long contexts be efficiently leveraged to achieve both consistency and scalability in data annotation?

## Challenge Details

- Participating teams are expected to independently design a complete LLM-based automatic data annotation pipeline and validate their approach under a unified dataset and evaluation protocol. Evaluation scores and rankings will be published on a standardized leaderboard.

- In addition to prediction results, teams must submit a technical report and fully reproducible source code in accordance with the competition requirements. The Organizing Committee will reproduce submitted solutions and review the technical design. The final score will be calculated as a weighted combination of prediction performance and technical solution evaluation, with detailed rules specified by the competition.

- Teams are required to submit their technical reports and complete source code to the official OpenSeek GitHub repository designated by the competition.

- For additional details, please refer to [FlagOS platform](https://flagos.io/RaceDetail?id=296fmsd8&lang=en). All competition-related information is subject to the announcements published on the official platform.
---

Comment on lines 12 to 41
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The removal of the "Introduction", "Objectives", and "Challenge Details" sections significantly reduces the contextual information available to users. These sections are crucial for new participants to understand the competition's purpose, goals, and detailed rules. While a "Quick Start" guide is valuable, a comprehensive overview is equally important for proper engagement and understanding of the project. Consider retaining a condensed version of these sections or providing a clear link to where this information can be found.

## Quick Start

Expand Down
Loading