|
2 | 2 |
|
3 | 3 | Large Language Models Automatic Data Annotation under Long-Context Scenarios. |
4 | 4 |
|
| 5 | +--- |
| 6 | + |
5 | 7 | ## News |
6 | 8 | <!-- BEGIN NEWS --> |
7 | 9 | - **[2026-01-20] `Release`:** The competition is now officially live on **Kaggle**. See details: [FlagOS Open Computing Global Challenge](https://www.kaggle.com/competitions/flag-os-open-computing-global-challenge). |
8 | 10 | - **[2026-01-06] `Release`:** The comprehensive competition **FlagOS Open Computing Global Challenge** was officially announced, co-hosted by the **FlagOS Community**, the **Beijing Academy of Artificial Intelligence (BAAI)**, and **CCF ODTC**. See details: |
9 | 11 | [FlagOS开放计算全球挑战赛- AI赛事通 | 数据算法赛](https://www.competehub.dev/zh/competitions/modelscope180) |
10 | 12 | <!-- END NEWS --> |
11 | 13 |
|
12 | | -## Introduction |
13 | | - |
14 | | -The LongContext-ICL-Annotation Challenge focuses on automatic data annotation under long-context settings using Large Language Models (LLMs). The competition is built upon the Qwen3-4B model and adopts the In-context Learning (ICL) paradigm to investigate scalable and high-quality automated annotation methods. |
15 | | - |
16 | | -Participating teams are required to use the officially provided datasets and design effective ICL-based annotation solutions tailored for ultra-long context scenarios. All submissions will be evaluated on a unified benchmark dataset. The Organizing Committee will conduct standardized evaluations and determine the final rankings based on the official evaluation results. |
17 | | - |
18 | | -## Objectives |
19 | | - |
20 | | -This challenge takes Large Language Models (LLMs) as the core technical foundation and targets automated data annotation under ultra-long context constraints, aiming to explore novel paradigms that balance annotation efficiency and annotation accuracy. The competition focuses on the following key scientific and engineering challenges: |
21 | | - |
22 | | -- 1. Instruction and Prompt Design: |
23 | | - |
24 | | - How can effective model instructions and prompt strategies be designed in ultra-long context scenarios to guide LLMs toward stable and high-quality data annotation? |
25 | | -- 2. Ultra-Long Context Construction: |
26 | | - |
27 | | - When the number of available annotation examples significantly exceeds the model’s context capacity, how can information-dense and structurally coherent ultra-long context inputs be constructed for target data annotation? |
28 | | -- 3. Multi-Turn and Continuous Annotation: |
29 | | - |
30 | | - In automated multi-round dialogue or continuous interaction settings, how can ultra-long contexts be efficiently leveraged to achieve both consistency and scalability in data annotation? |
31 | | - |
32 | | -## Challenge Details |
33 | | - |
34 | | -- Participating teams are expected to independently design a complete LLM-based automatic data annotation pipeline and validate their approach under a unified dataset and evaluation protocol. Evaluation scores and rankings will be published on a standardized leaderboard. |
35 | | - |
36 | | -- In addition to prediction results, teams must submit a technical report and fully reproducible source code in accordance with the competition requirements. The Organizing Committee will reproduce submitted solutions and review the technical design. The final score will be calculated as a weighted combination of prediction performance and technical solution evaluation, with detailed rules specified by the competition. |
37 | | - |
38 | | -- Teams are required to submit their technical reports and complete source code to the official OpenSeek GitHub repository designated by the competition. |
39 | | - |
40 | | -- For additional details, please refer to [FlagOS platform](https://flagos.io/RaceDetail?id=296fmsd8&lang=en). All competition-related information is subject to the announcements published on the official platform. |
| 14 | +--- |
41 | 15 |
|
42 | 16 | ## Quick Start |
43 | 17 |
|
|
0 commit comments