Skip to content

Add LongContext-ICL annotations#180

Merged
Haiming94 merged 1 commit intomainfrom
LongContext-ICL-Annotation
Jan 22, 2026
Merged

Add LongContext-ICL annotations#180
Haiming94 merged 1 commit intomainfrom
LongContext-ICL-Annotation

Conversation

@Haiming94
Copy link
Copy Markdown
Collaborator

No description provided.

Signed-off-by: Haiming94 <931284009@qq.com>
@Haiming94 Haiming94 merged commit 0c58a98 into main Jan 22, 2026
1 of 4 checks passed
@gemini-code-assist
Copy link
Copy Markdown

Summary of Changes

Hello @Haiming94, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request refactors the documentation structure for the LongContext-ICL-Annotation challenge. It simplifies both the English and Chinese README files by removing detailed introductory and challenge objective sections, focusing instead on providing concise news updates and quick start guides. The aim is to make the initial setup information more direct and accessible within the repository.

Highlights

  • Documentation Refactoring: The pull request streamlines the documentation for the LongContext-ICL-Annotation challenge by reorganizing its README files.
  • New Chinese README Added: A new Chinese README file (READMD_cn.md) has been added, focusing on quick start instructions for the challenge.
  • English README Simplified: The English README file (README.md) has been modified to remove detailed introductory sections, now concentrating on news and quick start guides.
  • Old Chinese README Removed: An older Chinese README file (READMD_zh_cn.md), which contained more extensive competition details, has been removed.

🧠 New Feature in Public Preview: You can now enable Memory to help Gemini Code Assist learn from your team's feedback. This makes future code reviews more consistent and personalized to your project's style. Click here to enable Memory in your admin console.

Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point by creating a comment using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands on the current page.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in pull request comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

Copy link
Copy Markdown

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

The pull request updates the documentation by removing detailed introductory, objective, and challenge detail sections from both the English and Chinese READMEs, focusing instead on a "Quick Start" guide. While streamlining documentation can be beneficial, the removal of these comprehensive sections might reduce the clarity and context for new users trying to understand the competition's scope and rules. It would be advisable to ensure that this critical information is still easily accessible, perhaps by linking to a dedicated competition page or a more detailed project overview.

Comment on lines 12 to 41
<!-- END NEWS -->

## Introduction

The LongContext-ICL-Annotation Challenge focuses on automatic data annotation under long-context settings using Large Language Models (LLMs). The competition is built upon the Qwen3-4B model and adopts the In-context Learning (ICL) paradigm to investigate scalable and high-quality automated annotation methods.

Participating teams are required to use the officially provided datasets and design effective ICL-based annotation solutions tailored for ultra-long context scenarios. All submissions will be evaluated on a unified benchmark dataset. The Organizing Committee will conduct standardized evaluations and determine the final rankings based on the official evaluation results.

## Objectives

This challenge takes Large Language Models (LLMs) as the core technical foundation and targets automated data annotation under ultra-long context constraints, aiming to explore novel paradigms that balance annotation efficiency and annotation accuracy. The competition focuses on the following key scientific and engineering challenges:

- 1. Instruction and Prompt Design:

How can effective model instructions and prompt strategies be designed in ultra-long context scenarios to guide LLMs toward stable and high-quality data annotation?
- 2. Ultra-Long Context Construction:

When the number of available annotation examples significantly exceeds the model’s context capacity, how can information-dense and structurally coherent ultra-long context inputs be constructed for target data annotation?
- 3. Multi-Turn and Continuous Annotation:

In automated multi-round dialogue or continuous interaction settings, how can ultra-long contexts be efficiently leveraged to achieve both consistency and scalability in data annotation?

## Challenge Details

- Participating teams are expected to independently design a complete LLM-based automatic data annotation pipeline and validate their approach under a unified dataset and evaluation protocol. Evaluation scores and rankings will be published on a standardized leaderboard.

- In addition to prediction results, teams must submit a technical report and fully reproducible source code in accordance with the competition requirements. The Organizing Committee will reproduce submitted solutions and review the technical design. The final score will be calculated as a weighted combination of prediction performance and technical solution evaluation, with detailed rules specified by the competition.

- Teams are required to submit their technical reports and complete source code to the official OpenSeek GitHub repository designated by the competition.

- For additional details, please refer to [FlagOS platform](https://flagos.io/RaceDetail?id=296fmsd8&lang=en). All competition-related information is subject to the announcements published on the official platform.
---

Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The removal of the "Introduction", "Objectives", and "Challenge Details" sections significantly reduces the contextual information available to users. These sections are crucial for new participants to understand the competition's purpose, goals, and detailed rules. While a "Quick Start" guide is valuable, a comprehensive overview is equally important for proper engagement and understanding of the project. Consider retaining a condensed version of these sections or providing a clear link to where this information can be found.

@Haiming94 Haiming94 deleted the LongContext-ICL-Annotation branch January 22, 2026 08:12
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant