Skip to content

Commit 08d516a

Browse files
committed
news5
1 parent 5e3d5d1 commit 08d516a

File tree

220 files changed

+70406
-13
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

220 files changed

+70406
-13
lines changed

Gemfile.lock

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -151,7 +151,7 @@ GEM
151151
rouge (4.1.3)
152152
ruby2_keywords (0.0.5)
153153
safe_yaml (1.0.5)
154-
sass-embedded (1.63.6)
154+
sass-embedded (1.64.0)
155155
google-protobuf (~> 3.23)
156156
rake (>= 13.0.0)
157157
sax-machine (1.3.2)

_i18n/en.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -75,6 +75,7 @@ news:
7575
announcement_2: A paper was accepted by SEKE 2023(ccf-c)!
7676
announcement_3: SHU_Chat is available now!
7777
announcement_4: Two papers have been accepted by EMNLP2023! 🎉🎉
78+
announcement_5: A paper accepted for LREC-COLING 2024! ✌✌
7879

7980
pagination:
8081
newer: Newer

_i18n/en/_news/announcement_5.md

Lines changed: 25 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,32 @@
1-
欧洲语言资源协会(ELRA)和国际计算语言学委员会(ICCL)将联合主办 2024 年计算语言学、语言资源和评估联合国际会议(LREC-COLING 2024),该会议将于 2024 年 5 月 20-25 日在意大利都灵举行。
1+
The European Language Resources Association (ELRA) and the International Council for Computational Linguistics (ICCL) are co-hosting the Joint International Conference on Computational Linguistics, Language Resources and Evaluation 2024 (LREC-COLING 2024), which will take place from 20-25 May 2024 in Turin, Italy.
22

3-
此次联合会议将聚焦计算语言学、语音处理、多模态和自然语言处理领域的研究人员和从业者,重点关注评估和资源开发,以支持这些领域的工作。延续 COLING LREC 的悠久传统,本次联合会议将突出重大挑战,提供口头报告和广泛的海报展示,为与会者提供充分交流的机会,并配以丰富的社交活动。
3+
The joint conference will focus on researchers and practitioners in the fields of computational linguistics, speech processing, multimodality and natural language processing, with an emphasis on evaluation and resource development to support work in these areas. Continuing the long tradition of COLING and LREC, this joint conference will highlight major challenges, offer oral presentations and extensive poster presentations, and provide ample opportunities for participants to network, with a rich social programme.
44

5-
此次上海大学自然语言处理与人机交互实验室(H!NTLAB)由李唐同学主要参与的工作1篇论文被LREC-COLING 2024接收。
6-
被录用论文的简要介绍如下:
7-
Towards Human-Like Machine Comprehension:
5+
This time, one paper of the work of the Natural Language Processing and Human-Computer Interaction Laboratory (H!NTLAB) of Shanghai University, in which Tang Li was mainly involved, was accepted by LREC-COLING 2024.
6+
The brief introduction of the accepted paper is as follows:
7+
Towards Human-Like Machine Comprehension.
88
Few-Shot Relational Learning in Visually-Rich Documents
99

10-
类型:Long Paper
10+
**Type:**
11+
Long Paper
1112

12-
作者:王昊 (讲师)、李唐 (2021级硕士研究生)等
13+
**Author(s):**
14+
Hao Wang (Lecturer), Tang Li (M.S. Candidate, Class of 2021), et al.
1315

14-
简介:键值关系在视觉丰富文档(VRD)中很普遍,通常在不同的空间区域中描述,并伴有特定的颜色和字体样式。这些非文本是重要的特征,极大地增强了人类对这种关系三元组的理解。然而,当前的文档AI方法无法考虑与视觉和空间特征相关的这些有价值的先验信息,导致性能次优,尤其是在处理有限的示例时。为了解决这一局限性,我们的研究重点是少样本关系学习,特别是针对VRD中键值关系三元组的提取。鉴于缺乏适用于该任务的数据集,我们引入了两个新的基于现有监督基准数据集的少样本基准。此外,我们提出了一种结合关系二维空间先验和原型校正技术的变分方法。这种方法旨在生成关系表示以类似于人类感知的方式意识到空间上下文和看不见的关系。实验结果证明了我们提出的方法的有效性,展示了其优于现有方法的能力。这项研究也为实际应用开辟了新的可能性。123
16+
**Introduction:**
17+
Key-value relationships are common in visually-rich documents (VRDs), which are usually described in different spatial regions accompanied by specific colours and font styles. These non-texts are important features that greatly enhance human understanding of this relational triad. However, current document AI approaches are unable to take into account this valuable a priori information related to visual and spatial features, resulting in sub-optimal performance, especially when dealing with a limited number of examples. To address this limitation, our research focuses on few-shot relational learning, especially for the extraction of key-value relation triples in VRD. In view of the lack of datasets suitable for this task, we have introduced two new few-shot datums based on the existing supervised datum datasets. In addition, we propose a variational method that combines relational two-dimensional space prior and prototype correction techniques. This approach aims to generate relational representations that are aware of spatial context and unseen relationships in a manner similar to human perception. Experimental results demonstrate the effectiveness of our proposed method and demonstrate its ability to outperform existing methods. This research also opens up new possibilities for practical applications.
1518

19+
<div align="center">
20+
<img src="../../assets/img/news/announcement_5_p1.png" width = "800" height = "500" alt="图1" align=center />
21+
</div>
22+
<div style="text-align:center">Figure 1: Visual rich document few-shot dataset with multimodal features</div>
23+
24+
25+
<div align="center">
26+
<img src="../../assets/img/news/announcement_5_p2.png" width = "800" height = "500" alt="图2" align=center />
27+
</div>
28+
<div style="text-align:center">Figure 2: Diagram of the model architecture combining relational 2D space prior and prototype correction techniques</div>
29+
<div align="center">
30+
<img src="../../assets/img/news/announcement_5_p3.png" width = "800" height = "500" alt="图3" align=center />
31+
</div>
32+
<div style="text-align:center">Figure 3: Comparison of the performance of different methods on different datasets</div>

_i18n/中文/_news/announcement_5.md

Lines changed: 22 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -7,9 +7,28 @@
77
Towards Human-Like Machine Comprehension:
88
Few-Shot Relational Learning in Visually-Rich Documents
99

10-
类型:Long Paper
10+
**类型:**
11+
Long Paper
1112

12-
作者:王昊 (讲师)、李唐 (2021级硕士研究生)等
13+
**作者:**
14+
王昊 (讲师)、李唐 (2021级硕士研究生)等
1315

14-
简介:键值关系在视觉丰富文档(VRD)中很普遍,通常在不同的空间区域中描述,并伴有特定的颜色和字体样式。这些非文本是重要的特征,极大地增强了人类对这种关系三元组的理解。然而,当前的文档AI方法无法考虑与视觉和空间特征相关的这些有价值的先验信息,导致性能次优,尤其是在处理有限的示例时。为了解决这一局限性,我们的研究重点是少样本关系学习,特别是针对VRD中键值关系三元组的提取。鉴于缺乏适用于该任务的数据集,我们引入了两个新的基于现有监督基准数据集的少样本基准。此外,我们提出了一种结合关系二维空间先验和原型校正技术的变分方法。这种方法旨在生成关系表示以类似于人类感知的方式意识到空间上下文和看不见的关系。实验结果证明了我们提出的方法的有效性,展示了其优于现有方法的能力。这项研究也为实际应用开辟了新的可能性。123
16+
**简介:**
17+
键值关系在视觉丰富文档(VRD)中很普遍,通常在不同的空间区域中描述,并伴有特定的颜色和字体样式。这些非文本是重要的特征,极大地增强了人类对这种关系三元组的理解。然而,当前的文档AI方法无法考虑与视觉和空间特征相关的这些有价值的先验信息,导致性能次优,尤其是在处理有限的示例时。为了解决这一局限性,我们的研究重点是少样本关系学习,特别是针对VRD中键值关系三元组的提取。鉴于缺乏适用于该任务的数据集,我们引入了两个新的基于现有监督基准数据集的少样本基准。此外,我们提出了一种结合关系二维空间先验和原型校正技术的变分方法。这种方法旨在生成关系表示以类似于人类感知的方式意识到空间上下文和看不见的关系。实验结果证明了我们提出的方法的有效性,展示了其优于现有方法的能力。这项研究也为实际应用开辟了新的可能性。
18+
19+
<div align="center">
20+
<img src="../../assets/img/news/announcement_5_p1.png" width = "800" height = "500" alt="图1" align=center />
21+
</div>
22+
23+
<div style="text-align:center">图1:具有多模态特征的视觉富文档少样本数据集</div>
24+
25+
<div align="center">
26+
<img src="../../assets/img/news/announcement_5_p2.png" width = "800" height = "500" alt="图2" align=center />
27+
</div>
28+
29+
<div style="text-align:center">图2:结合关系二维空间先验和原型校正技术的模型架构图</div>
30+
<div align="center">
31+
<img src="../../assets/img/news/announcement_5_p3.png" width = "800" height = "500" alt="图3" align=center />
32+
</div>
33+
<div style="text-align:center">图3:不同数据集不同方法的性能比较</div>
1534

_news/announcement_5.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,11 @@
11
---
22
layout: post
33
title: news.titles.announcement_5
4-
date: 2024-03-24 15:59:00-0400
4+
date: 2024-03-24 20:58:00+0800
55
inline: false
66
related_posts: false
77
---
88

99
{% translate_file _news/announcement_5.md %}
10+
11+

0 commit comments

Comments
 (0)