From e0077ec2fd14ad63b2d3b0fbec831e11c86e7193 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Fran=C3=A7ois=20Hernandez?= Date: Thu, 6 Feb 2025 23:21:39 +0100 Subject: [PATCH] Bump 0.1.1 (#205) --- CHANGELOG.md | 8 ++++++++ README.md | 4 ++-- setup.py | 2 +- 3 files changed, 11 insertions(+), 3 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index ca938771..1458a5fd 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2,6 +2,14 @@ This is just a centralised version of the Github automatically generated Release changelogs. +## 0.1.1 + +* fix rope when very long sequence precision is key by @vince62s in https://github.com/eole-nlp/eole/pull/200 +* Better fix for long rope (training was not optimized) by @vince62s in https://github.com/eole-nlp/eole/pull/201 +* add filtertooshort transform by @vince62s in https://github.com/eole-nlp/eole/pull/202 +* Basic pixtral support, paving the way for vision models 🖼️ by @francoishernandez in https://github.com/eole-nlp/eole/pull/153 +* Clean / rename / simplify by @vince62s in https://github.com/eole-nlp/eole/pull/203 + ## 0.1.0 * reinstate cuda rmsnorm (much faster in fp16/awq) + ct2 enc/dec config by @vince62s in https://github.com/eole-nlp/eole/pull/167 diff --git a/README.md b/README.md index 73733a81..ccd0a996 100644 --- a/README.md +++ b/README.md @@ -61,12 +61,12 @@ You can customize the workflow and build your own images based on specific needs To pull the Docker image: ```bash -docker pull ghcr.io/eole-nlp/eole:0.1.0-torch2.5.1-ubuntu22.04-cuda12.4 +docker pull ghcr.io/eole-nlp/eole:0.1.1-torch2.5.1-ubuntu22.04-cuda12.4 ``` Example one-liner to run a container and open a bash shell within it: ```bash -docker run --rm -it --runtime=nvidia ghcr.io/eole-nlp/eole:0.1.0-torch2.5.1-ubuntu22.04-cuda12.4 +docker run --rm -it --runtime=nvidia ghcr.io/eole-nlp/eole:0.1.1-torch2.5.1-ubuntu22.04-cuda12.4 ``` > **Note**: Ensure you have the [Nvidia Container Toolkit](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html) (formerly nvidia-docker) installed to take advantage of CUDA/GPU features. diff --git a/setup.py b/setup.py index 15c81da6..4e5571b9 100644 --- a/setup.py +++ b/setup.py @@ -11,7 +11,7 @@ description="Open language modeling toolkit based on PyTorch", long_description=long_description, long_description_content_type="text/markdown", - version="0.1.0", + version="0.1.1", packages=find_packages(), project_urls={ "Source": "https://github.com/eole-nlp/eole/",