Skip to content

Releases: intel/intel-optimization-for-horovod

v0.28.1.6

18 Dec 23:15
Compare
Choose a tag to compare

Intel® Optimization for Horovod* v0.28.1.6 Release Notes

Intel® Optimization for Horovod* is the Intel optimized distributed training framework to extend the official Horovod (based on v0.28.1) which aims at providing distributed ability to run TensorFlow workloads on Intel GPU clusters. This release contains the following major features:

v0.28.1.5

12 Aug 08:28
Compare
Choose a tag to compare

Intel® Optimization for Horovod* v0.28.1.5 Release Notes

Major Features and Improvements

Intel® Optimization for Horovod* is the Intel optimized distributed training framework to extend the official Horovod (based on v0.28.1) which aims at providing distributed ability to run TensorFlow workloads on Intel GPU clusters. This release contains the following major features:

v0.28.1.4

28 Mar 11:29
Compare
Choose a tag to compare

Intel® Optimization for Horovod* v0.28.1.4 Release Notes

Major Features and Improvements

Intel® Optimization for Horovod* is the Intel optimized distributed training framework to extend the official Horovod (based on v0.28.1) which aims at providing distributed ability to run TensorFlow workloads on Intel GPU clusters. This release contains the following major features:

Known Issues

Intel® Optimization for Horovod* 0.28.1.2

01 Dec 02:17
4d8b248
Compare
Choose a tag to compare

Major Features and Improvements

Intel® Optimization for Horovod* is Intel optimized distributed training framework to extend official Horovod (based on v0.28.1) which aims at providing distributed ability to run TensorFlow workloads in Intel GPU cluster. This release contains following major features:

  • Supported TorusAllreduce operation for cross node AllReduce. This collective operation development is based on oneAPI Collective Communications Library (oneCCL) to do inter-GPU communication primitives that are topology-aware and provide accelerated inter-GPU communication.
  • Supported TensorFlow 2.14.0 and Intel® Extension for TensorFlow* v2.14.0.0 in Intel® Optimization for Horovod*.
  • Supported scale-up and scale-out in Intel® Data Center Max GPU cluster.

Documentations to get started

Intel® Optimization for Horovod* 0.28.1.0

28 Jul 06:30
Compare
Choose a tag to compare

Major Features and Improvements

Intel® Optimization for Horovod* is Intel optimized distributed training framework to extend official Horovod (based on v0.28.1) which aims at providing distributed ability to run TensorFlow workloads in Intel GPU cluster. This release contains following major features:

  • Rebased Intel® Optimization for Horovod* to latest stock v0.28.1 Horovod. The main changes in this rebase includes:
    • Horovod API compatibility issues with tf.keras 2.11 are fixed. Now, HVD wrapper for keras optimizer can work correctly without legacy limitation.
    • Support new implemented methodology in reducescatter and enable batch memory copy for allgather/reducescatter.
  • Refined Intel ® Optimization for Horovod* version to four digits version format v0.28.1.0 by three digits from stock Horovod v0.28.1 and the last one initiated from 0 and increased. It will make it easier for users to understand Intel ® Optimization for Horovod* and stock Horovod version mapping relationship.
  • Supported TensorFlow 2.13.0 and Intel® Extension for TensorFlow* v2.13.0.0 in Intel® Optimization for Horovod*.
  • Supported scale-up and scale-out in Intel® Data Center Max GPU cluster.

Documentations to get started

Intel® Optimization for Horovod* 0.5.0

27 Apr 02:18
Compare
Choose a tag to compare
Pre-release

Major Features and Improvements

Intel® Optimization for Horovod* is Intel optimized distributed training framework to extend official Horovod (based on v0.26.1) which aims at providing distributed ability to run TensorFlow workloads in Intel GPU cluster. This release contains following major features:

  • Enabled All2All/(grouped)AllGather/(grouped)ReduceScatter/BroadcastInplace(Resource) operations for TensorFlow in Intel® Data Center Max GPU cluster. Those collective operation development is based on oneAPI Collective Communications Library (oneCCL) to do inter-GPU communication primitives that are topology-aware and provide accelerated inter-GPU communication.
  • Switched CXX compiler from dpcpp to icpx. New source code build command in how to build.
  • Supported TF2.12 and Intel® Extension for TensorFlow* v1.2 in Intel® Optimization for Horovod*.
  • Supported scale-up and scale-out in Intel® Data Center Max GPU cluster.

Documentations to get started

Intel® Optimization for Horovod* 0.4.0

06 Jan 08:40
64107e0
Compare
Choose a tag to compare
Pre-release

Major Features

Intel® Optimization for Horovod* is Intel optimized distributed training framework to extend official Horovod, which aims at providing distributed ability to run TensorFlow and PyTorch workloads in Intel GPU cluster. It's developed based on public Horovod latest release v0.26.1.

This release contains following major features:

  • Enabled AllReduce/GroupedAllreduce/BroadCast operations for TensorFlow and PyTorch in Intel® Data Center Max GPU Series cluster. Those collective operation development is based on Intel® oneAPI Collective Communications Library (oneCCL) to do inter-GPU communication primitives that are topology-aware and provide accelerated inter-GPU communication.
  • Supported scale-up and scale-out in Intel® Data Center Max GPU Series cluster.
  • Co-worked with Intel® Extension for TensorFlow* v1.1 and Intel® Extension for PyTorch* v1.13.10+xpu.

Documentations to get started