Skip to content

Commit

Permalink
[Version] v1.7.1. (#445)
Browse files Browse the repository at this point in the history
  • Loading branch information
Duyi-Wang authored Jun 12, 2024
1 parent c8ba661 commit 38658b1
Show file tree
Hide file tree
Showing 2 changed files with 11 additions and 1 deletion.
10 changes: 10 additions & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -1,4 +1,14 @@
# CHANGELOG
# [Version v1.7.0](https://github.com/intel/xFasterTransformer/releases/tag/v1.7.0)
v1.7.1 - Continuous batching feature supports ChatGLM2/3.

## Functionality
- Add continuous batching support of ChatGLM2/3 models.
- Qwen2Convert supports quantized Qwen2 models by GPTQ, such as GPTQ-Int8 and GPTQ-Int4, by param `from_quantized_model="gptq"`.

## BUG fix
- Fixed the segament fault error when running with more than 2 ranks in vllm-xft serving.

# [Version v1.7.0](https://github.com/intel/xFasterTransformer/releases/tag/v1.7.0)
v1.7.0 - Continuous batching feature supported.

Expand Down
2 changes: 1 addition & 1 deletion VERSION
Original file line number Diff line number Diff line change
@@ -1 +1 @@
1.7.0
1.7.1

0 comments on commit 38658b1

Please sign in to comment.