Skip to content

Commit

Permalink
[0.24.0 branch] Release 0.24.0 changes (#1154)
Browse files Browse the repository at this point in the history
  • Loading branch information
zachgk committed Oct 12, 2023
1 parent 9f0d840 commit 9481927
Show file tree
Hide file tree
Showing 11 changed files with 28 additions and 28 deletions.
12 changes: 6 additions & 6 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -50,20 +50,20 @@ brew services stop djl-serving
For Ubuntu

```
curl -O https://publish.djl.ai/djl-serving/djl-serving_0.23.0-1_all.deb
sudo dpkg -i djl-serving_0.23.0-1_all.deb
curl -O https://publish.djl.ai/djl-serving/djl-serving_0.24.0-1_all.deb
sudo dpkg -i djl-serving_0.24.0-1_all.deb
```

For Windows

We are considering to create a `chocolatey` package for Windows. For the time being, you can
download djl-serving zip file from [here](https://publish.djl.ai/djl-serving/serving-0.23.0.zip).
download djl-serving zip file from [here](https://publish.djl.ai/djl-serving/serving-0.24.0.zip).

```
curl -O https://publish.djl.ai/djl-serving/serving-0.23.0.zip
unzip serving-0.23.0.zip
curl -O https://publish.djl.ai/djl-serving/serving-0.24.0.zip
unzip serving-0.24.0.zip
# start djl-serving
serving-0.23.0\bin\serving.bat
serving-0.24.0\bin\serving.bat
```

### Docker
Expand Down
16 changes: 8 additions & 8 deletions benchmark/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -43,25 +43,25 @@ sudo snap alias djlbench djl-bench
- Or download .deb package from S3

```
curl -O https://publish.djl.ai/djl-bench/0.23.0/djl-bench_0.23.0-1_all.deb
sudo dpkg -i djl-bench_0.23.0-1_all.deb
curl -O https://publish.djl.ai/djl-bench/0.24.0/djl-bench_0.24.0-1_all.deb
sudo dpkg -i djl-bench_0.24.0-1_all.deb
```

For macOS, centOS or Amazon Linux 2

You can download djl-bench zip file from [here](https://publish.djl.ai/djl-bench/0.23.0/benchmark-0.23.0.zip).
You can download djl-bench zip file from [here](https://publish.djl.ai/djl-bench/0.24.0/benchmark-0.24.0.zip).

```
curl -O https://publish.djl.ai/djl-bench/0.23.0/benchmark-0.23.0.zip
unzip benchmark-0.23.0.zip
rm benchmark-0.23.0.zip
sudo ln -s $PWD/benchmark-0.23.0/bin/benchmark /usr/bin/djl-bench
curl -O https://publish.djl.ai/djl-bench/0.24.0/benchmark-0.24.0.zip
unzip benchmark-0.24.0.zip
rm benchmark-0.24.0.zip
sudo ln -s $PWD/benchmark-0.24.0/bin/benchmark /usr/bin/djl-bench
```

For Windows

We are considering to create a `chocolatey` package for Windows. For the time being, you can
download djl-bench zip file from [here](https://publish.djl.ai/djl-bench/0.23.0/benchmark-0.23.0.zip).
download djl-bench zip file from [here](https://publish.djl.ai/djl-bench/0.24.0/benchmark-0.24.0.zip).

Or you can run benchmark using gradle:

Expand Down
4 changes: 2 additions & 2 deletions engines/python/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -29,13 +29,13 @@ The javadocs output is generated in the `build/doc/javadoc` folder.
## Installation
You can pull the Python engine from the central Maven repository by including the following dependency:

- ai.djl.python:python:0.23.0
- ai.djl.python:python:0.24.0

```xml
<dependency>
<groupId>ai.djl.python</groupId>
<artifactId>python</artifactId>
<version>0.23.0</version>
<version>0.24.0</version>
<scope>runtime</scope>
</dependency>
```
Expand Down
2 changes: 1 addition & 1 deletion serving/docker/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@
# BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, express or implied. See the License for
# the specific language governing permissions and limitations under the License.
FROM ubuntu:20.04 AS base
ARG djl_version=0.24.0~SNAPSHOT
ARG djl_version=0.24.0

COPY scripts scripts/
RUN mkdir -p /opt/djl/conf && \
Expand Down
6 changes: 3 additions & 3 deletions serving/docker/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@ mkdir models
cd models
curl -O https://resources.djl.ai/test-models/pytorch/bert_qa_jit.tar.gz

docker run -it --rm -v $PWD:/opt/ml/model -p 8080:8080 deepjavalibrary/djl-serving:0.23.0
docker run -it --rm -v $PWD:/opt/ml/model -p 8080:8080 deepjavalibrary/djl-serving:0.24.0
```

### GPU
Expand All @@ -42,7 +42,7 @@ mkdir models
cd models
curl -O https://resources.djl.ai/test-models/pytorch/bert_qa_jit.tar.gz

docker run -it --runtime=nvidia --shm-size 2g -v $PWD:/opt/ml/model -p 8080:8080 deepjavalibrary/djl-serving:0.23.0-pytorch-cu118
docker run -it --runtime=nvidia --shm-size 2g -v $PWD:/opt/ml/model -p 8080:8080 deepjavalibrary/djl-serving:0.24.0-pytorch-cu118
```

### AWS Inferentia
Expand All @@ -52,5 +52,5 @@ mkdir models
cd models

curl -O https://resources.djl.ai/test-models/pytorch/resnet18_inf2_2_4.tar.gz
docker run --device /dev/neuron0 -it --rm -v $PWD:/opt/ml/model -p 8080:8080 deepjavalibrary/djl-serving:0.23.0-pytorch-inf2
docker run --device /dev/neuron0 -it --rm -v $PWD:/opt/ml/model -p 8080:8080 deepjavalibrary/djl-serving:0.24.0-pytorch-inf2
```
2 changes: 1 addition & 1 deletion serving/docker/aarch64.Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@
# BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, express or implied. See the License for
# the specific language governing permissions and limitations under the License.
FROM arm64v8/ubuntu:20.04
ARG djl_version=0.24.0~SNAPSHOT
ARG djl_version=0.24.0
ARG torch_version=2.0.1

EXPOSE 8080
Expand Down
6 changes: 3 additions & 3 deletions serving/docker/deepspeed.Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -11,18 +11,18 @@
# the specific language governing permissions and limitations under the License.
ARG version=11.8.0-cudnn8-devel-ubuntu20.04
FROM nvidia/cuda:$version
ARG djl_version=0.24.0~SNAPSHOT
ARG djl_version=0.24.0
ARG python_version=3.9
ARG torch_version=2.0.1
ARG torch_vision_version=0.15.2
ARG vllm_version=0.2.0
ARG deepspeed_wheel="https://publish.djl.ai/deepspeed/deepspeed-nightly-py2.py3-none-any.whl"
ARG deepspeed_wheel="https://publish.djl.ai/deepspeed/deepspeed-0.10.0-py2.py3-none-any.whl"
ARG flash_attn_wheel="https://publish.djl.ai/flash_attn/flash_attn_1-1.0.9-cp39-cp39-linux_x86_64.whl"
ARG dropout_layer_norm_wheel="https://publish.djl.ai/flash_attn/dropout_layer_norm-0.1-cp39-cp39-linux_x86_64.whl"
ARG rotary_emb_wheel="https://publish.djl.ai/flash_attn/rotary_emb-0.1-cp39-cp39-linux_x86_64.whl"
ARG flash_attn_2_wheel="https://publish.djl.ai/flash_attn/flash_attn-2.0.1-cp39-cp39-linux_x86_64.whl"
ARG lmi_vllm_wheel="https://publish.djl.ai/lmi_vllm/lmi_vllm-0.1.1-cp39-cp39-linux_x86_64.whl"
ARG lmi_dist_wheel="https://publish.djl.ai/lmi_dist/lmi_dist-nightly-py3-none-any.whl"
ARG lmi_dist_wheel="https://publish.djl.ai/lmi_dist/lmi_dist-0.9.4-py3-none-any.whl"
ARG seq_scheduler_wheel="https://publish.djl.ai/seq_scheduler/seq_scheduler-0.1.0-py3-none-any.whl"
ARG peft_wheel="https://publish.djl.ai/peft/peft-0.5.0alpha-py3-none-any.whl"
ARG mmaploader_wheel="https://publish.djl.ai/mmaploader/mmaploader-nightly-py3-none-any.whl"
Expand Down
2 changes: 1 addition & 1 deletion serving/docker/fastertransformer.Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
# the specific language governing permissions and limitations under the License.
ARG version=11.8.0-cudnn8-devel-ubuntu20.04
FROM nvidia/cuda:$version
ARG djl_version=0.24.0~SNAPSHOT
ARG djl_version=0.24.0
ARG python_version=3.9
ARG ft_version="llama"
ARG triton_version="r23.04"
Expand Down
2 changes: 1 addition & 1 deletion serving/docker/pytorch-cu118.Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ ARG version=11.8.0-cudnn8-devel-ubuntu20.04

FROM nvidia/cuda:$version as base

ARG djl_version=0.24.0~SNAPSHOT
ARG djl_version=0.24.0
ARG torch_version=2.0.1
ARG torch_vision_version=0.15.2
ARG python_version=3.9
Expand Down
2 changes: 1 addition & 1 deletion serving/docker/pytorch-inf2.Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@
# BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, express or implied. See the License for
# the specific language governing permissions and limitations under the License.
FROM ubuntu:20.04
ARG djl_version=0.24.0~SNAPSHOT
ARG djl_version=0.24.0
ARG torch_version=1.13.1
ARG python_version=3.8
ARG torch_neuronx_version=1.13.1.1.11.0
Expand Down
2 changes: 1 addition & 1 deletion wlm/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ You can pull the server from the central Maven repository by including the follo
<dependency>
<groupId>ai.djl.serving</groupId>
<artifactId>wlm</artifactId>
<version>0.23.0</version>
<version>0.24.0</version>
</dependency>
```

0 comments on commit 9481927

Please sign in to comment.