Skip to content

Commit

Permalink
update docker tag to 2.5.0-SNAPSHOT (intel#9443)
Browse files Browse the repository at this point in the history
  • Loading branch information
liu-shaojun authored Nov 13, 2023
1 parent ddb920a commit ee8ce18
Show file tree
Hide file tree
Showing 104 changed files with 195 additions and 195 deletions.
4 changes: 2 additions & 2 deletions .github/actions/ppml-occlum-exampletests-action/action.yml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ inputs:
image-tag:
description: 'image tag'
required: true
default: '2.4.0-SNAPSHOT'
default: '2.5.0-SNAPSHOT'
runs:
using: "composite"
steps:
Expand All @@ -31,7 +31,7 @@ runs:
export IMAGE=${{ env.DEFAULT_IMAGE }}
docker pull $IMAGE
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.4.0-SNAPSHOT
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.5.0-SNAPSHOT
docker stop $CONTAINER_NAME
docker rm -f $CONTAINER_NAME
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ inputs:
image-tag:
description: 'image tag'
required: true
default: '2.4.0-SNAPSHOT'
default: '2.5.0-SNAPSHOT'
runs:
using: "composite"
steps:
Expand All @@ -19,7 +19,7 @@ runs:
run: |
export IMAGE=${{ env.DEFAULT_IMAGE }}
docker pull $IMAGE
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.4.0-SNAPSHOT
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.5.0-SNAPSHOT
export NO_PROXY=10.239.45.10:8081,10.112.231.51,10.239.45.10,172.168.0.205
export kubernetes_master_url=172.168.0.205
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ inputs:
image-tag:
description: 'image tag'
required: true
default: '2.4.0-SNAPSHOT-EDMM6'
default: '2.5.0-SNAPSHOT-EDMM6'
runs:
using: "composite"
steps:
Expand All @@ -32,7 +32,7 @@ runs:
export IMAGE=${{ env.DEFAULT_IMAGE }}
docker pull $IMAGE
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.4.0-SNAPSHOT-EDMM
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.5.0-SNAPSHOT-EDMM
docker stop $CONTAINER_NAME
docker rm -f $CONTAINER_NAME
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ inputs:
image-tag:
description: 'image tag'
required: true
default: '2.4.0-SNAPSHOT-EDMM'
default: '2.5.0-SNAPSHOT-EDMM'
runs:
using: "composite"
steps:
Expand All @@ -20,7 +20,7 @@ runs:
whoami
export IMAGE=${{ env.DEFAULT_IMAGE }}
docker pull $IMAGE
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.4.0-SNAPSHOT-EDMM
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.5.0-SNAPSHOT-EDMM
export NO_PROXY=10.239.45.10:8081,10.112.231.51,10.239.45.10,172.168.0.205,172.168.0.210
export kubernetes_master_url=172.168.0.210
export SPARK_HOME=/opt/spark-3.1.2-bin-hadoop3.2
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/BiWeekly_docker.yml
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ jobs:
cd docker/bigdl-k8s
export IMAGE=intelanalytics/bigdl-k8s
export TIMESTAMP=`date '+%Y%m%d'`
export TAG=2.4.0-SNAPSHOT-${TIMESTAMP}
export TAG=2.5.0-SNAPSHOT-${TIMESTAMP}
echo "########################################"
echo "################# bigdl-k8s 3.1.3 #######"
echo "########################################"
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/license-scan.yml
Original file line number Diff line number Diff line change
Expand Up @@ -62,8 +62,8 @@ jobs:
echo "ls -d scala/assembly/target/*"
ls -d scala/assembly/target/*
echo ""
echo "zipinfo -1 scala/assembly/target/bigdl-assembly-spark_3.1.3-2.4.0-SNAPSHOT-dist-all.zip 'jars/*.jar'"
zipinfo -1 scala/assembly/target/bigdl-assembly-spark_3.1.3-2.4.0-SNAPSHOT-dist-all.zip 'jars/*.jar'
echo "zipinfo -1 scala/assembly/target/bigdl-assembly-spark_3.1.3-2.5.0-SNAPSHOT-dist-all.zip 'jars/*.jar'"
zipinfo -1 scala/assembly/target/bigdl-assembly-spark_3.1.3-2.5.0-SNAPSHOT-dist-all.zip 'jars/*.jar'
- name: Create Job Badge
if: ${{ always() }}
Expand Down
8 changes: 4 additions & 4 deletions .github/workflows/nightly_build.yml
Original file line number Diff line number Diff line change
Expand Up @@ -213,7 +213,7 @@ jobs:
docker tag ${IMAGE}-spark-3.1.3:${TAG} 10.239.45.10/arda/${IMAGE}-spark-3.1.3:${TAG}
docker push 10.239.45.10/arda/${IMAGE}-spark-3.1.3:${TAG}
# tag SNAPSHOT
export TAG_SNAPSHOT=2.4.0-SNAPSHOT
export TAG_SNAPSHOT=2.5.0-SNAPSHOT
docker tag 10.239.45.10/arda/${IMAGE}-spark-3.1.3:${TAG} ${IMAGE}-spark-3.1.3:${TAG_SNAPSHOT}
docker push ${IMAGE}-spark-3.1.3:${TAG_SNAPSHOT}
docker rmi -f ${IMAGE}-spark-3.1.3:${TAG} 10.239.45.10/arda/${IMAGE}-spark-3.1.3:${TAG} ${IMAGE}-spark-3.1.3:${TAG_SNAPSHOT}
Expand All @@ -240,7 +240,7 @@ jobs:
docker tag ${IMAGE}:${TAG} 10.239.45.10/arda/${IMAGE}-spark-3.1.3:${TAG}
docker push 10.239.45.10/arda/${IMAGE}-spark-3.1.3:${TAG}
# tag SNAPSHOT
export TAG_SNAPSHOT=2.4.0-SNAPSHOT
export TAG_SNAPSHOT=2.5.0-SNAPSHOT
docker tag 10.239.45.10/arda/${IMAGE}-spark-3.1.3:${TAG} ${IMAGE}:${TAG_SNAPSHOT}
docker push ${IMAGE}:${TAG_SNAPSHOT}
docker rmi -f ${IMAGE}-spark-3.1.3:${TAG} ${IMAGE}:${TAG} 10.239.45.10/arda/${IMAGE}-spark-3.1.3:${TAG} ${IMAGE}:${TAG_SNAPSHOT}
Expand All @@ -261,7 +261,7 @@ jobs:
docker tag ${IMAGE}:${TAG} 10.239.45.10/arda/${IMAGE}:${TAG}
docker push 10.239.45.10/arda/${IMAGE}:${TAG}
# tag SNAPSHOT
export TAG_SNAPSHOT=2.4.0-SNAPSHOT
export TAG_SNAPSHOT=2.5.0-SNAPSHOT
docker tag 10.239.45.10/arda/${IMAGE}:${TAG} ${IMAGE}:${TAG_SNAPSHOT}
docker push ${IMAGE}:${TAG_SNAPSHOT}
docker rmi -f ${IMAGE}:${TAG_SNAPSHOT} ${IMAGE}:${TAG} 10.239.45.10/arda/${IMAGE}:${TAG}
Expand All @@ -288,7 +288,7 @@ jobs:
docker tag ${IMAGE}:${TAG} 10.239.45.10/arda/${IMAGE}:${TAG}
docker push 10.239.45.10/arda/${IMAGE}:${TAG}
# tag SNAPSHOT
export TAG_SNAPSHOT=2.4.0-SNAPSHOT
export TAG_SNAPSHOT=2.5.0-SNAPSHOT
docker tag 10.239.45.10/arda/${IMAGE}:${TAG} ${IMAGE}:${TAG_SNAPSHOT}
docker push ${IMAGE}:${TAG_SNAPSHOT}
docker rmi -f ${IMAGE}:${TAG_SNAPSHOT} ${IMAGE}:${TAG} 10.239.45.10/arda/${IMAGE}:${TAG}
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/nightly_test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -1382,7 +1382,7 @@ jobs:
- name: set env
env:
DEFAULT_IMAGE: '10.239.45.10/arda/intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum'
DEFAULT_TAG: '2.4.0-SNAPSHOT'
DEFAULT_TAG: '2.5.0-SNAPSHOT'
run: |
echo "TAG=${{ github.event.inputs.tag || env.DEFAULT_TAG }}" >> $GITHUB_ENV
echo "IMAGE=${{ github.event.inputs.image || env.DEFAULT_IMAGE }}" >> $GITHUB_ENV
Expand Down Expand Up @@ -1416,7 +1416,7 @@ jobs:
- name: set env
env:
DEFAULT_IMAGE: '10.239.45.10/arda/intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum'
DEFAULT_TAG: '2.4.0-SNAPSHOT-EDMM'
DEFAULT_TAG: '2.5.0-SNAPSHOT-EDMM'
run: |
echo "TAG=${{ github.event.inputs.tag || env.DEFAULT_TAG }}" >> $GITHUB_ENV
echo "IMAGE=${{ github.event.inputs.image || env.DEFAULT_IMAGE }}" >> $GITHUB_ENV
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/weekly_test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,7 @@ jobs:
- name: set env
env:
DEFAULT_IMAGE: '10.239.45.10/arda/intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum'
DEFAULT_TAG: '2.4.0-SNAPSHOT'
DEFAULT_TAG: '2.5.0-SNAPSHOT'
run: |
echo "TAG=${{ github.event.inputs.tag || env.DEFAULT_TAG }}" >> $GITHUB_ENV
echo "IMAGE=${{ github.event.inputs.image || env.DEFAULT_IMAGE }}" >> $GITHUB_ENV
Expand Down Expand Up @@ -81,7 +81,7 @@ jobs:
- name: set env
env:
DEFAULT_IMAGE: '10.239.45.10/arda/intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum'
DEFAULT_TAG: '2.4.0-SNAPSHOT-EDMM'
DEFAULT_TAG: '2.5.0-SNAPSHOT-EDMM'
run: |
echo "TAG=${{ github.event.inputs.tag || env.DEFAULT_TAG }}" >> $GITHUB_ENV
echo "IMAGE=${{ github.event.inputs.image || env.DEFAULT_IMAGE }}" >> $GITHUB_ENV
Expand Down
2 changes: 1 addition & 1 deletion docker/bigdl-k8s/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ ARG HADOOP_VERSION=2.7
ARG SPARK_HOME=/opt/spark
ARG JDK_VERSION=8u192
ARG JDK_URL=your_jdk_url
ARG BIGDL_VERSION=2.4.0-SNAPSHOT
ARG BIGDL_VERSION=2.5.0-SNAPSHOT
ARG TINI_VERSION=v0.18.0
ARG DEBIAN_FRONTEND=noninteractive

Expand Down
2 changes: 1 addition & 1 deletion docker/bigdl/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ ARG HADOOP_VERSION=2.7
ARG SPARK_HOME=/opt/spark
ARG JDK_VERSION=8u192
ARG JDK_URL=your_jdk_url
ARG BIGDL_VERSION=2.4.0-SNAPSHOT
ARG BIGDL_VERSION=2.5.0-SNAPSHOT
ARG DEBIAN_FRONTEND=noninteractive

# stage.1 jdk & spark
Expand Down
8 changes: 4 additions & 4 deletions docker/friesian-serving/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -16,13 +16,13 @@


ARG SPARK_VERSION=2.4.6
ARG HADOOP=2.7
ARG BIGDL_VERSION=2.4.0-SNAPSHOT
ARG HADOOP_VERSION=2.7
ARG BIGDL_VERSION=2.5.0-SNAPSHOT

# stage.1 spark
FROM ubuntu:20.04 as spark
ARG SPARK_VERSION
ARG HADOOP
ARG HADOOP_VERSION
ARG SPARK_HOME=/opt/spark

RUN apt-get update --fix-missing && \
Expand Down Expand Up @@ -52,7 +52,7 @@ RUN apt-get update --fix-missing && \
FROM ubuntu:20.04

ARG SPARK_VERSION=2.4.6
ARG BIGDL_VERSION=2.4.0-SNAPSHOT
ARG BIGDL_VERSION=2.5.0-SNAPSHOT

ENV SPARK_HOME /opt/work/spark-${SPARK_VERSION}
ENV LD_LIBRARY_PATH /opt/work/lib
Expand Down
34 changes: 17 additions & 17 deletions docker/llm/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -27,11 +27,11 @@ The instructions for installing can be accessed from

To pull image from hub, you can execute command on console:
```bash
docker pull intelanalytics/bigdl-llm-cpu:2.4.0-SNAPSHOT
docker pull intelanalytics/bigdl-llm-cpu:2.5.0-SNAPSHOT
```
to check if the image is successfully downloaded, you can use:
```powershell
docker images | sls intelanalytics/bigdl-llm-cpu:2.4.0-SNAPSHOT
docker images | sls intelanalytics/bigdl-llm-cpu:2.5.0-SNAPSHOT
```


Expand All @@ -42,7 +42,7 @@ To run the image and do inference, you could create and run a bat script on Wind
An example on Windows could be:
```bat
@echo off
set DOCKER_IMAGE=intelanalytics/bigdl-llm-cpu:2.4.0-SNAPSHOT
set DOCKER_IMAGE=intelanalytics/bigdl-llm-cpu:2.5.0-SNAPSHOT
set CONTAINER_NAME=my_container
set MODEL_PATH=D:/llm/models[change to your model path]
Expand Down Expand Up @@ -112,7 +112,7 @@ Here is a demostration of how to use tutorial in explorer:
To run container on Linux/MacOS:
```bash
#/bin/bash
export DOCKER_IMAGE=intelanalytics/bigdl-llm-cpu:2.4.0-SNAPSHOT
export DOCKER_IMAGE=intelanalytics/bigdl-llm-cpu:2.5.0-SNAPSHOT
export CONTAINER_NAME=my_container
export MODEL_PATH=/llm/models[change to your model path]

Expand All @@ -136,13 +136,13 @@ Also, you could use chat.py and bigdl-llm-tutorial for development.

First, pull docker image from docker hub:
```
docker pull intelanalytics/bigdl-llm-xpu:2.4.0-SNAPSHOT
docker pull intelanalytics/bigdl-llm-xpu:2.5.0-SNAPSHOT
```
To map the xpu into the container, you need to specify --device=/dev/dri when booting the container.
An example could be:
```bash
#/bin/bash
export DOCKER_IMAGE=intelanalytics/bigdl-llm-xpu:2.4.0-SNAPSHOT
export DOCKER_IMAGE=intelanalytics/bigdl-llm-xpu:2.5.0-SNAPSHOT
export CONTAINER_NAME=my_container
export MODEL_PATH=/llm/models[change to your model path]

Expand Down Expand Up @@ -176,12 +176,12 @@ To run inference using `BigDL-LLM` using xpu, you could refer to this [documenta

Pull image:
```
docker pull intelanalytics/bigdl-llm-serving-cpu:2.4.0-SNAPSHOT
docker pull intelanalytics/bigdl-llm-serving-cpu:2.5.0-SNAPSHOT
```

You could use the following bash script to start the container. Please be noted that the CPU config is specified for Xeon CPUs, change it accordingly if you are not using a Xeon CPU.
```bash
export DOCKER_IMAGE=intelanalytics/bigdl-llm-serving-cpu:2.4.0-SNAPSHOT
export DOCKER_IMAGE=intelanalytics/bigdl-llm-serving-cpu:2.5.0-SNAPSHOT
export CONTAINER_NAME=my_container
export MODEL_PATH=/llm/models[change to your model path]

Expand Down Expand Up @@ -268,15 +268,15 @@ python3 -m fastchat.serve.openai_api_server --host localhost --port 8000

Pull image:
```
docker pull intelanalytics/bigdl-llm-serving-xpu:2.4.0-SNAPSHOT
docker pull intelanalytics/bigdl-llm-serving-xpu:2.5.0-SNAPSHOT
```

To map the `xpu` into the container, you need to specify `--device=/dev/dri` when booting the container.

An example could be:
```bash
#/bin/bash
export DOCKER_IMAGE=intelanalytics/bigdl-llm-serving-cpu:2.4.0-SNAPSHOT
export DOCKER_IMAGE=intelanalytics/bigdl-llm-serving-cpu:2.5.0-SNAPSHOT
export CONTAINER_NAME=my_container
export MODEL_PATH=/llm/models[change to your model path]
export SERVICE_MODEL_PATH=/llm/models/chatglm2-6b[a specified model path for running service]
Expand Down Expand Up @@ -362,7 +362,7 @@ python3 -m fastchat.serve.openai_api_server --host localhost --port 8000
You can download directly from Dockerhub like:

```bash
docker pull intelanalytics/bigdl-llm-finetune-lora-cpu:2.4.0-SNAPSHOT
docker pull intelanalytics/bigdl-llm-finetune-lora-cpu:2.5.0-SNAPSHOT
```

Or build the image from source:
Expand All @@ -374,7 +374,7 @@ export HTTPS_PROXY=your_https_proxy
docker build \
--build-arg http_proxy=${HTTP_PROXY} \
--build-arg https_proxy=${HTTPS_PROXY} \
-t intelanalytics/bigdl-llm-finetune-lora-cpu:2.4.0-SNAPSHOT \
-t intelanalytics/bigdl-llm-finetune-lora-cpu:2.5.0-SNAPSHOT \
-f ./Dockerfile .
```

Expand All @@ -390,7 +390,7 @@ docker run -itd \
-e WORKER_COUNT_DOCKER=your_worker_count \
-v your_downloaded_base_model_path:/bigdl/model \
-v your_downloaded_data_path:/bigdl/data/alpaca_data_cleaned_archive.json \
intelanalytics/bigdl-llm-finetune-cpu:2.4.0-SNAPSHOT \
intelanalytics/bigdl-llm-finetune-cpu:2.5.0-SNAPSHOT \
bash
```

Expand Down Expand Up @@ -449,7 +449,7 @@ The following shows how to fine-tune LLM with Quantization (QLoRA built on BigDL
You can download directly from Dockerhub like:

```bash
docker pull intelanalytics/bigdl-llm-finetune-qlora-xpu:2.4.0-SNAPSHOT
docker pull intelanalytics/bigdl-llm-finetune-qlora-xpu:2.5.0-SNAPSHOT
```

Or build the image from source:
Expand All @@ -461,7 +461,7 @@ export HTTPS_PROXY=your_https_proxy
docker build \
--build-arg http_proxy=${HTTP_PROXY} \
--build-arg https_proxy=${HTTPS_PROXY} \
-t intelanalytics/bigdl-llm-finetune-qlora-xpu:2.4.0-SNAPSHOT \
-t intelanalytics/bigdl-llm-finetune-qlora-xpu:2.5.0-SNAPSHOT \
-f ./Dockerfile .
```

Expand All @@ -485,7 +485,7 @@ docker run -itd \
-v $BASE_MODE_PATH:/model \
-v $DATA_PATH:/data/english_quotes \
--shm-size="16g" \
intelanalytics/bigdl-llm-fintune-qlora-xpu:2.4.0-SNAPSHOT
intelanalytics/bigdl-llm-fintune-qlora-xpu:2.5.0-SNAPSHOT
```

The download and mount of base model and data to a docker container demonstrates a standard fine-tuning process. You can skip this step for a quick start, and in this way, the fine-tuning codes will automatically download the needed files:
Expand All @@ -502,7 +502,7 @@ docker run -itd \
-e http_proxy=${HTTP_PROXY} \
-e https_proxy=${HTTPS_PROXY} \
--shm-size="16g" \
intelanalytics/bigdl-llm-fintune-qlora-xpu:2.4.0-SNAPSHOT
intelanalytics/bigdl-llm-fintune-qlora-xpu:2.5.0-SNAPSHOT
```

However, we do recommend you to handle them manually, because the automatical download can be blocked by Internet access and Huggingface authentication etc. according to different environment, and the manual method allows you to fine-tune in a custom way (with different base model and dataset).
Expand Down
6 changes: 3 additions & 3 deletions docker/llm/finetune/lora/cpu/docker/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
You can download directly from Dockerhub like:

```bash
docker pull intelanalytics/bigdl-llm-finetune-lora-cpu:2.4.0-SNAPSHOT
docker pull intelanalytics/bigdl-llm-finetune-lora-cpu:2.5.0-SNAPSHOT
```

Or build the image from source:
Expand All @@ -17,7 +17,7 @@ export HTTPS_PROXY=your_https_proxy
docker build \
--build-arg http_proxy=${HTTP_PROXY} \
--build-arg https_proxy=${HTTPS_PROXY} \
-t intelanalytics/bigdl-llm-finetune-lora-cpu:2.4.0-SNAPSHOT \
-t intelanalytics/bigdl-llm-finetune-lora-cpu:2.5.0-SNAPSHOT \
-f ./Dockerfile .
```

Expand All @@ -33,7 +33,7 @@ docker run -itd \
-e WORKER_COUNT_DOCKER=your_worker_count \
-v your_downloaded_base_model_path:/bigdl/model \
-v your_downloaded_data_path:/bigdl/data/alpaca_data_cleaned_archive.json \
intelanalytics/bigdl-llm-finetune-lora-cpu:2.4.0-SNAPSHOT \
intelanalytics/bigdl-llm-finetune-lora-cpu:2.5.0-SNAPSHOT \
bash
```

Expand Down
2 changes: 1 addition & 1 deletion docker/llm/finetune/lora/cpu/kubernetes/values.yaml
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
imageName: intelanalytics/bigdl-llm-finetune-lora-cpu:2.4.0-SNAPSHOT
imageName: intelanalytics/bigdl-llm-finetune-lora-cpu:2.5.0-SNAPSHOT
trainerNum: 8
microBatchSize: 8
nfsServerIp: your_nfs_server_ip
Expand Down
Loading

0 comments on commit ee8ce18

Please sign in to comment.