-
Notifications
You must be signed in to change notification settings - Fork 211
/
prepare_llm_models.sh
executable file
·55 lines (49 loc) · 2.4 KB
/
prepare_llm_models.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
#!/bin/bash
#
# Copyright 2024 Intel Corporation
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
EMBEDDING_MODEL="thenlper/gte-small"
if [ -d "$1/facebook/opt-125m" ] && [ -d "$1/$EMBEDDING_MODEL" ]; then
echo "Models directory $1 exists. Skipping downloading models."
exit 0
fi
if [ "$(python3 -c 'import sys; print(sys.version_info[1])')" -le "8" ]; then echo "Prepare models with python > 3.8."; exit 1 ; fi
echo "Downloading LLM testing models to directory $1"
export PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu"
if [ "$2" = "docker" ]; then
sed -i '/openvino~=/d' /openvino_tokenizers/pyproject.toml
python3 -m pip wheel -v --no-deps --wheel-dir wheel /openvino_tokenizers
python3 -m pip install $(find wheel -name 'openvino_tokenizers*.whl')
python3 -m pip install "optimum-intel"@git+https://github.com/huggingface/optimum-intel.git nncf
else
python3 -m venv .venv
. .venv/bin/activate
pip3 install -U pip
pip3 install --pre "optimum-intel[nncf,openvino]"@git+https://github.com/huggingface/optimum-intel.git openvino-tokenizers
fi
if [ -d "$1/facebook/opt-125m" ]; then
echo "Models directory $1/facebook/opt-125m exists. Skipping downloading models."
else
optimum-cli export openvino --disable-convert-tokenizer --model facebook/opt-125m --weight-format int8 $1/facebook/opt-125m
convert_tokenizer -o $1/facebook/opt-125m --with-detokenizer --skip-special-tokens --streaming-detokenizer --not-add-special-tokens facebook/opt-125m
fi
if [ -d "$1/$EMBEDDING_MODEL" ]; then
echo "Models directory $1/$EMBEDDING_MODEL exists. Skipping downloading models."
else
optimum-cli export openvino --model "$EMBEDDING_MODEL" --task feature-extraction "$1/$EMBEDDING_MODEL/embeddings/1"
convert_tokenizer -o "$1/$EMBEDDING_MODEL/tokenizer/1" "$EMBEDDING_MODEL"
rm "$1/$EMBEDDING_MODEL/embeddings/1/openvino_tokenizer.xml"
rm "$1/$EMBEDDING_MODEL/embeddings/1/openvino_tokenizer.bin"
fi