From 3c4e47b2893741c610f01849cc9b353a3c994bed Mon Sep 17 00:00:00 2001 From: lvliang-intel Date: Thu, 2 Nov 2023 16:29:21 +0800 Subject: [PATCH] Fix examples path change issue (#613) Signed-off-by: Lv, Liang1 Co-authored-by: kevinintel --- README.md | 2 +- .../neural_chat/docker/Dockerfile | 6 +++--- .../neural_chat/docker/finetuning/Dockerfile | 8 ++++---- .../notebooks/setup_talking_chatbot_service_on_spr.ipynb | 2 +- .../notebooks/setup_text_chatbot_service_on_spr.ipynb | 2 +- .../neural_chat/pipeline/plugins/retrieval/README.md | 2 +- 6 files changed, 11 insertions(+), 11 deletions(-) diff --git a/README.md b/README.md index 901337390c9..b5dd35ff1a3 100644 --- a/README.md +++ b/README.md @@ -114,7 +114,7 @@ Additionally, we are preparing to introduce Baichuan, Mistral, and other models Chatbot on Gaudi - Chatbot on Client + Chatbot on Client More Notebooks diff --git a/intel_extension_for_transformers/neural_chat/docker/Dockerfile b/intel_extension_for_transformers/neural_chat/docker/Dockerfile index fed3fce8351..cd027136517 100644 --- a/intel_extension_for_transformers/neural_chat/docker/Dockerfile +++ b/intel_extension_for_transformers/neural_chat/docker/Dockerfile @@ -75,7 +75,7 @@ RUN if [ -f /torchvision-0.16.0+cpu-cp3${PYTHON_VERSION##*.}-cp3${PYTHON_VERSION RUN source activate && conda activate neuralchat && pip install oneccl_bind_pt==2.0.0+cpu -f https://developer.intel.com/ipex-whl-stable-cpu && \ cd /intel-extension-for-transformers && pip install -r requirements.txt && pip install -v . && \ - cd ./intel_extension_for_transformers/neural_chat/examples/instruction_tuning && pip install -r requirements.txt && \ + cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install -r requirements.txt && \ cd /intel-extension-for-transformers/intel_extension_for_transformers/neural_chat && pip install -r requirements_cpu.txt && \ conda install astunparse ninja pyyaml mkl mkl-include setuptools cmake cffi future six requests dataclasses -y && \ conda install jemalloc gperftools -c conda-forge -y && \ @@ -133,7 +133,7 @@ RUN git clone https://github.com/huggingface/optimum-habana.git && \ pip install -e . # Install dependency -RUN cd /intel-extension-for-transformers/intel_extension_for_transformers/neural_chat/examples/instruction_tuning/ && \ +RUN cd /intel-extension-for-transformers/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && \ sed -i '/find-links https:\/\/download.pytorch.org\/whl\/torch_stable.html/d' requirements.txt && \ sed -i '/^torch/d;/^intel-extension-for-pytorch/d' requirements.txt && \ pip install -r requirements.txt && \ @@ -184,7 +184,7 @@ RUN conda init bash && \ source ~/.bashrc RUN source activate && conda activate neuralchat && cd /intel-extension-for-transformers && \ - pip install -r ./intel_extension_for_transformers/neural_chat/examples/instruction_tuning/requirements.txt && \ + pip install -r ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/requirements.txt && \ pip install -r ./intel_extension_for_transformers/neural_chat/requirements.txt WORKDIR /intel-extension-for-transformers/intel_extension_for_transformers/neural_chat/ diff --git a/intel_extension_for_transformers/neural_chat/docker/finetuning/Dockerfile b/intel_extension_for_transformers/neural_chat/docker/finetuning/Dockerfile index cdfd8f571d2..b893de129a7 100644 --- a/intel_extension_for_transformers/neural_chat/docker/finetuning/Dockerfile +++ b/intel_extension_for_transformers/neural_chat/docker/finetuning/Dockerfile @@ -67,7 +67,7 @@ RUN source activate && conda activate chatbot-finetuning && pip install oneccl_b pip install datasets torch accelerate SentencePiece evaluate nltk rouge_score protobuf==3.20.1 tokenizers einops && \ git clone https://github.com/huggingface/peft.git && cd peft && python setup.py install && \ cd /itrex && pip install -v . && \ - cd ./intel_extension_for_transformers/neural_chat/examples/instruction_tuning && pip install -r requirements.txt + cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install -r requirements.txt # Enable passwordless ssh for mpirun^M RUN mkdir /var/run/sshd @@ -79,7 +79,7 @@ RUN sed -i'' -e's/^#PermitRootLogin prohibit-password$/PermitRootLogin yes/' /et && echo "StrictHostKeyChecking no" >> /etc/ssh/ssh_config EXPOSE 22 -WORKDIR /itrex/neural_chat/examples/instruction_tuning/ +WORKDIR /itrex/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/ CMD ["/usr/sbin/sshd", "-D"] @@ -109,7 +109,7 @@ ARG ITREX_VER=main ARG REPO=https://github.com/intel/intel-extension-for-transformers.git RUN git clone --single-branch --branch=${ITREX_VER} ${REPO} itrex && \ - cd /itrex/intel_extension_for_transformers/neural_chat/examples/instruction_tuning/ && \ + cd /itrex/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/ && \ pip install -r requirements.txt # Build ITREX @@ -117,4 +117,4 @@ RUN cd /itrex && pip install -v . && \ pip install transformers==4.32.0 && \ pip install accelerate==0.22.0 -WORKDIR /itrex/neural_chat/examples/instruction_tuning +WORKDIR /itrex/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/ diff --git a/intel_extension_for_transformers/neural_chat/docs/notebooks/setup_talking_chatbot_service_on_spr.ipynb b/intel_extension_for_transformers/neural_chat/docs/notebooks/setup_talking_chatbot_service_on_spr.ipynb index e6ba93151d0..c0b1cf95dac 100644 --- a/intel_extension_for_transformers/neural_chat/docs/notebooks/setup_talking_chatbot_service_on_spr.ipynb +++ b/intel_extension_for_transformers/neural_chat/docs/notebooks/setup_talking_chatbot_service_on_spr.ipynb @@ -67,7 +67,7 @@ "metadata": {}, "outputs": [], "source": [ - "!curl -OL https://raw.githubusercontent.com/intel/intel-extension-for-transformers/main/intel_extension_for_transformers/neural_chat/examples/talkingbot/backend/talkingbot.yaml" + "!curl -OL https://raw.githubusercontent.com/intel/intel-extension-for-transformers/main/intel_extension_for_transformers/neural_chat/examples/deployment/textbot/backend/textbot.yaml" ] }, { diff --git a/intel_extension_for_transformers/neural_chat/docs/notebooks/setup_text_chatbot_service_on_spr.ipynb b/intel_extension_for_transformers/neural_chat/docs/notebooks/setup_text_chatbot_service_on_spr.ipynb index a7d3f31db2e..a1d9b93f1c5 100644 --- a/intel_extension_for_transformers/neural_chat/docs/notebooks/setup_text_chatbot_service_on_spr.ipynb +++ b/intel_extension_for_transformers/neural_chat/docs/notebooks/setup_text_chatbot_service_on_spr.ipynb @@ -56,7 +56,7 @@ "metadata": {}, "outputs": [], "source": [ - "!curl -OL https://raw.githubusercontent.com/intel/intel-extension-for-transformers/main/intel_extension_for_transformers/neural_chat/examples/textbot/backend/textbot.yaml" + "!curl -OL https://raw.githubusercontent.com/intel/intel-extension-for-transformers/main/intel_extension_for_transformers/neural_chat/examples/deployment/textbot/backend/textbot.yaml" ] }, { diff --git a/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/README.md b/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/README.md index 2d0bd5bccff..134c2835699 100644 --- a/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/README.md +++ b/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/README.md @@ -28,7 +28,7 @@ In addition, our plugin seamlessly integrates the online embedding model, Google The workflow of this plugin consists of three main operations: document indexing, intent detection, and retrieval. The `Agent_QA` initializes itself using the provided `input_path` to construct a local database. During a conversation, the user's query is first passed to the `IntentDetector` to determine whether the user intends to engage in chitchat or seek answers to specific questions. If the `IntentDetector` determines that the user's query requires an answer, the retriever is activated to search the database using the user's query. The documents retrieved from the database serve as reference context in the input prompt, assisting in generating responses using the Large Language Models (LLMs). # Usage -The most convenient way to use is this plugin is via our `build_chatbot` api as introduced in the [example code](https://github.com/intel/intel-extension-for-transformers/tree/main/intel_extension_for_transformers/neural_chat/examples/retrieval). The user could refer to it for a simple test. +The most convenient way to use is this plugin is via our `build_chatbot` api as introduced in the [example code](https://github.com/intel/intel-extension-for-transformers/tree/main/intel_extension_for_transformers/neural_chat/examples/plugins/retrieval). The user could refer to it for a simple test. We support multiple file formats for retrieval, including unstructured file formats such as pdf, docx, html, txt, and markdown, as well as structured file formats like jsonl and xlsx. For structured file formats, they must adhere to predefined structures.