diff --git a/.github/workflows/linux.yml b/.github/workflows/linux.yml index 0d7a5b7bae..8fc678fd85 100644 --- a/.github/workflows/linux.yml +++ b/.github/workflows/linux.yml @@ -374,7 +374,7 @@ jobs: - name: Test multinomial_causal_lm.py if: ${{ 'Release' == matrix.build-type }} # Python bindings can be built in Release only timeout-minutes: 1 - run: ${{ env.INSTALL_DIR }}/samples/python/multinomial_causal_lm/multinomial_causal_lm.py ./TinyLlama-1.1B-Chat-v1.0/ 0 + run: ${{ env.INSTALL_DIR }}/samples/python/text_generation/multinomial_causal_lm.py ./TinyLlama-1.1B-Chat-v1.0/ 0 working-directory: ${{ env.MODELS_DIR }} - name: Test whisper_speech_recognition.py diff --git a/.github/workflows/mac.yml b/.github/workflows/mac.yml index 062b83fc27..3662e9c72a 100644 --- a/.github/workflows/mac.yml +++ b/.github/workflows/mac.yml @@ -395,7 +395,7 @@ jobs: if: ${{ 'Release' == matrix.build-type }} # Python bindings can be built in Release only run: | source ${OV_INSTALL_DIR}/setupvars.sh - ${OV_INSTALL_DIR}/samples/python/multinomial_causal_lm/multinomial_causal_lm.py ./TinyLlama-1.1B-Chat-v1.0/ 0 + ${OV_INSTALL_DIR}/samples/python/text_generation/multinomial_causal_lm.py ./TinyLlama-1.1B-Chat-v1.0/ 0 timeout-minutes: 1 - name: Test python samples (whisper_speech_recognition) diff --git a/src/README.md b/src/README.md index 028277d4db..5d18d0b67b 100644 --- a/src/README.md +++ b/src/README.md @@ -231,7 +231,7 @@ custom_streamer = CustomStreamer() pipe.generate("The Sun is yellow because", max_new_tokens=15, streamer=custom_streamer) ``` -For fully implemented iterable CustomStreamer please refer to [multinomial_causal_lm](https://github.com/openvinotoolkit/openvino.genai/tree/releases/2024/3/samples/python/multinomial_causal_lm/README.md) sample. +For fully implemented iterable CustomStreamer please refer to [multinomial_causal_lm](https://github.com/openvinotoolkit/openvino.genai/tree/releases/2024/3/samples/python/text_generation/README.md) sample. Continuous batching with LLMPipeline: