Skip to content

Commit

Permalink
Tryied mistral as well
Browse files Browse the repository at this point in the history
  • Loading branch information
ilya-lavrenov committed May 6, 2024
1 parent cb13f2f commit e4700f4
Show file tree
Hide file tree
Showing 2 changed files with 10 additions and 9 deletions.
Original file line number Diff line number Diff line change
@@ -1,5 +1,9 @@
--extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly/
openvino
# we need at least openvino 2024.2
openvino >= 2024.2
openvino-tokenizers
# use latest released version once it's available
git+https://github.com/huggingface/optimum-intel.git@main
pytest
pytest
# set 'export HF_HUB_ENABLE_HF_TRANSFER=1' to benefits from hf_transfer
hf_transfer
Original file line number Diff line number Diff line change
Expand Up @@ -129,17 +129,14 @@ def run_continuous_batching(
pipe = ContinuousBatchingPipeline(model_path.absolute().as_posix(), scheduler_config)
return pipe.generate(prompts, generation_configs)

# export models via
# optimum-cli export openvino -m meta-llama/Llama-2-7b-chat-hf llama2
# optimum-cli export openvino -m meta-llama/Llama-2-7b-chat-hf --fp16 llama2-fp16

# tested models:
# - facebook/opt-125m (opt125)
# - meta-llama/Llama-2-7b-chat-hf (llama2 or llama2-fp16)
# - facebook/opt-125m
# - meta-llama/Llama-2-7b-chat-hf
# - mistralai/Mistral-7B-Instruct-v0.2

def test_check_greedy_search(tmp_path):
prompts, generation_configs = get_test_dataset()
model_id : str = "meta-llama/Llama-2-7b-chat-hf"
model_id : str = "facebook/opt-125m"

(hf_results, model_path) = run_hugging_face(model_id=model_id, prompts=prompts, generation_configs=generation_configs, tmp_path=tmp_path, use_optimum=True)
my_results : List[GenerationResult] = run_continuous_batching(model_path, get_scheduler_config(), prompts, generation_configs)
Expand Down

0 comments on commit e4700f4

Please sign in to comment.