Skip to content

Commit

Permalink
Merge pull request microsoft#1 from microsoft/dolly-hf
Browse files Browse the repository at this point in the history
Updated to point to Dolly 2 optimized version on HuggingFace
  • Loading branch information
jennifermarsman authored May 23, 2023
2 parents ebb8057 + ee83f11 commit 230e156
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 3 deletions.
5 changes: 3 additions & 2 deletions PodcastSocialMediaCopilot.py
Original file line number Diff line number Diff line change
Expand Up @@ -102,8 +102,9 @@

# Step 2 - Make a call to a local Dolly 2.0 model optimized for Windows to extract the name of who I'm interviewing from the transcript
print("Calling a local Dolly 2.0 model optimized for Windows to extract the name of the podcast guest...\n")
tokenizer = AutoTokenizer.from_pretrained("models/dolly_v2_7b_merged_fp16.0.0.1", padding_side="left")
model = ORTModelForCausalLM.from_pretrained("models/dolly_v2_7b_merged_fp16.0.0.1", provider="DmlExecutionProvider", use_cache=True, use_merged=True, use_io_binding=False)
repo_id = "microsoft/dolly-v2-7b-olive-optimized"
tokenizer = AutoTokenizer.from_pretrained(repo_id, padding_side="left")
model = ORTModelForCausalLM.from_pretrained(repo_id, provider="DmlExecutionProvider", use_cache=True, use_merged=True, use_io_binding=False)
streamer = TextStreamer(tokenizer, skip_prompt=True)
generate_text = InstructionTextGenerationPipeline(model=model, streamer=streamer, tokenizer=tokenizer, max_new_tokens=128, return_full_text=True, task="text-generation")
hf_pipeline = HuggingFacePipeline(pipeline=generate_text)
Expand Down
1 change: 0 additions & 1 deletion requirements.txt
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,6 @@ torch
transformers
optimum
onnx
wheels/dolly_v2_wheel.0.0.4/onnxruntime_directml-1.16.0-cp39-cp39-win_amd64.whl
openai-whisper
langchain
pydub
Expand Down

0 comments on commit 230e156

Please sign in to comment.