|  | --- | 
					
						
						|  | datasets: | 
					
						
						|  | - WizardLM/WizardLM_evol_instruct_V2_196k | 
					
						
						|  | - Open-Orca/OpenOrca | 
					
						
						|  | language: | 
					
						
						|  | - en | 
					
						
						|  | tags: | 
					
						
						|  | - chat | 
					
						
						|  | - palmyra | 
					
						
						|  | license: apache-2.0 | 
					
						
						|  | --- | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | **DEPRECATED MODEL NOTICE** | 
					
						
						|  | ========================== | 
					
						
						|  |  | 
					
						
						|  | Please note that this model is no longer maintained or supported by our team. We strongly advise against using it in production or for any critical applications. | 
					
						
						|  |  | 
					
						
						|  | Instead, we recommend using our latest and greatest models, which can be found at: | 
					
						
						|  |  | 
					
						
						|  | https://huggingface.co/collections/Writer/palmyra-writer-license-66476fa8156169f8720a2c89 | 
					
						
						|  |  | 
					
						
						|  | ========================== | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | # Writer/palmyra-20b-chat | 
					
						
						|  | --- | 
					
						
						|  |  | 
					
						
						|  | # Usage | 
					
						
						|  |  | 
					
						
						|  | ```py | 
					
						
						|  |  | 
					
						
						|  | import torch | 
					
						
						|  | from transformers import AutoTokenizer, AutoModelForCausalLM, TextStreamer | 
					
						
						|  |  | 
					
						
						|  | model_name = "Writer/palmyra-20b-chat" | 
					
						
						|  |  | 
					
						
						|  | tokenizer = AutoTokenizer.from_pretrained(model_name) | 
					
						
						|  |  | 
					
						
						|  | model = AutoModelForCausalLM.from_pretrained( | 
					
						
						|  | model_name, | 
					
						
						|  | torch_dtype=torch.float16, | 
					
						
						|  | device_map="auto", | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  | prompt = "What is the meaning of life?" | 
					
						
						|  |  | 
					
						
						|  | input_text = ( | 
					
						
						|  | "A chat between a curious user and an artificial intelligence assistant. " | 
					
						
						|  | "The assistant gives helpful, detailed, and polite answers to the user's questions. " | 
					
						
						|  | "USER: {prompt} " | 
					
						
						|  | "ASSISTANT:" | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  | model_inputs = tokenizer(input_text.format(prompt=prompt), return_tensors="pt").to( | 
					
						
						|  | "cuda" | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  | gen_conf = { | 
					
						
						|  | "top_k": 20, | 
					
						
						|  | "max_new_tokens": 2048, | 
					
						
						|  | "temperature": 0.6, | 
					
						
						|  | "do_sample": True, | 
					
						
						|  | "eos_token_id": tokenizer.eos_token_id, | 
					
						
						|  | } | 
					
						
						|  |  | 
					
						
						|  | streamer = TextStreamer(tokenizer) | 
					
						
						|  | if "token_type_ids" in model_inputs: | 
					
						
						|  | del model_inputs["token_type_ids"] | 
					
						
						|  |  | 
					
						
						|  | all_inputs = {**model_inputs, **gen_conf} | 
					
						
						|  | output = model.generate(**all_inputs, streamer=streamer) | 
					
						
						|  |  | 
					
						
						|  | print("-"*20) | 
					
						
						|  | print(output) | 
					
						
						|  |  | 
					
						
						|  | ``` | 
					
						
						|  | # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) | 
					
						
						|  | Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Writer__palmyra-20b-chat) | 
					
						
						|  |  | 
					
						
						|  | | Metric                | Value                     | | 
					
						
						|  | |-----------------------|---------------------------| | 
					
						
						|  | | Avg.                  | 38.97   | | 
					
						
						|  | | ARC (25-shot)         | 43.52          | | 
					
						
						|  | | HellaSwag (10-shot)   | 72.83    | | 
					
						
						|  | | MMLU (5-shot)         | 35.18         | | 
					
						
						|  | | TruthfulQA (0-shot)   | 43.17   | | 
					
						
						|  | | Winogrande (5-shot)   | 66.46   | | 
					
						
						|  | | GSM8K (5-shot)        | 3.94        | | 
					
						
						|  | | DROP (3-shot)         | 7.7         | |