|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | try: | 
					
						
						|  | import os | 
					
						
						|  | from openai import OpenAI | 
					
						
						|  |  | 
					
						
						|  | client = OpenAI( | 
					
						
						|  | base_url="https://router.huggingface.co/v1", | 
					
						
						|  | api_key=os.environ["HF_TOKEN"], | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  | completion = client.chat.completions.create( | 
					
						
						|  | model="Qwen/Qwen3-0.6B", | 
					
						
						|  | messages=[ | 
					
						
						|  | { | 
					
						
						|  | "role": "user", | 
					
						
						|  | "content": "What is the capital of France?" | 
					
						
						|  | } | 
					
						
						|  | ], | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  | print(completion.choices[0].message) | 
					
						
						|  | with open('Qwen_Qwen3-0.6B_3.txt', 'w', encoding='utf-8') as f: | 
					
						
						|  | f.write('Everything was good in Qwen_Qwen3-0.6B_3.txt') | 
					
						
						|  | except Exception as e: | 
					
						
						|  | import os | 
					
						
						|  | from slack_sdk import WebClient | 
					
						
						|  | client = WebClient(token=os.environ['SLACK_TOKEN']) | 
					
						
						|  | client.chat_postMessage( | 
					
						
						|  | channel='#exp-slack-alerts', | 
					
						
						|  | text='Problem in <https://huggingface.co/datasets/model-metadata/code_execution_files/blob/main/Qwen_Qwen3-0.6B_3.txt|Qwen_Qwen3-0.6B_3.txt>', | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  | with open('Qwen_Qwen3-0.6B_3.txt', 'a', encoding='utf-8') as f: | 
					
						
						|  | import traceback | 
					
						
						|  | f.write('''```CODE: | 
					
						
						|  | import os | 
					
						
						|  | from openai import OpenAI | 
					
						
						|  |  | 
					
						
						|  | client = OpenAI( | 
					
						
						|  | base_url="https://router.huggingface.co/v1", | 
					
						
						|  | api_key=os.environ["HF_TOKEN"], | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  | completion = client.chat.completions.create( | 
					
						
						|  | model="Qwen/Qwen3-0.6B", | 
					
						
						|  | messages=[ | 
					
						
						|  | { | 
					
						
						|  | "role": "user", | 
					
						
						|  | "content": "What is the capital of France?" | 
					
						
						|  | } | 
					
						
						|  | ], | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  | print(completion.choices[0].message) | 
					
						
						|  | ``` | 
					
						
						|  |  | 
					
						
						|  | ERROR: | 
					
						
						|  | ''') | 
					
						
						|  | traceback.print_exc(file=f) | 
					
						
						|  |  | 
					
						
						|  | finally: | 
					
						
						|  | from huggingface_hub import upload_file | 
					
						
						|  | upload_file( | 
					
						
						|  | path_or_fileobj='Qwen_Qwen3-0.6B_3.txt', | 
					
						
						|  | repo_id='model-metadata/code_execution_files', | 
					
						
						|  | path_in_repo='Qwen_Qwen3-0.6B_3.txt', | 
					
						
						|  | repo_type='dataset', | 
					
						
						|  | ) | 
					
						
						|  |  |