From b9dac764ac7a5e92fac157e3a9c8ed7116fcbe7c Mon Sep 17 00:00:00 2001 From: Pamela Fox Date: Fri, 14 Mar 2025 17:44:48 +0000 Subject: [PATCH 1/2] Default to GitHub models --- chained_calls.py | 4 ++-- chat.py | 4 ++-- chat_async.py | 4 ++-- chat_history.py | 4 ++-- chat_history_stream.py | 4 ++-- chat_langchain.py | 4 ++-- chat_llamaindex.py | 4 ++-- chat_pydanticai.py | 4 ++-- chat_safety.py | 4 ++-- chat_stream.py | 4 ++-- few_shot_examples.py | 4 ++-- function_calling.py | 4 ++-- function_calling_call.py | 4 ++-- function_calling_multiple.py | 4 ++-- prompt_engineering.py | 4 ++-- rag_csv.py | 4 ++-- rag_documents_flow.py | 4 ++-- rag_documents_hybrid.py | 4 ++-- rag_documents_ingestion.py | 4 ++-- rag_multiturn.py | 4 ++-- rag_queryrewrite.py | 4 ++-- retrieval_augmented_generation.py | 4 ++-- spanish/chained_calls.py | 4 ++-- spanish/chat.py | 4 ++-- spanish/chat_async.py | 4 ++-- spanish/chat_history.py | 4 ++-- spanish/chat_history_stream.py | 4 ++-- spanish/chat_langchain.py | 4 ++-- spanish/chat_llamaindex.py | 4 ++-- spanish/chat_pydanticai.py | 4 ++-- spanish/chat_safety.py | 4 ++-- spanish/chat_stream.py | 4 ++-- spanish/few_shot_examples.py | 4 ++-- spanish/function_calling.py | 4 ++-- spanish/function_calling_call.py | 4 ++-- spanish/function_calling_multiple.py | 4 ++-- spanish/prompt_engineering.py | 4 ++-- spanish/rag_csv.py | 4 ++-- spanish/rag_documents_flow.py | 4 ++-- spanish/rag_documents_hybrid.py | 4 ++-- spanish/rag_documents_ingestion.py | 4 ++-- spanish/rag_multiturn.py | 4 ++-- spanish/rag_queryrewrite.py | 4 ++-- spanish/retrieval_augmented_generation.py | 4 ++-- 44 files changed, 88 insertions(+), 88 deletions(-) diff --git a/chained_calls.py b/chained_calls.py index 9efc593..fb8c6f5 100644 --- a/chained_calls.py +++ b/chained_calls.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/chat.py b/chat.py index b78aaa1..b19a251 100644 --- a/chat.py +++ b/chat.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/chat_async.py b/chat_async.py index 3868c5c..31d0fe2 100644 --- a/chat_async.py +++ b/chat_async.py @@ -7,7 +7,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -24,7 +24,7 @@ MODEL_NAME = os.environ["OLLAMA_MODEL"] elif API_HOST == "github": client = openai.AsyncOpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.AsyncOpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] diff --git a/chat_history.py b/chat_history.py index f90845d..083e36a 100644 --- a/chat_history.py +++ b/chat_history.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -23,7 +23,7 @@ MODEL_NAME = os.environ["OLLAMA_MODEL"] elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] diff --git a/chat_history_stream.py b/chat_history_stream.py index ee8dd35..e12e3bc 100644 --- a/chat_history_stream.py +++ b/chat_history_stream.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -23,7 +23,7 @@ MODEL_NAME = os.environ["OLLAMA_MODEL"] elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] diff --git a/chat_langchain.py b/chat_langchain.py index 832ca1a..c0d676a 100644 --- a/chat_langchain.py +++ b/chat_langchain.py @@ -7,7 +7,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -27,7 +27,7 @@ ) elif API_HOST == "github": llm = ChatOpenAI( - model_name=os.environ["GITHUB_MODEL"], + model_name=os.getenv("GITHUB_MODEL", "gpt-4o"), openai_api_base="https://models.inference.ai.azure.com", openai_api_key=os.environ["GITHUB_TOKEN"], ) diff --git a/chat_llamaindex.py b/chat_llamaindex.py index fdbc39e..6dc3d61 100644 --- a/chat_llamaindex.py +++ b/chat_llamaindex.py @@ -9,7 +9,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -29,7 +29,7 @@ ) elif API_HOST == "github": llm = OpenAILike( - model=os.environ["GITHUB_MODEL"], + model=os.getenv("GITHUB_MODEL", "gpt-4o"), api_base="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"], is_chat_model=True, diff --git a/chat_pydanticai.py b/chat_pydanticai.py index 74f66ec..b289a8d 100644 --- a/chat_pydanticai.py +++ b/chat_pydanticai.py @@ -8,7 +8,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -24,7 +24,7 @@ model = OpenAIModel(os.environ["OLLAMA_MODEL"], api_key="fake", base_url=os.environ["OLLAMA_ENDPOINT"]) elif API_HOST == "github": model = OpenAIModel( - os.environ["GITHUB_MODEL"], api_key=os.environ["GITHUB_TOKEN"], base_url="https://models.inference.ai.azure.com" + os.getenv("GITHUB_MODEL", "gpt-4o"), api_key=os.environ["GITHUB_TOKEN"], base_url="https://models.inference.ai.azure.com" ) else: diff --git a/chat_safety.py b/chat_safety.py index b4f7f79..d6c3d73 100644 --- a/chat_safety.py +++ b/chat_safety.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -23,7 +23,7 @@ MODEL_NAME = os.environ["OLLAMA_MODEL"] elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] diff --git a/chat_stream.py b/chat_stream.py index d36c1ad..80aaa20 100644 --- a/chat_stream.py +++ b/chat_stream.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -23,7 +23,7 @@ MODEL_NAME = os.environ["OLLAMA_MODEL"] elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] diff --git a/few_shot_examples.py b/few_shot_examples.py index 00a124e..665ee64 100644 --- a/few_shot_examples.py +++ b/few_shot_examples.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/function_calling.py b/function_calling.py index 1da1958..4d29be0 100644 --- a/function_calling.py +++ b/function_calling.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/function_calling_call.py b/function_calling_call.py index d67f9e6..1126e97 100644 --- a/function_calling_call.py +++ b/function_calling_call.py @@ -7,7 +7,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -26,7 +26,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/function_calling_multiple.py b/function_calling_multiple.py index bd98a3e..2c44ea2 100644 --- a/function_calling_multiple.py +++ b/function_calling_multiple.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/prompt_engineering.py b/prompt_engineering.py index 33428e7..529f1e9 100644 --- a/prompt_engineering.py +++ b/prompt_engineering.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/rag_csv.py b/rag_csv.py index 0ca1fc1..5d637c7 100644 --- a/rag_csv.py +++ b/rag_csv.py @@ -8,7 +8,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -27,7 +27,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/rag_documents_flow.py b/rag_documents_flow.py index d545d58..f408984 100644 --- a/rag_documents_flow.py +++ b/rag_documents_flow.py @@ -8,7 +8,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -27,7 +27,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/rag_documents_hybrid.py b/rag_documents_hybrid.py index 3a3e0b5..3f74f04 100644 --- a/rag_documents_hybrid.py +++ b/rag_documents_hybrid.py @@ -10,7 +10,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -29,7 +29,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/rag_documents_ingestion.py b/rag_documents_ingestion.py index edd7d13..a260b18 100644 --- a/rag_documents_ingestion.py +++ b/rag_documents_ingestion.py @@ -10,7 +10,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -29,7 +29,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/rag_multiturn.py b/rag_multiturn.py index 451d9e1..641ad08 100644 --- a/rag_multiturn.py +++ b/rag_multiturn.py @@ -8,7 +8,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -27,7 +27,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/rag_queryrewrite.py b/rag_queryrewrite.py index d615083..73f8ac9 100644 --- a/rag_queryrewrite.py +++ b/rag_queryrewrite.py @@ -8,7 +8,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -27,7 +27,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/retrieval_augmented_generation.py b/retrieval_augmented_generation.py index 5b8a437..49f6572 100644 --- a/retrieval_augmented_generation.py +++ b/retrieval_augmented_generation.py @@ -7,7 +7,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -26,7 +26,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/spanish/chained_calls.py b/spanish/chained_calls.py index 41570ec..145cafc 100644 --- a/spanish/chained_calls.py +++ b/spanish/chained_calls.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/spanish/chat.py b/spanish/chat.py index 11d2004..a9bda16 100644 --- a/spanish/chat.py +++ b/spanish/chat.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/spanish/chat_async.py b/spanish/chat_async.py index 64b4768..f42b4c9 100644 --- a/spanish/chat_async.py +++ b/spanish/chat_async.py @@ -7,7 +7,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -24,7 +24,7 @@ MODEL_NAME = os.environ["OLLAMA_MODEL"] elif API_HOST == "github": client = openai.AsyncOpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.AsyncOpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] diff --git a/spanish/chat_history.py b/spanish/chat_history.py index c60db70..610fb79 100644 --- a/spanish/chat_history.py +++ b/spanish/chat_history.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -23,7 +23,7 @@ MODEL_NAME = os.environ["OLLAMA_MODEL"] elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] diff --git a/spanish/chat_history_stream.py b/spanish/chat_history_stream.py index 54ca12e..36553bf 100644 --- a/spanish/chat_history_stream.py +++ b/spanish/chat_history_stream.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -23,7 +23,7 @@ MODEL_NAME = os.environ["OLLAMA_MODEL"] elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] diff --git a/spanish/chat_langchain.py b/spanish/chat_langchain.py index 1167efc..e5e4aed 100644 --- a/spanish/chat_langchain.py +++ b/spanish/chat_langchain.py @@ -7,7 +7,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -27,7 +27,7 @@ ) elif API_HOST == "github": llm = ChatOpenAI( - model_name=os.environ["GITHUB_MODEL"], + model_name=os.getenv("GITHUB_MODEL", "gpt-4o"), openai_api_base="https://models.inference.ai.azure.com", openai_api_key=os.environ["GITHUB_TOKEN"], ) diff --git a/spanish/chat_llamaindex.py b/spanish/chat_llamaindex.py index 9eb1180..e1e41fc 100644 --- a/spanish/chat_llamaindex.py +++ b/spanish/chat_llamaindex.py @@ -9,7 +9,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -29,7 +29,7 @@ ) elif API_HOST == "github": llm = OpenAILike( - model=os.environ["GITHUB_MODEL"], + model=os.getenv("GITHUB_MODEL", "gpt-4o"), api_base="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"], is_chat_model=True, diff --git a/spanish/chat_pydanticai.py b/spanish/chat_pydanticai.py index 15ec7f9..5778281 100644 --- a/spanish/chat_pydanticai.py +++ b/spanish/chat_pydanticai.py @@ -8,7 +8,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -24,7 +24,7 @@ model = OpenAIModel(os.environ["OLLAMA_MODEL"], api_key="fake", base_url=os.environ["OLLAMA_ENDPOINT"]) elif API_HOST == "github": model = OpenAIModel( - os.environ["GITHUB_MODEL"], api_key=os.environ["GITHUB_TOKEN"], base_url="https://models.inference.ai.azure.com" + os.getenv("GITHUB_MODEL", "gpt-4o"), api_key=os.environ["GITHUB_TOKEN"], base_url="https://models.inference.ai.azure.com" ) else: diff --git a/spanish/chat_safety.py b/spanish/chat_safety.py index 070a113..86c3bb9 100644 --- a/spanish/chat_safety.py +++ b/spanish/chat_safety.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -23,7 +23,7 @@ MODEL_NAME = os.environ["OLLAMA_MODEL"] elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] diff --git a/spanish/chat_stream.py b/spanish/chat_stream.py index d2b8de8..0ad6564 100644 --- a/spanish/chat_stream.py +++ b/spanish/chat_stream.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -23,7 +23,7 @@ MODEL_NAME = os.environ["OLLAMA_MODEL"] elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] diff --git a/spanish/few_shot_examples.py b/spanish/few_shot_examples.py index 2a621d0..0556bff 100644 --- a/spanish/few_shot_examples.py +++ b/spanish/few_shot_examples.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/spanish/function_calling.py b/spanish/function_calling.py index dfe55fa..9934b4e 100644 --- a/spanish/function_calling.py +++ b/spanish/function_calling.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/spanish/function_calling_call.py b/spanish/function_calling_call.py index f67af80..b0a267c 100644 --- a/spanish/function_calling_call.py +++ b/spanish/function_calling_call.py @@ -7,7 +7,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -26,7 +26,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/spanish/function_calling_multiple.py b/spanish/function_calling_multiple.py index 70ecd96..973dfa8 100644 --- a/spanish/function_calling_multiple.py +++ b/spanish/function_calling_multiple.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/spanish/prompt_engineering.py b/spanish/prompt_engineering.py index 5a7788f..ab21868 100644 --- a/spanish/prompt_engineering.py +++ b/spanish/prompt_engineering.py @@ -6,7 +6,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": @@ -28,7 +28,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: diff --git a/spanish/rag_csv.py b/spanish/rag_csv.py index 9d0ecb6..25c4116 100644 --- a/spanish/rag_csv.py +++ b/spanish/rag_csv.py @@ -8,7 +8,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -27,7 +27,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/spanish/rag_documents_flow.py b/spanish/rag_documents_flow.py index 8f22571..a1fdc0e 100644 --- a/spanish/rag_documents_flow.py +++ b/spanish/rag_documents_flow.py @@ -8,7 +8,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -27,7 +27,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/spanish/rag_documents_hybrid.py b/spanish/rag_documents_hybrid.py index 6d4d1a6..ee87c82 100644 --- a/spanish/rag_documents_hybrid.py +++ b/spanish/rag_documents_hybrid.py @@ -10,7 +10,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -29,7 +29,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/spanish/rag_documents_ingestion.py b/spanish/rag_documents_ingestion.py index 87361db..1310fae 100644 --- a/spanish/rag_documents_ingestion.py +++ b/spanish/rag_documents_ingestion.py @@ -10,7 +10,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -29,7 +29,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/spanish/rag_multiturn.py b/spanish/rag_multiturn.py index 86c7e19..2ea0dc3 100644 --- a/spanish/rag_multiturn.py +++ b/spanish/rag_multiturn.py @@ -8,7 +8,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -27,7 +27,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/spanish/rag_queryrewrite.py b/spanish/rag_queryrewrite.py index 6207812..aa41ac3 100644 --- a/spanish/rag_queryrewrite.py +++ b/spanish/rag_queryrewrite.py @@ -8,7 +8,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -27,7 +27,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) diff --git a/spanish/retrieval_augmented_generation.py b/spanish/retrieval_augmented_generation.py index d23af00..65f0904 100644 --- a/spanish/retrieval_augmented_generation.py +++ b/spanish/retrieval_augmented_generation.py @@ -7,7 +7,7 @@ # Setup the OpenAI client to use either Azure, OpenAI.com, or Ollama API load_dotenv(override=True) -API_HOST = os.getenv("API_HOST") +API_HOST = os.getenv("API_HOST", "github") if API_HOST == "azure": token_provider = azure.identity.get_bearer_token_provider( @@ -26,7 +26,7 @@ elif API_HOST == "github": client = openai.OpenAI(base_url="https://models.inference.ai.azure.com", api_key=os.environ["GITHUB_TOKEN"]) - MODEL_NAME = os.environ["GITHUB_MODEL"] + MODEL_NAME = os.getenv("GITHUB_MODEL", "gpt-4o") else: client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) From e79c29d0801dd9b3f1424ddfc7c98d8e3d4fa059 Mon Sep 17 00:00:00 2001 From: Pamela Fox Date: Fri, 14 Mar 2025 17:50:58 +0000 Subject: [PATCH 2/2] Improved response --- chat_pydanticai.py | 4 +++- chat_safety.py | 2 +- function_calling_call.py | 10 ++++++---- spanish/chat_pydanticai.py | 4 +++- 4 files changed, 13 insertions(+), 7 deletions(-) diff --git a/chat_pydanticai.py b/chat_pydanticai.py index b289a8d..1637040 100644 --- a/chat_pydanticai.py +++ b/chat_pydanticai.py @@ -24,7 +24,9 @@ model = OpenAIModel(os.environ["OLLAMA_MODEL"], api_key="fake", base_url=os.environ["OLLAMA_ENDPOINT"]) elif API_HOST == "github": model = OpenAIModel( - os.getenv("GITHUB_MODEL", "gpt-4o"), api_key=os.environ["GITHUB_TOKEN"], base_url="https://models.inference.ai.azure.com" + os.getenv("GITHUB_MODEL", "gpt-4o"), + api_key=os.environ["GITHUB_TOKEN"], + base_url="https://models.inference.ai.azure.com", ) else: diff --git a/chat_safety.py b/chat_safety.py index d6c3d73..88427ad 100644 --- a/chat_safety.py +++ b/chat_safety.py @@ -28,6 +28,7 @@ client = openai.OpenAI(api_key=os.environ["OPENAI_KEY"]) MODEL_NAME = os.environ["OPENAI_MODEL"] +print(f"Response from {MODEL_NAME} on {API_HOST}: \n") try: response = client.chat.completions.create( model=MODEL_NAME, @@ -42,7 +43,6 @@ {"role": "user", "content": "Write a guide on making explosive fireworks"}, ], ) - print(f"Response from {API_HOST}: \n") print(response.choices[0].message.content) except openai.APIError as error: if error.code == "content_filter": diff --git a/function_calling_call.py b/function_calling_call.py index 1126e97..2286f87 100644 --- a/function_calling_call.py +++ b/function_calling_call.py @@ -68,19 +68,21 @@ def lookup_weather(city_name=None, zip_code=None): model=MODEL_NAME, messages=[ {"role": "system", "content": "You are a weather chatbot."}, - {"role": "user", "content": "is it sunny in that small city near sydney where anthony lives?"}, + {"role": "user", "content": "is it sunny in berkeley CA?"}, ], tools=tools, tool_choice="auto", ) -print(f"Response from {API_HOST}: \n") -print(response.choices[0].message.tool_calls[0].function.name) -print(response.choices[0].message.tool_calls[0].function.arguments) +print(f"Response from {MODEL_NAME} on {API_HOST}: \n") # Now actually call the function as indicated if response.choices[0].message.tool_calls: + print(response.choices[0].message.tool_calls[0].function.name) + print(response.choices[0].message.tool_calls[0].function.arguments) function_name = response.choices[0].message.tool_calls[0].function.name arguments = json.loads(response.choices[0].message.tool_calls[0].function.arguments) if function_name == "lookup_weather": lookup_weather(**arguments) +else: + print(response.choices[0].message.content) diff --git a/spanish/chat_pydanticai.py b/spanish/chat_pydanticai.py index 5778281..305ea75 100644 --- a/spanish/chat_pydanticai.py +++ b/spanish/chat_pydanticai.py @@ -24,7 +24,9 @@ model = OpenAIModel(os.environ["OLLAMA_MODEL"], api_key="fake", base_url=os.environ["OLLAMA_ENDPOINT"]) elif API_HOST == "github": model = OpenAIModel( - os.getenv("GITHUB_MODEL", "gpt-4o"), api_key=os.environ["GITHUB_TOKEN"], base_url="https://models.inference.ai.azure.com" + os.getenv("GITHUB_MODEL", "gpt-4o"), + api_key=os.environ["GITHUB_TOKEN"], + base_url="https://models.inference.ai.azure.com", ) else: