# Eval set to false, because the api key is stored in .env and thus can't be found when
# nbdev_test is run
= ModelConfig(
hf_config ="mistralai/Mistral-7B-Instruct-v0.3", # "Qwen/QwQ-32B" is another possibility, but with vision you need another messages format
model_name="huggingface",
provider="HF_API_KEY",
api_key_env_var="https://router.huggingface.co/hf-inference/v1",
api_base_url=100,
max_completion_tokens=0.7
temperature
)
# Create the client
= create_llm_client(hf_config) client
App logic
Import statement
from gradiochat.app import *
Explanation of new code
I used the Protocol
and runtime_checkable
for the first time. Here’s a short explainer.
Python’s Protocol System
Protocols were introduced in Python 3.8 through PEP 544 and are part of the typing
module. They provide a way to define interfaces that classes can implement without explicitly inheriting from them - this is called “structural typing” or “duck typing.”
Protocols vs Abstract Base Classes (ABCs)
Abstract Base Classes (Traditional Approach): - Require explicit inheritance (class MyClass(AbstractBaseClass):
) - Use the @abstractmethod
decorator to mark methods that must be implemented - Check for compatibility based on the class hierarchy (nominal typing) - Enforce implementation at class definition time
Protocols (New Approach): - Don’t require inheritance - classes just need to implement the required methods - Use the Protocol
class and @runtime_checkable
decorator - Check for compatibility based on method signatures (structural typing) - Can check compatibility at runtime with isinstance()
if marked as @runtime_checkable
How It Works
In our code:
@runtime_checkable
class LLMClientProtocol(Protocol):
def chat_completion(self, messages: List[Message], **kwargs) -> str:
...
def chat_completion_stream(self, messages: List[Message], **kwargs) -> Generator[str, None, None]:
...
This defines an interface that says “any class with methods named chat_completion
and chat_completion_stream
with these signatures is considered compatible with LLMClientProtocol
.”
The ...
in the method bodies is a special syntax that means “this method is required but not implemented here.” It’s similar to pass
but specifically for protocol definitions.
Benefits in Our Context
Flexibility: We can create any class that implements these methods, and it will be compatible with
LLMClientProtocol
without explicitly inheriting from it.Easy Testing: We can create mock implementations that automatically satisfy the protocol by just implementing the required methods.
Type Checking: Tools like mypy can verify that our classes implement all required methods with the correct signatures.
Runtime Checking: With
@runtime_checkable
, we can useisinstance(obj, LLMClientProtocol)
to check if an object implements the protocol.
Example of Use
def process_with_any_llm_client(client: LLMClientProtocol, messages: List[Message]):
# This function will accept any object that has the required methods,
# regardless of its class hierarchy
= client.chat_completion(messages)
response return response
This would accept our HuggingFaceClient
or any other class that implements the required methods, without forcing them to inherit from a common base class.
Define the general LLMClientProtocol structure
Which means it should have the methods defined in LLMClientProtocol
.
LLMClientProtocol
LLMClientProtocol (*args, **kwargs)
Protocol defining the interface for LLM clients
Define the LLM Clients
This should at least follow the structure of LLMClientProtocol
but can of course be expanded.
HuggingFaceClient
HuggingFaceClient
HuggingFaceClient (model_config:gradiochat.config.ModelConfig)
Client for interacting with HuggingFace models
TogetherAI
TogetherAiClient
TogetherAiClient (model_config:gradiochat.config.ModelConfig)
Client for interacting with models through the TogetherAI API server We use the openai package
Local Ollama client
OllamaClient
OllamaClient (model_config:gradiochat.config.ModelConfig)
Client for interacting with models through a local Ollama API server Uses the official Ollama Python library
Create the LLM client
This function creates the client using the available LLM Client classes. It gets the provider from the model_config
. If it finds a LLM Client Class for this provider, it returns that client. If it doesn’t find a LLM Client Class for that provider, it returns a ValueError.
create_llm_client
create_llm_client (model_config:gradiochat.config.ModelConfig)
Factory function to create an LLM client based on the provider.
The internal logic of the chat app
Now the BaseChatApp
class is defined. This class is used to instantiate the properties en methods for the internal workings of the chat app. The UI is defined in the ui
module.
BaseChatApp
BaseChatApp (config:gradiochat.config.ChatAppConfig)
Base class for creating configurable chat applications with Gradio
Create a HuggingFace test model config
Create a Together AI test model config
# Eval set to false, because the api key is stored in .env and thus can't be found when
# nbdev_test is run
= ModelConfig(
ta_config # model_name="mistralai/Mistral-Nemo-Instruct-2407",
="meta-llama/Llama-3.3-70B-Instruct-Turbo-Free",
model_name="togetherai",
provider="TG_API_KEY",
api_key_env_var
)
# Create the client
= create_llm_client(ta_config) client
--------------------------------------------------------------------------- NameError Traceback (most recent call last) Cell In[11], line 12 4 ta_config = ModelConfig( 5 # model_name="mistralai/Mistral-Nemo-Instruct-2407", 6 model_name="meta-llama/Llama-3.3-70B-Instruct-Turbo-Free", 7 provider="togetherai", 8 api_key_env_var="TG_API_KEY", 9 ) 11 # Create the client ---> 12 client = create_llm_client(ta_config) Cell In[5], line 9, in create_llm_client(model_config) 7 return HuggingFaceClient(model_config) 8 if model_config.provider.lower() == "togetherai": ----> 9 return TogetherAiClient(model_config) 10 if model_config.provider.lower() == "ollama": 11 return OllamaClient(model_config) NameError: name 'TogetherAiClient' is not defined
Create a Ollama test model config
# Eval set to false, because the api key is stored in .env and thus can't be found when
# nbdev_test is run
= ModelConfig(
olla_config ="nchapman/ministral-8b-instruct-2410",
model_name="ollama",
provider="OLLAMA_API_KEY",
api_key_env_var
)
# Create the client
= create_llm_client(olla_config) client
--------------------------------------------------------------------------- NameError Traceback (most recent call last) Cell In[10], line 11 4 olla_config = ModelConfig( 5 model_name="nchapman/ministral-8b-instruct-2410", 6 provider="ollama", 7 api_key_env_var="OLLAMA_API_KEY", 8 ) 10 # Create the client ---> 11 client = create_llm_client(olla_config) Cell In[5], line 11, in create_llm_client(model_config) 9 return TogetherAiClient(model_config) 10 if model_config.provider.lower() == "ollama": ---> 11 return OllamaClient(model_config) 12 else: 13 raise ValueError(f"Unsupported provider: {model_config.provider}") NameError: name 'OllamaClient' is not defined
= [
test_messages ="system", content="You are Aurelius Augustinus, helping me to think deeply and be humble and thankfull."),
Message(role="user", content="Why should I engage with the people around me?")
Message(role
]# Test with a simple prompt
try:
= client.chat_completion(test_messages)
response print(f"Response received: {response[:100]}...")
except Exception as e:
print(f"Error: {e}")
# Test with overriden parameters
try:
print("\nTesting with overridden parameters:")
= client.chat_completion(test_messages, max_completion_tokens=50, temperature=0.9)
response print(f"Response: {response[:100]}...") # Show first 100 chars
except Exception as e:
print(f"Error: {e}")
Error: Error code: 402 - {'error': 'You have exceeded your monthly included credits for Inference Providers. Subscribe to PRO to get 20x more monthly included credits.'}
Testing with overridden parameters:
Error: Error code: 402 - {'error': 'You have exceeded your monthly included credits for Inference Providers. Subscribe to PRO to get 20x more monthly included credits.'}