LlamaCpp
Prerequisites
Install Llama Cpp Python by following the instructions provided in the Llama Cpp Python repository.
alternatively, to install with CUDA support:
Initializing the Llama model
Initialize the model within your program with the desired parameters.
from llama_cpp import Llama
llm = Llama(
model_path="./sppo_finetuned_llama_3_8b.gguf",
n_gpu_layers=-1,
n_ctx=0,
verbose=False
)
Sending requests to the model
After initializing the Llama model, you can interact with it using the LlamaCpp
client.
import dspy
llamalm = dspy.LlamaCpp(model="llama", llama_model=llm, model_type="chat", temperature=0.4)
dspy.settings.configure(lm=llamalm)
#Define a simple signature for basic question answering
class BasicQA(dspy.Signature):
"""Answer questions with short factoid answers."""
question = dspy.InputField()
answer = dspy.OutputField(desc="often between 1 and 5 words")
#Pass signature to Predict module
generate_answer = dspy.Predict(BasicQA)
# Call the predictor on a particular input.
question='What is the color of the sky?'
pred = generate_answer(question=question)
print(f"Question: {question}")
print(f"Predicted Answer: {pred.answer}")