state = LazyState(problem="What is the result of f(x) = 3x + 2 when x = 5?")state.add_step("First, we need substitute x in the function with 5")state
LazyState(problem='What is the result of f(x) = 3x + 2 when x = 5?', steps=['What is the result of f(x) = 3x + 2 when x = 5?', 'First, we need substitute x in the function with 5'], current_step=1)
state.get_context()
"Problem: What is the result of f(x) = 3x + 2 when x = 5? \n Steps so far: ['What is the result of f(x) = 3x + 2 when x = 5?', 'First, we need substitute x in the function with 5']"
To keep things general, we’ll have a very simple data class called LLM that will give us an interface for the language model client and the model version.
This package will also keep support for our internal system prompt, called lazy_system_p
With this simple state manager, we have a way to track each step of the problem-solving process and get the current context to be used for a call to a language model.
Now, let’s set up a class LazyEvaluationClient that will do the heavy lifting of managing the state and calling the language model.
LazyEvaluationClient (llm:__main__.LLM, max_tokens:int=100,
state:Optional[__main__.LazyState]=None,
lazy_system_p:str='\n You are a helpful
assistant that can help with math problems.\n
You will be given a problem and a list of steps as
context, the format will be:\n \n
PROBLEM: <problem>\n STEPS: <steps>\n\n
Your job is to complete the next step and only the
next step in the problem-solving process. You
should never give more than one step.\n If
you evaluate that the problem is done, respond with
"PROBLEM DONE"\n ')
You are a helpful assistant that can help with math problems. You will be given a problem and a list of steps as context, the format will be:
PROBLEM: STEPS:
Your job is to complete the next step and only the next step in the problem-solving process. You should never give more than one step. If you evaluate that the problem is done, respond with “PROBLEM DONE”
lazy_state = LazyState(problem="What is the derivative of `2x^3 + x^2 + 2x + 1`? Give me the solution step-by-step")client = AnthropicVertex(project_id=project_id, region=location)llm = LLM(client=client, model=model)lazy_lm = LazyEvaluationClient(llm=llm, state=lazy_state)
Let’s grab the current step
lazy_lm.get_current_step()
'What is the derivative of `2x^3 + x^2 + 2x + 1`? Give me the solution step-by-step'
Let’s get the next step
lazy_lm.get_next_step()
"To find the derivative of the given function, we'll use the power rule and the constant rule of differentiation. Let's start with the first term:\n\nThe derivative of 2x^3 is:\n3 * 2x^(3-1) = 6x^2"
and the next…
lazy_lm.get_next_step()
"Let's proceed with the next step in finding the derivative:\n\nThe derivative of x^2 is:\n2 * x^(2-1) = 2x"
Let’s look at our current state
for step in lazy_lm.state.steps:print("-"*10)print(step)
----------
What is the derivative of `2x^3 + x^2 + 2x + 1`? Give me the solution step-by-step
----------
To find the derivative of the given function, we'll use the power rule and the constant rule of differentiation. Let's start with the first term:
The derivative of 2x^3 is:
3 * 2x^(3-1) = 6x^2
----------
Let's proceed with the next step in finding the derivative:
The derivative of x^2 is:
2 * x^(2-1) = 2x
We need a way to take the current reasoning step and query it without having the model advance to the next step in the problem-solving process.
*Allows the user to ask a question about the current step without affecting the model’s ability to generate the next step.
Args: question (str): The question the user wants to ask about the current step.
Returns: str: The model’s response to the question.*
lazy_lm.get_current_step()
"Let's proceed with the next step in finding the derivative:\n\nThe derivative of x^2 is:\n2 * x^(2-1) = 2x"
lazy_lm.ask_question("This does not make sense to me")
"I apologize for the confusion. Let me explain the current step without advancing further:\n\nThe step we're looking at is finding the derivative of the x^2 term in the original expression 2x^3 + x^2 + 2x + 1.\n\nTo find the derivative of x^2, we use the power rule of differentiation. The power rule states that for a term ax^n, the derivative is n * ax^(n-1"
The Lazy Evaluation Flow
This simple framework effectivly shows how we can wrape a language model capable of step-by-step reasoning to create a lazy evaluator.
This approach follows these system design steps:
Problem Initialization: A state manager is initalized with a problem
Prompting Strategy: Prompt the language model to generate the next step given the context in the state manager.
State Update: State Manager records the newly generated step and updates.
User Interaction: User interaction is held within a different state manager question_history which does not affect the overall state of the current problem.
Adaptive Response: Based on the current input, the Lazy Evaluator decides to either 1) Generate the next step or 2) Provide a response to the user’s question given the current state of the problem and the question history.
Finally, let’s patch different model provider’s APIs