MLflow
MLflow is a versatile, open-source platform for managing workflows and artifacts across the machine learning lifecycle. It has built-in integrations with many popular ML libraries, but can be used with any library, algorithm, or deployment tool. It is designed to be extensible, so you can write plugins to support new workflows, libraries, and tools.
In the context of LangChain integration, MLflow provides the following capabilities:
- Experiment Tracking: MLflow tracks and stores artifacts from your LangChain experiments, including models, code, prompts, metrics, and more.
- Dependency Management: MLflow automatically records model dependencies, ensuring consistency between development and production environments.
- Model Evaluation MLflow offers native capabilities for evaluating LangChain applications.
- Tracing: MLflow allows you to visually trace data flows through your LangChain chain, agent, retriever, or other components.
Note: The tracing capability is only available in MLflow versions 2.14.0 and later.
This notebook demonstrates how to track your LangChain experiments using MLflow. For more information about this feature and to explore tutorials and examples of using LangChain with MLflow, please refer to the MLflow documentation for LangChain integration.
Setupβ
Install MLflow Python package:
%pip install google-search-results num
%pip install mlflow -qU
This example utilizes the OpenAI LLM. Feel free to skip the command below and proceed with a different LLM if desired.
%pip install langchain-openai -qU
import os
# Set MLflow tracking URI if you have MLflow Tracking Server running
os.environ["MLFLOW_TRACKING_URI"] = ""
os.environ["OPENAI_API_KEY"] = ""
To begin, let's create a dedicated MLflow experiment in order track our model and artifacts. While you can opt to skip this step and use the default experiment, we strongly recommend organizing your runs and artifacts into separate experiments to avoid clutter and maintain a clean, structured workflow.
import mlflow
mlflow.set_experiment("LangChain MLflow Integration")
Overviewβ
Integrate MLflow with your LangChain Application using one of the following methods:
- Autologging: Enable seamless tracking with the
mlflow.langchain.autolog()
command, our recommended first option for leveraging the LangChain MLflow integration. - Manual Logging: Use MLflow APIs to log LangChain chains and agents, providing fine-grained control over what to track in your experiment.
- Custom Callbacks: Pass MLflow callbacks manually when invoking chains, allowing for semi-automated customization of your workload, such as tracking specific invocations.
Scenario 1: MLFlow Autologgingβ
To get started with autologging, simply call mlflow.langchain.autolog()
. In this example, we set the log_models
parameter to True
, which allows the chain definition and its dependency libraries to be recorded as an MLflow model, providing a comprehensive tracking experience.
import mlflow
mlflow.langchain.autolog(
# These are optional configurations to control what information should be logged automatically (default: False)
# For the full list of the arguments, refer to https://mlflow.org/docs/latest/llms/langchain/autologging.html#id1
log_models=True,
log_input_examples=True,
log_model_signatures=True,
)
Define a Chainβ
from langchain.schema.output_parser import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate
from langchain_openai import ChatOpenAI
llm = ChatOpenAI(model_name="gpt-4o", temperature=0)
prompt = ChatPromptTemplate.from_messages(
[
(
"system",
"You are a helpful assistant that translates {input_language} to {output_language}.",
),
("human", "{input}"),
]
)
parser = StrOutputParser()
chain = prompt | llm | parser
Invoke the Chainβ
Note that this step may take a few seconds longer than usual, as MLflow runs several background tasks in the background to log models, traces, and artifacts to the tracking server.
test_input = {
"input_language": "English",
"output_language": "German",
"input": "I love programming.",
}
chain.invoke(test_input)
'Ich liebe das Programmieren.'
Take a moment to explore the MLflow Tracking UI, where you can gain a deeper understanding of what information are being logged.
- Traces - Navigate to the "Traces" tab in the experiment and click the request ID link of the first row. The displayed trace tree visualizes the call stack of your chain invocation, providing you with a deep insight into how each component is executed within the chain.
- MLflow Model - As we set
log_model=True
, MLflow automatically creates an MLflow Run to track your chain definition. Navigate to the newest Run page and open the "Artifacts" tab, which lists file artifacts logged as an MLflow Model, including dependencies, input examples, model signatures, and more.
Invoke the Logged Chainβ
Next, let's load the model back and verify that we can reproduce the same prediction, ensuring consistency and reliability.
There are two ways to load the model
mlflow.langchain.load_model(MODEL_URI)
- This loads the model as the original LangChain object.mlflow.pyfunc.load_model(MODEL_URI)
- This loads the model within thePythonModel
wrapper and encapsulates the prediction logic with thepredict()
API, which contains additional logic such as schema enforcement.
# Replace YOUR_RUN_ID with the Run ID displayed on the MLflow UI
loaded_model = mlflow.langchain.load_model("runs:/{YOUR_RUN_ID}/model")
loaded_model.invoke(test_input)
'Ich liebe Programmieren.'
pyfunc_model = mlflow.pyfunc.load_model("runs:/{YOUR_RUN_ID}/model")
pyfunc_model.predict(test_input)
['Ich liebe das Programmieren.']
Configure Autologgingβ
The mlflow.langchain.autolog()
function offers several parameters that allow for fine-grained control over the artifacts logged to MLflow. For a comprehensive list of available configurations, please refer to the latest MLflow LangChain Autologging Documentation.
Scenario 2: Manually Logging an Agent from Codeβ
Prerequisitesβ
This example uses SerpAPI
, a search engine API, as a tool for the agent to retrieve Google Search results. LangChain is natively integrated with SerpAPI
, allowing you to configure the tool for your agent with just one line of code.
To get started:
- Install the required Python package via pip:
pip install google-search-results numexpr
. - Create an account at SerpAPI's Official Website and retrieve an API key.
- Set the API key in the environment variable:
os.environ["SERPAPI_API_KEY"] = "YOUR_API_KEY"
Define an Agentβ
In this example, we will log the agent definition as code, rather than directly feeding the Python object and saving it in a serialized format. This approach offers several benefits:
- No serialization required: By saving the model as code, we avoid the need for serialization, which can be problematic when working with components that don't natively support it. This approach also eliminates the risk of incompatibility issues when deserializing the model in a different environment.
- Better transparency: By inspecting the saved code file, you can gain valuable insights into what the model does. This is in contrast to serialized formats like pickle, where the model's behavior remains opaque until it's loaded back, potentially exposing security risks such as remote code execution.
First, create a separate .py
file that defines the agent instance.
In the interest of time, you can run the following cell to generate a Python file agent.py
, which contains the agent definition code. In actual dev scenario, you would define it in another notebook or hand-crafted python script.
script_content = """
from langchain.agents import AgentType, initialize_agent, load_tools
from langchain_openai import ChatOpenAI
import mlflow
llm = ChatOpenAI(model_name="gpt-4o", temperature=0)
tools = load_tools(["serpapi", "llm-math"], llm=llm)
agent = initialize_agent(tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION)
# IMPORTANT: call set_model() to register the instance to be logged.
mlflow.models.set_model(agent)
"""
with open("agent.py", "w") as f:
f.write(script_content)
Log the Agentβ
Return to the original notebook and run the following cell to log the agent you've defined in the agent.py
file.
question = "How long would it take to drive to the Moon with F1 racing cars?"
with mlflow.start_run(run_name="search-math-agent") as run:
info = mlflow.langchain.log_model(
lc_model="agent.py", # Specify the relative code path to the agent definition
artifact_path="model",
input_example=question,
)
print("The agent is successfully logged to MLflow!")
The agent is successfully logged to MLflow!
Now, open the MLflow UI and navigate to the "Artifacts" tab in the Run detail page. You should see that the agent.py
file has been successfully logged, along with other model artifacts, such as dependencies, input examples, and more.
Invoke the Logged Agentβ
Now load the agent back and invoke it. There are two ways to load the model
# Let's turn on the autologging with default configuration, so we can see the trace for the agent invocation.
mlflow.langchain.autolog()
# Load the model back
agent = mlflow.pyfunc.load_model(info.model_uri)
# Invoke
agent.predict(question)
Downloading artifacts: 100%|ββββββββββ| 10/10 [00:00<00:00, 331.57it/s]
['It would take approximately 1194.5 hours to drive to the Moon with an F1 racing car.']
Navigate to the "Traces" tab in the experiment and click the request ID link of the first row. The trace visualizes how the agent operate multiple tasks within the single prediction call:
- Determine what subtasks are required to answer the questions.
- Search for the speed of an F1 racing car.
- Search for the distance from Earth to Moon.
- Compute the division using LLM.
Scenario 3. Using MLflow Callbacksβ
MLflow Callbacks provide a semi-automated way to track your LangChain application in MLflow. There are two primary callbacks available:
MlflowLangchainTracer
: Primarily used for generating traces, available inmlflow >= 2.14.0
.MLflowCallbackHandler
: Logs metrics and artifacts to the MLflow tracking server.
MlflowLangchainTracerβ
When the chain or agent is invoked with the MlflowLangchainTracer
callback, MLflow will automatically generate a trace for the call stack and log it to the MLflow tracking server. The outcome is exactly same as mlflow.langchain.autolog()
, but this is particularly useful when you want to only trace specific invocation. Autologging is applied to all invocation in the same notebook/script, on the other hand.
from mlflow.langchain.langchain_tracer import MlflowLangchainTracer
mlflow_tracer = MlflowLangchainTracer()
# This call generates a trace
chain.invoke(test_input, config={"callbacks": [mlflow_tracer]})
# This call does not generate a trace
chain.invoke(test_input)
Where to Pass the Callbackβ
LangChain supports two ways of passing callback instances: (1) Request time callbacks - pass them to the invoke
method or bind with with_config()
(2) Constructor callbacks - set them in the chain constructor. When using the MlflowLangchainTracer
as a callback, you must use request time callbacks. Setting it in the constructor instead will only apply the callback to the top-level object, preventing it from being propagated to child components, resulting in incomplete traces. For more information on this behavior, please refer to Callbacks Documentation for more details.
# OK
chain.invoke(test_input, config={"callbacks": [mlflow_tracer]})
chain.with_config(callbacks=[mlflow_tracer])
# NG
chain = TheNameOfSomeChain(callbacks=[mlflow_tracer])
Supported Methodsβ
MlflowLangchainTracer
supports the following invocation methods from the Runnable Interfaces.
- Standard interfaces:
invoke
,stream
,batch
- Async interfaces:
astream
,ainvoke
,abatch
,astream_log
,astream_events
Other methods are not guaranteed to be fully compatible.
MlflowCallbackHandlerβ
MlflowCallbackHandler
is a callback handler that resides in the LangChain Community code base.
This callback can be passed for chain/agent invocation, but it must be explicitly finished by calling the flush_tracker()
method.
When a chain is invoked with the callback, it performs the following actions:
- Creates a new MLflow Run or retrieves an active one if available within the active MLflow Experiment.
- Logs metrics such as the number of LLM calls, token usage, and other relevant metrics. If the chain/agent includes LLM call and you have
spacy
library installed, it logs text complexity metrics such asflesch_kincaid_grade
. - Logs internal steps as a JSON file (this is a legacy version of traces).
- Logs chain input and output as a Pandas Dataframe.
- Calls the
flush_tracker()
method with a chain/agent instance, logging the chain/agent as an MLflow Model.
from langchain_community.callbacks import MlflowCallbackHandler
mlflow_callback = MlflowCallbackHandler()
chain.invoke("What is LangChain callback?", config={"callbacks": [mlflow_callback]})
mlflow_callback.flush_tracker()
Referencesβ
To learn more about the feature and visit tutorials and examples of using LangChain with MLflow, please refer to the MLflow documentation for LangChain integration.
MLflow
also provides several tutorials and examples for the LangChain
integration: