The Writer LangChain integration allows you to leverage Writer’s capabilities within the LangChain ecosystem, making it easy to build sophisticated AI applications. In this tutorial, you’ll explore each component of the integration and understand how they work.

Prerequisites

Before you begin, make sure you have:

Installation and setup

First, install the necessary packages:

pip install langchain-writer python-dotenv

Next, create a .env file with WRITER_API_KEY set to your Writer API key:

WRITER_API_KEY=<your-api-key>

Components of the Writer LangChain integration

The langchain-writer package provides several key components:

  1. ChatWriter for text generation
  2. Tool calling capabilities, including GraphTool for Knowledge Graph integration
  3. Additional tools like PDFParser for parsing PDFs and WriterTextSplitter for intelligent text splitting

ChatWriter

ChatWriter is a LangChain chat model that provides access to Writer’s AI capabilities for text generation. It supports streaming, non-streaming, batching, and asynchronous operations. You can use any of the Palmyra chat models available in AI Studio.

See the full documentation for ChatWriter to learn more about the available parameters.

Usage

This example uses ChatWriter to ask the Palmyra X 004 model to explain what LangChain is in plain terms.

from langchain_writer import ChatWriter
from dotenv import load_dotenv

load_dotenv()

# Initialize the chat model
# These are optional parameters with default values listed here
llm = ChatWriter(
    model="palmyra-x-004",  # default model
    temperature=0.7,        # controls randomness (0-1)
    max_tokens=None,        # maximum number of tokens to generate
    timeout=None,           # request timeout
    max_retries=2          # number of retries on failure
)

# Generate a response
response = llm.invoke("Explain what LangChain is in simple terms.")
print(response.content)

Streaming

Streaming allows you to receive the generated text in chunks as it’s being produced. This example shows how to stream a response from the Palmyra X 004 model using synchronous streaming:

from langchain_writer import ChatWriter
import asyncio
from dotenv import load_dotenv

load_dotenv()

llm = ChatWriter()

# Streaming (synchronous)
for chunk in llm.stream("Write a short poem about artificial intelligence."):
    print(chunk.content, end="")

You can also use async streaming with the async for loop and the astream method:

from langchain_writer import ChatWriter
import asyncio
from dotenv import load_dotenv

load_dotenv()

llm = ChatWriter()

# Streaming (asynchronous)
async def async_stream():
    async for chunk in llm.astream("Write a short poem about artificial intelligence."):
        print(chunk.content, end="")

asyncio.run(async_stream())

Batch processing

You can batch process multiple prompts for efficient processing. The following example batches three individual LLM invocations and runs them in parallel:

from langchain_writer import ChatWriter
from dotenv import load_dotenv

load_dotenv()

llm = ChatWriter()

questions = [
    "What is a three-sentence definition of retrieval-augmented generation?",
    "When did transformer architecture first appear in the literature?",
    "What are 3 benefits of using AI to assist in software development?"
]

# Process multiple prompts in parallel
responses = llm.batch(questions)

for question, response in zip(questions, responses):
    print(f"Q: {question}")
    print(f"A: {response.content}\n")

Note that batch returns results in the same order as the inputs. You can use batch_as_completed to return results as they complete. Results may arrive out of order, but each includes the input index for matching.

You can also optionally set the max_concurrency parameter to control the number of concurrent requests, which can be useful when you want to limit the number of parallel calls to prevent overloading a server or API:

responses = llm.batch(questions,
    config={"max_concurrency": 2}
)

See the LangChain documentation on parallel execution for more information.

Tool calling

ChatWriter supports tool calling, which allows the model to use external functions to enhance its capabilities. Tool calling is available with Palmyra X 004 and later.

Tool calling basics

To use tool calling, follow these steps:

  1. Define a function that will be called by the model and decorate it with the @tool decorator.
  2. Bind the tool to the chat model using the bind_tools method.
  3. Use the tool in a chat and append the response to the messages list.
  4. Execute the tool call with the arguments given by the model and append the response to the messages list.
  5. Invoke the chat model with the updated messages list to receive the final response.

Here’s an example of how to use tool calling:

from langchain_writer import ChatWriter
from langchain_core.tools import tool
from langchain_core.messages import HumanMessage
from dotenv import load_dotenv

load_dotenv()

@tool
def get_weather(location: str) -> str:
    """Get the current weather for a location."""
    # In a real application, you would call a weather API here
    return f"The weather in {location} is sunny and 75°F"

chat = ChatWriter()
chat.bind_tools([get_weather])

messages = [
    HumanMessage(
        "What's the weather like in San Francisco?"
    )
]

response = chat.invoke(messages)
messages.append(response)
print(response.tool_calls)

for tool_call in response.tool_calls:
    selected_tool = {
        "get_weather": get_weather,
    }[tool_call["name"].lower()]
    tool_msg = selected_tool.invoke(tool_call)
    messages.append(tool_msg)

response = chat.invoke(messages)
print(response.content)

GraphTool

GraphTool is a LangChain tool that allows you to retrieve information from a Knowledge Graph to enhance its responses. The tool executes remotely, so you simply need to provide the IDs of the Knowledge Graph you want to use. For more details on the built-in Knowledge Graph chat tool in Writer, see the Knowledge Graph chat support guide.

Usage

from langchain_writer import ChatWriter
from langchain_writer.tools import GraphTool
from dotenv import load_dotenv

load_dotenv()

# Initialize the chat model
chat = ChatWriter()

# Create a graph tool with your knowledge graph ID
graph_tool = GraphTool(graph_ids=["your-knowledge-graph-id"])

# Bind the tool to the chat model
chat.bind_tools([graph_tool])

# Ask a question that can be answered using the knowledge graph
response = chat.invoke("What information do you have about product X?")
print(response.content)

Additional tools

PDFParser

PDFParser is a document loader that uses Writer’s PDF parsing capabilities to extract text from PDF documents. It converts PDFs into LangChain Document objects that can be used in your applications. For more details on the underlying API, see the PDF parser API reference.

Usage

Here’s an example of how to use the PDFParser:

from langchain_writer import PDFParser
from langchain_core.documents.base import Blob
from dotenv import load_dotenv

load_dotenv()

# Initialize the PDF parser with the desired output format
parser = PDFParser(output_format="markdown")  # Options: "text", "markdown"

# Load a PDF file
file = Blob.from_path("path/to/your/document.pdf")

parsed_pages = parser.parse(blob=file)
print(parsed_pages)

Output formats

The PDFParser supports different output formats:

  • text: Plain text extraction
  • markdown: Structured markdown with preserved formatting
# For markdown output
markdown_parser = PDFParser(output_format="markdown")
markdown_docs = markdown_parser.load("path/to/your/document.pdf")

WriterTextSplitter

WriterTextSplitter is a text splitter that uses Writer’s context-aware splitting capabilities to divide documents into semantically meaningful chunks. This is particularly useful for preparing documents for retrieval systems. For more details on the underlying API, see the context-aware splitting tool API reference.

Usage

Here’s an example of how to use the WriterTextSplitter:

from langchain_writer import WriterTextSplitter
from dotenv import load_dotenv

load_dotenv()

# Initialize the text splitter with the desired strategy
text_splitter = WriterTextSplitter(strategy="hybrid_split")

# Long text you wish to split
text = "Long text you wish to split..."

# Split the document
chunks = text_splitter.split_text(text)

# Print the chunks
for i, chunk in enumerate(chunks):
    print(f"Chunk {i+1}: {chunk[:50]}...")

Splitting strategies

The WriterTextSplitter supports different splitting strategies:

  • llm_split: Uses language model for precise semantic splitting
  • fast_split: Uses heuristic-based approach for quick splitting
  • hybrid_split: Combines both approaches for a balance of speed and quality
# For language model-based splitting
llm_splitter = WriterTextSplitter(strategy="llm_split")

# For faster, heuristic-based splitting
fast_splitter = WriterTextSplitter(strategy="fast_split")

Conclusion

In this tutorial, you’ve explored the LangChain integration with Writer, covering each of its components:

  1. ChatWriter for text generation
  2. Tool calling capabilities, including GraphTool for Knowledge Graph integration
  3. Additional tools like PDFParser for parsing PDFs and WriterTextSplitter for intelligent text splitting

This integration provides a powerful foundation for building AI applications with Writer and LangChain. Check out the package README and documentation, as well as the LangChain Documentation for more information on how to use LangChain with Writer.