Navigate:
LangChain
~$LANGC0.6%

LangChain: Framework for building LLM applications

Modular framework for chaining LLMs with external data.

LIVE RANKINGS • 10:20 AM • STEADY
TOP 50TOP 50TOP 50TOP 50TOP 50TOP 50TOP 50TOP 50TOP 50TOP 50TOP 50TOP 50
OVERALL
#45
3
AI & ML
#26
3
30 DAY RANKING TREND
ovr#45
·AI#26
STARS
127.6K
FORKS
21.0K
7D STARS
+800
7D FORKS
+105
Tags:
See Repo:
Share:

Learn more about LangChain

LangChain is a Python framework for building applications that integrate large language models with external data sources and computational tools. It implements a modular architecture with standardized interfaces for models, embeddings, vector stores, and retrievers, enabling different component implementations to be substituted without modifying core application logic. The framework uses a chain-based composition pattern where operations are linked together sequentially or conditionally to create complex workflows such as retrieval-augmented generation systems and multi-step reasoning agents. Components communicate through a common data structure that preserves context and intermediate results as information flows through the processing pipeline. This abstraction layer allows developers to combine various language models, data retrieval mechanisms, and external APIs into unified applications while maintaining separation between business logic and infrastructure concerns.

LangChain

1

Provider-Agnostic Interfaces

Standardized abstractions let you swap LLM providers, vector stores, and embeddings without rewriting application logic. Switch between OpenAI, Anthropic, or local models by changing configuration, not code.

2

Pre-Built Connector Ecosystem

Includes native integrations for 100+ model providers, vector databases, and external APIs. Eliminates boilerplate for common data sources and reduces time spent on integration code.

3

Composable Chain Architecture

Links LLM calls, tools, and data retrieval into reusable chains with typed inputs and outputs. Build complex reasoning pipelines by composing simple, testable components.


from langgraph.graph import StateGraph, MessagesState, START, END

def chatbot(state: MessagesState):
    return {"messages": [{"role": "ai", "content": "Hello! How can I help?"}]}

graph = StateGraph(MessagesState)
graph.add_node("chatbot", chatbot)
graph.add_edge(START, "chatbot")
graph.add_edge("chatbot", END)
app = graph.compile()

response = app.invoke({"messages": [{"role": "user", "content": "Hi"}]})


vlangchain==1.2.9

Fixes response format schema normalization and improves token counting with threading context support.

  • Fix normalize raw schemas in middleware responseformat override
  • Support state updates from wrapmodelcall with commands
  • Bump min core version and improve approximate token counting
  • Threading context through create_agent flows + middleware
vlangchain-core==1.2.9

Adjusts token count scaling with caps and reverts regex pattern precompilation changes.

  • Adjust cap when scaling approximate token counts
  • Revert precompile hex color regex pattern at module level
  • Apply cap when scaling approximate token counts
  • Allow scaling by reported usage when counting tokens approximately
vlangchain==1.2.8

Fixes tool strategy reuse in agent factory and adds ToolCallRequest to middleware exports.

  • Fix reuse ToolStrategy in agent factory to prevent name mismatch
  • Add ToolCallRequest to middleware exports
  • Fix blocking unit test
  • Use blockbuster to detect blocking calls in the async event loop

See how people are using LangChain

Loading tweets...


[ EXPLORE MORE ]

Related Repositories

Discover similar tools and frameworks used by developers