feat: adding LlamaStack integration with chat, embeddings, and safety #32706
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description: This PR introduces a comprehensive LangChain integration for Llama Stack, providing full-featured chat completion, text embeddings, and safety checking capabilities. The integration supports multiple providers (Ollama, OpenAI, Together AI, Fireworks, etc.) and includes advanced shield management for content moderation.
Key features:
The integration follows LangChain patterns and includes proper error handling, logging, and comprehensive documentation.
Dependencies:
llama-stack-client>=0.0.40
(optional, for Llama Stack integration)httpx>=0.25.0
pydantic>=1.10.0
numpy
andscikit-learn
(optional, for similarity search)Additional Notes:
make format
,make lint
, andmake test
examples/
directory