Releases
v0.10.35
llama-index-agent-introspective
[0.1.0]
Add CRITIC and reflection agent integrations (#13108 )
llama-index-core
[0.10.35]
fix from_defaults()
erasing summary memory buffer history (#13325 )
use existing async event loop instead of asyncio.run()
in core (#13309 )
fix async streaming from query engine in condense question chat engine (#13306 )
Handle ValueError in extract_table_summaries in element node parsers (#13318 )
Handle llm properly for QASummaryQueryEngineBuilder and RouterQueryEngine (#13281 )
expand instrumentation payloads (#13302 )
Fix Bug in sql join statement missing schema (#13277 )
llama-index-embeddings-jinaai
[0.1.5]
add encoding_type parameters in JinaEmbedding class (#13172 )
fix encoding type access in JinaEmbeddings (#13315 )
llama-index-embeddings-nvidia
[0.1.0]
add nvidia nim embeddings support (#13177 )
llama-index-llms-mistralai
[0.1.12]
Fix async issue when streaming with Mistral AI (#13292 )
llama-index-llms-nvidia
[0.1.0]
add nvidia nim llm support (#13176 )
llama-index-postprocessor-nvidia-rerank
[0.1.0]
add nvidia nim rerank support (#13178 )
llama-index-readers-file
[0.1.21]
Update MarkdownReader to parse text before first header (#13327 )
llama-index-readers-web
[0.1.13]
feat: Spider Web Loader (#13200 )
llama-index-vector-stores-vespa
[0.1.0]
Add VectorStore integration for Vespa (#13213 )
llama-index-vector-stores-vertexaivectorsearch
[0.1.0]
Add support for Vertex AI Vector Search as Vector Store (#13186 )
You can’t perform that action at this time.