From d8bee71dac50b8cd6f324c9e42321c47154b1b54 Mon Sep 17 00:00:00 2001 From: Logan Date: Tue, 27 Aug 2024 20:40:08 -0600 Subject: [PATCH] v0.11.2 (#15682) --- CHANGELOG.md | 59 ++++++++++++++++++ docs/docs/CHANGELOG.md | 60 +++++++++++++++++++ .../api_reference/embeddings/xinference.md | 4 ++ .../postprocessor/xinference_rerank.md | 4 ++ docs/mkdocs.yml | 6 ++ llama-index-core/llama_index/core/__init__.py | 2 +- llama-index-core/pyproject.toml | 2 +- poetry.lock | 26 ++++---- pyproject.toml | 4 +- 9 files changed, 148 insertions(+), 19 deletions(-) create mode 100644 docs/docs/api_reference/embeddings/xinference.md create mode 100644 docs/docs/api_reference/postprocessor/xinference_rerank.md diff --git a/CHANGELOG.md b/CHANGELOG.md index 39e082282f7bf..276ea5661ecd2 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,64 @@ # ChangeLog +## [2024-07-27] + +### `llama-index-core` [0.11.2] + +- fix tool schemas generation for pydantic v2 to handle nested models (#15679) +- feat: support default values for nested workflows (#15660) +- feat: allow FunctionTool with just an async fn (#15638) +- feat: Allow streaming events from steps (#15488) +- fix auto-retriever pydantic indent error (#15648) +- Implement Router Query Engine example using workflows (#15635) +- Add multi step query engine example using workflows (#15438) +- start traces for llm-level operations (#15542) +- Pass callback_manager to init in CodeSplitter from_defaults (#15585) + +### `llama-index-embeddings-xinference` [0.1.0] + +- Add Xinference Embedding Class (#15579) + +### `llama-index-llms-ai21` [0.3.3] + +- Integrations: AI21 function calling Support (#15622) + +### `llama-index-llms-anthropic` [0.3.0] + +- Added support for anthropic models through GCP Vertex AI (#15661) + +### `llama-index-llms-cerebras` [0.1.0] + +- Implement Cerebras Integration (#15665) + +### `llama-index-postprocessor-nvidia-rerank` [0.3.1] + +- fix downloaded nim endpoint path (#15645) +- fix llama-index-postprocessor-nvidia-rerank tests (#15643) + +### `llama-index-postprocessor-xinference-rerank` [0.1.0] + +- add xinference rerank class (#15639) + +### `llama-index-vector-stores-alibabacloud-opensearch` [0.2.1] + +- fix set output fields in AlibabaCloudOpenSearchConfig (#15562) + +### `llama-index-vector-stores-azureaisearch` [0.2.1] + +- Upgrade azure-search-documents to 2024-07-01 GA API and Add Support for Scalar and Binary Quantization in Index Creation (#15650) + +### `llama-index-vector-stores-neo4j` [0.2.1] + +- Neo4j Vector Store: Make Embedding Dimension Check Optional (#15628) + +### `llama-inde-vector-stores-milvus` [0.2.1] + +- Change the default consistency level of Milvus (#15577) + +### `llama-index-vector-stores-elasticsearch` [0.3.2] + +- Fix the ElasticsearchStore key error (#15631) + ## [2024-08-23] ### `llama-index-core` [0.11.1] diff --git a/docs/docs/CHANGELOG.md b/docs/docs/CHANGELOG.md index 39e082282f7bf..5c7a3030fab15 100644 --- a/docs/docs/CHANGELOG.md +++ b/docs/docs/CHANGELOG.md @@ -1,5 +1,65 @@ # ChangeLog +## [2024-07-27] + +### `llama-index-core` [0.11.2] + +- fix tool schemas generation for pydantic v2 to handle nested models (#15679) +- feat: support default values for nested workflows (#15660) +- feat: allow FunctionTool with just an async fn (#15638) +- feat: Allow streaming events from steps (#15488) +- fix auto-retriever pydantic indent error (#15648) +- Implement Router Query Engine example using workflows (#15635) +- Add multi step query engine example using workflows (#15438) +- start traces for llm-level operations (#15542) +- Pass callback_manager to init in CodeSplitter from_defaults (#15585) + +### `llama-index-embeddings-xinference` [0.1.0] + +- Add Xinference Embedding Class (#15579) + +### `llama-index-llms-ai21` [0.3.3] + +- Integrations: AI21 function calling Support (#15622) + +### `llama-index-llms-anthropic` [0.3.0] + +- Added support for anthropic models through GCP Vertex AI (#15661) + +### `llama-index-llms-cerebras` [0.1.0] + +- Implement Cerebras Integration (#15665) + +### `llama-index-postprocessor-nvidia-rerank` [0.3.1] + +- fix downloaded nim endpoint path (#15645) +- fix llama-index-postprocessor-nvidia-rerank tests (#15643) + +### `llama-index-postprocessor-xinference-rerank` [0.1.0] + +- add xinference rerank class (#15639) + +### `llama-index-vector-stores-alibabacloud-opensearch` [0.2.1] + +- fix set output fields in AlibabaCloudOpenSearchConfig (#15562) + +### `llama-index-vector-stores-azureaisearch` [0.2.1] + +- Upgrade azure-search-documents to 2024-07-01 GA API and Add Support for Scalar and Binary Quantization in Index Creation (#15650) + +### `llama-index-vector-stores-neo4j` [0.2.1] + +- Neo4j Vector Store: Make Embedding Dimension Check Optional (#15628) + +### `llama-inde-vector-stores-milvus` [0.2.1] + +- Change the default consistency level of Milvus (#15577) + +### `llama-index-vector-stores-elasticsearch` [0.3.2] + +- Fix the ElasticsearchStore key error (#15631) + + ## [2024-08-23] ### `llama-index-core` [0.11.1] diff --git a/docs/docs/api_reference/embeddings/xinference.md b/docs/docs/api_reference/embeddings/xinference.md new file mode 100644 index 0000000000000..69b2d3ef385cd --- /dev/null +++ b/docs/docs/api_reference/embeddings/xinference.md @@ -0,0 +1,4 @@ +::: llama_index.embeddings.xinference + options: + members: + - XinferenceEmbedding diff --git a/docs/docs/api_reference/postprocessor/xinference_rerank.md b/docs/docs/api_reference/postprocessor/xinference_rerank.md new file mode 100644 index 0000000000000..11d71202e448f --- /dev/null +++ b/docs/docs/api_reference/postprocessor/xinference_rerank.md @@ -0,0 +1,4 @@ +::: llama_index.postprocessor.xinference_rerank + options: + members: + - XinferenceRerank diff --git a/docs/mkdocs.yml b/docs/mkdocs.yml index c1fa62ca1a54f..b29812ca011a7 100644 --- a/docs/mkdocs.yml +++ b/docs/mkdocs.yml @@ -657,11 +657,13 @@ nav: - ./examples/workflow/corrective_rag_pack.ipynb - ./examples/workflow/function_calling_agent.ipynb - ./examples/workflow/long_rag_pack.ipynb + - ./examples/workflow/multi_step_query_engine.ipynb - ./examples/workflow/multi_strategy_workflow.ipynb - ./examples/workflow/parallel_execution.ipynb - ./examples/workflow/rag.ipynb - ./examples/workflow/react_agent.ipynb - ./examples/workflow/reflection.ipynb + - ./examples/workflow/router_query_engine.ipynb - ./examples/workflow/self_discover_workflow.ipynb - ./examples/workflow/sub_question_query_engine.ipynb - ./examples/workflow/workflows_cookbook.ipynb @@ -867,6 +869,7 @@ nav: - ./api_reference/embeddings/upstage.md - ./api_reference/embeddings/vertex.md - ./api_reference/embeddings/voyageai.md + - ./api_reference/embeddings/xinference.md - ./api_reference/embeddings/yandexgpt.md - Evaluation: - ./api_reference/evaluation/answer_relevancy.md @@ -1131,6 +1134,7 @@ nav: - ./api_reference/postprocessor/tei_rerank.md - ./api_reference/postprocessor/time_weighted.md - ./api_reference/postprocessor/voyageai_rerank.md + - ./api_reference/postprocessor/xinference_rerank.md - Object Stores: - ./api_reference/objects/index.md - Output Parsers: @@ -2138,6 +2142,8 @@ plugins: - ../llama-index-integrations/llms/llama-index-llms-sambanova - ../llama-index-integrations/embeddings/llama-index-embeddings-gigachat - ../llama-index-integrations/llms/llama-index-llms-gigachat + - ../llama-index-integrations/embeddings/llama-index-embeddings-xinference + - ../llama-index-integrations/postprocessor/llama-index-postprocessor-xinference-rerank - redirects: redirect_maps: ./api/llama_index.vector_stores.MongoDBAtlasVectorSearch.html: api_reference/storage/vector_store/mongodb.md diff --git a/llama-index-core/llama_index/core/__init__.py b/llama-index-core/llama_index/core/__init__.py index 0da056dcabd41..61bc90bb2011b 100644 --- a/llama-index-core/llama_index/core/__init__.py +++ b/llama-index-core/llama_index/core/__init__.py @@ -1,6 +1,6 @@ """Init file of LlamaIndex.""" -__version__ = "0.11.1" +__version__ = "0.11.2" import logging from logging import NullHandler diff --git a/llama-index-core/pyproject.toml b/llama-index-core/pyproject.toml index b41575c42e182..98c114448cba1 100644 --- a/llama-index-core/pyproject.toml +++ b/llama-index-core/pyproject.toml @@ -43,7 +43,7 @@ name = "llama-index-core" packages = [{include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.11.1" +version = "0.11.2" [tool.poetry.dependencies] SQLAlchemy = {extras = ["asyncio"], version = ">=1.4.49"} diff --git a/poetry.lock b/poetry.lock index da26d4df56663..d859daa3e357c 100644 --- a/poetry.lock +++ b/poetry.lock @@ -1,4 +1,4 @@ -# This file is automatically @generated by Poetry 1.8.3 and should not be changed by hand. +# This file is automatically @generated by Poetry 1.5.1 and should not be changed by hand. [[package]] name = "aiohappyeyeballs" @@ -1145,13 +1145,13 @@ trio = ["trio (>=0.22.0,<0.26.0)"] [[package]] name = "httpx" -version = "0.27.0" +version = "0.27.2" description = "The next generation HTTP client." optional = false python-versions = ">=3.8" files = [ - {file = "httpx-0.27.0-py3-none-any.whl", hash = "sha256:71d5465162c13681bff01ad59b2cc68dd838ea1f10e51574bac27103f00c91a5"}, - {file = "httpx-0.27.0.tar.gz", hash = "sha256:a0cb88a46f32dc874e04ee956e4c2764aba2aa228f650b06788ba6bda2962ab5"}, + {file = "httpx-0.27.2-py3-none-any.whl", hash = "sha256:7bb2708e112d8fdd7829cd4243970f0c223274051cb35ee80c03301ee29a3df0"}, + {file = "httpx-0.27.2.tar.gz", hash = "sha256:f7c2be1d2f3c3c3160d441802406b206c2b76f5947b11115e6df10c6c65e66c2"}, ] [package.dependencies] @@ -1166,6 +1166,7 @@ brotli = ["brotli", "brotlicffi"] cli = ["click (==8.*)", "pygments (==2.*)", "rich (>=10,<14)"] http2 = ["h2 (>=3,<5)"] socks = ["socksio (==1.*)"] +zstd = ["zstandard (>=0.18.0)"] [[package]] name = "identify" @@ -1663,13 +1664,13 @@ llama-index-llms-openai = ">=0.2.0,<0.3.0" [[package]] name = "llama-index-core" -version = "0.11.1" +version = "0.11.2" description = "Interface between LLMs and your data" optional = false python-versions = "<4.0,>=3.8.1" files = [ - {file = "llama_index_core-0.11.1-py3-none-any.whl", hash = "sha256:7d86c72be272c13f2c8fd5c62f27aba08a38e23cc5d28e02509d69909859102d"}, - {file = "llama_index_core-0.11.1.tar.gz", hash = "sha256:5c8b5cb1313866ff38ce91bea53a72519a52c9dabc0ca9d5c8b63aa8d3b2dd23"}, + {file = "llama_index_core-0.11.2-py3-none-any.whl", hash = "sha256:6c55667c4943ba197199e21e9b0e4641449f5e5dca662b0c91f5306f8c114e4f"}, + {file = "llama_index_core-0.11.2.tar.gz", hash = "sha256:eec37976fe3b1baa3bb31bd3c5f6ea821555c7065ac6a55b71b5601a7e097977"}, ] [package.dependencies] @@ -2228,7 +2229,6 @@ optional = false python-versions = ">=3.6" files = [ {file = "mkdocs-redirects-1.2.1.tar.gz", hash = "sha256:9420066d70e2a6bb357adf86e67023dcdca1857f97f07c7fe450f8f1fb42f861"}, - {file = "mkdocs_redirects-1.2.1-py3-none-any.whl", hash = "sha256:497089f9e0219e7389304cffefccdfa1cac5ff9509f2cb706f4c9b221726dffb"}, ] [package.dependencies] @@ -2713,7 +2713,7 @@ files = [ [package.dependencies] numpy = [ {version = ">=1.20.3", markers = "python_version < \"3.10\""}, - {version = ">=1.21.0", markers = "python_version >= \"3.10\" and python_version < \"3.11\""}, + {version = ">=1.21.0", markers = "python_version >= \"3.10\""}, {version = ">=1.23.2", markers = "python_version >= \"3.11\""}, ] python-dateutil = ">=2.8.2" @@ -3875,17 +3875,13 @@ files = [ {file = "SQLAlchemy-2.0.32-cp312-cp312-win32.whl", hash = "sha256:31983018b74908ebc6c996a16ad3690301a23befb643093fcfe85efd292e384d"}, {file = "SQLAlchemy-2.0.32-cp312-cp312-win_amd64.whl", hash = "sha256:4363ed245a6231f2e2957cccdda3c776265a75851f4753c60f3004b90e69bfeb"}, {file = "SQLAlchemy-2.0.32-cp37-cp37m-macosx_10_9_x86_64.whl", hash = "sha256:b8afd5b26570bf41c35c0121801479958b4446751a3971fb9a480c1afd85558e"}, - {file = "SQLAlchemy-2.0.32-cp37-cp37m-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c750987fc876813f27b60d619b987b057eb4896b81117f73bb8d9918c14f1cad"}, {file = "SQLAlchemy-2.0.32-cp37-cp37m-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ada0102afff4890f651ed91120c1120065663506b760da4e7823913ebd3258be"}, - {file = "SQLAlchemy-2.0.32-cp37-cp37m-musllinux_1_2_aarch64.whl", hash = "sha256:78c03d0f8a5ab4f3034c0e8482cfcc415a3ec6193491cfa1c643ed707d476f16"}, {file = "SQLAlchemy-2.0.32-cp37-cp37m-musllinux_1_2_x86_64.whl", hash = "sha256:3bd1cae7519283ff525e64645ebd7a3e0283f3c038f461ecc1c7b040a0c932a1"}, {file = "SQLAlchemy-2.0.32-cp37-cp37m-win32.whl", hash = "sha256:01438ebcdc566d58c93af0171c74ec28efe6a29184b773e378a385e6215389da"}, {file = "SQLAlchemy-2.0.32-cp37-cp37m-win_amd64.whl", hash = "sha256:4979dc80fbbc9d2ef569e71e0896990bc94df2b9fdbd878290bd129b65ab579c"}, {file = "SQLAlchemy-2.0.32-cp38-cp38-macosx_10_9_x86_64.whl", hash = "sha256:6c742be912f57586ac43af38b3848f7688863a403dfb220193a882ea60e1ec3a"}, {file = "SQLAlchemy-2.0.32-cp38-cp38-macosx_11_0_arm64.whl", hash = "sha256:62e23d0ac103bcf1c5555b6c88c114089587bc64d048fef5bbdb58dfd26f96da"}, - {file = "SQLAlchemy-2.0.32-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:251f0d1108aab8ea7b9aadbd07fb47fb8e3a5838dde34aa95a3349876b5a1f1d"}, {file = "SQLAlchemy-2.0.32-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ef18a84e5116340e38eca3e7f9eeaaef62738891422e7c2a0b80feab165905f"}, - {file = "SQLAlchemy-2.0.32-cp38-cp38-musllinux_1_2_aarch64.whl", hash = "sha256:3eb6a97a1d39976f360b10ff208c73afb6a4de86dd2a6212ddf65c4a6a2347d5"}, {file = "SQLAlchemy-2.0.32-cp38-cp38-musllinux_1_2_x86_64.whl", hash = "sha256:0c1c9b673d21477cec17ab10bc4decb1322843ba35b481585facd88203754fc5"}, {file = "SQLAlchemy-2.0.32-cp38-cp38-win32.whl", hash = "sha256:c41a2b9ca80ee555decc605bd3c4520cc6fef9abde8fd66b1cf65126a6922d65"}, {file = "SQLAlchemy-2.0.32-cp38-cp38-win_amd64.whl", hash = "sha256:8a37e4d265033c897892279e8adf505c8b6b4075f2b40d77afb31f7185cd6ecd"}, @@ -3902,7 +3898,7 @@ files = [ ] [package.dependencies] -greenlet = {version = "!=0.4.17", optional = true, markers = "python_version < \"3.13\" and (platform_machine == \"aarch64\" or platform_machine == \"ppc64le\" or platform_machine == \"x86_64\" or platform_machine == \"amd64\" or platform_machine == \"AMD64\" or platform_machine == \"win32\" or platform_machine == \"WIN32\") or extra == \"asyncio\""} +greenlet = {version = "!=0.4.17", optional = true, markers = "python_version < \"3.13\" and (platform_machine == \"win32\" or platform_machine == \"WIN32\" or platform_machine == \"AMD64\" or platform_machine == \"amd64\" or platform_machine == \"x86_64\" or platform_machine == \"ppc64le\" or platform_machine == \"aarch64\") or extra == \"asyncio\""} typing-extensions = ">=4.6.0" [package.extras] @@ -4621,4 +4617,4 @@ type = ["pytest-mypy"] [metadata] lock-version = "2.0" python-versions = ">=3.8.1,<4.0" -content-hash = "d50a2e23d473f9acf7dfc9fcacdaae05dfc85a9eb6f473e0d2d79b201781659a" +content-hash = "73e0c44869cc23ac3e86022fdf8a81f3eb1fc2fa5643e4d1b0e2e11ddff32e98" diff --git a/pyproject.toml b/pyproject.toml index 4ff2362fb6a75..93d33c877848c 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -44,7 +44,7 @@ name = "llama-index" packages = [{from = "_llama-index", include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.11.1" +version = "0.11.2" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" @@ -57,7 +57,7 @@ llama-index-agent-openai = "^0.3.0" llama-index-readers-file = "^0.2.0" llama-index-readers-llama-parse = ">=0.2.0" llama-index-indices-managed-llama-cloud = ">=0.3.0" -llama-index-core = "^0.11.1" +llama-index-core = "^0.11.2" llama-index-multi-modal-llms-openai = "^0.2.0" llama-index-cli = "^0.3.0" nltk = ">3.8.1" # avoids a CVE, temp until next release, should be in llama-index-core