diff --git a/js/sdk/package-lock.json b/js/sdk/package-lock.json index 0ca55fd42..97dc6ced0 100644 --- a/js/sdk/package-lock.json +++ b/js/sdk/package-lock.json @@ -1,6 +1,6 @@ { "name": "r2r-js", - "version": "0.4.14", + "version": "0.4.17", "lockfileVersion": 3, "requires": true, "packages": { diff --git a/js/sdk/package.json b/js/sdk/package.json index 354655f76..839266503 100644 --- a/js/sdk/package.json +++ b/js/sdk/package.json @@ -1,6 +1,6 @@ { "name": "r2r-js", - "version": "0.4.16", + "version": "0.4.17", "description": "", "main": "dist/index.js", "browser": "dist/index.browser.js", diff --git a/py/core/configs/full_lm_studio.toml b/py/core/configs/full_lm_studio.toml index 344aa1624..7cd7024a3 100644 --- a/py/core/configs/full_lm_studio.toml +++ b/py/core/configs/full_lm_studio.toml @@ -31,6 +31,7 @@ provider = "postgres" max_knowledge_relationships = 100 max_description_input_length = 65536 generation_config = { model = "lm_studio/llama-3.2-3b-instruct" } # and other params, model used for relationshipt extraction + automatic_deduplication = false [database.graph_enrichment_settings] community_reports_prompt = "graphrag_community_reports" @@ -60,6 +61,7 @@ combine_under_n_chars = 128 overlap = 20 chunks_for_document_summary = 16 document_summary_model = "lm_studio/llama-3.2-3b-instruct" +automatic_extraction = false [orchestration] provider = "hatchet" diff --git a/py/core/configs/full_local_llm.toml b/py/core/configs/full_local_llm.toml index 2f312296a..7570a3ac2 100644 --- a/py/core/configs/full_local_llm.toml +++ b/py/core/configs/full_local_llm.toml @@ -31,6 +31,7 @@ provider = "postgres" max_knowledge_relationships = 100 max_description_input_length = 65536 generation_config = { model = "ollama/llama3.1" } # and other params, model used for relationshipt extraction + automatic_deduplication = false [database.graph_enrichment_settings] community_reports_prompt = "graphrag_community_reports" @@ -60,6 +61,7 @@ combine_under_n_chars = 128 overlap = 20 chunks_for_document_summary = 16 document_summary_model = "ollama/llama3.1" +automatic_extraction = false [orchestration] provider = "hatchet" diff --git a/py/core/configs/full_ollama.toml b/py/core/configs/full_ollama.toml index 2f312296a..7570a3ac2 100644 --- a/py/core/configs/full_ollama.toml +++ b/py/core/configs/full_ollama.toml @@ -31,6 +31,7 @@ provider = "postgres" max_knowledge_relationships = 100 max_description_input_length = 65536 generation_config = { model = "ollama/llama3.1" } # and other params, model used for relationshipt extraction + automatic_deduplication = false [database.graph_enrichment_settings] community_reports_prompt = "graphrag_community_reports" @@ -60,6 +61,7 @@ combine_under_n_chars = 128 overlap = 20 chunks_for_document_summary = 16 document_summary_model = "ollama/llama3.1" +automatic_extraction = false [orchestration] provider = "hatchet" diff --git a/py/core/configs/lm_studio.toml b/py/core/configs/lm_studio.toml index 096f46d1a..618e76839 100644 --- a/py/core/configs/lm_studio.toml +++ b/py/core/configs/lm_studio.toml @@ -37,6 +37,7 @@ provider = "postgres" max_knowledge_relationships = 100 max_description_input_length = 65536 generation_config = { model = "lm_studio/llama-3.2-3b-instruct" } # and other params, model used for relationshipt extraction + automatic_deduplication = false [database.graph_enrichment_settings] community_reports_prompt = "graphrag_community_reports" @@ -57,6 +58,7 @@ vision_img_model = "lm_studio/llama3.2-vision" vision_pdf_model = "lm_studio/llama3.2-vision" chunks_for_document_summary = 16 document_summary_model = "lm_studio/llama-3.2-3b-instruct" +automatic_extraction = false [ingestion.extra_parsers] pdf = "zerox" diff --git a/py/core/configs/local_llm.toml b/py/core/configs/local_llm.toml index 00ec18e94..59ec04e23 100644 --- a/py/core/configs/local_llm.toml +++ b/py/core/configs/local_llm.toml @@ -37,6 +37,7 @@ provider = "postgres" max_knowledge_relationships = 100 max_description_input_length = 65536 generation_config = { model = "ollama/llama3.1" } # and other params, model used for relationshipt extraction + automatic_deduplication = false [database.graph_enrichment_settings] community_reports_prompt = "graphrag_community_reports" @@ -57,6 +58,7 @@ vision_img_model = "ollama/llama3.2-vision" vision_pdf_model = "ollama/llama3.2-vision" chunks_for_document_summary = 16 document_summary_model = "ollama/llama3.1" +automatic_extraction = false [ingestion.extra_parsers] pdf = "zerox" diff --git a/py/core/configs/ollama.toml b/py/core/configs/ollama.toml index 00ec18e94..59ec04e23 100644 --- a/py/core/configs/ollama.toml +++ b/py/core/configs/ollama.toml @@ -37,6 +37,7 @@ provider = "postgres" max_knowledge_relationships = 100 max_description_input_length = 65536 generation_config = { model = "ollama/llama3.1" } # and other params, model used for relationshipt extraction + automatic_deduplication = false [database.graph_enrichment_settings] community_reports_prompt = "graphrag_community_reports" @@ -57,6 +58,7 @@ vision_img_model = "ollama/llama3.2-vision" vision_pdf_model = "ollama/llama3.2-vision" chunks_for_document_summary = 16 document_summary_model = "ollama/llama3.1" +automatic_extraction = false [ingestion.extra_parsers] pdf = "zerox"