-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Expand file tree
/
Copy pathpyproject.toml
More file actions
59 lines (53 loc) · 2.26 KB
/
pyproject.toml
File metadata and controls
59 lines (53 loc) · 2.26 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
[build-system]
requires = ["hatchling"]
build-backend = "hatchling.build"
[project]
name = "fedrag"
version = "1.0.0"
description = ""
license = "Apache-2.0"
dependencies = [
"flwr[simulation]>=1.27.0",
"numpy>=1.26.0,<2",
"faiss-cpu>=1.10.0",
"tqdm>=4.67",
"sentence_transformers>=3.4.1",
]
[tool.hatch.build.targets.wheel]
packages = ["."]
[tool.flwr.app]
publisher = "dimitris"
[tool.flwr.app.components]
serverapp = "fedrag.server_app:app"
clientapp = "fedrag.client_app:app"
[tool.flwr.app.config]
# Possible question answer datasets: PubMedQA, BioASQ, MMLU, MedQA, MedMCQA
# you can pass them as follows: "pubmedqa|bioasq|mmlu|medqa|medmcqa"
# the vertical bar separates the different benchmark datasets
server-qa-datasets = "pubmedqa|bioasq"
# Force to only evaluate the first 10 questions from each benchmark
# dataset, if this value is not set, then evaluate all questions
server-qa-num = 10
# Pass corpus as "StatPearls|Textbooks", the corpus will be assigned
# to each client participating in the federation in a Round-Robin fashion
# The supported corpora are:
# PubMed (23.9M docs), StatPearls (9.3k docs), Textbooks (18 docs), Wikipedia (6.5M docs)
# you can pass them as follows: "pubmed|statpearls|textbooks|wikipedia"
# the vertical bar separates the different corpus
clients-corpus-names = "statpearls|textbooks"
# The value of the Reciprocal Rank Fusion. The value will be used
# to merge multiple documents together at the server-side
# If the value is not set then only the retrieval scores are considered
k-rrf = 60
# The k-nearest-neighbor (most relevant) documents to the given search query
# used by clients to retrieve top-k relevant documents and at the server
# to select the final top-k documents after the merge operation
k-nn = 8
# The LLM used by the server to execute the RAG query.
# We assign the Hugging Face name of the LLM,
# e.g., "HuggingFaceTB/SmolLM2-1.7B-Instruct" or "meta-llama/Llama-3.2-1B-Instruct"
# By default, we use Hugging Face's SmolLM2-1.7B model. For Llama models, please
# ensure that you have accepted the terms and conditions, and have been granted access to the model.
server-llm-hfpath = "HuggingFaceTB/SmolLM2-1.7B-Instruct"
# Whether the server should use a GPU to load the llm and run the inference
server-llm-use-gpu = "false"