You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* docsum refine mode promt update
Signed-off-by: Xinyao Wang <[email protected]>
* docsum vllm requirement update
Signed-off-by: Xinyao Wang <[email protected]>
* docsum add auto mode
Signed-off-by: Xinyao Wang <[email protected]>
* fix bug
Signed-off-by: Xinyao Wang <[email protected]>
* fix bug
Signed-off-by: Xinyao Wang <[email protected]>
* fix readme
Signed-off-by: Xinyao Wang <[email protected]>
* [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
* refine
Signed-off-by: Xinyao Wang <[email protected]>
* [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
---------
Signed-off-by: Xinyao Wang <[email protected]>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
#### 3.2.2 Long context summarization with "summary_type"
128
128
129
-
"summary_type" is set to be "stuff" by default, which will let LLM generate summary based on complete input text. In this case please carefully set `MAX_INPUT_TOKENS` and `MAX_TOTAL_TOKENS` according to your model and device memory, otherwise it may exceed LLM context limit and raise error when meet long context.
129
+
**summary_type=auto**
130
130
131
-
When deal with long context, you can set "summary_type" to one of "truncate", "map_reduce" and "refine" for better performance.
131
+
"summary_type" is set to be "auto" by default, in this mode we will check input token length, if it exceed `MAX_INPUT_TOKENS`, `summary_type` will automatically be set to `refine` mode, otherwise will be set to `stuff` mode.
132
+
133
+
**summary_type=stuff**
134
+
135
+
In this mode LLM generate summary based on complete input text. In this case please carefully set `MAX_INPUT_TOKENS` and `MAX_TOTAL_TOKENS` according to your model and device memory, otherwise it may exceed LLM context limit and raise error when meet long context.
#### 3.2.2 Long context summarization with "summary_type"
127
127
128
-
"summary_type" is set to be "stuff" by default, which will let LLM generate summary based on complete input text. In this case please carefully set `MAX_INPUT_TOKENS` and `MAX_TOTAL_TOKENS` according to your model and device memory, otherwise it may exceed LLM context limit and raise error when meet long context.
128
+
**summary_type=auto**
129
129
130
-
When deal with long context, you can set "summary_type" to one of "truncate", "map_reduce" and "refine" for better performance.
130
+
"summary_type" is set to be "auto" by default, in this mode we will check input token length, if it exceed `MAX_INPUT_TOKENS`, `summary_type` will automatically be set to `refine` mode, otherwise will be set to `stuff` mode.
131
+
132
+
**summary_type=stuff**
133
+
134
+
In this mode LLM generate summary based on complete input text. In this case please carefully set `MAX_INPUT_TOKENS` and `MAX_TOTAL_TOKENS` according to your model and device memory, otherwise it may exceed LLM context limit and raise error when meet long context.
0 commit comments