@@ -40,8 +40,6 @@ public class LLM : MonoBehaviour
40
40
[ DynamicRange ( "minContextLength" , "maxContextLength" , false ) , Model ] public int contextSize = 8192 ;
41
41
/// <summary> Batch size for prompt processing. </summary>
42
42
[ ModelAdvanced ] public int batchSize = 512 ;
43
- /// <summary> a base prompt to use as a base for all LLMCaller objects </summary>
44
- [ TextArea ( 5 , 10 ) , ChatAdvanced ] public string basePrompt = "" ;
45
43
/// <summary> Boolean set to true if the server has started and is ready to receive requests, false otherwise. </summary>
46
44
public bool started { get ; protected set ; } = false ;
47
45
/// <summary> Boolean set to true if the server has failed to start. </summary>
@@ -132,7 +130,6 @@ public async void Awake()
132
130
await Task . Run ( ( ) => StartLLMServer ( arguments ) ) ;
133
131
if ( ! started ) return ;
134
132
if ( dontDestroyOnLoad ) DontDestroyOnLoad ( transform . root . gameObject ) ;
135
- if ( basePrompt != "" ) await SetBasePrompt ( basePrompt ) ;
136
133
}
137
134
138
135
/// <summary>
@@ -778,13 +775,6 @@ public async Task<string> Completion(string json, Callback<string> streamCallbac
778
775
return result ;
779
776
}
780
777
781
- public async Task SetBasePrompt ( string base_prompt )
782
- {
783
- AssertStarted ( ) ;
784
- SystemPromptRequest request = new SystemPromptRequest ( ) { system_prompt = base_prompt , prompt = " " , n_predict = 0 } ;
785
- await Completion ( JsonUtility . ToJson ( request ) ) ;
786
- }
787
-
788
778
/// <summary>
789
779
/// Allows to cancel the requests in a specific slot of the LLM
790
780
/// </summary>
0 commit comments