|
@@ -65,14 +65,15 @@
|
|
|
Task<ApiResponse> ChatWithFilesAsync(List<string> fileIds, string question, string model = "deepseek-chat", float temperature = 0.7f, int maxTokens = 4000);
|
|
Task<ApiResponse> ChatWithFilesAsync(List<string> fileIds, string question, string model = "deepseek-chat", float temperature = 0.7f, int maxTokens = 4000);
|
|
|
|
|
|
|
|
/// <summary>
|
|
/// <summary>
|
|
|
- /// 使用已上传的文件进行聊天
|
|
|
|
|
|
|
+ /// chat接口 - 直接提问(不使用文件上下文)
|
|
|
/// </summary>
|
|
/// </summary>
|
|
|
/// <param name="question">问题</param>
|
|
/// <param name="question">问题</param>
|
|
|
/// <param name="model">模型名称</param>
|
|
/// <param name="model">模型名称</param>
|
|
|
|
|
+ /// <param name="stream">是否流式输出</param>
|
|
|
/// <param name="temperature">温度参数</param>
|
|
/// <param name="temperature">温度参数</param>
|
|
|
/// <param name="maxTokens">最大token数</param>
|
|
/// <param name="maxTokens">最大token数</param>
|
|
|
/// <returns>聊天响应</returns>
|
|
/// <returns>聊天响应</returns>
|
|
|
- Task<ApiResponse> ChatAsync(string question, string model = "deepseek-chat", float temperature = 0.7f, int maxTokens = 4000);
|
|
|
|
|
|
|
+ Task<ApiResponse> ChatAsync(string question, bool stream = false, string model = "deepseek-chat", float temperature = 0.7f, int maxTokens = 4000);
|
|
|
|
|
|
|
|
/// <summary>
|
|
/// <summary>
|
|
|
/// 等待文件处理完成
|
|
/// 等待文件处理完成
|