2025-07-05 15:11:56 +08:00
|
|
|
|
using System.Collections.Concurrent;
|
2025-08-03 23:23:32 +08:00
|
|
|
|
using System.Diagnostics;
|
2025-06-25 17:12:09 +08:00
|
|
|
|
using System.Runtime.CompilerServices;
|
2025-07-05 15:11:56 +08:00
|
|
|
|
using System.Text;
|
2025-08-11 15:29:24 +08:00
|
|
|
|
using System.Text.Json;
|
2025-07-05 15:11:56 +08:00
|
|
|
|
using Microsoft.AspNetCore.Http;
|
2025-06-21 01:08:14 +08:00
|
|
|
|
using Microsoft.Extensions.DependencyInjection;
|
2025-07-05 15:11:56 +08:00
|
|
|
|
using Microsoft.Extensions.Logging;
|
|
|
|
|
|
using Newtonsoft.Json;
|
2025-12-11 01:17:31 +08:00
|
|
|
|
using Newtonsoft.Json.Linq;
|
2025-07-05 15:11:56 +08:00
|
|
|
|
using Newtonsoft.Json.Serialization;
|
2025-06-21 01:08:14 +08:00
|
|
|
|
using Volo.Abp.Domain.Services;
|
2025-07-17 23:10:26 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.AiGateWay;
|
2025-08-03 23:23:32 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.AiGateWay.Exceptions;
|
2025-12-25 23:25:54 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.Entities.Chat;
|
2025-06-27 22:13:26 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.Entities.Model;
|
2025-10-14 22:17:21 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.Shared.Consts;
|
2025-06-25 17:12:09 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.Shared.Dtos;
|
2025-10-11 15:25:43 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.Shared.Dtos.Anthropic;
|
2025-12-17 18:47:28 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.Shared.Dtos.Gemini;
|
2025-08-11 15:31:11 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.Shared.Dtos.OpenAi;
|
|
|
|
|
|
using Yi.Framework.AiHub.Domain.Shared.Dtos.OpenAi.Embeddings;
|
|
|
|
|
|
using Yi.Framework.AiHub.Domain.Shared.Dtos.OpenAi.Images;
|
2025-12-11 01:17:31 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.Shared.Dtos.OpenAi.Responses;
|
2025-10-14 22:17:21 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.Shared.Enums;
|
2025-12-11 17:16:21 +08:00
|
|
|
|
using Yi.Framework.AiHub.Domain.Shared.Extensions;
|
2025-08-03 23:23:32 +08:00
|
|
|
|
using Yi.Framework.Core.Extensions;
|
2025-06-25 17:12:09 +08:00
|
|
|
|
using Yi.Framework.SqlSugarCore.Abstractions;
|
2025-10-11 15:25:43 +08:00
|
|
|
|
using JsonSerializer = System.Text.Json.JsonSerializer;
|
|
|
|
|
|
using ThorJsonSerializer = Yi.Framework.AiHub.Domain.AiGateWay.ThorJsonSerializer;
|
2025-06-21 01:08:14 +08:00
|
|
|
|
|
|
|
|
|
|
namespace Yi.Framework.AiHub.Domain.Managers;
|
|
|
|
|
|
|
|
|
|
|
|
public class AiGateWayManager : DomainService
|
|
|
|
|
|
{
|
2025-06-25 17:12:09 +08:00
|
|
|
|
private readonly ISqlSugarRepository<AiAppAggregateRoot> _aiAppRepository;
|
2025-10-14 22:17:21 +08:00
|
|
|
|
private readonly ISqlSugarRepository<AiModelEntity> _aiModelRepository;
|
2025-07-05 15:11:56 +08:00
|
|
|
|
private readonly ILogger<AiGateWayManager> _logger;
|
|
|
|
|
|
private readonly AiMessageManager _aiMessageManager;
|
|
|
|
|
|
private readonly UsageStatisticsManager _usageStatisticsManager;
|
2025-07-18 20:46:30 +08:00
|
|
|
|
private readonly ISpecialCompatible _specialCompatible;
|
2025-10-12 20:07:58 +08:00
|
|
|
|
private PremiumPackageManager? _premiumPackageManager;
|
2025-12-25 23:25:54 +08:00
|
|
|
|
private readonly ISqlSugarRepository<ImageStoreTaskAggregateRoot> _imageStoreTaskRepository;
|
2025-10-14 22:17:21 +08:00
|
|
|
|
|
2025-07-05 15:11:56 +08:00
|
|
|
|
public AiGateWayManager(ISqlSugarRepository<AiAppAggregateRoot> aiAppRepository, ILogger<AiGateWayManager> logger,
|
2025-07-18 20:46:30 +08:00
|
|
|
|
AiMessageManager aiMessageManager, UsageStatisticsManager usageStatisticsManager,
|
2025-12-25 23:25:54 +08:00
|
|
|
|
ISpecialCompatible specialCompatible, ISqlSugarRepository<AiModelEntity> aiModelRepository,
|
|
|
|
|
|
ISqlSugarRepository<ImageStoreTaskAggregateRoot> imageStoreTaskRepository)
|
2025-06-21 01:08:14 +08:00
|
|
|
|
{
|
2025-06-25 17:12:09 +08:00
|
|
|
|
_aiAppRepository = aiAppRepository;
|
2025-07-05 15:11:56 +08:00
|
|
|
|
_logger = logger;
|
|
|
|
|
|
_aiMessageManager = aiMessageManager;
|
|
|
|
|
|
_usageStatisticsManager = usageStatisticsManager;
|
2025-07-18 20:46:30 +08:00
|
|
|
|
_specialCompatible = specialCompatible;
|
2025-10-14 22:17:21 +08:00
|
|
|
|
_aiModelRepository = aiModelRepository;
|
2025-12-25 23:25:54 +08:00
|
|
|
|
_imageStoreTaskRepository = imageStoreTaskRepository;
|
2025-06-21 01:08:14 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-10-12 20:07:58 +08:00
|
|
|
|
private PremiumPackageManager PremiumPackageManager =>
|
|
|
|
|
|
_premiumPackageManager ??= LazyServiceProvider.LazyGetRequiredService<PremiumPackageManager>();
|
|
|
|
|
|
|
2025-06-25 17:12:09 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 获取模型
|
|
|
|
|
|
/// </summary>
|
2025-10-14 22:17:21 +08:00
|
|
|
|
/// <param name="modelApiType"></param>
|
2025-06-25 17:12:09 +08:00
|
|
|
|
/// <param name="modelId"></param>
|
|
|
|
|
|
/// <returns></returns>
|
2025-12-23 17:08:42 +08:00
|
|
|
|
public async Task<AiModelDescribe> GetModelAsync(ModelApiTypeEnum modelApiType, string modelId)
|
2025-06-21 01:08:14 +08:00
|
|
|
|
{
|
2025-10-14 22:17:21 +08:00
|
|
|
|
var aiModelDescribe = await _aiModelRepository._DbQueryable
|
|
|
|
|
|
.LeftJoin<AiAppAggregateRoot>((model, app) => model.AiAppId == app.Id)
|
|
|
|
|
|
.Where((model, app) => model.ModelId == modelId)
|
|
|
|
|
|
.Where((model, app) => model.ModelApiType == modelApiType)
|
2026-01-05 00:11:06 +08:00
|
|
|
|
.Where((model, app) => model.IsEnabled)
|
2025-10-14 22:17:21 +08:00
|
|
|
|
.Select((model, app) =>
|
|
|
|
|
|
new AiModelDescribe
|
2025-06-25 17:12:09 +08:00
|
|
|
|
{
|
|
|
|
|
|
AppId = app.Id,
|
|
|
|
|
|
AppName = app.Name,
|
|
|
|
|
|
Endpoint = app.Endpoint,
|
|
|
|
|
|
ApiKey = app.ApiKey,
|
|
|
|
|
|
OrderNum = model.OrderNum,
|
|
|
|
|
|
HandlerName = model.HandlerName,
|
|
|
|
|
|
ModelId = model.ModelId,
|
|
|
|
|
|
ModelName = model.Name,
|
2025-07-17 23:16:16 +08:00
|
|
|
|
Description = model.Description,
|
2025-07-17 23:52:00 +08:00
|
|
|
|
AppExtraUrl = app.ExtraUrl,
|
2025-11-25 09:54:13 +08:00
|
|
|
|
ModelExtraInfo = model.ExtraInfo,
|
2026-01-04 00:08:08 +08:00
|
|
|
|
Multiplier = model.Multiplier,
|
2026-01-04 12:32:31 +08:00
|
|
|
|
IsPremium = model.IsPremium,
|
|
|
|
|
|
ModelType = model.ModelType
|
2025-10-14 22:17:21 +08:00
|
|
|
|
})
|
|
|
|
|
|
.FirstAsync();
|
|
|
|
|
|
if (aiModelDescribe is null)
|
|
|
|
|
|
{
|
|
|
|
|
|
throw new UserFriendlyException($"【{modelId}】模型当前版本【{modelApiType}】格式不支持");
|
2025-06-21 01:08:14 +08:00
|
|
|
|
}
|
2026-01-04 00:08:08 +08:00
|
|
|
|
|
2026-01-03 03:19:31 +08:00
|
|
|
|
// ✅ 统一处理 yi- 后缀(网关层模型规范化)
|
2025-12-27 22:50:36 +08:00
|
|
|
|
if (!string.IsNullOrEmpty(aiModelDescribe.ModelId) &&
|
2025-12-27 23:44:45 +08:00
|
|
|
|
aiModelDescribe.ModelId.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
2025-12-27 22:50:36 +08:00
|
|
|
|
{
|
2025-12-27 23:49:35 +08:00
|
|
|
|
aiModelDescribe.ModelId = aiModelDescribe.ModelId[3..];
|
2025-12-27 22:50:36 +08:00
|
|
|
|
}
|
2026-01-04 00:08:08 +08:00
|
|
|
|
|
2025-10-14 22:17:21 +08:00
|
|
|
|
return aiModelDescribe;
|
2025-06-25 17:12:09 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-07-09 19:12:53 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 聊天完成-非流式
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="httpContext"></param>
|
2025-07-17 23:10:26 +08:00
|
|
|
|
/// <param name="request"></param>
|
2025-07-09 19:12:53 +08:00
|
|
|
|
/// <param name="userId"></param>
|
|
|
|
|
|
/// <param name="sessionId"></param>
|
2025-11-27 19:01:16 +08:00
|
|
|
|
/// <param name="tokenId">Token Id(Web端传null或Guid.Empty)</param>
|
2025-07-09 19:12:53 +08:00
|
|
|
|
/// <param name="cancellationToken"></param>
|
|
|
|
|
|
/// <returns></returns>
|
2025-07-17 23:10:26 +08:00
|
|
|
|
public async Task CompleteChatForStatisticsAsync(HttpContext httpContext,
|
|
|
|
|
|
ThorChatCompletionsRequest request,
|
2025-07-09 19:12:53 +08:00
|
|
|
|
Guid? userId = null,
|
|
|
|
|
|
Guid? sessionId = null,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
Guid? tokenId = null,
|
2025-07-09 19:12:53 +08:00
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
2025-07-18 20:46:30 +08:00
|
|
|
|
_specialCompatible.Compatible(request);
|
2025-07-09 19:12:53 +08:00
|
|
|
|
var response = httpContext.Response;
|
|
|
|
|
|
// 设置响应头,声明是 json
|
2025-07-18 23:12:20 +08:00
|
|
|
|
//response.ContentType = "application/json; charset=UTF-8";
|
2026-01-08 21:38:36 +08:00
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Completions, request.Model);
|
2025-07-17 23:16:16 +08:00
|
|
|
|
var chatService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<IChatCompletionService>(modelDescribe.HandlerName);
|
2026-01-04 00:08:08 +08:00
|
|
|
|
|
2026-01-02 00:57:30 +08:00
|
|
|
|
var sourceModelId = request.Model;
|
|
|
|
|
|
if (!string.IsNullOrEmpty(request.Model) &&
|
|
|
|
|
|
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
|
|
|
|
|
{
|
|
|
|
|
|
request.Model = request.Model[3..];
|
2026-01-04 00:08:08 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-07-17 23:10:26 +08:00
|
|
|
|
var data = await chatService.CompleteChatAsync(modelDescribe, request, cancellationToken);
|
2025-11-25 09:54:13 +08:00
|
|
|
|
data.SupplementalMultiplier(modelDescribe.Multiplier);
|
2025-07-09 19:12:53 +08:00
|
|
|
|
if (userId is not null)
|
|
|
|
|
|
{
|
|
|
|
|
|
await _aiMessageManager.CreateUserMessageAsync(userId.Value, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-10-15 19:49:33 +08:00
|
|
|
|
Content = sessionId is null ? "不予存储" : request.Messages?.LastOrDefault().Content ?? string.Empty,
|
2026-01-02 00:57:30 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2025-07-17 23:10:26 +08:00
|
|
|
|
TokenUsage = data.Usage,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
}, tokenId);
|
2025-07-09 19:12:53 +08:00
|
|
|
|
|
|
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId.Value, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-10-15 19:49:33 +08:00
|
|
|
|
Content =
|
|
|
|
|
|
sessionId is null ? "不予存储" : data.Choices?.FirstOrDefault()?.Delta.Content ?? string.Empty,
|
2026-01-02 00:57:30 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2025-07-17 23:10:26 +08:00
|
|
|
|
TokenUsage = data.Usage
|
2025-11-27 19:01:16 +08:00
|
|
|
|
}, tokenId);
|
2025-07-09 19:12:53 +08:00
|
|
|
|
|
2026-01-02 00:57:30 +08:00
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId.Value, sourceModelId, data.Usage, tokenId);
|
2025-10-28 17:43:23 +08:00
|
|
|
|
|
2026-01-04 00:08:08 +08:00
|
|
|
|
if (modelDescribe.IsPremium)
|
2025-10-28 17:43:23 +08:00
|
|
|
|
{
|
|
|
|
|
|
var totalTokens = data.Usage?.TotalTokens ?? 0;
|
|
|
|
|
|
if (totalTokens > 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-07-09 19:12:53 +08:00
|
|
|
|
}
|
2025-07-21 21:15:02 +08:00
|
|
|
|
|
2025-07-18 23:12:20 +08:00
|
|
|
|
await response.WriteAsJsonAsync(data, cancellationToken);
|
2025-07-09 19:12:53 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-07-05 15:11:56 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 聊天完成-缓存处理
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="httpContext"></param>
|
2025-07-17 23:10:26 +08:00
|
|
|
|
/// <param name="request"></param>
|
|
|
|
|
|
/// <param name="userId"></param>
|
2025-07-05 15:11:56 +08:00
|
|
|
|
/// <param name="sessionId"></param>
|
2025-11-27 19:01:16 +08:00
|
|
|
|
/// <param name="tokenId">Token Id(Web端传null或Guid.Empty)</param>
|
2025-07-05 15:11:56 +08:00
|
|
|
|
/// <param name="cancellationToken"></param>
|
|
|
|
|
|
/// <returns></returns>
|
2025-07-09 19:12:53 +08:00
|
|
|
|
public async Task CompleteChatStreamForStatisticsAsync(
|
2025-07-05 15:11:56 +08:00
|
|
|
|
HttpContext httpContext,
|
2025-07-17 23:10:26 +08:00
|
|
|
|
ThorChatCompletionsRequest request,
|
2025-07-05 15:11:56 +08:00
|
|
|
|
Guid? userId = null,
|
|
|
|
|
|
Guid? sessionId = null,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
Guid? tokenId = null,
|
2025-07-05 15:11:56 +08:00
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
|
|
|
|
|
var response = httpContext.Response;
|
|
|
|
|
|
// 设置响应头,声明是 SSE 流
|
2025-07-21 21:15:02 +08:00
|
|
|
|
response.ContentType = "text/event-stream;charset=utf-8;";
|
|
|
|
|
|
response.Headers.TryAdd("Cache-Control", "no-cache");
|
|
|
|
|
|
response.Headers.TryAdd("Connection", "keep-alive");
|
2025-07-05 15:11:56 +08:00
|
|
|
|
|
|
|
|
|
|
|
2025-12-11 17:16:21 +08:00
|
|
|
|
_specialCompatible.Compatible(request);
|
2026-01-08 21:38:36 +08:00
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Completions, request.Model);
|
2025-12-11 17:16:21 +08:00
|
|
|
|
var chatService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<IChatCompletionService>(modelDescribe.HandlerName);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2026-01-02 00:57:30 +08:00
|
|
|
|
var sourceModelId = request.Model;
|
|
|
|
|
|
if (!string.IsNullOrEmpty(request.Model) &&
|
|
|
|
|
|
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
|
|
|
|
|
{
|
|
|
|
|
|
request.Model = request.Model[3..];
|
2026-01-04 00:08:08 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-12-25 23:25:54 +08:00
|
|
|
|
var completeChatResponse = chatService.CompleteChatStreamAsync(modelDescribe, request, cancellationToken);
|
2025-07-17 23:10:26 +08:00
|
|
|
|
var tokenUsage = new ThorUsageResponse();
|
2025-07-05 15:11:56 +08:00
|
|
|
|
|
|
|
|
|
|
//缓存队列算法
|
|
|
|
|
|
// 创建一个队列来缓存消息
|
|
|
|
|
|
var messageQueue = new ConcurrentQueue<string>();
|
|
|
|
|
|
|
|
|
|
|
|
StringBuilder backupSystemContent = new StringBuilder();
|
|
|
|
|
|
// 设置输出速率(例如每50毫秒输出一次)
|
2025-07-08 18:24:21 +08:00
|
|
|
|
var outputInterval = TimeSpan.FromMilliseconds(75);
|
2025-07-05 15:11:56 +08:00
|
|
|
|
// 标记是否完成接收
|
|
|
|
|
|
var isComplete = false;
|
|
|
|
|
|
// 启动一个后台任务来消费队列
|
|
|
|
|
|
var outputTask = Task.Run(async () =>
|
|
|
|
|
|
{
|
|
|
|
|
|
while (!(isComplete && messageQueue.IsEmpty))
|
|
|
|
|
|
{
|
|
|
|
|
|
if (messageQueue.TryDequeue(out var message))
|
|
|
|
|
|
{
|
2025-07-21 21:15:02 +08:00
|
|
|
|
await response.WriteAsync(message, Encoding.UTF8, cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
await response.Body.FlushAsync(cancellationToken).ConfigureAwait(false);
|
2025-07-05 15:11:56 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
if (!isComplete)
|
|
|
|
|
|
{
|
|
|
|
|
|
// 如果没有完成,才等待,已完成,全部输出
|
2025-07-21 21:15:02 +08:00
|
|
|
|
await Task.Delay(outputInterval, cancellationToken).ConfigureAwait(false);
|
2025-07-05 15:11:56 +08:00
|
|
|
|
}
|
2025-08-21 21:50:06 +08:00
|
|
|
|
else
|
|
|
|
|
|
{
|
|
|
|
|
|
//已经完成了,也等待,但是速度可以放快
|
|
|
|
|
|
await Task.Delay(10, cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
}
|
2025-07-05 15:11:56 +08:00
|
|
|
|
}
|
|
|
|
|
|
}, cancellationToken);
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
//IAsyncEnumerable 只能在最外层捕获异常(如果你有其他办法的话...)
|
|
|
|
|
|
try
|
|
|
|
|
|
{
|
|
|
|
|
|
await foreach (var data in completeChatResponse)
|
|
|
|
|
|
{
|
2025-12-11 17:16:21 +08:00
|
|
|
|
data.SupplementalMultiplier(modelDescribe.Multiplier);
|
2025-10-28 17:43:23 +08:00
|
|
|
|
if (data.Usage is not null && (data.Usage.CompletionTokens > 0 || data.Usage.OutputTokens > 0))
|
2025-07-05 15:11:56 +08:00
|
|
|
|
{
|
2025-07-17 23:10:26 +08:00
|
|
|
|
tokenUsage = data.Usage;
|
2025-07-05 15:11:56 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-12-24 00:22:46 +08:00
|
|
|
|
var message = JsonSerializer.Serialize(data, ThorJsonSerializer.DefaultOptions);
|
2025-07-17 23:10:26 +08:00
|
|
|
|
backupSystemContent.Append(data.Choices.FirstOrDefault()?.Delta.Content);
|
2025-07-05 15:11:56 +08:00
|
|
|
|
// 将消息加入队列而不是直接写入
|
2025-07-21 21:15:02 +08:00
|
|
|
|
messageQueue.Enqueue($"data: {message}\n\n");
|
2025-07-05 15:11:56 +08:00
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
catch (Exception e)
|
|
|
|
|
|
{
|
|
|
|
|
|
_logger.LogError(e, $"Ai对话异常");
|
2025-12-12 21:14:38 +08:00
|
|
|
|
var errorContent = $"对话Ai异常,异常信息:\n当前Ai模型:{request.Model}\n异常信息:{e.Message}\n异常堆栈:{e}";
|
2025-07-17 23:10:26 +08:00
|
|
|
|
var model = new ThorChatCompletionsResponse()
|
|
|
|
|
|
{
|
|
|
|
|
|
Choices = new List<ThorChatChoiceResponse>()
|
|
|
|
|
|
{
|
|
|
|
|
|
new ThorChatChoiceResponse()
|
|
|
|
|
|
{
|
|
|
|
|
|
Delta = new ThorChatMessage()
|
|
|
|
|
|
{
|
|
|
|
|
|
Content = errorContent
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
};
|
2025-07-05 15:11:56 +08:00
|
|
|
|
var message = JsonConvert.SerializeObject(model, new JsonSerializerSettings
|
|
|
|
|
|
{
|
|
|
|
|
|
ContractResolver = new CamelCasePropertyNamesContractResolver()
|
|
|
|
|
|
});
|
|
|
|
|
|
backupSystemContent.Append(errorContent);
|
2025-07-21 21:15:02 +08:00
|
|
|
|
messageQueue.Enqueue($"data: {message}\n\n");
|
2025-07-05 15:11:56 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//断开连接
|
2025-07-21 21:15:02 +08:00
|
|
|
|
messageQueue.Enqueue("data: [DONE]\n\n");
|
2025-07-05 15:11:56 +08:00
|
|
|
|
// 标记完成并发送结束标记
|
|
|
|
|
|
isComplete = true;
|
|
|
|
|
|
|
|
|
|
|
|
await outputTask;
|
|
|
|
|
|
|
|
|
|
|
|
|
2025-08-03 21:32:54 +08:00
|
|
|
|
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-12-14 13:07:44 +08:00
|
|
|
|
Content = sessionId is null ? "不予存储" : request.Messages?.LastOrDefault()?.MessagesStore ?? string.Empty,
|
2026-01-02 00:57:30 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2025-08-03 21:32:54 +08:00
|
|
|
|
TokenUsage = tokenUsage,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
}, tokenId);
|
2025-07-05 15:11:56 +08:00
|
|
|
|
|
2025-08-03 21:32:54 +08:00
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-10-15 19:49:33 +08:00
|
|
|
|
Content = sessionId is null ? "不予存储" : backupSystemContent.ToString(),
|
2026-01-02 00:57:30 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2025-08-03 21:32:54 +08:00
|
|
|
|
TokenUsage = tokenUsage
|
2025-11-27 19:01:16 +08:00
|
|
|
|
}, tokenId);
|
2025-08-03 21:32:54 +08:00
|
|
|
|
|
2026-01-02 00:57:30 +08:00
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId, sourceModelId, tokenUsage, tokenId);
|
2025-10-14 22:17:21 +08:00
|
|
|
|
|
|
|
|
|
|
// 扣减尊享token包用量
|
2026-01-01 00:44:02 +08:00
|
|
|
|
if (userId is not null)
|
2025-10-14 22:17:21 +08:00
|
|
|
|
{
|
2026-01-04 00:08:08 +08:00
|
|
|
|
if (modelDescribe.IsPremium)
|
2025-10-14 22:17:21 +08:00
|
|
|
|
{
|
2026-01-01 00:44:02 +08:00
|
|
|
|
var totalTokens = tokenUsage.TotalTokens ?? 0;
|
|
|
|
|
|
if (totalTokens > 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
|
|
|
|
|
|
}
|
2025-10-14 22:17:21 +08:00
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-07-05 15:11:56 +08:00
|
|
|
|
}
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
|
|
|
|
|
|
2025-08-03 23:23:32 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 图片生成
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="context"></param>
|
|
|
|
|
|
/// <param name="userId"></param>
|
|
|
|
|
|
/// <param name="sessionId"></param>
|
|
|
|
|
|
/// <param name="request"></param>
|
2025-11-27 19:01:16 +08:00
|
|
|
|
/// <param name="tokenId">Token Id(Web端传null或Guid.Empty)</param>
|
2025-08-03 23:23:32 +08:00
|
|
|
|
/// <exception cref="BusinessException"></exception>
|
|
|
|
|
|
/// <exception cref="Exception"></exception>
|
2025-08-11 15:29:24 +08:00
|
|
|
|
public async Task CreateImageForStatisticsAsync(HttpContext context, Guid? userId, Guid? sessionId,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
ImageCreateRequest request, Guid? tokenId = null)
|
2025-08-03 23:23:32 +08:00
|
|
|
|
{
|
|
|
|
|
|
try
|
|
|
|
|
|
{
|
|
|
|
|
|
var model = request.Model;
|
|
|
|
|
|
if (string.IsNullOrEmpty(model)) model = "dall-e-2";
|
2025-08-11 15:29:24 +08:00
|
|
|
|
|
2026-01-08 21:38:36 +08:00
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Completions, model);
|
2025-08-11 15:29:24 +08:00
|
|
|
|
|
2025-08-03 23:23:32 +08:00
|
|
|
|
// 获取渠道指定的实现类型的服务
|
|
|
|
|
|
var imageService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<IImageService>(modelDescribe.HandlerName);
|
2025-08-11 15:29:24 +08:00
|
|
|
|
|
2025-08-03 23:23:32 +08:00
|
|
|
|
var response = await imageService.CreateImage(request, modelDescribe);
|
|
|
|
|
|
|
|
|
|
|
|
if (response.Error != null || response.Results.Count == 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
throw new BusinessException(response.Error?.Message ?? "图片生成失败", response.Error?.Code?.ToString());
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
await context.Response.WriteAsJsonAsync(response);
|
2025-08-11 15:29:24 +08:00
|
|
|
|
|
2025-08-03 23:23:32 +08:00
|
|
|
|
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-10-15 19:49:33 +08:00
|
|
|
|
Content = sessionId is null ? "不予存储" : request.Prompt,
|
2025-08-03 23:23:32 +08:00
|
|
|
|
ModelId = model,
|
|
|
|
|
|
TokenUsage = response.Usage,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
}, tokenId);
|
2025-08-03 23:23:32 +08:00
|
|
|
|
|
|
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-10-15 19:49:33 +08:00
|
|
|
|
Content = sessionId is null ? "不予存储" : response.Results?.FirstOrDefault()?.Url,
|
2025-08-03 23:23:32 +08:00
|
|
|
|
ModelId = model,
|
|
|
|
|
|
TokenUsage = response.Usage
|
2025-11-27 19:01:16 +08:00
|
|
|
|
}, tokenId);
|
2025-08-03 23:23:32 +08:00
|
|
|
|
|
2025-11-27 19:01:16 +08:00
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId, model, response.Usage, tokenId);
|
2025-10-15 19:49:33 +08:00
|
|
|
|
|
2026-01-04 00:08:08 +08:00
|
|
|
|
// 直接扣减尊享token包用量
|
2026-01-01 00:44:02 +08:00
|
|
|
|
if (userId is not null)
|
2025-10-14 22:17:21 +08:00
|
|
|
|
{
|
2026-01-04 00:08:08 +08:00
|
|
|
|
var totalTokens = response.Usage.TotalTokens ?? 0;
|
|
|
|
|
|
if (totalTokens > 0)
|
2025-10-14 22:17:21 +08:00
|
|
|
|
{
|
2026-01-04 00:08:08 +08:00
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
|
2025-10-14 22:17:21 +08:00
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-08-03 23:23:32 +08:00
|
|
|
|
}
|
|
|
|
|
|
catch (Exception e)
|
|
|
|
|
|
{
|
2025-12-12 21:14:38 +08:00
|
|
|
|
var errorContent = $"图片生成Ai异常,异常信息:\n当前Ai模型:{request.Model}\n异常信息:{e.Message}\n异常堆栈:{e}";
|
2025-08-03 23:23:32 +08:00
|
|
|
|
throw new UserFriendlyException(errorContent);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
|
|
|
|
|
|
2025-08-11 15:29:24 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 向量生成
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="context"></param>
|
2025-11-27 19:01:16 +08:00
|
|
|
|
/// <param name="userId"></param>
|
2025-08-11 15:29:24 +08:00
|
|
|
|
/// <param name="sessionId"></param>
|
|
|
|
|
|
/// <param name="input"></param>
|
2025-11-27 19:01:16 +08:00
|
|
|
|
/// <param name="tokenId">Token Id(Web端传null或Guid.Empty)</param>
|
2025-08-11 15:29:24 +08:00
|
|
|
|
/// <exception cref="Exception"></exception>
|
|
|
|
|
|
/// <exception cref="BusinessException"></exception>
|
|
|
|
|
|
public async Task EmbeddingForStatisticsAsync(HttpContext context, Guid? userId, Guid? sessionId,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
ThorEmbeddingInput input, Guid? tokenId = null)
|
2025-08-11 15:29:24 +08:00
|
|
|
|
{
|
|
|
|
|
|
try
|
|
|
|
|
|
{
|
|
|
|
|
|
if (input == null) throw new Exception("模型校验异常");
|
|
|
|
|
|
|
|
|
|
|
|
using var embedding =
|
|
|
|
|
|
Activity.Current?.Source.StartActivity("向量模型调用");
|
|
|
|
|
|
|
2026-01-08 21:38:36 +08:00
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Completions, input.Model);
|
2025-08-11 15:29:24 +08:00
|
|
|
|
|
|
|
|
|
|
// 获取渠道指定的实现类型的服务
|
|
|
|
|
|
var embeddingService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<ITextEmbeddingService>(modelDescribe.HandlerName);
|
|
|
|
|
|
|
|
|
|
|
|
var embeddingCreateRequest = new EmbeddingCreateRequest
|
|
|
|
|
|
{
|
|
|
|
|
|
Model = input.Model,
|
|
|
|
|
|
EncodingFormat = input.EncodingFormat
|
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
//dto进行转换,支持多种格式
|
|
|
|
|
|
if (input.Input is JsonElement str)
|
|
|
|
|
|
{
|
2025-08-11 18:10:11 +08:00
|
|
|
|
if (str.ValueKind == JsonValueKind.String)
|
|
|
|
|
|
{
|
|
|
|
|
|
embeddingCreateRequest.Input = str.ToString();
|
|
|
|
|
|
}
|
|
|
|
|
|
else if (str.ValueKind == JsonValueKind.Array)
|
2025-08-11 15:29:24 +08:00
|
|
|
|
{
|
|
|
|
|
|
var inputString = str.EnumerateArray().Select(x => x.ToString()).ToArray();
|
|
|
|
|
|
embeddingCreateRequest.InputAsList = inputString.ToList();
|
|
|
|
|
|
}
|
|
|
|
|
|
else
|
|
|
|
|
|
{
|
|
|
|
|
|
throw new Exception("Input,输入格式错误,非string或Array类型");
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-08-11 18:05:33 +08:00
|
|
|
|
else if (input.Input is string strInput)
|
|
|
|
|
|
{
|
|
|
|
|
|
embeddingCreateRequest.Input = strInput;
|
|
|
|
|
|
}
|
2025-08-11 15:29:24 +08:00
|
|
|
|
else
|
|
|
|
|
|
{
|
|
|
|
|
|
throw new Exception("Input,输入格式错误,未找到类型");
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
var stream =
|
|
|
|
|
|
await embeddingService.EmbeddingAsync(embeddingCreateRequest, modelDescribe, context.RequestAborted);
|
|
|
|
|
|
|
|
|
|
|
|
var usage = new ThorUsageResponse()
|
|
|
|
|
|
{
|
2025-10-11 15:25:43 +08:00
|
|
|
|
PromptTokens = stream.Usage?.PromptTokens ?? 0,
|
2025-08-11 15:29:24 +08:00
|
|
|
|
InputTokens = stream.Usage?.InputTokens ?? 0,
|
|
|
|
|
|
CompletionTokens = 0,
|
|
|
|
|
|
TotalTokens = stream.Usage?.InputTokens ?? 0
|
|
|
|
|
|
};
|
|
|
|
|
|
await context.Response.WriteAsJsonAsync(new
|
|
|
|
|
|
{
|
|
|
|
|
|
input.Model,
|
|
|
|
|
|
stream.Data,
|
|
|
|
|
|
stream.Error,
|
2025-08-11 20:24:48 +08:00
|
|
|
|
Object = stream.ObjectTypeName,
|
2025-08-11 15:29:24 +08:00
|
|
|
|
Usage = usage
|
|
|
|
|
|
});
|
|
|
|
|
|
|
2025-08-11 18:05:33 +08:00
|
|
|
|
//知识库暂不使用message统计
|
|
|
|
|
|
// await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
|
|
|
|
|
|
// new MessageInputDto
|
|
|
|
|
|
// {
|
|
|
|
|
|
// Content = string.Empty,
|
|
|
|
|
|
// ModelId = input.Model,
|
|
|
|
|
|
// TokenUsage = usage,
|
|
|
|
|
|
// });
|
|
|
|
|
|
//
|
|
|
|
|
|
// await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
|
|
|
|
|
|
// new MessageInputDto
|
|
|
|
|
|
// {
|
|
|
|
|
|
// Content = string.Empty,
|
|
|
|
|
|
// ModelId = input.Model,
|
|
|
|
|
|
// TokenUsage = usage
|
|
|
|
|
|
// });
|
2025-08-11 15:29:24 +08:00
|
|
|
|
|
2025-11-27 19:01:16 +08:00
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId, input.Model, usage, tokenId);
|
2025-08-11 15:29:24 +08:00
|
|
|
|
}
|
|
|
|
|
|
catch (ThorRateLimitException)
|
|
|
|
|
|
{
|
|
|
|
|
|
context.Response.StatusCode = 429;
|
|
|
|
|
|
}
|
|
|
|
|
|
catch (UnauthorizedAccessException e)
|
|
|
|
|
|
{
|
|
|
|
|
|
context.Response.StatusCode = 401;
|
|
|
|
|
|
}
|
|
|
|
|
|
catch (Exception e)
|
|
|
|
|
|
{
|
2025-12-12 21:14:38 +08:00
|
|
|
|
var errorContent = $"嵌入Ai异常,异常信息:\n当前Ai模型:{input.Model}\n异常信息:{e.Message}\n异常堆栈:{e}";
|
2025-08-11 15:29:24 +08:00
|
|
|
|
throw new UserFriendlyException(errorContent);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-10-11 15:25:43 +08:00
|
|
|
|
|
|
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// Anthropic聊天完成-非流式
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="httpContext"></param>
|
|
|
|
|
|
/// <param name="request"></param>
|
|
|
|
|
|
/// <param name="userId"></param>
|
|
|
|
|
|
/// <param name="sessionId"></param>
|
2025-11-27 19:01:16 +08:00
|
|
|
|
/// <param name="tokenId">Token Id(Web端传null或Guid.Empty)</param>
|
2025-10-11 15:25:43 +08:00
|
|
|
|
/// <param name="cancellationToken"></param>
|
|
|
|
|
|
/// <returns></returns>
|
|
|
|
|
|
public async Task AnthropicCompleteChatForStatisticsAsync(HttpContext httpContext,
|
|
|
|
|
|
AnthropicInput request,
|
|
|
|
|
|
Guid? userId = null,
|
|
|
|
|
|
Guid? sessionId = null,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
Guid? tokenId = null,
|
2025-10-11 15:25:43 +08:00
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
|
|
|
|
|
_specialCompatible.AnthropicCompatible(request);
|
|
|
|
|
|
var response = httpContext.Response;
|
|
|
|
|
|
// 设置响应头,声明是 json
|
|
|
|
|
|
//response.ContentType = "application/json; charset=UTF-8";
|
2026-01-08 21:38:36 +08:00
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Messages, request.Model);
|
2025-12-28 01:04:58 +08:00
|
|
|
|
|
|
|
|
|
|
var sourceModelId = request.Model;
|
2025-12-27 23:49:35 +08:00
|
|
|
|
if (!string.IsNullOrEmpty(request.Model) &&
|
|
|
|
|
|
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
|
|
|
|
|
{
|
|
|
|
|
|
request.Model = request.Model[3..];
|
|
|
|
|
|
}
|
2026-01-04 00:08:08 +08:00
|
|
|
|
|
2025-12-27 23:53:25 +08:00
|
|
|
|
var chatService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<IAnthropicChatCompletionService>(modelDescribe.HandlerName);
|
|
|
|
|
|
var data = await chatService.ChatCompletionsAsync(modelDescribe, request, cancellationToken);
|
2026-01-04 00:08:08 +08:00
|
|
|
|
|
2026-01-05 19:34:48 +08:00
|
|
|
|
var currentUsage = data.Usage;
|
|
|
|
|
|
ThorUsageResponse tokenUsage = new ThorUsageResponse
|
|
|
|
|
|
{
|
|
|
|
|
|
InputTokens = (currentUsage?.InputTokens??0) + (currentUsage?.CacheCreationInputTokens??0)+ (currentUsage?.CacheReadInputTokens??0),
|
|
|
|
|
|
OutputTokens = (currentUsage?.OutputTokens??0),
|
|
|
|
|
|
TotalTokens = (currentUsage?.InputTokens??0) + (currentUsage?.CacheCreationInputTokens??0)+ (currentUsage?.CacheReadInputTokens??0)+(currentUsage?.OutputTokens??0)
|
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
|
2025-12-11 01:17:31 +08:00
|
|
|
|
|
2025-10-11 15:25:43 +08:00
|
|
|
|
if (userId is not null)
|
|
|
|
|
|
{
|
|
|
|
|
|
await _aiMessageManager.CreateUserMessageAsync(userId.Value, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-12-11 17:33:12 +08:00
|
|
|
|
Content = "不予存储",
|
2025-12-28 01:04:58 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2026-01-05 19:34:48 +08:00
|
|
|
|
TokenUsage = tokenUsage,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
}, tokenId);
|
2025-10-11 15:25:43 +08:00
|
|
|
|
|
|
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId.Value, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-12-11 17:33:12 +08:00
|
|
|
|
Content = "不予存储",
|
2025-12-28 01:04:58 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2026-01-05 19:34:48 +08:00
|
|
|
|
TokenUsage = tokenUsage
|
2025-11-27 19:01:16 +08:00
|
|
|
|
}, tokenId);
|
2025-10-11 15:25:43 +08:00
|
|
|
|
|
2026-01-05 19:34:48 +08:00
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId.Value, sourceModelId, tokenUsage, tokenId);
|
2025-10-12 20:07:58 +08:00
|
|
|
|
|
2026-01-04 00:08:08 +08:00
|
|
|
|
// 直接扣减尊享token包用量
|
2026-01-05 19:34:48 +08:00
|
|
|
|
var totalTokens = tokenUsage.TotalTokens ?? 0;
|
2025-10-12 20:07:58 +08:00
|
|
|
|
if (totalTokens > 0)
|
|
|
|
|
|
{
|
2025-10-14 22:34:05 +08:00
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
|
2025-10-12 20:07:58 +08:00
|
|
|
|
}
|
2025-10-11 15:25:43 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
await response.WriteAsJsonAsync(data, cancellationToken);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-10-11 15:25:43 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// Anthropic聊天完成-缓存处理
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="httpContext"></param>
|
|
|
|
|
|
/// <param name="request"></param>
|
|
|
|
|
|
/// <param name="userId"></param>
|
|
|
|
|
|
/// <param name="sessionId"></param>
|
2025-11-27 19:01:16 +08:00
|
|
|
|
/// <param name="tokenId">Token Id(Web端传null或Guid.Empty)</param>
|
2025-10-11 15:25:43 +08:00
|
|
|
|
/// <param name="cancellationToken"></param>
|
|
|
|
|
|
/// <returns></returns>
|
|
|
|
|
|
public async Task AnthropicCompleteChatStreamForStatisticsAsync(
|
|
|
|
|
|
HttpContext httpContext,
|
|
|
|
|
|
AnthropicInput request,
|
|
|
|
|
|
Guid? userId = null,
|
|
|
|
|
|
Guid? sessionId = null,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
Guid? tokenId = null,
|
2025-10-11 15:25:43 +08:00
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
|
|
|
|
|
var response = httpContext.Response;
|
|
|
|
|
|
// 设置响应头,声明是 SSE 流
|
|
|
|
|
|
response.ContentType = "text/event-stream;charset=utf-8;";
|
|
|
|
|
|
response.Headers.TryAdd("Cache-Control", "no-cache");
|
|
|
|
|
|
response.Headers.TryAdd("Connection", "keep-alive");
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-12-11 17:16:21 +08:00
|
|
|
|
_specialCompatible.AnthropicCompatible(request);
|
2026-01-08 21:38:36 +08:00
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Messages, request.Model);
|
2025-12-11 17:16:21 +08:00
|
|
|
|
var chatService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<IAnthropicChatCompletionService>(modelDescribe.HandlerName);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-12-28 01:04:58 +08:00
|
|
|
|
var sourceModelId = request.Model;
|
2025-12-27 23:21:49 +08:00
|
|
|
|
if (!string.IsNullOrEmpty(request.Model) &&
|
2025-12-27 23:44:45 +08:00
|
|
|
|
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
2025-12-27 23:21:49 +08:00
|
|
|
|
{
|
2025-12-27 23:49:35 +08:00
|
|
|
|
request.Model = request.Model[3..];
|
2025-12-27 23:21:49 +08:00
|
|
|
|
}
|
2026-01-04 00:08:08 +08:00
|
|
|
|
|
2025-12-25 23:25:54 +08:00
|
|
|
|
var completeChatResponse = chatService.StreamChatCompletionsAsync(modelDescribe, request, cancellationToken);
|
2026-01-05 19:34:48 +08:00
|
|
|
|
ThorUsageResponse? tokenUsage = new ThorUsageResponse();
|
2025-10-11 15:25:43 +08:00
|
|
|
|
try
|
|
|
|
|
|
{
|
|
|
|
|
|
await foreach (var responseResult in completeChatResponse)
|
|
|
|
|
|
{
|
2026-01-05 19:34:48 +08:00
|
|
|
|
//部分供应商message_start放一部分
|
|
|
|
|
|
if (responseResult.Item1.Contains("message_start"))
|
2025-10-12 14:38:26 +08:00
|
|
|
|
{
|
2026-01-05 19:34:48 +08:00
|
|
|
|
var currentTokenUsage = responseResult.Item2.Message.Usage;
|
|
|
|
|
|
if ((currentTokenUsage.InputTokens ?? 0) != 0)
|
2026-01-05 15:44:48 +08:00
|
|
|
|
{
|
2026-01-05 19:34:48 +08:00
|
|
|
|
tokenUsage.InputTokens = (currentTokenUsage?.InputTokens??0) + (currentTokenUsage?.CacheCreationInputTokens??0)+ (currentTokenUsage?.CacheReadInputTokens??0);
|
|
|
|
|
|
}
|
|
|
|
|
|
if ((currentTokenUsage.OutputTokens ?? 0) != 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage.OutputTokens = currentTokenUsage.OutputTokens;
|
2026-01-05 15:44:48 +08:00
|
|
|
|
}
|
2025-10-12 14:38:26 +08:00
|
|
|
|
}
|
2025-10-14 22:17:21 +08:00
|
|
|
|
|
2026-01-05 19:34:48 +08:00
|
|
|
|
//message_delta又放一部分
|
|
|
|
|
|
if (responseResult.Item1.Contains("message_delta"))
|
|
|
|
|
|
{
|
|
|
|
|
|
var currentTokenUsage = responseResult.Item2.Usage;
|
|
|
|
|
|
|
|
|
|
|
|
if ((currentTokenUsage.InputTokens ?? 0) != 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage.InputTokens = (currentTokenUsage?.InputTokens??0) + (currentTokenUsage?.CacheCreationInputTokens??0)+ (currentTokenUsage?.CacheReadInputTokens??0);;
|
|
|
|
|
|
}
|
|
|
|
|
|
if ((currentTokenUsage.OutputTokens ?? 0) != 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage.OutputTokens = currentTokenUsage.OutputTokens;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
2025-10-11 15:25:43 +08:00
|
|
|
|
await WriteAsEventStreamDataAsync(httpContext, responseResult.Item1, responseResult.Item2,
|
|
|
|
|
|
cancellationToken);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
catch (Exception e)
|
|
|
|
|
|
{
|
|
|
|
|
|
_logger.LogError(e, $"Ai对话异常");
|
2025-12-28 01:04:58 +08:00
|
|
|
|
var errorContent = $"对话Ai异常,异常信息:\n当前Ai模型:{sourceModelId}\n异常信息:{e.Message}\n异常堆栈:{e}";
|
2025-10-12 16:42:26 +08:00
|
|
|
|
throw new UserFriendlyException(errorContent);
|
2025-10-11 15:25:43 +08:00
|
|
|
|
}
|
2026-01-05 19:34:48 +08:00
|
|
|
|
tokenUsage.TotalTokens = (tokenUsage.InputTokens ?? 0) + (tokenUsage.OutputTokens ?? 0);
|
|
|
|
|
|
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
|
2025-10-11 15:25:43 +08:00
|
|
|
|
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-12-11 17:33:12 +08:00
|
|
|
|
Content = "不予存储",
|
2025-12-28 01:04:58 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2025-10-11 15:25:43 +08:00
|
|
|
|
TokenUsage = tokenUsage,
|
2025-11-27 19:01:16 +08:00
|
|
|
|
}, tokenId);
|
2025-10-11 15:25:43 +08:00
|
|
|
|
|
|
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-12-11 17:33:12 +08:00
|
|
|
|
Content = "不予存储",
|
2025-12-28 01:04:58 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2025-10-11 15:25:43 +08:00
|
|
|
|
TokenUsage = tokenUsage
|
2025-11-27 19:01:16 +08:00
|
|
|
|
}, tokenId);
|
2025-10-11 15:25:43 +08:00
|
|
|
|
|
2025-12-28 01:04:58 +08:00
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId, sourceModelId, tokenUsage, tokenId);
|
2025-10-12 20:07:58 +08:00
|
|
|
|
|
2026-01-04 00:08:08 +08:00
|
|
|
|
// 直接扣减尊享token包用量
|
2025-10-12 20:07:58 +08:00
|
|
|
|
if (userId.HasValue && tokenUsage is not null)
|
|
|
|
|
|
{
|
2025-10-14 22:17:21 +08:00
|
|
|
|
var totalTokens = tokenUsage.TotalTokens ?? 0;
|
2025-10-15 19:49:33 +08:00
|
|
|
|
if (tokenUsage.TotalTokens > 0)
|
2025-10-12 20:07:58 +08:00
|
|
|
|
{
|
2025-10-15 19:49:33 +08:00
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
|
2025-10-12 20:07:58 +08:00
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-10-11 15:25:43 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-12-11 01:17:31 +08:00
|
|
|
|
|
|
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// OpenAi 响应-非流式-缓存处理
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="httpContext"></param>
|
|
|
|
|
|
/// <param name="request"></param>
|
|
|
|
|
|
/// <param name="userId"></param>
|
|
|
|
|
|
/// <param name="sessionId"></param>
|
|
|
|
|
|
/// <param name="tokenId"></param>
|
|
|
|
|
|
/// <param name="cancellationToken"></param>
|
|
|
|
|
|
public async Task OpenAiResponsesAsyncForStatisticsAsync(HttpContext httpContext,
|
|
|
|
|
|
OpenAiResponsesInput request,
|
|
|
|
|
|
Guid? userId = null,
|
|
|
|
|
|
Guid? sessionId = null,
|
|
|
|
|
|
Guid? tokenId = null,
|
|
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
|
|
|
|
|
// _specialCompatible.AnthropicCompatible(request);
|
|
|
|
|
|
var response = httpContext.Response;
|
|
|
|
|
|
// 设置响应头,声明是 json
|
|
|
|
|
|
//response.ContentType = "application/json; charset=UTF-8";
|
2026-01-08 21:38:36 +08:00
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Responses, request.Model);
|
2025-12-11 01:17:31 +08:00
|
|
|
|
|
|
|
|
|
|
var chatService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<IOpenAiResponseService>(modelDescribe.HandlerName);
|
2026-01-02 00:57:30 +08:00
|
|
|
|
var sourceModelId = request.Model;
|
|
|
|
|
|
if (!string.IsNullOrEmpty(request.Model) &&
|
|
|
|
|
|
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
|
|
|
|
|
{
|
|
|
|
|
|
request.Model = request.Model[3..];
|
|
|
|
|
|
}
|
2026-01-04 00:08:08 +08:00
|
|
|
|
|
2025-12-11 01:17:31 +08:00
|
|
|
|
var data = await chatService.ResponsesAsync(modelDescribe, request, cancellationToken);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-12-11 17:16:21 +08:00
|
|
|
|
data.SupplementalMultiplier(modelDescribe.Multiplier);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
|
|
|
|
|
var tokenUsage = new ThorUsageResponse
|
2025-12-11 01:17:31 +08:00
|
|
|
|
{
|
2025-12-11 17:16:21 +08:00
|
|
|
|
InputTokens = data.Usage.InputTokens,
|
|
|
|
|
|
OutputTokens = data.Usage.OutputTokens,
|
|
|
|
|
|
TotalTokens = data.Usage.InputTokens + data.Usage.OutputTokens,
|
|
|
|
|
|
};
|
|
|
|
|
|
if (userId is not null)
|
|
|
|
|
|
{
|
|
|
|
|
|
await _aiMessageManager.CreateUserMessageAsync(userId.Value, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
|
|
|
|
|
Content = "不予存储",
|
2026-01-02 00:57:30 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2025-12-11 17:16:21 +08:00
|
|
|
|
TokenUsage = tokenUsage,
|
|
|
|
|
|
}, tokenId);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-12-11 17:16:21 +08:00
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId.Value, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
|
|
|
|
|
Content = "不予存储",
|
2026-01-02 00:57:30 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2025-12-11 17:16:21 +08:00
|
|
|
|
TokenUsage = tokenUsage
|
|
|
|
|
|
}, tokenId);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2026-01-02 00:57:30 +08:00
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId.Value, sourceModelId, tokenUsage, tokenId);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2026-01-04 00:08:08 +08:00
|
|
|
|
// 直接扣减尊享token包用量
|
2025-12-11 17:16:21 +08:00
|
|
|
|
var totalTokens = tokenUsage.TotalTokens ?? 0;
|
|
|
|
|
|
if (totalTokens > 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
|
|
|
|
|
|
}
|
2025-12-11 01:17:31 +08:00
|
|
|
|
}
|
2025-12-11 17:16:21 +08:00
|
|
|
|
|
|
|
|
|
|
await response.WriteAsJsonAsync(data, cancellationToken);
|
2025-12-11 01:17:31 +08:00
|
|
|
|
}
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
|
|
|
|
|
|
2025-12-11 01:17:31 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// OpenAi响应-流式-缓存处理
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="httpContext"></param>
|
|
|
|
|
|
/// <param name="request"></param>
|
|
|
|
|
|
/// <param name="userId"></param>
|
|
|
|
|
|
/// <param name="sessionId"></param>
|
|
|
|
|
|
/// <param name="tokenId">Token Id(Web端传null或Guid.Empty)</param>
|
|
|
|
|
|
/// <param name="cancellationToken"></param>
|
|
|
|
|
|
/// <returns></returns>
|
|
|
|
|
|
public async Task OpenAiResponsesStreamForStatisticsAsync(
|
|
|
|
|
|
HttpContext httpContext,
|
|
|
|
|
|
OpenAiResponsesInput request,
|
|
|
|
|
|
Guid? userId = null,
|
|
|
|
|
|
Guid? sessionId = null,
|
|
|
|
|
|
Guid? tokenId = null,
|
|
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
|
|
|
|
|
var response = httpContext.Response;
|
|
|
|
|
|
// 设置响应头,声明是 SSE 流
|
|
|
|
|
|
response.ContentType = "text/event-stream;charset=utf-8;";
|
|
|
|
|
|
response.Headers.TryAdd("Cache-Control", "no-cache");
|
|
|
|
|
|
response.Headers.TryAdd("Connection", "keep-alive");
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2026-01-08 21:38:36 +08:00
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Responses, request.Model);
|
2025-12-11 17:16:21 +08:00
|
|
|
|
var chatService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<IOpenAiResponseService>(modelDescribe.HandlerName);
|
2026-01-02 00:57:30 +08:00
|
|
|
|
var sourceModelId = request.Model;
|
|
|
|
|
|
if (!string.IsNullOrEmpty(request.Model) &&
|
|
|
|
|
|
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
|
|
|
|
|
{
|
|
|
|
|
|
request.Model = request.Model[3..];
|
|
|
|
|
|
}
|
2026-01-04 00:08:08 +08:00
|
|
|
|
|
2025-12-25 23:25:54 +08:00
|
|
|
|
var completeChatResponse = chatService.ResponsesStreamAsync(modelDescribe, request, cancellationToken);
|
2025-12-11 01:17:31 +08:00
|
|
|
|
ThorUsageResponse? tokenUsage = null;
|
|
|
|
|
|
try
|
|
|
|
|
|
{
|
|
|
|
|
|
await foreach (var responseResult in completeChatResponse)
|
|
|
|
|
|
{
|
|
|
|
|
|
//message_start是为了保底机制
|
|
|
|
|
|
if (responseResult.Item1.Contains("response.completed"))
|
|
|
|
|
|
{
|
2025-12-11 17:16:21 +08:00
|
|
|
|
var obj = responseResult.Item2!.Value;
|
2025-12-25 23:25:54 +08:00
|
|
|
|
int inputTokens = obj.GetPath("response", "usage", "input_tokens").GetInt();
|
|
|
|
|
|
int outputTokens = obj.GetPath("response", "usage", "output_tokens").GetInt();
|
|
|
|
|
|
inputTokens = Convert.ToInt32(inputTokens * modelDescribe.Multiplier);
|
|
|
|
|
|
outputTokens = Convert.ToInt32(outputTokens * modelDescribe.Multiplier);
|
2025-12-11 01:17:31 +08:00
|
|
|
|
tokenUsage = new ThorUsageResponse
|
|
|
|
|
|
{
|
2025-12-25 23:25:54 +08:00
|
|
|
|
PromptTokens = inputTokens,
|
2025-12-11 01:17:31 +08:00
|
|
|
|
InputTokens = inputTokens,
|
|
|
|
|
|
OutputTokens = outputTokens,
|
|
|
|
|
|
CompletionTokens = outputTokens,
|
2025-12-25 23:25:54 +08:00
|
|
|
|
TotalTokens = inputTokens + outputTokens,
|
2025-12-11 01:17:31 +08:00
|
|
|
|
};
|
|
|
|
|
|
}
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-12-11 01:17:31 +08:00
|
|
|
|
await WriteAsEventStreamDataAsync(httpContext, responseResult.Item1, responseResult.Item2,
|
|
|
|
|
|
cancellationToken);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
catch (Exception e)
|
|
|
|
|
|
{
|
|
|
|
|
|
_logger.LogError(e, $"Ai响应异常");
|
2025-12-12 21:14:38 +08:00
|
|
|
|
var errorContent = $"响应Ai异常,异常信息:\n当前Ai模型:{request.Model}\n异常信息:{e.Message}\n异常堆栈:{e}";
|
2025-12-11 01:17:31 +08:00
|
|
|
|
throw new UserFriendlyException(errorContent);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-12-25 23:25:54 +08:00
|
|
|
|
Content = "不予存储",
|
2026-01-02 00:57:30 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2025-12-11 01:17:31 +08:00
|
|
|
|
TokenUsage = tokenUsage,
|
|
|
|
|
|
}, tokenId);
|
|
|
|
|
|
|
|
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-12-25 23:25:54 +08:00
|
|
|
|
Content = "不予存储",
|
2026-01-02 00:57:30 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2025-12-11 01:17:31 +08:00
|
|
|
|
TokenUsage = tokenUsage
|
|
|
|
|
|
}, tokenId);
|
|
|
|
|
|
|
2026-01-02 00:57:30 +08:00
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId, sourceModelId, tokenUsage, tokenId);
|
2025-12-11 01:17:31 +08:00
|
|
|
|
|
2026-01-04 00:08:08 +08:00
|
|
|
|
// 直接扣减尊享token包用量
|
2025-12-11 01:17:31 +08:00
|
|
|
|
if (userId.HasValue && tokenUsage is not null)
|
|
|
|
|
|
{
|
|
|
|
|
|
var totalTokens = tokenUsage.TotalTokens ?? 0;
|
|
|
|
|
|
if (tokenUsage.TotalTokens > 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-12-17 18:47:28 +08:00
|
|
|
|
|
|
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// Gemini 生成-非流式-缓存处理
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="httpContext"></param>
|
|
|
|
|
|
/// <param name="modelId"></param>
|
|
|
|
|
|
/// <param name="request"></param>
|
|
|
|
|
|
/// <param name="userId"></param>
|
|
|
|
|
|
/// <param name="sessionId"></param>
|
|
|
|
|
|
/// <param name="tokenId"></param>
|
|
|
|
|
|
/// <param name="cancellationToken"></param>
|
2025-12-17 21:51:01 +08:00
|
|
|
|
public async Task GeminiGenerateContentForStatisticsAsync(HttpContext httpContext,
|
2025-12-17 18:47:28 +08:00
|
|
|
|
string modelId,
|
|
|
|
|
|
JsonElement request,
|
|
|
|
|
|
Guid? userId = null,
|
|
|
|
|
|
Guid? sessionId = null,
|
|
|
|
|
|
Guid? tokenId = null,
|
|
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
|
|
|
|
|
var response = httpContext.Response;
|
|
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.GenerateContent, modelId);
|
|
|
|
|
|
|
|
|
|
|
|
var chatService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<IGeminiGenerateContentService>(modelDescribe.HandlerName);
|
|
|
|
|
|
var data = await chatService.GenerateContentAsync(modelDescribe, request, cancellationToken);
|
|
|
|
|
|
|
2025-12-25 23:25:54 +08:00
|
|
|
|
var tokenUsage = GeminiGenerateContentAcquirer.GetUsage(data);
|
2026-01-04 12:32:31 +08:00
|
|
|
|
//如果是图片模型,单独扣费
|
|
|
|
|
|
if (modelDescribe.ModelType == ModelTypeEnum.Image)
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage = new ThorUsageResponse
|
|
|
|
|
|
{
|
|
|
|
|
|
InputTokens = (int)modelDescribe.Multiplier,
|
|
|
|
|
|
OutputTokens = (int)modelDescribe.Multiplier,
|
|
|
|
|
|
TotalTokens = (int)modelDescribe.Multiplier
|
|
|
|
|
|
};
|
|
|
|
|
|
}
|
|
|
|
|
|
else
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
|
|
|
|
|
|
}
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-12-17 18:47:28 +08:00
|
|
|
|
if (userId is not null)
|
|
|
|
|
|
{
|
|
|
|
|
|
await _aiMessageManager.CreateUserMessageAsync(userId.Value, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
|
|
|
|
|
Content = "不予存储",
|
|
|
|
|
|
ModelId = modelId,
|
|
|
|
|
|
TokenUsage = tokenUsage,
|
|
|
|
|
|
}, tokenId);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-12-17 18:47:28 +08:00
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId.Value, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
|
|
|
|
|
Content = "不予存储",
|
|
|
|
|
|
ModelId = modelId,
|
|
|
|
|
|
TokenUsage = tokenUsage
|
|
|
|
|
|
}, tokenId);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-12-17 18:47:28 +08:00
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId.Value, modelId, tokenUsage, tokenId);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2026-01-04 00:08:08 +08:00
|
|
|
|
// 直接扣减尊享token包用量
|
2025-12-17 18:47:28 +08:00
|
|
|
|
var totalTokens = tokenUsage.TotalTokens ?? 0;
|
|
|
|
|
|
if (totalTokens > 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
await response.WriteAsJsonAsync(data, cancellationToken);
|
|
|
|
|
|
}
|
2025-12-17 21:51:01 +08:00
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// Gemini 生成-流式-缓存处理
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="httpContext"></param>
|
|
|
|
|
|
/// <param name="modelId"></param>
|
|
|
|
|
|
/// <param name="request"></param>
|
|
|
|
|
|
/// <param name="userId"></param>
|
|
|
|
|
|
/// <param name="sessionId"></param>
|
|
|
|
|
|
/// <param name="tokenId">Token Id(Web端传null或Guid.Empty)</param>
|
|
|
|
|
|
/// <param name="cancellationToken"></param>
|
|
|
|
|
|
/// <returns></returns>
|
|
|
|
|
|
public async Task GeminiGenerateContentStreamForStatisticsAsync(
|
|
|
|
|
|
HttpContext httpContext,
|
|
|
|
|
|
string modelId,
|
|
|
|
|
|
JsonElement request,
|
|
|
|
|
|
Guid? userId = null,
|
|
|
|
|
|
Guid? sessionId = null,
|
|
|
|
|
|
Guid? tokenId = null,
|
|
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
|
|
|
|
|
var response = httpContext.Response;
|
|
|
|
|
|
// 设置响应头,声明是 SSE 流
|
|
|
|
|
|
response.ContentType = "text/event-stream;charset=utf-8;";
|
|
|
|
|
|
response.Headers.TryAdd("Cache-Control", "no-cache");
|
|
|
|
|
|
response.Headers.TryAdd("Connection", "keep-alive");
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-12-17 21:51:01 +08:00
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.GenerateContent, modelId);
|
|
|
|
|
|
var chatService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<IGeminiGenerateContentService>(modelDescribe.HandlerName);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
|
|
|
|
|
var completeChatResponse = chatService.GenerateContentStreamAsync(modelDescribe, request, cancellationToken);
|
2025-12-17 21:51:01 +08:00
|
|
|
|
ThorUsageResponse? tokenUsage = null;
|
|
|
|
|
|
try
|
|
|
|
|
|
{
|
|
|
|
|
|
await foreach (var responseResult in completeChatResponse)
|
|
|
|
|
|
{
|
2025-12-25 23:25:54 +08:00
|
|
|
|
if (responseResult!.Value.GetPath("candidates", 0, "finishReason").GetString() == "STOP")
|
2025-12-17 21:51:01 +08:00
|
|
|
|
{
|
|
|
|
|
|
tokenUsage = GeminiGenerateContentAcquirer.GetUsage(responseResult!.Value);
|
2026-01-04 12:32:31 +08:00
|
|
|
|
//如果是图片模型,单独扣费
|
|
|
|
|
|
if (modelDescribe.ModelType == ModelTypeEnum.Image)
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage = new ThorUsageResponse
|
|
|
|
|
|
{
|
|
|
|
|
|
InputTokens = (int)modelDescribe.Multiplier,
|
|
|
|
|
|
OutputTokens = (int)modelDescribe.Multiplier,
|
|
|
|
|
|
TotalTokens = (int)modelDescribe.Multiplier
|
|
|
|
|
|
};
|
|
|
|
|
|
}
|
|
|
|
|
|
else
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
|
|
|
|
|
|
}
|
2025-12-17 21:51:01 +08:00
|
|
|
|
}
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
|
|
|
|
|
await response.WriteAsync($"data: {JsonSerializer.Serialize(responseResult)}\n\n", Encoding.UTF8,
|
|
|
|
|
|
cancellationToken).ConfigureAwait(false);
|
2025-12-17 21:51:01 +08:00
|
|
|
|
await response.Body.FlushAsync(cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
catch (Exception e)
|
|
|
|
|
|
{
|
|
|
|
|
|
_logger.LogError(e, $"Ai生成异常");
|
|
|
|
|
|
var errorContent = $"生成Ai异常,异常信息:\n当前Ai模型:{modelId}\n异常信息:{e.Message}\n异常堆栈:{e}";
|
|
|
|
|
|
throw new UserFriendlyException(errorContent);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-12-25 23:25:54 +08:00
|
|
|
|
Content = "不予存储",
|
2025-12-17 21:51:01 +08:00
|
|
|
|
ModelId = modelId,
|
|
|
|
|
|
TokenUsage = tokenUsage,
|
|
|
|
|
|
}, tokenId);
|
|
|
|
|
|
|
|
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2025-12-25 23:25:54 +08:00
|
|
|
|
Content = "不予存储",
|
2025-12-17 21:51:01 +08:00
|
|
|
|
ModelId = modelId,
|
|
|
|
|
|
TokenUsage = tokenUsage
|
|
|
|
|
|
}, tokenId);
|
|
|
|
|
|
|
|
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId, modelId, tokenUsage, tokenId);
|
|
|
|
|
|
|
2026-01-04 00:08:08 +08:00
|
|
|
|
// 直接扣减尊享token包用量
|
2025-12-17 21:51:01 +08:00
|
|
|
|
if (userId.HasValue && tokenUsage is not null)
|
|
|
|
|
|
{
|
|
|
|
|
|
var totalTokens = tokenUsage.TotalTokens ?? 0;
|
|
|
|
|
|
if (tokenUsage.TotalTokens > 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2026-01-03 22:09:30 +08:00
|
|
|
|
private const string ImageStoreHost = "https://ccnetcore.com/prod-api";
|
2026-01-04 00:08:08 +08:00
|
|
|
|
|
2025-12-25 23:25:54 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// Gemini 生成(Image)-非流式-缓存处理
|
|
|
|
|
|
/// 返回图片绝对路径
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
/// <param name="taskId"></param>
|
|
|
|
|
|
/// <param name="modelId"></param>
|
|
|
|
|
|
/// <param name="request"></param>
|
|
|
|
|
|
/// <param name="userId"></param>
|
|
|
|
|
|
/// <param name="sessionId"></param>
|
|
|
|
|
|
/// <param name="tokenId"></param>
|
|
|
|
|
|
/// <param name="cancellationToken"></param>
|
|
|
|
|
|
public async Task GeminiGenerateContentImageForStatisticsAsync(
|
|
|
|
|
|
Guid taskId,
|
|
|
|
|
|
string modelId,
|
|
|
|
|
|
JsonElement request,
|
|
|
|
|
|
Guid userId,
|
|
|
|
|
|
Guid? sessionId = null,
|
|
|
|
|
|
Guid? tokenId = null,
|
|
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
|
|
|
|
|
var imageStoreTask = await _imageStoreTaskRepository.GetFirstAsync(x => x.Id == taskId);
|
|
|
|
|
|
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.GenerateContent, modelId);
|
|
|
|
|
|
|
|
|
|
|
|
var chatService =
|
|
|
|
|
|
LazyServiceProvider.GetRequiredKeyedService<IGeminiGenerateContentService>(modelDescribe.HandlerName);
|
|
|
|
|
|
var data = await chatService.GenerateContentAsync(modelDescribe, request, cancellationToken);
|
|
|
|
|
|
|
|
|
|
|
|
//解析json,获取base64字符串
|
2026-01-03 03:19:31 +08:00
|
|
|
|
var imagePrefixBase64 = GeminiGenerateContentAcquirer.GetImagePrefixBase64(data);
|
2026-01-04 12:32:31 +08:00
|
|
|
|
if (string.IsNullOrWhiteSpace(imagePrefixBase64))
|
|
|
|
|
|
{
|
2026-01-08 22:09:42 +08:00
|
|
|
|
_logger.LogError($"图片生成解析失败,模型id:,请求信息:【{request}】,请求响应信息:{imagePrefixBase64}");
|
2026-01-04 12:32:31 +08:00
|
|
|
|
throw new UserFriendlyException("大模型没有返回图片,请调整提示词或稍后再试");
|
|
|
|
|
|
}
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
2025-12-26 23:46:36 +08:00
|
|
|
|
//远程调用上传接口,将base64转换为URL
|
|
|
|
|
|
var httpClient = LazyServiceProvider.LazyGetRequiredService<IHttpClientFactory>().CreateClient();
|
2026-01-02 19:26:09 +08:00
|
|
|
|
// var uploadUrl = $"https://ccnetcore.com/prod-api/ai-hub/ai-image/upload-base64";
|
2026-01-02 21:32:48 +08:00
|
|
|
|
var uploadUrl = $"{ImageStoreHost}/ai-image/upload-base64";
|
2026-01-03 03:19:31 +08:00
|
|
|
|
var content = new StringContent(JsonSerializer.Serialize(imagePrefixBase64), Encoding.UTF8, "application/json");
|
2025-12-26 23:46:36 +08:00
|
|
|
|
var uploadResponse = await httpClient.PostAsync(uploadUrl, content, cancellationToken);
|
|
|
|
|
|
uploadResponse.EnsureSuccessStatusCode();
|
|
|
|
|
|
var storeUrl = await uploadResponse.Content.ReadAsStringAsync(cancellationToken);
|
2025-12-25 23:25:54 +08:00
|
|
|
|
|
|
|
|
|
|
var tokenUsage = new ThorUsageResponse
|
|
|
|
|
|
{
|
|
|
|
|
|
InputTokens = (int)modelDescribe.Multiplier,
|
|
|
|
|
|
OutputTokens = (int)modelDescribe.Multiplier,
|
|
|
|
|
|
TotalTokens = (int)modelDescribe.Multiplier,
|
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
|
|
|
|
|
Content = "不予存储",
|
|
|
|
|
|
ModelId = modelId,
|
|
|
|
|
|
TokenUsage = tokenUsage
|
|
|
|
|
|
}, tokenId);
|
|
|
|
|
|
|
|
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId, modelId, tokenUsage, tokenId);
|
|
|
|
|
|
|
2026-01-04 00:08:08 +08:00
|
|
|
|
// 直接扣减尊享token包用量
|
2025-12-25 23:25:54 +08:00
|
|
|
|
var totalTokens = tokenUsage.TotalTokens ?? 0;
|
|
|
|
|
|
if (totalTokens > 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId, totalTokens);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-12-26 23:46:36 +08:00
|
|
|
|
//设置存储base64和url
|
2026-01-02 21:32:48 +08:00
|
|
|
|
imageStoreTask.SetSuccess($"{ImageStoreHost}{storeUrl}");
|
2025-12-25 23:25:54 +08:00
|
|
|
|
await _imageStoreTaskRepository.UpdateAsync(imageStoreTask);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 统一流式处理 - 支持4种API类型的原封不动转发
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
public async Task UnifiedStreamForStatisticsAsync(
|
|
|
|
|
|
HttpContext httpContext,
|
|
|
|
|
|
ModelApiTypeEnum apiType,
|
|
|
|
|
|
JsonElement requestBody,
|
|
|
|
|
|
string modelId,
|
|
|
|
|
|
Guid? userId = null,
|
|
|
|
|
|
Guid? sessionId = null,
|
|
|
|
|
|
Guid? tokenId = null,
|
|
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
|
|
|
|
|
var response = httpContext.Response;
|
|
|
|
|
|
// 设置响应头,声明是 SSE 流
|
|
|
|
|
|
response.ContentType = "text/event-stream;charset=utf-8;";
|
|
|
|
|
|
response.Headers.TryAdd("Cache-Control", "no-cache");
|
|
|
|
|
|
response.Headers.TryAdd("Connection", "keep-alive");
|
|
|
|
|
|
|
|
|
|
|
|
var sourceModelId = modelId;
|
|
|
|
|
|
// 处理 yi- 前缀
|
|
|
|
|
|
if (!string.IsNullOrEmpty(modelId) &&
|
|
|
|
|
|
modelId.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
|
|
|
|
|
{
|
|
|
|
|
|
modelId = modelId[3..];
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
var modelDescribe = await GetModelAsync(apiType, sourceModelId);
|
|
|
|
|
|
|
|
|
|
|
|
// 公共缓存队列
|
|
|
|
|
|
var messageQueue = new ConcurrentQueue<string>();
|
|
|
|
|
|
var outputInterval = TimeSpan.FromMilliseconds(75);
|
|
|
|
|
|
var isComplete = false;
|
|
|
|
|
|
|
|
|
|
|
|
// 公共消费任务
|
|
|
|
|
|
var outputTask = Task.Run(async () =>
|
|
|
|
|
|
{
|
|
|
|
|
|
while (!(isComplete && messageQueue.IsEmpty))
|
|
|
|
|
|
{
|
|
|
|
|
|
if (messageQueue.TryDequeue(out var message))
|
|
|
|
|
|
{
|
|
|
|
|
|
await response.WriteAsync(message, Encoding.UTF8, cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
await response.Body.FlushAsync(cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
if (!isComplete)
|
|
|
|
|
|
{
|
|
|
|
|
|
await Task.Delay(outputInterval, cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
}
|
|
|
|
|
|
else
|
|
|
|
|
|
{
|
|
|
|
|
|
await Task.Delay(10, cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}, cancellationToken);
|
2026-01-11 13:48:20 +08:00
|
|
|
|
|
|
|
|
|
|
StreamProcessResult? processResult = null;
|
2026-01-10 00:22:57 +08:00
|
|
|
|
|
|
|
|
|
|
switch (apiType)
|
|
|
|
|
|
{
|
|
|
|
|
|
case ModelApiTypeEnum.Completions:
|
2026-01-11 13:48:20 +08:00
|
|
|
|
processResult = await ProcessCompletionsStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
|
2026-01-10 00:22:57 +08:00
|
|
|
|
break;
|
|
|
|
|
|
case ModelApiTypeEnum.Messages:
|
2026-01-11 13:48:20 +08:00
|
|
|
|
processResult = await ProcessAnthropicStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
|
2026-01-10 00:22:57 +08:00
|
|
|
|
break;
|
|
|
|
|
|
case ModelApiTypeEnum.Responses:
|
2026-01-11 13:48:20 +08:00
|
|
|
|
processResult = await ProcessOpenAiResponsesStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
|
2026-01-10 00:22:57 +08:00
|
|
|
|
break;
|
|
|
|
|
|
case ModelApiTypeEnum.GenerateContent:
|
2026-01-11 13:48:20 +08:00
|
|
|
|
processResult = await ProcessGeminiStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
|
2026-01-10 00:22:57 +08:00
|
|
|
|
break;
|
|
|
|
|
|
default:
|
|
|
|
|
|
throw new UserFriendlyException($"不支持的API类型: {apiType}");
|
|
|
|
|
|
}
|
2026-01-11 13:48:20 +08:00
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
// 标记完成并等待消费任务结束
|
|
|
|
|
|
isComplete = true;
|
|
|
|
|
|
await outputTask;
|
|
|
|
|
|
|
|
|
|
|
|
// 统一的统计处理
|
|
|
|
|
|
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2026-01-11 13:48:20 +08:00
|
|
|
|
Content = sessionId is null ? "不予存储" : processResult?.UserContent ?? string.Empty,
|
2026-01-10 00:22:57 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2026-01-11 13:48:20 +08:00
|
|
|
|
TokenUsage = processResult?.TokenUsage,
|
2026-01-10 00:22:57 +08:00
|
|
|
|
}, tokenId);
|
|
|
|
|
|
|
|
|
|
|
|
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
|
|
|
|
|
|
new MessageInputDto
|
|
|
|
|
|
{
|
2026-01-11 13:48:20 +08:00
|
|
|
|
Content = sessionId is null ? "不予存储" : processResult?.SystemContent ?? string.Empty,
|
2026-01-10 00:22:57 +08:00
|
|
|
|
ModelId = sourceModelId,
|
2026-01-11 13:48:20 +08:00
|
|
|
|
TokenUsage = processResult?.TokenUsage
|
2026-01-10 00:22:57 +08:00
|
|
|
|
}, tokenId);
|
|
|
|
|
|
|
2026-01-11 13:48:20 +08:00
|
|
|
|
await _usageStatisticsManager.SetUsageAsync(userId, sourceModelId, processResult?.TokenUsage, tokenId);
|
2026-01-10 00:22:57 +08:00
|
|
|
|
|
|
|
|
|
|
// 扣减尊享token包用量
|
2026-01-11 13:48:20 +08:00
|
|
|
|
if (userId.HasValue && processResult?.TokenUsage is not null && modelDescribe.IsPremium)
|
2026-01-10 00:22:57 +08:00
|
|
|
|
{
|
2026-01-11 13:48:20 +08:00
|
|
|
|
var totalTokens = processResult?.TokenUsage.TotalTokens ?? 0;
|
2026-01-10 00:22:57 +08:00
|
|
|
|
if (totalTokens > 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
#region 统一流式处理 - 各API类型的具体实现
|
|
|
|
|
|
|
2026-01-11 13:48:20 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 流式处理结果,包含用户输入、系统输出和 token 使用情况
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
private class StreamProcessResult
|
|
|
|
|
|
{
|
|
|
|
|
|
public string UserContent { get; set; } = string.Empty;
|
|
|
|
|
|
public string SystemContent { get; set; } = string.Empty;
|
|
|
|
|
|
public ThorUsageResponse TokenUsage { get; set; } = new();
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 处理 OpenAI Completions 格式流式响应
|
|
|
|
|
|
/// </summary>
|
2026-01-11 13:48:20 +08:00
|
|
|
|
private async Task<StreamProcessResult> ProcessCompletionsStreamAsync(
|
2026-01-10 00:22:57 +08:00
|
|
|
|
ConcurrentQueue<string> messageQueue,
|
|
|
|
|
|
JsonElement requestBody,
|
|
|
|
|
|
AiModelDescribe modelDescribe,
|
|
|
|
|
|
CancellationToken cancellationToken)
|
|
|
|
|
|
{
|
|
|
|
|
|
var request = requestBody.Deserialize<ThorChatCompletionsRequest>(ThorJsonSerializer.DefaultOptions)!;
|
|
|
|
|
|
_specialCompatible.Compatible(request);
|
|
|
|
|
|
|
2026-01-11 13:48:20 +08:00
|
|
|
|
// 提取用户最后一条消息
|
|
|
|
|
|
var userContent = request.Messages?.LastOrDefault()?.MessagesStore ?? string.Empty;
|
|
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
// 处理 yi- 前缀
|
|
|
|
|
|
if (!string.IsNullOrEmpty(request.Model) &&
|
|
|
|
|
|
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
|
|
|
|
|
{
|
|
|
|
|
|
request.Model = request.Model[3..];
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
var chatService = LazyServiceProvider.GetRequiredKeyedService<IChatCompletionService>(modelDescribe.HandlerName);
|
|
|
|
|
|
var completeChatResponse = chatService.CompleteChatStreamAsync(modelDescribe, request, cancellationToken);
|
|
|
|
|
|
var tokenUsage = new ThorUsageResponse();
|
2026-01-11 13:48:20 +08:00
|
|
|
|
var systemContentBuilder = new StringBuilder();
|
2026-01-10 00:22:57 +08:00
|
|
|
|
|
|
|
|
|
|
try
|
|
|
|
|
|
{
|
|
|
|
|
|
await foreach (var data in completeChatResponse)
|
|
|
|
|
|
{
|
|
|
|
|
|
data.SupplementalMultiplier(modelDescribe.Multiplier);
|
|
|
|
|
|
if (data.Usage is not null && (data.Usage.CompletionTokens > 0 || data.Usage.OutputTokens > 0))
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage = data.Usage;
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-11 13:48:20 +08:00
|
|
|
|
// 累加系统输出内容 (choices[].delta.content)
|
|
|
|
|
|
var deltaContent = data.Choices?.FirstOrDefault()?.Delta?.Content;
|
|
|
|
|
|
if (!string.IsNullOrEmpty(deltaContent))
|
|
|
|
|
|
{
|
|
|
|
|
|
systemContentBuilder.Append(deltaContent);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
var message = JsonSerializer.Serialize(data, ThorJsonSerializer.DefaultOptions);
|
|
|
|
|
|
messageQueue.Enqueue($"data: {message}\n\n");
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
catch (Exception e)
|
|
|
|
|
|
{
|
|
|
|
|
|
_logger.LogError(e, "Ai对话异常");
|
|
|
|
|
|
var errorContent = $"对话Ai异常,异常信息:\n当前Ai模型:{request.Model}\n异常信息:{e.Message}\n异常堆栈:{e}";
|
2026-01-11 13:48:20 +08:00
|
|
|
|
systemContentBuilder.Append(errorContent);
|
2026-01-10 00:22:57 +08:00
|
|
|
|
var model = new ThorChatCompletionsResponse()
|
|
|
|
|
|
{
|
|
|
|
|
|
Choices = new List<ThorChatChoiceResponse>()
|
|
|
|
|
|
{
|
|
|
|
|
|
new ThorChatChoiceResponse()
|
|
|
|
|
|
{
|
|
|
|
|
|
Delta = new ThorChatMessage()
|
|
|
|
|
|
{
|
|
|
|
|
|
Content = errorContent
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
};
|
|
|
|
|
|
var errorMessage = JsonConvert.SerializeObject(model, new JsonSerializerSettings
|
|
|
|
|
|
{
|
|
|
|
|
|
ContractResolver = new CamelCasePropertyNamesContractResolver()
|
|
|
|
|
|
});
|
|
|
|
|
|
messageQueue.Enqueue($"data: {errorMessage}\n\n");
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
messageQueue.Enqueue("data: [DONE]\n\n");
|
2026-01-11 13:48:20 +08:00
|
|
|
|
return new StreamProcessResult
|
|
|
|
|
|
{
|
|
|
|
|
|
UserContent = userContent,
|
|
|
|
|
|
SystemContent = systemContentBuilder.ToString(),
|
|
|
|
|
|
TokenUsage = tokenUsage
|
|
|
|
|
|
};
|
2026-01-10 00:22:57 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 处理 Anthropic Messages 格式流式响应
|
|
|
|
|
|
/// </summary>
|
2026-01-11 13:48:20 +08:00
|
|
|
|
private async Task<StreamProcessResult> ProcessAnthropicStreamAsync(
|
2026-01-10 00:22:57 +08:00
|
|
|
|
ConcurrentQueue<string> messageQueue,
|
|
|
|
|
|
JsonElement requestBody,
|
|
|
|
|
|
AiModelDescribe modelDescribe,
|
|
|
|
|
|
CancellationToken cancellationToken)
|
|
|
|
|
|
{
|
|
|
|
|
|
var request = requestBody.Deserialize<AnthropicInput>(ThorJsonSerializer.DefaultOptions)!;
|
|
|
|
|
|
_specialCompatible.AnthropicCompatible(request);
|
|
|
|
|
|
|
2026-01-11 13:48:20 +08:00
|
|
|
|
// 提取用户最后一条消息
|
|
|
|
|
|
var lastMessage = request.Messages?.LastOrDefault();
|
|
|
|
|
|
var userContent = lastMessage?.Content ?? string.Empty;
|
|
|
|
|
|
if (string.IsNullOrEmpty(userContent) && lastMessage?.Contents != null && lastMessage.Contents.Any())
|
|
|
|
|
|
{
|
|
|
|
|
|
// 如果是 Contents 数组,提取第一个 text 类型的内容
|
|
|
|
|
|
var textContent = lastMessage.Contents.FirstOrDefault(c => c.Type == "text");
|
|
|
|
|
|
userContent = textContent?.Text ?? System.Text.Json.JsonSerializer.Serialize(lastMessage.Contents);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
// 处理 yi- 前缀
|
|
|
|
|
|
if (!string.IsNullOrEmpty(request.Model) &&
|
|
|
|
|
|
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
|
|
|
|
|
{
|
|
|
|
|
|
request.Model = request.Model[3..];
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
var chatService = LazyServiceProvider.GetRequiredKeyedService<IAnthropicChatCompletionService>(modelDescribe.HandlerName);
|
|
|
|
|
|
var completeChatResponse = chatService.StreamChatCompletionsAsync(modelDescribe, request, cancellationToken);
|
|
|
|
|
|
var tokenUsage = new ThorUsageResponse();
|
2026-01-11 13:48:20 +08:00
|
|
|
|
var systemContentBuilder = new StringBuilder();
|
2026-01-10 00:22:57 +08:00
|
|
|
|
|
|
|
|
|
|
try
|
|
|
|
|
|
{
|
|
|
|
|
|
await foreach (var responseResult in completeChatResponse)
|
|
|
|
|
|
{
|
|
|
|
|
|
// 部分供应商message_start放一部分
|
|
|
|
|
|
if (responseResult.Item1.Contains("message_start"))
|
|
|
|
|
|
{
|
|
|
|
|
|
var currentTokenUsage = responseResult.Item2?.Message?.Usage;
|
|
|
|
|
|
if (currentTokenUsage != null)
|
|
|
|
|
|
{
|
|
|
|
|
|
if ((currentTokenUsage.InputTokens ?? 0) != 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage.InputTokens = (currentTokenUsage.InputTokens ?? 0) +
|
|
|
|
|
|
(currentTokenUsage.CacheCreationInputTokens ?? 0) +
|
|
|
|
|
|
(currentTokenUsage.CacheReadInputTokens ?? 0);
|
|
|
|
|
|
}
|
|
|
|
|
|
if ((currentTokenUsage.OutputTokens ?? 0) != 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage.OutputTokens = currentTokenUsage.OutputTokens;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// message_delta又放一部分
|
|
|
|
|
|
if (responseResult.Item1.Contains("message_delta"))
|
|
|
|
|
|
{
|
|
|
|
|
|
var currentTokenUsage = responseResult.Item2?.Usage;
|
|
|
|
|
|
if (currentTokenUsage != null)
|
|
|
|
|
|
{
|
|
|
|
|
|
if ((currentTokenUsage.InputTokens ?? 0) != 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage.InputTokens = (currentTokenUsage.InputTokens ?? 0) +
|
|
|
|
|
|
(currentTokenUsage.CacheCreationInputTokens ?? 0) +
|
|
|
|
|
|
(currentTokenUsage.CacheReadInputTokens ?? 0);
|
|
|
|
|
|
}
|
|
|
|
|
|
if ((currentTokenUsage.OutputTokens ?? 0) != 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage.OutputTokens = currentTokenUsage.OutputTokens;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-11 13:48:20 +08:00
|
|
|
|
// 累加系统输出内容 (delta.text)
|
|
|
|
|
|
var deltaText = responseResult.Item2?.Delta?.Text;
|
|
|
|
|
|
if (!string.IsNullOrEmpty(deltaText))
|
|
|
|
|
|
{
|
|
|
|
|
|
systemContentBuilder.Append(deltaText);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
// 序列化为SSE格式字符串
|
|
|
|
|
|
var data = JsonSerializer.Serialize(responseResult.Item2, ThorJsonSerializer.DefaultOptions);
|
|
|
|
|
|
messageQueue.Enqueue($"{responseResult.Item1.Trim()}\ndata: {data}\n\n");
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
catch (Exception e)
|
|
|
|
|
|
{
|
|
|
|
|
|
_logger.LogError(e, "Ai对话异常");
|
|
|
|
|
|
var errorContent = $"对话Ai异常,异常信息:\n当前Ai模型:{request.Model}\n异常信息:{e.Message}\n异常堆栈:{e}";
|
2026-01-11 13:48:20 +08:00
|
|
|
|
systemContentBuilder.Append(errorContent);
|
2026-01-10 00:22:57 +08:00
|
|
|
|
throw new UserFriendlyException(errorContent);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
tokenUsage.TotalTokens = (tokenUsage.InputTokens ?? 0) + (tokenUsage.OutputTokens ?? 0);
|
|
|
|
|
|
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
|
|
|
|
|
|
|
2026-01-11 13:48:20 +08:00
|
|
|
|
return new StreamProcessResult
|
|
|
|
|
|
{
|
|
|
|
|
|
UserContent = userContent,
|
|
|
|
|
|
SystemContent = systemContentBuilder.ToString(),
|
|
|
|
|
|
TokenUsage = tokenUsage
|
|
|
|
|
|
};
|
2026-01-10 00:22:57 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 处理 OpenAI Responses 格式流式响应
|
|
|
|
|
|
/// </summary>
|
2026-01-11 13:48:20 +08:00
|
|
|
|
private async Task<StreamProcessResult> ProcessOpenAiResponsesStreamAsync(
|
2026-01-10 00:22:57 +08:00
|
|
|
|
ConcurrentQueue<string> messageQueue,
|
|
|
|
|
|
JsonElement requestBody,
|
|
|
|
|
|
AiModelDescribe modelDescribe,
|
|
|
|
|
|
CancellationToken cancellationToken)
|
|
|
|
|
|
{
|
|
|
|
|
|
var request = requestBody.Deserialize<OpenAiResponsesInput>(ThorJsonSerializer.DefaultOptions)!;
|
|
|
|
|
|
|
2026-01-11 13:48:20 +08:00
|
|
|
|
// 提取用户输入内容 (input 字段可能是字符串或数组)
|
|
|
|
|
|
var userContent = string.Empty;
|
|
|
|
|
|
if (request.Input.ValueKind == JsonValueKind.String)
|
|
|
|
|
|
{
|
|
|
|
|
|
userContent = request.Input.GetString() ?? string.Empty;
|
|
|
|
|
|
}
|
|
|
|
|
|
else if (request.Input.ValueKind == JsonValueKind.Array)
|
|
|
|
|
|
{
|
|
|
|
|
|
// 获取最后一个 user 角色的消息
|
|
|
|
|
|
var inputArray = request.Input.EnumerateArray().ToList();
|
|
|
|
|
|
var lastUserMessage = inputArray.LastOrDefault(x =>
|
|
|
|
|
|
x.TryGetProperty("role", out var role) && role.GetString() == "user");
|
|
|
|
|
|
if (lastUserMessage.ValueKind != JsonValueKind.Undefined)
|
|
|
|
|
|
{
|
|
|
|
|
|
if (lastUserMessage.TryGetProperty("content", out var content))
|
|
|
|
|
|
{
|
|
|
|
|
|
userContent = content.GetString() ?? string.Empty;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
// 处理 yi- 前缀
|
|
|
|
|
|
if (!string.IsNullOrEmpty(request.Model) &&
|
|
|
|
|
|
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
|
|
|
|
|
|
{
|
|
|
|
|
|
request.Model = request.Model[3..];
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
var chatService = LazyServiceProvider.GetRequiredKeyedService<IOpenAiResponseService>(modelDescribe.HandlerName);
|
|
|
|
|
|
var completeChatResponse = chatService.ResponsesStreamAsync(modelDescribe, request, cancellationToken);
|
|
|
|
|
|
ThorUsageResponse? tokenUsage = null;
|
2026-01-11 13:48:20 +08:00
|
|
|
|
var systemContentBuilder = new StringBuilder();
|
2026-01-10 00:22:57 +08:00
|
|
|
|
|
|
|
|
|
|
try
|
|
|
|
|
|
{
|
|
|
|
|
|
await foreach (var responseResult in completeChatResponse)
|
|
|
|
|
|
{
|
2026-01-11 13:48:20 +08:00
|
|
|
|
// 提取输出文本内容 (response.output_text.delta 事件)
|
|
|
|
|
|
if (responseResult.Item1.Contains("response.output_text.delta"))
|
|
|
|
|
|
{
|
|
|
|
|
|
var delta = responseResult.Item2?.GetPath("delta").GetString();
|
|
|
|
|
|
if (!string.IsNullOrEmpty(delta))
|
|
|
|
|
|
{
|
|
|
|
|
|
systemContentBuilder.Append(delta);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
if (responseResult.Item1.Contains("response.completed"))
|
|
|
|
|
|
{
|
|
|
|
|
|
var obj = responseResult.Item2!.Value;
|
|
|
|
|
|
int inputTokens = obj.GetPath("response", "usage", "input_tokens").GetInt();
|
|
|
|
|
|
int outputTokens = obj.GetPath("response", "usage", "output_tokens").GetInt();
|
|
|
|
|
|
inputTokens = Convert.ToInt32(inputTokens * modelDescribe.Multiplier);
|
|
|
|
|
|
outputTokens = Convert.ToInt32(outputTokens * modelDescribe.Multiplier);
|
|
|
|
|
|
tokenUsage = new ThorUsageResponse
|
|
|
|
|
|
{
|
|
|
|
|
|
PromptTokens = inputTokens,
|
|
|
|
|
|
InputTokens = inputTokens,
|
|
|
|
|
|
OutputTokens = outputTokens,
|
|
|
|
|
|
CompletionTokens = outputTokens,
|
|
|
|
|
|
TotalTokens = inputTokens + outputTokens,
|
|
|
|
|
|
};
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// 序列化为SSE格式字符串
|
|
|
|
|
|
var data = JsonSerializer.Serialize(responseResult.Item2, ThorJsonSerializer.DefaultOptions);
|
|
|
|
|
|
messageQueue.Enqueue($"{responseResult.Item1.Trim()}\ndata: {data}\n\n");
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
catch (Exception e)
|
|
|
|
|
|
{
|
|
|
|
|
|
_logger.LogError(e, "Ai响应异常");
|
|
|
|
|
|
var errorContent = $"响应Ai异常,异常信息:\n当前Ai模型:{request.Model}\n异常信息:{e.Message}\n异常堆栈:{e}";
|
2026-01-11 13:48:20 +08:00
|
|
|
|
systemContentBuilder.Append(errorContent);
|
2026-01-10 00:22:57 +08:00
|
|
|
|
throw new UserFriendlyException(errorContent);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-11 13:48:20 +08:00
|
|
|
|
return new StreamProcessResult
|
|
|
|
|
|
{
|
|
|
|
|
|
UserContent = userContent,
|
|
|
|
|
|
SystemContent = systemContentBuilder.ToString(),
|
|
|
|
|
|
TokenUsage = tokenUsage ?? new ThorUsageResponse()
|
|
|
|
|
|
};
|
2026-01-10 00:22:57 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 处理 Gemini GenerateContent 格式流式响应
|
|
|
|
|
|
/// </summary>
|
2026-01-11 13:48:20 +08:00
|
|
|
|
private async Task<StreamProcessResult> ProcessGeminiStreamAsync(
|
2026-01-10 00:22:57 +08:00
|
|
|
|
ConcurrentQueue<string> messageQueue,
|
|
|
|
|
|
JsonElement requestBody,
|
|
|
|
|
|
AiModelDescribe modelDescribe,
|
|
|
|
|
|
CancellationToken cancellationToken)
|
|
|
|
|
|
{
|
2026-01-11 13:48:20 +08:00
|
|
|
|
// 提取用户最后一条消息 (contents[last].parts[last].text)
|
|
|
|
|
|
var userContent = GeminiGenerateContentAcquirer.GetLastUserContent(requestBody);
|
|
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
var chatService = LazyServiceProvider.GetRequiredKeyedService<IGeminiGenerateContentService>(modelDescribe.HandlerName);
|
|
|
|
|
|
var completeChatResponse = chatService.GenerateContentStreamAsync(modelDescribe, requestBody, cancellationToken);
|
|
|
|
|
|
ThorUsageResponse? tokenUsage = null;
|
2026-01-11 13:48:20 +08:00
|
|
|
|
var systemContentBuilder = new StringBuilder();
|
2026-01-10 00:22:57 +08:00
|
|
|
|
|
|
|
|
|
|
try
|
|
|
|
|
|
{
|
|
|
|
|
|
await foreach (var responseResult in completeChatResponse)
|
|
|
|
|
|
{
|
2026-01-11 13:48:20 +08:00
|
|
|
|
// 累加系统输出内容 (candidates[0].content.parts[].text,排除 thought)
|
|
|
|
|
|
var textContent = GeminiGenerateContentAcquirer.GetTextContent(responseResult!.Value);
|
|
|
|
|
|
if (!string.IsNullOrEmpty(textContent))
|
|
|
|
|
|
{
|
|
|
|
|
|
systemContentBuilder.Append(textContent);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-10 00:22:57 +08:00
|
|
|
|
if (responseResult!.Value.GetPath("candidates", 0, "finishReason").GetString() == "STOP")
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage = GeminiGenerateContentAcquirer.GetUsage(responseResult!.Value);
|
|
|
|
|
|
// 如果是图片模型,单独扣费
|
|
|
|
|
|
if (modelDescribe.ModelType == ModelTypeEnum.Image)
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage = new ThorUsageResponse
|
|
|
|
|
|
{
|
|
|
|
|
|
InputTokens = (int)modelDescribe.Multiplier,
|
|
|
|
|
|
OutputTokens = (int)modelDescribe.Multiplier,
|
|
|
|
|
|
TotalTokens = (int)modelDescribe.Multiplier
|
|
|
|
|
|
};
|
|
|
|
|
|
}
|
|
|
|
|
|
else
|
|
|
|
|
|
{
|
|
|
|
|
|
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
messageQueue.Enqueue($"data: {JsonSerializer.Serialize(responseResult)}\n\n");
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
catch (Exception e)
|
|
|
|
|
|
{
|
|
|
|
|
|
_logger.LogError(e, "Ai生成异常");
|
|
|
|
|
|
var errorContent = $"生成Ai异常,异常信息:\n当前Ai模型:{modelDescribe.ModelId}\n异常信息:{e.Message}\n异常堆栈:{e}";
|
2026-01-11 13:48:20 +08:00
|
|
|
|
systemContentBuilder.Append(errorContent);
|
2026-01-10 00:22:57 +08:00
|
|
|
|
throw new UserFriendlyException(errorContent);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2026-01-11 13:48:20 +08:00
|
|
|
|
return new StreamProcessResult
|
|
|
|
|
|
{
|
|
|
|
|
|
UserContent = userContent,
|
|
|
|
|
|
SystemContent = systemContentBuilder.ToString(),
|
|
|
|
|
|
TokenUsage = tokenUsage ?? new ThorUsageResponse()
|
|
|
|
|
|
};
|
2026-01-10 00:22:57 +08:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
#endregion
|
|
|
|
|
|
|
2025-12-11 17:16:21 +08:00
|
|
|
|
#region 流式传输格式Http响应
|
2025-10-11 15:25:43 +08:00
|
|
|
|
|
|
|
|
|
|
private static readonly byte[] EventPrefix = "event: "u8.ToArray();
|
|
|
|
|
|
private static readonly byte[] DataPrefix = "data: "u8.ToArray();
|
|
|
|
|
|
private static readonly byte[] NewLine = "\n"u8.ToArray();
|
|
|
|
|
|
private static readonly byte[] DoubleNewLine = "\n\n"u8.ToArray();
|
|
|
|
|
|
|
|
|
|
|
|
/// <summary>
|
|
|
|
|
|
/// 使用 JsonSerializer.SerializeAsync 直接序列化到响应流
|
|
|
|
|
|
/// </summary>
|
|
|
|
|
|
private static async ValueTask WriteAsEventStreamDataAsync<T>(
|
|
|
|
|
|
HttpContext context,
|
|
|
|
|
|
string @event,
|
|
|
|
|
|
T value,
|
|
|
|
|
|
CancellationToken cancellationToken = default)
|
|
|
|
|
|
{
|
|
|
|
|
|
var response = context.Response;
|
|
|
|
|
|
var bodyStream = response.Body;
|
|
|
|
|
|
// 确保 SSE Header 已经设置好
|
|
|
|
|
|
// e.g. Content-Type: text/event-stream; charset=utf-8
|
|
|
|
|
|
await response.StartAsync(cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
// 写事件类型
|
2026-01-03 12:49:32 +08:00
|
|
|
|
//此处事件前缀重复了
|
|
|
|
|
|
// await bodyStream.WriteAsync(EventPrefix, cancellationToken).ConfigureAwait(false);
|
2025-10-11 15:25:43 +08:00
|
|
|
|
await WriteUtf8StringAsync(bodyStream, @event.Trim(), cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
await bodyStream.WriteAsync(NewLine, cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
// 写 data: + JSON
|
|
|
|
|
|
await bodyStream.WriteAsync(DataPrefix, cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
await JsonSerializer.SerializeAsync(
|
|
|
|
|
|
bodyStream,
|
|
|
|
|
|
value,
|
|
|
|
|
|
ThorJsonSerializer.DefaultOptions,
|
|
|
|
|
|
cancellationToken
|
|
|
|
|
|
).ConfigureAwait(false);
|
|
|
|
|
|
// 事件结束 \n\n
|
|
|
|
|
|
await bodyStream.WriteAsync(DoubleNewLine, cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
// 及时把数据发送给客户端
|
|
|
|
|
|
await bodyStream.FlushAsync(cancellationToken).ConfigureAwait(false);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
private static async ValueTask WriteUtf8StringAsync(Stream stream, string value, CancellationToken token)
|
|
|
|
|
|
{
|
|
|
|
|
|
if (string.IsNullOrEmpty(value))
|
|
|
|
|
|
return;
|
|
|
|
|
|
var buffer = Encoding.UTF8.GetBytes(value);
|
|
|
|
|
|
await stream.WriteAsync(buffer, token).ConfigureAwait(false);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
#endregion
|
2025-06-21 01:08:14 +08:00
|
|
|
|
}
|