Files
Yi.Admin/Yi.Abp.Net8/module/ai-hub/Yi.Framework.AiHub.Domain/Managers/AiGateWayManager.cs

1650 lines
67 KiB
C#
Raw Normal View History

2025-07-05 15:11:56 +08:00
using System.Collections.Concurrent;
2025-08-03 23:23:32 +08:00
using System.Diagnostics;
2025-06-25 17:12:09 +08:00
using System.Runtime.CompilerServices;
2025-07-05 15:11:56 +08:00
using System.Text;
using System.Text.Json;
2025-07-05 15:11:56 +08:00
using Microsoft.AspNetCore.Http;
2025-06-21 01:08:14 +08:00
using Microsoft.Extensions.DependencyInjection;
2025-07-05 15:11:56 +08:00
using Microsoft.Extensions.Logging;
using Newtonsoft.Json;
2025-12-11 01:17:31 +08:00
using Newtonsoft.Json.Linq;
2025-07-05 15:11:56 +08:00
using Newtonsoft.Json.Serialization;
2025-06-21 01:08:14 +08:00
using Volo.Abp.Domain.Services;
2025-07-17 23:10:26 +08:00
using Yi.Framework.AiHub.Domain.AiGateWay;
2025-08-03 23:23:32 +08:00
using Yi.Framework.AiHub.Domain.AiGateWay.Exceptions;
2025-12-25 23:25:54 +08:00
using Yi.Framework.AiHub.Domain.Entities.Chat;
2025-06-27 22:13:26 +08:00
using Yi.Framework.AiHub.Domain.Entities.Model;
using Yi.Framework.AiHub.Domain.Shared.Consts;
2025-06-25 17:12:09 +08:00
using Yi.Framework.AiHub.Domain.Shared.Dtos;
2025-10-11 15:25:43 +08:00
using Yi.Framework.AiHub.Domain.Shared.Dtos.Anthropic;
2025-12-17 18:47:28 +08:00
using Yi.Framework.AiHub.Domain.Shared.Dtos.Gemini;
2025-08-11 15:31:11 +08:00
using Yi.Framework.AiHub.Domain.Shared.Dtos.OpenAi;
using Yi.Framework.AiHub.Domain.Shared.Dtos.OpenAi.Embeddings;
using Yi.Framework.AiHub.Domain.Shared.Dtos.OpenAi.Images;
2025-12-11 01:17:31 +08:00
using Yi.Framework.AiHub.Domain.Shared.Dtos.OpenAi.Responses;
using Yi.Framework.AiHub.Domain.Shared.Enums;
2025-12-11 17:16:21 +08:00
using Yi.Framework.AiHub.Domain.Shared.Extensions;
2025-08-03 23:23:32 +08:00
using Yi.Framework.Core.Extensions;
2025-06-25 17:12:09 +08:00
using Yi.Framework.SqlSugarCore.Abstractions;
2025-10-11 15:25:43 +08:00
using JsonSerializer = System.Text.Json.JsonSerializer;
using ThorJsonSerializer = Yi.Framework.AiHub.Domain.AiGateWay.ThorJsonSerializer;
2025-06-21 01:08:14 +08:00
namespace Yi.Framework.AiHub.Domain.Managers;
public class AiGateWayManager : DomainService
{
2025-06-25 17:12:09 +08:00
private readonly ISqlSugarRepository<AiAppAggregateRoot> _aiAppRepository;
private readonly ISqlSugarRepository<AiModelEntity> _aiModelRepository;
2025-07-05 15:11:56 +08:00
private readonly ILogger<AiGateWayManager> _logger;
private readonly AiMessageManager _aiMessageManager;
private readonly UsageStatisticsManager _usageStatisticsManager;
2025-07-18 20:46:30 +08:00
private readonly ISpecialCompatible _specialCompatible;
private PremiumPackageManager? _premiumPackageManager;
2025-12-25 23:25:54 +08:00
private readonly ISqlSugarRepository<ImageStoreTaskAggregateRoot> _imageStoreTaskRepository;
2025-07-05 15:11:56 +08:00
public AiGateWayManager(ISqlSugarRepository<AiAppAggregateRoot> aiAppRepository, ILogger<AiGateWayManager> logger,
2025-07-18 20:46:30 +08:00
AiMessageManager aiMessageManager, UsageStatisticsManager usageStatisticsManager,
2025-12-25 23:25:54 +08:00
ISpecialCompatible specialCompatible, ISqlSugarRepository<AiModelEntity> aiModelRepository,
ISqlSugarRepository<ImageStoreTaskAggregateRoot> imageStoreTaskRepository)
2025-06-21 01:08:14 +08:00
{
2025-06-25 17:12:09 +08:00
_aiAppRepository = aiAppRepository;
2025-07-05 15:11:56 +08:00
_logger = logger;
_aiMessageManager = aiMessageManager;
_usageStatisticsManager = usageStatisticsManager;
2025-07-18 20:46:30 +08:00
_specialCompatible = specialCompatible;
_aiModelRepository = aiModelRepository;
2025-12-25 23:25:54 +08:00
_imageStoreTaskRepository = imageStoreTaskRepository;
2025-06-21 01:08:14 +08:00
}
private PremiumPackageManager PremiumPackageManager =>
_premiumPackageManager ??= LazyServiceProvider.LazyGetRequiredService<PremiumPackageManager>();
2025-06-25 17:12:09 +08:00
/// <summary>
/// 获取模型
/// </summary>
/// <param name="modelApiType"></param>
2025-06-25 17:12:09 +08:00
/// <param name="modelId"></param>
/// <returns></returns>
2025-12-23 17:08:42 +08:00
public async Task<AiModelDescribe> GetModelAsync(ModelApiTypeEnum modelApiType, string modelId)
2025-06-21 01:08:14 +08:00
{
var aiModelDescribe = await _aiModelRepository._DbQueryable
.LeftJoin<AiAppAggregateRoot>((model, app) => model.AiAppId == app.Id)
.Where((model, app) => model.ModelId == modelId)
.Where((model, app) => model.ModelApiType == modelApiType)
2026-01-05 00:11:06 +08:00
.Where((model, app) => model.IsEnabled)
.Select((model, app) =>
new AiModelDescribe
2025-06-25 17:12:09 +08:00
{
AppId = app.Id,
AppName = app.Name,
Endpoint = app.Endpoint,
ApiKey = app.ApiKey,
OrderNum = model.OrderNum,
HandlerName = model.HandlerName,
ModelId = model.ModelId,
ModelName = model.Name,
2025-07-17 23:16:16 +08:00
Description = model.Description,
2025-07-17 23:52:00 +08:00
AppExtraUrl = app.ExtraUrl,
2025-11-25 09:54:13 +08:00
ModelExtraInfo = model.ExtraInfo,
Multiplier = model.Multiplier,
2026-01-04 12:32:31 +08:00
IsPremium = model.IsPremium,
ModelType = model.ModelType
})
.FirstAsync();
if (aiModelDescribe is null)
{
throw new UserFriendlyException($"【{modelId}】模型当前版本【{modelApiType}】格式不支持");
2025-06-21 01:08:14 +08:00
}
// ✅ 统一处理 yi- 后缀(网关层模型规范化)
if (!string.IsNullOrEmpty(aiModelDescribe.ModelId) &&
aiModelDescribe.ModelId.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
aiModelDescribe.ModelId = aiModelDescribe.ModelId[3..];
}
return aiModelDescribe;
2025-06-25 17:12:09 +08:00
}
2025-12-25 23:25:54 +08:00
2025-07-09 19:12:53 +08:00
/// <summary>
/// 聊天完成-非流式
/// </summary>
/// <param name="httpContext"></param>
2025-07-17 23:10:26 +08:00
/// <param name="request"></param>
2025-07-09 19:12:53 +08:00
/// <param name="userId"></param>
/// <param name="sessionId"></param>
2025-11-27 19:01:16 +08:00
/// <param name="tokenId">Token IdWeb端传null或Guid.Empty</param>
2025-07-09 19:12:53 +08:00
/// <param name="cancellationToken"></param>
/// <returns></returns>
2025-07-17 23:10:26 +08:00
public async Task CompleteChatForStatisticsAsync(HttpContext httpContext,
ThorChatCompletionsRequest request,
2025-07-09 19:12:53 +08:00
Guid? userId = null,
Guid? sessionId = null,
2025-11-27 19:01:16 +08:00
Guid? tokenId = null,
2025-07-09 19:12:53 +08:00
CancellationToken cancellationToken = default)
{
2025-07-18 20:46:30 +08:00
_specialCompatible.Compatible(request);
2025-07-09 19:12:53 +08:00
var response = httpContext.Response;
// 设置响应头,声明是 json
2025-07-18 23:12:20 +08:00
//response.ContentType = "application/json; charset=UTF-8";
2026-01-08 21:38:36 +08:00
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Completions, request.Model);
2025-07-17 23:16:16 +08:00
var chatService =
LazyServiceProvider.GetRequiredKeyedService<IChatCompletionService>(modelDescribe.HandlerName);
var sourceModelId = request.Model;
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
request.Model = request.Model[3..];
}
2025-07-17 23:10:26 +08:00
var data = await chatService.CompleteChatAsync(modelDescribe, request, cancellationToken);
2025-11-25 09:54:13 +08:00
data.SupplementalMultiplier(modelDescribe.Multiplier);
2025-07-09 19:12:53 +08:00
if (userId is not null)
{
await _aiMessageManager.CreateUserMessageAsync(userId.Value, sessionId,
new MessageInputDto
{
Content = sessionId is null ? "不予存储" : request.Messages?.LastOrDefault().Content ?? string.Empty,
ModelId = sourceModelId,
2025-07-17 23:10:26 +08:00
TokenUsage = data.Usage,
2025-11-27 19:01:16 +08:00
}, tokenId);
2025-07-09 19:12:53 +08:00
await _aiMessageManager.CreateSystemMessageAsync(userId.Value, sessionId,
new MessageInputDto
{
Content =
sessionId is null ? "不予存储" : data.Choices?.FirstOrDefault()?.Delta.Content ?? string.Empty,
ModelId = sourceModelId,
2025-07-17 23:10:26 +08:00
TokenUsage = data.Usage
2025-11-27 19:01:16 +08:00
}, tokenId);
2025-07-09 19:12:53 +08:00
await _usageStatisticsManager.SetUsageAsync(userId.Value, sourceModelId, data.Usage, tokenId);
if (modelDescribe.IsPremium)
{
var totalTokens = data.Usage?.TotalTokens ?? 0;
if (totalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
}
}
2025-07-09 19:12:53 +08:00
}
2025-07-21 21:15:02 +08:00
2025-07-18 23:12:20 +08:00
await response.WriteAsJsonAsync(data, cancellationToken);
2025-07-09 19:12:53 +08:00
}
2025-12-25 23:25:54 +08:00
2025-07-05 15:11:56 +08:00
/// <summary>
/// 聊天完成-缓存处理
/// </summary>
/// <param name="httpContext"></param>
2025-07-17 23:10:26 +08:00
/// <param name="request"></param>
/// <param name="userId"></param>
2025-07-05 15:11:56 +08:00
/// <param name="sessionId"></param>
2025-11-27 19:01:16 +08:00
/// <param name="tokenId">Token IdWeb端传null或Guid.Empty</param>
2025-07-05 15:11:56 +08:00
/// <param name="cancellationToken"></param>
/// <returns></returns>
2025-07-09 19:12:53 +08:00
public async Task CompleteChatStreamForStatisticsAsync(
2025-07-05 15:11:56 +08:00
HttpContext httpContext,
2025-07-17 23:10:26 +08:00
ThorChatCompletionsRequest request,
2025-07-05 15:11:56 +08:00
Guid? userId = null,
Guid? sessionId = null,
2025-11-27 19:01:16 +08:00
Guid? tokenId = null,
2025-07-05 15:11:56 +08:00
CancellationToken cancellationToken = default)
{
var response = httpContext.Response;
// 设置响应头,声明是 SSE 流
2025-07-21 21:15:02 +08:00
response.ContentType = "text/event-stream;charset=utf-8;";
response.Headers.TryAdd("Cache-Control", "no-cache");
response.Headers.TryAdd("Connection", "keep-alive");
2025-07-05 15:11:56 +08:00
2025-12-11 17:16:21 +08:00
_specialCompatible.Compatible(request);
2026-01-08 21:38:36 +08:00
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Completions, request.Model);
2025-12-11 17:16:21 +08:00
var chatService =
LazyServiceProvider.GetRequiredKeyedService<IChatCompletionService>(modelDescribe.HandlerName);
2025-12-25 23:25:54 +08:00
var sourceModelId = request.Model;
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
request.Model = request.Model[3..];
}
2025-12-25 23:25:54 +08:00
var completeChatResponse = chatService.CompleteChatStreamAsync(modelDescribe, request, cancellationToken);
2025-07-17 23:10:26 +08:00
var tokenUsage = new ThorUsageResponse();
2025-07-05 15:11:56 +08:00
//缓存队列算法
// 创建一个队列来缓存消息
var messageQueue = new ConcurrentQueue<string>();
StringBuilder backupSystemContent = new StringBuilder();
// 设置输出速率例如每50毫秒输出一次
2025-07-08 18:24:21 +08:00
var outputInterval = TimeSpan.FromMilliseconds(75);
2025-07-05 15:11:56 +08:00
// 标记是否完成接收
var isComplete = false;
// 启动一个后台任务来消费队列
var outputTask = Task.Run(async () =>
{
while (!(isComplete && messageQueue.IsEmpty))
{
if (messageQueue.TryDequeue(out var message))
{
2025-07-21 21:15:02 +08:00
await response.WriteAsync(message, Encoding.UTF8, cancellationToken).ConfigureAwait(false);
await response.Body.FlushAsync(cancellationToken).ConfigureAwait(false);
2025-07-05 15:11:56 +08:00
}
if (!isComplete)
{
// 如果没有完成,才等待,已完成,全部输出
2025-07-21 21:15:02 +08:00
await Task.Delay(outputInterval, cancellationToken).ConfigureAwait(false);
2025-07-05 15:11:56 +08:00
}
else
{
//已经完成了,也等待,但是速度可以放快
await Task.Delay(10, cancellationToken).ConfigureAwait(false);
}
2025-07-05 15:11:56 +08:00
}
}, cancellationToken);
//IAsyncEnumerable 只能在最外层捕获异常(如果你有其他办法的话...
try
{
await foreach (var data in completeChatResponse)
{
2025-12-11 17:16:21 +08:00
data.SupplementalMultiplier(modelDescribe.Multiplier);
if (data.Usage is not null && (data.Usage.CompletionTokens > 0 || data.Usage.OutputTokens > 0))
2025-07-05 15:11:56 +08:00
{
2025-07-17 23:10:26 +08:00
tokenUsage = data.Usage;
2025-07-05 15:11:56 +08:00
}
2025-12-24 00:22:46 +08:00
var message = JsonSerializer.Serialize(data, ThorJsonSerializer.DefaultOptions);
2025-07-17 23:10:26 +08:00
backupSystemContent.Append(data.Choices.FirstOrDefault()?.Delta.Content);
2025-07-05 15:11:56 +08:00
// 将消息加入队列而不是直接写入
2025-07-21 21:15:02 +08:00
messageQueue.Enqueue($"data: {message}\n\n");
2025-07-05 15:11:56 +08:00
}
}
catch (Exception e)
{
_logger.LogError(e, $"Ai对话异常");
2025-12-12 21:14:38 +08:00
var errorContent = $"对话Ai异常异常信息\n当前Ai模型{request.Model}\n异常信息{e.Message}\n异常堆栈{e}";
2025-07-17 23:10:26 +08:00
var model = new ThorChatCompletionsResponse()
{
Choices = new List<ThorChatChoiceResponse>()
{
new ThorChatChoiceResponse()
{
Delta = new ThorChatMessage()
{
Content = errorContent
}
}
}
};
2025-07-05 15:11:56 +08:00
var message = JsonConvert.SerializeObject(model, new JsonSerializerSettings
{
ContractResolver = new CamelCasePropertyNamesContractResolver()
});
backupSystemContent.Append(errorContent);
2025-07-21 21:15:02 +08:00
messageQueue.Enqueue($"data: {message}\n\n");
2025-07-05 15:11:56 +08:00
}
//断开连接
2025-07-21 21:15:02 +08:00
messageQueue.Enqueue("data: [DONE]\n\n");
2025-07-05 15:11:56 +08:00
// 标记完成并发送结束标记
isComplete = true;
await outputTask;
2025-08-03 21:32:54 +08:00
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
new MessageInputDto
{
2025-12-14 13:07:44 +08:00
Content = sessionId is null ? "不予存储" : request.Messages?.LastOrDefault()?.MessagesStore ?? string.Empty,
ModelId = sourceModelId,
2025-08-03 21:32:54 +08:00
TokenUsage = tokenUsage,
2025-11-27 19:01:16 +08:00
}, tokenId);
2025-07-05 15:11:56 +08:00
2025-08-03 21:32:54 +08:00
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
new MessageInputDto
{
Content = sessionId is null ? "不予存储" : backupSystemContent.ToString(),
ModelId = sourceModelId,
2025-08-03 21:32:54 +08:00
TokenUsage = tokenUsage
2025-11-27 19:01:16 +08:00
}, tokenId);
2025-08-03 21:32:54 +08:00
await _usageStatisticsManager.SetUsageAsync(userId, sourceModelId, tokenUsage, tokenId);
// 扣减尊享token包用量
if (userId is not null)
{
if (modelDescribe.IsPremium)
{
var totalTokens = tokenUsage.TotalTokens ?? 0;
if (totalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
}
}
}
2025-07-05 15:11:56 +08:00
}
2025-12-25 23:25:54 +08:00
2025-08-03 23:23:32 +08:00
/// <summary>
/// 图片生成
/// </summary>
/// <param name="context"></param>
/// <param name="userId"></param>
/// <param name="sessionId"></param>
/// <param name="request"></param>
2025-11-27 19:01:16 +08:00
/// <param name="tokenId">Token IdWeb端传null或Guid.Empty</param>
2025-08-03 23:23:32 +08:00
/// <exception cref="BusinessException"></exception>
/// <exception cref="Exception"></exception>
public async Task CreateImageForStatisticsAsync(HttpContext context, Guid? userId, Guid? sessionId,
2025-11-27 19:01:16 +08:00
ImageCreateRequest request, Guid? tokenId = null)
2025-08-03 23:23:32 +08:00
{
try
{
var model = request.Model;
if (string.IsNullOrEmpty(model)) model = "dall-e-2";
2026-01-08 21:38:36 +08:00
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Completions, model);
2025-08-03 23:23:32 +08:00
// 获取渠道指定的实现类型的服务
var imageService =
LazyServiceProvider.GetRequiredKeyedService<IImageService>(modelDescribe.HandlerName);
2025-08-03 23:23:32 +08:00
var response = await imageService.CreateImage(request, modelDescribe);
if (response.Error != null || response.Results.Count == 0)
{
throw new BusinessException(response.Error?.Message ?? "图片生成失败", response.Error?.Code?.ToString());
}
await context.Response.WriteAsJsonAsync(response);
2025-08-03 23:23:32 +08:00
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
new MessageInputDto
{
Content = sessionId is null ? "不予存储" : request.Prompt,
2025-08-03 23:23:32 +08:00
ModelId = model,
TokenUsage = response.Usage,
2025-11-27 19:01:16 +08:00
}, tokenId);
2025-08-03 23:23:32 +08:00
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
new MessageInputDto
{
Content = sessionId is null ? "不予存储" : response.Results?.FirstOrDefault()?.Url,
2025-08-03 23:23:32 +08:00
ModelId = model,
TokenUsage = response.Usage
2025-11-27 19:01:16 +08:00
}, tokenId);
2025-08-03 23:23:32 +08:00
2025-11-27 19:01:16 +08:00
await _usageStatisticsManager.SetUsageAsync(userId, model, response.Usage, tokenId);
// 直接扣减尊享token包用量
if (userId is not null)
{
var totalTokens = response.Usage.TotalTokens ?? 0;
if (totalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
}
}
2025-08-03 23:23:32 +08:00
}
catch (Exception e)
{
2025-12-12 21:14:38 +08:00
var errorContent = $"图片生成Ai异常异常信息\n当前Ai模型{request.Model}\n异常信息{e.Message}\n异常堆栈{e}";
2025-08-03 23:23:32 +08:00
throw new UserFriendlyException(errorContent);
}
}
2025-12-25 23:25:54 +08:00
/// <summary>
/// 向量生成
/// </summary>
/// <param name="context"></param>
2025-11-27 19:01:16 +08:00
/// <param name="userId"></param>
/// <param name="sessionId"></param>
/// <param name="input"></param>
2025-11-27 19:01:16 +08:00
/// <param name="tokenId">Token IdWeb端传null或Guid.Empty</param>
/// <exception cref="Exception"></exception>
/// <exception cref="BusinessException"></exception>
public async Task EmbeddingForStatisticsAsync(HttpContext context, Guid? userId, Guid? sessionId,
2025-11-27 19:01:16 +08:00
ThorEmbeddingInput input, Guid? tokenId = null)
{
try
{
if (input == null) throw new Exception("模型校验异常");
using var embedding =
Activity.Current?.Source.StartActivity("向量模型调用");
2026-01-08 21:38:36 +08:00
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Completions, input.Model);
// 获取渠道指定的实现类型的服务
var embeddingService =
LazyServiceProvider.GetRequiredKeyedService<ITextEmbeddingService>(modelDescribe.HandlerName);
var embeddingCreateRequest = new EmbeddingCreateRequest
{
Model = input.Model,
EncodingFormat = input.EncodingFormat
};
//dto进行转换支持多种格式
if (input.Input is JsonElement str)
{
if (str.ValueKind == JsonValueKind.String)
{
embeddingCreateRequest.Input = str.ToString();
}
else if (str.ValueKind == JsonValueKind.Array)
{
var inputString = str.EnumerateArray().Select(x => x.ToString()).ToArray();
embeddingCreateRequest.InputAsList = inputString.ToList();
}
else
{
throw new Exception("Input输入格式错误非string或Array类型");
}
}
else if (input.Input is string strInput)
{
embeddingCreateRequest.Input = strInput;
}
else
{
throw new Exception("Input输入格式错误未找到类型");
}
var stream =
await embeddingService.EmbeddingAsync(embeddingCreateRequest, modelDescribe, context.RequestAborted);
var usage = new ThorUsageResponse()
{
2025-10-11 15:25:43 +08:00
PromptTokens = stream.Usage?.PromptTokens ?? 0,
InputTokens = stream.Usage?.InputTokens ?? 0,
CompletionTokens = 0,
TotalTokens = stream.Usage?.InputTokens ?? 0
};
await context.Response.WriteAsJsonAsync(new
{
input.Model,
stream.Data,
stream.Error,
Object = stream.ObjectTypeName,
Usage = usage
});
//知识库暂不使用message统计
// await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
// new MessageInputDto
// {
// Content = string.Empty,
// ModelId = input.Model,
// TokenUsage = usage,
// });
//
// await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
// new MessageInputDto
// {
// Content = string.Empty,
// ModelId = input.Model,
// TokenUsage = usage
// });
2025-11-27 19:01:16 +08:00
await _usageStatisticsManager.SetUsageAsync(userId, input.Model, usage, tokenId);
}
catch (ThorRateLimitException)
{
context.Response.StatusCode = 429;
}
catch (UnauthorizedAccessException e)
{
context.Response.StatusCode = 401;
}
catch (Exception e)
{
2025-12-12 21:14:38 +08:00
var errorContent = $"嵌入Ai异常异常信息\n当前Ai模型{input.Model}\n异常信息{e.Message}\n异常堆栈{e}";
throw new UserFriendlyException(errorContent);
}
}
2025-12-25 23:25:54 +08:00
2025-10-11 15:25:43 +08:00
/// <summary>
/// Anthropic聊天完成-非流式
/// </summary>
/// <param name="httpContext"></param>
/// <param name="request"></param>
/// <param name="userId"></param>
/// <param name="sessionId"></param>
2025-11-27 19:01:16 +08:00
/// <param name="tokenId">Token IdWeb端传null或Guid.Empty</param>
2025-10-11 15:25:43 +08:00
/// <param name="cancellationToken"></param>
/// <returns></returns>
public async Task AnthropicCompleteChatForStatisticsAsync(HttpContext httpContext,
AnthropicInput request,
Guid? userId = null,
Guid? sessionId = null,
2025-11-27 19:01:16 +08:00
Guid? tokenId = null,
2025-10-11 15:25:43 +08:00
CancellationToken cancellationToken = default)
{
_specialCompatible.AnthropicCompatible(request);
var response = httpContext.Response;
// 设置响应头,声明是 json
//response.ContentType = "application/json; charset=UTF-8";
2026-01-08 21:38:36 +08:00
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Messages, request.Model);
var sourceModelId = request.Model;
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
request.Model = request.Model[3..];
}
var chatService =
LazyServiceProvider.GetRequiredKeyedService<IAnthropicChatCompletionService>(modelDescribe.HandlerName);
var data = await chatService.ChatCompletionsAsync(modelDescribe, request, cancellationToken);
var currentUsage = data.Usage;
ThorUsageResponse tokenUsage = new ThorUsageResponse
{
InputTokens = (currentUsage?.InputTokens??0) + (currentUsage?.CacheCreationInputTokens??0)+ (currentUsage?.CacheReadInputTokens??0),
OutputTokens = (currentUsage?.OutputTokens??0),
TotalTokens = (currentUsage?.InputTokens??0) + (currentUsage?.CacheCreationInputTokens??0)+ (currentUsage?.CacheReadInputTokens??0)+(currentUsage?.OutputTokens??0)
};
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
2025-12-11 01:17:31 +08:00
2025-10-11 15:25:43 +08:00
if (userId is not null)
{
await _aiMessageManager.CreateUserMessageAsync(userId.Value, sessionId,
new MessageInputDto
{
2025-12-11 17:33:12 +08:00
Content = "不予存储",
ModelId = sourceModelId,
TokenUsage = tokenUsage,
2025-11-27 19:01:16 +08:00
}, tokenId);
2025-10-11 15:25:43 +08:00
await _aiMessageManager.CreateSystemMessageAsync(userId.Value, sessionId,
new MessageInputDto
{
2025-12-11 17:33:12 +08:00
Content = "不予存储",
ModelId = sourceModelId,
TokenUsage = tokenUsage
2025-11-27 19:01:16 +08:00
}, tokenId);
2025-10-11 15:25:43 +08:00
await _usageStatisticsManager.SetUsageAsync(userId.Value, sourceModelId, tokenUsage, tokenId);
// 直接扣减尊享token包用量
var totalTokens = tokenUsage.TotalTokens ?? 0;
if (totalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
}
2025-10-11 15:25:43 +08:00
}
await response.WriteAsJsonAsync(data, cancellationToken);
}
2025-12-25 23:25:54 +08:00
2025-10-11 15:25:43 +08:00
/// <summary>
/// Anthropic聊天完成-缓存处理
/// </summary>
/// <param name="httpContext"></param>
/// <param name="request"></param>
/// <param name="userId"></param>
/// <param name="sessionId"></param>
2025-11-27 19:01:16 +08:00
/// <param name="tokenId">Token IdWeb端传null或Guid.Empty</param>
2025-10-11 15:25:43 +08:00
/// <param name="cancellationToken"></param>
/// <returns></returns>
public async Task AnthropicCompleteChatStreamForStatisticsAsync(
HttpContext httpContext,
AnthropicInput request,
Guid? userId = null,
Guid? sessionId = null,
2025-11-27 19:01:16 +08:00
Guid? tokenId = null,
2025-10-11 15:25:43 +08:00
CancellationToken cancellationToken = default)
{
var response = httpContext.Response;
// 设置响应头,声明是 SSE 流
response.ContentType = "text/event-stream;charset=utf-8;";
response.Headers.TryAdd("Cache-Control", "no-cache");
response.Headers.TryAdd("Connection", "keep-alive");
2025-12-25 23:25:54 +08:00
2025-12-11 17:16:21 +08:00
_specialCompatible.AnthropicCompatible(request);
2026-01-08 21:38:36 +08:00
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Messages, request.Model);
2025-12-11 17:16:21 +08:00
var chatService =
LazyServiceProvider.GetRequiredKeyedService<IAnthropicChatCompletionService>(modelDescribe.HandlerName);
2025-12-25 23:25:54 +08:00
var sourceModelId = request.Model;
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
request.Model = request.Model[3..];
}
2025-12-25 23:25:54 +08:00
var completeChatResponse = chatService.StreamChatCompletionsAsync(modelDescribe, request, cancellationToken);
ThorUsageResponse? tokenUsage = new ThorUsageResponse();
2025-10-11 15:25:43 +08:00
try
{
await foreach (var responseResult in completeChatResponse)
{
//部分供应商message_start放一部分
if (responseResult.Item1.Contains("message_start"))
{
var currentTokenUsage = responseResult.Item2.Message.Usage;
if ((currentTokenUsage.InputTokens ?? 0) != 0)
2026-01-05 15:44:48 +08:00
{
tokenUsage.InputTokens = (currentTokenUsage?.InputTokens??0) + (currentTokenUsage?.CacheCreationInputTokens??0)+ (currentTokenUsage?.CacheReadInputTokens??0);
}
if ((currentTokenUsage.OutputTokens ?? 0) != 0)
{
tokenUsage.OutputTokens = currentTokenUsage.OutputTokens;
2026-01-05 15:44:48 +08:00
}
}
//message_delta又放一部分
if (responseResult.Item1.Contains("message_delta"))
{
var currentTokenUsage = responseResult.Item2.Usage;
if ((currentTokenUsage.InputTokens ?? 0) != 0)
{
tokenUsage.InputTokens = (currentTokenUsage?.InputTokens??0) + (currentTokenUsage?.CacheCreationInputTokens??0)+ (currentTokenUsage?.CacheReadInputTokens??0);;
}
if ((currentTokenUsage.OutputTokens ?? 0) != 0)
{
tokenUsage.OutputTokens = currentTokenUsage.OutputTokens;
}
}
2025-10-11 15:25:43 +08:00
await WriteAsEventStreamDataAsync(httpContext, responseResult.Item1, responseResult.Item2,
cancellationToken);
}
}
catch (Exception e)
{
_logger.LogError(e, $"Ai对话异常");
var errorContent = $"对话Ai异常异常信息\n当前Ai模型{sourceModelId}\n异常信息{e.Message}\n异常堆栈{e}";
2025-10-12 16:42:26 +08:00
throw new UserFriendlyException(errorContent);
2025-10-11 15:25:43 +08:00
}
tokenUsage.TotalTokens = (tokenUsage.InputTokens ?? 0) + (tokenUsage.OutputTokens ?? 0);
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
2025-10-11 15:25:43 +08:00
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
new MessageInputDto
{
2025-12-11 17:33:12 +08:00
Content = "不予存储",
ModelId = sourceModelId,
2025-10-11 15:25:43 +08:00
TokenUsage = tokenUsage,
2025-11-27 19:01:16 +08:00
}, tokenId);
2025-10-11 15:25:43 +08:00
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
new MessageInputDto
{
2025-12-11 17:33:12 +08:00
Content = "不予存储",
ModelId = sourceModelId,
2025-10-11 15:25:43 +08:00
TokenUsage = tokenUsage
2025-11-27 19:01:16 +08:00
}, tokenId);
2025-10-11 15:25:43 +08:00
await _usageStatisticsManager.SetUsageAsync(userId, sourceModelId, tokenUsage, tokenId);
// 直接扣减尊享token包用量
if (userId.HasValue && tokenUsage is not null)
{
var totalTokens = tokenUsage.TotalTokens ?? 0;
if (tokenUsage.TotalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
}
}
2025-10-11 15:25:43 +08:00
}
2025-12-11 01:17:31 +08:00
/// <summary>
/// OpenAi 响应-非流式-缓存处理
/// </summary>
/// <param name="httpContext"></param>
/// <param name="request"></param>
/// <param name="userId"></param>
/// <param name="sessionId"></param>
/// <param name="tokenId"></param>
/// <param name="cancellationToken"></param>
public async Task OpenAiResponsesAsyncForStatisticsAsync(HttpContext httpContext,
OpenAiResponsesInput request,
Guid? userId = null,
Guid? sessionId = null,
Guid? tokenId = null,
CancellationToken cancellationToken = default)
{
// _specialCompatible.AnthropicCompatible(request);
var response = httpContext.Response;
// 设置响应头,声明是 json
//response.ContentType = "application/json; charset=UTF-8";
2026-01-08 21:38:36 +08:00
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Responses, request.Model);
2025-12-11 01:17:31 +08:00
var chatService =
LazyServiceProvider.GetRequiredKeyedService<IOpenAiResponseService>(modelDescribe.HandlerName);
var sourceModelId = request.Model;
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
request.Model = request.Model[3..];
}
2025-12-11 01:17:31 +08:00
var data = await chatService.ResponsesAsync(modelDescribe, request, cancellationToken);
2025-12-25 23:25:54 +08:00
2025-12-11 17:16:21 +08:00
data.SupplementalMultiplier(modelDescribe.Multiplier);
2025-12-25 23:25:54 +08:00
var tokenUsage = new ThorUsageResponse
2025-12-11 01:17:31 +08:00
{
2025-12-11 17:16:21 +08:00
InputTokens = data.Usage.InputTokens,
OutputTokens = data.Usage.OutputTokens,
TotalTokens = data.Usage.InputTokens + data.Usage.OutputTokens,
};
if (userId is not null)
{
await _aiMessageManager.CreateUserMessageAsync(userId.Value, sessionId,
new MessageInputDto
{
Content = "不予存储",
ModelId = sourceModelId,
2025-12-11 17:16:21 +08:00
TokenUsage = tokenUsage,
}, tokenId);
2025-12-25 23:25:54 +08:00
2025-12-11 17:16:21 +08:00
await _aiMessageManager.CreateSystemMessageAsync(userId.Value, sessionId,
new MessageInputDto
{
Content = "不予存储",
ModelId = sourceModelId,
2025-12-11 17:16:21 +08:00
TokenUsage = tokenUsage
}, tokenId);
2025-12-25 23:25:54 +08:00
await _usageStatisticsManager.SetUsageAsync(userId.Value, sourceModelId, tokenUsage, tokenId);
2025-12-25 23:25:54 +08:00
// 直接扣减尊享token包用量
2025-12-11 17:16:21 +08:00
var totalTokens = tokenUsage.TotalTokens ?? 0;
if (totalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
}
2025-12-11 01:17:31 +08:00
}
2025-12-11 17:16:21 +08:00
await response.WriteAsJsonAsync(data, cancellationToken);
2025-12-11 01:17:31 +08:00
}
2025-12-25 23:25:54 +08:00
2025-12-11 01:17:31 +08:00
/// <summary>
/// OpenAi响应-流式-缓存处理
/// </summary>
/// <param name="httpContext"></param>
/// <param name="request"></param>
/// <param name="userId"></param>
/// <param name="sessionId"></param>
/// <param name="tokenId">Token IdWeb端传null或Guid.Empty</param>
/// <param name="cancellationToken"></param>
/// <returns></returns>
public async Task OpenAiResponsesStreamForStatisticsAsync(
HttpContext httpContext,
OpenAiResponsesInput request,
Guid? userId = null,
Guid? sessionId = null,
Guid? tokenId = null,
CancellationToken cancellationToken = default)
{
var response = httpContext.Response;
// 设置响应头,声明是 SSE 流
response.ContentType = "text/event-stream;charset=utf-8;";
response.Headers.TryAdd("Cache-Control", "no-cache");
response.Headers.TryAdd("Connection", "keep-alive");
2025-12-25 23:25:54 +08:00
2026-01-08 21:38:36 +08:00
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.Responses, request.Model);
2025-12-11 17:16:21 +08:00
var chatService =
LazyServiceProvider.GetRequiredKeyedService<IOpenAiResponseService>(modelDescribe.HandlerName);
var sourceModelId = request.Model;
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
request.Model = request.Model[3..];
}
2025-12-25 23:25:54 +08:00
var completeChatResponse = chatService.ResponsesStreamAsync(modelDescribe, request, cancellationToken);
2025-12-11 01:17:31 +08:00
ThorUsageResponse? tokenUsage = null;
try
{
await foreach (var responseResult in completeChatResponse)
{
//message_start是为了保底机制
if (responseResult.Item1.Contains("response.completed"))
{
2025-12-11 17:16:21 +08:00
var obj = responseResult.Item2!.Value;
2025-12-25 23:25:54 +08:00
int inputTokens = obj.GetPath("response", "usage", "input_tokens").GetInt();
int outputTokens = obj.GetPath("response", "usage", "output_tokens").GetInt();
inputTokens = Convert.ToInt32(inputTokens * modelDescribe.Multiplier);
outputTokens = Convert.ToInt32(outputTokens * modelDescribe.Multiplier);
2025-12-11 01:17:31 +08:00
tokenUsage = new ThorUsageResponse
{
2025-12-25 23:25:54 +08:00
PromptTokens = inputTokens,
2025-12-11 01:17:31 +08:00
InputTokens = inputTokens,
OutputTokens = outputTokens,
CompletionTokens = outputTokens,
2025-12-25 23:25:54 +08:00
TotalTokens = inputTokens + outputTokens,
2025-12-11 01:17:31 +08:00
};
}
2025-12-25 23:25:54 +08:00
2025-12-11 01:17:31 +08:00
await WriteAsEventStreamDataAsync(httpContext, responseResult.Item1, responseResult.Item2,
cancellationToken);
}
}
catch (Exception e)
{
_logger.LogError(e, $"Ai响应异常");
2025-12-12 21:14:38 +08:00
var errorContent = $"响应Ai异常异常信息\n当前Ai模型{request.Model}\n异常信息{e.Message}\n异常堆栈{e}";
2025-12-11 01:17:31 +08:00
throw new UserFriendlyException(errorContent);
}
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
new MessageInputDto
{
2025-12-25 23:25:54 +08:00
Content = "不予存储",
ModelId = sourceModelId,
2025-12-11 01:17:31 +08:00
TokenUsage = tokenUsage,
}, tokenId);
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
new MessageInputDto
{
2025-12-25 23:25:54 +08:00
Content = "不予存储",
ModelId = sourceModelId,
2025-12-11 01:17:31 +08:00
TokenUsage = tokenUsage
}, tokenId);
await _usageStatisticsManager.SetUsageAsync(userId, sourceModelId, tokenUsage, tokenId);
2025-12-11 01:17:31 +08:00
// 直接扣减尊享token包用量
2025-12-11 01:17:31 +08:00
if (userId.HasValue && tokenUsage is not null)
{
var totalTokens = tokenUsage.TotalTokens ?? 0;
if (tokenUsage.TotalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
}
}
}
2025-12-17 18:47:28 +08:00
/// <summary>
/// Gemini 生成-非流式-缓存处理
/// </summary>
/// <param name="httpContext"></param>
/// <param name="modelId"></param>
/// <param name="request"></param>
/// <param name="userId"></param>
/// <param name="sessionId"></param>
/// <param name="tokenId"></param>
/// <param name="cancellationToken"></param>
2025-12-17 21:51:01 +08:00
public async Task GeminiGenerateContentForStatisticsAsync(HttpContext httpContext,
2025-12-17 18:47:28 +08:00
string modelId,
JsonElement request,
Guid? userId = null,
Guid? sessionId = null,
Guid? tokenId = null,
CancellationToken cancellationToken = default)
{
var response = httpContext.Response;
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.GenerateContent, modelId);
var chatService =
LazyServiceProvider.GetRequiredKeyedService<IGeminiGenerateContentService>(modelDescribe.HandlerName);
var data = await chatService.GenerateContentAsync(modelDescribe, request, cancellationToken);
2025-12-25 23:25:54 +08:00
var tokenUsage = GeminiGenerateContentAcquirer.GetUsage(data);
2026-01-04 12:32:31 +08:00
//如果是图片模型,单独扣费
if (modelDescribe.ModelType == ModelTypeEnum.Image)
{
tokenUsage = new ThorUsageResponse
{
InputTokens = (int)modelDescribe.Multiplier,
OutputTokens = (int)modelDescribe.Multiplier,
TotalTokens = (int)modelDescribe.Multiplier
};
}
else
{
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
}
2025-12-25 23:25:54 +08:00
2025-12-17 18:47:28 +08:00
if (userId is not null)
{
await _aiMessageManager.CreateUserMessageAsync(userId.Value, sessionId,
new MessageInputDto
{
Content = "不予存储",
ModelId = modelId,
TokenUsage = tokenUsage,
}, tokenId);
2025-12-25 23:25:54 +08:00
2025-12-17 18:47:28 +08:00
await _aiMessageManager.CreateSystemMessageAsync(userId.Value, sessionId,
new MessageInputDto
{
Content = "不予存储",
ModelId = modelId,
TokenUsage = tokenUsage
}, tokenId);
2025-12-25 23:25:54 +08:00
2025-12-17 18:47:28 +08:00
await _usageStatisticsManager.SetUsageAsync(userId.Value, modelId, tokenUsage, tokenId);
2025-12-25 23:25:54 +08:00
// 直接扣减尊享token包用量
2025-12-17 18:47:28 +08:00
var totalTokens = tokenUsage.TotalTokens ?? 0;
if (totalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
}
}
await response.WriteAsJsonAsync(data, cancellationToken);
}
2025-12-17 21:51:01 +08:00
/// <summary>
/// Gemini 生成-流式-缓存处理
/// </summary>
/// <param name="httpContext"></param>
/// <param name="modelId"></param>
/// <param name="request"></param>
/// <param name="userId"></param>
/// <param name="sessionId"></param>
/// <param name="tokenId">Token IdWeb端传null或Guid.Empty</param>
/// <param name="cancellationToken"></param>
/// <returns></returns>
public async Task GeminiGenerateContentStreamForStatisticsAsync(
HttpContext httpContext,
string modelId,
JsonElement request,
Guid? userId = null,
Guid? sessionId = null,
Guid? tokenId = null,
CancellationToken cancellationToken = default)
{
var response = httpContext.Response;
// 设置响应头,声明是 SSE 流
response.ContentType = "text/event-stream;charset=utf-8;";
response.Headers.TryAdd("Cache-Control", "no-cache");
response.Headers.TryAdd("Connection", "keep-alive");
2025-12-25 23:25:54 +08:00
2025-12-17 21:51:01 +08:00
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.GenerateContent, modelId);
var chatService =
LazyServiceProvider.GetRequiredKeyedService<IGeminiGenerateContentService>(modelDescribe.HandlerName);
2025-12-25 23:25:54 +08:00
var completeChatResponse = chatService.GenerateContentStreamAsync(modelDescribe, request, cancellationToken);
2025-12-17 21:51:01 +08:00
ThorUsageResponse? tokenUsage = null;
try
{
await foreach (var responseResult in completeChatResponse)
{
2025-12-25 23:25:54 +08:00
if (responseResult!.Value.GetPath("candidates", 0, "finishReason").GetString() == "STOP")
2025-12-17 21:51:01 +08:00
{
tokenUsage = GeminiGenerateContentAcquirer.GetUsage(responseResult!.Value);
2026-01-04 12:32:31 +08:00
//如果是图片模型,单独扣费
if (modelDescribe.ModelType == ModelTypeEnum.Image)
{
tokenUsage = new ThorUsageResponse
{
InputTokens = (int)modelDescribe.Multiplier,
OutputTokens = (int)modelDescribe.Multiplier,
TotalTokens = (int)modelDescribe.Multiplier
};
}
else
{
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
}
2025-12-17 21:51:01 +08:00
}
2025-12-25 23:25:54 +08:00
await response.WriteAsync($"data: {JsonSerializer.Serialize(responseResult)}\n\n", Encoding.UTF8,
cancellationToken).ConfigureAwait(false);
2025-12-17 21:51:01 +08:00
await response.Body.FlushAsync(cancellationToken).ConfigureAwait(false);
}
}
catch (Exception e)
{
_logger.LogError(e, $"Ai生成异常");
var errorContent = $"生成Ai异常异常信息\n当前Ai模型{modelId}\n异常信息{e.Message}\n异常堆栈{e}";
throw new UserFriendlyException(errorContent);
}
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
new MessageInputDto
{
2025-12-25 23:25:54 +08:00
Content = "不予存储",
2025-12-17 21:51:01 +08:00
ModelId = modelId,
TokenUsage = tokenUsage,
}, tokenId);
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
new MessageInputDto
{
2025-12-25 23:25:54 +08:00
Content = "不予存储",
2025-12-17 21:51:01 +08:00
ModelId = modelId,
TokenUsage = tokenUsage
}, tokenId);
await _usageStatisticsManager.SetUsageAsync(userId, modelId, tokenUsage, tokenId);
// 直接扣减尊享token包用量
2025-12-17 21:51:01 +08:00
if (userId.HasValue && tokenUsage is not null)
{
var totalTokens = tokenUsage.TotalTokens ?? 0;
if (tokenUsage.TotalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
}
}
}
2025-12-25 23:25:54 +08:00
private const string ImageStoreHost = "https://ccnetcore.com/prod-api";
2025-12-25 23:25:54 +08:00
/// <summary>
/// Gemini 生成(Image)-非流式-缓存处理
/// 返回图片绝对路径
/// </summary>
/// <param name="taskId"></param>
/// <param name="modelId"></param>
/// <param name="request"></param>
/// <param name="userId"></param>
/// <param name="sessionId"></param>
/// <param name="tokenId"></param>
/// <param name="cancellationToken"></param>
public async Task GeminiGenerateContentImageForStatisticsAsync(
Guid taskId,
string modelId,
JsonElement request,
Guid userId,
Guid? sessionId = null,
Guid? tokenId = null,
CancellationToken cancellationToken = default)
{
var imageStoreTask = await _imageStoreTaskRepository.GetFirstAsync(x => x.Id == taskId);
var modelDescribe = await GetModelAsync(ModelApiTypeEnum.GenerateContent, modelId);
var chatService =
LazyServiceProvider.GetRequiredKeyedService<IGeminiGenerateContentService>(modelDescribe.HandlerName);
var data = await chatService.GenerateContentAsync(modelDescribe, request, cancellationToken);
//解析json获取base64字符串
var imagePrefixBase64 = GeminiGenerateContentAcquirer.GetImagePrefixBase64(data);
2026-01-04 12:32:31 +08:00
if (string.IsNullOrWhiteSpace(imagePrefixBase64))
{
_logger.LogError($"图片生成解析失败模型id:,请求信息:【{request}】,请求响应信息:{imagePrefixBase64}");
2026-01-04 12:32:31 +08:00
throw new UserFriendlyException("大模型没有返回图片,请调整提示词或稍后再试");
}
2025-12-25 23:25:54 +08:00
2025-12-26 23:46:36 +08:00
//远程调用上传接口将base64转换为URL
var httpClient = LazyServiceProvider.LazyGetRequiredService<IHttpClientFactory>().CreateClient();
2026-01-02 19:26:09 +08:00
// var uploadUrl = $"https://ccnetcore.com/prod-api/ai-hub/ai-image/upload-base64";
var uploadUrl = $"{ImageStoreHost}/ai-image/upload-base64";
var content = new StringContent(JsonSerializer.Serialize(imagePrefixBase64), Encoding.UTF8, "application/json");
2025-12-26 23:46:36 +08:00
var uploadResponse = await httpClient.PostAsync(uploadUrl, content, cancellationToken);
uploadResponse.EnsureSuccessStatusCode();
var storeUrl = await uploadResponse.Content.ReadAsStringAsync(cancellationToken);
2025-12-25 23:25:54 +08:00
var tokenUsage = new ThorUsageResponse
{
InputTokens = (int)modelDescribe.Multiplier,
OutputTokens = (int)modelDescribe.Multiplier,
TotalTokens = (int)modelDescribe.Multiplier,
};
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
new MessageInputDto
{
Content = "不予存储",
ModelId = modelId,
TokenUsage = tokenUsage
}, tokenId);
await _usageStatisticsManager.SetUsageAsync(userId, modelId, tokenUsage, tokenId);
// 直接扣减尊享token包用量
2025-12-25 23:25:54 +08:00
var totalTokens = tokenUsage.TotalTokens ?? 0;
if (totalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId, totalTokens);
}
2025-12-26 23:46:36 +08:00
//设置存储base64和url
imageStoreTask.SetSuccess($"{ImageStoreHost}{storeUrl}");
2025-12-25 23:25:54 +08:00
await _imageStoreTaskRepository.UpdateAsync(imageStoreTask);
}
/// <summary>
/// 统一流式处理 - 支持4种API类型的原封不动转发
/// </summary>
public async Task UnifiedStreamForStatisticsAsync(
HttpContext httpContext,
ModelApiTypeEnum apiType,
JsonElement requestBody,
string modelId,
Guid? userId = null,
Guid? sessionId = null,
Guid? tokenId = null,
CancellationToken cancellationToken = default)
{
var response = httpContext.Response;
// 设置响应头,声明是 SSE 流
response.ContentType = "text/event-stream;charset=utf-8;";
response.Headers.TryAdd("Cache-Control", "no-cache");
response.Headers.TryAdd("Connection", "keep-alive");
var sourceModelId = modelId;
// 处理 yi- 前缀
if (!string.IsNullOrEmpty(modelId) &&
modelId.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
modelId = modelId[3..];
}
var modelDescribe = await GetModelAsync(apiType, sourceModelId);
// 公共缓存队列
var messageQueue = new ConcurrentQueue<string>();
var outputInterval = TimeSpan.FromMilliseconds(75);
var isComplete = false;
// 公共消费任务
var outputTask = Task.Run(async () =>
{
while (!(isComplete && messageQueue.IsEmpty))
{
if (messageQueue.TryDequeue(out var message))
{
await response.WriteAsync(message, Encoding.UTF8, cancellationToken).ConfigureAwait(false);
await response.Body.FlushAsync(cancellationToken).ConfigureAwait(false);
}
if (!isComplete)
{
await Task.Delay(outputInterval, cancellationToken).ConfigureAwait(false);
}
else
{
await Task.Delay(10, cancellationToken).ConfigureAwait(false);
}
}
}, cancellationToken);
StreamProcessResult? processResult = null;
switch (apiType)
{
case ModelApiTypeEnum.Completions:
processResult = await ProcessCompletionsStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
break;
case ModelApiTypeEnum.Messages:
processResult = await ProcessAnthropicStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
break;
case ModelApiTypeEnum.Responses:
processResult = await ProcessOpenAiResponsesStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
break;
case ModelApiTypeEnum.GenerateContent:
processResult = await ProcessGeminiStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
break;
default:
throw new UserFriendlyException($"不支持的API类型: {apiType}");
}
// 标记完成并等待消费任务结束
isComplete = true;
await outputTask;
// 统一的统计处理
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
new MessageInputDto
{
Content = sessionId is null ? "不予存储" : processResult?.UserContent ?? string.Empty,
ModelId = sourceModelId,
TokenUsage = processResult?.TokenUsage,
}, tokenId);
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
new MessageInputDto
{
Content = sessionId is null ? "不予存储" : processResult?.SystemContent ?? string.Empty,
ModelId = sourceModelId,
TokenUsage = processResult?.TokenUsage
}, tokenId);
await _usageStatisticsManager.SetUsageAsync(userId, sourceModelId, processResult?.TokenUsage, tokenId);
// 扣减尊享token包用量
if (userId.HasValue && processResult?.TokenUsage is not null && modelDescribe.IsPremium)
{
var totalTokens = processResult?.TokenUsage.TotalTokens ?? 0;
if (totalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
}
}
}
#region - API类型的具体实现
/// <summary>
/// 流式处理结果,包含用户输入、系统输出和 token 使用情况
/// </summary>
private class StreamProcessResult
{
public string UserContent { get; set; } = string.Empty;
public string SystemContent { get; set; } = string.Empty;
public ThorUsageResponse TokenUsage { get; set; } = new();
}
/// <summary>
/// 处理 OpenAI Completions 格式流式响应
/// </summary>
private async Task<StreamProcessResult> ProcessCompletionsStreamAsync(
ConcurrentQueue<string> messageQueue,
JsonElement requestBody,
AiModelDescribe modelDescribe,
CancellationToken cancellationToken)
{
var request = requestBody.Deserialize<ThorChatCompletionsRequest>(ThorJsonSerializer.DefaultOptions)!;
_specialCompatible.Compatible(request);
// 提取用户最后一条消息
var userContent = request.Messages?.LastOrDefault()?.MessagesStore ?? string.Empty;
// 处理 yi- 前缀
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
request.Model = request.Model[3..];
}
var chatService = LazyServiceProvider.GetRequiredKeyedService<IChatCompletionService>(modelDescribe.HandlerName);
var completeChatResponse = chatService.CompleteChatStreamAsync(modelDescribe, request, cancellationToken);
var tokenUsage = new ThorUsageResponse();
var systemContentBuilder = new StringBuilder();
try
{
await foreach (var data in completeChatResponse)
{
data.SupplementalMultiplier(modelDescribe.Multiplier);
if (data.Usage is not null && (data.Usage.CompletionTokens > 0 || data.Usage.OutputTokens > 0))
{
tokenUsage = data.Usage;
}
// 累加系统输出内容 (choices[].delta.content)
var deltaContent = data.Choices?.FirstOrDefault()?.Delta?.Content;
if (!string.IsNullOrEmpty(deltaContent))
{
systemContentBuilder.Append(deltaContent);
}
var message = JsonSerializer.Serialize(data, ThorJsonSerializer.DefaultOptions);
messageQueue.Enqueue($"data: {message}\n\n");
}
}
catch (Exception e)
{
_logger.LogError(e, "Ai对话异常");
var errorContent = $"对话Ai异常异常信息\n当前Ai模型{request.Model}\n异常信息{e.Message}\n异常堆栈{e}";
systemContentBuilder.Append(errorContent);
var model = new ThorChatCompletionsResponse()
{
Choices = new List<ThorChatChoiceResponse>()
{
new ThorChatChoiceResponse()
{
Delta = new ThorChatMessage()
{
Content = errorContent
}
}
}
};
var errorMessage = JsonConvert.SerializeObject(model, new JsonSerializerSettings
{
ContractResolver = new CamelCasePropertyNamesContractResolver()
});
messageQueue.Enqueue($"data: {errorMessage}\n\n");
}
messageQueue.Enqueue("data: [DONE]\n\n");
return new StreamProcessResult
{
UserContent = userContent,
SystemContent = systemContentBuilder.ToString(),
TokenUsage = tokenUsage
};
}
/// <summary>
/// 处理 Anthropic Messages 格式流式响应
/// </summary>
private async Task<StreamProcessResult> ProcessAnthropicStreamAsync(
ConcurrentQueue<string> messageQueue,
JsonElement requestBody,
AiModelDescribe modelDescribe,
CancellationToken cancellationToken)
{
var request = requestBody.Deserialize<AnthropicInput>(ThorJsonSerializer.DefaultOptions)!;
_specialCompatible.AnthropicCompatible(request);
// 提取用户最后一条消息
var lastMessage = request.Messages?.LastOrDefault();
var userContent = lastMessage?.Content ?? string.Empty;
if (string.IsNullOrEmpty(userContent) && lastMessage?.Contents != null && lastMessage.Contents.Any())
{
// 如果是 Contents 数组,提取第一个 text 类型的内容
var textContent = lastMessage.Contents.FirstOrDefault(c => c.Type == "text");
userContent = textContent?.Text ?? System.Text.Json.JsonSerializer.Serialize(lastMessage.Contents);
}
// 处理 yi- 前缀
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
request.Model = request.Model[3..];
}
var chatService = LazyServiceProvider.GetRequiredKeyedService<IAnthropicChatCompletionService>(modelDescribe.HandlerName);
var completeChatResponse = chatService.StreamChatCompletionsAsync(modelDescribe, request, cancellationToken);
var tokenUsage = new ThorUsageResponse();
var systemContentBuilder = new StringBuilder();
try
{
await foreach (var responseResult in completeChatResponse)
{
// 部分供应商message_start放一部分
if (responseResult.Item1.Contains("message_start"))
{
var currentTokenUsage = responseResult.Item2?.Message?.Usage;
if (currentTokenUsage != null)
{
if ((currentTokenUsage.InputTokens ?? 0) != 0)
{
tokenUsage.InputTokens = (currentTokenUsage.InputTokens ?? 0) +
(currentTokenUsage.CacheCreationInputTokens ?? 0) +
(currentTokenUsage.CacheReadInputTokens ?? 0);
}
if ((currentTokenUsage.OutputTokens ?? 0) != 0)
{
tokenUsage.OutputTokens = currentTokenUsage.OutputTokens;
}
}
}
// message_delta又放一部分
if (responseResult.Item1.Contains("message_delta"))
{
var currentTokenUsage = responseResult.Item2?.Usage;
if (currentTokenUsage != null)
{
if ((currentTokenUsage.InputTokens ?? 0) != 0)
{
tokenUsage.InputTokens = (currentTokenUsage.InputTokens ?? 0) +
(currentTokenUsage.CacheCreationInputTokens ?? 0) +
(currentTokenUsage.CacheReadInputTokens ?? 0);
}
if ((currentTokenUsage.OutputTokens ?? 0) != 0)
{
tokenUsage.OutputTokens = currentTokenUsage.OutputTokens;
}
}
}
// 累加系统输出内容 (delta.text)
var deltaText = responseResult.Item2?.Delta?.Text;
if (!string.IsNullOrEmpty(deltaText))
{
systemContentBuilder.Append(deltaText);
}
// 序列化为SSE格式字符串
var data = JsonSerializer.Serialize(responseResult.Item2, ThorJsonSerializer.DefaultOptions);
messageQueue.Enqueue($"{responseResult.Item1.Trim()}\ndata: {data}\n\n");
}
}
catch (Exception e)
{
_logger.LogError(e, "Ai对话异常");
var errorContent = $"对话Ai异常异常信息\n当前Ai模型{request.Model}\n异常信息{e.Message}\n异常堆栈{e}";
systemContentBuilder.Append(errorContent);
throw new UserFriendlyException(errorContent);
}
tokenUsage.TotalTokens = (tokenUsage.InputTokens ?? 0) + (tokenUsage.OutputTokens ?? 0);
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
return new StreamProcessResult
{
UserContent = userContent,
SystemContent = systemContentBuilder.ToString(),
TokenUsage = tokenUsage
};
}
/// <summary>
/// 处理 OpenAI Responses 格式流式响应
/// </summary>
private async Task<StreamProcessResult> ProcessOpenAiResponsesStreamAsync(
ConcurrentQueue<string> messageQueue,
JsonElement requestBody,
AiModelDescribe modelDescribe,
CancellationToken cancellationToken)
{
var request = requestBody.Deserialize<OpenAiResponsesInput>(ThorJsonSerializer.DefaultOptions)!;
// 提取用户输入内容 (input 字段可能是字符串或数组)
var userContent = string.Empty;
if (request.Input.ValueKind == JsonValueKind.String)
{
userContent = request.Input.GetString() ?? string.Empty;
}
else if (request.Input.ValueKind == JsonValueKind.Array)
{
// 获取最后一个 user 角色的消息
var inputArray = request.Input.EnumerateArray().ToList();
var lastUserMessage = inputArray.LastOrDefault(x =>
x.TryGetProperty("role", out var role) && role.GetString() == "user");
if (lastUserMessage.ValueKind != JsonValueKind.Undefined)
{
if (lastUserMessage.TryGetProperty("content", out var content))
{
userContent = content.GetString() ?? string.Empty;
}
}
}
// 处理 yi- 前缀
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
{
request.Model = request.Model[3..];
}
var chatService = LazyServiceProvider.GetRequiredKeyedService<IOpenAiResponseService>(modelDescribe.HandlerName);
var completeChatResponse = chatService.ResponsesStreamAsync(modelDescribe, request, cancellationToken);
ThorUsageResponse? tokenUsage = null;
var systemContentBuilder = new StringBuilder();
try
{
await foreach (var responseResult in completeChatResponse)
{
// 提取输出文本内容 (response.output_text.delta 事件)
if (responseResult.Item1.Contains("response.output_text.delta"))
{
var delta = responseResult.Item2?.GetPath("delta").GetString();
if (!string.IsNullOrEmpty(delta))
{
systemContentBuilder.Append(delta);
}
}
if (responseResult.Item1.Contains("response.completed"))
{
var obj = responseResult.Item2!.Value;
int inputTokens = obj.GetPath("response", "usage", "input_tokens").GetInt();
int outputTokens = obj.GetPath("response", "usage", "output_tokens").GetInt();
inputTokens = Convert.ToInt32(inputTokens * modelDescribe.Multiplier);
outputTokens = Convert.ToInt32(outputTokens * modelDescribe.Multiplier);
tokenUsage = new ThorUsageResponse
{
PromptTokens = inputTokens,
InputTokens = inputTokens,
OutputTokens = outputTokens,
CompletionTokens = outputTokens,
TotalTokens = inputTokens + outputTokens,
};
}
// 序列化为SSE格式字符串
var data = JsonSerializer.Serialize(responseResult.Item2, ThorJsonSerializer.DefaultOptions);
messageQueue.Enqueue($"{responseResult.Item1.Trim()}\ndata: {data}\n\n");
}
}
catch (Exception e)
{
_logger.LogError(e, "Ai响应异常");
var errorContent = $"响应Ai异常异常信息\n当前Ai模型{request.Model}\n异常信息{e.Message}\n异常堆栈{e}";
systemContentBuilder.Append(errorContent);
throw new UserFriendlyException(errorContent);
}
return new StreamProcessResult
{
UserContent = userContent,
SystemContent = systemContentBuilder.ToString(),
TokenUsage = tokenUsage ?? new ThorUsageResponse()
};
}
/// <summary>
/// 处理 Gemini GenerateContent 格式流式响应
/// </summary>
private async Task<StreamProcessResult> ProcessGeminiStreamAsync(
ConcurrentQueue<string> messageQueue,
JsonElement requestBody,
AiModelDescribe modelDescribe,
CancellationToken cancellationToken)
{
// 提取用户最后一条消息 (contents[last].parts[last].text)
var userContent = GeminiGenerateContentAcquirer.GetLastUserContent(requestBody);
var chatService = LazyServiceProvider.GetRequiredKeyedService<IGeminiGenerateContentService>(modelDescribe.HandlerName);
var completeChatResponse = chatService.GenerateContentStreamAsync(modelDescribe, requestBody, cancellationToken);
ThorUsageResponse? tokenUsage = null;
var systemContentBuilder = new StringBuilder();
try
{
await foreach (var responseResult in completeChatResponse)
{
// 累加系统输出内容 (candidates[0].content.parts[].text排除 thought)
var textContent = GeminiGenerateContentAcquirer.GetTextContent(responseResult!.Value);
if (!string.IsNullOrEmpty(textContent))
{
systemContentBuilder.Append(textContent);
}
if (responseResult!.Value.GetPath("candidates", 0, "finishReason").GetString() == "STOP")
{
tokenUsage = GeminiGenerateContentAcquirer.GetUsage(responseResult!.Value);
// 如果是图片模型,单独扣费
if (modelDescribe.ModelType == ModelTypeEnum.Image)
{
tokenUsage = new ThorUsageResponse
{
InputTokens = (int)modelDescribe.Multiplier,
OutputTokens = (int)modelDescribe.Multiplier,
TotalTokens = (int)modelDescribe.Multiplier
};
}
else
{
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
}
}
messageQueue.Enqueue($"data: {JsonSerializer.Serialize(responseResult)}\n\n");
}
}
catch (Exception e)
{
_logger.LogError(e, "Ai生成异常");
var errorContent = $"生成Ai异常异常信息\n当前Ai模型{modelDescribe.ModelId}\n异常信息{e.Message}\n异常堆栈{e}";
systemContentBuilder.Append(errorContent);
throw new UserFriendlyException(errorContent);
}
return new StreamProcessResult
{
UserContent = userContent,
SystemContent = systemContentBuilder.ToString(),
TokenUsage = tokenUsage ?? new ThorUsageResponse()
};
}
#endregion
2025-12-11 17:16:21 +08:00
#region Http响应
2025-10-11 15:25:43 +08:00
private static readonly byte[] EventPrefix = "event: "u8.ToArray();
private static readonly byte[] DataPrefix = "data: "u8.ToArray();
private static readonly byte[] NewLine = "\n"u8.ToArray();
private static readonly byte[] DoubleNewLine = "\n\n"u8.ToArray();
/// <summary>
/// 使用 JsonSerializer.SerializeAsync 直接序列化到响应流
/// </summary>
private static async ValueTask WriteAsEventStreamDataAsync<T>(
HttpContext context,
string @event,
T value,
CancellationToken cancellationToken = default)
{
var response = context.Response;
var bodyStream = response.Body;
// 确保 SSE Header 已经设置好
// e.g. Content-Type: text/event-stream; charset=utf-8
await response.StartAsync(cancellationToken).ConfigureAwait(false);
// 写事件类型
//此处事件前缀重复了
// await bodyStream.WriteAsync(EventPrefix, cancellationToken).ConfigureAwait(false);
2025-10-11 15:25:43 +08:00
await WriteUtf8StringAsync(bodyStream, @event.Trim(), cancellationToken).ConfigureAwait(false);
await bodyStream.WriteAsync(NewLine, cancellationToken).ConfigureAwait(false);
// 写 data: + JSON
await bodyStream.WriteAsync(DataPrefix, cancellationToken).ConfigureAwait(false);
await JsonSerializer.SerializeAsync(
bodyStream,
value,
ThorJsonSerializer.DefaultOptions,
cancellationToken
).ConfigureAwait(false);
// 事件结束 \n\n
await bodyStream.WriteAsync(DoubleNewLine, cancellationToken).ConfigureAwait(false);
// 及时把数据发送给客户端
await bodyStream.FlushAsync(cancellationToken).ConfigureAwait(false);
}
private static async ValueTask WriteUtf8StringAsync(Stream stream, string value, CancellationToken token)
{
if (string.IsNullOrEmpty(value))
return;
var buffer = Encoding.UTF8.GetBytes(value);
await stream.WriteAsync(buffer, token).ConfigureAwait(false);
}
#endregion
2025-06-21 01:08:14 +08:00
}