附录F:多模态基础#
F.1 多模态:VQVAE、VQGAN、扩散模型#
在构建故事讲述AI大语言模型的过程中,多模态能力的整合变得越来越重要。多模态AI系统能够理解和生成多种形式的信息,如文本、图像、音频和视频,从而创造更丰富、更沉浸式的故事体验。本节将深入探讨多模态AI的核心技术,包括VQVAE、VQGAN、扩散模型,以及如何基于LoRA技术训练和整合CLIP、BLIP2、LLaVA和Qwen-vl等多模态模型。
多模态AI的基础概念#
多模态AI系统能够处理和整合来自不同感知通道(模态)的信息,类似于人类同时使用视觉、听觉和语言能力来理解世界。
多模态表示学习#
多模态表示学习的目标是创建能够捕捉不同模态之间关系的联合表示空间。
联合嵌入空间: 将不同模态的数据映射到同一个语义空间,使得相关内容在该空间中彼此接近。
import torch import torch.nn as nn class JointEmbeddingModel(nn.Module): def __init__(self, text_dim, image_dim, joint_dim): super().__init__() self.text_encoder = nn.Linear(text_dim, joint_dim) self.image_encoder = nn.Linear(image_dim, joint_dim) def encode_text(self, text_features): return self.text_encoder(text_features) def encode_image(self, image_features): return self.image_encoder(image_features) def compute_similarity(self, text_features, image_features): text_embeddings = self.encode_text(text_features) image_embeddings = self.encode_image(image_features) # 归一化嵌入 text_embeddings = text_embeddings / text_embeddings.norm(dim=1, keepdim=True) image_embeddings = image_embeddings / image_embeddings.norm(dim=1, keepdim=True) # 计算余弦相似度 similarity = torch.matmul(text_embeddings, image_embeddings.t()) return similarity
跨模态注意力: 允许一个模态的表示关注另一个模态中的相关部分。
class CrossModalAttention(nn.Module): def __init__(self, query_dim, key_dim, value_dim, hidden_dim): super().__init__() self.query_proj = nn.Linear(query_dim, hidden_dim) self.key_proj = nn.Linear(key_dim, hidden_dim) self.value_proj = nn.Linear(value_dim, hidden_dim) self.scale = hidden_dim ** 0.5 def forward(self, query, key, value): # 投影查询、键、值 q = self.query_proj(query) # [batch_size, query_len, hidden_dim] k = self.key_proj(key) # [batch_size, key_len, hidden_dim] v = self.value_proj(value) # [batch_size, key_len, hidden_dim] # 计算注意力分数 attn_scores = torch.matmul(q, k.transpose(-1, -2)) / self.scale # 注意力权重 attn_weights = torch.softmax(attn_scores, dim=-1) # 应用注意力权重 output = torch.matmul(attn_weights, v) return output, attn_weights
多模态融合: 将不同模态的信息整合为统一的表示。
class MultimodalFusion(nn.Module): def __init__(self, text_dim, image_dim, fusion_dim): super().__init__() self.text_proj = nn.Linear(text_dim, fusion_dim) self.image_proj = nn.Linear(image_dim, fusion_dim) self.fusion_layer = nn.Sequential( nn.Linear(fusion_dim * 2, fusion_dim), nn.ReLU(), nn.Linear(fusion_dim, fusion_dim) ) def forward(self, text_features, image_features): # 投影特征 text_proj = self.text_proj(text_features) image_proj = self.image_proj(image_features) # 连接特征 concat_features = torch.cat([text_proj, image_proj], dim=-1) # 融合 fused_features = self.fusion_layer(concat_features) return fused_features
多模态预训练目标#
多模态模型通常使用以下预训练目标:
对比学习: 训练模型区分正样本对(相关的文本-图像对)和负样本对(不相关的文本-图像对)。
def contrastive_loss(text_embeddings, image_embeddings, temperature=0.07): # 归一化嵌入 text_embeddings = text_embeddings / text_embeddings.norm(dim=1, keepdim=True) image_embeddings = image_embeddings / image_embeddings.norm(dim=1, keepdim=True) # 计算相似度矩阵 similarity = torch.matmul(text_embeddings, image_embeddings.t()) / temperature # 对角线上的元素是正样本对 labels = torch.arange(similarity.size(0), device=similarity.device) # 计算文本到图像和图像到文本的损失 loss_t2i = nn.CrossEntropyLoss()(similarity, labels) loss_i2t = nn.CrossEntropyLoss()(similarity.t(), labels) # 总损失 loss = (loss_t2i + loss_i2t) / 2 return loss
掩码语言建模: 预测被掩码的文本标记,类似于BERT,但使用多模态上下文。
def masked_language_modeling(text_input, image_features, text_model, fusion_model, mask_token_id, vocab_size): # 创建掩码 masked_input = text_input.clone() mask = torch.rand(text_input.shape) < 0.15 # 掩码15%的标记 masked_input[mask] = mask_token_id # 获取文本特征 text_features = text_model(masked_input) # 融合文本和图像特征 fused_features = fusion_model(text_features, image_features) # 预测被掩码的标记 logits = nn.Linear(fused_features.size(-1), vocab_size)(fused_features) # 计算损失(仅对被掩码的位置) loss = nn.CrossEntropyLoss()(logits[mask], text_input[mask]) return loss
图像-文本匹配: 预测给定的图像-文本对是否匹配。
def image_text_matching(text_features, image_features, fusion_model, labels): # 融合特征 fused_features = fusion_model(text_features, image_features) # 二分类:匹配或不匹配 logits = nn.Linear(fused_features.size(-1), 2)(fused_features) # 计算损失 loss = nn.CrossEntropyLoss()(logits, labels) return loss
VQVAE:向量量化变分自编码器#
VQVAE(Vector Quantized Variational Autoencoder)是一种生成模型,它通过离散的潜在表示来压缩和重构数据,特别适用于图像和音频等高维数据。
VQVAE的核心原理#
VQVAE的核心思想是将输入数据编码为离散的潜在表示,然后从这些离散表示中重构原始数据。
编码器: 将输入数据映射到连续的潜在空间。
向量量化: 将连续的潜在表示映射到最近的离散码本向量。
解码器: 从量化的潜在表示重构原始数据。
import torch
import torch.nn as nn
import torch.nn.functional as F
class VectorQuantizer(nn.Module):
def __init__(self, num_embeddings, embedding_dim, commitment_cost):
super().__init__()
self.num_embeddings = num_embeddings
self.embedding_dim = embedding_dim
self.commitment_cost = commitment_cost
# 创建码本
self.codebook = nn.Embedding(num_embeddings, embedding_dim)
self.codebook.weight.data.uniform_(-1.0 / num_embeddings, 1.0 / num_embeddings)
def forward(self, z):
# z: [batch_size, embedding_dim, height, width]
# 调整z的形状以计算距离
z = z.permute(0, 2, 3, 1).contiguous() # [batch_size, height, width, embedding_dim]
z_flattened = z.view(-1, self.embedding_dim)
# 计算z_flattened和码本向量之间的距离
d = torch.sum(z_flattened ** 2, dim=1, keepdim=True) + \
torch.sum(self.codebook.weight ** 2, dim=1) - \
2 * torch.matmul(z_flattened, self.codebook.weight.t())
# 找到最近的码本条目
min_encoding_indices = torch.argmin(d, dim=1)
min_encodings = torch.zeros(min_encoding_indices.shape[0], self.num_embeddings)
min_encodings.scatter_(1, min_encoding_indices.unsqueeze(1), 1)
# 量化
z_q = torch.matmul(min_encodings, self.codebook.weight)
z_q = z_q.view(z.shape)
# 计算损失
# 1. 码本损失
codebook_loss = F.mse_loss(z_q.detach(), z)
# 2. 承诺损失
commitment_loss = F.mse_loss(z, z_q.detach())
# 3. 总损失
loss = codebook_loss + self.commitment_cost * commitment_loss
# 直通估计器
z_q = z + (z_q - z).detach()
# 调整z_q的形状以匹配编码器输出
z_q = z_q.permute(0, 3, 1, 2).contiguous()
return z_q, loss, min_encoding_indices
VQVAE的完整实现#
下面是一个简化的VQVAE实现,适用于图像数据:
class Encoder(nn.Module):
def __init__(self, in_channels, hidden_channels, embedding_dim):
super().__init__()
self.conv1 = nn.Conv2d(in_channels, hidden_channels, kernel_size=4, stride=2, padding=1)
self.conv2 = nn.Conv2d(hidden_channels, hidden_channels, kernel_size=4, stride=2, padding=1)
self.conv3 = nn.Conv2d(hidden_channels, hidden_channels, kernel_size=4, stride=2, padding=1)
self.conv4 = nn.Conv2d(hidden_channels, embedding_dim, kernel_size=3, stride=1, padding=1)
def forward(self, x):
x = F.relu(self.conv1(x))
x = F.relu(self.conv2(x))
x = F.relu(self.conv3(x))
x = self.conv4(x)
return x
class Decoder(nn.Module):
def __init__(self, embedding_dim, hidden_channels, out_channels):
super().__init__()
self.conv1 = nn.Conv2d(embedding_dim, hidden_channels, kernel_size=3, stride=1, padding=1)
self.conv2 = nn.ConvTranspose2d(hidden_channels, hidden_channels, kernel_size=4, stride=2, padding=1)
self.conv3 = nn.ConvTranspose2d(hidden_channels, hidden_channels, kernel_size=4, stride=2, padding=1)
self.conv4 = nn.ConvTranspose2d(hidden_channels, out_channels, kernel_size=4, stride=2, padding=1)
def forward(self, x):
x = F.relu(self.conv1(x))
x = F.relu(self.conv2(x))
x = F.relu(self.conv3(x))
x = torch.tanh(self.conv4(x))
return x
class VQVAE(nn.Module):
def __init__(self, in_channels, hidden_channels, embedding_dim, num_embeddings, commitment_cost):
super().__init__()
self.encoder = Encoder(in_channels, hidden_channels, embedding_dim)
self.vector_quantizer = VectorQuantizer(num_embeddings, embedding_dim, commitment_cost)
self.decoder = Decoder(embedding_dim, hidden_channels, in_channels)
def forward(self, x):
z = self.encoder(x)
z_q, vq_loss, indices = self.vector_quantizer(z)
x_recon = self.decoder(z_q)
# 计算重构损失
recon_loss = F.mse_loss(x_recon, x)
# 总损失
loss = recon_loss + vq_loss
return x_recon, loss, indices
VQVAE的应用#
VQVAE在多模态AI中有多种应用:
图像压缩和重构: 将图像压缩为离散的潜在表示,然后重构原始图像。
条件图像生成: 基于文本描述生成相应的图像。
图像编辑和操作: 在潜在空间中编辑图像特征,然后重构修改后的图像。
多模态表示学习: 学习不同模态(如文本和图像)之间的共享离散表示。
VQGAN:结合GAN的VQVAE#
VQGAN(Vector Quantized Generative Adversarial Network)是VQVAE的扩展,它结合了GAN(生成对抗网络)的训练方法,以生成更高质量、更真实的图像。
VQGAN的核心原理#
VQGAN在VQVAE的基础上添加了以下关键组件:
判别器: 评估生成图像的真实性,帮助生成器产生更真实的图像。
感知损失: 使用预训练的视觉网络(如VGG)计算特征空间中的损失,而不仅仅是像素空间。
对抗训练: 生成器和判别器进行对抗训练,相互改进。
class Discriminator(nn.Module):
def __init__(self, in_channels, hidden_channels):
super().__init__()
self.layers = nn.Sequential(
nn.Conv2d(in_channels, hidden_channels, kernel_size=4, stride=2, padding=1),
nn.LeakyReLU(0.2),
nn.Conv2d(hidden_channels, hidden_channels * 2, kernel_size=4, stride=2, padding=1),
nn.LeakyReLU(0.2),
nn.Conv2d(hidden_channels * 2, hidden_channels * 4, kernel_size=4, stride=2, padding=1),
nn.LeakyReLU(0.2),
nn.Conv2d(hidden_channels * 4, 1, kernel_size=4, stride=1, padding=0)
)
def forward(self, x):
return self.layers(x)
class PerceptualLoss(nn.Module):
def __init__(self):
super().__init__()
# 加载预训练的VGG16
vgg = torchvision.models.vgg16(pretrained=True)
self.feature_extractor = nn.Sequential(*list(vgg.features.children())[:29])
# 冻结参数
for param in self.feature_extractor.parameters():
param.requires_grad = False
def forward(self, x, y):
# 提取特征
x_features = self.feature_extractor(x)
y_features = self.feature_extractor(y)
# 计算特征空间中的MSE损失
return F.mse_loss(x_features, y_features)
class VQGAN(nn.Module):
def __init__(self, in_channels, hidden_channels, embedding_dim, num_embeddings, commitment_cost):
super().__init__()
self.vqvae = VQVAE(in_channels, hidden_channels, embedding_dim, num_embeddings, commitment_cost)
self.discriminator = Discriminator(in_channels, hidden_channels)
self.perceptual_loss = PerceptualLoss()
def forward(self, x):
x_recon, vq_loss, indices = self.vqvae(x)
# 计算重构损失
recon_loss = F.mse_loss(x_recon, x)
# 计算感知损失
p_loss = self.perceptual_loss(x_recon, x)
# 总损失(不包括对抗损失,它在训练循环中计算)
loss = recon_loss + vq_loss + p_loss
return x_recon, loss, indices
VQGAN的训练过程#
VQGAN的训练涉及生成器(VQVAE部分)和判别器的交替优化:
def train_vqgan(vqgan, dataloader, num_epochs, device, lr=2e-4):
# 优化器
optimizer_g = torch.optim.Adam(vqgan.vqvae.parameters(), lr=lr)
optimizer_d = torch.optim.Adam(vqgan.discriminator.parameters(), lr=lr)
# 对抗损失
adversarial_criterion = nn.BCEWithLogitsLoss()
for epoch in range(num_epochs):
for batch in dataloader:
real_images = batch.to(device)
# 训练判别器
optimizer_d.zero_grad()
# 生成重构图像
with torch.no_grad():
reconstructed_images, _, _ = vqgan.vqvae(real_images)
# 真实图像的判别器输出
real_preds = vqgan.discriminator(real_images)
real_targets = torch.ones_like(real_preds)
real_loss = adversarial_criterion(real_preds, real_targets)
# 重构图像的判别器输出
fake_preds = vqgan.discriminator(reconstructed_images.detach())
fake_targets = torch.zeros_like(fake_preds)
fake_loss = adversarial_criterion(fake_preds, fake_targets)
# 判别器总损失
d_loss = real_loss + fake_loss
d_loss.backward()
optimizer_d.step()
# 训练生成器
optimizer_g.zero_grad()
# 重新生成重构图像
reconstructed_images, vq_loss, _ = vqgan.vqvae(real_images)
# 重构损失
recon_loss = F.mse_loss(reconstructed_images, real_images)
# 感知损失
p_loss = vqgan.perceptual_loss(reconstructed_images, real_images)
# 对抗损失(欺骗判别器)
fake_preds = vqgan.discriminator(reconstructed_images)
g_adversarial_loss = adversarial_criterion(fake_preds, torch.ones_like(fake_preds))
# 生成器总损失
g_loss = recon_loss + vq_loss + p_loss + 0.1 * g_adversarial_loss
g_loss.backward()
optimizer_g.step()
print(f"Epoch [{epoch+1}/{num_epochs}], G Loss: {g_loss.item():.4f}, D Loss: {d_loss.item():.4f}")
VQGAN的应用#
VQGAN在多模态AI中的应用包括:
高质量图像生成: 生成比VQVAE更真实、更高质量的图像。
文本到图像生成: 结合语言模型,根据文本描述生成相应的图像。
图像编辑和操作: 在潜在空间中进行语义编辑,生成修改后的图像。
风格迁移: 将一种图像风格转换为另一种风格,同时保持内容不变。
扩散模型#
扩散模型是近年来在图像生成领域取得巨大成功的模型类型,如DALL-E 2、Stable Diffusion和Midjourney等都基于扩散模型。
扩散模型的核心原理#
扩散模型基于两个过程:
前向扩散过程: 逐步向数据添加噪声,直到数据变为纯噪声。
反向扩散过程: 学习如何逐步去除噪声,从纯噪声恢复原始数据。
import torch
import torch.nn as nn
import torch.nn.functional as F
class SimpleUNet(nn.Module):
"""简化的U-Net模型,用于扩散模型的噪声预测"""
def __init__(self, in_channels, out_channels, hidden_channels):
super().__init__()
# 下采样路径
self.down1 = nn.Conv2d(in_channels, hidden_channels, 3, padding=1)
self.down2 = nn.Conv2d(hidden_channels, hidden_channels * 2, 4, stride=2, padding=1)
self.down3 = nn.Conv2d(hidden_channels * 2, hidden_channels * 4, 4, stride=2, padding=1)
# 时间嵌入
self.time_mlp = nn.Sequential(
nn.Linear(1, hidden_channels * 4),
nn.ReLU(),
nn.Linear(hidden_channels * 4, hidden_channels * 4)
)
# 中间层
self.mid = nn.Sequential(
nn.Conv2d(hidden_channels * 4, hidden_channels * 4, 3, padding=1),
nn.ReLU(),
nn.Conv2d(hidden_channels * 4, hidden_channels * 4, 3, padding=1)
)
# 上采样路径
self.up1 = nn.ConvTranspose2d(hidden_channels * 8, hidden_channels * 2, 4, stride=2, padding=1)
self.up2 = nn.ConvTranspose2d(hidden_channels * 4, hidden_channels, 4, stride=2, padding=1)
self.up3 = nn.Conv2d(hidden_channels * 2, out_channels, 3, padding=1)
def forward(self, x, t):
# 下采样
x1 = F.relu(self.down1(x))
x2 = F.relu(self.down2(x1))
x3 = F.relu(self.down3(x2))
# 时间嵌入
t = self.time_mlp(t.unsqueeze(1).float())
t = t.unsqueeze(-1).unsqueeze(-1)
# 中间层
x3 = x3 + t
x3 = self.mid(x3)
# 上采样
x = F.relu(self.up1(torch.cat([x3, x3], dim=1)))
x = F.relu(self.up2(torch.cat([x, x2], dim=1)))
x = self.up3(torch.cat([x, x1], dim=1))
return x
class DiffusionModel(nn.Module):
def __init__(self, model, beta_start=1e-4, beta_end=0.02, timesteps=1000):
super().__init__()
self.model = model
self.timesteps = timesteps
# 定义噪声调度
self.betas = torch.linspace(beta_start, beta_end, timesteps)
self.alphas = 1 - self.betas
self.alphas_cumprod = torch.cumprod(self.alphas, dim=0)
def forward(self, x_0, t):
"""预测添加到x_0的噪声"""
# 获取批次大小
batch_size = x_0.shape[0]
# 采样噪声
epsilon = torch.randn_like(x_0)
# 选择对应时间步的alpha_cumprod
alpha_cumprod_t = self.alphas_cumprod[t]
alpha_cumprod_t = alpha_cumprod_t.view(-1, 1, 1, 1)
# 对x_0添加噪声
x_t = torch.sqrt(alpha_cumprod_t) * x_0 + torch.sqrt(1 - alpha_cumprod_t) * epsilon
# 预测噪声
predicted_noise = self.model(x_t, t)
# 计算损失
loss = F.mse_loss(predicted_noise, epsilon)
return loss
@torch.no_grad()
def sample(self, batch_size, image_size, channels, device):
"""从噪声生成图像"""
# 从纯噪声开始
x = torch.randn(batch_size, channels, image_size, image_size, device=device)
# 逐步去噪
for t in reversed(range(self.timesteps)):
t_batch = torch.ones(batch_size, device=device).long() * t
# 预测噪声
predicted_noise = self.model(x, t_batch)
# 计算去噪参数
alpha = self.alphas[t]
alpha_cumprod = self.alphas_cumprod[t]
beta = self.betas[t]
# 无噪声添加的情况
if t > 0:
noise = torch.randn_like(x)
else:
noise = torch.zeros_like(x)
# 更新x
x = (1 / torch.sqrt(alpha)) * (x - ((1 - alpha) / torch.sqrt(1 - alpha_cumprod)) * predicted_noise) + torch.sqrt(beta) * noise
return x
条件扩散模型#
条件扩散模型允许基于某些条件(如文本描述)生成图像:
class ConditionalUNet(nn.Module):
"""条件U-Net模型,用于基于文本条件的扩散模型"""
def __init__(self, in_channels, out_channels, hidden_channels, context_dim):
super().__init__()
# 下采样路径
self.down1 = nn.Conv2d(in_channels, hidden_channels, 3, padding=1)
self.down2 = nn.Conv2d(hidden_channels, hidden_channels * 2, 4, stride=2, padding=1)
self.down3 = nn.Conv2d(hidden_channels * 2, hidden_channels * 4, 4, stride=2, padding=1)
# 时间嵌入
self.time_mlp = nn.Sequential(
nn.Linear(1, hidden_channels * 4),
nn.ReLU(),
nn.Linear(hidden_channels * 4, hidden_channels * 4)
)
# 上下文嵌入(用于文本条件)
self.context_proj = nn.Linear(context_dim, hidden_channels * 4)
# 中间层(包含交叉注意力)
self.mid_block1 = nn.Conv2d(hidden_channels * 4, hidden_channels * 4, 3, padding=1)
self.mid_attn = CrossAttention(hidden_channels * 4, context_dim)
self.mid_block2 = nn.Conv2d(hidden_channels * 4, hidden_channels * 4, 3, padding=1)
# 上采样路径
self.up1 = nn.ConvTranspose2d(hidden_channels * 8, hidden_channels * 2, 4, stride=2, padding=1)
self.up2 = nn.ConvTranspose2d(hidden_channels * 4, hidden_channels, 4, stride=2, padding=1)
self.up3 = nn.Conv2d(hidden_channels * 2, out_channels, 3, padding=1)
def forward(self, x, t, context):
# 下采样
x1 = F.relu(self.down1(x))
x2 = F.relu(self.down2(x1))
x3 = F.relu(self.down3(x2))
# 时间嵌入
t_emb = self.time_mlp(t.unsqueeze(1).float())
t_emb = t_emb.unsqueeze(-1).unsqueeze(-1)
# 上下文处理
context_emb = self.context_proj(context)
# 中间层(添加时间嵌入)
h = x3 + t_emb
h = F.relu(self.mid_block1(h))
# 应用交叉注意力
b, c, h, w = h.shape
h_flat = h.view(b, c, -1).transpose(1, 2) # [B, H*W, C]
h_flat = self.mid_attn(h_flat, context) # 应用交叉注意力
h = h_flat.transpose(1, 2).view(b, c, h, w) # 恢复形状
h = F.relu(self.mid_block2(h))
# 上采样
h = F.relu(self.up1(torch.cat([h, x3], dim=1)))
h = F.relu(self.up2(torch.cat([h, x2], dim=1)))
h = self.up3(torch.cat([h, x1], dim=1))
return h
class CrossAttention(nn.Module):
"""交叉注意力模块,用于融合图像和文本特征"""
def __init__(self, query_dim, context_dim, heads=8, dim_head=64):
super().__init__()
inner_dim = dim_head * heads
self.heads = heads
self.scale = dim_head ** -0.5
self.to_q = nn.Linear(query_dim, inner_dim, bias=False)
self.to_k = nn.Linear(context_dim, inner_dim, bias=False)
self.to_v = nn.Linear(context_dim, inner_dim, bias=False)
self.to_out = nn.Linear(inner_dim, query_dim)
def forward(self, x, context):
h = self.heads
q = self.to_q(x)
k = self.to_k(context)
v = self.to_v(context)
q, k, v = map(lambda t: t.reshape(*t.shape[:-1], h, -1).transpose(-3, -2), (q, k, v))
# 注意力计算
sim = torch.einsum('b h i d, b h j d -> b h i j', q, k) * self.scale
attn = sim.softmax(dim=-1)
out = torch.einsum('b h i j, b h j d -> b h i d', attn, v)
out = out.transpose(-3, -2).reshape(*x.shape[:-1], -1)
return self.to_out(out)
class ConditionalDiffusionModel(nn.Module):
def __init__(self, model, beta_start=1e-4, beta_end=0.02, timesteps=1000):
super().__init__()
self.model = model
self.timesteps = timesteps
# 定义噪声调度
self.betas = torch.linspace(beta_start, beta_end, timesteps)
self.alphas = 1 - self.betas
self.alphas_cumprod = torch.cumprod(self.alphas, dim=0)
def forward(self, x_0, t, context):
"""预测添加到x_0的噪声,基于上下文条件"""
# 获取批次大小
batch_size = x_0.shape[0]
# 采样噪声
epsilon = torch.randn_like(x_0)
# 选择对应时间步的alpha_cumprod
alpha_cumprod_t = self.alphas_cumprod[t]
alpha_cumprod_t = alpha_cumprod_t.view(-1, 1, 1, 1)
# 对x_0添加噪声
x_t = torch.sqrt(alpha_cumprod_t) * x_0 + torch.sqrt(1 - alpha_cumprod_t) * epsilon
# 预测噪声
predicted_noise = self.model(x_t, t, context)
# 计算损失
loss = F.mse_loss(predicted_noise, epsilon)
return loss
@torch.no_grad()
def sample(self, context, batch_size, image_size, channels, device):
"""基于上下文条件从噪声生成图像"""
# 从纯噪声开始
x = torch.randn(batch_size, channels, image_size, image_size, device=device)
# 逐步去噪
for t in reversed(range(self.timesteps)):
t_batch = torch.ones(batch_size, device=device).long() * t
# 预测噪声
predicted_noise = self.model(x, t_batch, context)
# 计算去噪参数
alpha = self.alphas[t]
alpha_cumprod = self.alphas_cumprod[t]
beta = self.betas[t]
# 无噪声添加的情况
if t > 0:
noise = torch.randn_like(x)
else:
noise = torch.zeros_like(x)
# 更新x
x = (1 / torch.sqrt(alpha)) * (x - ((1 - alpha) / torch.sqrt(1 - alpha_cumprod)) * predicted_noise) + torch.sqrt(beta) * noise
return x
扩散变换器(Diffusion Transformer)#
扩散变换器是将Transformer架构应用于扩散模型的方法,特别适用于高分辨率图像生成和多模态任务。
class DiffusionTransformer(nn.Module):
"""基于Transformer的扩散模型"""
def __init__(self, img_size, patch_size, in_channels, hidden_dim, num_heads, num_layers, context_dim=None):
super().__init__()
self.img_size = img_size
self.patch_size = patch_size
self.in_channels = in_channels
# 计算序列长度
self.seq_length = (img_size // patch_size) ** 2
# 图像分块和嵌入
self.patch_embed = nn.Conv2d(in_channels, hidden_dim, kernel_size=patch_size, stride=patch_size)
# 位置嵌入
self.pos_embed = nn.Parameter(torch.zeros(1, self.seq_length, hidden_dim))
# 时间嵌入
self.time_embed = nn.Sequential(
nn.Linear(1, hidden_dim),
nn.SiLU(),
nn.Linear(hidden_dim, hidden_dim)
)
# Transformer块
self.transformer_blocks = nn.ModuleList([
TransformerBlock(hidden_dim, num_heads, context_dim) for _ in range(num_layers)
])
# 输出头
self.to_out = nn.Sequential(
nn.LayerNorm(hidden_dim),
nn.Linear(hidden_dim, patch_size * patch_size * in_channels)
)
def forward(self, x, t, context=None):
# x: [B, C, H, W]
batch_size = x.shape[0]
# 图像分块和嵌入
x = self.patch_embed(x) # [B, hidden_dim, H/patch_size, W/patch_size]
x = x.flatten(2).transpose(1, 2) # [B, seq_len, hidden_dim]
# 添加位置嵌入
x = x + self.pos_embed
# 时间嵌入
t_emb = self.time_embed(t.unsqueeze(1).float()) # [B, hidden_dim]
# 添加时间嵌入到每个位置
x = x + t_emb.unsqueeze(1)
# 通过Transformer块
for block in self.transformer_blocks:
x = block(x, context)
# 输出头
x = self.to_out(x) # [B, seq_len, patch_size*patch_size*in_channels]
# 重塑为图像
x = x.reshape(batch_size, self.img_size // self.patch_size, self.img_size // self.patch_size,
self.patch_size, self.patch_size, self.in_channels)
x = x.permute(0, 5, 1, 3, 2, 4).contiguous()
x = x.reshape(batch_size, self.in_channels, self.img_size, self.img_size)
return x
class TransformerBlock(nn.Module):
"""Transformer块,支持自注意力和交叉注意力"""
def __init__(self, hidden_dim, num_heads, context_dim=None):
super().__init__()
self.norm1 = nn.LayerNorm(hidden_dim)
self.attn1 = nn.MultiheadAttention(hidden_dim, num_heads, batch_first=True)
self.has_cross_attn = context_dim is not None
if self.has_cross_attn:
self.norm2 = nn.LayerNorm(hidden_dim)
self.attn2 = nn.MultiheadAttention(hidden_dim, num_heads, batch_first=True)
self.context_proj = nn.Linear(context_dim, hidden_dim) if context_dim != hidden_dim else nn.Identity()
self.norm3 = nn.LayerNorm(hidden_dim)
self.mlp = nn.Sequential(
nn.Linear(hidden_dim, hidden_dim * 4),
nn.GELU(),
nn.Linear(hidden_dim * 4, hidden_dim)
)
def forward(self, x, context=None):
# 自注意力
x = x + self.attn1(self.norm1(x), self.norm1(x), self.norm1(x))[0]
# 交叉注意力(如果有上下文)
if self.has_cross_attn and context is not None:
context = self.context_proj(context)
x = x + self.attn2(self.norm2(x), context, context)[0]
# 前馈网络
x = x + self.mlp(self.norm3(x))
return x
扩散模型的应用#
扩散模型在多模态AI中有广泛的应用:
文本到图像生成: 基于文本描述生成高质量图像,如Stable Diffusion。
图像编辑和操作: 在保持图像整体结构的同时修改特定区域或属性。
图像到图像转换: 将一种类型的图像转换为另一种类型,如草图到照片。
3D内容生成: 生成3D模型或从不同角度生成一致的图像。
视频生成: 生成连贯的视频序列,保持时间一致性。
多模态模型训练与整合#
在构建故事讲述AI系统时,我们需要训练和整合多种多模态模型,如CLIP、BLIP2、LLaVA和Qwen-vl等。这些模型可以通过LoRA等参数高效微调技术进行定制,以适应特定的故事讲述任务。
基于LoRA的多模态模型训练#
LoRA(Low-Rank Adaptation)是一种参数高效的微调方法,它通过添加低秩矩阵来适应预训练模型,而不需要更新所有参数。
import torch
import torch.nn as nn
import torch.nn.functional as F
class LoRALayer(nn.Module):
"""LoRA适配层"""
def __init__(self, in_features, out_features, rank=4, alpha=1.0):
super().__init__()
self.rank = rank
self.alpha = alpha
self.scaling = alpha / rank
# 低秩矩阵
self.lora_A = nn.Parameter(torch.zeros(in_features, rank))
self.lora_B = nn.Parameter(torch.zeros(rank, out_features))
# 初始化
nn.init.kaiming_uniform_(self.lora_A, a=math.sqrt(5))
nn.init.zeros_(self.lora_B)
def forward(self, x):
# 低秩更新
return self.scaling * (x @ self.lora_A @ self.lora_B)
class LoRALinear(nn.Module):
"""带LoRA的线性层"""
def __init__(self, linear_layer, rank=4, alpha=1.0):
super().__init__()
self.linear = linear_layer
self.lora = LoRALayer(linear_layer.in_features, linear_layer.out_features, rank, alpha)
def forward(self, x):
return self.linear(x) + self.lora(x)
使用LoRA微调CLIP模型#
CLIP(Contrastive Language-Image Pretraining)是一个强大的多模态模型,可以理解图像和文本之间的关系。
from transformers import CLIPModel, CLIPProcessor
def apply_lora_to_clip(clip_model, rank=4, alpha=1.0, target_modules=None):
"""将LoRA应用到CLIP模型的特定模块"""
if target_modules is None:
# 默认目标模块
target_modules = [
"text_model.encoder.layers.{}.self_attn.q_proj",
"text_model.encoder.layers.{}.self_attn.k_proj",
"text_model.encoder.layers.{}.self_attn.v_proj",
"text_model.encoder.layers.{}.self_attn.out_proj",
"vision_model.encoder.layers.{}.self_attn.q_proj",
"vision_model.encoder.layers.{}.self_attn.k_proj",
"vision_model.encoder.layers.{}.self_attn.v_proj",
"vision_model.encoder.layers.{}.self_attn.out_proj"
]
# 冻结所有参数
for param in clip_model.parameters():
param.requires_grad = False
# 应用LoRA到目标模块
for name, module in clip_model.named_modules():
if any(target.format(i) == name for target in target_modules for i in range(12)):
if isinstance(module, nn.Linear):
parent_name = '.'.join(name.split('.')[:-1])
layer_name = name.split('.')[-1]
parent = clip_model.get_submodule(parent_name)
# 替换为LoRA版本
setattr(parent, layer_name, LoRALinear(module, rank, alpha))
return clip_model
def train_clip_with_lora(clip_model, train_dataloader, num_epochs, learning_rate=1e-4):
"""使用LoRA训练CLIP模型"""
# 应用LoRA
clip_model = apply_lora_to_clip(clip_model)
# 获取可训练参数
trainable_params = [p for p in clip_model.parameters() if p.requires_grad]
# 优化器
optimizer = torch.optim.AdamW(trainable_params, lr=learning_rate)
# 训练循环
for epoch in range(num_epochs):
for batch in train_dataloader:
# 获取输入
input_ids = batch["input_ids"].to(clip_model.device)
attention_mask = batch["attention_mask"].to(clip_model.device)
pixel_values = batch["pixel_values"].to(clip_model.device)
# 前向传播
outputs = clip_model(input_ids=input_ids,
attention_mask=attention_mask,
pixel_values=pixel_values,
return_loss=True)
loss = outputs.loss
# 反向传播
optimizer.zero_grad()
loss.backward()
optimizer.step()
print(f"Epoch {epoch+1}/{num_epochs}, Loss: {loss.item():.4f}")
return clip_model
使用LoRA微调BLIP2模型#
BLIP2(Bootstrapping Language-Image Pre-training)是一个先进的视觉-语言模型,结合了冻结的图像编码器和大型语言模型。
from transformers import Blip2Model, Blip2Processor
def apply_lora_to_blip2(blip2_model, rank=4, alpha=1.0):
"""将LoRA应用到BLIP2模型的语言模型部分"""
# 冻结所有参数
for param in blip2_model.parameters():
param.requires_grad = False
# 应用LoRA到语言模型的注意力层
for name, module in blip2_model.language_model.named_modules():
if "self_attn" in name and isinstance(module, nn.Linear):
parent_name = '.'.join(name.split('.')[:-1])
layer_name = name.split('.')[-1]
parent = blip2_model.language_model.get_submodule(parent_name)
# 替换为LoRA版本
setattr(parent, layer_name, LoRALinear(module, rank, alpha))
return blip2_model
def train_blip2_with_lora(blip2_model, train_dataloader, num_epochs, learning_rate=1e-4):
"""使用LoRA训练BLIP2模型"""
# 应用LoRA
blip2_model = apply_lora_to_blip2(blip2_model)
# 获取可训练参数
trainable_params = [p for p in blip2_model.parameters() if p.requires_grad]
# 优化器
optimizer = torch.optim.AdamW(trainable_params, lr=learning_rate)
# 训练循环
for epoch in range(num_epochs):
for batch in train_dataloader:
# 获取输入
input_ids = batch["input_ids"].to(blip2_model.device)
attention_mask = batch["attention_mask"].to(blip2_model.device)
pixel_values = batch["pixel_values"].to(blip2_model.device)
labels = batch["labels"].to(blip2_model.device)
# 前向传播
outputs = blip2_model(input_ids=input_ids,
attention_mask=attention_mask,
pixel_values=pixel_values,
labels=labels)
loss = outputs.loss
# 反向传播
optimizer.zero_grad()
loss.backward()
optimizer.step()
print(f"Epoch {epoch+1}/{num_epochs}, Loss: {loss.item():.4f}")
return blip2_model
多模态模型整合框架#
为了构建一个完整的故事讲述AI系统,我们需要整合多种多模态模型,形成一个统一的框架。
class MultimodalStorytellerFramework:
"""多模态故事讲述框架,整合多种视觉-语言模型"""
def __init__(self, clip_model, blip2_model, llava_model, qwen_vl_model, device="cuda"):
self.clip_model = clip_model.to(device)
self.blip2_model = blip2_model.to(device)
self.llava_model = llava_model.to(device)
self.qwen_vl_model = qwen_vl_model.to(device)
self.clip_processor = CLIPProcessor.from_pretrained("openai/clip-vit-base-patch32")
self.blip2_processor = Blip2Processor.from_pretrained("Salesforce/blip2-opt-2.7b")
self.llava_processor = AutoProcessor.from_pretrained("llava-hf/llava-1.5-7b")
self.qwen_vl_processor = AutoProcessor.from_pretrained("Qwen/Qwen-VL")
self.device = device
def generate_image_from_text(self, text, model_type="stable_diffusion"):
"""基于文本生成图像"""
# 这里可以集成Stable Diffusion或其他文本到图像模型
# 简化示例,实际实现需要集成适当的生成模型
pass
def generate_caption(self, image, max_length=30):
"""为图像生成描述性文本"""
# 使用BLIP2生成图像描述
inputs = self.blip2_processor(images=image, return_tensors="pt").to(self.device)
generated_ids = self.blip2_model.generate(
**inputs,
max_length=max_length,
num_beams=5,
early_stopping=True
)
caption = self.blip2_processor.batch_decode(generated_ids, skip_special_tokens=True)[0]
return caption
def answer_visual_question(self, image, question, model="llava"):
"""回答关于图像的问题"""
if model == "llava":
# 使用LLaVA回答视觉问题
inputs = self.llava_processor(images=image, text=question, return_tensors="pt").to(self.device)
generated_ids = self.llava_model.generate(
**inputs,
max_length=100,
num_beams=5
)
answer = self.llava_processor.batch_decode(generated_ids, skip_special_tokens=True)[0]
return answer
elif model == "qwen_vl":
# 使用Qwen-VL回答视觉问题
inputs = self.qwen_vl_processor(images=image, text=question, return_tensors="pt").to(self.device)
generated_ids = self.qwen_vl_model.generate(
**inputs,
max_length=100,
num_beams=5
)
answer = self.qwen_vl_processor.batch_decode(generated_ids, skip_special_tokens=True)[0]
return answer
def retrieve_similar_images(self, text_query, image_database):
"""基于文本查询检索相似图像"""
# 使用CLIP计算文本和图像之间的相似度
text_inputs = self.clip_processor(text=text_query, return_tensors="pt", padding=True).to(self.device)
text_features = self.clip_model.get_text_features(**text_inputs)
# 计算所有图像的特征
image_features_list = []
for image in image_database:
image_inputs = self.clip_processor(images=image, return_tensors="pt").to(self.device)
image_features = self.clip_model.get_image_features(**image_inputs)
image_features_list.append(image_features)
# 堆叠所有图像特征
all_image_features = torch.cat(image_features_list, dim=0)
# 计算相似度
text_features = text_features / text_features.norm(dim=1, keepdim=True)
all_image_features = all_image_features / all_image_features.norm(dim=1, keepdim=True)
similarity = torch.matmul(text_features, all_image_features.t())
# 获取最相似的图像索引
top_indices = similarity.argsort(descending=True)[0]
return [image_database[idx] for idx in top_indices]
def generate_story_with_images(self, prompt, num_images=3, max_length=1000):
"""生成包含图像的故事"""
# 第一步:生成故事文本
# 这里假设我们使用Qwen-VL生成故事文本
story_prompt = f"Write a creative story based on this prompt: {prompt}"
inputs = self.qwen_vl_processor(text=story_prompt, return_tensors="pt").to(self.device)
story_ids = self.qwen_vl_model.generate(
**inputs,
max_length=max_length,
num_beams=5,
early_stopping=True
)
story_text = self.qwen_vl_processor.batch_decode(story_ids, skip_special_tokens=True)[0]
# 第二步:为故事生成配图
# 将故事分成几个部分
story_parts = story_text.split("\n\n")
# 选择一些关键段落生成图像
selected_parts = story_parts[:min(num_images, len(story_parts))]
images = []
for part in selected_parts:
# 生成图像(简化示例)
image = self.generate_image_from_text(part)
images.append(image)
return story_text, images
def enhance_story_with_visual_details(self, story_text, reference_image):
"""基于参考图像增强故事细节"""
# 首先生成图像描述
image_caption = self.generate_caption(reference_image)
# 使用图像描述和原始故事作为提示
enhancement_prompt = f"""
Original story: {story_text}
Reference image description: {image_caption}
Enhance the story with visual details inspired by the reference image.
"""
# 使用Qwen-VL生成增强的故事
inputs = self.qwen_vl_processor(
images=reference_image,
text=enhancement_prompt,
return_tensors="pt"
).to(self.device)
enhanced_story_ids = self.qwen_vl_model.generate(
**inputs,
max_length=len(story_text) * 2,
num_beams=5,
early_stopping=True
)
enhanced_story = self.qwen_vl_processor.batch_decode(enhanced_story_ids, skip_special_tokens=True)[0]
return enhanced_story
多模态故事讲述应用示例#
以下是一个使用多模态框架创建交互式故事讲述应用的示例:
class InteractiveStorytellingApp:
"""交互式故事讲述应用"""
def __init__(self, multimodal_framework):
self.framework = multimodal_framework
self.story_state = {
"title": "",
"text": "",
"images": [],
"characters": {},
"settings": {},
"plot_points": []
}
def start_new_story(self, title, initial_prompt, setting_image=None):
"""开始一个新故事"""
self.story_state["title"] = title
# 如果提供了设置图像,使用它来增强提示
if setting_image is not None:
setting_description = self.framework.generate_caption(setting_image)
enhanced_prompt = f"{initial_prompt}\nSetting: {setting_description}"
else:
enhanced_prompt = initial_prompt
# 生成初始故事
story_text, story_images = self.framework.generate_story_with_images(enhanced_prompt)
self.story_state["text"] = story_text
self.story_state["images"] = story_images
# 提取角色和设置
self._extract_story_elements()
return self.story_state
def _extract_story_elements(self):
"""从故事中提取角色和设置"""
# 使用LLaVA分析故事文本
analysis_prompt = f"""
Analyze the following story and identify:
1. Main characters
2. Settings
3. Key plot points
Story: {self.story_state["text"]}
"""
# 使用第一张图像作为视觉上下文
if self.story_state["images"]:
analysis = self.framework.answer_visual_question(
self.story_state["images"][0],
analysis_prompt,
model="llava"
)
else:
# 如果没有图像,只使用文本
inputs = self.framework.llava_processor(text=analysis_prompt, return_tensors="pt").to(self.framework.device)
analysis_ids = self.framework.llava_model.generate(
**inputs,
max_length=500,
num_beams=5
)
analysis = self.framework.llava_processor.batch_decode(analysis_ids, skip_special_tokens=True)[0]
# 解析分析结果(简化示例)
# 实际实现需要更复杂的解析逻辑
if "Characters:" in analysis:
characters_text = analysis.split("Characters:")[1].split("Settings:")[0]
characters = [c.strip() for c in characters_text.split("-") if c.strip()]
self.story_state["characters"] = {c: {} for c in characters}
if "Settings:" in analysis:
settings_text = analysis.split("Settings:")[1].split("Plot points:")[0]
settings = [s.strip() for s in settings_text.split("-") if s.strip()]
self.story_state["settings"] = {s: {} for s in settings}
if "Plot points:" in analysis:
plot_text = analysis.split("Plot points:")[1]
plot_points = [p.strip() for p in plot_text.split("-") if p.strip()]
self.story_state["plot_points"] = plot_points
def continue_story(self, user_input, user_image=None):
"""继续故事,基于用户输入和可选的图像"""
# 准备提示
continuation_prompt = f"""
Current story: {self.story_state["text"]}
User input: {user_input}
Continue the story based on the user's input.
"""
# 如果用户提供了图像,使用它来增强故事
if user_image is not None:
# 生成图像描述
image_caption = self.framework.generate_caption(user_image)
# 增强提示
continuation_prompt += f"\nImage description: {image_caption}"
# 使用Qwen-VL生成故事续写
inputs = self.framework.qwen_vl_processor(
images=user_image,
text=continuation_prompt,
return_tensors="pt"
).to(self.framework.device)
else:
# 只使用文本
inputs = self.framework.qwen_vl_processor(
text=continuation_prompt,
return_tensors="pt"
).to(self.framework.device)
# 生成续写
continuation_ids = self.framework.qwen_vl_model.generate(
**inputs,
max_length=len(self.story_state["text"]) + 500,
num_beams=5,
early_stopping=True
)
continuation = self.framework.qwen_vl_processor.batch_decode(continuation_ids, skip_special_tokens=True)[0]
# 提取新添加的部分
new_content = continuation[len(continuation_prompt):]
# 更新故事状态
self.story_state["text"] += "\n\n" + new_content
# 为新内容生成配图
new_image = self.framework.generate_image_from_text(new_content)
self.story_state["images"].append(new_image)
# 更新故事元素
self._extract_story_elements()
return self.story_state
def visualize_character(self, character_name):
"""为故事中的角色生成视觉形象"""
if character_name not in self.story_state["characters"]:
return None
# 从故事中提取角色描述
character_prompt = f"""
Based on the story: {self.story_state["text"]}
Generate a detailed visual description of the character: {character_name}
"""
# 使用LLaVA生成角色描述
inputs = self.framework.llava_processor(text=character_prompt, return_tensors="pt").to(self.framework.device)
description_ids = self.framework.llava_model.generate(
**inputs,
max_length=200,
num_beams=5
)
character_description = self.framework.llava_processor.batch_decode(description_ids, skip_special_tokens=True)[0]
# 生成角色图像
character_image = self.framework.generate_image_from_text(character_description)
# 更新角色信息
self.story_state["characters"][character_name]["description"] = character_description
self.story_state["characters"][character_name]["image"] = character_image
return character_image
def export_illustrated_story(self):
"""导出带插图的故事"""
# 将故事文本分成段落
paragraphs = self.story_state["text"].split("\n\n")
# 准备导出数据
illustrated_story = {
"title": self.story_state["title"],
"content": []
}
# 交替添加文本和图像
for i, paragraph in enumerate(paragraphs):
illustrated_story["content"].append({"type": "text", "data": paragraph})
# 如果有对应的图像,添加它
if i < len(self.story_state["images"]):
illustrated_story["content"].append({"type": "image", "data": self.story_state["images"][i]})
# 添加角色图库
character_gallery = []
for name, info in self.story_state["characters"].items():
if "image" in info:
character_gallery.append({
"name": name,
"description": info.get("description", ""),
"image": info["image"]
})
illustrated_story["character_gallery"] = character_gallery
return illustrated_story
多模态技术在故事讲述中的应用#
多模态技术可以极大地增强故事讲述体验,使故事更加生动、沉浸和个性化。
视觉增强故事#
自动插图生成:
为故事关键场景生成配图
根据文本描述可视化角色和场景
创建一致的视觉风格
交互式视觉探索:
允许用户探索故事世界的不同部分
提供场景的多角度视图
可视化故事中的地图和位置
角色可视化:
创建角色的视觉形象
展示角色随故事发展的变化
可视化角色关系和互动
多模态故事互动#
图像引导的故事分支:
用户可以通过提供图像来影响故事方向
系统分析图像内容并将其整合到故事中
创建基于视觉输入的个性化故事体验
视觉问答增强:
允许用户询问关于故事场景的问题
提供基于图像的额外上下文和细节
深化对故事世界的理解
情感响应生成:
基于用户表情或情绪调整故事语调
生成与用户情感状态共鸣的内容
创建更具情感连接的故事体验
多模态故事记忆和一致性#
视觉记忆库:
维护故事元素的视觉表示
确保角色和场景的视觉一致性
随着故事发展更新视觉表示
多模态上下文跟踪:
跟踪文本和视觉元素之间的关系
确保新生成的内容与之前的视觉和文本保持一致
管理长期故事的连贯性
风格一致性维护:
保持一致的视觉和文本风格
适应用户偏好的艺术风格
在整个故事中维持统一的美学
未来发展趋势#
多模态AI在故事讲述领域的未来发展趋势包括:
更深层次的模态融合:
超越简单的文本-图像对齐
理解模态之间的复杂关系和互动
创建真正统一的多模态表示
更多模态的整合:
添加音频(音乐、音效、语音)
整合视频和动画
探索触觉和其他感官模态
更强的上下文理解:
更长的上下文窗口,理解整个故事
更好的长期一致性和连贯性
更深入的叙事结构理解
更高效的训练方法:
改进的参数高效微调技术
更好的知识迁移和模型压缩
更高效的多模态预训练目标
更个性化的体验:
适应用户偏好和兴趣
学习用户的视觉和文本风格
创建真正个性化的故事体验
总结#
多模态AI技术为故事讲述带来了革命性的变化,使AI能够创造更丰富、更沉浸、更个性化的故事体验。通过整合VQVAE、VQGAN、扩散模型等技术,并使用LoRA等参数高效微调方法训练CLIP、BLIP2、LLaVA和Qwen-vl等多模态模型,我们可以构建强大的故事讲述AI系统。
这些系统能够理解和生成文本和图像,创建视觉增强的故事,支持多模态互动,并维护故事的视觉和文本一致性。随着技术的不断发展,多模态故事讲述AI将变得更加强大、自然和引人入胜,为用户提供前所未有的创意体验。
在构建故事讲述AI系统时,理解和应用这些多模态技术至关重要,它们不仅增强了系统的表达能力,还创造了新的交互和体验维度,使AI生成的故事更加生动和令人难忘。