## 主要更新 - ✨ 更新所有依赖到最新稳定版本 - 📝 添加详细的项目文档和模型推荐 - 🔧 配置 VSCode Cloud Studio 预览功能 - 🐛 修复 PyTorch API 弃用警告 ## 依赖更新 - diffusers: 0.27.2 → 0.35.2 - gradio: 4.21.0 → 5.46.0 - peft: 0.7.1 → 0.18.0 - Pillow: 9.5.0 → 11.3.0 - fastapi: 0.108.0 → 0.116.2 ## 新增文件 - CLAUDE.md - 项目架构和开发指南 - UPGRADE_NOTES.md - 详细的升级说明 - .vscode/preview.yml - 预览配置 - .vscode/LAUNCH_GUIDE.md - 启动指南 - .gitignore - 更新的忽略规则 ## 代码修复 - 修复 iopaint/model/ldm.py 中的 torch.cuda.amp.autocast() 弃用警告 ## 文档更新 - README.md - 添加模型推荐和使用指南 - 完整的项目源码(iopaint/) - Web 前端源码(web_app/) 🤖 Generated with Claude Code
41 lines
1.4 KiB
Python
41 lines
1.4 KiB
Python
import torch
|
|
from transformers import PreTrainedModel
|
|
|
|
from ..utils import torch_gc
|
|
|
|
|
|
class CPUTextEncoderWrapper(PreTrainedModel):
|
|
def __init__(self, text_encoder, torch_dtype):
|
|
super().__init__(text_encoder.config)
|
|
self.config = text_encoder.config
|
|
self._device = text_encoder.device
|
|
# cpu not support float16
|
|
self.text_encoder = text_encoder.to(torch.device("cpu"), non_blocking=True)
|
|
self.text_encoder = self.text_encoder.to(torch.float32, non_blocking=True)
|
|
self.torch_dtype = torch_dtype
|
|
del text_encoder
|
|
torch_gc()
|
|
|
|
def __call__(self, x, **kwargs):
|
|
input_device = x.device
|
|
original_output = self.text_encoder(x.to(self.text_encoder.device), **kwargs)
|
|
for k, v in original_output.items():
|
|
if isinstance(v, tuple):
|
|
original_output[k] = [
|
|
v[i].to(input_device).to(self.torch_dtype) for i in range(len(v))
|
|
]
|
|
else:
|
|
original_output[k] = v.to(input_device).to(self.torch_dtype)
|
|
return original_output
|
|
|
|
@property
|
|
def dtype(self):
|
|
return self.torch_dtype
|
|
|
|
@property
|
|
def device(self) -> torch.device:
|
|
"""
|
|
`torch.device`: The device on which the module is (assuming that all the module parameters are on the same
|
|
device).
|
|
"""
|
|
return self._device |