RobustVideoMatting 预测图片
改为了推理图片,文件夹的图片尺寸必须一样,否则会报错
针对复杂场景,效果也不好,比如被另一个人遮挡,前面还挂了围脖,背了包包,抱着小孩
。
"""
python inference.py \--variant mobilenetv3 \--checkpoint "CHECKPOINT" \--device cuda \--input-source "input.mp4" \--output-type video \--output-composition "composition.mp4" \--output-alpha "alpha.mp4" \--output-foreground "foreground.mp4" \--output-video-mbps 4 \--seq-chunk 1
"""import torch
import os
from torch.utils.data import DataLoader
from torchvision import transforms
from typing import Optional, Tuple
from tqdm.auto import tqdmfrom inference_utils import VideoReader, VideoWriter, ImageSequenceReader, ImageSequenceWriterdef convert_video(model,input_source: str,input_resize: Optional[Tuple[int, int]] = None,downsample_ratio: Optional[float] = None,output_type: str = 'video',output_composition: Optional[str] = None,output_alpha: Optional[str] = None,output_foreground: Optional[str] = None,output_video_mbps: Optional[float] = None,seq_chunk: int = 1,num_workers: int = 0,progress: bool = True,device: Optional[str] = None,dtype: Optional[torch.dtype] = None):assert downsample_ratio is None or (downsample_ratio > 0 and downsample_ratio <= 1), 'Downsample ratio must be between 0 (exclusive) and 1 (inclusive).'assert any([output_composition, output_alpha, output_foreground]), 'Must provide at least one output.'assert output_type in ['video', 'png_sequence'], 'Only support "video" and "png_sequence" output modes.'assert seq_chunk >= 1, 'Sequence chunk must be >= 1'assert num_workers >= 0, 'Number of workers must be >= 0'# Initialize transformif input_resize is not None:transform = transforms.Compose([transforms.Resize(input_resize[::-1]),transforms.ToTensor()])else:transform = transforms.ToTensor()# Initialize readerif os.path.isfile(input_source):source = VideoReader(input_source, transform)else:source = ImageSequenceReader(input_source, transform)reader = DataLoader(source, batch_size=seq_chunk, pin_memory=True, num_workers=num_workers)# Initialize writersif output_type == 'video':frame_rate = source.frame_rate if isinstance(source, VideoReader) else 30output_video_mbps = 1 if output_video_mbps is None else output_video_mbpsif output_composition is not None:writer_com = VideoWriter(path=output_composition,frame_rate=frame_rate,bit_rate=int(output_video_mbps * 1000000))if output_alpha is not None:writer_pha = VideoWriter(path=output_alpha,frame_rate=frame_rate,bit_rate=int(output_video_mbps * 1000000))if output_foreground is not None:writer_fgr = VideoWriter(path=output_foreground,frame_rate=frame_rate,bit_rate=int(output_video_mbps * 1000000))else:if output_composition is not None:writer_com = ImageSequenceWriter(output_composition, 'png')if output_alpha is not None:writer_pha = ImageSequenceWriter(output_alpha, 'png')if output_foreground is not None:writer_fgr = ImageSequenceWriter(output_foreground, 'png')# Inferencemodel = model.eval()if device is None or dtype is None:param = next(model.parameters())dtype = param.dtypedevice = param.deviceif (output_composition is not None) and (output_type == 'video'):bgr = torch.tensor([120, 255, 155], device=device, dtype=dtype).div(255).view(1, 1, 3, 1, 1)try:with torch.no_grad():bar = tqdm(total=len(source), disable=not progress, dynamic_ncols=True)rec = [None] * 4for src in reader:if downsample_ratio is None:downsample_ratio = auto_downsample_ratio(*src.shape[2:])src = src.to(device, dtype, non_blocking=True).unsqueeze(0) # [B, T, C, H, W]fgr, pha, *rec = model(src, *rec, downsample_ratio)if output_foreground is not None:writer_fgr.write(fgr[0])if output_alpha is not None:writer_pha.write(pha[0])if output_composition is not None:if output_type == 'video':com = fgr * pha + bgr * (1 - pha)else:fgr = fgr * pha.gt(0)com = torch.cat([fgr, pha], dim=-3)writer_com.write(com[0])bar.update(src.size(1))finally:# Clean upif output_composition is not None:writer_com.close()if output_alpha is not None:writer_pha.close()if output_foreground is not None:writer_fgr.close()def auto_downsample_ratio(h, w):"""Automatically find a downsample ratio so that the largest side of the resolution be 512px."""return min(512 / max(h, w), 1)class Converter:def __init__(self, variant: str, checkpoint: str, device: str):self.model = MattingNetwork(variant).eval().to(device)self.model.load_state_dict(torch.load(checkpoint, map_location=device))self.model = torch.jit.script(self.model)self.model = torch.jit.freeze(self.model)self.device = devicedef convert(self, *args, **kwargs):convert_video(self.model, device=self.device, dtype=torch.float32, *args, **kwargs)if __name__ == '__main__':import argparsefrom model import MattingNetwork"""python inference.py \--variant mobilenetv3 \--checkpoint "CHECKPOINT" \--device cuda \--input-source "input.mp4" \--output-type video \--output-composition "composition.mp4" \--output-alpha "alpha.mp4" \--output-foreground "foreground.mp4" \--output-video-mbps 4 \--seq-chunk 1"""parser = argparse.ArgumentParser()parser.add_argument('--variant', type=str, default='resnet50', choices=['mobilenetv3', 'resnet50'])parser.add_argument('--checkpoint', type=str, default=r'D:\project\fenge\jacke121-rvm_128_json\model_a\rvm_resnet50.pth')parser.add_argument('--device', type=str,default='cuda')parser.add_argument('--input-source', type=str, default=r'C:\Users\Administrator\Documents\WeChat Files\libanggeng\FileStorage\File\2023-11\koutu\weilanliandai\aa')parser.add_argument('--input-resize', type=int, default=None, nargs=2)parser.add_argument('--downsample-ratio', type=float)parser.add_argument('--output-composition', type=str,default='output-composition')parser.add_argument('--output-alpha', type=str,default='output-alpha')parser.add_argument('--output-foreground', type=str,default='output-foreground')parser.add_argument('--output-type', type=str, default='png_sequence', choices=['video', 'png_sequence'])parser.add_argument('--output-video-mbps', type=int, default=1)parser.add_argument('--seq-chunk', type=int, default=1)parser.add_argument('--num-workers', type=int, default=0)parser.add_argument('--disable-progress', action='store_true')args = parser.parse_args()converter = Converter(args.variant, args.checkpoint, args.device)converter.convert(input_source=args.input_source,input_resize=args.input_resize,downsample_ratio=args.downsample_ratio,output_type=args.output_type,output_composition=args.output_composition,output_alpha=args.output_alpha,output_foreground=args.output_foreground,output_video_mbps=args.output_video_mbps,seq_chunk=args.seq_chunk,num_workers=args.num_workers,progress=not args.disable_progress)