Files
ieemoo-ai-isempty/testsingle.py
2022-10-26 02:06:34 +00:00

78 lines
2.6 KiB
Python
Executable File

# coding=utf-8
import os
import torch
import numpy as np
from PIL import Image
from torchvision import transforms
import argparse
from models.modeling import VisionTransformer, CONFIGS
import time
#模型测试单张图片
parser = argparse.ArgumentParser()
parser.add_argument("--dataset", choices=["emptyJudge2"], default="emptyJudge2", help="Which dataset.")
parser.add_argument("--img_size", default=600, type=int, help="Resolution size")
parser.add_argument('--split', type=str, default='overlap', help="Split method") # non-overlap
parser.add_argument('--slide_step', type=int, default=2, help="Slide step for overlap split")
parser.add_argument('--smoothing_value', type=float, default=0.0, help="Label smoothing value\n")
#parser.add_argument("--pretrained_model", type=str, default="../module/ieemoo-ai-isempty/model/now/emptyjudge5_checkpoint.bin", help="load pretrained model")
parser.add_argument("--pretrained_model", type=str, default="output/ieemooempty_vit_checkpoint.pth", help="load pretrained model") #使用自定义VIT
args = parser.parse_args()
args.device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
args.nprocs = torch.cuda.device_count()
# 准备模型
config = CONFIGS["ViT-B_16"]
config.split = args.split
config.slide_step = args.slide_step
num_classes = 5
cls_dict = {0: "noemp", 1: "yesemp"}
model = None
#model = VisionTransformer(config, args.img_size, zero_head=True, num_classes=num_classes, smoothing_value=args.smoothing_value)
if args.pretrained_model is not None:
model = torch.load(args.pretrained_model) #自己预训练模型
model.to(args.device)
model.eval()
test_transform = transforms.Compose([transforms.Resize((600, 600), Image.BILINEAR),
transforms.ToTensor(),
transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])])
#自定义Vit模型
# test_transform = transforms.Compose([transforms.Resize((320, 320), Image.BILINEAR),
# transforms.ToTensor(),
# transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])])
#img = Image.open("img.jpg")
img = Image.open("img.jpg")
x = test_transform(img)
startime = time.process_time()
part_logits = model(x.unsqueeze(0).to(args.device))
probs = torch.nn.Softmax(dim=-1)(part_logits)
top5 = torch.argsort(probs, dim=-1, descending=True)
print("Prediction Label\n")
for idx in top5[0, :5]:
print(f'{probs[0, idx.item()]:.5f} : {cls_dict[idx.item()]}', end='\n')
endtime = time.process_time()
print("Time cost:"+ str(endtime - startime)) #评估一张图片耗时2.8秒