深度学习训练八股

发布于:2024-05-08 ⋅ 阅读:(29) ⋅ 点赞:(0)

一、模型中的函数的定义

1.torchmetrics.AUROC

(1).binary
>>> from torch import tensor
>>> preds = tensor([0.13, 0.26, 0.08, 0.19, 0.34])
>>> target = tensor([0, 0, 1, 1, 1])
>>> auroc = AUROC(task="binary")
>>> auroc(preds, target)
tensor(0.5000)
(2).multiclass
>>> preds = tensor([[0.90, 0.05, 0.05],
...                       [0.05, 0.90, 0.05],
...                       [0.05, 0.05, 0.90],
...                       [0.85, 0.05, 0.10],
...                       [0.10, 0.10, 0.80]])
>>> target = tensor([0, 1, 1, 2, 2])
>>> auroc = AUROC(task="multiclass", num_classes=3)
>>> auroc(preds, target)
tensor(0.7778)

注意函数中average参数的默认值为“macro”。

二、test_k_fold_test_copy.py—.logs_k_fold/result_draw

# Test script
def test(model, test_loader, writer, device,criterion,roc_path,fold):
    model.eval()
    accuracy = Accuracy(task='multiclass', num_classes=2).to(device)
    precision = Precision(task='multiclass', average='macro', num_classes=2).to(device)
    recall = Recall(task='multiclass', average='macro', num_classes=2).to(device)
    auroc = AUROC(task='multiclass',num_classes=2).to(device)
    f1 = F1Score(num_classes=2, task='multiclass', average='macro').to(device)
    specificity=Specificity(num_classes=2, task='multiclass', average='macro').to(device)

    pred_scores = [] 
    true_labels = []
    pred_labels = []
    

    fold_results={}
    with torch.no_grad():
        for images, coords, labels, _, _  in test_loader:
            images = images.to(device)
            labels = labels.to(device) 
            outputs = model(images,coords)
            _, predicted = torch.max(outputs.data, 1)
                    
            accuracy(predicted, labels.data)
            precision(predicted, labels.data)
            recall(predicted, labels.data)
            f1(predicted, labels.data)
            #auroc(predicted, labels.data)
            specificity(predicted, labels.data)
            auroc(outputs, labels.data)
            pred_labels.extend(predicted.cpu().numpy())
            pred_scores.extend(outputs.cpu().numpy()) 
            true_labels.extend(labels.cpu().numpy())

    acc = accuracy.compute().item() 
    prec = precision.compute().item() 
    rec = recall.compute().item() 
    f1_score = f1.compute().item()
    auroc_score = auroc.compute().item()
    spec=specificity.compute().item()

    fold_results['fold']=fold
    fold_results['accuracy'] = acc
    fold_results['precision'] = prec
    fold_results['recall'] = rec
    fold_results['f1_score'] = f1_score
    fold_results['auroc_score'] = auroc_score
    fold_results['specificity'] = spec
    
    logging.info(f"Test Accuracy: {acc:.4f}, Test precision: {prec:.4f}, Test recall: {rec:.4f}, Test f1: {f1_score:.4f}, Test auroc: {auroc_score:4f},Test specificity:{spec:.4f}")
    logging.error("This is a fatal log!")   
    
    roc = MulticlassROC(num_classes=2, thresholds=None)
    pred_scores = torch.Tensor(pred_scores).to(device)
    true_labels = torch.Tensor(true_labels).int().to(device)
    fpr, tpr, thresholds = roc(pred_scores, true_labels)
    
    draw_fold_path = Path(os.path.join(fprs_tprs_path, f'fold_{fold}'))
    draw_fold_path.mkdir(parents=True, exist_ok=True)
    torch.save(tpr,os.path.join(draw_fold_path,"tpr.pt"))
    torch.save(fpr,os.path.join(draw_fold_path,"fpr.pt"))
   
    return fold_results, fpr, tpr