计算两个文件夹对应图像的SSIM和 PSNR

       在计算两个文件夹对应图像的SSIM和PSNR时,在网上找了很多资料,大多数都是计算两幅图片之间的值,而批量计算两个文件夹中图像的metrics的资料很少,要么就是跑出来有点问题。经过修改,该文章的代码是可以运行的。

代码如下:

 __init__.py

file:__init__.py

 evaluate.py

#  file: evaluate.py
import argparse
import json
import logging
import os

import sys
sys.path.append("/media/grapemaid/本地磁盘/grapemaidaaaaa/SSMI&PSNR")
import cv2
import numpy as np

try:
    import rasterio
except ImportError:
    rasterio = None

from image_similarity_measures.quality_metrics import metric_functions

logger = logging.getLogger(__name__)


def read_image(path):
    logger.info("Reading image %s", os.path.basename(path))
    if rasterio and (path.endswith(".tif") or path.endswith(".tiff")):
        return np.rollaxis(rasterio.open(path).read(), 0, 3)
    return cv2.imread(path)


def evaluation(org_img_path, pred_img_path, metrics):
    output_dict = {}
    org_img = read_image(org_img_path)
    pred_img = read_image(pred_img_path)

    for metric in metrics:
        metric_func = metric_functions[metric]
        out_value = float(metric_func(org_img, pred_img))
        logger.info(f"{metric.upper()} value is: {out_value}")
        output_dict[metric] = out_value
    return output_dict

def rename2PicsToSameNames(path1, path2):
    pics1 = os.listdir(path1)
    pics2 = os.listdir(path2)

    sum = 0

    for pic in pics1:
        print(pic)
        try:
            flag = pics2.index(pic.replace('_real_B', '_fake_B'))
        except IOError:
            print('There is not any image with the same name as ' + str(pic) + ' in ' + str(path2))
        p1_name = "{:0>6d}".format(sum) + pic[-4:]
        p2_name = p1_name
        src_p1 = os.path.join(path1, pic)
        dst_p1 = os.path.join(path1, p1_name)
        os.rename(src_p1, dst_p1)
        print(str(src_p1) + ' --> ' + str(dst_p1) + '(renamed)')

        src_p2 = os.path.join(path2, pic.replace('_real_B', '_fake_B'))
        dst_p2 = os.path.join(path2, p2_name)
        os.rename(src_p2, dst_p2)
        print(str(src_p2) + ' --> ' + str(dst_p2) + '(renamed)')
        sum += 1

def main():
    logging.basicConfig(
        format="%(asctime)s - %(name)s - %(levelname)s - %(message)s",
        level=logging.INFO,
    )
    all_metrics = sorted(metric_functions.keys())
    parser = argparse.ArgumentParser(description="Evaluates an Image Super Resolution Model")
    parser.add_argument("--org_img_path", help="Path to original input image", required=True, metavar="FOLDER")
    parser.add_argument("--pred_img_path", help="Path to predicted image", required=True, metavar="FOLDER")
    parser.add_argument("--metric", dest="metrics", action="append",
                        choices=all_metrics + ['all'], metavar="METRIC",
                        help="select an evaluation metric (%(choices)s) (can be repeated)")
    args = parser.parse_args()
    if not args.metrics:
        args.metrics = ["psnr"]
    if "all" in args.metrics:
        args.metrics = all_metrics

    path1 = args.org_img_path
    path2 = args.pred_img_path

    # rename2PicsToSameNames(path1, path2)

    pics1 = os.listdir(path1)

    tot_ssim = 0.0
    tot_psnr = 0.0
    tot = 0

    parser.set_defaults(metrics=['ssim'])
    args = parser.parse_args()

    for pic in pics1:
        p1 = os.path.join(path1, pic)
        p2 = os.path.join(path2, pic)
        result_dict = evaluation(p1, p2, args.metrics)
        #print(result_dict['ssim'])

        #print(result_dict['psnr'])

        tot_ssim += result_dict['ssim']

        tot_psnr += result_dict['psnr']
        tot += 1





    # parser.set_defaults(metrics=['psnr'])
    # tot_psnr = 0.0
    # tot = 0
    #
    # for pic in pics1:
    #     p1 = os.path.join(path1, pic)
    #     p2 = os.path.join(path2, pic)
    #     result_dict = evaluation(p1, p2, args.metrics)
    #
    #     print(result_dict['psnr'])
    #
    #     tot_psnr += result_dict['psnr']
    #     tot += 1


    print('***************')
    print('Average SSIM: ' + str(tot_ssim / tot )) # origin: tot_ssim / tot
    print('Average PSNR: ' + str(tot_psnr / tot )) # origin: tot_psnr / tot


if __name__ == "__main__":
    main()

 quality_metrics.py

# file: quality_metrics.py
"""
This module is a collection of metrics to assess the similarity between two images.
PSNR, SSIM, FSIM and ISSM are the current metrics that are implemented in this module.
"""
import math

import numpy as np
from skimage.metrics import structural_similarity
import phasepack.phasecong as pc
import cv2


def _assert_image_shapes_equal(org_img: np.ndarray, pred_img: np.ndarray, metric: str):
    msg = (f"Cannot calculate {metric}. Input shapes not identical. y_true shape ="
           f"{str(org_img.shape)}, y_pred shape = {str(pred_img.shape)}")

    assert org_img.shape == pred_img.shape, msg

def rmse(org_img: np.ndarray, pred_img: np.ndarray, max_p=4095) -> float:
    """
    Root Mean Squared Error

    Calculated individually for all bands, then averaged
    """
    _assert_image_shapes_equal(org_img, pred_img, "RMSE")

    org_img = org_img.astype(np.float32)

    rmse_bands = []
    for i in range(org_img.shape[2]):
        dif = np.subtract(org_img, pred_img)
        m = np.mean(np.square( dif / max_p))
        s = np.sqrt(m)
        rmse_bands.append(s)

    return np.mean(rmse_bands)


def psnr(org_img: np.ndarray, pred_img: np.ndarray, max_p=4095) -> float:
    """
    Peek Signal to Noise Ratio, implemented as mean squared error converted to dB.

    It can be calculated as
    PSNR = 20 * log10(MAXp) - 10 * log10(MSE)

    When using 12-bit imagery MaxP is 4095, for 8-bit imagery 255. For floating point imagery using values between
    0 and 1 (e.g. unscaled reflectance) the first logarithmic term can be dropped as it becomes 0
    """
    _assert_image_shapes_equal(org_img, pred_img, "PSNR")

    org_img = org_img.astype(np.float32)

    mse_bands = []
    for i in range(org_img.shape[2]):
        mse_bands.append(np.mean(np.square(org_img[:, :, i] - pred_img[:, :, i])))

    return 20 * np.log10(max_p) - 10. * np.log10(np.mean(mse_bands))


def _similarity_measure(x, y, constant):
    """
    Calculate feature similarity measurement between two images
    """
    numerator = 2 * x * y + constant
    denominator = x ** 2 + y ** 2 + constant

    return numerator / denominator


def _gradient_magnitude(img: np.ndarray, img_depth):
    """
    Calculate gradient magnitude based on Scharr operator
    """
    scharrx = cv2.Scharr(img, img_depth, 1, 0)
    scharry = cv2.Scharr(img, img_depth, 0, 1)

    return np.sqrt(scharrx ** 2 + scharry ** 2)


def fsim(org_img: np.ndarray, pred_img: np.ndarray, T1=0.85, T2=160) -> float:
    """
    Feature-based similarity index, based on phase congruency (PC) and image gradient magnitude (GM)

    There are different ways to implement PC, the authors of the original FSIM paper use the method
    defined by Kovesi (1999). The Python phasepack project fortunately provides an implementation
    of the approach.

    There are also alternatives to implement GM, the FSIM authors suggest to use the Scharr
    operation which is implemented in OpenCV.

    Note that FSIM is defined in the original papers for grayscale as well as for RGB images. Our use cases
    are mostly multi-band images e.g. RGB + NIR. To accommodate for this fact, we compute FSIM for each individual
    band and then take the average.

    Note also that T1 and T2 are constants depending on the dynamic range of PC/GM values. In theory this parameters
    would benefit from fine-tuning based on the used data, we use the values found in the original paper as defaults.

    Args:
        org_img -- numpy array containing the original image
        pred_img -- predicted image
        T1 -- constant based on the dynamic range of PC values
        T2 -- constant based on the dynamic range of GM values
    """
    _assert_image_shapes_equal(org_img, pred_img, "FSIM")

    alpha = beta = 1  # parameters used to adjust the relative importance of PC and GM features
    fsim_list = []
    for i in range(org_img.shape[2]):
        # Calculate the PC for original and predicted images
        pc1_2dim = pc(org_img[:, :, i], nscale=4, minWaveLength=6, mult=2, sigmaOnf=0.5978)
        pc2_2dim = pc(pred_img[:, :, i], nscale=4, minWaveLength=6, mult=2, sigmaOnf=0.5978)

        # pc1_2dim and pc2_2dim are tuples with the length 7, we only need the 4th element which is the PC.
        # The PC itself is a list with the size of 6 (number of orientation). Therefore, we need to
        # calculate the sum of all these 6 arrays.
        pc1_2dim_sum = np.zeros((org_img.shape[0], org_img.shape[1]), dtype=np.float64)
        pc2_2dim_sum = np.zeros((pred_img.shape[0], pred_img.shape[1]), dtype=np.float64)
        for orientation in range(6):
            pc1_2dim_sum += pc1_2dim[4][orientation]
            pc2_2dim_sum += pc2_2dim[4][orientation]

        # Calculate GM for original and predicted images based on Scharr operator
        gm1 = _gradient_magnitude(org_img[:, :, i], cv2.CV_16U)
        gm2 = _gradient_magnitude(pred_img[:, :, i], cv2.CV_16U)

        # Calculate similarity measure for PC1 and PC2
        S_pc = _similarity_measure(pc1_2dim_sum, pc2_2dim_sum, T1)
        # Calculate similarity measure for GM1 and GM2
        S_g = _similarity_measure(gm1, gm2, T2)

        S_l = (S_pc ** alpha) * (S_g ** beta)

        numerator = np.sum(S_l * np.maximum(pc1_2dim_sum, pc2_2dim_sum))
        denominator = np.sum(np.maximum(pc1_2dim_sum, pc2_2dim_sum))
        fsim_list.append(numerator / denominator)

    return np.mean(fsim_list)


def _ehs(x, y):
    """
    Entropy-Histogram Similarity measure
    """
    H = (np.histogram2d(x.flatten(), y.flatten()))[0]

    return -np.sum(np.nan_to_num(H * np.log2(H)))


def _edge_c(x, y):
    """
    Edge correlation coefficient based on Canny detector
    """
    # Use 100 and 200 as thresholds, no indication in the paper what was used
    g = cv2.Canny((x * 0.0625).astype(np.uint8), 100, 200)
    h = cv2.Canny((y * 0.0625).astype(np.uint8), 100, 200)

    g0 = np.mean(g)
    h0 = np.mean(h)

    numerator = np.sum((g - g0) * (h - h0))
    denominator = np.sqrt(np.sum(np.square(g-g0)) * np.sum(np.square(h-h0)))

    return numerator / denominator


def issm(org_img: np.ndarray, pred_img: np.ndarray) -> float:
    """
    Information theoretic-based Statistic Similarity Measure

    Note that the term e which is added to both the numerator as well as the denominator is not properly
    introduced in the paper. We assume the authers refer to the Euler number.
    """
    _assert_image_shapes_equal(org_img, pred_img, "ISSM")

    # Variable names closely follow original paper for better readability
    x = org_img
    y = pred_img
    A = 0.3
    B = 0.5
    C = 0.7

    ehs_val = _ehs(x, y)
    canny_val = _edge_c(x, y)

    numerator = canny_val * ehs_val * (A + B) + math.e
    denominator = A * canny_val * ehs_val + B * ehs_val + C * ssim(x, y) + math.e

    return np.nan_to_num(numerator / denominator)


def ssim(org_img: np.ndarray, pred_img: np.ndarray, max_p=4095) -> float:
    """
    Structural SIMularity index
    """
    _assert_image_shapes_equal(org_img, pred_img, "SSIM")

    return structural_similarity(org_img, pred_img, data_range=max_p, channel_axis=2)


def sliding_window(image, stepSize, windowSize):
    # slide a window across the image
    for y in range(0, image.shape[0], stepSize):
        for x in range(0, image.shape[1], stepSize):
            # yield the current window
            yield (x, y, image[y:y + windowSize[1], x:x + windowSize[0]])


def uiq(org_img: np.ndarray, pred_img: np.ndarray, step_size=1, window_size=8):
    """
    Universal Image Quality index
    """
    # TODO: Apply optimization, right now it is very slow
    _assert_image_shapes_equal(org_img, pred_img, "UIQ")

    org_img = org_img.astype(np.float32)
    pred_img = pred_img.astype(np.float32)

    q_all = []
    for (x, y, window_org), (x, y, window_pred) in zip(sliding_window(org_img, stepSize=step_size,
                                                                      windowSize=(window_size, window_size)),
                                                       sliding_window(pred_img, stepSize=step_size,
                                                                      windowSize=(window_size, window_size))):
        # if the window does not meet our desired window size, ignore it
        if window_org.shape[0] != window_size or window_org.shape[1] != window_size:
            continue

        for i in range(org_img.shape[2]):
            org_band = window_org[:, :, i]
            pred_band = window_pred[:, :, i]
            org_band_mean = np.mean(org_band)
            pred_band_mean = np.mean(pred_band)
            org_band_variance = np.var(org_band)
            pred_band_variance = np.var(pred_band)
            org_pred_band_variance = np.mean((org_band - org_band_mean) * (pred_band - pred_band_mean))

            numerator = 4 * org_pred_band_variance * org_band_mean * pred_band_mean
            denominator = (org_band_variance + pred_band_variance) * (org_band_mean**2 + pred_band_mean**2)

            if denominator != 0.0:
                q = numerator / denominator
                q_all.append(q)

    if not np.any(q_all):
        raise ValueError(f"Window size ({window_size}) is too big for image with shape "
                         f"{org_img.shape[0:2]}, please use a smaller window size.")

    return np.mean(q_all)


def sam(org_img: np.ndarray, pred_img: np.ndarray, convert_to_degree=True):
    """
    Spectral Angle Mapper which defines the spectral similarity between two spectra
    """

    _assert_image_shapes_equal(org_img, pred_img, "SAM")

    # Spectral angles are first computed for each pair of pixels
    numerator = np.sum(np.multiply(pred_img, org_img), axis=2)
    denominator = np.linalg.norm(org_img, axis=2) * np.linalg.norm(pred_img, axis=2)
    val = np.clip(numerator / denominator, -1, 1)
    sam_angles = np.arccos(val)
    if convert_to_degree:
        sam_angles = sam_angles * 180.0 / np.pi

    # The original paper states that SAM values are expressed as radians, while e.g. Lanares
    # et al. (2018) use degrees. We therefore made this configurable, with degree the default
    return np.mean(np.nan_to_num(sam_angles))


def sre(org_img: np.ndarray, pred_img: np.ndarray):
    """
    signal to reconstruction error ratio
    """
    _assert_image_shapes_equal(org_img, pred_img, "SRE")

    org_img = org_img.astype(np.float32)

    sre_final = []
    for i in range(org_img.shape[2]):
        numerator = np.square(np.mean(org_img[:, :, i]))
        denominator = (np.linalg.norm(org_img[:, :, i] - pred_img[:, :, i])) /\
                      (org_img.shape[0] * org_img.shape[1])
        sre_final.append(numerator/denominator)

    return 10 * np.log10(np.mean(sre_final))

metric_functions = {
    "fsim": fsim,
    "issm": issm,
    "psnr": psnr,
    "rmse": rmse,
    "sam": sam,
    "sre": sre,
    "ssim": ssim,
    "uiq": uiq,
}

 终端运行:

 python evaluate.py --org_img_path "/media/grapemaid/本地磁盘/grapemaidaaaaa/SSMI&PSNR/real/" --pred_img_path "/media/grapemaid/本地磁盘/grapemaidaaaaa/SSMI&PSNR/generated/" --metric psnr
2024-03-15 14:40:25,213 - __main__ - INFO - Reading image 1.png
2024-03-15 14:40:25,219 - __main__ - INFO - Reading image 1.png
2024-03-15 14:40:25,298 - __main__ - INFO - SSIM value is: 0.8564206996036278
2024-03-15 14:40:25,301 - __main__ - INFO - PSNR value is: 38.358364140300324
2024-03-15 14:40:25,301 - __main__ - INFO - Reading image 2.png
2024-03-15 14:40:25,307 - __main__ - INFO - Reading image 2.png
2024-03-15 14:40:25,384 - __main__ - INFO - SSIM value is: 0.718175835507345
2024-03-15 14:40:25,387 - __main__ - INFO - PSNR value is: 32.463593518046906
2024-03-15 14:40:25,387 - __main__ - INFO - Reading image 3.png
2024-03-15 14:40:25,392 - __main__ - INFO - Reading image 3.png
2024-03-15 14:40:25,471 - __main__ - INFO - SSIM value is: 0.8717178098908634
2024-03-15 14:40:25,473 - __main__ - INFO - PSNR value is: 35.22717240980106
2024-03-15 14:40:25,474 - __main__ - INFO - Reading image 4.png
2024-03-15 14:40:25,478 - __main__ - INFO - Reading image 4.png
2024-03-15 14:40:25,555 - __main__ - INFO - SSIM value is: 0.9065380455568989
2024-03-15 14:40:25,558 - __main__ - INFO - PSNR value is: 36.84613469770389
2024-03-15 14:40:25,558 - __main__ - INFO - Reading image 5.png
2024-03-15 14:40:25,565 - __main__ - INFO - Reading image 5.png
2024-03-15 14:40:25,645 - __main__ - INFO - SSIM value is: 0.7740617226963673
2024-03-15 14:40:25,648 - __main__ - INFO - PSNR value is: 32.94374231031375
2024-03-15 14:40:25,648 - __main__ - INFO - Reading image 6.png
2024-03-15 14:40:25,653 - __main__ - INFO - Reading image 6.png
2024-03-15 14:40:25,732 - __main__ - INFO - SSIM value is: 0.7547088723759559
2024-03-15 14:40:25,735 - __main__ - INFO - PSNR value is: 31.90708163907962
***************
Average SSIM: 0.8136038309385096
Average PSNR: 34.62434811920759

PS:1.  evaluate.py中的 "sys.path.append"换成自己项目的路径就行。

sys.path.append(" ")

2. 两文件夹中对应图像文件名和后缀名都要相同

参考资料:

参考博客

原始代码

  • 10
    点赞
  • 8
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值