Disclosure of Invention
The invention aims to solve the technical problem of providing a three-dimensional image quality objective evaluation method based on feature fusion, which can effectively improve the correlation between objective evaluation results and subjective perception.
The technical scheme adopted by the invention for solving the technical problems is as follows: a stereo image quality objective evaluation method based on feature fusion is characterized in that the processing process is as follows: firstly, obtaining a single eye diagram of an original undistorted stereo image according to even symmetric frequency response and odd symmetric frequency response of each pixel point in a left viewpoint image and a right viewpoint image of the original undistorted stereo image in different scales and directions and a distorted image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; obtaining a single eye diagram of the distorted stereo image to be evaluated according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the distorted stereo image to be evaluated in different scales and directions and a distorted image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; secondly, obtaining an objective evaluation metric value of each pixel point in the single eye diagram of the distorted three-dimensional image to be evaluated according to the mean value and the standard deviation of each pixel point in the two single eye diagrams; thirdly, obtaining a corresponding saliency map according to the amplitude and the phase of the single eye map of the original undistorted stereo image; obtaining a corresponding saliency map according to the amplitude and the phase of the single eye map of the distorted stereo image to be evaluated; then, according to the two saliency maps and the distortion map between the two single eye maps, fusing objective evaluation metric values of each pixel point in the single eye map of the distorted three-dimensional image to be evaluated to obtain an objective evaluation prediction value of the image quality of the distorted three-dimensional image to be evaluated; and finally, obtaining the image quality objective evaluation predicted value of the distorted three-dimensional images with different distortion types and different distortion degrees according to the processing process.
The invention relates to a method for objectively evaluating the quality of a stereo image based on feature fusion, which comprises the following specific steps:
making SorgFor original undistorted stereo image, let SdisFor the distorted stereo image to be evaluated, SorgIs noted as { Lorg(x, y) }, adding SorgIs noted as { Rorg(x, y) }, adding SdisIs noted as { Ldis(x, y) }, adding SdisIs noted as { Rdis(x, y) }, where (x, y) here denotes the coordinate positions of pixel points in the left viewpoint image and the right viewpoint image, x is 1. ltoreq. W, y is 1. ltoreq. H, W denotes the widths of the left viewpoint image and the right viewpoint image, H denotes the heights of the left viewpoint image and the right viewpoint image, Lorg(x, y) represents { L }orgThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rorg(x, y) represents { RorgThe pixel value L of the pixel point with the coordinate position (x, y) in (x, y) } isdis(x, y) represents { L }disThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rdis(x, y) represents { RdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
② according to { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are correspondingly obtained to obtain { L }org(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisThe amplitude of each pixel in (x, y) } is then based on { L }org(x, y) } and { R }orgAmplitude of each pixel in (x, y) } and { Lorg(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imageorgThe eye-independent diagram of (2), is marked as { CMorg(x, y) }, and according to { L }dis(x, y) } and { R }disAmplitude of each pixel in (x, y) } and { Lorg(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imagedisThe eye-independent diagram of (2), is marked as { CMdis(x, y) }, wherein CMorg(x, y) denotes { CMorgThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) }, CMdis(x, y) denotes { CMdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
③ according to { CMorg(x, y) } and { CMdisCalculating the mean value and standard deviation of each pixel point in (x, y) } and calculating the (CM)dis(x, y) } the objective evaluation metric for each pixel point, will be { CMdisThe objective evaluation metric value of the pixel point with the coordinate position (x, y) in (x, y) is recorded as Qimage(x,y);
Fourthly, according to { CMorg(x, y) } amplitude and phase, calculating { CMorg(x, y) } significant graph, denoted as { SMorg(x, y) }, according to { CMdis(x, y) } amplitude and phase, calculating { CMdis(x, y) } significant graph, denoted as { SMdis(x, y) }, wherein, SMorg(x, y) denotes { SMorgThe pixel value of the pixel point whose coordinate position is (x, y), SMdis(x, y) denotes { SMdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
fifthly, calculating { CMorg(x, y) } and { CMdisA distortion map between (x, y) } is recorded as { DM (x, y) }, and the pixel value of a pixel point with a coordinate position (x, y) in { DM (x, y) } is recorded as DM (x, y), and DM (x, y) = (CM)org(x,y)-CMdis(x,y))2;
Root of Chinese ThalictrifoliateAccording to { SMorg(x, y) } and { SMdis(x, y) } and { DM (x, y) }, for { CMdisThe objective evaluation metric values of all pixel points in (x, y) are fused to obtain SdisThe image quality objective evaluation predicted value of (1) is marked as Q, where Ω represents a pixel domain range, SM (x, y) = max (SM)org(x,y),SMdis(x, y)), max () is a function taking the maximum value, and γ and β are weight coefficients;
adopting n original undistorted stereo images, establishing a distorted stereo image set of the undistorted stereo images under different distortion types and different distortion degrees, wherein the distorted stereo image set comprises a plurality of distorted stereo images, respectively obtaining an average subjective score difference value of each distorted stereo image in the distorted stereo image set by using a subjective quality evaluation method, and marking the average subjective score difference value as DMOS, DMOS =100-MOS, wherein MOS represents a subjective score mean value, DMOS belongs to [0,100], and n is more than or equal to 1;
calculating S according to the steps from the first step to the sixth stepdisThe image quality objective evaluation prediction value of each distorted stereo image in the distorted stereo image set is calculated respectively through the operation of the image quality objective evaluation prediction value.
The concrete process of the second step is as follows:
2-1, pair { Lorg(x, y) is filtered to obtain { L }orgEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are converted into { L }orgEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions are recorded as eα,θ(x, y) will { LorgThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is marked as oα,θ(x, y), wherein alpha represents the scale factor of the filter used for filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and theta is more than or equal to 1 and less than or equal to 4;
2 according to { LorgCalculating the even symmetric frequency response and the odd symmetric frequency response of each pixel point in (x, y) in different scales and directions, and calculating the { L }orgThe amplitude of each pixel in (x, y) } will be { LorgThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
② -3, obtaining { L ] according to the steps from ② -1 to ② -2
orgOperation of amplitude of each pixel in (x, y) } acquires { R } in the same manner
org(x,y)}、{L
dis(x, y) } and { R }
dis(x, y) } the amplitude of each pixel point in the (x, y) } will be { R
orgThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
Will { L
disThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
Will { R
disThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
② 4, calculating { L by adopting a block matching method
org(x, y) } and { R }
org(x, y) } parallax images, noted as
Wherein,
to represent
The middle coordinate position is the pixel value of the pixel point of (x, y);
② 5 according to { L
org(x, y) } and { R }
orgAmplitude of each pixel in (x, y) } and
calculating the pixel value of each pixel point in S
orgThe eye-independent diagram of (2), is marked as { CM
org(x, y) }, will { CM
orgThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) is marked as CM
org(x,y),
Wherein,
represents { R
org(x, y) } coordinate position of
The amplitude of the pixel points of (a) is,
represents { R
org(x, y) } coordinate position of
The pixel value of the pixel point of (1);
② 6 according to { L
dis(x, y) } and { R }
disAmplitude of each pixel in (x, y) } and
calculating the pixel value of each pixel point in S
disThe eye-independent diagram of (2), is marked as { CM
dis(x, y) }, will { CM
disThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) is marked as CM
dis(x,y),
Wherein,
represents { R
dis(x, y) } coordinate position of
The amplitude of the pixel points of (a) is,
represents { R
dis(x, y) } coordinate position of
The pixel value of the pixel point of (1).
In the step II-1, the pairs { Lorg(x, y) } the filter used for the filtering process is a log-Garbor filter.
The concrete process of the step III is as follows:
③ 1, calculate { CMorg(x, y) } and { CMdisThe mean and standard deviation of each pixel in (x, y) } will be { CMorgThe coordinate position in (x, y) } is (x)1,y1) The mean value and the standard difference of the pixel points are respectively recorded as muorg(x1,y1) And σorg(x1,y1) Will { CMdisThe coordinate position in (x, y) } is (x)1,y1) The mean value and the standard difference of the pixel points are respectively recorded as mudis(x1,y1) And σdis(x1,y1), Wherein x is more than or equal to 11≤W,1≤y1≤H,N(x1,y1) The coordinate position is shown as (x)1,y1) The pixel point of (a) is an 8 x 8 neighborhood window with the center, M represents N (x)1,y1) Number of inner pixels, CMorg(x1,y1) Representation { CMorgThe coordinate position in (x, y) } is (x)1,y1) Pixel value of the pixel point, CMdis(x1,y1) Representation { CMdisThe coordinate position in (x, y) } is (x)1,y1) The pixel value of the pixel point of (1);
③ 2 according to { CMorg(x, y) } and { CMdisCalculating the mean value and standard deviation of each pixel point in (x, y) } and calculating the (CM)dis(x, y) } the objective evaluation metric for each pixel point, will be { CMdisThe coordinate position in (x, y) } is (x)1,y1) The objective evaluation metric value of the pixel point is recorded as Qimage(x1,y1), Wherein C is a control parameter.
The specific process of the step IV is as follows:
tetra-1, pair { CMorg(x, y) performing discrete Fourier transform to obtain (CM)orgThe amplitude and phase of (x, y) } are denoted as { M, respectivelyorg(u, v) } and { Aorg(u, v) }, where u denotes the width of the amplitude or phase of the transform domain, v denotes the height of the amplitude or phase of the transform domain, 1. ltoreq. u.ltoreq.W, 1. ltoreq. v.ltoreq.H, Morg(u, v) represents { M }org(u, v) } the amplitude value of the pixel point with the coordinate position of (u, v), Aorg(u, v) represents { A }org(u, v) } the coordinate position is the phase value of the pixel point with (u, v);
fourthly-2, calculating { Morg(u, v) } amplitude of the high frequency component, denoted as { R }org(u, v) }, will { RorgThe amplitude value of the high-frequency component of the pixel point with the coordinate position (u, v) in (u, v) is recorded as Rorg(u,v),Rorg(u,v)=log(Morg(u,v))-hm(u,v)*log(Morg(u, v)), wherein log () is a logarithmic function based on e, e =2.718281828, "+" is the convolution operator symbol, hm(u, v) represents an m × m mean filtering;
fourthly-3, according to { Rorg(u, v) } and { Aorg(u, v) } inverse discrete Fourier transform, and the obtained inverse transform image is taken as { CMorg(x, y) } significant graph, denoted as { SMorg(x, y) }, wherein, SMorg(x, y) denotes { SMorgThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
fourthly-4, obtaining the { CM according to the steps from the fourth step-1 to the fourth step-3orgOperation of the saliency map of (x, y) } acquisition of { CM in the same mannerdis(x, y) } significant graph, denoted as { SMdis(x, y) }, wherein, SMdis(x, y) denotes { SMdisAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
Compared with the prior art, the invention has the advantages that:
1) the method of the invention respectively calculates the single eye diagram of the original undistorted stereo image and the single eye diagram of the distorted stereo image to be evaluated, and directly evaluates the single eye diagram of the distorted stereo image, thereby effectively simulating the binocular stereo fusion process and avoiding the process of linearly weighting the objective evaluation metric values of the left viewpoint image and the right viewpoint image.
2) The method can ensure that the evaluation result is more perceptually accordant with the human visual system by calculating the single eye diagram of the original undistorted stereo image, the salient diagram of the single eye diagram of the distorted stereo image to be evaluated and the distortion diagram between the two single eye diagrams and fusing the objective evaluation metric value of each pixel point in the single eye diagram of the distorted stereo image to be evaluated, thereby effectively improving the correlation between the objective evaluation result and the subjective perception.
Detailed Description
The invention is described in further detail below with reference to the accompanying examples.
The invention provides a method for objectively evaluating the quality of a stereo image based on feature fusion, the overall implementation block diagram of which is shown in figure 1, and the processing process is as follows: firstly, obtaining a single eye diagram of an original undistorted stereo image according to even symmetric frequency response and odd symmetric frequency response of each pixel point in a left viewpoint image and a right viewpoint image of the original undistorted stereo image in different scales and directions and a distorted image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; obtaining a single eye diagram of the distorted stereo image to be evaluated according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the distorted stereo image to be evaluated in different scales and directions and a distorted image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; secondly, obtaining an objective evaluation metric value of each pixel point in the single eye diagram of the distorted three-dimensional image to be evaluated according to the mean value and the standard deviation of each pixel point in the two single eye diagrams; thirdly, obtaining a corresponding saliency map according to the amplitude and the phase of the single eye map of the original undistorted stereo image; obtaining a corresponding saliency map according to the amplitude and the phase of the single eye map of the distorted stereo image to be evaluated; then, according to the two saliency maps and the distortion map between the two single eye maps, fusing objective evaluation metric values of each pixel point in the single eye map of the distorted three-dimensional image to be evaluated to obtain an objective evaluation prediction value of the image quality of the distorted three-dimensional image to be evaluated; and finally, obtaining the image quality objective evaluation predicted value of the distorted three-dimensional images with different distortion types and different distortion degrees according to the processing process.
The method specifically comprises the following steps:
making SorgFor original undistorted stereo image, let SdisFor the distorted stereo image to be evaluated, SorgIs noted as { Lorg(x, y) }, adding SorgIs noted as { Rorg(x, y) }, adding SdisIs noted as { Ldis(x, y) }, adding SdisIs noted as { Rdis(x, y) }, where (x, y) here denotes the coordinate positions of pixel points in the left viewpoint image and the right viewpoint image, x is 1. ltoreq. W, y is 1. ltoreq. H, W denotes the widths of the left viewpoint image and the right viewpoint image, H denotes the heights of the left viewpoint image and the right viewpoint image, Lorg(x, y) represents { L }orgThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rorg(x, y) represents { RorgThe pixel value L of the pixel point with the coordinate position (x, y) in (x, y) } isdis(x, y) represents { L }disThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rdis(x, y) represents { RdisAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
② according to { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are correspondingly obtained to obtain { L }org(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisThe amplitude of each pixel in (x, y) } is then based on { L }org(x, y) } and { R }orgAmplitude of each pixel in (x, y) } and { Lorg(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imageorgIs denoted as { CM (cyclopean map) }org(x, y) }, and according to { L }dis(x, y) } and { R }disAmplitude of each pixel in (x, y) } and { Lorg(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imagedisThe eye-independent diagram of (2), is marked as { CMdis(x, y) }, wherein CMorg(x, y) denotes { CMorgThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) }, CMdis(x, y) denotes { CMdisAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
In this embodiment, the specific process of step two is:
2-1, pair { Lorg(x, y) is filtered to obtain { L }orgEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are converted into { L }orgEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions are recorded as eα,θ(x, y) will { LorgThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is marked as oα,θ(x, y), wherein alpha represents the scale factor of the filter used for filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and theta is more than or equal to 1 and less than or equal to 4.
Here, for { Lorg(x, y) } the filter used for the filtering process is a log-Garbor filter.
2 according to { LorgCalculating the even symmetric frequency response and the odd symmetric frequency response of each pixel point in (x, y) in different scales and directions, and calculating the { L }orgThe amplitude of each pixel in (x, y) } will be { LorgThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
② -3, obtaining { L ] according to the steps from ② -1 to ② -2
orgOperation of amplitude of each pixel in (x, y) } acquires { R } in the same manner
org(x,y)}、{L
dis(x, y) } and { R }
dis(x, y) } the amplitude of each pixel point in the (x, y) } will be { R
orgThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
Will { L
disThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
Will { R
disThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
If obtain { L
disThe operation process of the amplitude of each pixel point in (x, y) } is as follows: 1) for { L
dis(x, y) is filtered to obtain { L }
disEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are converted into { L }
disEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions are recorded as e
α,θ' (x, y), will { L
disThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is marked as o
α,θ' (x, y), wherein alpha represents the scale factor of the filter used for filtering, 1 is less than or equal to alpha and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and 1 is less than or equal to theta and less than or equal to 4; 2) according to { L
disCalculating the even symmetric frequency response and the odd symmetric frequency response of each pixel point in (x, y) in different scales and directions, and calculating the { L }
disThe amplitude of each pixel in (x, y) } will be { L
disThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
② 4, calculating { L by adopting a block matching method
org(x, y) } and { R }
org(x, y) } parallax images, noted as
Wherein,
to represent
The middle coordinate position is the pixel value of the pixel point of (x, y).
② 5 according to { L
org(x, y) } and { R }
orgAmplitude of each pixel in (x, y) } and
calculating the pixel value of each pixel point in S
orgThe eye-independent diagram of (2), is marked as { CM
org(x, y) }, will { CM
orgThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) is marked as CM
org(x,y),
Wherein,
represents { R
org(x, y) } coordinate position of
The amplitude of the pixel points of (a) is,
represents { R
org(x, y) } coordinate position of
The pixel value of the pixel point of (1).
② 6 according to { L
dis(x, y) } and { R }
disAmplitude of each pixel in (x, y) } and
calculating the pixel value of each pixel point in S
disThe eye-independent diagram of (2), is marked as { CM
dis(x, y) }, will { CM
disThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) is marked as CM
dis(x,y),
Wherein,
represents { R
dis(x, y) } coordinate position of
The amplitude of the pixel points of (a) is,
represents { R
dis(x, y) } coordinate position of
The pixel value of the pixel point of (1).
③ according to { CMorg(x, y) } and { CMdisCalculating the mean value and standard deviation of each pixel point in (x, y) } and calculating the (CM)dis(x, y) } the objective evaluation metric for each pixel point, will be { CMdisThe objective evaluation metric value of the pixel point with the coordinate position (x, y) in (x, y) is recorded as Qimsge(x, y), will { CMdisThe objective evaluation metric values of all the pixel points in (x, y) are collectively expressed as { Q }image(x,y)}。
In this embodiment, the specific process of step three is:
③ 1, calculate { CMorg(x, y) } and { CMdisThe mean and standard deviation of each pixel in (x, y) } will be { CMorgThe coordinate position in (x, y) } is (x)1,y1) The mean value and the standard difference of the pixel points are respectively recorded as muorg(x1,y1) And σorg(x1,y1) Will { CMdisThe coordinate position in (x, y) } is (x)1,y1) The mean value and the standard difference of the pixel points are respectively recorded as mudis(x1,y1) And σdis(x1,y1), Wherein x is more than or equal to 11≤W,1≤y1≤H,N(x1,y1) The coordinate position is shown as (x)1,y1) The pixel point of (a) is an 8 x 8 neighborhood window with the center, M represents N (x)1,y1) Number of inner pixels, CMorg(x1,y1) Representation { CMorgThe coordinate position in (x, y) } is (x)1,y1) Pixel value of the pixel point, CMdis(x1,y1) Representation { CMdisThe coordinate position in (x, y) } is (x)1,y1) The pixel value of the pixel point of (1).
③ 2 according to { CMorg(x, y) } and { CMdisCalculating the mean value and standard deviation of each pixel point in (x, y) } and calculating the (CM)dis(x, y) } the objective evaluation metric for each pixel point, will be { CMdisThe coordinate position in (x, y) } is (x)1,y1) The objective evaluation metric value of the pixel point is recorded as Qimage(x1,y1), Where C is a control parameter, in this embodiment, C =0.01 is taken.
Fourthly, according to { CMorgSpectral redundancy characteristics of (x, y) }, i.e. according to { CMorg(x, y) } amplitude and phase, calculating { CMorg(x, y) } saliency map, denoted as { SM }org(x, y) }, according to { CMdisSpectral redundancy characteristics of (x, y) }, i.e. according to { CMdis(x, y) } amplitude and phase, calculating { CMdis(x, y) } significant graph, denoted as { SMdis(x, y) }, wherein, SMorg(x, y) denotes { SMorgThe pixel value of the pixel point whose coordinate position is (x, y), SMdis(x, y) denotes { SMdisAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
In this embodiment, the specific process of step iv is:
tetra-1, pair { CMorg(x, y) performing discrete Fourier transform to obtain (CM)orgThe amplitude and phase of (x, y) } are denoted as { M, respectivelyorg(u, v) } and { Aorg(u, v) }, where u denotes the width of the amplitude or phase of the transform domain, v denotes the height of the amplitude or phase of the transform domain, 1. ltoreq. u.ltoreq.W, 1. ltoreq. v.ltoreq.H, Morg(u, v) represents { M }org(u, v) } the amplitude value of the pixel point with the coordinate position of (u, v), Aorg(u, v) represents { A }orgAnd (u, v) the coordinate position in the (u, v) is the phase value of the pixel point of (u, v).
Fourthly-2, calculating { Morg(u, v) } amplitude of the high frequency component, denoted as { R }org(u, v) }, will { RorgThe amplitude value of the high-frequency component of the pixel point with the coordinate position (u, v) in (u, v) is recorded as Rorg(u,v),Rorg(u,v)=log(Morg(u,v))-hm(u,v)*log(Morg(u, v)), wherein log () is a logarithmic function based on e, e =2.718281828, "+" is the convolution operator symbol, hm(u, v) represents m × m mean filtering, and in the present embodiment, m =3 is taken.
Fourthly-3, according to { Rorg(u, v) } and { Aorg(u, v) } inverse discrete Fourier transform, and the obtained inverse transform image is taken as { CMorg(x, y) } significant graph, denoted as { SMorg(x, y) }, wherein, SMorg(x, y) denotes { SMorgAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
Fourthly-4, obtaining the { CM according to the steps from the fourth step-1 to the fourth step-3orgOperation of the saliency map of (x, y) } acquisition of { CM in the same mannerdis(x, y) } significant graph, denoted as { SMdis(x, y) }, wherein, SMdis(x, y) denotes { SMdisAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
Fifthly, calculating { CMorg(x, y) } and { CMdisA distortion map (distortion map) between (x, y) } is recorded as { DM (x, y) }, and a pixel value of a pixel point with a coordinate position (x, y) in { DM (x, y) } is recorded as DM (x, y), DM (x, y) = (CM)org(x,y)-CMdis(x,y))2。
According to { SMorg(x, y) } and { SMdis(x, y) } and { DM (x, y) }, for { CMdisThe objective evaluation metric values of all pixel points in (x, y) are fused to obtain SdisThe image quality objective evaluation predicted value of (1) is marked as Q, where Ω represents a pixel domain range, SM (x, y) = max (SM)org(x,y),SMdis(x, y)), max () is a function taking the maximum value, γ and β are weight coefficients, and in the present embodiment, γ =1.601 and β =0.501 are taken.
Adopting n original undistorted stereo images, establishing a distorted stereo image set of the undistorted stereo images under different distortion types and different distortion degrees, wherein the distorted stereo image set comprises a plurality of distorted stereo images, respectively obtaining an average subjective score difference value of each distorted stereo image in the distorted stereo image set by using a subjective quality evaluation method, and marking the average subjective score difference value as DMOS, DMOS =100-MOS, wherein MOS represents a subjective score mean value, DMOS belongs to [0,100], and n is more than or equal to 1.
In the present embodiment, a set of distorted stereoscopic images at different distortion degrees of different distortion types is established by using the stereoscopic images composed of fig. 2a and 2b, fig. 3a and 3b, fig. 4a and 4b, fig. 5a and 5b, fig. 6a and 6b, fig. 7a and 7b, fig. 8a and 8b, fig. 9a and 9b, fig. 10a and 10b, fig. 11a and 11b, fig. 12a and 12b, fig. 13a and 13b, and n =12, the set of distorted stereoscopic images including 252 distorted stereoscopic images of 4 distortion types, the total number of JPEG-compressed distorted stereoscopic images is 60, JPEG 2000-compressed distorted stereoscopic images is 60, Gaussian Blur (Gaussian Blur) distorted stereoscopic images is 60, and h.264-encoded distorted stereoscopic images is 72.
Calculating S according to the steps from the first step to the sixth stepdisThe image quality objective evaluation prediction value of each distorted stereo image in the distorted stereo image set is calculated respectively through the operation of the image quality objective evaluation prediction value.
The correlation between the objective evaluation prediction value of the image quality of the distorted stereo image obtained in the present embodiment and the average subjective score difference value is analyzed by using 252 distorted stereo images of 12 undistorted stereo images shown in fig. 2a to 13b under different degrees of JPEG compression, JPEG2000 compression, gaussian blur and h.264 coding distortion. Here, 4 common objective parameters of the evaluation method for evaluating image quality are used as evaluation indexes, that is, Pearson correlation coefficient (PLCC), Spearman correlation coefficient (SROCC), Kendall correlation coefficient (KROCC), mean square error (RMSE), stereo image evaluation objective model accuracy in which PLCC and RMSE reflect distortion, and SROCC and KROCC reflect monotonicity thereof under nonlinear regression conditions. The image quality objective evaluation predicted value of the distorted stereo image calculated according to the method is subjected to five-parameter Logistic function nonlinear fitting, and the higher the PLCC, SROCC and KROCC values are, the lower the RMSE value is, the better the correlation between the objective evaluation method and the average subjective score difference is. The Pearson correlation coefficient, the Spearman correlation coefficient, the Kendall correlation coefficient and the mean square error between the objective evaluation predicted value of the image quality and the subjective score of the distorted three-dimensional image obtained by respectively adopting the method and not adopting the method are compared, the comparison results are shown in tables 1, 2, 3 and 4, and as can be seen from tables 1, 2, 3 and 4, the correlation between the final objective evaluation predicted value of the image quality and the average subjective score difference value of the distorted three-dimensional image obtained by adopting the method is very high, which indicates that the objective evaluation result is more consistent with the result of human eye subjective perception, and the effectiveness of the method is enough to explain.
Fig. 14 shows a scatter diagram of the difference between the objective evaluation prediction value of the image quality of each distorted stereoscopic image in the distorted stereoscopic image set and the average subjective score, and the more concentrated the scatter is, the better the consistency between the objective evaluation result and the subjective perception is. As can be seen from fig. 14, the scatter diagram obtained by the method of the present invention is more concentrated, and the goodness of fit with the subjective evaluation data is higher.
TABLE 1 Pearson correlation coefficient comparison between objective evaluation prediction value and subjective score of image quality for distorted stereoscopic images obtained without and with the method of the present invention
TABLE 2 comparison of Spearman correlation coefficient between objective evaluation prediction value and subjective score of image quality for distorted stereo images obtained without and with the method of the present invention
TABLE 3 comparison of Kendall correlation coefficients between objective evaluation prediction values and subjective scores for image quality of distorted stereo images obtained without using the method of the present invention
TABLE 4 comparison of mean square error between objective evaluation prediction and subjective score of image quality for distorted stereoscopic images obtained with and without the method of the present invention