CN102158710B - Depth view encoding rate distortion judgment method for virtual view quality - Google Patents
Depth view encoding rate distortion judgment method for virtual view quality Download PDFInfo
- Publication number
- CN102158710B CN102158710B CN 201110140492 CN201110140492A CN102158710B CN 102158710 B CN102158710 B CN 102158710B CN 201110140492 CN201110140492 CN 201110140492 CN 201110140492 A CN201110140492 A CN 201110140492A CN 102158710 B CN102158710 B CN 102158710B
- Authority
- CN
- China
- Prior art keywords
- present encoding
- encoding piece
- dis
- distortion
- rec
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Images
Landscapes
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
The invention discloses a depth view encoding rate distortion judgment method for virtual view quality. The method comprises the steps of performing predication on a current encoding block to obtain a predicated block; computing the difference between the current encoding block and the predicated block, and performing discrete cosine transform, quantification and entropy coding on the difference to obtain the code rate of the current encoding block; converting pixel gray values of the current encoding block and the predicated block into parallax values; then computing the distortion of the current encoding block according to the converted parallax values; and finally computing the rate-distortion cost of the current encoding block according to the distortion and the code rate. The method better reflects the influence on the composite virtual view quality by the compression distortion of a depth view according to the influence on the composite virtual view quality by the compression distortion of the depth view, improves the encoding code rate of three-dimensional video, and can be applied to the coding standard of three-dimensional videos.
Description
Technical field
The present invention relates to a kind of rate distortion judgment criterion that improves depth map encoding efficient, belong to the depth map encoding technical field in the 3 D stereo video coding standard.
Background technology
The 3 D stereo video is meant that as main video applications in future the user can enjoy real 3 D stereo video content through 3 D stereo video display device.The correlation technique of 3 D video, such as, the technology such as demonstration of the collection of 3 D stereo video, 3 D stereo video coding, 3 D stereo video are paid close attention to widely.In order to promote the standardization of 3 D stereo video technology, 2002, (Motion Picture Experts Group MPEG) proposed any viewpoint TV (Free View Television, notion FTV) in Motion Picture Experts Group.It can provide vividly real, interactively 3 D stereo audiovisual system.The user can watch the 3 D stereo video of this angle from different angles, makes the user have and incorporates the sense of reality in the video scene.FTV can be widely used in fields such as broadcast communication, amusement, education, medical treatment and video monitoring.In order to make the user can watch 3 D stereo video at any angle, FTV system service end uses the video camera array of having demarcated to obtain the video on certain viewpoint.And, utilize corrected video information to generate the virtual view of virtual view through the virtual view synthetic technology to the video correction on the different points of view.MPEG suggestion is at present specifically used based on the degree of depth-image (Depth-Image Based Rendering, virtual view synthetic technology DIBR).Depth information is generally represented through depth map.The main process that virtual view synthesizes is following:
1). confirm to want the relative position of virtual view in video camera array.
2). confirm to be used for the texture video of synthetic virtual view.
3). confirm step 2) the corresponding depth map of texture video.
4). according to step 2) with 3) and in texture video and depth map, adopt the DIBR technology, synthetic virtual view.
The standardization effort of FTV is divided into two stages to carry out.Phase I is 2006 to 2008 the expansion scheme-MVC H.264/AVC (Multi-View Video Coding) that formulated by JVT (Joint Video Team, joint video code sets).MVC can encode to many viewpoints texture video.But to finally realize the function of FTV system, also must encode depth information.The standardization formulation work of FTV has at present got into second stage, i.e. 3DVC (Three Dimensional Video Coding).3DVC mainly pays close attention to the expression and the coding of depth information, and the combined coding of texture video and depth information.Among the 3DVC, depth information is represented through depth map.
The leading indicator of weighing the 3DVC performance is the quality of synthetic virtual view, and the encoder bit rate of texture video, depth map.The quality of virtual view:
Usually (Peak Signal-to-Noise Ratio PSNR) weighs the quality of video to adopt Y-PSNR.The computing formula of PSNR is shown below,
(1) MSE representes the mean square error between original view and the synthetic virtual view in the formula, is used for weighing the distortion of virtual view, and the coding distortion of the coding distortion of texture video, depth map.
In practical application, the view of virtual view is non-existent, does not also promptly have original view.But,, at first adopt the existing texture video of un-encoded and the corresponding synthetic virtual view V of depth map thereof therefore for weighing the performance of 3DVC because 3DVC mainly pays close attention to coding efficiency
Orig, the depth map that adopts the texture video of the reconstruction after process is encoded and the back of encoding to rebuild then synthesizes virtual view V
Rec, at last through calculating V
RecWith V
OrigBetween MSE, and then obtain PSNR, to weigh the performance of 3DVC.
Fig. 1 has provided existing rate distortion criterion calculation flow chart H.264/AVC, in video encoding standard H.264/AVC, at first the present encoding piece is predicted, calculates Mean Square Error MSE between present encoding piece and the predict blocks then as distortion D
H264Then the grey scale pixel value of present encoding piece is deducted the grey scale pixel value of predict blocks, obtain prediction difference, prediction difference is carried out discrete cosine transform, quantize, entropy coding obtains the encoder bit rate R of present encoding piece then
H264, adopt the rate distortion costs J of computes present encoding piece at last
H264: J
H264=D
H266+ λ R
H264, wherein λ is a Lagrange multiplier.
H.264/AVC the rate distortion judgment criterion that adopts is not considered the influence of the compression artefacts of depth map to the quality of synthetic virtual view, and the efficient that therefore adopts rate distortion judgment criterion H.264/AVC that depth map is encoded is not high.Can know that by theory analysis the distortion of synthetic virtual view is directly by the parallax decision, and the quality of depth map can have influence on the accuracy of parallax.In the compression process of depth map, the distortion of depth map itself not necessarily can cause the inaccurate of parallax; But when the depth map entirely accurate, parallax also is an entirely accurate.
Summary of the invention
Do not consider of the influence of the compression artefacts of depth map to the quality of synthetic virtual view to the rate distortion judgment criterion that H.264/AVC adopts; The present invention proposes a kind of rate distortion determination methods towards the virtual view quality that is applicable to depth map encoding according to the influence to synthetic virtual view quality of the compression artefacts of depth map.
The present invention towards the depth map encoding rate distortion determination methods of virtual view quality is:
The present encoding piece is predicted, obtained predict blocks; Calculate the difference of present encoding piece and predict blocks, and difference is carried out discrete cosine transform, quantification and entropy coding, obtain the code check of present encoding piece; Convert the grey scale pixel value of present encoding piece and predict blocks into parallax value; Calculate the distortion of present encoding piece then according to the parallax value after the conversion; At last according to the distortion of gained and the rate distortion costs of code check calculating present encoding piece; Concrete steps are following:
(1) the present encoding piece of depth map is predicted that obtain the predict blocks of present encoding piece, the gray value of each pixel in the present encoding piece is used L
iExpression, each grey scale pixel value in the predict blocks is used L
P, iExpression, i ∈ 1 ... N}, N are the pixel quantity in the present encoding piece;
The present encoding piece of depth map is carried out forecast method to be adopted in the various frames of H.264/AVC standard code or inter-frame prediction method.
(2) with L
iWith L
P, iSubtract each other, obtain the difference of present encoding piece, and difference is carried out discrete cosine transform, quantification and entropy coding, confirm the code check R of present encoding piece;
(3) the entropy coding data are decoded, and the data of decoding are carried out inverse quantization and inverse discrete cosine transformation, rebuild difference signal;
(4) with the difference signal addition of present encoding piece and reconstruction, rebuild current block;
(5) convert each grey scale pixel value in the reconstruction current block of present encoding piece and step (4) gained into parallax value respectively according to following formula,
Dis wherein
iThe pairing parallax value of gray value of i pixel in the expression present encoding piece, dis
Rec, iThe pairing parallax value of gray value of i pixel in the current block, L are rebuild in expression
Rec, iThe grey scale pixel value of current block is rebuild in expression, and f representes focus of camera, and b is the spacing between the adjacent camera, Z
NearThe real depth value of the object point that expression is nearest apart from video camera, Z
FarExpression is apart from the real depth value of video camera object point farthest;
(6) according to dis
iWith dis
Rec, iCalculate the distortion D of present encoding piece,
Wherein, work as dis
iWith dis
Rec, iWhen equating, δ
iEqual 0, work as dis
iWith dis
Rec, iWhen unequal, δ
iEqual L
i-L
Rec, i
(7) according to the code check R of the present encoding piece of step (2) gained and the distortion D of step (6) gained, and according to the rate distortion costs J of computes present encoding piece, J=D+ λ R, wherein λ is a Lagrange multiplier.Lagrange multiplier λ adopts the numerical value of H.264/AVC standard code.
The present invention confirms rate distortion according to the compression artefacts of depth map to the influence of synthetic virtual view quality, can better reflect the influence of the compression artefacts of depth map to synthetic virtual view quality, improves the efficient of 3 D stereo video coding; Only needing degree of depth grayvalue transition is corresponding parallax value, and according to the parallax value calculated distortion, calculates and simply be easy to realization, can not increase encoder complexity.
Description of drawings
Fig. 1 is a rate distortion criterion calculation flow chart H.264/AVC;
Fig. 2 is a depth map encoding rate distortion determination methods flow chart of the present invention;
Fig. 3 is the rate distortion curve comparison diagram that adopts respectively after method of the present invention and method are H.264/AVC encoded to depth map.
Embodiment
Depth map encoding rate distortion determination methods towards the virtual view quality of the present invention, as shown in Figure 2, specifically comprise the steps:
(1) the present encoding piece of depth map is predicted that obtain the predict blocks of present encoding piece, the gray value of each pixel in the present encoding piece is used L
iExpression, each grey scale pixel value in the predict blocks is used L
P, iExpression, i ∈ 1 ... N}, N are the pixel quantity in the present encoding piece.The present encoding piece of depth map is carried out forecast method to be adopted in the various frames of H.264/AVC standard code or inter-frame prediction method.
(2) with L
iWith L
P, iSubtract each other, obtain the difference of present encoding piece, and difference is carried out discrete cosine transform, quantize, entropy coding is confirmed the code check R of present encoding piece;
(3) the entropy coding data are decoded, and the data of decoding are carried out inverse quantization and inverse discrete cosine transformation, rebuild difference signal;
(4) with the difference signal addition of present encoding piece and reconstruction, rebuild current block;
(5) convert each grey scale pixel value in the predict blocks of present encoding piece and step 1 gained into parallax value respectively according to following formula,
Dis wherein
iThe pairing parallax value of gray value of i pixel in the expression present encoding piece, dis
Rec, iThe pairing parallax value of gray value of i pixel in the expression reconstructed block, L
Rec, iThe grey scale pixel value of expression reconstructed block, f representes focus of camera, b is the spacing between the adjacent camera, Z
NearThe real depth value of the object point that expression is nearest apart from video camera, Z
FarExpression is apart from the real depth value of video camera object point farthest;
(6) according to dis
iWith dis
Rec, iCalculate the distortion D of present encoding piece,
Wherein, work as dis
iWith dis
Rec, iWhen equating, δ
iEqual 0, work as dis
iWith dis
Rec, iWhen unequal, δ
iEqual L
i-L
P, i
(7) according to the code check R of the present encoding piece of step 2 gained and the distortion D of step 6 gained, and according to the rate distortion costs J of computes present encoding piece,
J=D+λ·R,
Wherein λ is a Lagrange multiplier.Lagrange multiplier λ adopts the numerical value of H.264/AVC standard code.
Effect of the present invention can further specify through experiment.
Experiment test under the different quantized parameters conditions, the encoder bit rate after adopting the present invention that depth map is encoded and the objective quality PSNR of synthetic virtual view.Fig. 3 has compared the rate distortion curve after adopting the present invention and existing method H.264/AVC to depth map encoding.Wherein Fig. 3 (a) is the experimental result that the depth map of 3 D video sequence B ookarrival is encoded; Fig. 3 (b) is the experimental result that the depth map of 3 D video sequence Kendo is encoded, and Fig. 3 (c) is the experimental result that the depth map of 3 D video sequence Lovebirdl is encoded.Visible by Fig. 3, and H.264/AVC compare, adopt the present invention to encode after, under the identical condition of the encoder bit rate of depth map, the objective quality of synthetic virtual view is higher, explains that the present invention has improved the code efficiency of depth map.As far as 3 D video sequence B ookarrival, the objective quality of synthetic virtual view on average increases 0.215dB; As far as 3 D video sequence Kendo, the objective quality of synthetic virtual view on average increases 0.237dB; As far as 3 D video sequence Lovebird1, the objective quality of synthetic virtual view on average increases 0.45dB.
Claims (3)
1. depth map encoding rate distortion determination methods towards the virtual view quality is characterized in that:
The present encoding piece is predicted, obtained predict blocks; Calculate the difference of present encoding piece and predict blocks, and difference is carried out discrete cosine transform, quantification and entropy coding, obtain the code check of present encoding piece; Convert the grey scale pixel value of present encoding piece and predict blocks into parallax value; Calculate the distortion of present encoding piece then according to the parallax value after the conversion; At last according to the distortion of gained and the rate distortion costs of code check calculating present encoding piece; Concrete steps are following:
(1) the present encoding piece of depth map is predicted, obtained the predict blocks of present encoding piece, the gray value of i pixel in the present encoding piece representes that with Li i grey scale pixel value in the predict blocks used L
P, iExpression, i ∈ 1 ... N}, N are the pixel quantity in the present encoding piece;
(2) with L
iWith L
P, iSubtract each other, obtain the difference of present encoding piece, and difference is carried out discrete cosine transform, quantification and entropy coding, confirm the code check R of present encoding piece;
(3) the entropy coding data are decoded, and the data of decoding are carried out inverse quantization and inverse discrete cosine transformation, rebuild difference signal;
(4) with the difference signal addition of present encoding piece and reconstruction, rebuild current block;
(5) convert each grey scale pixel value in the reconstruction current block of present encoding piece and step (4) gained into parallax value respectively according to following formula,
Dis wherein
iThe pairing parallax value of gray value of i pixel in the expression present encoding piece, dis
Rec, iThe pairing parallax value of gray value of i pixel in the current block, L are rebuild in expression
Rec, iI grey scale pixel value in the current block rebuild in expression, and f representes focus of camera, and b is the spacing between the adjacent camera, Z
NearThe real depth value of the object point that expression is nearest apart from video camera, Z
FarExpression is apart from the real depth value of video camera object point farthest;
(6) according to dis
iWith dis
Rec, iCalculate the distortion D of present encoding piece,
Wherein, work as dis
iWith dis
Rec, iWhen equating, δ
iEqual 0, work as dis
iWith ds
Rec, iWhen unequal, δ
iEqual L
i-L
Rec, i
(7) according to the code check R of the present encoding piece of step (2) gained and the distortion D of step (6) gained, and according to the rate distortion costs J of computes present encoding piece, J=D+ λ R, wherein λ is a Lagrange multiplier.
2. the depth map encoding rate distortion determination methods towards the virtual view quality according to claim 1; It is characterized in that, in the said step (1) the present encoding piece of depth map is carried out forecast method and adopt in the various frames of H.264/AVC standard code or inter-frame prediction method.
3. the depth map encoding rate distortion determination methods towards the virtual view quality according to claim 1 is characterized in that the Lagrange multiplier λ in the said step (7) adopts the numerical value of H.264/AVC standard code.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 201110140492 CN102158710B (en) | 2011-05-27 | 2011-05-27 | Depth view encoding rate distortion judgment method for virtual view quality |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 201110140492 CN102158710B (en) | 2011-05-27 | 2011-05-27 | Depth view encoding rate distortion judgment method for virtual view quality |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102158710A CN102158710A (en) | 2011-08-17 |
CN102158710B true CN102158710B (en) | 2012-12-26 |
Family
ID=44439851
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN 201110140492 Expired - Fee Related CN102158710B (en) | 2011-05-27 | 2011-05-27 | Depth view encoding rate distortion judgment method for virtual view quality |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102158710B (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013113134A1 (en) * | 2012-02-02 | 2013-08-08 | Nokia Corporation | An apparatus, a method and a computer program for video coding and decoding |
CN102595166B (en) * | 2012-03-05 | 2014-03-05 | 山东大学 | Lagrange factor calculation method applied for depth image encoding |
CN103826135B (en) * | 2013-12-24 | 2017-02-08 | 浙江大学 | Three-dimensional video depth map coding method based on just distinguishable parallax error estimation |
CN104506856B (en) * | 2015-01-14 | 2017-03-22 | 山东大学 | Method of estimating quality of virtual view applicable to 3D (Three-dimensional) video system |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101478677A (en) * | 2008-12-24 | 2009-07-08 | 西安交通大学 | Scalable multi-description video encoding structure design method based on code rate control |
CN101888566A (en) * | 2010-06-30 | 2010-11-17 | 清华大学 | Estimation method of distortion performance of stereo video encoding rate |
CN102065296A (en) * | 2011-01-06 | 2011-05-18 | 宁波大学 | Three-dimensional video coding method |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2092747B1 (en) * | 2006-12-14 | 2015-08-12 | Thomson Licensing | Method and apparatus for encoding and/or decoding bit depth scalable video data using adaptive enhancement layer prediction |
-
2011
- 2011-05-27 CN CN 201110140492 patent/CN102158710B/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101478677A (en) * | 2008-12-24 | 2009-07-08 | 西安交通大学 | Scalable multi-description video encoding structure design method based on code rate control |
CN101888566A (en) * | 2010-06-30 | 2010-11-17 | 清华大学 | Estimation method of distortion performance of stereo video encoding rate |
CN102065296A (en) * | 2011-01-06 | 2011-05-18 | 宁波大学 | Three-dimensional video coding method |
Non-Patent Citations (1)
Title |
---|
杨海涛,等.三维电视系统中的视频-深度联合预测编码.《光学学报》.2009,第29卷(第12期), * |
Also Published As
Publication number | Publication date |
---|---|
CN102158710A (en) | 2011-08-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102209243B (en) | Depth map intra prediction method based on linear model | |
CN101159875B (en) | Double forecast video coding/decoding method and apparatus | |
US10009611B2 (en) | Visual quality measure for real-time video processing | |
CN102281446B (en) | Visual-perception-characteristic-based quantification method in distributed video coding | |
CN101835056A (en) | Allocation method for optimal code rates of texture video and depth map based on models | |
CN102413353B (en) | Method for allocating code rates of multi-view video and depth graph in stereo video encoding process | |
Yuan et al. | Rate distortion optimized inter-view frame level bit allocation method for MV-HEVC | |
CN107277550A (en) | Multi-view signal codec | |
CN102625102B (en) | H.264/scalable video coding medius-grain scalability (SVC MGS) coding-oriented rate distortion mode selection method | |
CN102158710B (en) | Depth view encoding rate distortion judgment method for virtual view quality | |
CN103546758A (en) | Rapid depth map sequence interframe mode selection fractal coding method | |
CN107864380A (en) | 3D HEVC fast intra-mode prediction decision-making techniques based on DCT | |
CN102308583A (en) | Apparatus and method for encoding and decoding multi-view image | |
CN102291582A (en) | Distributed video encoding method based on motion compensation refinement | |
US20140340478A1 (en) | Method and apparatus for depth video coding using endurable view synthesis distortion | |
CN101854555B (en) | Video coding system based on prediction residual self-adaptation regulation | |
CN103634600B (en) | A kind of Video Encoding Mode system of selection based on SSIM evaluation, system | |
CN101309404A (en) | Resolution descending video transcoding method and transcoding device thereof | |
CN101883283B (en) | Control method for code rate of three-dimensional video based on SAQD domain | |
CN106534855B (en) | A kind of Lagrange factor calculation method towards SATD | |
CN102595166B (en) | Lagrange factor calculation method applied for depth image encoding | |
CN102572440B (en) | Multi-viewpoint video transmission method based on depth map and distributed video coding | |
CN104282030A (en) | Image compression device and method | |
CN103379349A (en) | Viewpoint composite predictive encoding method, decoding method, corresponding devices and code stream | |
CN102790881A (en) | Transform domain distributed video coder based on frame-level coding end speed control |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20121226 Termination date: 20160527 |