Biography
Marco Cagnazzo obtained the Laurea (equivalent to the M.S.) degree in Telecommunication Engineering from the Federico II University (Napoli, Italy) in 2002 and the Ph.D. degree in Information and Communication Technology from the Federico II University and the University of Nice-Sophia Antipolis (France) in 2005, defending a doctoral thesis entitled “Wavelet transform and three-dimensional data compression”.
He obtained research grants from the CNIT national laboratory of multimedia communication and from the Centro Regionale di Competenze sull’ICT, and has been a post-doc fellow with the team CReATIVe at the I3S laboratory in Sophia Antipolis (France). Marco Cagnazzo has been a full professor at Telecom-Paris since December 2018, within the Multimedia team where he arrived as maître de conférences (roughly equivalent to associate professor) in 2008. He has achieved the Habilitation à Diriger des Recherches de l’Université Pierre et Marie Curie (Paris) in September 2013. He has been responsible for the Multimedia team from December 2015 to September 2021. Since December 2021 he is an Associate Professor at the University of Padua and an Invited Full Professor at Telecom-Paris.
Marco Cagnazzo has been an Associate Editor for IEEE Transactions on Circuits and Systems for Video Technology and for IEEE Signal Prcessing Letters and is Area Editor for Elsevier Signal Processing: Image Communication. Since 2017, he is a member of the IEEE SPS Technical Committee on Image, Video, and Multidimensional Signal Processing (IVMSP) and of the board of the main French society on signal processing, the GRETSI (Groupe de Recherche sur le Traitement du Signal et des Images).
Moreover, he regularly serves as a reviewer for several journals (among which IEEE Transactions on Image Processing, IEEE Transactions on Circuits and Systems for Video Technology, and IEEE Transactions on Signal Processing), and as reviewer and/or area chair for such conferences as IEEE ICIP, IEEE ICASSP, Eurasip EUSIPCO, IEEE MMSP, IEEE ICME.
He is IEEE Senior Member since February 2011.
He has been on the organizing committees of the 2010 IEEE International Workshop MMSP (Electronic Media Chair) and of the European Workshop on Visual Information Processing EUVIP (Local Arrangement). Moreover, he is Publicity Chair for EUSIPCO 2012 (Bucharest) and Award Chair for ICIP 2014 (Paris).
Marco Cagnazzo speaks Italian (his mother tongue), English, and French.
Research areas
Publications
2015
Meddeb, M.; Cagnazzo, M.; Pesquet-Popescu, B.
ROI-based rate control using tiles for an HEVC encoded video stream over a lossy network Conference
Proceedings – International Conference on Image Processing, ICIP, IEEE Computer Society, 2015.
@conference{11577_3469673,
title = {ROI-based rate control using tiles for an HEVC encoded video stream over a lossy network},
author = { M. Meddeb and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/ICIP.2015.7351028},
year = {2015},
date = {2015-01-01},
booktitle = {Proceedings - International Conference on Image Processing, ICIP},
pages = {1389--1393},
publisher = {IEEE Computer Society},
abstract = {The growth in the use of high definition (HD) and above video resolutions streams has outstripped the rate at which network infrastructure has been deployed. Video streaming applications require appropriate rate control techniques that make use of the specific characteristics of the video content, such as the regions of interest (ROI). With the introduction of high efficiency video coding (HEVC) streams, we consider new coding features to make a novel ROI-based rate control (RC) algorithm. The proposed approach introduces tiling in a ROI-based rate control scheme. It aims at enhancing the quality of important regions (i.e. faces for a videoconferencing system) considering independently coded regions lying within an ROI and helps evaluating the ROI quality under poor channel conditions. Our work consists of two major steps. First, we designed a RC algorithm based on an independent processing of tiles of different regions. Second, we investigate the effect of ROI- and tile-based rate control algorithm on the decoded quality of the stream transmitted over a lossy channel.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Calemme, M.; Cagnazzo, M.; Pesquet-Popescu, B.
Depth coding and perceived quality for 3D video Conference
2015 7th International Workshop on Quality of Multimedia Experience, QoMEX 2015, Institute of Electrical and Electronics Engineers Inc., 2015, ISBN: 978-1-4799-8958-4.
@conference{11577_3469681,
title = {Depth coding and perceived quality for 3D video},
author = { M. Calemme and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/QoMEX.2015.7148116},
isbn = {978-1-4799-8958-4},
year = {2015},
date = {2015-01-01},
booktitle = {2015 7th International Workshop on Quality of Multimedia Experience, QoMEX 2015},
pages = {1--2},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
abstract = {Multi-view video plus depth is emerging as the most flexible format for 3D video representation, as witnessed by the current standardization efforts by ISO and ITU. The depth information allows synthesizing virtual view points, and for its compression various techniques have been proposed. We make a preliminary investigation of the effects on the synthesized views of two different approaches: object-based and block-based, from a perceptual point of view.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Purica, A. I.; Mora, E. G.; Pesquet-Popescu, B.; Cagnazzo, M.; Ionescu, B.
Improved view synthesis by motion warping and temporal hole filling Conference
ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing – Proceedings, vol. 2015-, Institute of Electrical and Electronics Engineers Inc., 2015.
@conference{11577_3469670,
title = {Improved view synthesis by motion warping and temporal hole filling},
author = { A. I. Purica and E. G. Mora and B. Pesquet-Popescu and M. Cagnazzo and B. Ionescu},
doi = {10.1109/ICASSP.2015.7178158},
year = {2015},
date = {2015-01-01},
booktitle = {ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings},
volume = {2015-},
pages = {1191--1195},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
abstract = {View synthesis received increasing attention over the last years, as it offers a wide range of practical applications like Free Viewpoint Television, 3D video, video gaming, etc. The main issues in view synthesis are the filling of disoccluded areas and the warping of real views. In this paper we propose a new hole filling method, it uses temporal correlations in the real views to extract information on disoccluded areas from different time instants in the synthetic view. We also propose a sub-pixel warping technique that takes into account depth and can be used for both the warping of the real view as well as for motion compensation. Our method is proved to bring gains of up to 0.31dB in average over several multiview test sequences.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Abou-Elailah, A.; Dufaux, F.; Farah, J.; Cagnazzo, M.; Srivastava, A.; Pesquet-Popescu, B.
Fusion of global and local motion estimation using foreground objects for distributed video coding Journal Article
In: IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, vol. 25, no. 6, pp. 973–987, 2015.
@article{11577_3469369,
title = {Fusion of global and local motion estimation using foreground objects for distributed video coding},
author = { A. Abou-Elailah and F. Dufaux and J. Farah and M. Cagnazzo and A. Srivastava and B. Pesquet-Popescu},
doi = {10.1109/TCSVT.2014.2358872},
year = {2015},
date = {2015-01-01},
journal = {IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY},
volume = {25},
number = {6},
pages = {973--987},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
abstract = {The side information (SI) in Distributed Video Coding (DVC) is estimated using the available decoded frames and exploited for the decoding and reconstruction of other frames. The quality of the SI has a strong impact on the performance of DVC. Here, we propose a new approach that combines both global and local SI to improve coding performance. Since the background pixels in a frame are assigned to global estimation and the foreground objects to local estimation, one needs to estimate foreground objects in the SI using the backward and forward foreground objects, the background pixels are directly taken from the global SI. Specifically, elastic curves and local motion compensation are used to generate the foreground objects masks in the SI. Experimental results show that, as far as the rate-distortion performance is concerned, the proposed approach can achieve a PSNR improvement of up to 1.39 dB for a group of picture (GOP) size of 2, and up to 4.73 dB for larger GOP sizes, with respect to the reference DISCOVER codec.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Dricot, A.; Jung, J.; Cagnazzo, M.; Pesquet, B.; Dufaux, F.
Full parallax 3D video content compression Book Chapter
In: Novel 3D Media Technologies, pp. 49–70, Springer New York, 2015, ISBN: 978-1-4939-2026-6.
@inbook{11577_3469270,
title = {Full parallax 3D video content compression},
author = { A. Dricot and J. Jung and M. Cagnazzo and B. Pesquet and F. Dufaux},
doi = {10.1007/978-1-4939-2026-6_4},
isbn = {978-1-4939-2026-6},
year = {2015},
date = {2015-01-01},
booktitle = {Novel 3D Media Technologies},
pages = {49--70},
publisher = {Springer New York},
abstract = {Motion parallax is a key cue in the perception of the depth that current 3D stereoscopy and auto-stereoscopy technologies are not able to reproduce. Integral imaging and Super Multi-View video (SMV) are 3D technologies that allow creating a light-field representation of a scene with a smooth full motion parallax (i.e., in horizontal and vertical directions). However the large amount of data required is challenging and implies a need for new efficient coding technologies. This chapter first describes integral imaging and SMV content, acquisition, and display. Then it provides an overview of state-of-the-art methods for full parallax 3D content compression. Finally, several coding schemes are compared and a coding structure that exploits inter-view correlations in both horizontal and vertical directions is proposed. The new structure provides a rate reduction (for the same quality) up to 29.1% when compared to a basic anchor structure. Neighboring Block Disparity Vector (NBDV) and Inter-View Motion Prediction (IVMP) coding tools are further improved to efficiently exploit coding structures in two dimensions, with rate reduction up to 4.2% with respect to the reference 3D-HEVC encoder.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Cagnazzo, M.; Kieffer, M.
Shannon-Kotelnikov mappings for softcast-based joint source-channel video coding Conference
Proceedings – International Conference on Image Processing, ICIP, vol. 2015-, IEEE Computer Society, 2015.
@conference{11577_3469513,
title = {Shannon-Kotelnikov mappings for softcast-based joint source-channel video coding},
author = { M. Cagnazzo and M. Kieffer},
doi = {10.1109/ICIP.2015.7350967},
year = {2015},
date = {2015-01-01},
booktitle = {Proceedings - International Conference on Image Processing, ICIP},
volume = {2015-},
pages = {1085--1089},
publisher = {IEEE Computer Society},
abstract = {This paper introduces Shannon-Kotelnikov (SK) mapping in the SoftCast joint source-channel video coding scheme. On bandwidth constrained channels, the performance of SoftCast saturates, due to the large amount of data (chunks) dropped to match the bandwidth requirements. Using SK mapping, it is possible to increase the number of chunks that may be transmitted without increasing the bandwidth requirements. The resulting scheme has an increased number of design parameters for which we present a transmission-power constrained optimization. This extends range of channel SNRs over which the PSNR gracefully increases and improves the end-to-end performance at medium to high SNRs. The price to be paid is a performance degradation at low SNRs.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Dricot, A.; Jung, J.; Cagnazzo, M.; Pesquet, B.; Dufaux, F.
Integral images compression scheme based on view extraction Conference
2015 23rd European Signal Processing Conference, EUSIPCO 2015, Institute of Electrical and Electronics Engineers Inc., 2015, ISBN: 978-0-9928-6263-3.
@conference{11577_3469680,
title = {Integral images compression scheme based on view extraction},
author = { A. Dricot and J. Jung and M. Cagnazzo and B. Pesquet and F. Dufaux},
doi = {10.1109/EUSIPCO.2015.7362353},
isbn = {978-0-9928-6263-3},
year = {2015},
date = {2015-01-01},
booktitle = {2015 23rd European Signal Processing Conference, EUSIPCO 2015},
pages = {101--105},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
abstract = {Integral imaging is a glasses-free 3D video technology that captures a light-field representation of a scene. This representation eliminates many of the limitations of current stereoscopic and autostereoscopic techniques. However, integral images have a large resolution and a structure based on microimages which is challenging to encode. In this paper a compression scheme for integral images based on view extraction is proposed. Average BD-rate gains of 15.7% and up to 31.3% are reported over HEVC. Parameters of the proposed coding scheme can take a large range of values. Results are first provided with an exhaustive search of the best configuration. Then an RD criterion is proposed to avoid exhaustive search methods, saving runtime while preserving the gains. Finally, additional runtime savings are reported by exploring how the different parameters interact.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
2014
Pesquet-Popescu, Béatrice; Cagnazzo, Marco; Dufaux, Frédéric
Motion Estimation—A Video Coding Viewpoint Book Chapter
In: Academic Press Library in Signal Processing: Image and Video Compression and Multimedia, vol. 5, pp. 27–92, Elsevier, 2014, ISBN: 9780124201491.
@inbook{11577_3471590,
title = {Motion Estimation—A Video Coding Viewpoint},
author = { Béatrice Pesquet-Popescu and Marco Cagnazzo and Frédéric Dufaux},
doi = {10.1016/B978-0-12-420149-1.00002-8},
isbn = {9780124201491},
year = {2014},
date = {2014-01-01},
booktitle = {Academic Press Library in Signal Processing: Image and Video Compression and Multimedia},
volume = {5},
pages = {27--92},
publisher = {Elsevier},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Mora, Elie Gabriel; Cagnazzo, Marco; Dufaux, Frédéric
METHOD OF TRANSCODING VIDEO DATA WITH FUSION OF CODING UNITS, COMPUTER PROGRAM, TRANSCODING MODULE AND TELECOMMUNICATIONS EQUIPMENT ASSOCIATED THEREWITH Miscellaneous
2014.
@misc{11577_3472310,
title = {METHOD OF TRANSCODING VIDEO DATA WITH FUSION OF CODING UNITS, COMPUTER PROGRAM, TRANSCODING MODULE AND TELECOMMUNICATIONS EQUIPMENT ASSOCIATED THEREWITH},
author = { Elie Gabriel Mora and Marco Cagnazzo and Frédéric Dufaux},
year = {2014},
date = {2014-01-01},
abstract = {Method of transcoding video data with fusion of coding units, computer program, transcoding module and telecommunications equipment associated therewith. Method of transcoding video data between a first and a second format (F1, F2), the method comprising a step of decoding the binary stream (FB1) providing decoded video data, data representative of the coding structure of the frames in the first format (F1) and, for all or some of the first coding units, prediction data, and a step of re-encoding in the course of which the decoded video data are encoded in the second format (F2). During the re-encoding step, an intermediate coding structure is constructed, comprising intermediate coding units constructed so as to correspond to the fusion of one or more first coding units, prediction data are allocated to each of the intermediate coding units, and the decoded video data are re-encoded in the second format (F2) as a function of the intermediate coding structure.},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Greco, C.; Nemoianu, I. D.; Cagnazzo, M.; Feuvre, J. Le; Dufaux, F.; Pesquet-Popescu, B.
Multimedia Streaming Book Chapter
In: Academic Press Library in Signal Processing: Image and Video Compression and Multimedia, vol. 5, pp. 327–381, Elsevier, 2014, ISBN: 9780124201491.
@inbook{11577_3471588,
title = {Multimedia Streaming},
author = { C. Greco and I. D. Nemoianu and M. Cagnazzo and J. Le Feuvre and F. Dufaux and B. Pesquet-Popescu},
doi = {10.1016/B978-0-12-420149-1.00011-9},
isbn = {9780124201491},
year = {2014},
date = {2014-01-01},
booktitle = {Academic Press Library in Signal Processing: Image and Video Compression and Multimedia},
volume = {5},
pages = {327--381},
publisher = {Elsevier},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Mora, E. G.; Jung, J.; Cagnazzo, M.; Pesquet-Popescu, B.
Initialization, limitation, and predictive coding of the depth and texture quadtree in 3D-HEVC Journal Article
In: IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, vol. 24, no. 9, pp. 1554–1565, 2014.
@article{11577_3469371,
title = {Initialization, limitation, and predictive coding of the depth and texture quadtree in 3D-HEVC},
author = { E. G. Mora and J. Jung and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/TCSVT.2013.2283110},
year = {2014},
date = {2014-01-01},
journal = {IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY},
volume = {24},
number = {9},
pages = {1554--1565},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
abstract = {The 3D video extension of High Efficiency Video Coding (3D-HEVC) exploits texture-depth redundancies in 3D videos using intercomponent coding tools. It also inherits the same quadtree coding structure as HEVC for both components. The current software implementation of 3D-HEVC includes encoder shortcuts that speed up the quadtree construction process, but those are always accompanied by coding losses. Furthermore, since the texture and its associated depth represent the same scene, at the same time instant and view point, their quadtrees are closely linked. In this paper, an intercomponent tool is proposed in which this link is exploited to save both runtime and bits through a joint coding of the quadtrees. If depth is coded before the texture, the texture quadtree is initialized from the coded depth quadtree. Otherwise, the depth quadtree is limited to the coded texture quadtree. A 31% encoder runtime saving, a-0.3% gain for coded and synthesized views and a-1.8% gain for coded views are reported for the second method.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Fiengo, A.; Chierchia, G.; Cagnazzo, M.; Pesquet-Popescu, B.
A convex-optimization framework for frame-level optimal rate allocation in predictive video coding Conference
ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing – Proceedings, Institute of Electrical and Electronics Engineers Inc., 2014, ISBN: 978-1-4799-2893-4.
@conference{11577_3469674,
title = {A convex-optimization framework for frame-level optimal rate allocation in predictive video coding},
author = { A. Fiengo and G. Chierchia and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/ICASSP.2014.6855023},
isbn = {978-1-4799-2893-4},
year = {2014},
date = {2014-01-01},
booktitle = {ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings},
pages = {7328--7332},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
abstract = {Optimal rate allocation is among the most challenging tasks to perform in the context of predictive video coding, because of the dependencies between frames induced by motion compensation. In this paper, we derive an analytical rate-distortion model that explicitly takes into account the dependencies between frames. The proposed approach allows us to formulate the frame-level optimal rate allocation as a convex optimization problem. Within this framework, we are able to achieve the exact solution in limited time (even for large-size problems), thanks to the flexibility offered by recent convex optimization techniques. Experiments on standard sequences demonstrate the interest of considering the proposed rate-distortion model and confirm that the optimal rate allocation ensures a better distribution of the total bit budget, with superior results (in the rate-distortion sense) with respect to the standard H.264/AVC rate control. © 2014 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Dricot, A.; Jung, J.; Cagnazzo, M.; Pesquet, B.; Dufaux, F.
Full parallax super multi-view video coding Conference
2014 IEEE International Conference on Image Processing, ICIP 2014, Institute of Electrical and Electronics Engineers Inc., 2014, ISBN: 978-1-4799-5751-4.
@conference{11577_3469677,
title = {Full parallax super multi-view video coding},
author = { A. Dricot and J. Jung and M. Cagnazzo and B. Pesquet and F. Dufaux},
doi = {10.1109/ICIP.2014.7025026},
isbn = {978-1-4799-5751-4},
year = {2014},
date = {2014-01-01},
booktitle = {2014 IEEE International Conference on Image Processing, ICIP 2014},
pages = {135--139},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
abstract = {Super Multi-View (SMV) video is a key enabler for future 3D video services that allows a glasses-free visualization and eliminates many causes of discomfort existing in current available 3D video technologies. SMV video content is composed of tens or hundreds of views, that can be aligned in horizontal only or both horizontal and vertical directions, providing respectively horizontal parallax or full parallax. This paper compares several coding schemes and coding orders, and proposes a coding structure that exploits inter-view correlations in the two directions, providing BD-rate gains up to 29.1% when compared to a basic anchor structure. Additionally, Neighboring Block Disparity Vector (NBDV) and Inter-View Motion Prediction (IVMP) coding tools are further improved to efficiently exploit coding structures in two dimensions, with BD-rate gains up to 4.2% reported over the reference 3D-HEVC encoder.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Mora, E. G.; Jung, J.; Cagnazzo, M.; Pesquet-Popescu, B.
Depth video coding based on Intra mode inheritance from texture Journal Article
In: APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING, vol. 3, no. 1, 2014.
@article{11577_3469370,
title = {Depth video coding based on Intra mode inheritance from texture},
author = { E. G. Mora and J. Jung and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1017/ATSIP.2013.8},
year = {2014},
date = {2014-01-01},
journal = {APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING},
volume = {3},
number = {1},
publisher = {Cambridge University Press},
abstract = {With the recent development of new three-dimensional (3D) multimedia services such as 3D television or free viewpoint television, a new 3D video format, called multiview video + depth (MVD) is currently being investigated. MVD allows synthesizing as many views as required at the receiver side, thus providing smooth scene transitions and the ability to experience a new 3D perspective with each different viewing point. The format introduces, alongside traditional 2D image sequences, sequences of depth maps, which must be efficiently coded to achieve good quality for the synthesized views. One approach to code depth videos is to exploit the correlations between texture and depth. In this work, we propose a new tool to code depth videos in which the texture Intra modes are inherited and used as predictors for the depth Intra modes, hence reducing themode signaling bitrate. The tool is only used in prediction units where texture and depth Intra directions, or modes, are expected to match. Two criteria that exploit the statistical dependency between the texture and depth Intramodes are studied in this work: GradientMax and DominantAngle. Average bitrate reductions of 1.3 and 1.6% on synthesized sequences are reported for GradientMax and DominantAngle, respectively. The latter method additionally achieves 2.3% bitrate reduction on depth sequences. © The Authors, 2014.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Petrazzuoli, G.; Maugey, T.; Cagnazzo, M.; Pesquet-Popescu, B.
Depth-based multiview distributed video coding Journal Article
In: IEEE TRANSACTIONS ON MULTIMEDIA, vol. 16, no. 7, pp. 1834–1848, 2014.
@article{11577_3469374,
title = {Depth-based multiview distributed video coding},
author = { G. Petrazzuoli and T. Maugey and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/TMM.2014.2342201},
year = {2014},
date = {2014-01-01},
journal = {IEEE TRANSACTIONS ON MULTIMEDIA},
volume = {16},
number = {7},
pages = {1834--1848},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
abstract = {Multiview distributed video coding (DVC) has gained much attention in the last few years because of its potential in avoiding communication between cameras without decreasing the coding performance. However, the current results are not matching the expectations mainly due to the fact that some theoretical assumptions are not satisfied in the current implementations. For example, in distributed source coding the encoder must know the correlation between the sources, which cannot be achieved in the traditional DVC systems without having a communication between the cameras. In this work, we propose a novel multiview distributed video coding scheme in which the depth maps are used to estimate the way two views are correlated with no exchanges between the cameras. Only their relative positions are known. We design the complete scheme and further propose a rate allocation algorithm to efficiently share the bit budget between the different components of our scheme. Then, a rate allocation algorithm for depth maps is proposed in order to maximize the quality of synthesized virtual views. We show, through detailed experiments, that our scheme significantly outperforms the state-of-the-art DVC system.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Meddeb, M.; Cagnazzo, M.; Pesquet-Popescu, B.
Region-of-interest-based rate control scheme for high-efficiency video coding Journal Article
In: APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING, vol. 3, no. 1, 2014.
@article{11577_3469372,
title = {Region-of-interest-based rate control scheme for high-efficiency video coding},
author = { M. Meddeb and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1017/ATSIP.2014.15},
year = {2014},
date = {2014-01-01},
journal = {APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING},
volume = {3},
number = {1},
publisher = {Cambridge University Press},
abstract = {This paper presents a novel rate control scheme designed for the newest high efficiency video coding (HEVC) standard, and aimed at enhancing the quality of regions of interest (ROI) for a videoconferencing system. It is designed to consider the different regions at both frame level and coding tree unit (CTU) level. The proposed approach allocates a higher bit rate to the region of interest while keeping the global bit rate close to the assigned target value. The ROIs, typically faces in this application, are automatically detected and each CTU is classified in a region of interest map. This binary map is given as input to the rate control algorithm and the bit allocation is made accordingly. The algorithm is tested, first, using the initial version of the controller introduced in HEVC test model (HM.10), then, extended in HM.13. In this work, we first investigate the impact of differentiated bit allocation between the two regions using a fixed bit rate ratio in intra-coded frames (I-frames) and Bidirectionally predicted frames (B-frames). Then, unit quantization parameters (QPs) are computed independently for CTUs of different regions. The proposed approach has been compared to the reference controller implemented in HM and to a ROI-based rate control algorithm initially proposed for H.264 that we adopted to HEVC and implemented in HM.9. Experimental results show that our scheme has comparable performances with the ROI-based controller proposed for H.264. It achieves accurate target bit rates and provides an improvement in region of interest quality, both in objective metrics (up to 2 dB in PSNR) and based on subjective quality evaluation.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Calemme, M.; Cagnazzo, M.; Pesquet-Popescu, B.
Lossless contour coding using elastic curves in multiview video plus depth Journal Article
In: APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING, vol. 3, no. 1, pp. 1–10, 2014.
@article{11577_3469375,
title = {Lossless contour coding using elastic curves in multiview video plus depth},
author = { M. Calemme and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1017/ATSIP.2014.20},
year = {2014},
date = {2014-01-01},
journal = {APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING},
volume = {3},
number = {1},
pages = {1--10},
publisher = {Cambridge University Press},
abstract = {Multiview video plus depth is emerging as the most flexible format for three-dimensional video representation, as witnessed by the current standardization efforts by ISO and ITU. In particular, in depth representation, arguably the most important information lies in object contours. As a consequence, an interesting approach consists in performing a lossless coding of object contours, possibly followed by a lossy coding of per-object depth values. In this context, we propose a new technique for lossless coding of object contours, based on the elastic deformation of curves. Using the square-root velocity representation for the elements of the space of curves, we can model a continuous evolution of elastic deformations between two reference contour curves. An elastically deformed version of the reference contours can be sent to the decoder with a reduced coding cost and used as side information to improve the lossless coding of the actual contour. Experimental results on several multiview video sequences show remarkable gains with respect to the reference techniques and to the state of the art.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Nemoianu, I. -D.; Greco, C.; Cagnazzo, M.; Pesquet-Popescu, B.
On a hashing-based enhancement of source separation algorithms over finite fields with network coding perspectives Journal Article
In: IEEE TRANSACTIONS ON MULTIMEDIA, vol. 16, no. 7, pp. 2011–2024, 2014.
@article{11577_3469373,
title = {On a hashing-based enhancement of source separation algorithms over finite fields with network coding perspectives},
author = { I. -D. Nemoianu and C. Greco and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/TMM.2014.2341923},
year = {2014},
date = {2014-01-01},
journal = {IEEE TRANSACTIONS ON MULTIMEDIA},
volume = {16},
number = {7},
pages = {2011--2024},
publisher = {Institute of Electrical and Electronics Engineers Inc.},
abstract = {Blind Source Separation (BSS) deals with the recovery of source signals from a set of observed mixtures, when little or no knowledge of the mixing process is available. BSS can find an application in the context of network coding, where relaying linear combinations of packets maximizes the throughput and increases the loss immunity. By relieving the nodes from the need to send the combination coefficients, the overhead cost is largely reduced. However, the scaling ambiguity of the technique and the quasi-uniformity of compressed media sources makes it unfit, at its present state, for multimedia transmission. In order to open new practical applications for BSS in the context of multimedia transmission, we have recently proposed to use a non-linear encoding to increase the discriminating power of the classical entropy-based separation methods. Here, we propose to append to each source a non-linear message digest, which offers an overhead smaller than a per-symbol encoding and that can be more easily tuned. Our results prove that our algorithm is able to provide high decoding rates for different media types such as image, audio, and video, when the transmitted messages are less than 1.5 kilobytes, which is typically the case in a realistic transmission scenario.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
2013
Mora, ELIE GABRIEL; Jung, Joël; PESQUET-POPESCU, Béatrice; Cagnazzo, Marco
VIDEO ENCODING AND DECODING BY INHERITANCE OF A FIELD OF MOTION VECTORS Miscellaneous
2013.
@misc{11577_3472316,
title = {VIDEO ENCODING AND DECODING BY INHERITANCE OF A FIELD OF MOTION VECTORS},
author = { ELIE GABRIEL Mora and Joël Jung and Béatrice PESQUET-POPESCU and Marco Cagnazzo},
year = {2013},
date = {2013-01-01},
abstract = {The invention concerns a method of encoding a current image (le) comprising a step (C2) of determining, in a first image (IR-i) different from the current image, a group of k' pixels corresponding to a current group of k pixels (k'≤k) to be encoded in the current image (lc), said encoding method being characterised in that it comprises the following steps: - calculating (C3) a motion vector between each of the k' pixels of the first image (I R-1) and a corresponding pixel of a second image (I R2) different from the current image, after which a field of k' motion vectors is obtained, - predicting (C4) the pixels or the motion of the current group of k pixels in said current image (le) from the obtained field of k' motion vectors.},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Dufaux, F.; Pesquet-Popescu, B.; Cagnazzo, M.
Emerging Technologies for 3D Video: Creation, Coding, Transmission and Rendering Book
John Wiley and Sons, 2013, ISBN: 9781118355114.
@book{11577_3469671,
title = {Emerging Technologies for 3D Video: Creation, Coding, Transmission and Rendering},
author = { F. Dufaux and B. Pesquet-Popescu and M. Cagnazzo},
doi = {10.1002/9781118583593},
isbn = {9781118355114},
year = {2013},
date = {2013-01-01},
publisher = {John Wiley and Sons},
abstract = {With the expectation of greatly enhanced user experience, 3D video is widely perceived as the next major advancement in video technology. In order to fulfil the expectation of enhanced user experience, 3D video calls for new technologies addressing efficient content creation, representation/coding, transmission and display. Emerging Technologies for 3D Video will deal with all aspects involved in 3D video systems and services, including content acquisition and creation, data representation and coding, transmission, view synthesis, rendering, display technologies, human perception of depth and quality assessment. Key features: Offers an overview of key existing technologies for 3D video. Provides a discussion of advanced research topics and future technologies. Reviews relevant standardization efforts. Addresses applications and implementation issues. Includes contributions from leading researchers. The book is a comprehensive guide to 3D video systems and services suitable for all those involved in this field, including engineers, practitioners, researchers as well as professors, graduate and undergraduate students, and managers making technological decisions about 3D video. © 2013 by John Wiley & Sons, Ltd.},
keywords = {},
pubstate = {published},
tppubtype = {book}
}
Mora, E. G.; Jung, J.; Pesquet-Popescu, B.; Cagnazzo, M.
Modification of the disparity vector derivation process in 3D-HEVC Conference
2013 IEEE International Workshop on Multimedia Signal Processing, MMSP 2013, 2013, ISBN: 978-1-4799-0125-8.
@conference{11577_3469683,
title = {Modification of the disparity vector derivation process in 3D-HEVC},
author = { E. G. Mora and J. Jung and B. Pesquet-Popescu and M. Cagnazzo},
doi = {10.1109/MMSP.2013.6659289},
isbn = {978-1-4799-0125-8},
year = {2013},
date = {2013-01-01},
booktitle = {2013 IEEE International Workshop on Multimedia Signal Processing, MMSP 2013},
pages = {206--211},
abstract = {The up-and-coming extension of HEVC for 3D video (3D-HEVC) includes various tools to exploit different redundancies in a 3D video signal. Inter-view redundancies are in particular exploited using Inter-View Motion Prediction (IVMP) and Inter-View Residual Prediction (IVRP). Both of these tools compensate disparity-wise the current prediction unit (PU) in order to find its corresponding PU in a base view, from which some prediction information for the current PU is retrieved. The disparity vector (DV) used for disparity compensation is currently derived using a neighboring search process (NBDV) for a DV across spatial and temporal neighbors. The first DV found is selected as the final DV used in IVMP and IVRP, with no guarantee of optimality. In this paper, the NBDV derivation process is changed: all found DVs from different neighbors are stored in a list. Redundant vectors in this list are removed, and a median computation on the remaining vectors is performed. The resulting DV is set as the DV used for IVMP. Average bitrate reductions of 0.6% and 0.8% for the two dependent views and 0.2% on synthesized views are reported with only a slight increase in encoder and decoder runtimes. © 2013 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Abou-Elailah, A.; Dufaux, F.; Farah, J.; Cagnazzo, M.; Pesquet-Popescu, B.
Fusion of global and local motion estimation for distributed video coding Journal Article
In: IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, vol. 23, no. 1, pp. 158–172, 2013.
@article{11577_3469377,
title = {Fusion of global and local motion estimation for distributed video coding},
author = { A. Abou-Elailah and F. Dufaux and J. Farah and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/TCSVT.2012.2203211},
year = {2013},
date = {2013-01-01},
journal = {IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY},
volume = {23},
number = {1},
pages = {158--172},
abstract = {The quality of side information plays a key role in distributed video coding. In this paper, we propose a new approach that consists of combining global and local motion compensation at the decoder side. The parameters of the global motion are estimated at the encoder using scale invariant feature transform features. Those estimated parameters are sent to the decoder in order to generate a globally motion compensated side information. Conversely, a locally motion compensated side information is generated at the decoder based on motion-compensated temporal interpolation of neighboring reference frames. Moreover, an improved fusion of global and local side information during the decoding process is achieved using the partially decoded Wyner-Ziv frame and decoded reference frames. The proposed technique improves significantly the quality of the side information, especially for sequences containing high global motion. Experimental results show that, as far as the rate-distortion performance is concerned, the proposed approach can achieve a PSNR improvement of up to 1.9 dB for a Group of Pictures (GOP) size of 2, and up to 4.65 dB for larger GOP sizes, with respect to the reference DISCOVER codec. © 1991-2012 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Petrazzuoli, G.; Macovei, C.; Nicolae, I. -E.; Cagnazzo, M.; Dufaux, F.; Pesquet-Popescu, B.
Versatile multiview layered video based on distributed source coding Conference
International Workshop on Image Analysis for Multimedia Interactive Services, 2013, ISBN: 978-1-4799-0833-2.
@conference{11577_3469676,
title = {Versatile multiview layered video based on distributed source coding},
author = { G. Petrazzuoli and C. Macovei and I. -E. Nicolae and M. Cagnazzo and F. Dufaux and B. Pesquet-Popescu},
doi = {10.1109/WIAMIS.2013.6616157},
isbn = {978-1-4799-0833-2},
year = {2013},
date = {2013-01-01},
booktitle = {International Workshop on Image Analysis for Multimedia Interactive Services},
pages = {1--4},
abstract = {Video content constitutes today a large part of the data traffic on the Internet. This is allowed by the capillary spreading of video codec technologies: nowadays, every computer, tablet and smart phone is equipped with video encoding and decoding technologies. As a matter of fact, the video content often exists in different formats, that, even though they can be incompatible to each other, still have a significant mutual redun-dancy. The incompatibility prevents an efficient exploitation of the scalability, which on the other hand is a very important characteristic when it comes to efficient network use. An interesting alternative to classical scalable video is to use distributed video coding (DVC) for the enhancement layers. In the envisaged scenario, clients have different decoders for the base layer, adapted to the characteristics of their de-vice. However they can share the same enhancement layer, since DVC allows encoding frames independently from the reference that will be employed at the decoder. This approach has been considered in the past in order to improve temporal and spatial scalability. In this work we review the existing approaches, improve them using more recent DVC techniques and perform a new analysis for the emerging multi-view applications. © 2013 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Maugey, T.; Gauthier, J.; Cagnazzo, M.; Pesquet-Popescu, B.
Evaluation of side information effectiveness in distributed video coding Journal Article
In: IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, vol. 23, no. 12, pp. 2116–2126, 2013.
@article{11577_3469376,
title = {Evaluation of side information effectiveness in distributed video coding},
author = { T. Maugey and J. Gauthier and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/TCSVT.2013.2273623},
year = {2013},
date = {2013-01-01},
journal = {IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY},
volume = {23},
number = {12},
pages = {2116--2126},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Nemoianu, I.; Greco, C.; Castella, M.; Pesquet-Popescu, B.; Cagnazzo, M.
On a practical approach to source separation over finite fields for network coding applications Conference
ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing – Proceedings, 2013, ISBN: 978-1-4799-0356-6.
@conference{11577_3471522,
title = {On a practical approach to source separation over finite fields for network coding applications},
author = { I. Nemoianu and C. Greco and M. Castella and B. Pesquet-Popescu and M. Cagnazzo},
doi = {10.1109/ICASSP.2013.6637868},
isbn = {978-1-4799-0356-6},
year = {2013},
date = {2013-01-01},
booktitle = {ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings},
pages = {1335--1339},
abstract = {In Blind Source Separation, or BSS, a set of source signals are recovered from a set of mixed observations without knowledge of the mixing parameters. Originated for real signals, BSS has recently been applied to finite fields, enabling more practical applications. However, classical entropy-based techniques do not perform well in finite fields. Here, we propose a non-linear encoding of the sources to increase the discriminating power of the separation methods. Our results show that the encoding improves the success rate of the separation for sources with few samples in large finite fields, both conditions met in practical networking applications. Our results open new possibilities in the context of network coding-wherein linear combinations of packets are sent in order to maximize throughput and increase loss immunity- by relieving the nodes from the need to send the combination coefficients, thus reducing the overhead cost. © 2013 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Mora, E. G.; Valenzise, G.; Jung, J.; Pesquet-Popescu, B.; Cagnazzo, M.; Dufaux, F.
Depth Video Coding Technologies Book Chapter
In: Emerging Technologies for 3D Video: Creation, Coding, Transmission and Rendering, pp. 121–138, John Wiley and Sons, 2013, ISBN: 9781118355114.
@inbook{11577_3469389,
title = {Depth Video Coding Technologies},
author = { E. G. Mora and G. Valenzise and J. Jung and B. Pesquet-Popescu and M. Cagnazzo and F. Dufaux},
doi = {10.1002/9781118583593.ch7},
isbn = {9781118355114},
year = {2013},
date = {2013-01-01},
booktitle = {Emerging Technologies for 3D Video: Creation, Coding, Transmission and Rendering},
pages = {121--138},
publisher = {John Wiley and Sons},
abstract = {The emergence of new three-dimensional (3D) multimedia services, such as 3D television and free viewpoint television, created a need for multiple-view multiplexing on 3D displays for more fluidity and scalability. However, coding and transmitting more views is costly. Depth-based formats allow synthesizing the required number of views at a reduced cost compared with video-only formats. Nevertheless, depth videos, alongside texture or color videos, should also be efficiently compressed. A number of depth video coding technologies have been proposed in recent years, both in academic and industrial contexts. These technologies can be mainly categorized into three families. We have tools that exploit some inherent characteristics of depth maps, tools that exploit the correlations with the associated texture, and tools that optimize the depth map coding for the quality of the synthesis. This chapter is intended as a tutorial, illustrating those different depth coding tools, while giving corresponding examples of each. © 2013 by John Wiley & Sons, Ltd.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Mora, E. -G.; Jung, J.; Cagnazzo, M.; Pesquet-Popescu, B.
Modification of the merge candidate list for dependent views in 3D-HEVC Conference
2013 IEEE International Conference on Image Processing, ICIP 2013 – Proceedings, IEEE Computer Society, 2013, ISBN: 978-1-4799-2341-0.
@conference{11577_3471465,
title = {Modification of the merge candidate list for dependent views in 3D-HEVC},
author = { E. -G. Mora and J. Jung and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/ICIP.2013.6738352},
isbn = {978-1-4799-2341-0},
year = {2013},
date = {2013-01-01},
booktitle = {2013 IEEE International Conference on Image Processing, ICIP 2013 - Proceedings},
pages = {1709--1713},
publisher = {IEEE Computer Society},
abstract = {A test model for an HEVC-based 3D video coding standard (3D-HEVC) has recently been drafted. 3D-HEVC exploits inter-view redundancies by including disparity-compensated prediction (DCP) for efficient dependent view coding. It also uses the Merge coding mode to reduce the cost of motion / disparity parameters. However, the candidates in the Merge list are mostly temporal motion vectors. DCP does not often benefit from accurate predictors and is thus costly. Consequently, motion-compensated prediction (MCP) remains largely preferred. In this paper, we propose to reduce the cost of DCP by modifying the Merge candidate list to always include a disparity vector candidate. Two methods are proposed: the new candidate is either added in the secondary or in the primary list of candidates. The latter method, which achieves average bitrate reductions of 0.6% for dependent views, and 0.2% for coded and synthesized views, was adopted in both the 3D-HEVC working draft and software. © 2013 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Petrazzuoli, G.; Cagnazzo, M.; Pesquet-Popescu, B.
Novel solutions for side information generation and fusion in multiview DVC Journal Article
In: EURASIP JOURNAL ON ADVANCES IN SIGNAL PROCESSING, vol. 2013, no. 1, 2013.
@article{11577_3469780,
title = {Novel solutions for side information generation and fusion in multiview DVC},
author = { G. Petrazzuoli and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1186/1687-6180-2013-154},
year = {2013},
date = {2013-01-01},
journal = {EURASIP JOURNAL ON ADVANCES IN SIGNAL PROCESSING},
volume = {2013},
number = {1},
publisher = {Springer International Publishing},
abstract = {One of the key problems in distributed video coding is the generation of side information. This task consists of producing an estimate of an image with some neighboring ones, such as those taken by the same camera at different time instants, or, in the case of multiview setups, images taken at the same time instant by different cameras. If both estimates are available, a further problem arises, which is how to merge them in order to create a single side information. This problem is very relevant since a good estimate of the unknown image will require only a few bits to be corrected. Considering a multiview distributed video-coding setup, we propose a novel technique for inter-view interpolation based on occlusion prediction, a new fusion technique from multiple estimates, and finally an adaptive validation step for switching among the three possible side information images: temporal, inter-view, and fusion. We provide a comprehensive set of experimental results, which indicate bit rate reductions of more than 9% in average; moreover, we observe much more consistent results with respect to state-of-the-art techniques.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Kaaniche, M.; Gaetano, R.; Cagnazzo, M.; Pesquet-Popescu, B.
Disparity Estimation Techniques Book Chapter
In: Emerging Technologies for 3D Video: Creation, Coding, Transmission and Rendering, pp. 81–101, John Wiley and Sons, 2013, ISBN: 9781118355114.
@inbook{11577_3469391,
title = {Disparity Estimation Techniques},
author = { M. Kaaniche and R. Gaetano and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1002/9781118583593.ch5},
isbn = {9781118355114},
year = {2013},
date = {2013-01-01},
booktitle = {Emerging Technologies for 3D Video: Creation, Coding, Transmission and Rendering},
pages = {81--101},
publisher = {John Wiley and Sons},
abstract = {Disparity information has a great interest in different application fields such as view synthesis, 3D reconstruction and stereo/multiview video coding. For this reason, many research studies have been focused on the estimation of a consistent disparity map. This chapter deals with this estimation problem and it is divided into three main parts. The first part is devoted to analyzing the geometrical relationship between the 3D scene and the generated stereo images. In the second part, the formulation of the disparity estimation problem and the main difficulties that arise in solving this problem are addressed. Finally, in the third part, we present the most important techniques which aim at estimating the disparity map. © 2013 by John Wiley & Sons, Ltd.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Mora, E. G.; Jung, J.; Pesquet-Popescu, B.; Cagnazzo, M.
Coding Methods for Depth Videos Book Chapter
In: 3D Video: From Capture to Diffusion, pp. 229–247, wiley, 2013, ISBN: 9781848215078.
@inbook{11577_3469387,
title = {Coding Methods for Depth Videos},
author = { E. G. Mora and J. Jung and B. Pesquet-Popescu and M. Cagnazzo},
doi = {10.1002/9781118761915.ch12},
isbn = {9781848215078},
year = {2013},
date = {2013-01-01},
booktitle = {3D Video: From Capture to Diffusion},
pages = {229--247},
publisher = {wiley},
abstract = {This chapter presents the different methods or tools that have been developed to code depth videos. A section presents an analysis of depth maps and their specific characteristics. It also includes a discussion of different types of redundancies, which may be exploited in order to maximize compression levels of depth videos. Three main categories of depth coding tools are defined in the next section: (i) methods using the intrinsic characteristics of depth maps such as high-level coding tools and block-based coding tools; (ii) methods exploiting correlation with associated textures such as inheritance/selection of prediction modes and inheritance of prediction information; and (iii) methods optimizing depth coding for the quality of synthesized views such as view synthesis optimization and distortion models.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Petrazzuoli, G.; Maugey, T.; Cagnazzo, M.; Pesquet-Popescu, B.
A distributed video coding system for multi view video plus depth Conference
Conference Record – Asilomar Conference on Signals, Systems and Computers, IEEE Computer Society, 2013, ISBN: 978-1-4799-2388-5.
@conference{11577_3471469,
title = {A distributed video coding system for multi view video plus depth},
author = { G. Petrazzuoli and T. Maugey and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/ACSSC.2013.6810373},
isbn = {978-1-4799-2388-5},
year = {2013},
date = {2013-01-01},
booktitle = {Conference Record - Asilomar Conference on Signals, Systems and Computers},
pages = {699--703},
publisher = {IEEE Computer Society},
abstract = {Multi-view video plus depth (MVD) is gathering huge attention, as witnessed by the recent standardization activity, since its rich information about the geometry of the scene allows high-quality synthesis of virtual viewpoints. Distributed video coding of such kind of content is a challenging problem whose solution could enable new services as interactive multi-view streaming. In this work we propose to exploit the geometrical information of the MVD format in order to estimate inter-view occlusions without communication among cameras. Experimental results show a bit rate reduction up to 77% for low bit rate w.r.t. state-of-the-art architectures. © 2013 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Cagnazzo, M.; Pesquet-Popescu, B.; Dufaux, F.
3D Video Representation and Formats Book Chapter
In: Emerging Technologies for 3D Video: Creation, Coding, Transmission and Rendering, no. 2, pp. 102–120, John Wiley and Sons, 2013, ISBN: 9781118355114.
@inbook{11577_3469390,
title = {3D Video Representation and Formats},
author = { M. Cagnazzo and B. Pesquet-Popescu and F. Dufaux},
doi = {10.1002/9781118583593.ch6},
isbn = {9781118355114},
year = {2013},
date = {2013-01-01},
booktitle = {Emerging Technologies for 3D Video: Creation, Coding, Transmission and Rendering},
number = {2},
pages = {102--120},
publisher = {John Wiley and Sons},
abstract = {Efficient data representation is needed for successful three-dimensional (3D) video applications. This element is closely intertwined with the other components of a 3D video system: content acquisition, transmission, rendering, and display. It also has a significant impact on the overall performance of the system, as well as constraints such as backward compatibility with existing equipment and infrastructure. In the current eco-system, a variety of 3D video formats, coding schemes, and display technologies coexist. In this context, standardization is key to guarantee interoperability and support mass adoption. In this chapter, we first describe 3D video representations, including stereoscopic 3D video, multiview video, video-plus-depth, multiview video-plus-depth, and layered depth video. Next, we discuss existing coding formats, such as simulcast, frame-compatible stereo interleaving, multiview video coding, MPEG-4 multiple auxiliary components, MPEG-C Part 3, and MPEG-2 multiview profile. All these representations and formats are discussed in details, and their respective strengths and shortcomings are analyzed. © 2013 by John Wiley & Sons, Ltd.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
Battin, B.; Vautrot, P.; Cagnazzo, M.; Dufaux, F.
Multiview Video Coding (MVC) Book Chapter
In: 3D Video: From Capture to Diffusion, pp. 193–209, wiley, 2013, ISBN: 9781848215078.
@inbook{11577_3469388,
title = {Multiview Video Coding (MVC)},
author = { B. Battin and P. Vautrot and M. Cagnazzo and F. Dufaux},
doi = {10.1002/9781118761915.ch10},
isbn = {9781848215078},
year = {2013},
date = {2013-01-01},
booktitle = {3D Video: From Capture to Diffusion},
pages = {193--209},
publisher = {wiley},
abstract = {This chapter presents 3D formats and coding techniques for stereoscopic vision as well as for multiview examples. With a view to improving performance and increasing functionality in current formats, MPEG has recently undergone a new normalization phase for 3D video coding (3DVC). Three approaches are currently used in normalization. The first is a backward compatible extension of multiview video coding (MVC). Here, a second stream encodes the depth information independently of the stream representing textural information. The second approach involves a backward compatible extension of H.264/MPEG-4 AVC, where a basic video stream encodes texture information of a view with H.264/MPEG-4 AVC. Finally, a third approach uses a backward compatible extension of high efficiency video coding (HEVC) in which a simple multiview extension of HEVC is made using a schema identical to MVC.},
keywords = {},
pubstate = {published},
tppubtype = {inbook}
}
2012
Abou-Elailah, A.; Dufaux, F.; Farah, J.; Cagnazzo, M.
Fusion of global and local side information using Support Vector Machine in transform-domain DVC Conference
European Signal Processing Conference, 2012.
@conference{11577_3469693,
title = {Fusion of global and local side information using Support Vector Machine in transform-domain DVC},
author = { A. Abou-Elailah and F. Dufaux and J. Farah and M. Cagnazzo},
year = {2012},
date = {2012-01-01},
booktitle = {European Signal Processing Conference},
pages = {574--578},
abstract = {Side information has a strong impact on the performance of Distributed Video Coding. Commonly, side information is generated using motion compensated temporal interpolation. In this paper, we propose a new method for the fusion of global and local side information using Support Vector Machine. The global side information is generated at the decoder using global motion parameters estimated at the encoder using the Scale-Invariant Feature Transform. Experimental results show that the proposed approach can achieve a PSNR improvement of up to 1.7 dB for a GOP size of 2 and up to 3.78 dB for larger GOP sizes, with respect to the reference DISCOVER codec. © 2012 EURASIP.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Abou-Elailah, A.; Petrazzuoli, G.; Farah, J.; Cagnazzo, M.; Pesquet-Popescu, B.; Dufaux, F.
Side Information improvement in transform-domain distributed video coding Conference
Proceedings of SPIE – The International Society for Optical Engineering, vol. 8499, 2012.
@conference{11577_3471539,
title = {Side Information improvement in transform-domain distributed video coding},
author = { A. Abou-Elailah and G. Petrazzuoli and J. Farah and M. Cagnazzo and B. Pesquet-Popescu and F. Dufaux},
doi = {10.1117/12.930620},
year = {2012},
date = {2012-01-01},
booktitle = {Proceedings of SPIE - The International Society for Optical Engineering},
volume = {8499},
abstract = {Side Information (SI) has a strong impact on the rate-distortion performance in distributed video coding. The quality of the SI can be impaired when the temporal distance between the neighboring reference frames increases. In this paper, we introduce two novel methods that allow improving the quality of the SI. In the first approach, we propose a new estimation method for the initial SI using backward and forward motion estimation. The second one consists in re-estimating the SI after decoding all WZFs within the current Group of Pictures (GOP). For this purpose, the SI is first successively refined after each decoded DCT band. Then, after decoding all WZFs within the GOP, we adapt the search area to the motion content. Finally, each already decoded WZF is used, along with the neighboring ones, to estimate a new SI closer to the original WZF. This new SI is then used to reconstruct again the WZF with better quality. The experimental results show that, compared to the DISCOVER codec, the proposed method reaches an improvement of up to 3.53 dB in rate-distortion performance (measured with the Bjontegaard metric) for a GOP size of 8. © 2012 SPIE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Greco, C.; Nemoianu, I. Delia; Cagnazzo, M.; Pesquet-Popescu, B.
A network coding scheduling for Multiple Description video streaming over wireless networks Conference
European Signal Processing Conference, 2012.
@conference{11577_3469697,
title = {A network coding scheduling for Multiple Description video streaming over wireless networks},
author = { C. Greco and I. Delia Nemoianu and M. Cagnazzo and B. Pesquet-Popescu},
year = {2012},
date = {2012-01-01},
booktitle = {European Signal Processing Conference},
pages = {1915--1919},
abstract = {Network coding (NC) can achieve the maximum information flow in a network by allowing nodes to combine received packets before retransmission. Several works have shown NC to be beneficial in wireless networks, but the delay introduced by buffering before decoding raises a problem in real-time streaming applications. Here, we propose a framework for video delivery in wireless networks that combines Expanding Window Network Coding (EWNC), Multiple Description Coding (MDC), and a novel Rate-Distortion Optimised (RDO) scheduling algorithm that optimises the order in which the video packets are included in the coding window at the current sending opportunity. Results show that our approach consistently outperforms both EWNC applied on single description coding (SDC) and EWNC applied to MDC when a trivial scheduling is used. © 2012 EURASIP.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Mora, E. G.; Greco, C.; Pesquet-Popescu, B.; Cagnazzo, M.; Farah, J.
Cedar: An optimized network-aware solution for P2P video multicast Conference
2012 19th International Conference on Telecommunications, ICT 2012, 2012, ISBN: 978-1-4673-0746-8.
@conference{11577_3471542,
title = {Cedar: An optimized network-aware solution for P2P video multicast},
author = { E. G. Mora and C. Greco and B. Pesquet-Popescu and M. Cagnazzo and J. Farah},
doi = {10.1109/ICTEL.2012.6221262},
isbn = {978-1-4673-0746-8},
year = {2012},
date = {2012-01-01},
booktitle = {2012 19th International Conference on Telecommunications, ICT 2012},
pages = {1--6},
abstract = {Video multicast is an important application for P2P networks, which can benefit from the load repartitioning and large scale distribution properties they offer. In this context, the Orchard algorithm was developed in 2007, with an aim to build a P2P system meeting video multicast requirements. However, Orchard suffers from a slow overlay construction and a potentially high end-to-end delay. In this work, we propose a new algorithm that we call Cedar. Based on Orchard, it integrates original functionalities to speed-up the tree construction and builds a topologically-aware overlay network in order to minimize end-to-end delay. Simulation results showed that Cedar fulfilled its requirements, as the tree construction speed was increased by at least 64%, and the end-to-end delays reduced from around 20 seconds to a maximum of 5 seconds. The Network Awareness functionality also reduced end-to-end delays by 33%. © 2012 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Valenzise, G.; Cheung, G.; Galvao, R.; Cagnazzo, M.; Pesquet-Popescu, B.; Ortega, A.
Motion prediction of depth video for depth-image-based rendering using don’t care regions Conference
2012 Picture Coding Symposium, PCS 2012, Proceedings, 2012, ISBN: 978-1-4577-2048-2.
@conference{11577_3469694,
title = {Motion prediction of depth video for depth-image-based rendering using don't care regions},
author = { G. Valenzise and G. Cheung and R. Galvao and M. Cagnazzo and B. Pesquet-Popescu and A. Ortega},
doi = {10.1109/PCS.2012.6213294},
isbn = {978-1-4577-2048-2},
year = {2012},
date = {2012-01-01},
booktitle = {2012 Picture Coding Symposium, PCS 2012, Proceedings},
pages = {93--96},
abstract = {To enable synthesis of any desired intermediate view between two captured views at decoder via depth-image-based rendering (DIBR), both texture and depth maps from the captured viewpoints must be encoded and transmitted in a format known as texture-plus-depth. In this paper, we focus on the compression of depth maps across time to lower the overall bitrate in texture-plus-depth format. We observe that depth maps are not directly viewed, but are only used to provide geometric information of the captured scene for view synthesis at decoder. Thus, as long as the resulting geometric error does not lead to unacceptable synthesized view quality, each depth pixel only needs to be reconstructed at the decoder coarsely within a tolerable range. We first formalize the notion of tolerable range per depth pixel as don't care region (DCR), by studying the synthesized view distortion sensitivity to the pixel value a sensitive depth pixel will have a narrow DCR, and vice versa. Given per-pixel DCRs, we then modify inter-prediction modes during motion prediction to search for a predictor block matching per-pixel DCRs in a target block (rather than the fixed ground truth depth signal in a target block), in order to lower the energy of the prediction residual for the block. We implemented our DCR-based motion prediction scheme inside H.264; our encoded bitstreams remain 100% standard compliant. We show experimentally that our proposed encoding scheme can reduce the bitrate of depth maps coded with baseline H.264 by over 28%. © 2012 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Greco, C.; Cagnazzo, M.; Pesquet-Popescu, B.
Low-latency video streaming with congestion control in mobile Ad-Hoc networks Journal Article
In: IEEE TRANSACTIONS ON MULTIMEDIA, vol. 14, no. 4, pp. 1337–1350, 2012.
@article{11577_3469518,
title = {Low-latency video streaming with congestion control in mobile Ad-Hoc networks},
author = { C. Greco and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/TMM.2012.2195480},
year = {2012},
date = {2012-01-01},
journal = {IEEE TRANSACTIONS ON MULTIMEDIA},
volume = {14},
number = {4},
pages = {1337--1350},
abstract = {In this paper, we address the challenge of delivering a video stream, encoded with multiple descriptions, in a mobile ad-hoc environment with low-latency constraints. This kind of application is meant to provide an efficient and reliable video communication tool in scenarios where the deployment of an infrastructure is not feasible, such as military and disaster relief applications. First, we present a recently proposed protocol that employs a reliable form of one-hop broadcast to build an efficient overlay network according to a multi-objective function that minimizes the number of packets injected in the network and maximizes the path diversity among descriptions. Then, we introduce the main contribution of this paper: a cross-layer congestion control strategy where the MAC layer is video-coding aware and adjusts its transmission parameters (namely, the RTS retry limit) via congestion/distortion optimization. The main challenge in this approach is providing a reliable estimation of congestion and distortion, given the limited information available at each node. Our simulations show that, if a stringent constraint of low delay is imposed, our technique grants a consistent gain in terms of both PSNR and delay reduction, for bitrates up to a few megabits per second. © 2012 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Nemoianu, I. Delia; Greco, C.; Cagnazzo, M.; Pesquet-Popescu, B.
2012 IEEE Visual Communications and Image Processing, VCIP 2012, 2012, ISBN: 978-1-4673-4406-7.
@conference{11577_3469879,
title = {Multi-view video streaming over wireless networks with RD-optimized scheduling of network coded packets},
author = { I. Delia Nemoianu and C. Greco and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/VCIP.2012.6410748},
isbn = {978-1-4673-4406-7},
year = {2012},
date = {2012-01-01},
booktitle = {2012 IEEE Visual Communications and Image Processing, VCIP 2012},
pages = {1--6},
abstract = {Multi-view video streaming is an emerging video paradigm that enables new interactive services, such as free viewpoint television and immersive teleconferencing. However, it comes with a high bandwidth cost, as the equivalent of many single-view streams has to be transmitted. Network coding (NC) can improve the performance of the network by allowing nodes to combine received packets before retransmission. Several works have shown NC to be beneficiai in wireless networks, but the delay introduced by buffering before decoding raises a problem in real-time streaming applications. Here, we propose to use Expanding Window NC (EWNC) for multi-view streaming to allow immediate decoding of the received packets. The order in which the packets are included in the coding window is chosen via RD-optimization for the current sending opportunity. Results show that our approach consistently outperforms both classical NC applied on each view independently and transmission without NC. © 2012 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Nemoianu, I. D.; Greco, C.; Cagnazzo, M.; Pesquet-Popescu, B.
A framework for joint multiple description coding and network coding over wireless ad-hoc networks Conference
ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing – Proceedings, 2012, ISBN: 978-1-4673-0044-5.
@conference{11577_3471540,
title = {A framework for joint multiple description coding and network coding over wireless ad-hoc networks},
author = { I. D. Nemoianu and C. Greco and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/ICASSP.2012.6288376},
isbn = {978-1-4673-0044-5},
year = {2012},
date = {2012-01-01},
booktitle = {ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings},
pages = {2309--2312},
abstract = {Network coding (NC) can achieve the maximum information flow in the network by allowing nodes to combine received packets before retransmission. Several papers have shown NC to be beneficial in mobile ad-hoc networks, but the delay introduced by buffered decoding raises a problem in real-time streaming applications. Here we propose to use NC jointly with multiple description coding (MDC) to allow instant decoding of the received packets. The optimal encoding coefficients are chosen via distributed optimisation of the expected video quality. Nodes receive up-to-date information about the network topology through a recently proposed protocol, originally designed for real-time streaming of MDC video. Results show that, due to the limitations imposed by instant decoding to the coding window size, our approach consistently outperforms the popular technique of random linear network coding. © 2012 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Abou-Elailah, A.; Dufaux, F.; Cagnazzo, M.; Pesquet-Popescu, B.; Farah, J.
2012 19th International Conference on Telecommunications, ICT 2012, 2012, ISBN: 978-1-4673-0746-8.
@conference{11577_3471474,
title = {Successive refinement of side information using adaptive search area for long duration GOPs in distributed video coding},
author = { A. Abou-Elailah and F. Dufaux and M. Cagnazzo and B. Pesquet-Popescu and J. Farah},
doi = {10.1109/ICTEL.2012.6221275},
isbn = {978-1-4673-0746-8},
year = {2012},
date = {2012-01-01},
booktitle = {2012 19th International Conference on Telecommunications, ICT 2012},
pages = {1--6},
abstract = {In distributed video coding, the reference frames are used to generate a side information at the decoder in order to decode the Wyner-Ziv frame. The quality of side information has a strong impact on the coding efficiency of distributed video coding. The estimation of the side information becomes less accurate when the temporal distance between the neighboring reference frames increases or when the sequence contains fast motion. In this paper, we propose a new method based on successive refinement of the side information by adapting the motion search area after decoding the first DCT band. More specifically, different search areas are initially set, according to the temporal distance between the neighboring reference frames. Furthermore, the size of the search area is adapted to the motion content after decoding the first DCT band by using the partially decoded Wyner-Ziv frame. This adapted search area is used in order to refine the side information after decoding each remaining DCT band. The experimental results show that the proposed technique allows an improvement in rate distortion performance that can reach 0.7 dB for a GOP size of 8, compared to the method where a constant search area is used, and a significant gain up to 3.23 dB, with respect to DISCOVER codec. The improvement in the quality of the final side information reaches 5.6 dB for some frames, compared to constant search area. Moreover, the proposed method reduces significantly the decoding time for all test sequences. © 2012 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
2011
Petrazzuoli, G.; Cagnazzo, M.; Dufaux, F.; Pesquet-Popescu, B.
Wyner-Ziv coding for depth maps in multiview video-plus-depth Conference
Proceedings – International Conference on Image Processing, ICIP, 2011, ISBN: 978-1-4577-1302-6.
@conference{11577_3469686,
title = {Wyner-Ziv coding for depth maps in multiview video-plus-depth},
author = { G. Petrazzuoli and M. Cagnazzo and F. Dufaux and B. Pesquet-Popescu},
doi = {10.1109/ICIP.2011.6115817},
isbn = {978-1-4577-1302-6},
year = {2011},
date = {2011-01-01},
booktitle = {Proceedings - International Conference on Image Processing, ICIP},
pages = {1817--1820},
abstract = {Three dimensional digital video services are gathering a lot of attention in recent years, thanks to the introduction of new and efficient acquisition and rendering devices. In particular, 3D video is often represented by a single view and a so called depth map, which gives information about the distance between the point of view and the objects. This representation can be extended to multiple views, each with its own depth map. Efficient compression of this kind of data is of course a very important topic in sight of a massive deployment of services such as 3D-TV and FTV (free viewpoint TV). In this paper we consider the application of distributed coding techniques to the coding of depth maps, in order to reduce the complexity of single view or multi view encoders and to enhance interactive multiview video streaming. We start from state-of-the-art distributed video coding techniques and we improve them by using high order motion interpolation and by exploiting texture motion information to encode the depth maps. The experiments reported here show that the proposed method achieves a rate reduction up to 11.06% compared to state-of-the-art distributed video coding technique. © 2011 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Agostini-Vautard, M. A.; Cagnazzo, M.; Antonini, M.; Laroche, G.; Jung, J.
A new coding mode for hybrid video coders based on quantized motion vectors Journal Article
In: IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, vol. 21, no. 7, pp. 946–956, 2011.
@article{11577_3469379,
title = {A new coding mode for hybrid video coders based on quantized motion vectors},
author = { M. A. Agostini-Vautard and M. Cagnazzo and M. Antonini and G. Laroche and J. Jung},
doi = {10.1109/TCSVT.2011.2133590},
year = {2011},
date = {2011-01-01},
journal = {IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY},
volume = {21},
number = {7},
pages = {946--956},
abstract = {The rate allocation tradeoff between motion vectors and transform coefficients has a major importance when it comes to efficient video compression. This paper introduces a new coding mode for an H.264/AVC-like video coder, which improves the management of this resource allocation. The proposed technique can be used within any hybrid video encoder allowing a different coding mode for any macroblock. The key tool of the new mode is the lossy coding of motion vectors, obtained via quantization: while the transformed motion-compensated residual is computed with a high-precision motion vector, the motion vector itself is quantized before being sent to the decoder, in a rate/distortion optimized way. Several problems have to be faced with in order to get an efficient implementation of the coding mode, especially the coding and prediction of the quantized motion vectors, and the selection and encoding of the quantization steps. This new coding mode improves the performance of the hybrid video encoder over several sequences at different resolutions. © 2011 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Abou-Elailah, A.; Dufaux, F.; Farah, J.; Cagnazzo, M.; Pesquet-Popescu, B.
Successive refinement of motion compensated interpolation for transform-domain distributed video coding Conference
European Signal Processing Conference, 2011.
@conference{11577_3469811,
title = {Successive refinement of motion compensated interpolation for transform-domain distributed video coding},
author = { A. Abou-Elailah and F. Dufaux and J. Farah and M. Cagnazzo and B. Pesquet-Popescu},
year = {2011},
date = {2011-01-01},
booktitle = {European Signal Processing Conference},
pages = {11--15},
abstract = {In distributed video coding, the estimation of the side information at the decoder plays a key role in the final ratedistortion performance of the codec. The side information is commonly generated by motion-compensated temporal interpolation of the neighboring reference frames. In this paper, we propose a successive refinement after the decoding of each DCT subband to improve the accuracy of motion compensation between reference frames, in order to obtain a new side information estimation closer to the originalWyner-Ziv frame. The experimental results show that the proposed scheme can achieve up to 0.9 dB of improvement in ratedistortion performance for a GOP size of 2 and 2.4 dB for a GOP size of 8 for sequences containing high motion with respect to state-of-the-art techniques. © EURASIP, 2011.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Cagnazzo, M.; Pesquet-Popescu, B.
Depth map coding by dense disparity estimation for MVD compression Conference
17th DSP 2011 International Conference on Digital Signal Processing, Proceedings, 2011, ISBN: 978-1-4577-0273-0.
@conference{11577_3469809,
title = {Depth map coding by dense disparity estimation for MVD compression},
author = { M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/ICDSP.2011.6005011},
isbn = {978-1-4577-0273-0},
year = {2011},
date = {2011-01-01},
booktitle = {17th DSP 2011 International Conference on Digital Signal Processing, Proceedings},
pages = {1--6},
abstract = {In multiview video representation, one of the most popular format is the so-called multiple view video plus depth. This representation is made up of N image sequences, each accompanied by a sequence of depth maps, telling the distance of each represented pixel from the observing camera. The depth maps are needed at the decoder side in order to generate intermediate views and therefore to enrich the user experience. This format is very flexible but also very demanding, in terms of storage space or and transmission bandwidth. Therefore, compression is needed. At this end, one of the key steps is an efficient representation of depth maps. In this work we build over a proposed method for multiple view video coding, based on dense disparity estimation between views. This allows us to obtain a compact and high-quality depthmap representation. In particular we explore the complex relationship between estimation and encoding parameters, showing that an optimal parameter set exist, that allows a fine-tuning of the estimation phase and an adaption of its results to the subsequent compression phase. Experiments are encouraging, showing remarkable gain over simple methods such as H.264/AVC simulcast, and even some gain with respect to more sophisticated techniques such as MVC. © 2011 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Petrazzuoli, G.; Cagnazzo, M.; Dufaux, F.; Pesquet-Popescu, B.
Proceedings – International Conference on Image Processing, ICIP, 2011, ISBN: 978-1-4577-1302-6.
@conference{11577_3469695,
title = {Using distributed source coding and depth image based rendering to improve interactive multiview video access},
author = { G. Petrazzuoli and M. Cagnazzo and F. Dufaux and B. Pesquet-Popescu},
doi = {10.1109/ICIP.2011.6116620},
isbn = {978-1-4577-1302-6},
year = {2011},
date = {2011-01-01},
booktitle = {Proceedings - International Conference on Image Processing, ICIP},
pages = {597--600},
abstract = {Multiple-views video is commonly believed to be the next significant achievement in video communications, since it enables new exciting interactive services such as free viewpoint television and immersive teleconferencing. However the interactivity requirement (i.e. allowing the user to change the viewpoint during video streaming) involves a trade-off between storage and bandwidth costs. Several solutions have been proposed in the literature, using redundant predictive frames, Wyner-Ziv frames, or a combination of them. In this paper, we adopt distributed video coding for interactive multiview video plus depth (MVD), taking advantage of depth image based rendering (DIBR) and depth-aided inpainting to fill the occlusion areas. To the authors' best knowledge, very few works in interactive MVD consider the problem of continuity of the playback during the switching among streams. Therefore we survey the existing solutions, we propose a set of techniques for MVD coding and we compare them. As main results, we observe that DIBR can help in rate reduction (up to 13.36% for the texture video and up to 8.67% for the depth map, wrt the case where DIBR is not used), and we also note that the optimal strategy to combine DIBR and distributed video coding depends on the position of the switching time into the group of pictures. Choosing the best technique on a frame-to-frame basis can further reduce the rate from 1% to 6%. © 2011 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Abou-Elailah, A.; Farah, J.; Cagnazzo, M.; Pesquet-Popescu, B.; Dufaux, F.
Improved side information generation for distributed video coding Conference
3rd European Workshop on Visual Information Processing, EUVIP 2011 – Final Program, 2011, ISBN: 978-1-4577-0072-9.
@conference{11577_3469687,
title = {Improved side information generation for distributed video coding},
author = { A. Abou-Elailah and J. Farah and M. Cagnazzo and B. Pesquet-Popescu and F. Dufaux},
doi = {10.1109/EuVIP.2011.6045553},
isbn = {978-1-4577-0072-9},
year = {2011},
date = {2011-01-01},
booktitle = {3rd European Workshop on Visual Information Processing, EUVIP 2011 - Final Program},
pages = {42--49},
abstract = {In distributed video coding, the side information is commonly generated by motion-compensated temporal interpolation of the neighboring reference frames at the decoder side. The side information quality has a strong impact on the final rate-distortion performance of the codec. In this paper, we propose a successive refinement after the decoding of each DCT subband to improve the accuracy of motion compensation between reference frames, in order to obtain a new side information estimation closer to the original Wyner-Ziv frame. Here, we propose two different algorithms in the refinement process of the side information. The experimental results show that the proposed techniques allow an improvement in rate-distortion performance that can reach 1.15 dB for GOP size of 2 and 3.2 dB for longer GOP size, with respect to state-of-the-art techniques, for sequences containing high motion. © 2011 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}
Greco, C.; Cagnazzo, M.
A cross-layer protocol for cooperative content delivery over mobile ad-hoc networks Journal Article
In: INTERNATIONAL JOURNAL OF COMMUNICATION NETWORKS AND DISTRIBUTED SYSTEMS, vol. 7, no. 1-2, pp. 49–63, 2011.
@article{11577_3469524,
title = {A cross-layer protocol for cooperative content delivery over mobile ad-hoc networks},
author = { C. Greco and M. Cagnazzo},
doi = {10.1504/IJCNDS.2011.040977},
year = {2011},
date = {2011-01-01},
journal = {INTERNATIONAL JOURNAL OF COMMUNICATION NETWORKS AND DISTRIBUTED SYSTEMS},
volume = {7},
number = {1-2},
pages = {49--63},
publisher = {Inderscience Publishers},
abstract = {Real-time multimedia streaming over MANETs is an active research field since this technology promises scalable and robust audio/video delivery without infrastructure. Even though this problem has several commune features with peer-to-peer routing, the additional sensible parameters of MANETs make wired solutions unfit to this case; therefore, we propose a content routing/delivery protocol inherently designed for the ad-hoc wireless case, exploiting the intrinsic broadcast property of the medium. We provided an implementation of this protocol and we tested it in several use-cases, observing how it assures availability, robustness and scalability. Copyright © 2011 Inderscience Enterprises Ltd.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Greco, C.; Petrazzuoli, G.; Cagnazzo, M.; Pesquet-Popescu, B.
An MDC-based video streaming architecture for mobile networks Conference
MMSP 2011 – IEEE International Workshop on Multimedia Signal Processing, 2011, ISBN: 978-1-4577-1433-7.
@conference{11577_3469685,
title = {An MDC-based video streaming architecture for mobile networks},
author = { C. Greco and G. Petrazzuoli and M. Cagnazzo and B. Pesquet-Popescu},
doi = {10.1109/MMSP.2011.6093829},
isbn = {978-1-4577-1433-7},
year = {2011},
date = {2011-01-01},
booktitle = {MMSP 2011 - IEEE International Workshop on Multimedia Signal Processing},
pages = {1--6},
abstract = {Multiple description coding (MDC) is a framework designed to improve the robustness of video content transmission in lossy environments. In this work, we propose an MDC technique using a legacy coder to produce two descriptions, based on separation of even and odd frames. If only one description is received, the missing frames are reconstructed using temporal high-order motion interpolation (HOMI), a technique originally proposed for distributed video coding. If both descriptions are received, the frames are reconstructed as a block-wise linear combination of the two descriptions, with the coefficient computed at the encoder in a RD-optimised fashion, encoded with a context-adaptive arithmetic coder, and sent as side information. We integrated the proposed technique in a mobile ad-hoc streaming protocol, and tested it using a group mobility model. The results show a non-negligible gain for the expected video quality, with respect to the reference technique. © 2011 IEEE.},
keywords = {},
pubstate = {published},
tppubtype = {conference}
}