TY - JOUR
T1 - Stream encoder identification in green video context
AU - Allouche, Mohamed
AU - Cole, Elliot
AU - Zoughebi, Mateo
AU - Trias, Carl De Sousa
AU - Mitrea, Mihai
N1 - Publisher Copyright:
© 2025 Society for Imaging Science and Technology.
PY - 2025/1/1
Y1 - 2025/1/1
N2 - Video streaming hits more than 80% of the carbon emissions generated by worldwide digital technologies consumption that, in their turn, account for 5% of worldwide carbon emissions. Hence, green video encoding emerges as a research field devoted to reducing the size of the video streams and the complexity of the decoding/encoding operations, while keeping a preestablished visual quality. Having the specific view of tracking green encoded video streams, the present paper studies the possibility of identifying the last video encoder considered in the case of multiple reencoding distribution scenarios. To this end, classification solutions backboned by the VGG, ResNet and MobileNet families are considered to discriminate among MPEG-4 AVC stream syntax elements, such as luma/chroma coefficients or intra prediction modes. The video content sums-up to 2 hours and is structured in two databases. Three encoders are alternatively studied, namely a proprietary green-encoder solution, and the two by-default encoders available on a large video sharing platform and on a popular social media, respectively. The quantitative results show classification accuracy ranging between 75% to 100%, according to the specific architecture, sub-set of classified elements, and dataset.
AB - Video streaming hits more than 80% of the carbon emissions generated by worldwide digital technologies consumption that, in their turn, account for 5% of worldwide carbon emissions. Hence, green video encoding emerges as a research field devoted to reducing the size of the video streams and the complexity of the decoding/encoding operations, while keeping a preestablished visual quality. Having the specific view of tracking green encoded video streams, the present paper studies the possibility of identifying the last video encoder considered in the case of multiple reencoding distribution scenarios. To this end, classification solutions backboned by the VGG, ResNet and MobileNet families are considered to discriminate among MPEG-4 AVC stream syntax elements, such as luma/chroma coefficients or intra prediction modes. The video content sums-up to 2 hours and is structured in two databases. Three encoders are alternatively studied, namely a proprietary green-encoder solution, and the two by-default encoders available on a large video sharing platform and on a popular social media, respectively. The quantitative results show classification accuracy ranging between 75% to 100%, according to the specific architecture, sub-set of classified elements, and dataset.
UR - https://www.scopus.com/pages/publications/105000822724
U2 - 10.2352/EI.2025.37.10.IPAS-234
DO - 10.2352/EI.2025.37.10.IPAS-234
M3 - Conference article
AN - SCOPUS:105000822724
SN - 2470-1173
VL - 37
JO - IS and T International Symposium on Electronic Imaging Science and Technology
JF - IS and T International Symposium on Electronic Imaging Science and Technology
IS - 10
M1 - IPAS-234
T2 - IS and T International Symposium on Electronic Imaging 2025: 23rd Image Processing: Algorithms and Systems, IPAS 2025
Y2 - 2 February 2025 through 6 February 2025
ER -