(19)
(11)EP 3 513 326 B1

(12)EUROPEAN PATENT SPECIFICATION

(45)Mention of the grant of the patent:
02.11.2022 Bulletin 2022/44

(21)Application number: 17781339.1

(22)Date of filing:  28.09.2017
(51)International Patent Classification (IPC): 
G06F 16/783(2019.01)
H04N 13/194(2018.01)
H04N 21/234(2011.01)
H04N 21/81(2011.01)
G06F 21/10(2013.01)
H04N 13/161(2018.01)
H04N 21/231(2011.01)
H04N 21/2743(2011.01)
G06V 20/40(2022.01)
H04N 21/835(2011.01)
(52)Cooperative Patent Classification (CPC):
G06F 21/10; H04N 21/835; H04N 13/194; H04N 21/23109; H04N 21/23418; H04N 21/2743; H04N 21/816; G06F 16/783; H04N 13/161; G06V 20/46
(86)International application number:
PCT/US2017/054075
(87)International publication number:
WO 2018/111379 (21.06.2018 Gazette  2018/25)

(54)

METHODS, SYSTEMS, AND MEDIA FOR DETECTING STEREOSCOPIC VIDEOS BY GENERATING FINGERPRINTS FOR MULTIPLE PORTIONS OF A VIDEO FRAME

VERFAHREN, SYSTEME UND MEDIEN ZUR ERKENNUNG VON STEREOSKOPISCHEN VIDEOS DURCH ERZEUGUNG VON FINGERABDRÜCKEN FÜR MEHRERE TEILE EINES VIDEOFRAMES

PROCÉDÉS, SYSTÈMES ET SUPPORTS POUR DÉTECTER DES VIDÉOS STÉRÉOSCOPIQUES PAR GÉNÉRATION D'EMPREINTES DIGITALES POUR DE MULTIPLES PARTIES D'UNE TRAME VIDÉO


(84)Designated Contracting States:
AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

(30)Priority: 16.12.2016 US 201615382113

(43)Date of publication of application:
24.07.2019 Bulletin 2019/30

(73)Proprietor: Google LLC
Mountain View, CA 94043 (US)

(72)Inventors:
  • ZAMARAIEV, Valerii
    8002 Zurich (CH)
  • PAVETIC, Filip
    8002 Zurich (CH)

(74)Representative: Watkin, Timothy Lawrence Harvey 
Marks & Clerk LLP Fletcher House The Oxford Science Park Heatley Road
Oxford OX4 4GE
Oxford OX4 4GE (GB)


(56)References cited: : 
US-A1- 2009 165 031
US-B1- 8 340 449
US-A1- 2015 110 340
  
  • CHUNHUA WU ET AL: "Differential watermarking scheme of stereo video", COMMUNICATION TECHNOLOGY (ICCT), 2012 IEEE 14TH INTERNATIONAL CONFERENCE ON, IEEE, 9 November 2012 (2012-11-09), pages 744-748, XP032397987, DOI: 10.1109/ICCT.2012.6511303 ISBN: 978-1-4673-2100-6
  • BITAGHSIR SEPIDEH AKHAVAN ET AL: "Stereo image watermarking method based on binocular just noticeable difference", 2014 11TH INTERNATIONAL ISC CONFERENCE ON INFORMATION SECURITY AND CRYPTOLOGY, IEEE, 3 September 2014 (2014-09-03), pages 33-38, XP032712227, DOI: 10.1109/ISCISC.2014.6994018 [retrieved on 2014-12-18]
  • ZHANG YANA ET AL: "Binocular suppression based visual masking model for stereo image watermarking", 2015 IEEE 16TH INTERNATIONAL CONFERENCE ON COMMUNICATION TECHNOLOGY (ICCT), IEEE, 18 October 2015 (2015-10-18), pages 32-36, XP032860438, DOI: 10.1109/ICCT.2015.7399788 ISBN: 978-1-4673-7004-2 [retrieved on 2016-02-04]
  
Note: Within nine months from the publication of the mention of the grant of the European patent, any person may give notice to the European Patent Office of opposition to the European patent granted. Notice of opposition shall be filed in a written reasoned statement. It shall not be deemed to have been filed until the opposition fee has been paid. (Art. 99(1) European Patent Convention).


Description

Technical Field



[0001] The disclosed subject matter relates to methods, systems, and media for identifying content in stereoscopic videos, for example, for detecting where content may have been the subject of unauthorized distribution or compromised digital rights management.

Background



[0002] Video content providers can receive uploaded video content, store the uploaded video content, and then provide the uploaded video content to many users, for example, by streaming the video content to multiple user devices. These video content providers may determine whether the uploaded video content matches video content in a reference database, for example, of copyrighted content. With the advent of immersive or 360-degree spherical video content, these video content providers have started to receive uploaded video content containing three-dimensional video content. However, it can be difficult to determine whether such three-dimensional video content matches video content in a reference database.

[0003] Accordingly, it is desirable to provide methods, systems, and media for identifying content in stereoscopic videos and, more particularly, for detecting "abusive" stereoscopic videos (for example, where content may have been the subject of unauthorized distribution or compromised digital rights management) by generating fingerprints for multiple portions of a video frame. Chunhua wu et al "Differential watermarking scheme of stereo video" Communication Technology, International conference on, pp 744-748, 9 November 2012 discloses a watermarking method for stereo video. US 2015/110340 discloses a method of fingerprinting based on luminance and colour. US 2009/165031 disclose a method of identifying video.

Summary



[0004] The invention is defined by the independent claims. The dependent claims define advantageous embodiments. Embodiments described herein that are outside of the claims are to be considered comparative examples.

Brief Description of the Drawings



[0005] Various objects, features, and advantages of the disclosed subject matter can be more fully appreciated with reference to the following detailed description of the disclosed subject matter when considered in connection with the following drawings, in which like reference numerals identify like elements.

FIG. 1 shows an illustrative example of a user interface for receiving a video content upload from a content creator in accordance with some embodiments of the disclosed subject matter.

FIGS. 2A and 2B show illustrative examples of schematic diagrams for generating fingerprints associated with a whole frame and two encoded portions of a frame of a video content item in accordance with some embodiments of the disclosed subject matter.

FIG. 3 shows a schematic diagram of an illustrative system suitable for implementation of mechanisms described herein for detecting abusive stereoscopic videos by generating fingerprints for multiple portions of a video frame in accordance with some embodiments of the disclosed subject matter.

FIG. 4 shows a detailed example of hardware that can be used in a server and/or a user device of FIG. 3 in accordance with some embodiments of the disclosed subject matter.

FIG. 5 shows an illustrative example of a process for detecting abusive stereoscopic videos by generating fingerprints for multiple portions of a video frame in accordance with some embodiments of the disclosed subject matter.


Detailed Description



[0006] In accordance with various embodiments, mechanisms (which can include methods, systems, and media) for identifying content in stereoscopic videos and, more particularly, for detecting abusive stereoscopic videos by generating fingerprints for multiple portions of a video frame are provided.

[0007] In some embodiments, the mechanisms described herein can be used to detect particular types of video content in an uploaded video content item. For example, in some embodiments, the mechanisms can be used to detect potentially copyrighted video content or potentially copyrighted audio content within the uploaded video content item. It should be noted that various fingerprinting techniques have been used to identify copyrighted content, for example, by matching a portion of a content item to a reference content item in a database of reference content items. However, these fingerprinting techniques are sometimes unable to identify copyrighted video content when an uploaded video content item is three-dimensional video content. For example, in instances where the uploaded video content item includes three-dimensional content that is encoded as stereoscopic video content (e.g., with video content to be delivered to a left and right eye of a user encoded as a left and right half of a frame, and/or encoded in any other suitable manner), the fingerprinting techniques may be unable to detect a match to video content in the reference database. In particular, in instances where the fingerprinting techniques are not aware that the uploaded video content item includes stereoscopic video content (e.g., based on a user who uploads the video content item not indicating that the uploaded content includes stereoscopic video content, and/or based on any other suitable information), the fingerprinting techniques may fingerprint an entire frame of the video content item, and may therefore not detect a match to reference video content items. Conversely, in instances where the uploaded video content item includes two-dimensional video content but is tagged or otherwise indicated as being three-dimensional video content (e.g., by a user uploading the video content item, and/or in any other suitable manner), the fingerprinting techniques may try to match half of a frame of the video content item with reference video content items, and may therefore not detect matching video content in the reference database.

[0008] In some embodiments, the mechanisms described herein can generate fingerprints for a whole frame of the uploaded video content item, a first encoded portion of the frame for the uploaded video content item, and a second encoded portion of the frame for the uploaded video content item. For example, in some embodiments, the first encoded portion of the frame can correspond to a first half of the frame that is to be delivered to a left eye of a viewer if the video is delivered as a stereoscopically encoded three-dimensional video content item, and the second encoded portion of the frame can correspond to a second half of the frame that is to be delivered to a right eye of the viewer if the video is delivered as a stereoscopically encoded three-dimensional video content item. As a more particular example, in some embodiments, the first half of the frame and the second half of the frame can correspond to a left half of the frame and a right half of the frame, respectively. Additionally or alternatively, in some embodiments, the first half of the frame and the second half of the frame can be any suitable arrangement of stereoscopic video content, such as a top half and a bottom half of a frame, alternating pixels of a frame (e.g., in a checkerboard pattern), alternating frames of the video content item, and/or any other suitable arrangement.

[0009] In some embodiments, the mechanisms can then compare the fingerprints generated from the whole frame and the fingerprints generated from each encoded portion of the frame with fingerprints of video content items stored in a reference database, and can detect whether any of the generated fingerprints match the reference fingerprints. It should be noted that, in some embodiments, the mechanisms can therefore detect whether the uploaded video content item matches reference video content irrespective of whether the uploaded video content item includes two-dimensional video content or three-dimensional video content. It should also be noted that, in some embodiments, the mechanisms can therefore detect whether the uploaded video content item matches reference video content irrespective of whether the uploaded video content item that was uploaded by a user has correctly indicated or designated the video content item as being a two-dimensional video content item or a three-dimensional video content item.

[0010] In some embodiments, if the mechanisms described herein detect a match between the uploaded video content item and a reference video content item (e.g., indicating that the uploaded video content item is potentially copyrighted video content), the mechanisms can cause an indication of the match to be presented on a user device that transmitted and/or uploaded the video content item. For example, in some embodiments, a message can be presented on the user device indicating that the uploaded content matches reference video content. Additionally or alternatively, in some embodiments, the mechanisms can block storage of the uploaded video content item in response to detecting a match to a reference video content item. Furthermore, in some embodiments, the mechanisms can cause an uploaded video content item that has matched reference video content to be flagged for manual review.

[0011] Note that, in some embodiments, the uploaded video content item can be any suitable type of video content, including videos, movies, television programs, live-streamed content, and/or any other suitable video content. For example, in an instance where the uploaded video content item is live-streamed video content, a user device can transmit live video content to a server, which can host the received video content and cause the video content to be streamed to other user devices, for example, in response to a request to view the video content. In instances where the uploaded video content item is live-streamed video content, the mechanisms described herein can generate fingerprints for portions of the live-steamed content as it is received. For example, in some embodiments, where the live-steamed video content item is being transmitted from a user device to a server that is hosting the video content, the mechanisms can generate fingerprints for portions of the received video content of any suitable duration (e.g., one second, two seconds, five seconds, one minute, and/or any other suitable duration) as they are received by the server. In instances where a match of live-streamed content to reference video content in a reference database, the mechanisms described herein can present an indication that a match of the live-streamed content to reference video content in the reference database on the user device transmitting the live-streamed content to the server. For example, in some embodiments, the mechanisms can cause a message to be presented on the user device indicating that the live-streamed content has been determined to match a particular video content item in the reference database and can, in some embodiments, identify a name of the particular video content item in the reference database. Additionally or alternatively, in some embodiments, the mechanisms can block further streaming of the live-streamed video content in response to detecting a match to reference video content.

[0012] Turning to FIG. 1, an example 100 of a user interface for receiving an uploaded video content item and information about the video content item to be uploaded to a content provider is shown in accordance with some embodiments of the disclosed subject matter. As illustrated, user interface 100 can include a content item selection interface 102 and a content item information interface 104.

[0013] In some embodiments, content item selection interface 102 can include any suitable user interface controls for allowing a user of user interface 100 to select a content item for upload. For example, as shown in FIG. 1, in some embodiments, content item selection interface 102 can include a push button that, when selected, can cause a directory listing of a user device associated with user interface 100 to be presented. In some embodiments, a selected content item can then be indicated in a text box. In some embodiments, content item selection interface 102 can be omitted.

[0014] In some embodiments, content item information interface 104 can be used to allow a content creator uploading a particular video content item to indicate any suitable information about the video content item. For example, as shown in FIG. 1, in some embodiments, content item information interface 104 can include user interface controls to allow the content creator to indicate whether or not the uploaded video content item is a stereoscopic video content item. In some embodiments, any other suitable information can be indicated via content item information interface 104. For example, in some embodiments, the information can include a name of the video content item, a name of a creator of the video content item, credit information or collaboration information associated with the video content item, keywords associated with the video content item, access control or privacy settings associated with the video content item, and/or any other suitable information.

[0015] Turning to FIG. 3, an example of an illustrative system 300 suitable for implementation of mechanisms described herein for identifying content in stereoscopic videos and, more particularly, for detecting abusive stereoscopic videos by generating fingerprints for multiple portions of a video frame is shown in accordance with some embodiments of the disclosed subject matter is shown. As illustrated, hardware 300 can include one or more servers, such as a server 302, a communication network 304, and/or one or more user devices 306, such as user devices 308 and 310.

[0016] In some embodiments, server(s) 302 can be any suitable server(s) for receiving an upload of a video content item, determining whether the uploaded video content item matches a reference video content item, and/or performing any other suitable functions. For example, as described below in connection with FIG. 5, server(s) 302 can perform any suitable fingerprinting technique(s) to determine whether an uploaded video content item matches a video content item in a reference database. As a more particular example, in some embodiments, server(s) 302 can generate fingerprints of an entire frame of the video content item as well as a first half and a second half of the frame to determine whether the video content item matches a reference video content item, regardless of whether the uploaded video content item is a stereoscopic video content item. In some embodiments, server(s) 302 can be omitted.

[0017] Communication network 304 can be any suitable combination of one or more wired and/or wireless networks in some embodiments. For example, communication network 306 can include any one or more of the Internet, an intranet, a wide-area network (WAN), a local-area network (LAN), a wireless network, a digital subscriber line (DSL) network, a frame relay network, an asynchronous transfer mode (ATM) network, a virtual private network (VPN), and/or any other suitable communication network. User devices 306 can be connected by one or more communications links 312 and/or 314 to communication network 304 that can be linked via one or more communications links (e.g., communications link 316) to server(s) 302. Communications links 312, 314, and/or 316 can be any communications links suitable for communicating data among user devices 306 and server(s) 302 such as network links, dial-up links, wireless links, hard-wired links, any other suitable communications links, or any suitable combination of such links.

[0018] In some embodiments, user devices 306 can include one or more computing devices suitable for transmitting a video content item to server(s) 302, transmitting information related to the video content item to server(s) 302, and/or any other suitable functions. For example, in some embodiments, user devices 306 can be implemented as a mobile device, such as a smartphone, mobile phone, a tablet computer, a laptop computer, a vehicle (e.g., a car, a boat, an airplane, or any other suitable vehicle) entertainment system, a portable media player, and/or any other suitable mobile device. As another example, in some embodiments, user devices 306 can be implemented as a non-mobile device such as a desktop computer, a set-top box, a television, a streaming media player, a game console, and/or any other suitable non-mobile device.

[0019] Although server 302 is illustrated as a single device, the functions performed by server 302 can be performed using any suitable number of devices in some embodiments. For example, in some embodiments, multiple devices can be used to implement the functions performed by server 302.

[0020] Although two user devices 308 and 310 are shown in FIG. 3, any suitable number of user devices, and/or any suitable types of user devices, can be used in some embodiments.

[0021] Server(s) 302 and user devices 306 can be implemented using any suitable hardware in some embodiments. For example, in some embodiments, devices 302 and 306 can be implemented using any suitable general purpose computer or special purpose computer. For example, a server may be implemented using a special purpose computer. Any such general purpose computer or special purpose computer can include any suitable hardware. For example, as illustrated in example hardware 400 of FIG. 4, such hardware can include hardware processor 402, memory and/or storage 404, an input device controller 406, an input device 408, display/audio drivers 410, display and audio output circuitry 412, message interface(s) 414, an antenna 416, and a bus 418.

[0022] Hardware processor 402 can include any suitable hardware processor, such as a microprocessor, a micro-controller, digital signal processor(s), dedicated logic, and/or any other suitable circuitry for controlling the functioning of a general purpose computer or a special purpose computer in some embodiments. In some embodiments, hardware processor 402 can be controlled by a server program stored in memory and/or storage 404 of a server (e.g., such as server 302). For example, the server program can cause hardware processor 402 fingerprint one or more frames of an uploaded video content item, determine whether the fingerprint(s) match a video content item in a reference database, block an uploaded video content item in response to determining the video content item matches a video content item in a reference database, and/or perform any other suitable actions. In some embodiments, hardware processor 402 can be controlled by a computer program stored in memory and/or storage 404 of user device 306. For example, the computer program can cause hardware processor 402 to transmit a video content item to server(s) 302, and/or perform any other suitable actions.

[0023] Memory and/or storage 404 can be any suitable memory and/or storage for storing programs, data, media content, advertisements, and/or any other suitable information in some embodiments. For example, memory and/or storage 404 can include random access memory, read-only memory, flash memory, hard disk storage, optical media, and/or any other suitable memory.

[0024] Input device controller 406 can be any suitable circuitry for controlling and receiving input from one or more input devices 408 in some embodiments. For example, input device controller 406 can be circuitry for receiving input from a touchscreen, from a keyboard, from a mouse, from one or more buttons, from a voice recognition circuit, from a microphone, from a camera, from an optical sensor, from an accelerometer, from a temperature sensor, from a near field sensor, and/or any other type of input device. In another example, input device controller 406 can be circuitry for receiving input from a head-mountable device (e.g., for presenting virtual reality content or augmented reality content).

[0025] Display/audio drivers 410 can be any suitable circuitry for controlling and driving output to one or more display/audio output devices 412 in some embodiments. For example, display/audio drivers 410 can be circuitry for driving a touchscreen, a flat-panel display, a cathode ray tube display, a projector, a speaker or speakers, and/or any other suitable display and/or presentation devices.

[0026] Communication interface(s) 414 can be any suitable circuitry for interfacing with one or more communication networks, such as network 304 as shown in FIG. 3. For example, interface(s) 414 can include network interface card circuitry, wireless communication circuitry, and/or any other suitable type of communication network circuitry.

[0027] Antenna 416 can be any suitable one or more antennas for wirelessly communicating with a communication network (e.g., communication network 304) in some embodiments. In some embodiments, antenna 416 can be omitted.

[0028] Bus 418 can be any suitable mechanism for communicating between two or more components 402, 404, 406, 410, and 414 in some embodiments.

[0029] Any other suitable components can be included in hardware 400 in accordance with some embodiments.

[0030] Turning to FIG. 5, an example 500 of a process for identifying content in stereoscopic videos and, more particularly, for detecting abusive stereoscopic videos by generating fingerprints for multiple portions of a video frame is shown in accordance with some embodiments of the disclosed subject matter. In some embodiments, blocks of process 500 can be executed on server(s) 302.

[0031] Process 500 can begin by receiving an uploaded video content item at 502. In some embodiments, the video content item can be uploaded by any suitable entity. For example, in some embodiments, the video content item can be uploaded by a creator of the video content item, and/or any other suitable entity. In another example, in some embodiments, multiple video content items can be uploaded by an entity to a content provider (e.g., using an application program interface). In some embodiments, the video content item can be any suitable type of video content, such as a television program, a movie, live-streamed content (e.g., where the uploaded video content item is a video identifier of the live-streamed content), a video advertisement, and/or any other suitable type of video content. Additionally, in some embodiments, the video content item can include three-dimensional video content, which can be encoded in any suitable manner. For example, in some embodiments, the three-dimensional video content can be encoded as stereoscopic video content, with left and right images encoded in any suitable format. As a more particular example, in some embodiments, left and right images can be encoded side by side within each frame of the video content item, in a top half and a bottom half of each frame of the video content item, through interleaved pixels within each frame of the video content item (e.g., in a checkerboard pattern), as sequential frames of the video content item, and/or in any other suitable format.

[0032] In some embodiments, process 500 can additionally receive any suitable information about the uploaded video content item. For example, as shown in and described above in connection with FIG. 1, process 500 can receive information from a user uploading the video content item that indicates whether the video content item includes three-dimensional content that is stereoscopically encoded. In some embodiments, process 500 can additionally or alternatively receive any other suitable information, such as a name of the video content item, a name of a creator of the video content item, credit information or collaboration information associated with the video content item, keywords associated with the video content item, access control or privacy settings associated with the video content item, and/or any other suitable information.

[0033] Note that, in some embodiments, process 500 can determine whether the uploaded video content item includes three-dimensional video content that is stereoscopically encoded using any suitable technique or combination of techniques. For example, in some embodiments, process 500 can determine whether image content in a pixel in a left half of a frame corresponds to image content in a corresponding pixel in a right half of the frame to determine whether the frame includes image content to be delivered to left and right eyes of a viewer. As another example, in some embodiments, process 500 can determine whether image content in a pixel in a top half of a frame corresponds to image content in a corresponding pixel in a bottom half of the frame to determine whether the frame includes image content to be delivered to left and right eyes of a viewer. Additionally or alternatively, in instances where process 500 receives an explicit indication from a user (e.g., via user interface 100, and/or in any other suitable manner) of whether the uploaded video content item includes three-dimensional video content, process 500 can verify the explicit user input using any suitable technique or combination of techniques.

[0034] At 504, in some embodiments, process 500 can select a frame of the uploaded video content item to be fingerprinted. In some embodiments, the frame can be selected in any suitable manner. For example, in some embodiments, process 500 can select a frame that is determined to be a key frame in the video content item. As a more particular example, in some embodiments, the selected key frame can be one that is identified based on a motion analysis of the video content item and includes a time point with content in motion. As another more particular example, in some embodiments, the selected key frame can be one that is a boundary between two different scenes in the video content item. As another example, in some embodiments, process 500 can randomly select a frame of the uploaded video content item. As yet another example, in some embodiments, process 500 can generate fingerprints corresponding to a series of frames (e.g., every frame, every other frame, every tenth frame, and/or any other series) of the uploaded video content item until either a match to a video content item in a reference database is found or all of the frames of the uploaded video content item have been analyzed.

[0035] In some embodiments, process 500 can employ a plurality of approaches for selecting a frame from the uploaded video content item. For example, process 500 can determine which of the plurality of approaches (e.g., motion analysis, boundary analysis, random selection, etc.) has resulted in a generated fingerprint that matches a fingerprint of a reference video content item with the fewest number of selected frames and, in response, assign process 500 to use that approach for selecting the frame from the uploaded video content item. In another example, process 500 can select multiple approaches for selecting a frame from the uploaded video content item, where one or more fingerprints are generated for a frame selected based on a motion analysis and one or more fingerprints are generated for a frame selected based on boundary analysis.

[0036] In some embodiments, process 500 can generate one or more fingerprints of the entire selected frame at 506. Turning to FIG. 2A, an example of a frame 200 is shown in accordance with some embodiments of the disclosed subject matter. As illustrated, frame 200 can include any suitable number of pixels (such as pixel 202) that can be arranged in any suitable manner. In some embodiments, process 500 can generate one or more fingerprints of frame 200 by creating one or more signatures that represent the image content of frame 200. For example, in some embodiments, the one or more signatures can include a histogram of color, saturation, or hue information of frame 200. In some embodiments, the one or more signatures can be based on any other suitable information relating to frame 200. In some embodiments, process 500 can divide frame 200 into two or more spatial blocks and can calculate one or more signatures for each of the spatial blocks. In some embodiments, a first fingerprint generated by process 500 corresponding to the entire selected frame can be based on any of the generated signatures or any combination of the signatures.

[0037] In some embodiments, process 500 can generate one or more fingerprints of a first encoded portion of the selected frame and a second encoded portion of the selected frame at 508. In some embodiments, the first encoded portion of the selected frame can be a first stereoscopic image to be delivered to one eye of a viewer and the second encoded portion of the selected frame can be a second stereoscopic image to be delivered to the other eye of the viewer. In some embodiments, the first encoded portion and the second encoded portion of the selected frame can be encoded in any suitable format. For example, as shown in FIG. 2B, frame 250 can have a left half frame 252, which can correspond to the first encoded portion of frame 250, and a right half frame 254, which can correspond to the second encoded portion of frame 250. Note that, in some embodiments, the first encoded portion of the selected frame and the second encoded portion of the selected frame can be in any suitable arrangement, such as left and right halves (as shown in FIG. 2B), top and bottom halves, interleaved on a pixel-by-pixel basis (e.g., in a checkerboard pattern), interleaved on a frame-by-frame basis, and/or arranged in any other suitable manner. Additionally, note that, in some embodiments, process 500 can generate one or more fingerprints of the first encoded portion of the selected frame and the second encoded portion of the selected frame irrespective of whether or not frame 250 includes stereoscopic content.

[0038] It should be noted that process 500 can use any suitable technique or combination of techniques to generate one or more fingerprints of the first encoded portion of the selected frame and the second encoded portion of the selected frame. For example, in some embodiments, process 500 can use the techniques described above in connection with block 506 to generate a first fingerprint that represents the first encoded portion of the selected frame and a second fingerprint that represents the second encoded portion of the selected frame. As discussed above in connection with block 506, in some embodiments, the first fingerprint and the second fingerprint can be based on any suitable information or signatures for each encoded portion of the selected frame, such as histograms of color, saturation, or hue, and/or any other suitable information.

[0039] In some embodiments, process 500 can determine whether at least one of the fingerprints generated at blocks 506 and 508 matches a fingerprint of a reference video content item stored in a reference database at 510. Process 500 can determine whether there is a match using any suitable technique or combination of techniques. For example, in some embodiments, process 500 can determine whether a distance (e.g., a Euclidean distance, and/or any other suitable distance metric) between a pattern of any of the generated fingerprints associated with the uploaded video content item and a pattern of fingerprints for a reference video content item is less than a predetermined threshold. As another example, in some embodiments, process 500 can determine whether a correlation between any of the generated fingerprints associated with the uploaded video content item and a fingerprint of a reference video content item exceeds a predetermined threshold. Note that, in some embodiments, process 500 can additionally or alternatively calculate a probability that at least one of the fingerprints generated at blocks 506 and 508 match a fingerprint of a video content item in the reference database. In some such embodiments, process 500 can calculate the probability of a match using any suitable technique or combination of techniques. For example, in some embodiments, process 500 can calculate the probability of a match based on a distance metric between the fingerprints, a correlation between the fingerprints, and/or any other suitable factor or combination of factors (e.g., where the resulting probability score is between 0 and 1).

[0040] Note that, in some embodiments, process 500 can determine whether any of the generated fingerprints match a fingerprint of a video content item in the reference database based on fingerprints generated from multiple frames of the uploaded video content item. For example, in some embodiments, process 500 can generate fingerprints for multiple frames of the uploaded video content item and can calculate any suitable metrics (e.g., a distance metric, a correlation metric, and/or any other suitable metric) indicating a closeness of each of the generated fingerprints to fingerprints of a reference video content item. Process 500 can then combine the metrics for each frame to calculate a probability that the uploaded video content item matches the reference video content item. In some embodiments, process 500 can combine fingerprint information across multiple frames of the video content item using whole frames of the video content item (e.g., as generated at block 506), and/or one or both encoded portions of the selected frame (e.g., as generated at block 508).

[0041] In some embodiments, process 500 can additionally or alternatively generate one or more audio fingerprints (e.g., at blocks 506 and/or at 508). Process 500 can then determine whether the generated audio fingerprint(s) correspond to a reference audio fingerprint of a reference video in the reference database using any suitable metrics as described above. In some such embodiments, process 500 can determine a probability that the uploaded video content item matches the reference video based on any of the audio fingerprint, the fingerprint(s) corresponding to the entire selected frame of the uploaded video content item, the fingerprint corresponding to the first encoded portion of the uploaded video content item, and/or the fingerprint corresponding to the second encoded portion of the uploaded video content item (e.g., where each of the one or more fingerprints receives a probability score when compared to a reference fingerprint and where a combined probability score can be generated that incorporates each individual probability score).

[0042] If, at 510, process 500 determines that one or more of the fingerprints match fingerprints of a reference video content item ("yes" at 510), process 500 can proceed to block 512 and can block storage of the uploaded video content item. In some embodiments, process 500 can cause any suitable message indicating that the uploaded video content item has been blocked from storage to be presented on a user device that transmitted the video content item. For example, in some embodiments, the message can indicate that the uploaded video content item has been blocked because the uploaded video content item has been determined to match a video content item in a reference database (e.g., copyrighted video content, and/or any other suitable video content). In some embodiments, the message can include any other suitable information, such as a name or other identifier of the reference video that matched the uploaded video content item, a name of the creator of the reference video, one or more time points at which the uploaded video content item matched the reference video, and/or any other suitable information. Note that, in some embodiments, process 500 can allow the uploaded video content item to be stored, but can flag the uploaded video content item for further manual review. Alternatively, in some embodiments, process 500 can block storage of the uploaded video content item until further manual review has been completed.

[0043] If, at 510, process 500 determines that none of the generated fingerprints match fingerprints of a reference video content item ("no" at 510), process 500 can proceed to block 514 and can allow storage of the uploaded video content item. In some embodiments, the uploaded video content item can then be presented at a later time, for example, on a user device in response to receiving a request for the uploaded video content item. In some embodiments, the uploaded video content item can be stored on server(s) 302. In some embodiments, the uploaded video content item can be stored in connection with any suitable information, such as a name of a creator of the uploaded video content item, a date the video content item was uploaded to server(s) 302, keywords and/or topics associated with the uploaded video content item, and/or any other suitable information. In some embodiments, the stored information can further indicate that the uploaded video content item was not found to match any video content in a reference database and can indicate a time and/or a date on which the uploaded video content item was most recently compared to video content in the reference database.

[0044] Note that, in some embodiments, process 500 can repeat blocks 504-510 at any suitable time. For example, in some embodiments, process 500 can repeat blocks 504-510 at a later time (e.g., after the uploaded video content item has been allowed to be stored on server(s) 302) to determine if the uploaded video content item matches video content added to a reference database since the uploaded video content item was approved for storage at block 514. As another example, in some embodiments, process 500 can repeat blocks 504-510 in response to receiving an indication from a user who uploaded the video content item that information about the video content item has changed. As a more particular example, in an instance where the user had previously indicated that the uploaded video content item contains three-dimensional video content that is encoded as stereoscopic video content (e.g., as shown in and described above in connection with FIG. 1A), and subsequently indicates that the uploaded video content is not three-dimensional content, process 500 can repeat blocks 504-510 to determine if one or more fingerprints of whole frames of the video content item match fingerprints of a reference video content item in a reference database.

[0045] In some embodiments, at least some of the above described blocks of the process of FIG. 5 can be executed or performed in any order or sequence not limited to the order and sequence shown in and described in connection with the figure. Also, some of the above blocks of FIG. 5 can be executed or performed substantially simultaneously where appropriate or in parallel to reduce latency and processing times. Additionally or alternatively, some of the above described blocks of the process of FIG. 5 can be omitted.

[0046] In some embodiments, any suitable computer readable media can be used for storing instructions for performing the functions and/or processes herein. For example, in some embodiments, computer readable media can be transitory or non-transitory. For example, non-transitory computer readable media can include media such as magnetic media (such as hard disks, floppy disks, and/or any other suitable magnetic media), optical media (such as compact discs, digital video discs, Blu-ray discs, and/or any other suitable optical media), semiconductor media (such as flash memory, electrically programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), and/or any other suitable semiconductor media), any suitable media that is not fleeting or devoid of any semblance of permanence during transmission, and/or any suitable tangible media. As another example, transitory computer readable media can include signals on networks, in wires, conductors, optical fibers, circuits, any suitable media that is fleeting and devoid of any semblance of permanence during transmission, and/or any suitable intangible media.

[0047] In situations in which the systems described herein collect personal information about users, or make use of personal information, the users may be provided with an opportunity to control whether programs or features collect user information (e.g., information about a user's social network, social actions or activities, profession, a user's preferences, or a user's current location). In addition, certain data may be treated in one or more ways before it is stored or used, so that personal information is removed. For example, a user's identity may be treated so that no personal information can be determined for the user, or a user's geographic location may be generalized where location information is obtained (such as to a city, ZIP code, or state level), so that a particular location of a user cannot be determined. Thus, the user may have control over how information is collected about the user and used by a content server.

[0048] Accordingly, methods, systems, and media for identifying content in stereoscopic videos are provided.


Claims

1. A computer-implemented method for identifying content in stereoscopic videos, the method comprising:

receiving, from a user device, a video content item for uploading to a content provider;

selecting a frame from a plurality of frames of the video content item for generating one or more fingerprints corresponding to the video content item;

generating a first fingerprint corresponding to the selected frame, a second fingerprint corresponding to a first encoded portion of the selected frame, and a third fingerprint corresponding to a second encoded portion of the selected frame;

comparing each of the first fingerprint, the second fingerprint, and the third fingerprint to a plurality of reference fingerprints corresponding to reference video content items;

determining whether at least one of the first fingerprint, the second fingerprint, and the third fingerprint match a reference fingerprint of the plurality of reference fingerprints; and

in response to determining that at least one of the first fingerprint, the second fingerprint, and the third fingerprint match the reference fingerprint, causing an indication of the match to be presented on the user device, wherein the first encoded portion of the selected frame and the second encoded portion of the selected frame correspond to a first image to be presented to a left eye of a viewer and a second image to be presented to a right eye of the viewer when the video content item is presented stereoscopically.


 
2. The method of claim 1, further comprising causing storage of the video content item to be blocked in response to determining that at least one of the first fingerprint, the second fingerprint, and the third fingerprint match the reference fingerprint.
 
3. The method of claim 2, wherein the first encoded portion of the selected frame is a left half of the selected frame and the second encoded portion of the selected frame is a right half of the selected frame.
 
4. The method of claim 2, wherein the first encoded portion of the selected frame is a top half of the selected frame and the second encoded portion of the selected frame is a bottom half of the selected frame.
 
5. The method of any preceding claim, further comprising receiving, from the user device, an indication that the video content item contains three-dimensional video content, wherein the first fingerprint and the second fingerprint are generated in response to receiving the indication that the video content item contains three-dimensional video content.
 
6. A system for identifying content in stereoscopic videos, the system comprising:
a hardware processor that is configured to:

receive, from a user device, a video content item for uploading to a content provider;

select a frame from a plurality of frames of the video content item for generating one or more fingerprints corresponding to the video content item;

generate a first fingerprint corresponding to the selected frame, a second fingerprint corresponding to a first encoded portion of the selected frame, and a third fingerprint corresponding to a second encoded portion of the selected frame;

compare each of the first fingerprint, the second fingerprint, and the third fingerprint to a plurality of reference fingerprints corresponding to reference video content items;

determine whether at least one of the first fingerprint, the second fingerprint, and the third fingerprint match a reference fingerprint of the plurality of reference fingerprints; and

in response to determining that at least one of the first fingerprint, the second fingerprint, and the third fingerprint match the reference fingerprint, cause an indication of the match to be presented on the user device, wherein the first encoded portion of the selected frame and the second encoded portion of the selected frame correspond to a first image to be presented to a left eye of a viewer and a second image to be presented to a right eye of the viewer when the video content item is presented stereoscopically.


 
7. The system of claim 6, wherein the hardware processor is further configured to cause storage of the video content item to be blocked in response to determining that at least one of the first fingerprint, the second fingerprint, and the third fingerprint match the reference fingerprint.
 
8. The system of claim 7, wherein the first encoded portion of the selected frame is a left half of the selected frame and the second encoded portion of the selected frame is a right half of the selected frame.
 
9. The system of claim 7, wherein the first encoded portion of the selected frame is a top half of the selected frame and the second encoded portion of the selected frame is a bottom half of the selected frame.
 
10. The system of any of claims 6 to 9, wherein the hardware processor is further configured to receive, from the user device, an indication that the video content item contains three-dimensional video content, wherein the first fingerprint and the second fingerprint are generated in response to receiving the indication that the video content item contains three-dimensional video content.
 
11. A non-transitory computer-readable medium containing computer-executable instructions that, when executed by a processor, cause the processor to perform a method according to claim 1.
 
12. The non-transitory computer-readable medium of claim 11, wherein the method further comprises causing storage of the video content item to be blocked in response to determining that at least one of the first fingerprint, the second fingerprint, and the third fingerprint match the reference fingerprint.
 
13. The non-transitory computer-readable medium of claim 12, wherein the first encoded portion of the selected frame is a left half of the selected frame and the second encoded portion of the selected frame is a right half of the selected frame.
 
14. The non-transitory computer-readable medium of claim 12, wherein the first encoded portion of the selected frame is a top half of the selected frame and the second encoded portion of the selected frame is a bottom half of the selected frame.
 
15. The non-transitory computer-readable medium of any of claims 11 to 14, wherein the method further comprises receiving, from the user device, an indication that the video content item contains three-dimensional video content, wherein the first fingerprint and the second fingerprint are generated in response to receiving the indication that the video content item contains three-dimensional video content.
 


Ansprüche

1. Computerimplementiertes Verfahren zum Identifizieren von Inhalt in stereoskopischen Videos, wobei das Verfahren umfasst:

Empfangen eines Videoinhaltselements von einer Benutzervorrichtung zum Hochladen zu einem Inhaltsanbieter;

Auswählen eines Einzelbildes aus einer Vielzahl von Einzelbildern des Videoinhaltselements zum Erzeugen von einem oder mehreren Fingerabdrücken, die dem Videoinhaltselement entsprechen;

Erzeugen eines ersten Fingerabdrucks, der dem ausgewählten Einzelbild entspricht, eines zweiten Fingerabdrucks, der einem ersten codierten Abschnitt des ausgewählten Einzelbilds entspricht, und eines dritten Fingerabdrucks, der einem zweiten codierten Abschnitt des ausgewählten Einzelbilds entspricht;

Vergleichen von jedem des ersten Fingerabdrucks, des zweiten Fingerabdrucks und des dritten Fingerabdrucks mit einer Vielzahl von Referenz-Fingerabdrücken, die Referenz-Videoinhaltselementen entsprechen;

Bestimmen, ob mindestens einer des ersten Fingerabdrucks, des zweiten Fingerabdrucks und des dritten Fingerabdrucks mit einem Referenz-Fingerabdruck aus der Vielzahl von Referenz-Fingerabdrücken übereinstimmt; und

als Antwort auf das Bestimmen, dass mindestens einer des ersten Fingerabdrucks, des zweiten Fingerabdrucks und des dritten Fingerabdrucks mit dem Referenz-Fingerabdruck übereinstimmt, erfolgendes Bewirken, dass auf der Benutzervorrichtung ein Hinweis auf die Übereinstimmung präsentiert wird, wobei der erste codierte Abschnitt des ausgewählten Einzelbildes und der zweite codierte Abschnitt des ausgewählten Einzelbildes einem ersten Bild, das einem linken Auge eines Betrachters präsentiert werden soll, und einem zweiten Bild, das einem rechten Auge des Betrachters präsentiert werden soll, entsprechen, wenn das Videoinhaltselement stereoskopisch präsentiert wird.


 
2. Verfahren nach Anspruch 1, ferner umfassend: als Antwort auf das Bestimmen, dass mindestens einer des ersten Fingerabdrucks, des zweiten Fingerabdrucks und des dritten Fingerabdrucks mit dem Referenz-Fingerabdruck übereinstimmt, erfolgendes Bewirken, dass die Speicherung des Videoinhaltselements blockiert wird.
 
3. Verfahren nach Anspruch 2, wobei der erste codierte Abschnitt des ausgewählten Einzelbildes eine linke Hälfte des ausgewählten Einzelbildes ist und der zweite codierte Abschnitt des ausgewählten Einzelbildes eine rechte Hälfte des ausgewählten Einzelbildes ist.
 
4. Verfahren nach Anspruch 2, wobei der erste codierte Abschnitt des ausgewählten Einzelbildes eine obere Hälfte des ausgewählten Einzelbildes ist und der zweite codierte Abschnitt des ausgewählten Einzelbildes eine untere Hälfte des ausgewählten Einzelbildes ist.
 
5. Verfahren nach einem der vorhergehenden Ansprüche, ferner umfassend:
Empfangen eines Hinweises von der Benutzervorrichtung, dass das Videoinhaltselement dreidimensionalen Videoinhalt enthält, wobei der erste Fingerabdruck und der zweite Fingerabdruck als Antwort auf das Empfangen des Hinweises, dass das Videoinhaltselement dreidimensionalen Videoinhalt enthält, erzeugt werden.
 
6. System zum Identifizieren von Inhalt in stereoskopischen Videos, wobei das System umfasst:
einen Hardware-Prozessor, der dafür konfiguriert ist:

von einer Benutzervorrichtung ein Videoinhaltselement zum Hochladen zu einem Inhaltsanbieter zu empfangen;

ein Einzelbild aus einer Vielzahl von Einzelbildern des Videoinhaltselements auszuwählen, um einen oder mehrere Fingerabdrücke zu erzeugen, die dem Videoinhaltselement entsprechen;

einen ersten Fingerabdruck, der dem ausgewählten Einzelbild entspricht, einen zweiten Fingerabdruck, der einem ersten codierten Abschnitt des ausgewählten Einzelbilds entspricht, und einen dritten Fingerabdruck, der einem zweiten codierten Abschnitt des ausgewählten Einzelbilds entspricht, zu erzeugen;

jeden des ersten Fingerabdrucks, des zweiten Fingerabdrucks und des dritten Fingerabdrucks mit einer Vielzahl von Referenz-Fingerabdrücken, die Referenz-Videoinhaltselementen entsprechen, zu vergleichen;

zu bestimmen, ob mindestens einer des ersten Fingerabdrucks, des zweiten Fingerabdrucks und des dritten Fingerabdrucks mit einem Referenz-Fingerabdruck der Vielzahl von Referenz-Fingerabdrücken übereinstimmt; und

als Antwort auf das Bestimmen, dass mindestens einer des ersten Fingerabdrucks, des zweiten Fingerabdrucks und des dritten Fingerabdrucks mit dem Referenz-Fingerabdruck übereinstimmt, zu bewirken, dass auf der Benutzervorrichtung ein Hinweis auf die Übereinstimmung präsentiert wird, wobei der erste codierte Abschnitt des ausgewählten Einzelbildes und der zweite codierte Abschnitt des ausgewählten Einzelbildes einem ersten Bild, das einem linken Auge eines Betrachters präsentiert werden soll, und einem zweiten Bild, das einem rechten Auge des Betrachters präsentiert werden soll, entsprechen, wenn das Videoinhaltselement stereoskopisch präsentiert wird.


 
7. System nach Anspruch 6, wobei der Hardware-Prozessor ferner dafür konfiguriert ist, als Antwort auf das Bestimmen, dass mindestens einer des ersten Fingerabdrucks, des zweiten Fingerabdrucks und des dritten Fingerabdrucks mit dem Referenz-Fingerabdruck übereinstimmt, zu bewirken, dass die Speicherung des Videoinhaltselements blockiert wird.
 
8. System nach Anspruch 7, wobei der erste codierte Abschnitt des ausgewählten Einzelbildes eine linke Hälfte des ausgewählten Einzelbildes ist und der zweite codierte Abschnitt des ausgewählten Einzelbildes eine rechte Hälfte des ausgewählten Einzelbildes ist.
 
9. System nach Anspruch 7, wobei der erste codierte Abschnitt des ausgewählten Einzelbildes eine obere Hälfte des ausgewählten Einzelbildes ist und der zweite codierte Abschnitt des ausgewählten Einzelbildes eine untere Hälfte des ausgewählten Einzelbildes ist.
 
10. System nach einem der Ansprüche 6 bis 9, wobei der Hardware-Prozessor ferner dafür konfiguriert ist, von der Benutzervorrichtung einen Hinweis zu empfangen, dass das Videoinhaltselement dreidimensionalen Videoinhalt enthält, wobei der erste Fingerabdruck und der zweite Fingerabdruck als Antwort auf das Empfangen des Hinweises, dass das Videoinhaltselement dreidimensionalen Videoinhalt enthält, erzeugt werden.
 
11. Nichtflüchtiges computerlesbares Medium, computerausführbare Anweisungen enthaltend, die, wenn sie durch einen Prozessor ausgeführt werden, bewirken, dass der Prozessor ein Verfahren nach Anspruch 1 durchführt.
 
12. Nichtflüchtiges computerlesbares Medium nach Anspruch 11, wobei das Verfahren ferner umfasst: als Antwort auf das Bestimmen, dass mindestens einer des ersten Fingerabdrucks, des zweiten Fingerabdrucks und des dritten Fingerabdrucks mit dem Referenz-Fingerabdruck übereinstimmt, erfolgendes Bewirken, dass die Speicherung des Videoinhaltselements blockiert wird.
 
13. Nichtflüchtiges computerlesbares Medium nach Anspruch 12, wobei der erste codierte Abschnitt des ausgewählten Einzelbildes eine linke Hälfte des ausgewählten Einzelbildes ist und der zweite codierte Abschnitt des ausgewählten Einzelbildes eine rechte Hälfte des ausgewählten Einzelbildes ist.
 
14. Nichtflüchtiges computerlesbares Medium nach Anspruch 12, wobei der erste codierte Abschnitt des ausgewählten Bildes eine obere Hälfte des ausgewählten Bildes und der zweite codierte Abschnitt des ausgewählten Bildes eine untere Hälfte des ausgewählten Bildes ist.
 
15. Nichtflüchtiges computerlesbares Medium nach einem der Ansprüche 11 bis 14, wobei das Verfahren ferner umfasst: Empfangen eines Hinweises von der Benutzervorrichtung, dass das Videoinhaltselement dreidimensionalen Videoinhalt enthält, wobei der erste Fingerabdruck und der zweite Fingerabdruck als Antwort auf das Empfangen des Hinweises, dass das Videoinhaltselement dreidimensionalen Videoinhalt enthält, erzeugt werden.
 


Revendications

1. Procédé mis en œuvre par ordinateur pour identifier un contenu dans des vidéos stéréoscopiques, le procédé comprenant :

la réception, à partir d'un dispositif utilisateur, d'un élément de contenu vidéo à télécharger vers un fournisseur de contenu ;

la sélection d'une trame parmi une pluralité de trames de l'élément de contenu vidéo pour générer une ou plusieurs empreintes digitales correspondant à l'élément de contenu vidéo ;

la génération d'une première empreinte digitale correspondant à la trame sélectionnée, d'une deuxième empreinte digitale correspondant à une première partie codée de la trame sélectionnée, et d'une troisième empreinte digitale correspondant à une deuxième partie codée de la trame sélectionnée ;

la comparaison de chacune de la première empreinte digitale, de la deuxième empreinte digitale et de la troisième empreinte digitale à une pluralité d'empreintes digitales de référence correspondant à des éléments de contenu vidéo de référence ;

la détermination du fait qu'au moins l'une de la première empreinte digitale, de la deuxième empreinte digitale et de la troisième empreinte digitale correspond à une empreinte digitale de référence de la pluralité d'empreintes digitales de référence ; et

en réponse à la détermination du fait qu'au moins l'une de la première empreinte digitale, de la deuxième empreinte digitale et de la troisième empreinte digitale correspond à l'empreinte digitale de référence, le fait de provoquer la présentation d'une indication de la correspondance sur le dispositif utilisateur, dans lequel la première partie codée de la trame sélectionnée et la deuxième partie codée de la trame sélectionnée correspondent à une première trame à présenter à un œil gauche d'un spectateur et à une deuxième trame à présenter à un œil droit du spectateur lorsque l'élément de contenu vidéo est présenté de manière stéréoscopique.


 
2. Procédé selon la revendication 1, comprenant en outre le fait de provoquer le blocage du stockage de l'élément de contenu vidéo en réponse à la détermination du fait qu'au moins l'une de la première empreinte digitale, de la deuxième empreinte digitale et de la troisième empreinte digitale correspond à l'empreinte digitale de référence.
 
3. Procédé selon la revendication 2, dans lequel la première partie codée de la trame sélectionnée est une moitié gauche de la trame sélectionnée et la deuxième partie codée de la trame sélectionnée est une moitié droite de la trame sélectionnée.
 
4. Procédé selon la revendication 2, dans lequel la première partie codée de la trame sélectionnée est une moitié supérieure de la trame sélectionnée et la deuxième partie codée de la trame sélectionnée est une moitié inférieure de la trame sélectionnée.
 
5. Procédé selon l'une quelconque des revendications précédentes, comprenant en outre la réception, depuis le dispositif utilisateur, d'une indication que l'élément de contenu vidéo contient un contenu vidéo tridimensionnel, dans lequel la première empreinte digitale et la deuxième empreinte digitale sont générées en réponse à la réception de l'indication que l'élément de contenu vidéo contient un contenu vidéo tridimensionnel.
 
6. Système pour identifier un contenu dans des vidéos stéréoscopiques, le système comprenant :
un processeur matériel qui est configuré pour :

recevoir, à partir d'un dispositif utilisateur, un élément de contenu vidéo à télécharger vers un fournisseur de contenu ;

sélectionner une trame parmi une pluralité de trames de l'élément de contenu vidéo pour générer une ou plusieurs empreintes digitales correspondant à l'élément de contenu vidéo ;

générer une première empreinte digitale correspondant à la trame sélectionnée, d'une deuxième empreinte digitale correspondant à une première partie codée de la trame sélectionnée, et d'une troisième empreinte digitale correspondant à une deuxième partie codée de la trame sélectionnée ;

comparer chacune de la première empreinte digitale, de la deuxième empreinte digitale et de la troisième empreinte digitale à une pluralité d'empreintes digitales de référence correspondant à des éléments de contenu vidéo de référence ;

déterminer si au moins l'une de la première empreinte digitale, de la deuxième empreinte digitale et de la troisième empreinte digitale correspond à une empreinte digitale de référence de la pluralité d'empreintes digitales de référence ; et

en réponse à la détermination du fait qu'au moins l'une de la première empreinte digitale, de la deuxième empreinte digitale et de la troisième empreinte digitale correspond à l'empreinte digitale de référence, provoquer la présentation d'une indication de la correspondance sur le dispositif utilisateur, dans lequel la première partie codée de l'image sélectionnée et la deuxième partie codée de l'image sélectionnée correspondent à une première image à présenter à un œil gauche d'un spectateur et à une deuxième image à présenter à un œil droit du spectateur lorsque l'élément de contenu vidéo est présenté de manière stéréoscopique.


 
7. Système selon la revendication 6, dans lequel le processeur matériel est en outre configuré pour provoquer le blocage du stockage de l'élément de contenu vidéo en réponse à la détermination du fait qu'au moins l'une de la première empreinte digitale, de la deuxième empreinte digitale et de la troisième empreinte digitale correspond à l'empreinte digitale de référence.
 
8. Système selon la revendication 7, dans lequel la première partie codée de la trame sélectionnée est une moitié gauche de la trame sélectionnée et la deuxième partie codée de la trame sélectionnée est une moitié droite de la trame sélectionnée.
 
9. Système selon la revendication 7, dans lequel la première partie codée de la trame sélectionnée est une moitié supérieure de la trame sélectionnée et la deuxième partie codée de la trame sélectionnée est une moitié inférieure de la trame sélectionnée.
 
10. Système selon l'une quelconque des revendications 6 à 9, dans lequel le processeur matériel est en outre configuré pour recevoir, du dispositif utilisateur, une indication que l'élément de contenu vidéo contient un contenu vidéo tridimensionnel, dans lequel la première empreinte digitale et la deuxième empreinte digitale sont générées en réponse à la réception de l'indication que l'élément de contenu vidéo contient un contenu vidéo tridimensionnel.
 
11. Support lisible par ordinateur non transitoire contenant des instructions exécutables par ordinateur qui, lorsqu'elles sont exécutées par un processeur, amènent le processeur à réaliser un procédé selon la revendication 1.
 
12. Support lisible par ordinateur non transitoire selon la revendication 11, dans lequel le procédé comprend en outre le fait de provoquer le blocage du stockage de l'élément de contenu vidéo en réponse à la détermination du fait qu'au moins l'une de la première empreinte digitale, de la deuxième empreinte digitale et de la troisième empreinte digitale correspond à l'empreinte digitale de référence.
 
13. Support lisible par ordinateur non transitoire selon la revendication 12, dans lequel la première partie codée de la trame sélectionnée est une moitié gauche de la trame sélectionnée et la deuxième partie codée de la trame sélectionnée est une moitié droite de la trame sélectionnée.
 
14. Support lisible par ordinateur non transitoire selon la revendication 12, dans lequel la première partie codée de la trame sélectionnée est une moitié supérieure de la trame sélectionnée et la deuxième partie codée de la trame sélectionnée est une moitié inférieure de la trame sélectionnée.
 
15. Support lisible par ordinateur non transitoire selon l'une quelconque des revendications 11 à 14, dans lequel le procédé comprend en outre la réception, depuis le dispositif utilisateur, d'une indication que l'élément de contenu vidéo contient un contenu vidéo tridimensionnel, dans lequel la première empreinte digitale et la deuxième empreinte digitale sont générées en réponse à la réception de l'indication que l'élément de contenu vidéo contient un contenu vidéo tridimensionnel.
 




Drawing




















Cited references

REFERENCES CITED IN THE DESCRIPTION



This list of references cited by the applicant is for the reader's convenience only. It does not form part of the European patent document. Even though great care has been taken in compiling the references, errors or omissions cannot be excluded and the EPO disclaims all liability in this regard.

Patent documents cited in the description




Non-patent literature cited in the description