WO2016041311A1 - 视频浏览方法及装置 - Google Patents
视频浏览方法及装置 Download PDFInfo
- Publication number
- WO2016041311A1 WO2016041311A1 PCT/CN2015/071247 CN2015071247W WO2016041311A1 WO 2016041311 A1 WO2016041311 A1 WO 2016041311A1 CN 2015071247 W CN2015071247 W CN 2015071247W WO 2016041311 A1 WO2016041311 A1 WO 2016041311A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image frame
- target block
- information
- current image
- video
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 38
- 230000001133 acceleration Effects 0.000 claims description 14
- 238000006073 displacement reaction Methods 0.000 claims description 14
- 230000008859 change Effects 0.000 claims description 11
- 238000012216 screening Methods 0.000 claims description 9
- 230000035945 sensitivity Effects 0.000 claims description 7
- 238000001914 filtration Methods 0.000 claims description 3
- 230000000694 effects Effects 0.000 abstract description 6
- 238000012545 processing Methods 0.000 description 13
- 238000004891 communication Methods 0.000 description 10
- 238000010586 diagram Methods 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 6
- 238000001514 detection method Methods 0.000 description 5
- 230000008569 process Effects 0.000 description 5
- 230000003287 optical effect Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 239000002699 waste material Substances 0.000 description 3
- 230000003993 interaction Effects 0.000 description 2
- 238000007726 management method Methods 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 210000000056 organ Anatomy 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44004—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving video buffer management, e.g. video decoder buffer or video display buffer
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/74—Browsing; Visualisation therefor
- G06F16/745—Browsing; Visualisation therefor the internal structure of a single video sequence
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4312—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/442—Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
- H04N21/44213—Monitoring of end-user related data
- H04N21/44218—Detecting physical presence or behaviour of the user, e.g. using sensors to detect if the user is leaving the room or changes his face expression during a TV program
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/47202—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting content on demand, e.g. video on demand
Definitions
- the present disclosure relates to the field of computer technologies, and in particular, to a video browsing method and apparatus.
- the terminal segments the video into at least two shots during the extraction of the key frames of the video; in response to changes in image quality and video content according to the boundaries of each shot Select a set of keyframes for each shot; reproduce these keyframes on the screen at the appropriate rate for video browsing.
- the disclosed person finds that at least the following defects exist in the related art: the terminal needs to extract a key frame according to the boundary of the lens, and when the video is a continuous lens, the key frame cannot be extracted from the video.
- the present disclosure provides a video browsing method and apparatus.
- a video browsing method including:
- the target block information is performed on the target block included in the current image frame and the target block included in the previous image frame Information obtained after comparison;
- the comparing the current image frame in the video with the previous image frame to obtain the target block information including:
- each target block that does not exist is determined as a new target block, and a flag bit is set for the newly added target block.
- Information get each new item Target block information of the block, the flag bit information is the number of key frames that have been determined plus one;
- the current image frame includes at least one target block existing in the previous image frame, determining each target block that exists as an existing target block, and calculating the existing target block in the current image frame Obtaining target block information of each existing target block based on the movement information of the corresponding target block in the previous image frame;
- each target block that does not exist is determined as a disappearing target block, and target block information of each disappearing target block is obtained.
- determining the current image frame as a key frame includes:
- the current image frame is determined as the key frame.
- determining the current image frame as a key frame includes:
- the current image frame is determined as the key frame.
- the method further includes:
- the sensor information including at least one of global positioning system GPS information, acceleration information, and sensitivity ISO information;
- determining the current image frame as a key frame including:
- the current image frame is determined as a key frame.
- a video browsing apparatus including:
- An image comparison module configured to compare a current image frame in the video with a previous image frame to obtain target block information, the target block information being a target block and the previous image included in the current image frame Information obtained by comparing the target blocks included in the frame;
- An image determining module configured to satisfy the predetermined condition by the target block information obtained by the image comparison module Determining the current image frame as a key frame
- An operation execution module configured to determine a subsequent image frame of the current image frame as a current image frame, and continue to perform the operation of comparing the current image frame in the video with the previous image frame to obtain target block information, Stops after comparing the last image frame in the video;
- an image playing module configured to play at least one key frame determined by the image determining module.
- the image comparison module includes:
- a target comparison submodule configured to compare a target block in the previous image frame with a target block in the current image frame
- a first determining submodule configured to determine, when the target comparison submodule compares at least one target block that does not exist in the previous image frame in the current image frame, determine each target block that does not exist In order to add a target block, and set flag information for the newly added target block, obtain target block information of each newly added target block, where the flag bit information is the number of key frames that have been determined plus one;
- a second determining submodule configured to determine, when the target comparison submodule compares at least one target block existing in the previous image frame in the current image frame, to determine each target block that is present as The target block is calculated, and the existing target block in the current image frame is calculated based on the movement information of the corresponding target block in the previous image frame, and the target block information of each existing target block is obtained;
- a third determining submodule configured to determine, when the target comparison submodule compares at least one target block that does not exist in the current image frame in the previous image frame, determine each target block that does not exist For the disappearing target block, the target block information of each disappearing target block is obtained.
- the image determining module includes:
- a first reading submodule configured to read movement information of each existing target block from target block information of the current image frame, where the movement information includes movement times and movement displacements of the existing target block And at least one of the areas moved to;
- a first detecting submodule configured to detect whether the number of pieces of mobile information satisfying the moving condition is greater than a predetermined threshold
- a fourth determining submodule configured to determine the current image frame as the key frame when the first detecting submodule detects that the number of pieces of mobile information satisfying the moving condition is greater than the predetermined threshold.
- the image determining module includes:
- a second reading submodule configured to read flag bit information of each target block from target block information of the current image frame
- a second detecting submodule configured to detect whether there is the same flag bit information in the respective flag bit information as the flag bit information of the disappearing target block
- a fifth determining submodule configured to determine the current image frame as the key frame when the second detecting submodule detects that there is no flag information identical to the flag information of the vanishing target block .
- the device further includes:
- An information acquisition module configured to acquire each image frame in the video and sense when the image frame is captured Device information, the sensor information including at least one of global positioning system GPS information, acceleration information, and sensitivity ISO information;
- the image filtering module is configured to discard the image frame that does not satisfy the screening condition according to the sensor information acquired in the information acquiring module, and trigger the execution of the comparing the current image frame in the video with the previous image frame to obtain a target.
- the operation of the block information is configured to discard the image frame that does not satisfy the screening condition according to the sensor information acquired in the information acquiring module, and trigger the execution of the comparing the current image frame in the video with the previous image frame to obtain a target.
- the image determining module includes:
- a third detecting submodule configured to detect, when the character is included in the current image frame, whether a change occurs in the character image information of the current image frame relative to the previous image frame, and detect the current image frame Whether the target block information other than the person satisfies the predetermined condition;
- a sixth determining submodule configured to detect, in the third detecting submodule, that the character posture information is changed, and target block information other than the character in the current image frame satisfies the predetermined condition
- the current image frame is determined as a key frame.
- a video browsing apparatus including:
- a memory for storing processor executable instructions
- processor is configured to:
- target block information by comparing a current image frame in the video with a previous image frame, the target block information being obtained by comparing the target block included in the current image frame with the target block included in the previous image frame.
- Information if the target block information satisfies a predetermined condition, the current image frame is determined as a key frame, and the target block information can be obtained by comparing the target blocks of the current image frame in the video, and then the key frame is filtered according to the target block information, instead of
- the key frame needs to be determined through the boundary of the lens, which solves the problem that the inability to browse the video caused by the key frame cannot be extracted from the continuous lens, and the effect of improving the playback efficiency of the video is achieved.
- FIG. 1 is a flowchart of a video browsing method according to an exemplary embodiment.
- FIG. 2 is a flowchart of a video browsing method according to another exemplary embodiment.
- FIG. 3 is a block diagram of a video browsing apparatus, according to an exemplary embodiment.
- FIG. 4 is a block diagram of a video browsing apparatus according to an exemplary embodiment.
- FIG. 5 is a block diagram of an apparatus for video browsing, according to an exemplary embodiment.
- FIG. 1 is a flowchart of a video browsing method according to an exemplary embodiment.
- the video browsing method is applied to a terminal. As shown in FIG. 1 , the video browsing method includes the following steps.
- step 101 the current image frame in the video is compared with the previous image frame to obtain target block information, which is performed on the target block included in the current image frame and the target block included in the previous image frame. The information obtained after comparison.
- An image frame is a unit picture of a video.
- the target block is obtained by dividing the content in the image frame.
- the target block may include at least one object, and may also include a portion of the object. This embodiment does not limit the manner in which the target block is divided.
- the target block information refers to information obtained by comparing a target block of a current image frame in a video with a target block of a previous image frame.
- step 102 if the target block information satisfies a predetermined condition, the current image frame is determined as a key frame.
- a key frame is an image frame in which the key action of an object or object moves throughout the video, and can represent the video content of a time period in the video.
- the terminal can detect whether the image frame is a key frame according to the target block information, Thereby extracting at least one key frame from the video.
- step 103 the next image frame of the current image frame is determined as the current image frame, and the step of comparing the current image frame in the video with the previous image frame to obtain the target block information is performed, until the comparison in the video is completed. Stop after the last image frame.
- the terminal In order to extract all the key frames in the video, the terminal needs to compare all the image frames, so after comparing the current image frame, the next image frame of the current image frame can be determined as the new current image frame, and the The current image frame is determined as the previous image frame of the new current image frame, and step 101 is continued until the last image frame in the video is compared as the current image frame, and the detection flow is ended.
- step 104 the determined at least one key frame is played.
- each key frame can represent the video content of a time period
- at least one determined key frame is played in the terminal at an appropriate frame rate in order, without reducing the effective information to be expressed by the video. Compress the playback time of the video to achieve a quick view of the video content and improve the video playback efficiency.
- the video browsing method obtains target block information by comparing a current image frame in a video with a previous image frame, and the target block information is a target block included in the current image frame and the previous one.
- the key frame is filtered according to the target block information, and the key frame is not determined by the boundary of the lens, which solves the problem that the uncapable video cannot be extracted from the continuous lens, and the effect of improving the video playing efficiency is achieved.
- FIG. 2 is a flowchart of a video browsing method according to another exemplary embodiment.
- the video browsing method is applied to a terminal. As shown in FIG. 2, the video browsing method includes the following steps.
- each image frame in the video and sensor information when the image frame is captured are acquired, and the sensor information includes at least one of GPS (Global Positioning System) information, acceleration information, and sensitivity ISO information. And discarding the image frame that does not satisfy the screening condition according to the sensor information, and triggering step 202 is performed.
- GPS Global Positioning System
- step 201 is an optional step. That is, the terminal can capture the video through the camera, and after performing the video, step 202 is performed; or the terminal can filter the image frame in the video after the video is acquired, and discard the image frame that meets the discarding condition to save the pair.
- the waste of the image frame that satisfies the discarding condition is wasteful, and the processing efficiency of the image frame is improved.
- the image frame is a unit picture of the video.
- the sensor information of each image frame may be collected by the sensor when the video is captured by the camera, and whether each image frame satisfies the discarding condition is detected according to the sensor information.
- the sensor information may be at least one of GPS information, acceleration information, and sensitivity ISO information.
- the terminal can set different discarding conditions according to different sensor information.
- the sensor information is GPS information
- the shooting scene may be shooting video during the movement. If the terminal stays in the same geographical position for a long time, the image frame captured at this time may be repeated. Therefore, the discarding condition can be set: moving
- the distance between the GPS information and the GPS information of the previous image frame is less than the image frame of the GPS threshold.
- the sensor information is acceleration information
- the shooting scene may be shooting video during the movement. If the acceleration of the terminal movement is fast, the image frame captured at this time may be blurred. Therefore, the discarding condition may be set: the acceleration exceeds the acceleration threshold.
- Image frame when the sensor information is ISO information, if the ISO is high, there are more noises in the image frame captured at this time. Therefore, it is possible to set an image frame in which the ISO exceeds the ISO threshold.
- the terminal may combine the discarding conditions corresponding to each of the sensor information, and when at least one of the sensor information satisfies the discarding condition, determine to discard the discarding condition. Image frame.
- step 202 the current image frame in the video is compared with the previous image frame to obtain target block information.
- the target block information is information obtained by comparing a target block included in the current image frame with a target block included in the previous image frame.
- the target block is obtained by dividing the content in the image frame.
- the target block may include at least one object, and may also include a portion of the object, which is determined according to the manner in which the image frame is divided.
- the terminal may perform equal operation on the image frame. For example, the terminal divides the image frame into M in the length direction and divides N in the width direction to obtain M ⁇ N target blocks.
- the terminal may divide the target block according to the smallest target of interest in the image frame, ie, each target block includes at least one minimum target.
- the terminal may divide the target block or the like according to the experience value of the developer, and the embodiment does not limit the manner of dividing the target block.
- the terminal may compare the target block of the current image frame with the target block of the previous image frame to obtain target block information.
- the target block information is used to represent changes in target blocks in two adjacent image frames.
- the current image frame in the video is compared with the previous image frame to obtain target block information, including:
- each target block that does not exist is determined as a new target block, and flag information is set for the newly added target block, and each Target block information of the newly added target block, the flag bit information is the number of key frames that have been determined plus one;
- each target block existing is determined as an existing target block, and the existing target block in the current image frame is calculated based on the previous image frame.
- Corresponding moving information of the target block obtaining target block information of each existing target block;
- each target block that does not exist in the current image frame is included in the previous image frame, each target block that does not exist is determined as the disappearing target block, and the target block information of each disappearing target block is obtained.
- the target block in the current image frame can be compared with the target block in the previous image frame.
- the feature can be a color histogram, a color histogram and a texture feature, and the comparison method can use the Euclidean distance. There are three outcomes for comparing target blocks. The three results are described below.
- the current image frame includes at least one target block that does not exist in the previous image frame.
- the terminal may determine the at least one target block that does not exist as the newly added target block, and set the new target block.
- the flag information is identified.
- the terminal may set the flag information of the newly added target block to the number of the determined key frames plus one. Assuming that s key frames have been determined before the kth image frame, the flag information of the newly added target block in the kth image frame may be set to s+1; if the kth image frame is not a key frame, The flag bit of the newly added target block in the k+1th image frame can still be set to s+1. That is, the flag bit information of the newly added target block in the image frame between the sth key frame and the s+1th key frame is s+1.
- the terminal may further obtain information such as a location and an identifier of the target block, and determine information such as the identifier, the flag bit information, and the location as the target block information of the newly added target block.
- the current image frame includes at least one target block existing in the previous image frame.
- the terminal may determine the existing at least one target block as an existing target block, and acquire the movement information of the existing target block.
- the movement information may be a movement displacement of the target block in the current image frame relative to a corresponding target block in the previous image frame, and a movement of the target block in the current image frame relative to a corresponding target block in the previous image frame. The number of times and at least one of the regions to which the target block in the current image frame is moved.
- the motion information includes a motion displacement
- the coordinates of the target block in the previous image frame are (10 cm, 13 cm)
- the coordinates of the target block in the current image frame are (10 cm, 15 cm).
- the displacement of the target block is 2 cm.
- the movement information includes the number of movements
- it is determined that the number of movements of the target block in the previous image frame is 2 and the position of the target block in the current image frame is different from the position of the target block in the previous image frame, then determining The number of movements of this target block is 3.
- the movement information includes the area to which the movement is made, if the target block is located in the central area of the current image frame, it is determined that the area to which the target block is moved is the central area.
- the terminal may further obtain the identifier and flag information of the target block, and determine the identifier, the flag bit information, and the movement information as the target block information of the existing target block.
- the previous image frame includes at least one target block that does not exist in the current image frame.
- the terminal may determine the at least one target block that does not exist as the disappearing target block, and identify and mark the target block.
- the bit information and the movement information are determined as the target block information of the vanishing target block.
- step 203 if the target block information satisfies a predetermined condition, the current image frame is determined as a key frame.
- a key frame is an image frame in which the key action of an object or object moves throughout the video, and can represent the video content of a time period in the video. Since the object in the image frame can be represented by the target block, and the target block information can be used to represent the change of the target block in the adjacent two image frames, the terminal can detect whether the current image frame is a key frame according to the target block information. , thereby extracting at least one key frame from the video.
- the terminal may determine the key frame according to different target block information.
- the process of determining the key frame is illustrated by using the target block information to include the mobile information and the flag bit information respectively.
- the current image frame is determined as a key frame, including:
- the current image frame is determined as a key frame.
- the above method of determining a key frame is applicable to a scene in which the foreground object is moving and the background does not change much, or is applicable to a scene in which the background is scaled and there is no foreground target.
- the movement information may be at least one of a movement number, a movement displacement, and an area to be moved
- the movement condition corresponding to the movement information is different.
- the movement condition may be that the movement number is greater than the number threshold; when the movement information includes the movement displacement, the movement condition may be that the movement displacement is greater than the displacement threshold; when the movement information includes the moved area, the movement condition It can be the destination area.
- moving information Including at least two of the number of movements, the movement displacement, and the area to which the movement is made, the movement conditions corresponding to each of the movement information may be combined, and when each movement information in the movement information satisfies the respective movement condition, the determination is made.
- the mobile information satisfies the moving conditions.
- the movement condition is that the number of movements exceeds 4 times and the target area is the central area
- the terminal detects that the number of movements recorded in a certain movement information is 5 and the area moved to is the central area, it is determined that the movement information satisfies the movement. condition.
- the terminal may also perform statistics on the number of pieces of mobile information that satisfy the moving condition, and detect whether the number of times obtained by the statistics is greater than a predetermined threshold. If the number of times that the terminal detects the statistics is greater than the predetermined threshold, determining that the target information of the majority of the target blocks in the current image frame changes relative to the previous key frame, the current image frame may be determined as a key frame; if the terminal detects the statistics, If the number of times is less than the predetermined threshold, it is determined that the target information of a few target blocks in the current image frame has changed with respect to the previous key frame, and the current image frame may not be determined as the key frame.
- the current image frame is determined as a key frame, including:
- the current image frame is determined as a key frame.
- the above method of determining a key frame is applicable to a scene in which the background is translated at a predetermined speed and has no foreground target, or a scene in which the background is translated at a predetermined speed and the foreground target is not changed.
- the target block information further includes flag bit information
- the terminal may acquire flag bit information of the vanishing target block in the current image frame, and detect whether there is a flag corresponding to the flag bit information of the vanishing target block in each flag bit information included in the current image frame. Bit information.
- the terminal When the terminal detects that the flag information of the flag of the disappearing target block does not exist in each of the flag information included in the current image frame, it determines that the flag information completely disappears; when the terminal detects each flag included in the current image frame When there is the same flag bit information as the flag bit information of the disappearing target block in the bit information, it is determined that the flag bit information does not completely disappear.
- the current image frame is determined as a key frame, including:
- the current image frame is determined as the key frame.
- the terminal can separate the characters and the background in the current image frame, analyze the posture of the character, and divide the background.
- the target block is detected according to the target block information, and whether the image is a key frame is determined according to the analysis result and the detection result.
- the character posture information includes a facial expression or a character gesture.
- the terminal can use the ASM (Active Shape Model) algorithm or the AAM (Active Appearance Model) algorithm to locate the feature points of the face, and then extract the features of the key organs, according to the expression. Classify features.
- ASM Active Shape Model
- AAM Active Appearance Model
- the terminal detects that the expression classification of the feature in the current image frame is different from the expression classification of the feature in the previous image frame, it is determined that the current image frame is a key frame.
- the process of detecting the target block information by the terminal is the same as the above detection process, and is not described here.
- the terminal can also determine the key frame based on the sensor information. For example, when the sensor information is GPS information, the GPS information may be compared with the GPS information of the previous image frame. When the distance between the GPS information and the GPS information of the previous image frame is greater than the GPS threshold, the GPS information is determined to be corresponding.
- the current image frame is a key frame.
- step 204 the next image frame of the current image frame is determined as the current image frame, and step 202 is continued until the last image frame in the video is compared.
- the terminal In order to extract all the key frames in the video, the terminal needs to compare all the image frames, so after comparing the current image frame, the next image frame of the current image frame can be determined as the new current image frame, and the The current image frame is determined as the previous image frame of the new current image frame, and step 202 is continued until the last image frame in the video is compared as the current image frame, and the detection flow is ended.
- step 205 the determined at least one key frame is played.
- each key frame can represent the video content of a time period
- at least one determined key frame is played in the terminal at an appropriate frame rate in order, without reducing the effective information to be expressed by the video. Compress the playback time of the video to achieve a quick view of the video content and improve the video playback efficiency.
- the terminal may also store at least one key frame to convert the video into a short video.
- the video browsing method obtains target block information by comparing a current image frame in a video with a previous image frame, and the target block information is a target block and a front block included in the current image frame.
- the key frame is filtered according to the target block information, and the key frame is not determined by the boundary of the lens, which solves the problem that the uncapable video cannot be extracted from the continuous lens, and the effect of improving the video playing efficiency is achieved.
- the image frames that do not satisfy the screening condition are discarded according to the sensor information, thereby saving waste of resources caused by processing the image frames satisfying the discarding conditions. Improve the processing efficiency of image frames.
- FIG. 3 is a block diagram of a video browsing apparatus.
- the video browsing apparatus is applied to a terminal.
- the video browsing apparatus includes: an image comparison module 310, an image determining module 320, and The execution module 330 and the image playback module 340 are operated.
- the image comparison module 310 is configured to compare the current image frame in the video with the previous image frame to obtain target block information, where the target block information is included in the target block and the previous image frame included in the current image frame. Information obtained after comparing the target blocks;
- the image determining module 320 is configured to determine the current image frame as a key frame when the target block information obtained by the image comparison module 310 satisfies a predetermined condition;
- the operation execution module 330 is configured to determine the next image frame of the current image frame as the current image frame, and continue to perform the operation of comparing the current image frame in the video with the previous image frame to obtain the target block information, until the comparison is performed. Stop after the last image frame in the video;
- the image playing module 340 is configured to play at least one key frame determined by the image determining module 320.
- the video browsing apparatus obtains target block information by comparing a current image frame in a video with a previous image frame, and the target block information is a target block and a front part included in the current image frame.
- the key frame is filtered according to the target block information, and the key frame is not determined by the boundary of the lens, which solves the problem that the uncapable video cannot be extracted from the continuous lens, and the effect of improving the video playing efficiency is achieved.
- FIG. 4 is a block diagram of a video browsing apparatus.
- the video browsing apparatus is applied to a terminal.
- the video browsing apparatus includes: an image comparison module 310, an image determining module 320, and The execution module 330 and the image playback module 340 are operated.
- the image comparison module 310 is configured to compare the current image frame in the video with the previous image frame to obtain target block information, where the target block information is included in the target block and the previous image frame included in the current image frame. Information obtained after comparing the target blocks;
- the image determining module 320 is configured to determine the current image frame as a key frame when the target block information obtained by the image comparison module 310 satisfies a predetermined condition;
- the operation execution module 330 is configured to determine the next image frame of the current image frame as the current image frame, and continue to perform the operation of comparing the current image frame in the video with the previous image frame to obtain the target block information, until the comparison is performed. Stop after the last image frame in the video;
- the image playing module 340 is configured to play at least one key frame determined in the image determining module 320.
- the image comparison module 310 includes: a target comparison sub-module 311, a first determining sub-module 312, a second determining sub-module 313, and a third determining sub-module 314;
- the target comparison sub-module 311 is configured to compare the target block in the previous image frame with the target block in the current image frame;
- the first determining sub-module 312 is configured to determine, when the target comparison sub-module 311 compares at least one target block that does not exist in the previous image frame in the current image frame, each target block that does not exist is determined to be newly added.
- Target block, and Setting flag bit information for the newly added target block, obtaining target block information of each newly added target block, and the flag bit information is the number of key frames that have been determined plus one;
- the second determining sub-module 313 is configured to determine each target block that exists as an existing target block when the target comparison sub-module 311 compares at least one target block existing in the previous image frame in the current image frame. Calculating the moving information of the existing target block in the current image frame based on the corresponding target block in the previous image frame, and obtaining the target block information of each existing target block;
- the third determining sub-module 314 is configured to determine, when the target comparison sub-module 311 compares at least one target block that does not exist in the current image frame in the previous image frame, each target block that does not exist as the disappearing target block. Get the target block information for each vanishing target block.
- the image determining module 320 includes: a first reading sub-module 321, a first detecting sub-module 322, and a fourth determining sub-module 323;
- the first reading sub-module 321 is configured to read the movement information of each existing target block from the target block information of the current image frame, where the movement information includes the number of movements of the existing target block, the movement displacement, and the movement to At least one of the regions;
- the first detecting submodule 322 is configured to detect whether the number of pieces of mobile information satisfying the moving condition is greater than a predetermined threshold
- the fourth determining sub-module 323 is configured to determine the current image frame as a key frame when the first detecting sub-module 322 detects that the number of pieces of mobile information satisfying the moving condition is greater than a predetermined threshold.
- the image determining module 320 includes: a second reading submodule 324, a second detecting submodule 325, and a fifth determining submodule 326;
- the second read submodule 324 is configured to read the flag bit information of each target block from the target block information of the current image frame;
- the second detecting sub-module 325 is configured to detect whether there is the same flag bit information in the respective flag bit information as the flag bit information of the disappearing target block;
- the fifth determining sub-module 326 is configured to determine the current image frame as a key frame when the second detecting sub-module 325 detects that there is no flag information identical to the flag information of the vanishing target block.
- the video browsing device further includes: an information acquiring module 350, an image screening module 360;
- the information acquiring module 350 is configured to acquire each image frame in the video and sensor information when capturing the image frame, the sensor information including at least one of global positioning system GPS information, acceleration information, and sensitivity ISO information;
- the image filtering module 360 is configured to discard the image frame that does not satisfy the screening condition according to the sensor information acquired by the information acquiring module 350, and trigger execution to compare the current image frame in the video with the previous image frame to obtain a target block. The operation of the information.
- the image determining module 320 includes: a third detecting submodule 327, a sixth determining submodule 328;
- the third detecting sub-module 327 is configured to detect that the current image frame is relative when the character is included in the current image frame. Whether the character posture information of the previous image frame changes, and detecting whether the target block information other than the person in the current image frame satisfies a predetermined condition;
- the sixth determining sub-module 328 is configured to detect, in the third detecting sub-module 327, that the character posture information has changed, and the current image frame is in the current image frame, except that the target block information other than the person satisfies a predetermined condition. Determined as a keyframe.
- the video browsing apparatus obtains target block information by comparing a current image frame in a video with a previous image frame, and the target block information is a target block and a front part included in the current image frame.
- the key frame is filtered according to the target block information, and the key frame is not determined by the boundary of the lens, which solves the problem that the uncapable video cannot be extracted from the continuous lens, and the effect of improving the video playing efficiency is achieved.
- the image frames that do not satisfy the screening condition are discarded according to the sensor information, thereby saving waste of resources caused by processing the image frames satisfying the discarding conditions. Improve the processing efficiency of image frames.
- FIG. 5 is a block diagram of a video browsing device 500, according to an exemplary embodiment.
- device 500 can be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a gaming console, a tablet device, a medical device, a fitness device, a personal digital assistant, and the like.
- apparatus 500 can include one or more of the following components: processing component 502, memory 504, power component 506, multimedia component 508, audio component 510, input/output (I/O) interface 512, sensor component 514, And a communication component 516.
- Processing component 502 typically controls the overall operation of device 500, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations.
- Processing component 502 can include one or more processors 518 to execute instructions to perform all or part of the steps described above.
- processing component 502 can include one or more modules to facilitate interaction between component 502 and other components.
- processing component 502 can include a multimedia module to facilitate interaction between multimedia component 508 and processing component 502.
- Memory 504 is used to store various types of data to support operation at device 500. Examples of such data include instructions for any application or method operating on device 500, contact data, phone book data, messages, pictures, videos, and the like.
- the memory 504 can be implemented by any type of volatile or non-volatile storage device, or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), erasable.
- SRAM static random access memory
- EEPROM electrically erasable programmable read only memory
- EPROM Electrically erasable programmable read only memory
- PROM Programmable Read Only Memory
- ROM Read Only Memory
- Magnetic Memory Flash Memory
- Disk Disk or Optical Disk.
- Power component 506 provides power to various components of device 500.
- the power component 506 can include a power management system.
- the multimedia component 508 includes a screen between the device 500 and the user that provides an output interface.
- the screen can include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen can be implemented as a touch screen to receive input signals from the user.
- the touch panel includes one or more touch sensors to sense touches, slides, and gestures on the touch panel. The touch sensor may sense not only the boundary of the touch or sliding action, but also the duration and pressure associated with the touch or slide operation.
- the multimedia component 508 includes a front camera and/or a rear camera. When the device 500 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera can receive external multimedia data. Each front and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
- the audio component 510 is for outputting and/or inputting audio signals.
- audio component 510 includes a microphone (MIC) for receiving an external audio signal when device 500 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode.
- the received audio signal may be further stored in memory 504 or transmitted via communication component 516.
- audio component 510 also includes a speaker for outputting an audio signal.
- the I/O interface 512 provides an interface between the processing component 502 and the peripheral interface module, which may be a keyboard, a click wheel, a button, or the like. These buttons may include, but are not limited to, a home button, a volume button, a start button, and a lock button.
- Sensor assembly 514 includes one or more sensors for providing device 500 with various aspects of status assessment.
- sensor assembly 514 can detect an open/closed state of device 500, a relative positioning of components, such as the display and keypad of device 500, and sensor component 514 can also detect a change in position of one component of device 500 or device 500. The presence or absence of user contact with device 500, device 500 orientation or acceleration/deceleration, and temperature variation of device 500.
- Sensor assembly 514 can include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
- Sensor assembly 514 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
- the sensor component 514 can also include an acceleration sensor, a gyro sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
- Communication component 516 is used to facilitate wired or wireless communication between device 500 and other devices.
- the device 500 can access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof.
- communication component 516 receives broadcast signals or broadcast associated information from an external broadcast management system via a broadcast channel.
- the communication component 516 also includes a near field communication (NFC) module to facilitate short range communication.
- NFC near field communication
- the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
- RFID radio frequency identification
- IrDA infrared data association
- UWB ultra-wideband
- Bluetooth Bluetooth
- apparatus 500 may be implemented by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable A gate array (FPGA), controller, microcontroller, microprocessor, or other electronic component implementation for performing the above methods.
- ASICs application specific integrated circuits
- DSPs digital signal processors
- DSPDs digital signal processing devices
- PLDs programmable logic devices
- FPGA field programmable A gate array
- controller microcontroller, microprocessor, or other electronic component implementation for performing the above methods.
- non-transitory computer readable storage medium comprising instructions, such as a package
- a memory 504 of instructions is executable by processor 518 of apparatus 500 to perform the above method.
- the non-transitory computer readable storage medium may be a ROM, a random access memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, and an optical data storage device.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Social Psychology (AREA)
- Computer Networks & Wireless Communication (AREA)
- Image Analysis (AREA)
- Television Signal Processing For Recording (AREA)
- Studio Devices (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Computer Security & Cryptography (AREA)
Abstract
Description
Claims (13)
- 一种视频浏览方法,其特征在于,包括:将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,所述目标块信息是对所述当前图像帧中包括的目标块和所述前一个图像帧中包括的目标块进行比较后得到的信息;若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧;将所述当前图像帧的后一个图像帧确定为当前图像帧,继续执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤,直至比较完所述视频中的最后一个图像帧后停止;对确定的至少一个关键帧进行播放。
- 根据权利要求1所述的方法,其特征在于,所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,包括:比较所述前一个图像帧中的目标块与所述当前图像帧中的目标块;若所述当前图像帧中包括所述前一个图像帧中不存在的至少一个目标块,则将不存在的每个目标块确定为新增目标块,并为所述新增目标块设置标志位信息,得到每个新增目标块的目标块信息,所述标志位信息是已经确定的关键帧的个数加1;若所述当前图像帧中包括所述前一个图像帧中存在的至少一个目标块,则将存在的每个目标块确定为已有目标块,计算所述当前图像帧中所述已有目标块基于所述前一个图像帧中对应的目标块的移动信息,得到每个已有目标块的目标块信息;若所述前一个图像帧中包括所述当前图像帧中不存在的至少一个目标块,则将不存在的每个目标块确定为消失目标块,得到每个消失目标块的目标块信息。
- 根据权利要求2所述的方法,其特征在于,所述若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧,包括:从所述当前图像帧的目标块信息中读取每个已有目标块的移动信息,所述移动信息包括所述已有目标块的移动次数、移动位移和移动至的区域中的至少一种;检测满足移动条件的移动信息的条数是否大于预定阈值;若检测满足所述移动条件的移动信息的条数大于所述预定阈值,则将所述当前图像帧确定为所述关键帧。
- 根据权利要求2所述的方法,其特征在于,所述若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧,包括:从所述当前图像帧的目标块信息中读取每个目标块的标志位信息;检测各个标志位信息中是否存在与所述消失目标块的标志位信息相同的标志位信息;若检测出不存在与所述消失目标块的标志位信息相同的标志位信息,则将所述当前图像帧确定为所述关键帧。
- 根据权利要求1至4任一项所述的方法,其特征在于,所述方法,还包括:获取所述视频中的每个图像帧以及拍摄所述图像帧时的传感器信息,所述传感器信息包括全球定位系统GPS信息、加速度信息和感光度ISO信息中的至少一种;根据所述传感器信息对不满足筛选条件的图像帧进行丢弃,触发执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤。
- 根据权利要求1所述的方法,其特征在于,所述若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧,包括:若所述当前图像帧中包括人物,则检测所述当前图像帧相对于所述前一个图像帧的人物姿态信息是否产生变化,并检测所述当前图像帧中除所述人物之外的目标块信息是否满足所述预定条件;若检测出所述人物姿态信息产生了变化,且所述当前图像帧中除所述人物之外的目标块信息满足所述预定条件,则将所述当前图像帧确定为关键帧。
- 一种视频浏览装置,其特征在于,包括:图像比较模块,被配置为将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,所述目标块信息是对所述当前图像帧中包括的目标块和所述前一个图像帧中包括的目标块进行比较后得到的信息;图像确定模块,被配置为在所述图像比较模块得到的所述目标块信息满足预定条件时,将所述当前图像帧确定为关键帧;操作执行模块,被配置为将所述当前图像帧的后一个图像帧确定为当前图像帧,继续执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的操作,直至比较完所述视频中的最后一个图像帧后停止;图像播放模块,被配置为对所述图像确定模块确定的至少一个关键帧进行播放。
- 根据权利要求7所述的装置,其特征在于,所述图像比较模块,包括:目标比较子模块,被配置为比较所述前一个图像帧中的目标块与所述当前图像帧中的目标块;第一确定子模块,被配置为在所述目标比较子模块比较出所述当前图像帧中包括所述前一个图像帧中不存在的至少一个目标块时,将不存在的每个目标块确定为新增目标块,并为所述新增目标块设置标志位信息,得到每个新增目标块的目标块信息,所述标志位信息是已经确定的关键帧的个数加1;第二确定子模块,被配置为在所述目标比较子模块比较出所述当前图像帧中包括所述前一个图像帧中存在的至少一个目标块时,将存在的每个目标块确定为已有目标块,计算所述当前图像帧中所述已有目标块基于所述前一个图像帧中对应的目标块的移动信息,得到每个已有目标块的目标块信息;第三确定子模块,被配置为在所述目标比较子模块比较出所述前一个图像帧中包括所述当前图像帧中不存在的至少一个目标块时,将不存在的每个目标块确定为消失目标块,得到每个消失目标块的目标块信息。
- 根据权利要求8所述的装置,其特征在于,所述图像确定模块,包括:第一读取子模块,被配置为从所述当前图像帧的目标块信息中读取每个已有目标块的移动信息,所述移动信息包括所述已有目标块的移动次数、移动位移和移动至的区域中的至少一种;第一检测子模块,被配置为检测满足移动条件的移动信息的条数是否大于预定阈值;第四确定子模块,被配置为在所述第一检测子模块检测出满足所述移动条件的移动信息的条数大于所述预定阈值时,将所述当前图像帧确定为所述关键帧。
- 根据权利要求8所述的装置,其特征在于,所述图像确定模块,包括:第二读取子模块,被配置为从所述当前图像帧的目标块信息中读取每个目标块的标志位信息;第二检测子模块,被配置为检测各个标志位信息中是否存在与所述消失目标块的标志位信息相同的标志位信息;第五确定子模块,被配置为在所述第二检测子模块检测出不存在与所述消失目标块的标志位信息相同的标志位信息时,将所述当前图像帧确定为所述关键帧。
- 根据权利要求7至10任一项所述的装置,其特征在于,所述装置,还包括:信息获取模块,被配置为获取所述视频中的每个图像帧以及拍摄所述图像帧时的传感器信息,所述传感器信息包括全球定位系统GPS信息、加速度信息和感光度ISO信息中的至少一种;图像筛选模块,被配置为根据所述信息获取模块获取的传感器信息对不满足筛选条件的图像帧进行丢弃,触发执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的操作。
- 根据权利要求7所述的装置,其特征在于,所述图像确定模块,包括:第三检测子模块,被配置为在所述当前图像帧中包括人物时,检测所述当前图像帧相对于所述前一个图像帧的人物姿态信息是否产生变化,并检测所述当前图像帧中除所述人 物之外的目标块信息是否满足所述预定条件;第六确定子模块,被配置为在所述第三检测子模块检测出所述人物姿态信息产生了变化,且所述当前图像帧中除所述人物之外的目标块信息满足所述预定条件时,将所述当前图像帧确定为关键帧。
- 一种视频浏览装置,其特征在于,包括:处理器;用于存储处理器可执行指令的存储器;其中,所述处理器被配置为:将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,所述目标块信息是对所述图像帧中包括的目标块和所述前一个图像帧中包括的目标块进行比较后得到的信息;若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧;将所述当前图像帧的后一个图像帧确定为当前图像帧,继续执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤,直至比较完所述视频中的最后一个图像帧后停止;对确定的至少一个关键帧进行播放。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2016548409A JP6211715B2 (ja) | 2014-09-17 | 2015-01-21 | ビデオのブラウジング方法、その装置、プログラム及び記録媒体 |
RU2015111193A RU2015111193A (ru) | 2014-09-17 | 2015-01-21 | Способ и устройство для просмотра видео |
KR1020157007200A KR101677607B1 (ko) | 2014-09-17 | 2015-01-21 | 동영상 브라우징 방법, 장치, 프로그램 및 기록매체 |
MX2015003761A MX361664B (es) | 2014-09-17 | 2015-01-21 | Metodo y dispositivo para navegar videos. |
BR112015006631A BR112015006631A2 (pt) | 2014-09-17 | 2015-01-21 | método e dispositivo para navegação de vídeo |
US14/696,691 US9799376B2 (en) | 2014-09-17 | 2015-04-27 | Method and device for video browsing based on keyframe |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410475165.4A CN104284240B (zh) | 2014-09-17 | 2014-09-17 | 视频浏览方法及装置 |
CN201410475165.4 | 2014-09-17 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/696,691 Continuation US9799376B2 (en) | 2014-09-17 | 2015-04-27 | Method and device for video browsing based on keyframe |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016041311A1 true WO2016041311A1 (zh) | 2016-03-24 |
Family
ID=52258644
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2015/071247 WO2016041311A1 (zh) | 2014-09-17 | 2015-01-21 | 视频浏览方法及装置 |
Country Status (8)
Country | Link |
---|---|
EP (1) | EP2998960B1 (zh) |
JP (1) | JP6211715B2 (zh) |
KR (1) | KR101677607B1 (zh) |
CN (1) | CN104284240B (zh) |
BR (1) | BR112015006631A2 (zh) |
MX (1) | MX361664B (zh) |
RU (1) | RU2015111193A (zh) |
WO (1) | WO2016041311A1 (zh) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113507571A (zh) * | 2021-06-30 | 2021-10-15 | 深圳市路卓科技有限公司 | 视频防剪辑方法、装置、设备、可读存储介质及程序产品 |
CN113516006A (zh) * | 2021-04-01 | 2021-10-19 | 广州云硕科技发展有限公司 | 一种用于智能机器人的高效信息处理方法及系统 |
Families Citing this family (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104284240B (zh) * | 2014-09-17 | 2018-02-02 | 小米科技有限责任公司 | 视频浏览方法及装置 |
US9799376B2 (en) | 2014-09-17 | 2017-10-24 | Xiaomi Inc. | Method and device for video browsing based on keyframe |
CN105426515B (zh) * | 2015-12-01 | 2018-12-18 | 小米科技有限责任公司 | 视频归类方法及装置 |
CN105516675B (zh) * | 2015-12-28 | 2018-11-16 | 同方威视技术股份有限公司 | 图像显示方法及用于图像显示的系统 |
KR102374747B1 (ko) | 2017-12-15 | 2022-03-15 | 삼성전자주식회사 | 객체를 인식하는 장치 및 방법 |
CN108563651B (zh) * | 2017-12-19 | 2021-05-11 | 深圳云天励飞技术有限公司 | 一种多视频的目标搜索方法、装置及设备 |
CN109993025B (zh) * | 2017-12-29 | 2021-07-06 | 中移(杭州)信息技术有限公司 | 一种关键帧提取方法及设备 |
CN108470077B (zh) * | 2018-05-28 | 2023-07-28 | 广东工业大学 | 一种视频关键帧提取方法、系统及设备和存储介质 |
CN113965814B (zh) * | 2021-08-30 | 2023-07-04 | 国网山东省电力公司信息通信公司 | 基于视频会议场景的多会场关键帧提取方法及系统 |
CN115474084B (zh) * | 2022-08-10 | 2023-10-31 | 北京奇艺世纪科技有限公司 | 一种视频封面图像的生成方法、装置、设备和存储介质 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050226331A1 (en) * | 2004-03-31 | 2005-10-13 | Honeywell International Inc. | Identifying key video frames |
CN101360184A (zh) * | 2008-09-22 | 2009-02-04 | 腾讯科技(深圳)有限公司 | 提取视频关键帧的系统及方法 |
CN101398855A (zh) * | 2008-10-24 | 2009-04-01 | 清华大学 | 一种视频关键帧提取方法和系统 |
CN101479729A (zh) * | 2006-06-29 | 2009-07-08 | 皇家飞利浦电子股份有限公司 | 提取关键帧的系统和方法 |
CN103533237A (zh) * | 2013-09-29 | 2014-01-22 | 清华大学 | 一种从视频中提取视频关键帧的方法 |
CN104284240A (zh) * | 2014-09-17 | 2015-01-14 | 小米科技有限责任公司 | 视频浏览方法及装置 |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6055025A (en) * | 1993-12-21 | 2000-04-25 | Lucent Technologies, Inc. | Method and apparatus for detecting abrupt and gradual scene changes in image sequences |
JP2914170B2 (ja) * | 1994-04-18 | 1999-06-28 | 松下電器産業株式会社 | 映像変化点検出方法 |
JP3528524B2 (ja) * | 1997-07-10 | 2004-05-17 | ソニー株式会社 | 記録再生装置および記録再生方法、並びに記録媒体 |
US5956026A (en) * | 1997-12-19 | 1999-09-21 | Sharp Laboratories Of America, Inc. | Method for hierarchical summarization and browsing of digital video |
US7280753B2 (en) * | 2003-09-03 | 2007-10-09 | Canon Kabushiki Kaisha | Display apparatus, image processing apparatus, and image processing system |
JP2005159781A (ja) * | 2003-11-27 | 2005-06-16 | Seiko Epson Corp | 画像選択処理装置、画像選択方法およびそのプログラム、記録媒体 |
US20070237225A1 (en) * | 2006-03-30 | 2007-10-11 | Eastman Kodak Company | Method for enabling preview of video files |
CN101184226A (zh) * | 2006-11-14 | 2008-05-21 | 腾讯科技(深圳)有限公司 | 一种插入视频关键帧的方法及装置 |
JP4433027B2 (ja) * | 2007-09-21 | 2010-03-17 | ソニー株式会社 | 信号処理装置、信号処理方法、およびプログラム |
JP5630541B2 (ja) * | 2009-01-20 | 2014-11-26 | ソニー株式会社 | 情報処理装置、情報処理方法およびプログラム |
JP5393237B2 (ja) * | 2009-04-24 | 2014-01-22 | キヤノン株式会社 | 画像表示装置及びその制御方法、並びにプログラム及び記憶媒体 |
WO2011099299A1 (ja) * | 2010-02-10 | 2011-08-18 | パナソニック株式会社 | 映像抽出装置、撮影装置、プログラム及び記録媒体 |
JP5537285B2 (ja) * | 2010-06-28 | 2014-07-02 | 日本放送協会 | 要約映像生成装置及び要約映像生成プログラム |
CN102348115B (zh) * | 2010-08-02 | 2014-04-16 | 南京壹进制信息技术有限公司 | 一种去除视频中冗余图像的方法和装置 |
CN109542849B (zh) * | 2012-09-16 | 2021-09-24 | 吴东辉 | 一种图像文件格式及生成方法及装置及应用 |
CN103345492A (zh) * | 2013-06-25 | 2013-10-09 | 无锡赛思汇智科技有限公司 | 一种视频浓缩的方法和系统 |
-
2014
- 2014-09-17 CN CN201410475165.4A patent/CN104284240B/zh active Active
-
2015
- 2015-01-21 JP JP2016548409A patent/JP6211715B2/ja active Active
- 2015-01-21 MX MX2015003761A patent/MX361664B/es active IP Right Grant
- 2015-01-21 RU RU2015111193A patent/RU2015111193A/ru unknown
- 2015-01-21 KR KR1020157007200A patent/KR101677607B1/ko active IP Right Grant
- 2015-01-21 WO PCT/CN2015/071247 patent/WO2016041311A1/zh active Application Filing
- 2015-01-21 BR BR112015006631A patent/BR112015006631A2/pt not_active IP Right Cessation
- 2015-09-17 EP EP15185682.0A patent/EP2998960B1/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050226331A1 (en) * | 2004-03-31 | 2005-10-13 | Honeywell International Inc. | Identifying key video frames |
CN101479729A (zh) * | 2006-06-29 | 2009-07-08 | 皇家飞利浦电子股份有限公司 | 提取关键帧的系统和方法 |
CN101360184A (zh) * | 2008-09-22 | 2009-02-04 | 腾讯科技(深圳)有限公司 | 提取视频关键帧的系统及方法 |
CN101398855A (zh) * | 2008-10-24 | 2009-04-01 | 清华大学 | 一种视频关键帧提取方法和系统 |
CN103533237A (zh) * | 2013-09-29 | 2014-01-22 | 清华大学 | 一种从视频中提取视频关键帧的方法 |
CN104284240A (zh) * | 2014-09-17 | 2015-01-14 | 小米科技有限责任公司 | 视频浏览方法及装置 |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113516006A (zh) * | 2021-04-01 | 2021-10-19 | 广州云硕科技发展有限公司 | 一种用于智能机器人的高效信息处理方法及系统 |
CN113507571A (zh) * | 2021-06-30 | 2021-10-15 | 深圳市路卓科技有限公司 | 视频防剪辑方法、装置、设备、可读存储介质及程序产品 |
CN113507571B (zh) * | 2021-06-30 | 2023-10-24 | 深圳市路卓科技有限公司 | 视频防剪辑方法、装置、设备、可读存储介质 |
Also Published As
Publication number | Publication date |
---|---|
EP2998960B1 (en) | 2020-06-17 |
MX2015003761A (es) | 2016-08-01 |
JP2016541214A (ja) | 2016-12-28 |
CN104284240B (zh) | 2018-02-02 |
BR112015006631A2 (pt) | 2017-07-04 |
RU2015111193A (ru) | 2016-10-20 |
EP2998960A1 (en) | 2016-03-23 |
MX361664B (es) | 2018-12-13 |
KR101677607B1 (ko) | 2016-11-29 |
KR20160043523A (ko) | 2016-04-21 |
JP6211715B2 (ja) | 2017-10-11 |
CN104284240A (zh) | 2015-01-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2016041311A1 (zh) | 视频浏览方法及装置 | |
US20200077035A1 (en) | Video recording method and apparatus, electronic device and readable storage medium | |
WO2016029641A1 (zh) | 照片获取方法及装置 | |
CN106375674B (zh) | 寻找和使用与相邻静态图像相关的视频部分的方法和装置 | |
CN105072337B (zh) | 图片处理方法及装置 | |
JP6335289B2 (ja) | 画像フィルタを生成する方法及び装置 | |
TWI702544B (zh) | 圖像處理方法、電子設備和電腦可讀儲存介質 | |
EP3136391B1 (en) | Method, device and terminal device for video effect processing | |
WO2017071085A1 (zh) | 报警方法及装置 | |
US10115019B2 (en) | Video categorization method and apparatus, and storage medium | |
CN105069083B (zh) | 关联用户的确定方法及装置 | |
WO2017088470A1 (zh) | 图像分类方法及装置 | |
WO2017124773A1 (zh) | 手势识别方法及装置 | |
WO2020259073A1 (zh) | 图像处理方法及装置、电子设备和存储介质 | |
EP3327590A1 (en) | Method and device for adjusting video playback position | |
US20170154206A1 (en) | Image processing method and apparatus | |
US9799376B2 (en) | Method and device for video browsing based on keyframe | |
KR102545408B1 (ko) | 차폐된 이미지 검출 방법, 장치 및 매체 | |
CN105631803B (zh) | 滤镜处理的方法和装置 | |
WO2018095252A1 (zh) | 视频录制方法及装置 | |
CN105528078A (zh) | 控制电子设备的方法及装置 | |
WO2017143778A1 (zh) | 照片的自动获取方法及装置、电子设备 | |
CN110266914A (zh) | 图像拍摄方法、装置及计算机可读存储介质 | |
WO2022198821A1 (zh) | 人脸和人体匹配的方法、装置、电子设备、存储介质及程序 | |
CN114598923A (zh) | 视频文字清除方法、装置、电子设备及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2016548409 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20157007200 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/A/2015/003761 Country of ref document: MX |
|
ENP | Entry into the national phase |
Ref document number: 2015111193 Country of ref document: RU Kind code of ref document: A |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112015006631 Country of ref document: BR |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15842239 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 112015006631 Country of ref document: BR Kind code of ref document: A2 Effective date: 20150325 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15842239 Country of ref document: EP Kind code of ref document: A1 |