Nothing Special   »   [go: up one dir, main page]

WO2016041311A1 - 视频浏览方法及装置 - Google Patents

视频浏览方法及装置 Download PDF

Info

Publication number
WO2016041311A1
WO2016041311A1 PCT/CN2015/071247 CN2015071247W WO2016041311A1 WO 2016041311 A1 WO2016041311 A1 WO 2016041311A1 CN 2015071247 W CN2015071247 W CN 2015071247W WO 2016041311 A1 WO2016041311 A1 WO 2016041311A1
Authority
WO
WIPO (PCT)
Prior art keywords
image frame
target block
information
current image
video
Prior art date
Application number
PCT/CN2015/071247
Other languages
English (en)
French (fr)
Inventor
王百超
秦秋平
侯文迪
Original Assignee
小米科技有限责任公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 小米科技有限责任公司 filed Critical 小米科技有限责任公司
Priority to JP2016548409A priority Critical patent/JP6211715B2/ja
Priority to RU2015111193A priority patent/RU2015111193A/ru
Priority to KR1020157007200A priority patent/KR101677607B1/ko
Priority to MX2015003761A priority patent/MX361664B/es
Priority to BR112015006631A priority patent/BR112015006631A2/pt
Priority to US14/696,691 priority patent/US9799376B2/en
Publication of WO2016041311A1 publication Critical patent/WO2016041311A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44004Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving video buffer management, e.g. video decoder buffer or video display buffer
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/74Browsing; Visualisation therefor
    • G06F16/745Browsing; Visualisation therefor the internal structure of a single video sequence
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/102Programmed access in sequence to addressed parts of tracks of operating record carriers
    • G11B27/105Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44213Monitoring of end-user related data
    • H04N21/44218Detecting physical presence or behaviour of the user, e.g. using sensors to detect if the user is leaving the room or changes his face expression during a TV program
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47202End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting content on demand, e.g. video on demand

Definitions

  • the present disclosure relates to the field of computer technologies, and in particular, to a video browsing method and apparatus.
  • the terminal segments the video into at least two shots during the extraction of the key frames of the video; in response to changes in image quality and video content according to the boundaries of each shot Select a set of keyframes for each shot; reproduce these keyframes on the screen at the appropriate rate for video browsing.
  • the disclosed person finds that at least the following defects exist in the related art: the terminal needs to extract a key frame according to the boundary of the lens, and when the video is a continuous lens, the key frame cannot be extracted from the video.
  • the present disclosure provides a video browsing method and apparatus.
  • a video browsing method including:
  • the target block information is performed on the target block included in the current image frame and the target block included in the previous image frame Information obtained after comparison;
  • the comparing the current image frame in the video with the previous image frame to obtain the target block information including:
  • each target block that does not exist is determined as a new target block, and a flag bit is set for the newly added target block.
  • Information get each new item Target block information of the block, the flag bit information is the number of key frames that have been determined plus one;
  • the current image frame includes at least one target block existing in the previous image frame, determining each target block that exists as an existing target block, and calculating the existing target block in the current image frame Obtaining target block information of each existing target block based on the movement information of the corresponding target block in the previous image frame;
  • each target block that does not exist is determined as a disappearing target block, and target block information of each disappearing target block is obtained.
  • determining the current image frame as a key frame includes:
  • the current image frame is determined as the key frame.
  • determining the current image frame as a key frame includes:
  • the current image frame is determined as the key frame.
  • the method further includes:
  • the sensor information including at least one of global positioning system GPS information, acceleration information, and sensitivity ISO information;
  • determining the current image frame as a key frame including:
  • the current image frame is determined as a key frame.
  • a video browsing apparatus including:
  • An image comparison module configured to compare a current image frame in the video with a previous image frame to obtain target block information, the target block information being a target block and the previous image included in the current image frame Information obtained by comparing the target blocks included in the frame;
  • An image determining module configured to satisfy the predetermined condition by the target block information obtained by the image comparison module Determining the current image frame as a key frame
  • An operation execution module configured to determine a subsequent image frame of the current image frame as a current image frame, and continue to perform the operation of comparing the current image frame in the video with the previous image frame to obtain target block information, Stops after comparing the last image frame in the video;
  • an image playing module configured to play at least one key frame determined by the image determining module.
  • the image comparison module includes:
  • a target comparison submodule configured to compare a target block in the previous image frame with a target block in the current image frame
  • a first determining submodule configured to determine, when the target comparison submodule compares at least one target block that does not exist in the previous image frame in the current image frame, determine each target block that does not exist In order to add a target block, and set flag information for the newly added target block, obtain target block information of each newly added target block, where the flag bit information is the number of key frames that have been determined plus one;
  • a second determining submodule configured to determine, when the target comparison submodule compares at least one target block existing in the previous image frame in the current image frame, to determine each target block that is present as The target block is calculated, and the existing target block in the current image frame is calculated based on the movement information of the corresponding target block in the previous image frame, and the target block information of each existing target block is obtained;
  • a third determining submodule configured to determine, when the target comparison submodule compares at least one target block that does not exist in the current image frame in the previous image frame, determine each target block that does not exist For the disappearing target block, the target block information of each disappearing target block is obtained.
  • the image determining module includes:
  • a first reading submodule configured to read movement information of each existing target block from target block information of the current image frame, where the movement information includes movement times and movement displacements of the existing target block And at least one of the areas moved to;
  • a first detecting submodule configured to detect whether the number of pieces of mobile information satisfying the moving condition is greater than a predetermined threshold
  • a fourth determining submodule configured to determine the current image frame as the key frame when the first detecting submodule detects that the number of pieces of mobile information satisfying the moving condition is greater than the predetermined threshold.
  • the image determining module includes:
  • a second reading submodule configured to read flag bit information of each target block from target block information of the current image frame
  • a second detecting submodule configured to detect whether there is the same flag bit information in the respective flag bit information as the flag bit information of the disappearing target block
  • a fifth determining submodule configured to determine the current image frame as the key frame when the second detecting submodule detects that there is no flag information identical to the flag information of the vanishing target block .
  • the device further includes:
  • An information acquisition module configured to acquire each image frame in the video and sense when the image frame is captured Device information, the sensor information including at least one of global positioning system GPS information, acceleration information, and sensitivity ISO information;
  • the image filtering module is configured to discard the image frame that does not satisfy the screening condition according to the sensor information acquired in the information acquiring module, and trigger the execution of the comparing the current image frame in the video with the previous image frame to obtain a target.
  • the operation of the block information is configured to discard the image frame that does not satisfy the screening condition according to the sensor information acquired in the information acquiring module, and trigger the execution of the comparing the current image frame in the video with the previous image frame to obtain a target.
  • the image determining module includes:
  • a third detecting submodule configured to detect, when the character is included in the current image frame, whether a change occurs in the character image information of the current image frame relative to the previous image frame, and detect the current image frame Whether the target block information other than the person satisfies the predetermined condition;
  • a sixth determining submodule configured to detect, in the third detecting submodule, that the character posture information is changed, and target block information other than the character in the current image frame satisfies the predetermined condition
  • the current image frame is determined as a key frame.
  • a video browsing apparatus including:
  • a memory for storing processor executable instructions
  • processor is configured to:
  • target block information by comparing a current image frame in the video with a previous image frame, the target block information being obtained by comparing the target block included in the current image frame with the target block included in the previous image frame.
  • Information if the target block information satisfies a predetermined condition, the current image frame is determined as a key frame, and the target block information can be obtained by comparing the target blocks of the current image frame in the video, and then the key frame is filtered according to the target block information, instead of
  • the key frame needs to be determined through the boundary of the lens, which solves the problem that the inability to browse the video caused by the key frame cannot be extracted from the continuous lens, and the effect of improving the playback efficiency of the video is achieved.
  • FIG. 1 is a flowchart of a video browsing method according to an exemplary embodiment.
  • FIG. 2 is a flowchart of a video browsing method according to another exemplary embodiment.
  • FIG. 3 is a block diagram of a video browsing apparatus, according to an exemplary embodiment.
  • FIG. 4 is a block diagram of a video browsing apparatus according to an exemplary embodiment.
  • FIG. 5 is a block diagram of an apparatus for video browsing, according to an exemplary embodiment.
  • FIG. 1 is a flowchart of a video browsing method according to an exemplary embodiment.
  • the video browsing method is applied to a terminal. As shown in FIG. 1 , the video browsing method includes the following steps.
  • step 101 the current image frame in the video is compared with the previous image frame to obtain target block information, which is performed on the target block included in the current image frame and the target block included in the previous image frame. The information obtained after comparison.
  • An image frame is a unit picture of a video.
  • the target block is obtained by dividing the content in the image frame.
  • the target block may include at least one object, and may also include a portion of the object. This embodiment does not limit the manner in which the target block is divided.
  • the target block information refers to information obtained by comparing a target block of a current image frame in a video with a target block of a previous image frame.
  • step 102 if the target block information satisfies a predetermined condition, the current image frame is determined as a key frame.
  • a key frame is an image frame in which the key action of an object or object moves throughout the video, and can represent the video content of a time period in the video.
  • the terminal can detect whether the image frame is a key frame according to the target block information, Thereby extracting at least one key frame from the video.
  • step 103 the next image frame of the current image frame is determined as the current image frame, and the step of comparing the current image frame in the video with the previous image frame to obtain the target block information is performed, until the comparison in the video is completed. Stop after the last image frame.
  • the terminal In order to extract all the key frames in the video, the terminal needs to compare all the image frames, so after comparing the current image frame, the next image frame of the current image frame can be determined as the new current image frame, and the The current image frame is determined as the previous image frame of the new current image frame, and step 101 is continued until the last image frame in the video is compared as the current image frame, and the detection flow is ended.
  • step 104 the determined at least one key frame is played.
  • each key frame can represent the video content of a time period
  • at least one determined key frame is played in the terminal at an appropriate frame rate in order, without reducing the effective information to be expressed by the video. Compress the playback time of the video to achieve a quick view of the video content and improve the video playback efficiency.
  • the video browsing method obtains target block information by comparing a current image frame in a video with a previous image frame, and the target block information is a target block included in the current image frame and the previous one.
  • the key frame is filtered according to the target block information, and the key frame is not determined by the boundary of the lens, which solves the problem that the uncapable video cannot be extracted from the continuous lens, and the effect of improving the video playing efficiency is achieved.
  • FIG. 2 is a flowchart of a video browsing method according to another exemplary embodiment.
  • the video browsing method is applied to a terminal. As shown in FIG. 2, the video browsing method includes the following steps.
  • each image frame in the video and sensor information when the image frame is captured are acquired, and the sensor information includes at least one of GPS (Global Positioning System) information, acceleration information, and sensitivity ISO information. And discarding the image frame that does not satisfy the screening condition according to the sensor information, and triggering step 202 is performed.
  • GPS Global Positioning System
  • step 201 is an optional step. That is, the terminal can capture the video through the camera, and after performing the video, step 202 is performed; or the terminal can filter the image frame in the video after the video is acquired, and discard the image frame that meets the discarding condition to save the pair.
  • the waste of the image frame that satisfies the discarding condition is wasteful, and the processing efficiency of the image frame is improved.
  • the image frame is a unit picture of the video.
  • the sensor information of each image frame may be collected by the sensor when the video is captured by the camera, and whether each image frame satisfies the discarding condition is detected according to the sensor information.
  • the sensor information may be at least one of GPS information, acceleration information, and sensitivity ISO information.
  • the terminal can set different discarding conditions according to different sensor information.
  • the sensor information is GPS information
  • the shooting scene may be shooting video during the movement. If the terminal stays in the same geographical position for a long time, the image frame captured at this time may be repeated. Therefore, the discarding condition can be set: moving
  • the distance between the GPS information and the GPS information of the previous image frame is less than the image frame of the GPS threshold.
  • the sensor information is acceleration information
  • the shooting scene may be shooting video during the movement. If the acceleration of the terminal movement is fast, the image frame captured at this time may be blurred. Therefore, the discarding condition may be set: the acceleration exceeds the acceleration threshold.
  • Image frame when the sensor information is ISO information, if the ISO is high, there are more noises in the image frame captured at this time. Therefore, it is possible to set an image frame in which the ISO exceeds the ISO threshold.
  • the terminal may combine the discarding conditions corresponding to each of the sensor information, and when at least one of the sensor information satisfies the discarding condition, determine to discard the discarding condition. Image frame.
  • step 202 the current image frame in the video is compared with the previous image frame to obtain target block information.
  • the target block information is information obtained by comparing a target block included in the current image frame with a target block included in the previous image frame.
  • the target block is obtained by dividing the content in the image frame.
  • the target block may include at least one object, and may also include a portion of the object, which is determined according to the manner in which the image frame is divided.
  • the terminal may perform equal operation on the image frame. For example, the terminal divides the image frame into M in the length direction and divides N in the width direction to obtain M ⁇ N target blocks.
  • the terminal may divide the target block according to the smallest target of interest in the image frame, ie, each target block includes at least one minimum target.
  • the terminal may divide the target block or the like according to the experience value of the developer, and the embodiment does not limit the manner of dividing the target block.
  • the terminal may compare the target block of the current image frame with the target block of the previous image frame to obtain target block information.
  • the target block information is used to represent changes in target blocks in two adjacent image frames.
  • the current image frame in the video is compared with the previous image frame to obtain target block information, including:
  • each target block that does not exist is determined as a new target block, and flag information is set for the newly added target block, and each Target block information of the newly added target block, the flag bit information is the number of key frames that have been determined plus one;
  • each target block existing is determined as an existing target block, and the existing target block in the current image frame is calculated based on the previous image frame.
  • Corresponding moving information of the target block obtaining target block information of each existing target block;
  • each target block that does not exist in the current image frame is included in the previous image frame, each target block that does not exist is determined as the disappearing target block, and the target block information of each disappearing target block is obtained.
  • the target block in the current image frame can be compared with the target block in the previous image frame.
  • the feature can be a color histogram, a color histogram and a texture feature, and the comparison method can use the Euclidean distance. There are three outcomes for comparing target blocks. The three results are described below.
  • the current image frame includes at least one target block that does not exist in the previous image frame.
  • the terminal may determine the at least one target block that does not exist as the newly added target block, and set the new target block.
  • the flag information is identified.
  • the terminal may set the flag information of the newly added target block to the number of the determined key frames plus one. Assuming that s key frames have been determined before the kth image frame, the flag information of the newly added target block in the kth image frame may be set to s+1; if the kth image frame is not a key frame, The flag bit of the newly added target block in the k+1th image frame can still be set to s+1. That is, the flag bit information of the newly added target block in the image frame between the sth key frame and the s+1th key frame is s+1.
  • the terminal may further obtain information such as a location and an identifier of the target block, and determine information such as the identifier, the flag bit information, and the location as the target block information of the newly added target block.
  • the current image frame includes at least one target block existing in the previous image frame.
  • the terminal may determine the existing at least one target block as an existing target block, and acquire the movement information of the existing target block.
  • the movement information may be a movement displacement of the target block in the current image frame relative to a corresponding target block in the previous image frame, and a movement of the target block in the current image frame relative to a corresponding target block in the previous image frame. The number of times and at least one of the regions to which the target block in the current image frame is moved.
  • the motion information includes a motion displacement
  • the coordinates of the target block in the previous image frame are (10 cm, 13 cm)
  • the coordinates of the target block in the current image frame are (10 cm, 15 cm).
  • the displacement of the target block is 2 cm.
  • the movement information includes the number of movements
  • it is determined that the number of movements of the target block in the previous image frame is 2 and the position of the target block in the current image frame is different from the position of the target block in the previous image frame, then determining The number of movements of this target block is 3.
  • the movement information includes the area to which the movement is made, if the target block is located in the central area of the current image frame, it is determined that the area to which the target block is moved is the central area.
  • the terminal may further obtain the identifier and flag information of the target block, and determine the identifier, the flag bit information, and the movement information as the target block information of the existing target block.
  • the previous image frame includes at least one target block that does not exist in the current image frame.
  • the terminal may determine the at least one target block that does not exist as the disappearing target block, and identify and mark the target block.
  • the bit information and the movement information are determined as the target block information of the vanishing target block.
  • step 203 if the target block information satisfies a predetermined condition, the current image frame is determined as a key frame.
  • a key frame is an image frame in which the key action of an object or object moves throughout the video, and can represent the video content of a time period in the video. Since the object in the image frame can be represented by the target block, and the target block information can be used to represent the change of the target block in the adjacent two image frames, the terminal can detect whether the current image frame is a key frame according to the target block information. , thereby extracting at least one key frame from the video.
  • the terminal may determine the key frame according to different target block information.
  • the process of determining the key frame is illustrated by using the target block information to include the mobile information and the flag bit information respectively.
  • the current image frame is determined as a key frame, including:
  • the current image frame is determined as a key frame.
  • the above method of determining a key frame is applicable to a scene in which the foreground object is moving and the background does not change much, or is applicable to a scene in which the background is scaled and there is no foreground target.
  • the movement information may be at least one of a movement number, a movement displacement, and an area to be moved
  • the movement condition corresponding to the movement information is different.
  • the movement condition may be that the movement number is greater than the number threshold; when the movement information includes the movement displacement, the movement condition may be that the movement displacement is greater than the displacement threshold; when the movement information includes the moved area, the movement condition It can be the destination area.
  • moving information Including at least two of the number of movements, the movement displacement, and the area to which the movement is made, the movement conditions corresponding to each of the movement information may be combined, and when each movement information in the movement information satisfies the respective movement condition, the determination is made.
  • the mobile information satisfies the moving conditions.
  • the movement condition is that the number of movements exceeds 4 times and the target area is the central area
  • the terminal detects that the number of movements recorded in a certain movement information is 5 and the area moved to is the central area, it is determined that the movement information satisfies the movement. condition.
  • the terminal may also perform statistics on the number of pieces of mobile information that satisfy the moving condition, and detect whether the number of times obtained by the statistics is greater than a predetermined threshold. If the number of times that the terminal detects the statistics is greater than the predetermined threshold, determining that the target information of the majority of the target blocks in the current image frame changes relative to the previous key frame, the current image frame may be determined as a key frame; if the terminal detects the statistics, If the number of times is less than the predetermined threshold, it is determined that the target information of a few target blocks in the current image frame has changed with respect to the previous key frame, and the current image frame may not be determined as the key frame.
  • the current image frame is determined as a key frame, including:
  • the current image frame is determined as a key frame.
  • the above method of determining a key frame is applicable to a scene in which the background is translated at a predetermined speed and has no foreground target, or a scene in which the background is translated at a predetermined speed and the foreground target is not changed.
  • the target block information further includes flag bit information
  • the terminal may acquire flag bit information of the vanishing target block in the current image frame, and detect whether there is a flag corresponding to the flag bit information of the vanishing target block in each flag bit information included in the current image frame. Bit information.
  • the terminal When the terminal detects that the flag information of the flag of the disappearing target block does not exist in each of the flag information included in the current image frame, it determines that the flag information completely disappears; when the terminal detects each flag included in the current image frame When there is the same flag bit information as the flag bit information of the disappearing target block in the bit information, it is determined that the flag bit information does not completely disappear.
  • the current image frame is determined as a key frame, including:
  • the current image frame is determined as the key frame.
  • the terminal can separate the characters and the background in the current image frame, analyze the posture of the character, and divide the background.
  • the target block is detected according to the target block information, and whether the image is a key frame is determined according to the analysis result and the detection result.
  • the character posture information includes a facial expression or a character gesture.
  • the terminal can use the ASM (Active Shape Model) algorithm or the AAM (Active Appearance Model) algorithm to locate the feature points of the face, and then extract the features of the key organs, according to the expression. Classify features.
  • ASM Active Shape Model
  • AAM Active Appearance Model
  • the terminal detects that the expression classification of the feature in the current image frame is different from the expression classification of the feature in the previous image frame, it is determined that the current image frame is a key frame.
  • the process of detecting the target block information by the terminal is the same as the above detection process, and is not described here.
  • the terminal can also determine the key frame based on the sensor information. For example, when the sensor information is GPS information, the GPS information may be compared with the GPS information of the previous image frame. When the distance between the GPS information and the GPS information of the previous image frame is greater than the GPS threshold, the GPS information is determined to be corresponding.
  • the current image frame is a key frame.
  • step 204 the next image frame of the current image frame is determined as the current image frame, and step 202 is continued until the last image frame in the video is compared.
  • the terminal In order to extract all the key frames in the video, the terminal needs to compare all the image frames, so after comparing the current image frame, the next image frame of the current image frame can be determined as the new current image frame, and the The current image frame is determined as the previous image frame of the new current image frame, and step 202 is continued until the last image frame in the video is compared as the current image frame, and the detection flow is ended.
  • step 205 the determined at least one key frame is played.
  • each key frame can represent the video content of a time period
  • at least one determined key frame is played in the terminal at an appropriate frame rate in order, without reducing the effective information to be expressed by the video. Compress the playback time of the video to achieve a quick view of the video content and improve the video playback efficiency.
  • the terminal may also store at least one key frame to convert the video into a short video.
  • the video browsing method obtains target block information by comparing a current image frame in a video with a previous image frame, and the target block information is a target block and a front block included in the current image frame.
  • the key frame is filtered according to the target block information, and the key frame is not determined by the boundary of the lens, which solves the problem that the uncapable video cannot be extracted from the continuous lens, and the effect of improving the video playing efficiency is achieved.
  • the image frames that do not satisfy the screening condition are discarded according to the sensor information, thereby saving waste of resources caused by processing the image frames satisfying the discarding conditions. Improve the processing efficiency of image frames.
  • FIG. 3 is a block diagram of a video browsing apparatus.
  • the video browsing apparatus is applied to a terminal.
  • the video browsing apparatus includes: an image comparison module 310, an image determining module 320, and The execution module 330 and the image playback module 340 are operated.
  • the image comparison module 310 is configured to compare the current image frame in the video with the previous image frame to obtain target block information, where the target block information is included in the target block and the previous image frame included in the current image frame. Information obtained after comparing the target blocks;
  • the image determining module 320 is configured to determine the current image frame as a key frame when the target block information obtained by the image comparison module 310 satisfies a predetermined condition;
  • the operation execution module 330 is configured to determine the next image frame of the current image frame as the current image frame, and continue to perform the operation of comparing the current image frame in the video with the previous image frame to obtain the target block information, until the comparison is performed. Stop after the last image frame in the video;
  • the image playing module 340 is configured to play at least one key frame determined by the image determining module 320.
  • the video browsing apparatus obtains target block information by comparing a current image frame in a video with a previous image frame, and the target block information is a target block and a front part included in the current image frame.
  • the key frame is filtered according to the target block information, and the key frame is not determined by the boundary of the lens, which solves the problem that the uncapable video cannot be extracted from the continuous lens, and the effect of improving the video playing efficiency is achieved.
  • FIG. 4 is a block diagram of a video browsing apparatus.
  • the video browsing apparatus is applied to a terminal.
  • the video browsing apparatus includes: an image comparison module 310, an image determining module 320, and The execution module 330 and the image playback module 340 are operated.
  • the image comparison module 310 is configured to compare the current image frame in the video with the previous image frame to obtain target block information, where the target block information is included in the target block and the previous image frame included in the current image frame. Information obtained after comparing the target blocks;
  • the image determining module 320 is configured to determine the current image frame as a key frame when the target block information obtained by the image comparison module 310 satisfies a predetermined condition;
  • the operation execution module 330 is configured to determine the next image frame of the current image frame as the current image frame, and continue to perform the operation of comparing the current image frame in the video with the previous image frame to obtain the target block information, until the comparison is performed. Stop after the last image frame in the video;
  • the image playing module 340 is configured to play at least one key frame determined in the image determining module 320.
  • the image comparison module 310 includes: a target comparison sub-module 311, a first determining sub-module 312, a second determining sub-module 313, and a third determining sub-module 314;
  • the target comparison sub-module 311 is configured to compare the target block in the previous image frame with the target block in the current image frame;
  • the first determining sub-module 312 is configured to determine, when the target comparison sub-module 311 compares at least one target block that does not exist in the previous image frame in the current image frame, each target block that does not exist is determined to be newly added.
  • Target block, and Setting flag bit information for the newly added target block, obtaining target block information of each newly added target block, and the flag bit information is the number of key frames that have been determined plus one;
  • the second determining sub-module 313 is configured to determine each target block that exists as an existing target block when the target comparison sub-module 311 compares at least one target block existing in the previous image frame in the current image frame. Calculating the moving information of the existing target block in the current image frame based on the corresponding target block in the previous image frame, and obtaining the target block information of each existing target block;
  • the third determining sub-module 314 is configured to determine, when the target comparison sub-module 311 compares at least one target block that does not exist in the current image frame in the previous image frame, each target block that does not exist as the disappearing target block. Get the target block information for each vanishing target block.
  • the image determining module 320 includes: a first reading sub-module 321, a first detecting sub-module 322, and a fourth determining sub-module 323;
  • the first reading sub-module 321 is configured to read the movement information of each existing target block from the target block information of the current image frame, where the movement information includes the number of movements of the existing target block, the movement displacement, and the movement to At least one of the regions;
  • the first detecting submodule 322 is configured to detect whether the number of pieces of mobile information satisfying the moving condition is greater than a predetermined threshold
  • the fourth determining sub-module 323 is configured to determine the current image frame as a key frame when the first detecting sub-module 322 detects that the number of pieces of mobile information satisfying the moving condition is greater than a predetermined threshold.
  • the image determining module 320 includes: a second reading submodule 324, a second detecting submodule 325, and a fifth determining submodule 326;
  • the second read submodule 324 is configured to read the flag bit information of each target block from the target block information of the current image frame;
  • the second detecting sub-module 325 is configured to detect whether there is the same flag bit information in the respective flag bit information as the flag bit information of the disappearing target block;
  • the fifth determining sub-module 326 is configured to determine the current image frame as a key frame when the second detecting sub-module 325 detects that there is no flag information identical to the flag information of the vanishing target block.
  • the video browsing device further includes: an information acquiring module 350, an image screening module 360;
  • the information acquiring module 350 is configured to acquire each image frame in the video and sensor information when capturing the image frame, the sensor information including at least one of global positioning system GPS information, acceleration information, and sensitivity ISO information;
  • the image filtering module 360 is configured to discard the image frame that does not satisfy the screening condition according to the sensor information acquired by the information acquiring module 350, and trigger execution to compare the current image frame in the video with the previous image frame to obtain a target block. The operation of the information.
  • the image determining module 320 includes: a third detecting submodule 327, a sixth determining submodule 328;
  • the third detecting sub-module 327 is configured to detect that the current image frame is relative when the character is included in the current image frame. Whether the character posture information of the previous image frame changes, and detecting whether the target block information other than the person in the current image frame satisfies a predetermined condition;
  • the sixth determining sub-module 328 is configured to detect, in the third detecting sub-module 327, that the character posture information has changed, and the current image frame is in the current image frame, except that the target block information other than the person satisfies a predetermined condition. Determined as a keyframe.
  • the video browsing apparatus obtains target block information by comparing a current image frame in a video with a previous image frame, and the target block information is a target block and a front part included in the current image frame.
  • the key frame is filtered according to the target block information, and the key frame is not determined by the boundary of the lens, which solves the problem that the uncapable video cannot be extracted from the continuous lens, and the effect of improving the video playing efficiency is achieved.
  • the image frames that do not satisfy the screening condition are discarded according to the sensor information, thereby saving waste of resources caused by processing the image frames satisfying the discarding conditions. Improve the processing efficiency of image frames.
  • FIG. 5 is a block diagram of a video browsing device 500, according to an exemplary embodiment.
  • device 500 can be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a gaming console, a tablet device, a medical device, a fitness device, a personal digital assistant, and the like.
  • apparatus 500 can include one or more of the following components: processing component 502, memory 504, power component 506, multimedia component 508, audio component 510, input/output (I/O) interface 512, sensor component 514, And a communication component 516.
  • Processing component 502 typically controls the overall operation of device 500, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations.
  • Processing component 502 can include one or more processors 518 to execute instructions to perform all or part of the steps described above.
  • processing component 502 can include one or more modules to facilitate interaction between component 502 and other components.
  • processing component 502 can include a multimedia module to facilitate interaction between multimedia component 508 and processing component 502.
  • Memory 504 is used to store various types of data to support operation at device 500. Examples of such data include instructions for any application or method operating on device 500, contact data, phone book data, messages, pictures, videos, and the like.
  • the memory 504 can be implemented by any type of volatile or non-volatile storage device, or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), erasable.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read only memory
  • EPROM Electrically erasable programmable read only memory
  • PROM Programmable Read Only Memory
  • ROM Read Only Memory
  • Magnetic Memory Flash Memory
  • Disk Disk or Optical Disk.
  • Power component 506 provides power to various components of device 500.
  • the power component 506 can include a power management system.
  • the multimedia component 508 includes a screen between the device 500 and the user that provides an output interface.
  • the screen can include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen can be implemented as a touch screen to receive input signals from the user.
  • the touch panel includes one or more touch sensors to sense touches, slides, and gestures on the touch panel. The touch sensor may sense not only the boundary of the touch or sliding action, but also the duration and pressure associated with the touch or slide operation.
  • the multimedia component 508 includes a front camera and/or a rear camera. When the device 500 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera can receive external multimedia data. Each front and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
  • the audio component 510 is for outputting and/or inputting audio signals.
  • audio component 510 includes a microphone (MIC) for receiving an external audio signal when device 500 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode.
  • the received audio signal may be further stored in memory 504 or transmitted via communication component 516.
  • audio component 510 also includes a speaker for outputting an audio signal.
  • the I/O interface 512 provides an interface between the processing component 502 and the peripheral interface module, which may be a keyboard, a click wheel, a button, or the like. These buttons may include, but are not limited to, a home button, a volume button, a start button, and a lock button.
  • Sensor assembly 514 includes one or more sensors for providing device 500 with various aspects of status assessment.
  • sensor assembly 514 can detect an open/closed state of device 500, a relative positioning of components, such as the display and keypad of device 500, and sensor component 514 can also detect a change in position of one component of device 500 or device 500. The presence or absence of user contact with device 500, device 500 orientation or acceleration/deceleration, and temperature variation of device 500.
  • Sensor assembly 514 can include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
  • Sensor assembly 514 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
  • the sensor component 514 can also include an acceleration sensor, a gyro sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
  • Communication component 516 is used to facilitate wired or wireless communication between device 500 and other devices.
  • the device 500 can access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof.
  • communication component 516 receives broadcast signals or broadcast associated information from an external broadcast management system via a broadcast channel.
  • the communication component 516 also includes a near field communication (NFC) module to facilitate short range communication.
  • NFC near field communication
  • the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
  • RFID radio frequency identification
  • IrDA infrared data association
  • UWB ultra-wideband
  • Bluetooth Bluetooth
  • apparatus 500 may be implemented by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable A gate array (FPGA), controller, microcontroller, microprocessor, or other electronic component implementation for performing the above methods.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA field programmable A gate array
  • controller microcontroller, microprocessor, or other electronic component implementation for performing the above methods.
  • non-transitory computer readable storage medium comprising instructions, such as a package
  • a memory 504 of instructions is executable by processor 518 of apparatus 500 to perform the above method.
  • the non-transitory computer readable storage medium may be a ROM, a random access memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, and an optical data storage device.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Social Psychology (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Image Analysis (AREA)
  • Television Signal Processing For Recording (AREA)
  • Studio Devices (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Computer Security & Cryptography (AREA)

Abstract

本公开关于一种视频浏览方法及装置,属于计算机技术领域。所述方法包括:将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息;若目标块信息满足预定条件,则将当前图像帧确定为关键帧;将当前图像帧的后一个图像帧确定为当前图像帧,继续执行将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤,直至比较完视频中的最后一个图像帧后停止;对确定的至少一个关键帧进行播放。所述装置包括:图像比较模块、图像确定模块、操作执行模块和图像播放模块。本公开解决了无法从连续镜头的视频中提取关键帧导致的无法浏览视频的问题,达到了提高视频的播放效率的效果。

Description

视频浏览方法及装置
本申请基于申请号为201410475165.4、申请日为2014年9月17日的中国专利申请提出,并要求该中国专利申请的优先权,该中国专利申请的全部内容在此引入本申请作为参考。
技术领域
本公开涉及计算机技术领域,特别涉及一种视频浏览方法及装置。
背景技术
随着互联网的发展,网络视频被越来越频繁的使用。当我们获取到一段视频时,需要快速浏览视频来了解视频内容,因此,如何对视频进行快速浏览成为了人们关注的问题。
由于视频是由连续的图像帧组成的,因此,在提取视频的关键帧的过程中,终端将视频分段为至少两个镜头;根据每个镜头的边界,响应于图像质量和视频内容的改变为每个镜头选择一组关键帧;将这些关键帧以恰当的速率在屏幕上再现,实现视频浏览。
公开人在实现本公开的过程中,发现相关技术中至少存在以下缺陷:终端需要根据镜头的边界提取关键帧,当视频是一个连续的镜头时,无法从视频中提取关键帧。
发明内容
为解决无法从连续镜头的视频中提取关键帧导致的无法浏览视频的问题,本公开提供了一种视频浏览方法及装置。
根据本公开实施例的第一方面,提供一种视频浏览方法,包括:
将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,所述目标块信息是对所述当前图像帧中包括的目标块和所述前一个图像帧中包括的目标块进行比较后得到的信息;
若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧;
将所述当前图像帧的后一个图像帧确定为当前图像帧,继续执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤,直至比较完所述视频中的最后一个图像帧后停止;
对确定的至少一个关键帧进行播放。
可选的,所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,包括:
比较所述前一个图像帧中的目标块与所述当前图像帧中的目标块;
若所述当前图像帧中包括所述前一个图像帧中不存在的至少一个目标块,则将不存在的每个目标块确定为新增目标块,并为所述新增目标块设置标志位信息,得到每个新增目 标块的目标块信息,所述标志位信息是已经确定的关键帧的个数加1;
若所述当前图像帧中包括所述前一个图像帧中存在的至少一个目标块,则将存在的每个目标块确定为已有目标块,计算所述当前图像帧中所述已有目标块基于所述前一个图像帧中对应的目标块的移动信息,得到每个已有目标块的目标块信息;
若所述前一个图像帧中包括所述当前图像帧中不存在的至少一个目标块,则将不存在的每个目标块确定为消失目标块,得到每个消失目标块的目标块信息。
可选的,所述若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧,包括:
从所述当前图像帧的目标块信息中读取每个已有目标块的移动信息,所述移动信息包括所述已有目标块的移动次数、移动位移和移动至的区域中的至少一种;
检测满足移动条件的移动信息的条数是否大于预定阈值;
若检测满足所述移动条件的移动信息的条数大于所述预定阈值,则将所述当前图像帧确定为所述关键帧。
可选的,所述若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧,包括:
从所述当前图像帧的目标块信息中读取每个目标块的标志位信息;
检测各个标志位信息中是否存在与所述消失目标块的标志位信息相同的标志位信息;
若检测出不存在与所述消失目标块的标志位信息相同的标志位信息,则将所述当前图像帧确定为所述关键帧。
可选的,所述方法,还包括:
获取所述视频中的每个图像帧以及拍摄所述图像帧时的传感器信息,所述传感器信息包括全球定位系统GPS信息、加速度信息和感光度ISO信息中的至少一种;
根据所述传感器信息对不满足筛选条件的图像帧进行丢弃,触发执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤。
可选的,若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧,包括:
若所述当前图像帧中包括人物,则检测所述当前图像帧相对于所述前一个图像帧的人物姿态信息是否产生变化,并检测所述当前图像帧中除所述人物之外的目标块信息是否满足所述预定条件;
若检测出所述人物姿态信息产生了变化,且所述当前图像帧中除所述人物之外的目标块信息满足所述预定条件,则将所述当前图像帧确定为关键帧。
根据本公开实施例的第二方面,提供一种视频浏览装置,包括:
图像比较模块,被配置为将视频中的当前图像帧与前一个图像帧进行比较,得到目标块信息,所述目标块信息是对所述当前图像帧中包括的目标块和所述前一个图像帧中包括的目标块进行比较后得到的信息;
图像确定模块,被配置为在所述图像比较模块得到的所述目标块信息满足预定条件 时,将所述当前图像帧确定为关键帧;
操作执行模块,被配置为将所述当前图像帧的后一个图像帧确定为当前图像帧,继续执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的操作,直至比较完所述视频中的最后一个图像帧后停止;
图像播放模块,被配置为对所述图像确定模块确定的至少一个关键帧进行播放。
可选的,所述图像比较模块,包括:
目标比较子模块,被配置为比较所述前一个图像帧中的目标块与所述当前图像帧中的目标块;
第一确定子模块,被配置为在所述目标比较子模块比较出所述当前图像帧中包括所述前一个图像帧中不存在的至少一个目标块时,将不存在的每个目标块确定为新增目标块,并为所述新增目标块设置标志位信息,得到每个新增目标块的目标块信息,所述标志位信息是已经确定的关键帧的个数加1;
第二确定子模块,被配置为在所述目标比较子模块比较出所述当前图像帧中包括所述前一个图像帧中存在的至少一个目标块时,将存在的每个目标块确定为已有目标块,计算所述当前图像帧中所述已有目标块基于所述前一个图像帧中对应的目标块的移动信息,得到每个已有目标块的目标块信息;
第三确定子模块,被配置为在所述目标比较子模块比较出所述前一个图像帧中包括所述当前图像帧中不存在的至少一个目标块时,将不存在的每个目标块确定为消失目标块,得到每个消失目标块的目标块信息。
可选的,所述图像确定模块,包括:
第一读取子模块,被配置为从所述当前图像帧的目标块信息中读取每个已有目标块的移动信息,所述移动信息包括所述已有目标块的移动次数、移动位移和移动至的区域中的至少一种;
第一检测子模块,被配置为检测满足移动条件的移动信息的条数是否大于预定阈值;
第四确定子模块,被配置为在所述第一检测子模块检测出满足所述移动条件的移动信息的条数大于所述预定阈值时,将所述当前图像帧确定为所述关键帧。
可选的,所述图像确定模块,包括:
第二读取子模块,被配置为从所述当前图像帧的目标块信息中读取每个目标块的标志位信息;
第二检测子模块,被配置为检测各个标志位信息中是否存在与所述消失目标块的标志位信息相同的标志位信息;
第五确定子模块,被配置为在所述第二检测子模块检测出不存在与所述消失目标块的标志位信息相同的标志位信息时,将所述当前图像帧确定为所述关键帧。
可选的,所述装置,还包括:
信息获取模块,被配置为获取所述视频中的每个图像帧以及拍摄所述图像帧时的传感 器信息,所述传感器信息包括全球定位系统GPS信息、加速度信息和感光度ISO信息中的至少一种;
图像筛选模块,被配置为根据所述信息获取模块中获取的传感器信息对不满足筛选条件的图像帧进行丢弃,触发执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的操作。
可选的,所述图像确定模块,包括:
第三检测子模块,被配置为在所述当前图像帧中包括人物时,检测所述当前图像帧相对于所述前一个图像帧的人物姿态信息是否产生变化,并检测所述当前图像帧中除所述人物之外的目标块信息是否满足所述预定条件;
第六确定子模块,被配置为在所述第三检测子模块检测出所述人物姿态信息产生了变化,且所述当前图像帧中除所述人物之外的目标块信息满足所述预定条件时,将所述当前图像帧确定为关键帧。
根据本公开实施例的第三方面,提供一种视频浏览装置,包括:
处理器;
用于存储处理器可执行指令的存储器;
其中,所述处理器被配置为:
将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,所述目标块信息是对所述图像帧中包括的目标块和所述前一个图像帧中包括的目标块进行比较后得到的信息;
若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧;
将所述当前图像帧的后一个图像帧确定为当前图像帧,继续执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤,直至比较完所述视频中的最后一个图像帧后停止;
对确定的至少一个关键帧进行播放。
本公开的实施例提供的技术方案可以包括以下有益效果:
通过将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,该目标块信息是对当前图像帧中包括的目标块和前一个图像帧中包括的目标块进行比较后得到的信息;若目标块信息满足预定条件,则将当前图像帧确定为关键帧,可以通过对视频中当前图像帧的目标块进行比较得到目标块信息,再根据目标块信息筛选出关键帧,而不需要通过镜头的边界确定关键帧,解决了无法从连续的镜头中提取关键帧导致的无法浏览视频的问题,达到了提高视频的播放效率的效果。
应当理解的是,以上的一般描述和后文的细节描述仅是示例性的,并不能限制本公开。
附图说明
此处的附图被并入说明书中并构成本说明书的一部分,示出了符合本公开的实施例, 并于说明书一起用于解释本公开的原理。
图1是根据一示例性实施例示出的一种视频浏览方法的流程图。
图2是根据另一示例性实施例示出的一种视频浏览方法的流程图。
图3是根据一示例性实施例示出的一种视频浏览装置的框图。
图4是根据一示例性实施例示出的一种视频浏览装置的框图。
图5是根据一示例性实施例示出的一种用于视频浏览的装置的框图。
具体实施方式
这里将详细地对示例性实施例进行说明,其示例表示在附图中。下面的描述涉及附图时,除非另有表示,不同附图中的相同数字表示相同或相似的要素。以下示例性实施例中所描述的实施方式并不代表与本公开相一致的所有实施方式。相反,它们仅是与如所附权利要求书中所详述的、本公开的一些方面相一致的装置和方法的例子。
图1是根据一示例性实施例示出的一种视频浏览方法的流程图,该视频浏览方法应用于终端中,如图1所示,该视频浏览方法包括以下步骤。
在步骤101中,将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,该目标块信息是对当前图像帧中包括的目标块和前一个图像帧中包括的目标块进行比较后得到的信息。
图像帧是视频的单位画面。目标块是对图像帧中的内容进行划分后得到的,目标块可以包括至少一个物体,也可以包括一个物体的部分,本实施例不限定对目标块的划分方式。目标块信息指将视频中的当前图像帧的目标块与其前一个图像帧的目标块进行比较后得到的信息。
在步骤102中,若目标块信息满足预定条件,则将当前图像帧确定为关键帧。
关键帧是指在整个视频中物体或物体运动中关键动作所在的一个图像帧,可以代表视频中一个时间段的视频内容。
由于图像帧中的物体可以用目标块表示,且目标块信息可以用来表示相邻两个图像帧中目标块的变化,因此,终端可以根据目标块信息对图像帧是否是关键帧进行检测,从而从视频中提取出至少一个关键帧。
在步骤103中,将当前图像帧的后一个图像帧确定为当前图像帧,继续执行将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤,直至比较完视频中的最后一个图像帧后停止。
为了提取到视频中的所有关键帧,终端需要对所有图像帧进行比较,因此,在比较完当前图像帧之后,可以将当前图像帧的后一个图像帧确定为新的当前图像帧,并将该当前图像帧确定为新的当前图像帧的前一个图像帧,继续执行步骤101,直至将视频中的最后一个图像帧作为当前图像帧进行比较后,结束检测流程。
在步骤104中,对确定的至少一个关键帧进行播放。
由于每一个关键帧都可以代表一个时间段的视频内容,因此,将确定的至少一个关键帧按照顺序以适当的帧速度在终端中播放,可以在不缩减视频所要表达的有效信息的前提下,对视频的播放时长进行压缩,实现对视频内容的快速浏览,提高了视频播放效率。
综上所述,本公开提供的视频浏览方法,通过将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,目标块信息是对当前图像帧中包括的目标块和前一个图像帧中包括的目标块进行比较后得到的信息;若目标块信息满足预定条件,则将当前图像帧确定为关键帧,可以通过对视频中当前图像帧的目标块进行比较得到目标块信息,再根据目标块信息筛选出关键帧,而不需要通过镜头的边界确定关键帧,解决了无法从连续的镜头中提取关键帧导致的无法浏览视频的问题,达到了提高视频的播放效率的效果。
图2是根据另一示例性实施例示出的一种视频浏览方法的流程图,该视频浏览方法应用于终端中,如图2所示,该视频浏览方法包括如下步骤。
在步骤201中,获取视频中的每个图像帧以及拍摄图像帧时的传感器信息,该传感器信息包括GPS(Global Positioning System,全球定位系统)信息、加速度信息和感光度ISO信息中的至少一种,根据传感器信息对不满足筛选条件的图像帧进行丢弃,触发执行步骤202。
本实施例中,步骤201是可选步骤。即,终端可以通过摄像头拍摄视频,在获取到视频后执行步骤202;或,终端可以在获取到视频后,对视频中的图像帧进行筛选,对满足丢弃条件的图像帧进行丢弃,以节省对满足丢弃条件的图像帧进行处理造成的资源浪费,提高对图像帧的处理效率。其中,图像帧是视频的单位画面。
若终端需要对图像帧进行筛选,则可以在通过摄像头拍摄视频时,通过传感器采集每个图像帧的传感器信息,根据传感器信息检测每个图像帧是否满足丢弃条件。其中,传感器信息可以是GPS信息、加速度信息和感光度ISO信息中的至少一种。
其中,终端可以根据不同的传感器信息设置不同的丢弃条件。当传感器信息是GPS信息时,拍摄场景可以是在移动过程中拍摄视频,若终端长时间停留在同一个地理位置,此时拍摄的图像帧可能会重复,因此,可以设置丢弃条件是:在移动过程中拍摄视频时,GPS信息与前一个图像帧的GPS信息之间的距离小于GPS阈值的图像帧。当传感器信息是加速度信息时,拍摄场景可以是在移动过程中拍摄视频,若终端移动的加速度较快,此时拍摄的图像帧会模糊,因此,可以设置丢弃条件是:加速度超于加速度阈值的图像帧;当传感器信息是ISO信息时,若ISO较高,此时拍摄的图像帧中的噪点较多,因此,可以设置丢弃条件是:ISO超过ISO阈值的图像帧。当传感器信息包括GPS信息、加速度信息和ISO信息中的至少两种时,终端可以将每种传感器信息各自对应的丢弃条件进行合并,在传感器信息中的至少一种满足丢弃条件时,确定丢弃该图像帧。
在步骤202中,将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息, 该目标块信息是对当前图像帧中包括的目标块和前一个图像帧中包括的目标块进行比较后得到的信息。
目标块是对图像帧中的内容进行划分后得到的,目标块可以包括至少一个物体,也可以包括一个物体的部分,是根据对图像帧的划分方式确定的。其中,终端可以对图像帧进行等分操作,比如,终端将图像帧在长度方向上M等分,在宽度方向上N等分,得到M×N个目标块。或,终端可以根据图像帧中感兴趣的最小目标划分目标块,即,每个目标块包括至少一个最小目标。或,终端可以根据开发人员的经验值来划分目标块等等,本实施例不限定对目标块的划分方式。
在得到当前图像帧的目标块后,终端可以将该当前图像帧的目标块与其前一个图像帧的目标块进行比较,得到目标块信息。该目标块信息用于表示相邻两个图像帧中目标块的变化。
本实施例中,将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,包括:
1)比较前一个图像帧中的目标块与当前图像帧中的目标块;
2)若当前图像帧中包括前一个图像帧中不存在的至少一个目标块,则将不存在的每个目标块确定为新增目标块,并为新增目标块设置标志位信息,得到每个新增目标块的目标块信息,该标志位信息是已经确定的关键帧的个数加1;
3)若当前图像帧中包括前一个图像帧中存在的至少一个目标块,则将存在的每个目标块确定为已有目标块,计算当前图像帧中已有目标块基于前一个图像帧中对应的目标块的移动信息,得到每个已有目标块的目标块信息;
4)若前一个图像帧中包括当前图像帧中不存在的至少一个目标块,则将不存在的每个目标块确定为消失目标块,得到每个消失目标块的目标块信息。
其中,当前图像帧中的目标块与前一个图像帧中的目标块可以进行特征对比,该特征可以是颜色直方图,也可以是颜色直方图和纹理特征,比对方式可以使用欧式距离。进行目标块比较可以有三种结果,下面分别对这三种结果进行描述。
第一,当前图像帧中包括前一个图像帧中不存在的至少一个目标块,此时,终端可以将该不存在的至少一个目标块确定为新增目标块,并对该新增目标块设置标志位信息进行标识。
其中,终端可以将新增目标块的标志位信息设置为已经确定的关键帧的个数加1。假设在第k个图像帧之前已经确定了s个关键帧,则可以将第k个图像帧中新增目标块的标志位信息设置为s+1;若第k个图像帧不是关键帧,则可以将第k+1个图像帧中的新增目标块的标志位仍然设置为s+1。即,在第s个关键帧和第s+1个关键帧之间的图像帧中的新增目标块的标志位信息都是s+1。
可选的,终端还可以获取目标块的位置和标识等信息,并将标识、标志位信息和位置等信息确定为该新增目标块的目标块信息。
第二,当前图像帧中包括前一个图像帧中存在的至少一个目标块,此时,终端可以将该存在的至少一个目标块确定为已有目标块,并获取已有目标块的移动信息。
其中,移动信息可以是该当前图像帧中的目标块相对于前一个图像帧中对应的目标块的移动位移、该当前图像帧中的目标块相对于前一个图像帧中对应的目标块的移动次数和该当前图像帧中的目标块移动到至的区域中的至少一种。
比如,当移动信息包括移动位移时,假设该目标块在前一个图像帧中的坐标是(10厘米,13厘米),该目标块在当前图像帧中的坐标是(10厘米,15厘米),此时该目标块的位移是2厘米。当移动信息包括移动次数时,假设该目标块在前一个图像帧中的移动次数是2且该目标块在当前图像帧中的位置与该目标块在前一个图像帧中的位置不同,则确定该目标块的移动次数是3。当移动信息包括移动至的区域时,若该目标块位于当前图像帧的中央区域,则确定该目标块移动到的区域是中央区域。
可选的,终端还可以获取该目标块的标识和标志位信息,并将标识、标志位信息和移动信息确定为该已有目标块的目标块信息。
第三,前一个图像帧中包括当前图像帧中不存在的至少一个目标块,此时,终端可以将该不存在的至少一个目标块确定为消失目标块,并将该目标块的标识、标志位信息和移动信息确定为该消失目标块的目标块信息。
在步骤203中,若目标块信息满足预定条件,则将当前图像帧确定为关键帧。
关键帧是指在整个视频中物体或物体运动中关键动作所在的一个图像帧,可以代表视频中一个时间段的视频内容。由于图像帧中的物体可以用目标块表示,且目标块信息可以用来表示相邻两个图像帧中目标块的变化,因此,终端可以根据目标块信息对当前图像帧是否是关键帧进行检测,从而从视频中提取出至少一个关键帧。
其中,终端可以根据不同的目标块信息确定关键帧,本实施例以目标块信息分别包括移动信息和标志位信息对确定关键帧的流程进行举例说明。
第一,若目标块信息满足预定条件,则将当前图像帧确定为关键帧,包括:
1)从当前图像帧的目标块信息中读取每个已有目标块的移动信息,该移动信息包括已有目标块的移动次数、移动位移和移动至的区域中的至少一种;
2)检测满足移动条件的移动信息的条数是否大于预定阈值;
3)若检测满足移动条件的移动信息的条数大于预定阈值,则将当前图像帧确定为关键帧。
上述确定关键帧的方法适用于前景目标在移动且背景变化不大的场景,或,适用于缩放背景且没有前景目标的场景。
由于移动信息可以是移动次数、移动位移和移动至的区域中的至少一种,当移动信息不同时,与移动信息所对应的移动条件不同。比如,当移动信息包括移动次数时,移动条件可以是移动次数大于次数阈值;当移动信息包括移动位移时,移动条件可以是移动位移大于位移阈值;当移动信息包括移动至的区域时,移动条件可以是目的区域。当移动信息 包括移动次数、移动位移和移动至的区域中的至少两种,可以将每种移动信息各自对应的移动条件进行合并,在移动信息中的每种移动信息均满足各自的移动条件时,确定该移动信息满足移动条件。
比如,当移动条件是移动次数超过4次且目标区域是中央区域,则当终端检测出某一个移动信息中记录的移动次数为5且移动至的区域是中央区域时,确定该移动信息满足移动条件。
在确定出移动信息满足移动条件后,终端还可以对满足移动条件的移动信息的条数进行统计,并检测统计得到的次数是否大于预定阈值。若终端检测出统计得到的次数大于预定阈值,则确定当前图像帧中多数目标块的目标信息相对于前一个关键帧发生了变化,可以将当前图像帧确定为关键帧;若终端检测出统计得到的次数小于预定阈值,则确定当前图像帧中少数目标块的目标信息相对于前一个关键帧发生了变化,可以不将当前图像帧确定为关键帧。
第二,若目标块信息满足预定条件,则将当前图像帧确定为关键帧,包括:
1)从当前图像帧的目标块信息中读取每个目标块的标志位信息;
2)检测各个标志位信息中是否存在与消失目标块的标志位信息相同的标志位信息;
3)若检测出不存在与消失目标块的标志位信息相同的标志位信息,则将当前图像帧确定为关键帧。
上述确定关键帧的方法适用于背景以预定速度平移且没有前景目标的场景,或,适用于背景以预定速度平移且前景目标没有变化的场景。
由于目标块信息还包括标志位信息,则当图像帧中某一个标志位信息完全消失时,确定该图像帧是关键帧。在检测标志位信息是否完全消失时,终端可以获取当前图像帧中的消失目标块的标志位信息,检测当前图像帧包括的各个标志位信息中是否存在与消失目标块的标志位信息相同的标志位信息。当终端检测出当前图像帧包括的各个标志位信息中不存在与消失目标块的标志位信息相同的标志位信息时,确定该标志位信息完全消失;当终端检测出当前图像帧包括的各个标志位信息中存在与消失目标块的标志位信息相同的标志位信息时,确定该标志位信息没有完全消失。
对于包括人物的图像帧,有时候需要捕捉人物细微的姿态变化,此时,若终端将人物划分成目标块进行检测,可能会由于目标块信息的变化不大导致将重要的图像帧错误的确定为非关键帧,造成重要信息的丢失。
因此,若目标块信息满足预定条件,则将当前图像帧确定为关键帧,包括:
1)若当前图像帧中包括人物,则检测当前图像帧相对于前一个图像帧的人物姿态信息是否产生变化,并检测当前图像帧中除人物之外的目标块信息是否满足预定条件;
2)若检测出人物姿态信息产生了变化,且当前图像帧中除人物之外的目标块信息满足预定条件,则将当前图像帧确定为关键帧。
终端可以将当前图像帧中的人物和背景进行分离,对人物姿态进行分析,将背景划分 成目标块,根据目标块信息进行检测,根据分析结果和检测结果确定图像是否为关键帧。其中,人物姿态信息包括人脸表情或人物姿态。
终端对人物姿态进行分析的方法有很多种,比如,终端可以采用ASM(Active Shape Model)算法或AAM(Active Appearance Model)算法进行人脸特征点定位,再对关键器官的特征进行提取,按照表情对特征进行分类。在终端检测到当前图像帧中特征的表情分类与前一个图像帧中特征的表情分类不同时,确定该当前图像帧是关键帧。终端对目标块信息的检测过程和上述检测过程相同,此处不赘述。
需要补充说明的是,终端还可以根据传感器信息确定关键帧。比如,当传感器信息是GPS信息时,可以将GPS信息与前一个图像帧的GPS信息进行比较,当GPS信息与前一个图像帧的GPS信息之间的距离大于GPS阈值时,确定该GPS信息对应的当前图像帧为关键帧。
在步骤204中,将当前图像帧的后一个图像帧确定为当前图像帧,继续执行步骤202,直至比较完视频中的最后一个图像帧后停止。
为了提取到视频中的所有关键帧,终端需要对所有图像帧进行比较,因此,在比较完当前图像帧之后,可以将当前图像帧的后一个图像帧确定为新的当前图像帧,并将该当前图像帧确定为新的当前图像帧的前一个图像帧,继续执行步骤202,直至将视频中的最后一个图像帧作为当前图像帧进行比较后,结束检测流程。
在步骤205中,对确定的至少一个关键帧进行播放。
由于每一个关键帧都可以代表一个时间段的视频内容,因此,将确定的至少一个关键帧按照顺序以适当的帧速度在终端中播放,可以在不缩减视频所要表达的有效信息的前提下,对视频的播放时长进行压缩,实现对视频内容的快速浏览,提高了视频播放效率。
可选的,终端还可以对至少一个关键帧进行存储,从而将视频转换为短视频。
综上所述,本公开提供的视频浏览方法,通过将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,该目标块信息是对当前图像帧中包括的目标块和前一个图像帧中包括的目标块进行比较后得到的信息;若目标块信息满足预定条件,则将当前图像帧确定为关键帧,可以通过对视频中图像帧的目标块进行比较得到目标块信息,再根据目标块信息筛选出关键帧,而不需要通过镜头的边界确定关键帧,解决了无法从连续的镜头中提取关键帧导致的无法浏览视频的问题,达到了提高视频的播放效率的效果。
另外,通过获取视频中的每个图像帧以及拍摄图像帧时的传感器信息,根据传感器信息对不满足筛选条件的图像帧进行丢弃,可以节省对满足丢弃条件的图像帧进行处理造成的资源浪费,提高对图像帧的处理效率。
图3是根据一示例性实施例示出的一种视频浏览装置的框图,该视频浏览装置应用于终端中,如图3所示,该视频浏览装置包括:图像比较模块310、图像确定模块320、操作执行模块330和图像播放模块340。
该图像比较模块310,被配置为将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,目标块信息是对当前图像帧中包括的目标块和前一个图像帧中包括的目标块进行比较后得到的信息;
该图像确定模块320,被配置为在图像比较模块310得到的目标块信息满足预定条件时,将当前图像帧确定为关键帧;
该操作执行模块330,被配置为将当前图像帧的后一个图像帧确定为当前图像帧,继续执行将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的操作,直至比较完视频中的最后一个图像帧后停止;
该图像播放模块340,被配置为对图像确定模块320确定的至少一个关键帧进行播放。
综上所述,本公开提供的视频浏览装置,通过将视频中的当前图像帧与前一个图像帧进行比较,得到目标块信息,该目标块信息是对当前图像帧中包括的目标块和前一个图像帧中包括的目标块进行比较后得到的信息;若目标块信息满足预定条件,则将当前图像帧确定为关键帧,可以通过对视频中图像帧的目标块进行比较得到目标块信息,再根据目标块信息筛选出关键帧,而不需要通过镜头的边界确定关键帧,解决了无法从连续的镜头中提取关键帧导致的无法浏览视频的问题,达到了提高视频的播放效率的效果。
图4是根据一示例性实施例示出的一种视频浏览装置的框图,该视频浏览装置应用于终端中,如图4所示,该视频浏览装置包括:图像比较模块310、图像确定模块320、操作执行模块330和图像播放模块340。
该图像比较模块310,被配置为将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,目标块信息是对当前图像帧中包括的目标块和前一个图像帧中包括的目标块进行比较后得到的信息;
该图像确定模块320,被配置为在图像比较模块310得到的目标块信息满足预定条件时,将当前图像帧确定为关键帧;
该操作执行模块330,被配置为将当前图像帧的后一个图像帧确定为当前图像帧,继续执行将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的操作,直至比较完视频中的最后一个图像帧后停止;
该图像播放模块340,被配置为对图像确定模块320中确定的至少一个关键帧进行播放。
可选的,图像比较模块310,包括:目标比较子模块311、第一确定子模块312、第二确定子模块313、第三确定子模块314;
该目标比较子模块311,被配置为比较前一个图像帧中的目标块与当前图像帧中的目标块;
该第一确定子模块312,被配置为在目标比较子模块311比较出当前图像帧中包括前一个图像帧中不存在的至少一个目标块时,将不存在的每个目标块确定为新增目标块,并 为新增目标块设置标志位信息,得到每个新增目标块的目标块信息,标志位信息是已经确定的关键帧的个数加1;
该第二确定子模块313,被配置为在目标比较子模块311比较出当前图像帧中包括前一个图像帧中存在的至少一个目标块时,将存在的每个目标块确定为已有目标块,计算当前图像帧中已有目标块基于前一个图像帧中对应的目标块的移动信息,得到每个已有目标块的目标块信息;
该第三确定子模块314,被配置为在目标比较子模块311比较出前一个图像帧中包括当前图像帧中不存在的至少一个目标块时,将不存在的每个目标块确定为消失目标块,得到每个消失目标块的目标块信息。
可选的,图像确定模块320,包括:第一读取子模块321、第一检测子模块322、第四确定子模块323;
该第一读取子模块321,被配置为从当前图像帧的目标块信息中读取每个已有目标块的移动信息,移动信息包括已有目标块的移动次数、移动位移和移动至的区域中的至少一种;
该第一检测子模块322,被配置为检测满足移动条件的移动信息的条数是否大于预定阈值;
该第四确定子模块323,被配置为在第一检测子模块322检测出满足移动条件的移动信息的条数大于预定阈值时,将当前图像帧确定为关键帧。
可选的,图像确定模块320,包括:第二读取子模块324、第二检测子模块325、第五确定子模块326;
该第二读取子模块324,被配置为从当前图像帧的目标块信息中读取每个目标块的标志位信息;
该第二检测子模块325,被配置为检测各个标志位信息中是否存在与消失目标块的标志位信息相同的标志位信息;
该第五确定子模块326,被配置为在第二检测子模块325检测出不存在与消失目标块的标志位信息相同的标志位信息时,将当前图像帧确定为关键帧。
可选的,视频浏览装置,还包括:信息获取模块350、图像筛选模块360;
该信息获取模块350,被配置为获取视频中的每个图像帧以及拍摄图像帧时的传感器信息,传感器信息包括全球定位系统GPS信息、加速度信息和感光度ISO信息中的至少一种;
该图像筛选模块360,被配置为根据信息获取模块350中获取的传感器信息对不满足筛选条件的图像帧进行丢弃,触发执行将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的操作。
可选的,图像确定模块320,包括:第三检测子模块327、第六确定子模块328;
该第三检测子模块327,被配置为在当前图像帧中包括人物时,检测当前图像帧相对 于前一个图像帧的人物姿态信息是否产生变化,并检测当前图像帧中除人物之外的目标块信息是否满足预定条件;
该第六确定子模块328,被配置为在第三检测子模块327中检测出人物姿态信息产生了变化,且当前图像帧中除人物之外的目标块信息满足预定条件时,将当前图像帧确定为关键帧。
综上所述,本公开提供的视频浏览装置,通过将视频中的当前图像帧与前一个图像帧进行比较,得到目标块信息,该目标块信息是对当前图像帧中包括的目标块和前一个图像帧中包括的目标块进行比较后得到的信息;若目标块信息满足预定条件,则将当前图像帧确定为关键帧,可以通过对视频中图像帧的目标块进行比较得到目标块信息,再根据目标块信息筛选出关键帧,而不需要通过镜头的边界确定关键帧,解决了无法从连续的镜头中提取关键帧导致的无法浏览视频的问题,达到了提高视频的播放效率的效果。
另外,通过获取视频中的每个图像帧以及拍摄图像帧时的传感器信息,根据传感器信息对不满足筛选条件的图像帧进行丢弃,可以节省对满足丢弃条件的图像帧进行处理造成的资源浪费,提高对图像帧的处理效率。
关于上述实施例中的装置,其中各个模块执行操作的具体方式已经在有关该方法的实施例中进行了详细描述,此处将不做详细阐述说明。
图5是根据一示例性实施例示出的一种用于视频浏览装置500的框图。例如,装置500可以是移动电话,计算机,数字广播终端,消息收发设备,游戏控制台,平板设备,医疗设备,健身设备,个人数字助理等。
参照图5,装置500可以包括以下一个或多个组件:处理组件502,存储器504,电源组件506,多媒体组件508,音频组件510,输入/输出(I/O)的接口512,传感器组件514,以及通信组件516。
处理组件502通常控制装置500的整体操作,诸如与显示,电话呼叫,数据通信,相机操作和记录操作相关联的操作。处理组件502可以包括一个或多个处理器518来执行指令,以完成上述的方法的全部或部分步骤。此外,处理组件502可以包括一个或多个模块,便于处理组件502和其他组件之间的交互。例如,处理组件502可以包括多媒体模块,以方便多媒体组件508和处理组件502之间的交互。
存储器504用于存储各种类型的数据以支持在装置500的操作。这些数据的示例包括用于在装置500上操作的任何应用程序或方法的指令,联系人数据,电话簿数据,消息,图片,视频等。存储器504可以由任何类型的易失性或非易失性存储设备或者它们的组合实现,如静态随机存取存储器(SRAM),电可擦除可编程只读存储器(EEPROM),可擦除可编程只读存储器(EPROM),可编程只读存储器(PROM),只读存储器(ROM),磁存储器,快闪存储器,磁盘或光盘。
电源组件506为装置500的各种组件提供电力。电源组件506可以包括电源管理系统, 一个或多个电源,及其他与为装置500生成、管理和分配电力相关联的组件。
多媒体组件508包括在所述装置500和用户之间的提供一个输出接口的屏幕。在一些实施例中,屏幕可以包括液晶显示器(LCD)和触摸面板(TP)。如果屏幕包括触摸面板,屏幕可以被实现为触摸屏,以接收来自用户的输入信号。触摸面板包括一个或多个触摸传感器以感测触摸、滑动和触摸面板上的手势。所述触摸传感器可以不仅感测触摸或滑动动作的边界,而且还检测与所述触摸或滑动操作相关的持续时间和压力。在一些实施例中,多媒体组件508包括一个前置摄像头和/或后置摄像头。当装置500处于操作模式,如拍摄模式或视频模式时,前置摄像头和/或后置摄像头可以接收外部的多媒体数据。每个前置摄像头和后置摄像头可以是一个固定的光学透镜系统或具有焦距和光学变焦能力。
音频组件510用于输出和/或输入音频信号。例如,音频组件510包括一个麦克风(MIC),当装置500处于操作模式,如呼叫模式、记录模式和语音识别模式时,麦克风用于接收外部音频信号。所接收的音频信号可以被进一步存储在存储器504或经由通信组件516发送。在一些实施例中,音频组件510还包括一个扬声器,用于输出音频信号。
I/O接口512为处理组件502和外围接口模块之间提供接口,上述外围接口模块可以是键盘,点击轮,按钮等。这些按钮可包括但不限于:主页按钮、音量按钮、启动按钮和锁定按钮。
传感器组件514包括一个或多个传感器,用于为装置500提供各个方面的状态评估。例如,传感器组件514可以检测到装置500的打开/关闭状态,组件的相对定位,例如所述组件为装置500的显示器和小键盘,传感器组件514还可以检测装置500或装置500一个组件的位置改变,用户与装置500接触的存在或不存在,装置500方位或加速/减速和装置500的温度变化。传感器组件514可以包括接近传感器,被配置用来在没有任何的物理接触时检测附近物体的存在。传感器组件514还可以包括光传感器,如CMOS或CCD图像传感器,用于在成像应用中使用。在一些实施例中,该传感器组件514还可以包括加速度传感器,陀螺仪传感器,磁传感器,压力传感器或温度传感器。
通信组件516用于便于装置500和其他设备之间有线或无线方式的通信。装置500可以接入基于通信标准的无线网络,如WiFi,2G或3G,或它们的组合。在一个示例性实施例中,通信组件516经由广播信道接收来自外部广播管理系统的广播信号或广播相关信息。在一个示例性实施例中,所述通信组件516还包括近场通信(NFC)模块,以促进短程通信。例如,在NFC模块可基于射频识别(RFID)技术,红外数据协会(IrDA)技术,超宽带(UWB)技术,蓝牙(BT)技术和其他技术来实现。
在示例性实施例中,装置500可以被一个或多个应用专用集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理设备(DSPD)、可编程逻辑器件(PLD)、现场可编程门阵列(FPGA)、控制器、微控制器、微处理器或其他电子元件实现,用于执行上述方法。
在示例性实施例中,还提供了一种包括指令的非临时性计算机可读存储介质,例如包 括指令的存储器504,上述指令可由装置500的处理器518执行以完成上述方法。例如,所述非临时性计算机可读存储介质可以是ROM、随机存取存储器(RAM)、CD-ROM、磁带、软盘和光数据存储设备等。
本领域技术人员在考虑说明书及实践这里的公开的后,将容易想到本的其它实施方案。本申请旨在涵盖本的任何变型、用途或者适应性变化,这些变型、用途或者适应性变化遵循本的一般性原理并包括本公开未公开的本技术领域中的公知常识或惯用技术手段。说明书和实施例仅被视为示例性的,本的真正范围和精神由下面的权利要求指出。
应当理解的是,本公开并不局限于上面已经描述并在附图中示出的精确结构,并且可以在不脱离其范围进行各种修改和改变。本公开的范围仅由所附的权利要求来限制。

Claims (13)

  1. 一种视频浏览方法,其特征在于,包括:
    将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,所述目标块信息是对所述当前图像帧中包括的目标块和所述前一个图像帧中包括的目标块进行比较后得到的信息;
    若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧;
    将所述当前图像帧的后一个图像帧确定为当前图像帧,继续执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤,直至比较完所述视频中的最后一个图像帧后停止;
    对确定的至少一个关键帧进行播放。
  2. 根据权利要求1所述的方法,其特征在于,所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,包括:
    比较所述前一个图像帧中的目标块与所述当前图像帧中的目标块;
    若所述当前图像帧中包括所述前一个图像帧中不存在的至少一个目标块,则将不存在的每个目标块确定为新增目标块,并为所述新增目标块设置标志位信息,得到每个新增目标块的目标块信息,所述标志位信息是已经确定的关键帧的个数加1;
    若所述当前图像帧中包括所述前一个图像帧中存在的至少一个目标块,则将存在的每个目标块确定为已有目标块,计算所述当前图像帧中所述已有目标块基于所述前一个图像帧中对应的目标块的移动信息,得到每个已有目标块的目标块信息;
    若所述前一个图像帧中包括所述当前图像帧中不存在的至少一个目标块,则将不存在的每个目标块确定为消失目标块,得到每个消失目标块的目标块信息。
  3. 根据权利要求2所述的方法,其特征在于,所述若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧,包括:
    从所述当前图像帧的目标块信息中读取每个已有目标块的移动信息,所述移动信息包括所述已有目标块的移动次数、移动位移和移动至的区域中的至少一种;
    检测满足移动条件的移动信息的条数是否大于预定阈值;
    若检测满足所述移动条件的移动信息的条数大于所述预定阈值,则将所述当前图像帧确定为所述关键帧。
  4. 根据权利要求2所述的方法,其特征在于,所述若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧,包括:
    从所述当前图像帧的目标块信息中读取每个目标块的标志位信息;
    检测各个标志位信息中是否存在与所述消失目标块的标志位信息相同的标志位信息;
    若检测出不存在与所述消失目标块的标志位信息相同的标志位信息,则将所述当前图像帧确定为所述关键帧。
  5. 根据权利要求1至4任一项所述的方法,其特征在于,所述方法,还包括:
    获取所述视频中的每个图像帧以及拍摄所述图像帧时的传感器信息,所述传感器信息包括全球定位系统GPS信息、加速度信息和感光度ISO信息中的至少一种;
    根据所述传感器信息对不满足筛选条件的图像帧进行丢弃,触发执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤。
  6. 根据权利要求1所述的方法,其特征在于,所述若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧,包括:
    若所述当前图像帧中包括人物,则检测所述当前图像帧相对于所述前一个图像帧的人物姿态信息是否产生变化,并检测所述当前图像帧中除所述人物之外的目标块信息是否满足所述预定条件;
    若检测出所述人物姿态信息产生了变化,且所述当前图像帧中除所述人物之外的目标块信息满足所述预定条件,则将所述当前图像帧确定为关键帧。
  7. 一种视频浏览装置,其特征在于,包括:
    图像比较模块,被配置为将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,所述目标块信息是对所述当前图像帧中包括的目标块和所述前一个图像帧中包括的目标块进行比较后得到的信息;
    图像确定模块,被配置为在所述图像比较模块得到的所述目标块信息满足预定条件时,将所述当前图像帧确定为关键帧;
    操作执行模块,被配置为将所述当前图像帧的后一个图像帧确定为当前图像帧,继续执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的操作,直至比较完所述视频中的最后一个图像帧后停止;
    图像播放模块,被配置为对所述图像确定模块确定的至少一个关键帧进行播放。
  8. 根据权利要求7所述的装置,其特征在于,所述图像比较模块,包括:
    目标比较子模块,被配置为比较所述前一个图像帧中的目标块与所述当前图像帧中的目标块;
    第一确定子模块,被配置为在所述目标比较子模块比较出所述当前图像帧中包括所述前一个图像帧中不存在的至少一个目标块时,将不存在的每个目标块确定为新增目标块,并为所述新增目标块设置标志位信息,得到每个新增目标块的目标块信息,所述标志位信息是已经确定的关键帧的个数加1;
    第二确定子模块,被配置为在所述目标比较子模块比较出所述当前图像帧中包括所述前一个图像帧中存在的至少一个目标块时,将存在的每个目标块确定为已有目标块,计算所述当前图像帧中所述已有目标块基于所述前一个图像帧中对应的目标块的移动信息,得到每个已有目标块的目标块信息;
    第三确定子模块,被配置为在所述目标比较子模块比较出所述前一个图像帧中包括所述当前图像帧中不存在的至少一个目标块时,将不存在的每个目标块确定为消失目标块,得到每个消失目标块的目标块信息。
  9. 根据权利要求8所述的装置,其特征在于,所述图像确定模块,包括:
    第一读取子模块,被配置为从所述当前图像帧的目标块信息中读取每个已有目标块的移动信息,所述移动信息包括所述已有目标块的移动次数、移动位移和移动至的区域中的至少一种;
    第一检测子模块,被配置为检测满足移动条件的移动信息的条数是否大于预定阈值;
    第四确定子模块,被配置为在所述第一检测子模块检测出满足所述移动条件的移动信息的条数大于所述预定阈值时,将所述当前图像帧确定为所述关键帧。
  10. 根据权利要求8所述的装置,其特征在于,所述图像确定模块,包括:
    第二读取子模块,被配置为从所述当前图像帧的目标块信息中读取每个目标块的标志位信息;
    第二检测子模块,被配置为检测各个标志位信息中是否存在与所述消失目标块的标志位信息相同的标志位信息;
    第五确定子模块,被配置为在所述第二检测子模块检测出不存在与所述消失目标块的标志位信息相同的标志位信息时,将所述当前图像帧确定为所述关键帧。
  11. 根据权利要求7至10任一项所述的装置,其特征在于,所述装置,还包括:
    信息获取模块,被配置为获取所述视频中的每个图像帧以及拍摄所述图像帧时的传感器信息,所述传感器信息包括全球定位系统GPS信息、加速度信息和感光度ISO信息中的至少一种;
    图像筛选模块,被配置为根据所述信息获取模块获取的传感器信息对不满足筛选条件的图像帧进行丢弃,触发执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的操作。
  12. 根据权利要求7所述的装置,其特征在于,所述图像确定模块,包括:
    第三检测子模块,被配置为在所述当前图像帧中包括人物时,检测所述当前图像帧相对于所述前一个图像帧的人物姿态信息是否产生变化,并检测所述当前图像帧中除所述人 物之外的目标块信息是否满足所述预定条件;
    第六确定子模块,被配置为在所述第三检测子模块检测出所述人物姿态信息产生了变化,且所述当前图像帧中除所述人物之外的目标块信息满足所述预定条件时,将所述当前图像帧确定为关键帧。
  13. 一种视频浏览装置,其特征在于,包括:
    处理器;
    用于存储处理器可执行指令的存储器;
    其中,所述处理器被配置为:
    将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息,所述目标块信息是对所述图像帧中包括的目标块和所述前一个图像帧中包括的目标块进行比较后得到的信息;
    若所述目标块信息满足预定条件,则将所述当前图像帧确定为关键帧;
    将所述当前图像帧的后一个图像帧确定为当前图像帧,继续执行所述将视频中的当前图像帧与其前一个图像帧进行比较,得到目标块信息的步骤,直至比较完所述视频中的最后一个图像帧后停止;
    对确定的至少一个关键帧进行播放。
PCT/CN2015/071247 2014-09-17 2015-01-21 视频浏览方法及装置 WO2016041311A1 (zh)

Priority Applications (6)

Application Number Priority Date Filing Date Title
JP2016548409A JP6211715B2 (ja) 2014-09-17 2015-01-21 ビデオのブラウジング方法、その装置、プログラム及び記録媒体
RU2015111193A RU2015111193A (ru) 2014-09-17 2015-01-21 Способ и устройство для просмотра видео
KR1020157007200A KR101677607B1 (ko) 2014-09-17 2015-01-21 동영상 브라우징 방법, 장치, 프로그램 및 기록매체
MX2015003761A MX361664B (es) 2014-09-17 2015-01-21 Metodo y dispositivo para navegar videos.
BR112015006631A BR112015006631A2 (pt) 2014-09-17 2015-01-21 método e dispositivo para navegação de vídeo
US14/696,691 US9799376B2 (en) 2014-09-17 2015-04-27 Method and device for video browsing based on keyframe

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201410475165.4A CN104284240B (zh) 2014-09-17 2014-09-17 视频浏览方法及装置
CN201410475165.4 2014-09-17

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/696,691 Continuation US9799376B2 (en) 2014-09-17 2015-04-27 Method and device for video browsing based on keyframe

Publications (1)

Publication Number Publication Date
WO2016041311A1 true WO2016041311A1 (zh) 2016-03-24

Family

ID=52258644

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2015/071247 WO2016041311A1 (zh) 2014-09-17 2015-01-21 视频浏览方法及装置

Country Status (8)

Country Link
EP (1) EP2998960B1 (zh)
JP (1) JP6211715B2 (zh)
KR (1) KR101677607B1 (zh)
CN (1) CN104284240B (zh)
BR (1) BR112015006631A2 (zh)
MX (1) MX361664B (zh)
RU (1) RU2015111193A (zh)
WO (1) WO2016041311A1 (zh)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113507571A (zh) * 2021-06-30 2021-10-15 深圳市路卓科技有限公司 视频防剪辑方法、装置、设备、可读存储介质及程序产品
CN113516006A (zh) * 2021-04-01 2021-10-19 广州云硕科技发展有限公司 一种用于智能机器人的高效信息处理方法及系统

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104284240B (zh) * 2014-09-17 2018-02-02 小米科技有限责任公司 视频浏览方法及装置
US9799376B2 (en) 2014-09-17 2017-10-24 Xiaomi Inc. Method and device for video browsing based on keyframe
CN105426515B (zh) * 2015-12-01 2018-12-18 小米科技有限责任公司 视频归类方法及装置
CN105516675B (zh) * 2015-12-28 2018-11-16 同方威视技术股份有限公司 图像显示方法及用于图像显示的系统
KR102374747B1 (ko) 2017-12-15 2022-03-15 삼성전자주식회사 객체를 인식하는 장치 및 방법
CN108563651B (zh) * 2017-12-19 2021-05-11 深圳云天励飞技术有限公司 一种多视频的目标搜索方法、装置及设备
CN109993025B (zh) * 2017-12-29 2021-07-06 中移(杭州)信息技术有限公司 一种关键帧提取方法及设备
CN108470077B (zh) * 2018-05-28 2023-07-28 广东工业大学 一种视频关键帧提取方法、系统及设备和存储介质
CN113965814B (zh) * 2021-08-30 2023-07-04 国网山东省电力公司信息通信公司 基于视频会议场景的多会场关键帧提取方法及系统
CN115474084B (zh) * 2022-08-10 2023-10-31 北京奇艺世纪科技有限公司 一种视频封面图像的生成方法、装置、设备和存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050226331A1 (en) * 2004-03-31 2005-10-13 Honeywell International Inc. Identifying key video frames
CN101360184A (zh) * 2008-09-22 2009-02-04 腾讯科技(深圳)有限公司 提取视频关键帧的系统及方法
CN101398855A (zh) * 2008-10-24 2009-04-01 清华大学 一种视频关键帧提取方法和系统
CN101479729A (zh) * 2006-06-29 2009-07-08 皇家飞利浦电子股份有限公司 提取关键帧的系统和方法
CN103533237A (zh) * 2013-09-29 2014-01-22 清华大学 一种从视频中提取视频关键帧的方法
CN104284240A (zh) * 2014-09-17 2015-01-14 小米科技有限责任公司 视频浏览方法及装置

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6055025A (en) * 1993-12-21 2000-04-25 Lucent Technologies, Inc. Method and apparatus for detecting abrupt and gradual scene changes in image sequences
JP2914170B2 (ja) * 1994-04-18 1999-06-28 松下電器産業株式会社 映像変化点検出方法
JP3528524B2 (ja) * 1997-07-10 2004-05-17 ソニー株式会社 記録再生装置および記録再生方法、並びに記録媒体
US5956026A (en) * 1997-12-19 1999-09-21 Sharp Laboratories Of America, Inc. Method for hierarchical summarization and browsing of digital video
US7280753B2 (en) * 2003-09-03 2007-10-09 Canon Kabushiki Kaisha Display apparatus, image processing apparatus, and image processing system
JP2005159781A (ja) * 2003-11-27 2005-06-16 Seiko Epson Corp 画像選択処理装置、画像選択方法およびそのプログラム、記録媒体
US20070237225A1 (en) * 2006-03-30 2007-10-11 Eastman Kodak Company Method for enabling preview of video files
CN101184226A (zh) * 2006-11-14 2008-05-21 腾讯科技(深圳)有限公司 一种插入视频关键帧的方法及装置
JP4433027B2 (ja) * 2007-09-21 2010-03-17 ソニー株式会社 信号処理装置、信号処理方法、およびプログラム
JP5630541B2 (ja) * 2009-01-20 2014-11-26 ソニー株式会社 情報処理装置、情報処理方法およびプログラム
JP5393237B2 (ja) * 2009-04-24 2014-01-22 キヤノン株式会社 画像表示装置及びその制御方法、並びにプログラム及び記憶媒体
WO2011099299A1 (ja) * 2010-02-10 2011-08-18 パナソニック株式会社 映像抽出装置、撮影装置、プログラム及び記録媒体
JP5537285B2 (ja) * 2010-06-28 2014-07-02 日本放送協会 要約映像生成装置及び要約映像生成プログラム
CN102348115B (zh) * 2010-08-02 2014-04-16 南京壹进制信息技术有限公司 一种去除视频中冗余图像的方法和装置
CN109542849B (zh) * 2012-09-16 2021-09-24 吴东辉 一种图像文件格式及生成方法及装置及应用
CN103345492A (zh) * 2013-06-25 2013-10-09 无锡赛思汇智科技有限公司 一种视频浓缩的方法和系统

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050226331A1 (en) * 2004-03-31 2005-10-13 Honeywell International Inc. Identifying key video frames
CN101479729A (zh) * 2006-06-29 2009-07-08 皇家飞利浦电子股份有限公司 提取关键帧的系统和方法
CN101360184A (zh) * 2008-09-22 2009-02-04 腾讯科技(深圳)有限公司 提取视频关键帧的系统及方法
CN101398855A (zh) * 2008-10-24 2009-04-01 清华大学 一种视频关键帧提取方法和系统
CN103533237A (zh) * 2013-09-29 2014-01-22 清华大学 一种从视频中提取视频关键帧的方法
CN104284240A (zh) * 2014-09-17 2015-01-14 小米科技有限责任公司 视频浏览方法及装置

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113516006A (zh) * 2021-04-01 2021-10-19 广州云硕科技发展有限公司 一种用于智能机器人的高效信息处理方法及系统
CN113507571A (zh) * 2021-06-30 2021-10-15 深圳市路卓科技有限公司 视频防剪辑方法、装置、设备、可读存储介质及程序产品
CN113507571B (zh) * 2021-06-30 2023-10-24 深圳市路卓科技有限公司 视频防剪辑方法、装置、设备、可读存储介质

Also Published As

Publication number Publication date
EP2998960B1 (en) 2020-06-17
MX2015003761A (es) 2016-08-01
JP2016541214A (ja) 2016-12-28
CN104284240B (zh) 2018-02-02
BR112015006631A2 (pt) 2017-07-04
RU2015111193A (ru) 2016-10-20
EP2998960A1 (en) 2016-03-23
MX361664B (es) 2018-12-13
KR101677607B1 (ko) 2016-11-29
KR20160043523A (ko) 2016-04-21
JP6211715B2 (ja) 2017-10-11
CN104284240A (zh) 2015-01-14

Similar Documents

Publication Publication Date Title
WO2016041311A1 (zh) 视频浏览方法及装置
US20200077035A1 (en) Video recording method and apparatus, electronic device and readable storage medium
WO2016029641A1 (zh) 照片获取方法及装置
CN106375674B (zh) 寻找和使用与相邻静态图像相关的视频部分的方法和装置
CN105072337B (zh) 图片处理方法及装置
JP6335289B2 (ja) 画像フィルタを生成する方法及び装置
TWI702544B (zh) 圖像處理方法、電子設備和電腦可讀儲存介質
EP3136391B1 (en) Method, device and terminal device for video effect processing
WO2017071085A1 (zh) 报警方法及装置
US10115019B2 (en) Video categorization method and apparatus, and storage medium
CN105069083B (zh) 关联用户的确定方法及装置
WO2017088470A1 (zh) 图像分类方法及装置
WO2017124773A1 (zh) 手势识别方法及装置
WO2020259073A1 (zh) 图像处理方法及装置、电子设备和存储介质
EP3327590A1 (en) Method and device for adjusting video playback position
US20170154206A1 (en) Image processing method and apparatus
US9799376B2 (en) Method and device for video browsing based on keyframe
KR102545408B1 (ko) 차폐된 이미지 검출 방법, 장치 및 매체
CN105631803B (zh) 滤镜处理的方法和装置
WO2018095252A1 (zh) 视频录制方法及装置
CN105528078A (zh) 控制电子设备的方法及装置
WO2017143778A1 (zh) 照片的自动获取方法及装置、电子设备
CN110266914A (zh) 图像拍摄方法、装置及计算机可读存储介质
WO2022198821A1 (zh) 人脸和人体匹配的方法、装置、电子设备、存储介质及程序
CN114598923A (zh) 视频文字清除方法、装置、电子设备及存储介质

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 2016548409

Country of ref document: JP

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 20157007200

Country of ref document: KR

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: MX/A/2015/003761

Country of ref document: MX

ENP Entry into the national phase

Ref document number: 2015111193

Country of ref document: RU

Kind code of ref document: A

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112015006631

Country of ref document: BR

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15842239

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 112015006631

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20150325

122 Ep: pct application non-entry in european phase

Ref document number: 15842239

Country of ref document: EP

Kind code of ref document: A1