WO2021073204A1 - 对象的显示方法、装置、电子设备及计算机可读存储介质 - Google Patents
对象的显示方法、装置、电子设备及计算机可读存储介质 Download PDFInfo
- Publication number
- WO2021073204A1 WO2021073204A1 PCT/CN2020/106818 CN2020106818W WO2021073204A1 WO 2021073204 A1 WO2021073204 A1 WO 2021073204A1 CN 2020106818 W CN2020106818 W CN 2020106818W WO 2021073204 A1 WO2021073204 A1 WO 2021073204A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- target object
- video image
- shape
- response
- identifying
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 53
- 230000004044 response Effects 0.000 claims abstract description 57
- 230000008859 change Effects 0.000 claims description 27
- 230000000694 effects Effects 0.000 abstract description 13
- 230000006870 function Effects 0.000 description 10
- 238000004364 calculation method Methods 0.000 description 8
- 238000010586 diagram Methods 0.000 description 8
- 238000004891 communication Methods 0.000 description 7
- 238000004590 computer program Methods 0.000 description 7
- 238000012545 processing Methods 0.000 description 7
- 238000013528 artificial neural network Methods 0.000 description 6
- 230000000295 complement effect Effects 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 6
- 238000004422 calculation algorithm Methods 0.000 description 4
- 230000008034 disappearance Effects 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 238000001514 detection method Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 239000013307 optical fiber Substances 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012827 research and development Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000009967 tasteless effect Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/04—Context-preserving transformations, e.g. by using an importance map
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/24—Aligning, centring, orientation detection or correction of the image
- G06V10/245—Aligning, centring, orientation detection or correction of the image by locating a pattern; Special marks for positioning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/41—Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/2621—Cameras specially adapted for the electronic generation of special effects during image pickup, e.g. digital cameras, camcorders, video cameras having integrated special effects capability
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/255—Detecting or recognising potential candidate objects based on visual cues, e.g. shapes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Definitions
- the present disclosure relates to the field of image processing, and in particular to an object display method, device, electronic equipment, and computer-readable storage medium.
- the current video special effects production is generally completed by post-production after recording the video.
- the displayed special effects are fixed and can only be played until the end according to the pre-time logic; and the threshold of post-production is high, and ordinary users cannot generate it quickly.
- embodiments of the present disclosure provide an object display method, including:
- the first object and the second object are combined to make the first object and the second object form a first Three shapes, wherein the third shape is formed by combining the first shape and the second shape.
- an object display device including:
- Image receiving module for receiving video images
- the target object recognition module is used to recognize the target object in the video image
- the first object display module is configured to display the first object having the first shape at the first position of the video image in response to identifying the first target object;
- a second object display module configured to display a second object having a second shape at a second position of the video image in response to identifying the second target object
- the combining module is used to combine the first object and the second object so that the first object and the second object are combined when the distance between the first target object and the second target object is less than a first threshold.
- the second object forms a third shape, wherein the third shape is formed by combining the first shape and the second shape.
- an embodiment of the present disclosure provides an electronic device, including: at least one processor; and,
- a memory that is communicatively connected to the at least one processor; wherein the memory stores instructions that can be executed by the at least one processor, and the instructions are executed by the at least one processor to enable the at least one processing
- the device can execute any of the object display methods in the foregoing first aspect.
- embodiments of the present disclosure provide a non-transitory computer-readable storage medium, characterized in that the non-transitory computer-readable storage medium stores computer instructions, and the computer instructions are used to make a computer execute the aforementioned first aspect The display method of any of the described objects.
- the present disclosure discloses an object display method, device, electronic equipment and computer readable storage medium.
- the method for displaying the object includes: receiving a video image, the video image including a plurality of image frames; identifying the target object in the image frame of the video image; in response to identifying the first target object, in the video image A first object with a first shape is displayed at the first position; in response to identifying a second target object, the second object with a second shape is displayed at the second position of the video image; when the first The distance between the target object and the second target object is less than a first threshold, and the first object and the second object are combined to make the first object and the second object form a third shape.
- FIG. 1 is a flowchart of an embodiment of a method for displaying objects provided by the present disclosure
- step S103 is a flow chart of a specific example of step S103 in the embodiment of the object display method provided by the present disclosure
- FIG. 3 is a flowchart of a further embodiment of the object display method embodiment provided by the present disclosure.
- FIG. 4 is a schematic flowchart of an example of an object state change in an embodiment of an object display method provided by the present disclosure
- FIG. 5 is a schematic structural diagram of an embodiment of an object display device provided by an embodiment of the disclosure.
- Fig. 6 is a schematic structural diagram of an electronic device provided according to an embodiment of the present disclosure.
- FIG. 1 is a flowchart of an embodiment of an object display method provided by an embodiment of the disclosure.
- the object display method provided in this embodiment may be executed by a display device of an object, and the display device of the object may be implemented as software, or Realized as a combination of software and hardware, the display device of the object can be integrated in a certain device in the display system of the object, such as the display server of the object or the display terminal device of the object.
- the method includes the following steps:
- Step S101 receiving a video image
- the receiving a video image is receiving a video image from an image sensor.
- the image sensor refers to various devices that can collect images, and typical image sensors are video cameras, cameras, cameras, and so on.
- the image sensor may be a camera on a terminal device, such as a front or rear camera on a smart phone, and the image collected by the camera may be directly displayed on the display screen of the mobile phone.
- receiving The video image taken by the image sensor is used as the display background image of the object.
- the receiving the video image is receiving a video image file from a memory, and the video image file is a pre-recorded video and is stored in the memory.
- the storage is a local storage or a network storage.
- the video image file sent from the memory is received, and the target display background image is displayed on the display device of the device.
- the video image may include a human face.
- at least one of the multiple image frames in the video image includes a human face.
- the user turns on the camera of the mobile terminal device to take a video selfie of himself, and the video image obtained by the camera includes the user's face image. Or an image of a human face is recorded in the video image file.
- Step S102 Identify the target object in the video image
- the target object may be any preset object that needs to be recognized from the video image, and a typical target object may be a human face.
- any target recognition algorithm can be used to recognize the target object in the video image.
- the neural network can be trained using images with human faces, and the trained neural network can be used to classify each image frame in the video image to determine Whether the image frame contains a human face, when it is judged that the image frame contains a human face, then a face detection algorithm is used to detect the key points of the human face to determine the position of the human face; or a mark can be used to circumscribe the human face
- the neural network is trained using the trained neural network for each image frame in the video image to reduce the range of the face, and then the face detection algorithm is used to detect the key points of the face To determine the position of the human face; or use the image marked with key points of the human face to train the neural network, and use the trained neural network to perform face key point regression on each image frame in the video image to determine the image frame Whether to include the face and the position of the key points of the face.
- the aforementioned face and recognition method are only examples and do not constitute a limitation to the present disclosure.
- the target object and the appropriate recognition algorithm for the target object can be pre-selected according to the effect to be achieved and the scene.
- Step S103 in response to identifying the first target object, display the first object having the first shape at the first position of the video image;
- the object is a virtual material that is used to be displayed in an actual video image.
- a typical virtual material may be a texture, and the texture may be a static texture or a sequence frame texture.
- the first object corresponding to the first target object is acquired, and the first object is compared with the first object in the video image.
- the first object has a first shape.
- the corresponding relationship between the target object and the object is preset, and when the target object appears in the video image, the object corresponding to the target object is obtained from the corresponding relationship, wherein the corresponding relationship can be saved in a configuration file ,
- the target object to be recognized, the type of the object, and the corresponding relationship between the target object and the object can be configured through the configuration file.
- displaying the first object at the first position of the video image includes:
- Step S201 in response to identifying the first target object, obtain a tracking point of the first target object
- Step S202 Calculate the first position according to the tracking point of the first target object
- Step S203 Display the first object at the first position.
- the key points of the first target object are also detected.
- One or more of the key points are preset as a tracking point of the first target object, or the tracking point is determined according to one or more of the key points, and the tracking point is used to determine the first target object.
- the first position is calculated according to the tracking point of the first target object.
- the first target object is a human face.
- the tracking point is a point on an extension line passing through the line between the key point of the nose tip of the face and the key point of the edge, and is located outside the face area.
- the extension ratio is set in advance, and the ratio can be calculated from the edge.
- step S203 the specific point of the first object is overlapped with the tracking point, and the first object is displayed at the first position.
- the specific point of the first object is the The center point of the first object, the position of the center point is set at the position of the tracking point, so that the first object is displayed on the first position. Therefore, when the position of the first target object changes in different image frames, the position of the first object also changes accordingly, achieving the effect that the first object moves following the target object.
- the acquiring the tracking point of the first target object in response to identifying the first target object includes: in response to identifying the first target object, determining that the first target object is in the video image The position of the first target object is obtained according to the position of the first target object in the video image.
- the position of the first target object in the video image is first determined.
- the first target object is a human face
- the edge key points on the right side of the human face and the key points of the nose tip of the human face in the video image are acquired to Make the tracking point located on the right side of the face (relative to the viewing angle); similarly, when it is determined that the first target object is located on the right side of the video image, obtain the left edge of the face in the video image Key points and key points of the nose tip of the human face, so that the tracking point is located on the left side of the human face.
- the first object has a first shape.
- the first shape is a part of a complete shape.
- the first shape may be a part of a circle, a part of a heart shape, a part of a rectangle, etc., and the first shape may also be a part of a complete irregular shape. Such as part of the cloud, part of the mountain and so on.
- target objects and key points and calculation of tracking points are examples and do not constitute a limitation to the present disclosure.
- any calculation method of target objects, key points, and tracking points can be applied to the technology of the present disclosure. In the program.
- Step S104 in response to identifying the second target object, display the second object with the second shape at the second position of the video image;
- the displaying the second object on the second position of the video image in response to identifying the second target object includes: in response to identifying the second target object, acquiring the second target object The tracking point; the second position is calculated according to the tracking point of the second target object; the second object is displayed at the second position.
- acquiring the tracking point of the second target object includes: in response to identifying the second target object, determining that the second target object is in the video image The position of the second target object is obtained according to the position of the second target object in the video image.
- the second shape and the first shape are complementary shapes.
- the first shape is half a heart shape
- the second shape is the other half heart shape.
- Step S105 When the distance between the first target object and the second target object is less than a first threshold, the first object and the second object are combined to make the first object and the second object The object forms a third shape.
- the method when the distance between the first target object and the second target object is less than a first threshold, the method includes: obtaining a first anchor point of the first target object; obtaining a second anchor point of the second target object Anchor point; calculate the first distance between the first anchor point and the second anchor point; when the first distance is less than the first threshold, determine whether the first target object and the second target object are The distance between is smaller than the first threshold.
- the anchor point may be a key point of the target object or a point calculated from the key point of the target object.
- the anchor point is a key point of the nose tip of a human face, or the anchor point It is the midpoint of the triangle formed by the key points of the eyes and the nose of the human face.
- the first distance is calculated according to the coordinates of the first anchor point and the coordinates of the second anchor point, and the first distance represents the first target object and the second anchor point.
- the distance between the two target objects when the first distance is less than the preset first threshold, the distance between the distance between the first target object and the second target object is less than the first threshold.
- combining the first object and the second object to make the first object and the second object form a third shape includes: combining the first shape of the first object with the second shape
- the second shape of the object is combined according to complementary positions to form a third shape.
- the first shape and the second shape may be complementary half-heart shapes.
- the two objects can be combined together by calculating the distance between the two target objects in the image. There is no need to record the video in advance and make special effects in the video. It can be rendered in real time. Any video image.
- the method further includes:
- Step S301 when the distance between the first target object and the second target object is less than a second threshold, the first object and the second object are made to disappear;
- Step S302 displaying a first sequence of frames, the first sequence of frames including a third object having a third shape.
- the first distance between the first anchor point and the second anchor point can be calculated according to the method in step S105.
- the first distance is less than the second threshold
- the first object and the second object are controlled to disappear in the video image.
- the disappearance may be that the first object and the second object are no longer rendered.
- the transparency of the first object and the second object may be set to 1 to make it transparent and invisible.
- a first sequence of frames is acquired, and the first sequence of frames is displayed at the position where the first object and the second object disappear.
- the first sequence of frames includes a third object having a third shape, specifically, the first sequence of frames includes a third object, and each frame in the first sequence of frames represents a state change of the third object
- the third object is a virtual heart
- the first sequence of frames represents the dynamic change of the virtual heart, such as gradually changing from white to red from small to large, and so on.
- the first sequence of frames represents the change state of the third object represented by the combination of the first object and the second object.
- step S104 it may further include:
- Step S401 Calculate the distance between the first target object and the second target object
- Step S402 Obtain change parameters of the first object and the second object according to the distance
- Step S403 Change the display state of the first object and the second object according to the change parameter.
- step S401 the method described in step S105 can also be used to calculate the distance between the first target object and the second target object, which will not be repeated here.
- a change parameter is obtained.
- the change parameter is related to the distance and may be a preset correspondence relationship.
- the first value of the distance corresponds to the first color value
- the second value corresponds to the second color value. Therefore, the color change of the object and the like can be controlled through the distance.
- step S403 the display states of the first object and the second object are changed according to the change parameter.
- the change parameter determines the color of the first object and the second object, and the color of the first object and the second object is changed by the change parameter.
- the display state of the object is changed by the distance between the two target objects.
- the first target object and the second target object are human faces
- the first object and the second object are respectively two semi-cardioid shapes with complementary shapes.
- the present disclosure discloses an object display method, device, electronic equipment and computer readable storage medium.
- the method for displaying the object includes: receiving a video image, the video image including a plurality of image frames; identifying the target object in the image frame of the video image; in response to identifying the first target object, in the video image A first object with a first shape is displayed at the first position; in response to identifying a second target object, the second object with a second shape is displayed at the second position of the video image; when the first The distance between the target object and the second target object is less than a first threshold, and the first object and the second object are combined to make the first object and the second object form a third shape.
- FIG. 5 is a schematic structural diagram of an embodiment of an object display device provided by an embodiment of the disclosure.
- the device 500 includes: an image receiving module 501, a target object recognition module 502, a first object display module 503, and a second object display module 503.
- the image receiving module 501 is used to receive video images
- the target object recognition module 502 is configured to recognize the target object in the video image
- the first object display module 503 is configured to display a first object having a first shape at a first position of the video image in response to identifying a first target object;
- the second object display module 504 is configured to display the second object with the second shape in the second position of the video image in response to identifying the second target object;
- the combining module 505 is configured to combine the first object and the second object so that the first object and the second object are separated from each other when the distance between the first target object and the second target object is less than a first threshold.
- the second object forms a third shape, wherein the third shape is formed by combining the first shape and the second shape.
- the display device 500 of the object further includes:
- the disappearance control module is configured to make the first object and the second object disappear when the distance between the first target object and the second target object is less than a second threshold;
- the sequence frame display module is configured to display a first sequence of frames, and the first sequence of frames includes a third object with a third shape.
- the first object display module 503 further includes:
- the first tracking point acquisition module is configured to acquire the tracking point of the first target object in response to identifying the first target object
- a first position calculation module configured to calculate the first position according to the tracking point of the first target object
- the first object display sub-module is configured to display the first object at the first position.
- the first tracking point acquisition module further includes:
- the first target object position determination module is configured to determine the position of the first target object in the video image in response to identifying the first target object
- the first tracking point acquisition sub-module is configured to acquire the tracking point of the first target object according to the position of the first target object in the video image.
- the second object display module 504 further includes:
- the second tracking point acquiring module is configured to acquire the tracking point of the second target object in response to identifying the second target object;
- a second position calculation module configured to calculate the second position according to the tracking point of the second target object
- the second object display sub-module is configured to display the second object at the second position.
- the second tracking point acquisition module further includes:
- the second target object position determination module is configured to determine the position of the second target object in the video image in response to identifying the second target object;
- the second tracking point acquisition sub-module is configured to acquire the tracking point of the second target object according to the position of the second target object in the video image.
- the display device 500 of the object further includes:
- a distance calculation module configured to calculate the distance between the first target object and the second target object
- a change parameter acquisition module configured to acquire change parameters of the first object and the second object according to the distance
- the display state change module is configured to change the display states of the first object and the second object according to the change parameter.
- the device shown in FIG. 5 can execute the method of the embodiment shown in FIG. 1 to FIG. 4.
- parts that are not described in detail in this embodiment please refer to the related description of the embodiment shown in FIG. 1 to FIG. 4.
- the description in the embodiment shown in FIG. 1 to FIG. 4 please refer to the description in the embodiment shown in FIG. 1 to FIG. 4, which will not be repeated here.
- FIG. 6 shows a schematic structural diagram of an electronic device 600 suitable for implementing embodiments of the present disclosure.
- the terminal devices in the embodiments of the present disclosure may include, but are not limited to, mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PAD (tablet computers), PMPs (portable multimedia players), vehicle-mounted terminals (e.g. Mobile terminals such as car navigation terminals) and fixed terminals such as digital TVs, desktop computers, etc.
- the electronic device shown in FIG. 6 is only an example, and should not bring any limitation to the function and scope of use of the embodiments of the present disclosure.
- the electronic device 600 may include a processing device (such as a central processing unit, a graphics processor, etc.) 601, which may be loaded into a random access device according to a program stored in a read-only memory (ROM) 602 or from a storage device 606.
- the program in the memory (RAM) 603 executes various appropriate actions and processing.
- various programs and data required for the operation of the electronic device 600 are also stored.
- the processing device 601, the ROM 602, and the RAM 603 are connected to each other through a bus 604.
- An input/output (I/O) interface 605 is also connected to the bus 604.
- the following devices can be connected to the I/O interface 605: including input devices 606 such as touch screen, touch pad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, liquid crystal display (LCD), speakers, vibration An output device 607 such as a device; a storage device 606 such as a magnetic tape and a hard disk; and a communication device 609.
- the communication device 609 may allow the electronic device 600 to perform wireless or wired communication with other devices to exchange data.
- FIG. 6 shows an electronic device 600 having various devices, it should be understood that it is not required to implement or have all the illustrated devices. It may alternatively be implemented or provided with more or fewer devices.
- the process described above with reference to the flowchart can be implemented as a computer software program.
- the embodiments of the present disclosure include a computer program product, which includes a computer program carried on a non-transitory computer readable medium, and the computer program contains program code for executing the method shown in the flowchart.
- the computer program may be downloaded and installed from the network through the communication device 609, or installed from the storage device 606, or installed from the ROM 602.
- the processing device 601 the above-mentioned functions defined in the method of the embodiment of the present disclosure are executed.
- the aforementioned computer-readable medium in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the two.
- the computer-readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or a combination of any of the above. More specific examples of computer-readable storage media may include, but are not limited to: electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable Programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
- a computer-readable storage medium may be any tangible medium that contains or stores a program, and the program may be used by or in combination with an instruction execution system, apparatus, or device.
- a computer-readable signal medium may include a data signal propagated in a baseband or as a part of a carrier wave, and a computer-readable program code is carried therein. This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
- the computer-readable signal medium may also be any computer-readable medium other than the computer-readable storage medium.
- the computer-readable signal medium may send, propagate or transmit the program for use by or in combination with the instruction execution system, apparatus, or device .
- the program code contained on the computer-readable medium can be transmitted by any suitable medium, including but not limited to: wire, optical cable, RF (Radio Frequency), etc., or any suitable combination of the above.
- the client and server can communicate with any currently known or future developed network protocol such as HTTP (HyperText Transfer Protocol), and can communicate with digital data in any form or medium.
- Communication e.g., communication network
- Examples of communication networks include local area networks (“LAN”), wide area networks (“WAN”), the Internet (for example, the Internet), and end-to-end networks (for example, ad hoc end-to-end networks), as well as any currently known or future research and development network of.
- the above-mentioned computer-readable medium may be included in the above-mentioned electronic device; or it may exist alone without being assembled into the electronic device.
- the aforementioned computer-readable medium carries one or more programs, and when the aforementioned one or more programs are executed by the electronic device, the electronic device: receives a video image; recognizes the target object in the video image; and responds to the recognition A first target object, displaying a first object having a first shape at a first position of the video image; in response to recognizing a second target object, displaying a first object having a second shape at a second position of the video image Shape of the second object; when the distance between the first target object and the second target object is less than the first threshold, the first object and the second object are combined to make the first object and the
- the second object forms a third shape, wherein the third shape is formed by combining the first shape and the second shape.
- the computer program code used to perform the operations of the present disclosure may be written in one or more programming languages or a combination thereof.
- the above-mentioned programming languages include, but are not limited to, object-oriented programming languages such as Java, Smalltalk, C++, and Including conventional procedural programming languages-such as "C" language or similar programming languages.
- the program code can be executed entirely on the user's computer, partly on the user's computer, executed as an independent software package, partly on the user's computer and partly executed on a remote computer, or entirely executed on the remote computer or server.
- the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (for example, using an Internet service provider to pass Internet connection).
- LAN local area network
- WAN wide area network
- each block in the flowchart or block diagram may represent a module, program segment, or part of code, and the module, program segment, or part of code contains one or more for realizing the specified logical function Executable instructions.
- the functions marked in the block may also occur in a different order from the order marked in the drawings. For example, two blocks shown in succession can actually be executed substantially in parallel, and they can sometimes be executed in the reverse order, depending on the functions involved.
- each block in the block diagram and/or flowchart, and the combination of the blocks in the block diagram and/or flowchart can be implemented by a dedicated hardware-based system that performs the specified functions or operations Or it can be realized by a combination of dedicated hardware and computer instructions.
- the units described in the embodiments of the present disclosure can be implemented in software or hardware. Among them, the name of the unit does not constitute a limitation on the unit itself under certain circumstances.
- exemplary types of hardware logic components include: Field Programmable Gate Array (FPGA), Application Specific Integrated Circuit (ASIC), Application Specific Standard Product (ASSP), System on Chip (SOC), Complex Programmable Logical device (CPLD) and so on.
- FPGA Field Programmable Gate Array
- ASIC Application Specific Integrated Circuit
- ASSP Application Specific Standard Product
- SOC System on Chip
- CPLD Complex Programmable Logical device
- a machine-readable medium may be a tangible medium, which may contain or store a program for use by the instruction execution system, apparatus, or device or in combination with the instruction execution system, apparatus, or device.
- the machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
- the machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or any suitable combination of the foregoing.
- machine-readable storage media would include electrical connections based on one or more wires, portable computer disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable programmable read-only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
- RAM random access memory
- ROM read-only memory
- EPROM or flash memory erasable programmable read-only memory
- CD-ROM compact disk read-only memory
- magnetic storage device or any suitable combination of the above.
- an object display method including:
- the first object and the second object are combined to make the first object and the second object form a first Three shapes, wherein the third shape is formed by combining the first shape and the second shape.
- the method further includes:
- a first sequence of frames is displayed, and the first sequence of frames includes a third object having a third shape.
- displaying the first object at the first position of the video image includes:
- the first object is displayed at the first position.
- acquiring the tracking point of the first target object includes:
- displaying the second object at the second position of the video image includes:
- the second object is displayed at the second position.
- acquiring the tracking point of the second target object includes:
- the method further includes:
- the display states of the first object and the second object are changed according to the change parameter.
- an object display device including:
- Image receiving module for receiving video images
- the target object recognition module is used to recognize the target object in the video image
- the first object display module is configured to display the first object having the first shape at the first position of the video image in response to identifying the first target object;
- a second object display module configured to display a second object having a second shape at a second position of the video image in response to identifying the second target object
- the combining module is used to combine the first object and the second object so that the first object and the second object are combined when the distance between the first target object and the second target object is less than a first threshold.
- the second object forms a third shape, wherein the third shape is formed by combining the first shape and the second shape.
- the display device of the object further includes:
- the disappearance control module is configured to make the first object and the second object disappear when the distance between the first target object and the second target object is less than a second threshold;
- the sequence frame display module is configured to display a first sequence of frames, and the first sequence of frames includes a third object with a third shape.
- the first object display module further includes:
- the first tracking point acquisition module is configured to acquire the tracking point of the first target object in response to identifying the first target object
- a first position calculation module configured to calculate the first position according to the tracking point of the first target object
- the first object display sub-module is configured to display the first object at the first position.
- the first tracking point acquisition module further includes:
- the first target object position determination module is configured to determine the position of the first target object in the video image in response to identifying the first target object
- the first tracking point acquisition sub-module is configured to acquire the tracking point of the first target object according to the position of the first target object in the video image.
- the second object display module further includes:
- the second tracking point acquiring module is configured to acquire the tracking point of the second target object in response to identifying the second target object;
- a second position calculation module configured to calculate the second position according to the tracking point of the second target object
- the second object display sub-module is configured to display the second object at the second position.
- the second tracking point acquisition module further includes:
- the second target object position determination module is configured to determine the position of the second target object in the video image in response to identifying the second target object;
- the second tracking point acquisition sub-module is configured to acquire the tracking point of the second target object according to the position of the second target object in the video image.
- the display device of the object further includes:
- a distance calculation module configured to calculate the distance between the first target object and the second target object
- a change parameter acquisition module configured to acquire change parameters of the first object and the second object according to the distance
- the display state change module is configured to change the display states of the first object and the second object according to the change parameter.
- an electronic device including: at least one processor; and,
- the device can execute any of the aforementioned object display methods.
- non-transitory computer-readable storage medium characterized in that the non-transitory computer-readable storage medium stores computer instructions, and the computer instructions are used to make a computer execute the foregoing Any display method of the object.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Signal Processing (AREA)
- Image Analysis (AREA)
- User Interface Of Digital Computer (AREA)
- Studio Devices (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
Description
Claims (10)
- 一种对象的显示方法,包括:接收视频图像;识别所述视频图像中的目标对象;响应于识别出第一目标对象,在所述视频图像的第一位置上显示具有第一形状的第一对象;响应于识别出第二目标对象,在所述视频图像的第二位置上显示具有第二形状的第二对象;当所述第一目标对象和所述第二目标对象之间的距离小于第一阈值,将所述第一对象和第二对象进行结合以使所述第一对象和所述第二对象形成第三形状,其中所述第三形状由所述第一形状和所述第二形状结合而成。
- 如权利要求1所述的对象的显示方法,所述方法还包括:当所述第一目标对象和所述第二目标对象之间的距离小于第二阈值,使第一对象和第二对象消失;显示第一序列帧,所述第一序列帧中包括具有第三形状的第三对象。
- 如权利要求1所述的对象的显示方法,其中所述响应于识别出第一目标对象,在所述视频图像的第一位置上显示所述第一对象,包括:响应于识别出第一目标对象,获取所述第一目标对象的跟踪点;根据所述第一目标对象的跟踪点计算所述第一位置;在所述第一位置上显示所述第一对象。
- 如权利要求3所述的对象的显示方法,其中所述响应于识别出第一目标对象,获取所述第一目标对象的跟踪点,包括:响应于识别出第一目标对象,判断所述第一目标对象在所述视频图像中的位置;根据所述第一目标对象在所述视频图像中的位置获取所述第一目标对象的跟踪点。
- 如权利要求1所述的对象的显示方法,其中所述响应于识别出第二目标对象,在所述视频图像的第二位置上显示所述第二对象,包括:响应于识别出第二目标对象,获取所述第二目标对象的跟踪点;根据所述第二目标对象的跟踪点计算所述第二位置;在所述第二位置上显示所述第二对象。
- 如权利要求5所述的对象的显示方法,其中所述响应于识别出第二目标对象,获取所述第二目标对象的跟踪点,包括:响应于识别出第二目标对象,判断所述第二目标对象在所述视频图像中的位置;根据所述第二目标对象在所述视频图像中的位置获取所述第二目标对象的跟踪点。
- 如权利要求1所述的对象的显示方法,其中在所述响应于识别出第二目标对象,在所述视频图像的第二位置上显示所述具有第二形状的第二对象之后,还包括:计算所述第一目标对象和所述第二目标对象之间的距离;根据所述距离获取所述第一对象和第二对象的变化参数;根据所述变化参数改变所述第一对象和所述第二对象的显示状态。
- 一种对象的显示装置,包括:图像接收模块,用于接收视频图像;目标对象识别模块,用于识别所述视频图像中的目标对象;第一对象显示模块,用于响应于识别出第一目标对象,在所述视频图像的第一位置上显示具有第一形状的第一对象;第二对象显示模块,用于响应于识别出第二目标对象,在所述视频图像的第二位置上显示具有第二形状的第二对象;结合模块,用于当所述第一目标对象和所述第二目标对象之间的距离小于第一阈值,将所述第一对象和第二对象进行结合以使所述第一对象和所述第二对象形成第三形状,其中所述第三形状由所述第一形状和所述第二形状结合而成。
- 一种电子设备,包括:存储器,用于存储计算机可读指令;以及处理器,用于运行所述计算机可读指令,使得所述处理器运行时实现根据权利要求1-7中任意一项所述的对象的显示方法。
- 一种非暂态计算机可读存储介质,用于存储计算机可读指令,当所述计算机可读指令由计算机执行时,使得所述计算机执行权利要求1-7中任意一项所述的对象的显示方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2022522321A JP7560207B2 (ja) | 2019-10-14 | 2020-08-04 | オブジェクトの表示方法、装置、電子機器及びコンピュータ可読型記憶媒体 |
US17/720,232 US11810336B2 (en) | 2019-10-14 | 2022-04-13 | Object display method and apparatus, electronic device, and computer readable storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910974919.3A CN110807728B (zh) | 2019-10-14 | 2019-10-14 | 对象的显示方法、装置、电子设备及计算机可读存储介质 |
CN201910974919.3 | 2019-10-14 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/720,232 Continuation US11810336B2 (en) | 2019-10-14 | 2022-04-13 | Object display method and apparatus, electronic device, and computer readable storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021073204A1 true WO2021073204A1 (zh) | 2021-04-22 |
Family
ID=69488395
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/106818 WO2021073204A1 (zh) | 2019-10-14 | 2020-08-04 | 对象的显示方法、装置、电子设备及计算机可读存储介质 |
Country Status (4)
Country | Link |
---|---|
US (1) | US11810336B2 (zh) |
JP (1) | JP7560207B2 (zh) |
CN (1) | CN110807728B (zh) |
WO (1) | WO2021073204A1 (zh) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110807728B (zh) * | 2019-10-14 | 2022-12-13 | 北京字节跳动网络技术有限公司 | 对象的显示方法、装置、电子设备及计算机可读存储介质 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108289180A (zh) * | 2018-01-30 | 2018-07-17 | 广州市百果园信息技术有限公司 | 根据肢体动作处理视频的方法、介质和终端装置 |
CN109391792A (zh) * | 2017-08-03 | 2019-02-26 | 腾讯科技(深圳)有限公司 | 视频通信的方法、装置、终端及计算机可读存储介质 |
CN110807728A (zh) * | 2019-10-14 | 2020-02-18 | 北京字节跳动网络技术有限公司 | 对象的显示方法、装置、电子设备及计算机可读存储介质 |
Family Cites Families (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005107848A (ja) * | 2003-09-30 | 2005-04-21 | Lic Corporation:Kk | 子供画像生成装置 |
JP4448983B2 (ja) * | 2004-08-05 | 2010-04-14 | 株式会社国際電気通信基礎技術研究所 | 画像合成装置 |
GB2465791A (en) * | 2008-11-28 | 2010-06-02 | Sony Corp | Rendering shadows in augmented reality scenes |
CN101477690B (zh) * | 2008-12-30 | 2011-04-27 | 清华大学 | 一种视频帧序列中物体轮廓跟踪的方法和装置 |
JP5024465B2 (ja) * | 2010-03-26 | 2012-09-12 | 株式会社ニコン | 画像処理装置、電子カメラ、画像処理プログラム |
JP5684507B2 (ja) * | 2010-07-30 | 2015-03-11 | 株式会社バンダイナムコゲームス | 撮影装置、プログラム、情報記憶媒体及び撮影方法 |
JP5304751B2 (ja) * | 2010-08-06 | 2013-10-02 | フリュー株式会社 | 画像作成装置および方法、並びにプログラム |
US8527445B2 (en) * | 2010-12-02 | 2013-09-03 | Pukoa Scientific, Llc | Apparatus, system, and method for object detection and identification |
JP5912062B2 (ja) * | 2012-05-24 | 2016-04-27 | オリンパス株式会社 | 撮影機器及び動画像データの記録方法 |
JP5962547B2 (ja) * | 2013-03-08 | 2016-08-03 | ソニー株式会社 | 情報処理装置、情報処理方法およびプログラム |
JP5505751B2 (ja) * | 2013-11-11 | 2014-05-28 | フリュー株式会社 | 画像編集装置および方法、並びにプログラム |
US9947139B2 (en) * | 2014-06-20 | 2018-04-17 | Sony Interactive Entertainment America Llc | Method and apparatus for providing hybrid reality environment |
US20160012594A1 (en) * | 2014-07-10 | 2016-01-14 | Ditto Labs, Inc. | Systems, Methods, And Devices For Image Matching And Object Recognition In Images Using Textures |
US10007996B2 (en) * | 2015-03-02 | 2018-06-26 | Lawrence Livermore National Security, Llc | System for detecting objects in streaming 3D images formed from data acquired with a medium penetrating sensor |
CN107852443B (zh) * | 2015-07-21 | 2020-01-07 | 索尼公司 | 信息处理设备、信息处理方法和程序 |
CN105335064B (zh) * | 2015-09-29 | 2017-08-15 | 腾讯科技(深圳)有限公司 | 一种信息处理方法和终端 |
US10152630B2 (en) * | 2016-08-09 | 2018-12-11 | Qualcomm Incorporated | Methods and systems of performing blob filtering in video analytics |
US20180254065A1 (en) * | 2017-03-03 | 2018-09-06 | Qualcomm Incorporated | Methods and systems for splitting non-rigid objects for video analytics |
CN107633526B (zh) * | 2017-09-04 | 2022-10-14 | 腾讯科技(深圳)有限公司 | 一种图像跟踪点获取方法及设备、存储介质 |
CN107993191B (zh) * | 2017-11-30 | 2023-03-21 | 腾讯科技(深圳)有限公司 | 一种图像处理方法和装置 |
CN109068053B (zh) * | 2018-07-27 | 2020-12-04 | 香港乐蜜有限公司 | 图像特效展示方法、装置和电子设备 |
CN108958610A (zh) * | 2018-07-27 | 2018-12-07 | 北京微播视界科技有限公司 | 基于人脸的特效生成方法、装置和电子设备 |
CN110058685B (zh) * | 2019-03-20 | 2021-07-09 | 北京字节跳动网络技术有限公司 | 虚拟对象的显示方法、装置、电子设备和计算机可读存储介质 |
CN110062176B (zh) * | 2019-04-12 | 2020-10-30 | 北京字节跳动网络技术有限公司 | 生成视频的方法、装置、电子设备和计算机可读存储介质 |
CN110213638B (zh) * | 2019-06-05 | 2021-10-08 | 北京达佳互联信息技术有限公司 | 动画显示方法、装置、终端及存储介质 |
CN110321865A (zh) * | 2019-07-09 | 2019-10-11 | 北京字节跳动网络技术有限公司 | 头部特效处理方法及装置、存储介质 |
-
2019
- 2019-10-14 CN CN201910974919.3A patent/CN110807728B/zh active Active
-
2020
- 2020-08-04 WO PCT/CN2020/106818 patent/WO2021073204A1/zh active Application Filing
- 2020-08-04 JP JP2022522321A patent/JP7560207B2/ja active Active
-
2022
- 2022-04-13 US US17/720,232 patent/US11810336B2/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109391792A (zh) * | 2017-08-03 | 2019-02-26 | 腾讯科技(深圳)有限公司 | 视频通信的方法、装置、终端及计算机可读存储介质 |
CN108289180A (zh) * | 2018-01-30 | 2018-07-17 | 广州市百果园信息技术有限公司 | 根据肢体动作处理视频的方法、介质和终端装置 |
CN110807728A (zh) * | 2019-10-14 | 2020-02-18 | 北京字节跳动网络技术有限公司 | 对象的显示方法、装置、电子设备及计算机可读存储介质 |
Non-Patent Citations (2)
Title |
---|
LI XIAO,BAI WAN, LI BAI: "Xiao Hu Cat Sates: That's How You Couples Make Heart Shape with Arms, hmm...I'm an outsider. Love Looks Like Us", 16 May 2019 (2019-05-16), CN, pages 1 - 1, XP054981800, Retrieved from the Internet <URL:https://www.iqiyi.com/w_19s99nbngt.html> * |
LIAN YOU , XIAO CAO: "How to Film "Looks Like Love" and How to Add Special Effect of "Come Closer" on TikTok", 13 February 2019 (2019-02-13), pages 1 - 2, XP009527513, Retrieved from the Internet <URL:https://jingyan.baidu.com/article/ad310e8006b33f1849f49eb1.html> * |
Also Published As
Publication number | Publication date |
---|---|
JP7560207B2 (ja) | 2024-10-02 |
US11810336B2 (en) | 2023-11-07 |
CN110807728A (zh) | 2020-02-18 |
US20220245920A1 (en) | 2022-08-04 |
CN110807728B (zh) | 2022-12-13 |
JP2022551671A (ja) | 2022-12-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7181375B2 (ja) | 目標対象の動作認識方法、装置及び電子機器 | |
WO2023051185A1 (zh) | 图像处理方法、装置、电子设备及存储介质 | |
WO2021254502A1 (zh) | 目标对象显示方法、装置及电子设备 | |
WO2022007565A1 (zh) | 增强现实的图像处理方法、装置、电子设备及存储介质 | |
WO2016192325A1 (zh) | 视频文件的标识处理方法及装置 | |
US11734804B2 (en) | Face image processing method and apparatus, electronic device, and storage medium | |
CN110796664B (zh) | 图像处理方法、装置、电子设备及计算机可读存储介质 | |
WO2021139382A1 (zh) | 人脸图像的处理方法、装置、可读介质和电子设备 | |
WO2021170013A1 (zh) | 图像特效处理方法及装置 | |
CN115002359B (zh) | 视频处理方法、装置、电子设备及存储介质 | |
WO2021104130A1 (zh) | 在视频中显示对象的方法、装置、电子设备及计算机可读存储介质 | |
WO2020244553A1 (zh) | 字幕越界的处理方法、装置和电子设备 | |
US12019669B2 (en) | Method, apparatus, device, readable storage medium and product for media content processing | |
CN110070496A (zh) | 图像特效的生成方法、装置和硬件装置 | |
WO2023169305A1 (zh) | 特效视频生成方法、装置、电子设备及存储介质 | |
WO2021027596A1 (zh) | 图像特效处理方法、装置、电子设备和计算机可读存储介质 | |
US11880919B2 (en) | Sticker processing method and apparatus | |
WO2023138441A1 (zh) | 视频生成方法、装置、设备及存储介质 | |
WO2021227953A1 (zh) | 图像特效配置方法、图像识别方法、装置及电子设备 | |
WO2021073204A1 (zh) | 对象的显示方法、装置、电子设备及计算机可读存储介质 | |
WO2020155908A1 (zh) | 用于生成信息的方法和装置 | |
WO2023025181A1 (zh) | 图像识别方法、装置和电子设备 | |
CN113905177B (zh) | 视频生成方法、装置、设备及存储介质 | |
US12022162B2 (en) | Voice processing method and apparatus, electronic device, and computer readable storage medium | |
WO2021027632A1 (zh) | 图像特效处理方法、装置、电子设备和计算机可读存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20877198 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2022522321 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 03.08.2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20877198 Country of ref document: EP Kind code of ref document: A1 |