CN116175541A - Grabbing control method, grabbing control device, electronic equipment and storage medium - Google Patents
Grabbing control method, grabbing control device, electronic equipment and storage medium Download PDFInfo
- Publication number
- CN116175541A CN116175541A CN202111429141.1A CN202111429141A CN116175541A CN 116175541 A CN116175541 A CN 116175541A CN 202111429141 A CN202111429141 A CN 202111429141A CN 116175541 A CN116175541 A CN 116175541A
- Authority
- CN
- China
- Prior art keywords
- orientation
- value
- grabbing
- grabbed
- article
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 80
- 238000003860 storage Methods 0.000 title claims abstract description 16
- 230000001629 suppression Effects 0.000 claims abstract description 72
- 230000005764 inhibitory process Effects 0.000 claims abstract description 38
- 238000004590 computer program Methods 0.000 claims description 9
- 230000000694 effects Effects 0.000 abstract description 14
- 239000000463 material Substances 0.000 description 40
- 239000011159 matrix material Substances 0.000 description 25
- 238000012545 processing Methods 0.000 description 18
- 230000008569 process Effects 0.000 description 14
- 230000006870 function Effects 0.000 description 13
- 238000004519 manufacturing process Methods 0.000 description 9
- 238000010586 diagram Methods 0.000 description 8
- 235000013305 food Nutrition 0.000 description 8
- 238000004364 calculation method Methods 0.000 description 7
- 238000004891 communication Methods 0.000 description 7
- 235000013361 beverage Nutrition 0.000 description 6
- 238000012163 sequencing technique Methods 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 4
- 238000003491 array Methods 0.000 description 3
- 230000004888 barrier function Effects 0.000 description 3
- 230000007547 defect Effects 0.000 description 3
- 230000002829 reductive effect Effects 0.000 description 3
- 238000003466 welding Methods 0.000 description 3
- 238000007405 data analysis Methods 0.000 description 2
- 239000012636 effector Substances 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 230000001788 irregular Effects 0.000 description 2
- 230000000670 limiting effect Effects 0.000 description 2
- 239000008267 milk Substances 0.000 description 2
- 210000004080 milk Anatomy 0.000 description 2
- 235000013336 milk Nutrition 0.000 description 2
- 238000005096 rolling process Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 241000287196 Asthenes Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000003592 biomimetic effect Effects 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000000739 chaotic effect Effects 0.000 description 1
- 238000004140 cleaning Methods 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 239000002537 cosmetic Substances 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000002401 inhibitory effect Effects 0.000 description 1
- 239000007788 liquid Substances 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000003801 milling Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000011664 nicotinic acid Substances 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 238000011022 operating instruction Methods 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 238000004806 packaging method and process Methods 0.000 description 1
- 230000036961 partial effect Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 235000011888 snacks Nutrition 0.000 description 1
- 239000007921 spray Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1612—Programme controls characterised by the hand, wrist, grip control
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1602—Programme controls characterised by the control system, structure, architecture
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/02—Total factory control, e.g. smart factories, flexible manufacturing systems [FMS] or integrated manufacturing systems [IMS]
Landscapes
- Engineering & Computer Science (AREA)
- Robotics (AREA)
- Mechanical Engineering (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Orthopedic Medicine & Surgery (AREA)
- Automation & Control Theory (AREA)
- Manipulator (AREA)
Abstract
The application discloses a grabbing control method, a grabbing control device, electronic equipment and a storage medium. The grabbing control method comprises the following steps: acquiring position characteristics and orientation characteristics of an object to be grabbed; determining an orientation suppression value of the object to be grabbed based on the position feature and the orientation feature; wherein the orientation suppression value is such that the gripping feature value when the orientation of the article is away from the obstacle is greater than the gripping feature value when the orientation of the article is directed toward the obstacle; determining a grabbing characteristic value of the object to be grabbed based on the position characteristic of the object to be grabbed, the orientation characteristic and the orientation inhibition value; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped. The invention is specially used for a scene that a large number of articles are scattered and placed near the obstacle which affects grabbing, so that the articles can obtain grabbing characteristic values higher than the articles when the articles face the obstacle when the articles face away from the obstacle, and then the clamp is controlled to grab based on the grabbing characteristic values, so that the grabbing effect is good.
Description
Technical Field
The present application relates to the field of automatic control of a robot arm or a gripper, program control B25J, and more particularly, to a gripping control method, apparatus, electronic device, and storage medium.
Background
At present, in the fields of markets, supermarkets, logistics and the like, robots are gradually used to replace manual operations for sorting, carrying, placing and the like of goods, and the traditional robots are limited to only operate in a predetermined mode or operate in a limited intelligent mode, so that in the scenes, the positions and the placing of the objects to be operated have higher requirements. For example, for sorting tasks in a supermarket, the task requirement is to take out the articles to be sorted placed in the material frame and carry them to a specified location. In this task, the robot visually recognizes the position of each article in the material frame and takes out the article and places it at a designated position, in such a task, in order to ensure that the robot can grasp each article smoothly, the existing scheme requires that the worker first place the article in the material frame neatly, and each article in the material frame needs to be placed in a specific posture, for example, the canned drink, the boxed food, the bagged food, etc., all require that the opening face up, and then the material frame with a large number of articles placed neatly is transported to the robot work area, and the grasping work is performed by the robot.
For such a scenario that a large number of articles need to be sorted and grabbed to a designated position, after all the articles are identified, conventional schemes generally need to determine the grabbing order of the articles based on the height of the articles or the size of a clamp configured to grab the articles, and control the clamp to grab the articles based on the determined order, and in some schemes, whether the articles are folded or not may also be considered in determining the grabbing order. However, in the above-mentioned grabbing scenario, if a large number of articles are not placed in a regular posture, but stacked together in a disordered and scattered manner, when the existing grabbing scheme is used, unexpected situations of pushing the articles over, carrying the articles, and even failing to grab the articles successfully occur, especially when the articles to be grabbed are located beside the frame wall or other particularly high objects and other obstacles, the movement of the clamp and the clamping process of the clamp are blocked due to the obstacles, and the existing grabbing control scheme does not consider such an operation scenario, so that the grabbing effect is poor. Therefore, a gripping control scheme with high gripping success rate is needed to solve various problems possibly occurring in the scene that the dense articles are scattered and placed and the articles are gripped by using the clamp.
Disclosure of Invention
The present invention has been made in view of the above problems, and aims to overcome or at least partially solve the above problems. Specifically, firstly, when the gripping scheme of the invention is used for gripping by controlling the clamp, the orientation characteristics of the objects to be gripped are considered, compared with the existing scheme, the gripping difficulty of the objects with different orientations can be more accurately determined, the possibility of gripping failure is reduced, particularly when the gripping is performed in an industrial scene with a large number of objects being scattered and piled up, the existing scheme has poor robot operation effect in such a scene because the influence of the orientation characteristics of the objects is not considered, and the invention can greatly improve the robot gripping effect in such a scene; secondly, the invention also provides a grabbing control scheme which comprehensively considers the orientation characteristics of the articles and the position characteristics of the articles, when grabbing, firstly judging whether the articles are in an easy-to-grab area, and based on whether the articles are in the area or not, adopting different grabbing schemes, so that under certain scenes, for example, when a large number of articles are scattered in a container or a large number of articles are located in an area with firm barriers which can influence grabbing, compared with the scheme which only considers the orientation characteristics of the articles, the grabbing sequencing is more accurate, and the grabbing effect of a robot is further improved; thirdly, the invention provides a scheme for determining whether the object to be grabbed is positioned in the specific area based on the numerical value, and the position suppression value is preset, so that whether the object to be grabbed is positioned in the specific area can be determined only based on the position characteristic value of the object to be grabbed and the relation between the position characteristic value and the position suppression value; finally, the invention also provides a method for controlling the clamp and executing the grabbing under the scene that a large number of articles are scattered and placed near the obstacle which affects grabbing, the method determines the grabbing characteristic value of the articles in a numerical mode based on the orientation inhibition value, so that the grabbing characteristic value of the articles when the articles deviate from the obstacle is higher than that of the articles when the articles face the obstacle, and the clamp is controlled to grab based on the grabbing characteristic value, so that the clamp can grab the articles which are least prone to grabbing failure first, and the grabbing effect is improved.
All of the solutions disclosed in the claims and the description of the present application have one or more of the innovations described above, and accordingly, one or more of the technical problems described above can be solved. Specifically, the application provides a grabbing control method, a grabbing control device, electronic equipment and a storage medium.
The grabbing control method of the embodiment of the application comprises the following steps:
acquiring position characteristics and orientation characteristics of an object to be grabbed;
determining an orientation suppression value of the object to be grabbed based on the position feature and the orientation feature; wherein the orientation suppression value is such that the gripping feature value when the orientation of the article is away from the obstacle is greater than the gripping feature value when the orientation of the article is directed toward the obstacle;
determining a grabbing characteristic value of the object to be grabbed based on the position characteristic of the object to be grabbed, the orientation characteristic and the orientation inhibition value; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped.
In some embodiments, the orientation suppression value is greater when the orientation of the item is away from the obstacle than when the orientation of the item is directed toward the obstacle.
In some embodiments, the determining the gripping feature value of the article to be gripped includes: and respectively calculating the grabbing characteristic value of the X axis and the grabbing characteristic value of the Y axis of the object to be grabbed, and taking the larger one as the grabbing characteristic value of the object.
In certain embodiments, further comprising: and scaling the orientation inhibition value and/or the grabbing characteristic value.
The grasping control device according to an embodiment of the present application includes:
the characteristic acquisition module is used for acquiring the position characteristic and the orientation characteristic of the object to be grabbed;
the orientation inhibition value determining module is used for determining an orientation inhibition value of the object to be grabbed based on the position characteristics and the orientation characteristics; wherein the orientation suppression value is such that the gripping feature value when the orientation of the article is away from the obstacle is greater than the gripping feature value when the orientation of the article is directed toward the obstacle;
the grabbing feature value determining module is used for determining grabbing feature values of the to-be-grabbed objects based on the position features of the to-be-grabbed objects, the orientation features and the orientation inhibition values; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped.
In some embodiments, the orientation suppression value is greater when the orientation of the item is away from the obstacle than when the orientation of the item is directed toward the obstacle.
In some embodiments, the determining the gripping feature value of the article to be gripped includes: and respectively calculating the grabbing characteristic value of the X axis and the grabbing characteristic value of the Y axis of the object to be grabbed, and taking the larger one as the grabbing characteristic value of the object.
In certain embodiments, further comprising: and scaling the orientation inhibition value and/or the grabbing characteristic value.
The electronic device of the embodiment of the application comprises a memory, a processor and a computer program stored on the memory and capable of running on the processor, wherein the processor realizes the grabbing control method of any embodiment when executing the computer program.
The computer-readable storage medium of the embodiments of the present application has stored thereon a computer program which, when executed by a processor, implements the grab control method of any of the embodiments described above.
Additional aspects and advantages of the application will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the application.
Drawings
The foregoing and/or additional aspects and advantages of the present application will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings, in which:
fig. 1 is a schematic diagram of a relationship between an object to be grasped and a material frame in an object sorting scene of the present application;
FIG. 2 is a flow chart of a method of article orientation-based grip control according to certain embodiments of the present application;
FIG. 3 is a schematic illustration of frame parameters according to certain embodiments of the present application;
FIG. 4 is a schematic illustration of mask pretreatment according to certain embodiments of the present application;
FIG. 5 is a schematic illustration of pitch, roll and yaw axes associated with a rotation matrix;
FIG. 6 is a flow chart of a method of gripping control based on relative positional relationship and orientation of items according to certain embodiments of the present application;
FIG. 7 is a schematic diagram of a camera coordinate system directly above a frame as a reference coordinate system;
FIG. 8 is a flow chart of a method of determining the location of an item according to certain embodiments of the present application;
FIG. 9 is a flow chart of a method of grip control for an object beside an obstacle according to some embodiments of the present application;
FIG. 10 is a schematic structural view of an article orientation-based grip control device according to certain embodiments of the present application;
FIG. 11 is a schematic structural view of a gripping control device based on relative article positioning and orientation in accordance with certain embodiments of the present disclosure;
FIG. 12 is a schematic structural view of an apparatus for determining the location of an article according to certain embodiments of the present application;
FIG. 13 is a schematic structural view of a grip control device for objects beside an obstacle according to certain embodiments of the present application;
Fig. 14 is a schematic structural diagram of an electronic device according to some embodiments of the present application.
Detailed Description
Exemplary embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the disclosure to those skilled in the art.
In the description of the specific embodiments, it should be understood that the terms "center," "longitudinal," "transverse," "upper," "lower," "front," "rear," "left," "right," "vertical," "horizontal," "top," "bottom," "inner," "outer," and the like indicate orientations or positional relationships based on the orientation or positional relationships shown in the drawings, merely to facilitate describing the invention and simplify the description, and do not indicate or imply that the devices or elements referred to must have a specific orientation, be configured and operated in a specific orientation, and therefore should not be construed as limiting the invention.
Furthermore, the terms "first," "second," "third," and the like are used for descriptive purposes only and are not to be construed as indicating or implying a relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defining "a first", "a second", etc. may explicitly or implicitly include one or more such feature. In the description of the present invention, unless otherwise indicated, the meaning of "a plurality" is two or more.
The invention can be used in industrial robot control scenes based on visual identification. A typical vision-based industrial robot control scenario includes devices for capturing images, control devices such as hardware for a production line and a PLC for the production line, robot components for performing tasks, and operating systems or software for controlling these devices. The means for capturing images may include a 2D or 3D smart/non-smart industrial camera, which may include an area camera, a line camera, a black and white camera, a color camera, a CCD camera, a CMOS camera, an analog camera, a digital camera, a visible light camera, an infrared camera, an ultraviolet camera, etc., depending on different functions and application scenarios; the production line can comprise a packaging production line, a sorting production line, a logistics production line, a processing production line and the like which need robots; the robot parts used in the industrial scene for performing tasks may be biomimetic robots, such as a human-type robot or a dog-type robot, or may be conventional industrial robots, such as a mechanical arm, etc.; the industrial robot may be an operation type robot, a program controlled type robot, a teaching reproduction type robot, a numerical control type robot, a sensory control type robot, an adaptation control type robot, a learning control type robot, an intelligent robot, or the like; the mechanical arm can be a ball-and-socket type mechanical arm, a multi-joint mechanical arm, a rectangular coordinate mechanical arm, a cylindrical coordinate mechanical arm, a polar coordinate mechanical arm and the like according to the working principle, and can be a grabbing mechanical arm, a stacking mechanical arm, a welding mechanical arm and an industrial mechanical arm according to the functions of the mechanical arm; the end of the mechanical arm can be provided with an end effector, and the end effector can use a robot clamp, a robot gripper, a robot tool quick-change device, a robot collision sensor, a robot rotary connector, a robot pressure tool, a compliance device, a robot spray gun, a robot burr cleaning tool, a robot arc welding gun, a robot electric welding gun and the like according to the requirements of tasks; the robot clamp can be various universal clamps, and the universal clamps refer to clamps with standardized structures and wide application range, such as a three-jaw chuck and a four-jaw chuck for a lathe, a flat tongs and an index head for a milling machine, and the like. For another example, the clamp may be classified into a manual clamp, a pneumatic clamp, a hydraulic clamp, a gas-liquid linkage clamp, an electromagnetic clamp, a vacuum clamp, etc. or other bionic devices capable of picking up an article, according to a clamping power source used for the clamp. The device for collecting images, the control devices such as hardware for a production line, a PLC (programmable logic controller) for the production line and the like, the robot parts for executing tasks and the operating system or software for controlling the devices can communicate based on TCP (transmission control protocol), HTTP (hyper text transfer protocol) and GRPC (generic personal computer) protocols (Google Remote Procedure Call Protocol ) so as to transmit various control instructions or commands. The operating system or software may be disposed in any electronic device, typically such electronic devices include industrial computers, personal computers, notebook computers, tablet computers, cell phones, etc., which may communicate with other devices or systems by wired or wireless means. Further, the gripping appearing in the present invention refers to any gripping action capable of controlling an article to change the position of the article in a broad sense, and is not limited to gripping the article in a narrow sense in a "gripping" manner, in other words, gripping the article in a manner such as suction, lifting, tightening, or the like, and also falls within the scope of the gripping of the present invention. The articles to be grasped in the present invention may be cartons, plastic soft packs (including but not limited to snack packages, milk tetra pillow packages, milk plastic packages, etc.), cosmeceutical bottles, cosmeceuticals, and/or irregular toys, etc., which may be placed in a floor, tray, conveyor belt, and/or material basket.
The inventors have found that in a dense, random-to-stack scenario, such as stacking a plurality of articles to be grasped in a material frame, the existing solution suffers because each article has a unique orientation, as shown in fig. 1, assuming that it is necessary to grasp a plurality of articles from a deep material frame, some of these articles have their graspable areas facing the material frame opening and some of these articles have their graspable areas facing the material frame wall, it is apparent that the articles facing the material frame opening will grasp better, especially when the articles are located near the material frame edge and the graspable areas are facing the material frame wall beside it, and the clamp may not grasp the articles even. In the existing scheme, when the grabbing sequence is determined, the influence of the size of the sucker on the grabbing difficulty degree is not considered, namely the object grade gesture is not considered, or more specifically, the influence of the object orientation on the grabbing difficulty degree is not considered, so that the grabbing effect in a grabbing scene of dense objects and scattered stacking is poor, and the grabbing effect on the objects is greatly influenced in such a scene. The inventors have thus found that a solution to this technical problem is to control gripping at least in dependence of the orientation characteristics of the article.
Fig. 2 shows a flow diagram of a method of controlling gripping of an item based on the pose orientation of the item to be gripped, according to an embodiment of the invention. As shown in fig. 2, the method includes:
step S100, obtaining image data comprising at least one object to be grabbed;
step S110, processing the image data to acquire orientation characteristics of the object to be grabbed, wherein the orientation characteristics relate to the orientation;
and step S120, controlling the clamp to perform grabbing of at least one object to be grabbed at least based on the orientation characteristics of the object to be grabbed.
With respect to step S100, the type of image data and the acquisition method are not limited in the present embodiment. As an example, the acquired image data may include a point cloud or an RGB color map, the point cloud information may be acquired through a 3D industrial camera, and the 3D industrial camera is generally equipped with two lenses, which capture the object group to be grabbed from different angles, respectively, and the three-dimensional image of the object can be displayed after processing. And placing the object group to be grabbed below the vision sensor, shooting by two lenses at the same time, and calculating X, Y, Z coordinate values of each point and coordinate directions of each point of the glass to be glued by using a universal binocular stereoscopic vision algorithm according to the obtained relative attitude parameters of the two images so as to convert the X, Y, Z coordinate values and the coordinate directions of each point into point cloud data of the object group to be grabbed. In the specific implementation, the point cloud can also be generated by using elements such as a laser detector, a visible light detector such as an LED, an infrared detector, a radar detector and the like, and the specific implementation of the invention is not limited.
The point cloud data acquired in the mode is three-dimensional data, so that the data corresponding to the dimension with small influence on grabbing is filtered, the data processing amount is reduced, the data processing speed is further increased, the efficiency is improved, and the acquired three-dimensional point cloud data of the object group to be grabbed can be orthographically mapped to a two-dimensional plane.
As an example, a depth map corresponding to the orthographic projection may also be generated. A two-dimensional color map corresponding to the three-dimensional object region and a depth map corresponding to the two-dimensional color map may be acquired in a direction perpendicular to the depth of the object. Wherein the two-dimensional color map corresponds to an image of a planar area perpendicular to a preset depth direction; each pixel point in the depth map corresponding to the two-dimensional color map corresponds to each pixel point in the two-dimensional color map one by one, and the value of each pixel point is the depth value of the pixel point.
Articles to be grasped are often piled in boxes for transportation to the site, and such boxes for piled articles are often called material frames, and when grasping is performed, a mechanical arm or a clamp may touch the material frames during movement, so that the material frames and the placement positions of the articles in the material frames have important influence on grasping. As a preferred embodiment, parameters of the frame may be obtained. As shown in fig. 3, the frame data may be processed to extract or generate auxiliary parameters that have an effect on grabbing, such parameters including: the height of the material frame, the width of the material frame, the length of the material frame, and the grid obtained by dividing the width and the length of the material frame. It should be understood that the height, width and length are all determined values, and the dividing mode and number of the grids are determined by the skilled person according to the actual conditions of the used fixture, the grabbing mode, the characteristics of the objects to be grabbed and the like, and the grids can be used for conveniently calibrating the positions of the objects to be grabbed. The frame data may be preset or acquired by a camera.
It will be appreciated that the article in the present invention may also be an article grippable region, since the gripper needs to perform a grip in the grippable region of the article at the time of actual gripping, the non-grippable region has no substantial effect on the grip. The grippable area of the article refers to a part on the surface of the article, which can be gripped by the clamp, in an industrial scene, the articles to be gripped can be placed in a orderly and orderly manner, and at the moment, the grippable area of each article is basically the same, and the manner of determining the grippable area is simpler; it is also possible to pile together in a chaotic and unordered manner, where the grippable area of each item is random and it is necessary to determine the grippable area in a complex manner. The present embodiment is not limited to a specific use scenario and a specific method of determining the graspable region, as long as the graspable region can be acquired.
One possible embodiment of determining the grabber area and generating the mask may be to first, after acquiring image data comprising one or more objects to be grabbed, process the image data to identify each pixel in the image, e.g. for a 256 x 256 image 256 x 65536 pixels should be identified; and classifying all the pixel points included in the whole image based on the characteristics of each pixel point, wherein the characteristics of the pixel points mainly refer to RGB values of the pixel points, and in an actual application scene, RGB color images can be processed into gray images for conveniently classifying the characteristics, and the gray images can be classified by using the gray values. For classification of the pixel points, it may be predetermined which class the pixel points need to be classified into, for example, a large stack of beverage cans, food boxes and frames is included in the RGB image obtained by photographing, so if the purpose is to generate a mask in which the beverage cans, food boxes and frames are to be generated, the predetermined classification may be beverage cans, food boxes and frames. The three different classifications can be provided with a label, wherein the label can be a number, for example, a beverage can is 1, a food box is 2, a material frame is 3, or the label can be a color, for example, a beverage can is red, a food box is blue, and a material frame is green, so that after the classification and the processing are carried out, the beverage can is marked with 1 or red, the food box is marked with 2 or blue, and the material frame is marked with 3 or green in a finally obtained image. In this embodiment, the mask of the grippable region of the object is to be generated, so that only the grippable region is classified, for example, blue, and the blue region in the image processed in this way is the mask of the grippable region of the object to be grippable; a channel of image output is then created for each class, the channel acting to extract as output all class-dependent features in the input image. For example, after we create a channel of image output for the class of grippable region, the acquired RGB color image is input into the channel, and then the image from which the features of the grippable region are extracted can be acquired from the output of the channel. Finally, the feature image of the grippable region obtained by the processing is combined with the original RGB image to generate the composite image data with the grippable region mask identified.
Masks generated in this manner are sometimes unsuitable, e.g., some masks are of a size and shape that is inconvenient to follow. For another example, some areas may have masks generated, but the clamps may not be able to perform a grab at the mask locations. An unsuitable mask can have a significant impact on subsequent processing, and therefore requires pretreatment of the resulting mask for further steps. As shown in fig. 4, the preprocessing of the mask may include: 1. and (3) performing expansion treatment on the mask to fill in defects such as missing and irregular mask images. For example, for each pixel point on the mask, a certain number of points, e.g., 8-25 points, around the point may be set to be the same color as the point. This step corresponds to filling the periphery of each pixel, so if there is a defect in the object mask, the missing part will be filled completely, after this, the object mask will become complete, there is no defect, and the mask will become slightly "fat" due to expansion, and proper expansion will help to follow-up further image processing operation; 2. judging whether the area of the mask meets the preset condition, and if not, eliminating the mask. First, smaller mask areas are likely to be erroneous because of the continuity of the image data, one grabbed area will typically include a large number of pixels with similar characteristics, and mask areas formed by discrete small pixels may not be truly grabbed areas; secondly, the robot end actuating mechanism, namely the clamp, needs to have a certain area in the foot falling position when the grabbing task is executed, if the area of the grabbing area is too small, the clamp cannot drop the foot in the area at all, and therefore the object cannot be grabbed, and therefore too small mask is meaningless. The predetermined condition may be set according to the size of the jig and the size of the noise, and the value thereof may be a determined size, or the number of the included pixels, or a ratio, for example, the predetermined condition may be set to 0.1%, that is, when the ratio of the mask area to the whole image area is less than 0.1%, the mask is considered to be unusable, and then is removed from the image; 3. and judging whether the number of the point clouds in the mask is less than the preset minimum number of the point clouds. The number of the point clouds reflects the quality of the acquisition of the camera, and if the number of the point clouds in a certain grippable area is too small, the shooting of the area is not accurate enough. The point cloud may be used to control the gripper to perform the gripping, and too small a number may have an impact on the gripper's control process. Thus, the number of point clouds that should be included at least in a certain mask area may be set, for example: and when the number of the point clouds covered in a certain grabbing area is less than 10, eliminating the mask from the image data or randomly adding the point clouds for the grabbing area until the number reaches 10.
The image, pose, rotation matrix, orientation, position, etc. of the object of the present invention may be an image, pose, rotation matrix, orientation, position, etc. of a graspable region of the object. The following schemes will not be described specifically based on "graspable areas of items", and those skilled in the art will be able to understand which "items" present in embodiments of the present invention may be replaced with "graspable areas of items".
For step S110, as shown in fig. 1, when the object is oriented directly above, it is most convenient for the gripper to grasp, and the more the orientation is toward the XY plane, the more difficult it is to grasp. The orientation characteristic of an article is used to reflect the degree to which the orientation of the article is biased toward the XY plane. The feature reflecting the direction or rotation of the article may be used as an orientation feature, such as an angle, or a particular projection value, etc., and is not limited in this embodiment. As a preferred embodiment, the orientation characteristics of the object may be obtained based on a rotation matrix of the object. When an article with a specific orientation rotates, the article is converted into another specific orientation, and the rotation matrix is used for expressing what kind of rotation is performed on the article. Essentially, the rotation matrix reflects the transformation relationship represented by coordinates in one coordinate system in another coordinate system.
In one embodiment, the reference article pose is assumed to have a right-side-up orientation, i.e., an orientation such that the graspable region of the article is perpendicular to the Z-axis, while the pose of the article to be grasped is obtained after rotation from the reference pose. Assuming that a rotation matrix from a reference pose to a current pose of an article isThen to be graspedThe orientation characteristics of the object can be obtained according to R. In one embodiment, the orientation feature of the object may be (X vector ,Y vector ,Z vector ) Wherein X is vector ,Y vector ,Z vector Values of the first, second, and third columns, respectively, of the third row of the rotation matrix, i.e. X vector =x 3 ,Y vector =y 3 ,Z vector =z 3 。
There are various forms of rotation matrices in the prior art, and the invention is not limited in this regard. Alternatively, the rotation matrix of the present invention may be a rotation matrix obtained based on euler angles. Any one rotation may be expressed as a combination of three angles, in turn, around three axes of rotation, which are known as euler angles. As shown in fig. 5, the rotation of an article is described by 3 rotation components, which can be understood as an X-axis, a Y-axis and a Z-axis in a cartesian coordinate system, wherein the X-axis is a pitch axis, and the clockwise rotation angle along the X-axis is a pitch angle, denoted as α; the Y axis is a yaw axis, and the clockwise rotation angle along the Y axis is a yaw angle and is marked as beta; the Z axis is a rolling axis, and the angle along the Z axis rotating clockwise is a rolling angle and is marked as gamma. Any one rotation can be considered a combination of three rotation means, for example, if an article is rotated in XYZ, this means that the article is rotated clockwise by a along the X axis, then by β along the Y axis, and finally by γ along the Z axis. The rotation matrix is different in different rotation modes, and the total rotation modes are 12 rotation modes. Preferably, the article can be rotated from the reference direction to the current state in a ZYX manner, and accordingly, the rotation matrix of the article to be grasped can be
For step S120, in one embodiment, based on the orientation feature of the article and the jig used, the pose of the jig when performing gripping, including the rotation angle of the jig, the pose of the jig, etc., may be calculated to control the jig to perform gripping of the article at a certain angle or pose in the grippable region of the article. In another embodimentIn an embodiment, the orientation feature value calculated by the orientation feature of at least one object to be grabbed may also be used for sorting the grabbing difficulty of the plurality of objects to be grabbed, that is, sorting all objects to be grabbed based on the obtained orientation feature value, and controlling the clamp to grab according to the sorting order. Preferably, when an object has an orientation feature (X vector ,Y vector ,Z vector ) The orientation characteristic value of the object can be Max { X vector ,Y vector ,Z vector }。
The solution disclosed in the above embodiments can be used for gripping items placed in any way, however in an industrial scenario as shown in fig. 1, the items to be gripped are placed in a deeper material frame. Other articles scattered around the articles to be grabbed are generally small in obstacle caused by grabbing, and the clamp can push the articles away in the grabbing process or slightly deform the articles to be grabbed, so that the grabbing process is not affected obviously. Unlike the easy or other stronger barriers like deep material frames, which have high walls and are difficult to move or deform, the clamps can be obstructed from moving and grabbing, and even can fail to grab. As shown in fig. 1, if the object to be grabbed is located near the frame wall of the material frame and the grabbed area of the object to be grabbed faces the frame wall, the clamp is likely to collide with the frame wall during grabbing, so that grabbing is failed; however, if the object to be grasped is located in the central area of the material frame, grasping difficulty is uniform regardless of the direction to which it is directed. Thus, in a similar scenario, it is not sufficient to consider only the orientation characteristics of the item when gripping. In order to solve the problem, the inventor has studied and proposed a method for controlling grabbing based on the comprehensive pose of the object, namely the position of the object and the orientation of the object, which is also one of the important points of the invention.
Fig. 6 shows a flow diagram of a method of controlling gripping of an item based on the position and orientation of the item to be gripped, according to one embodiment of the invention. As shown in fig. 6, the method at least comprises the following steps:
step S200, obtaining the position characteristics and the orientation characteristics of the object to be grabbed;
step S210, based on the position characteristics, determining the relation between the position of the object to be grabbed and the interest area of the reference area where the object to be grabbed is located;
step S220, determining a grabbing characteristic value based on the relation between the position of the object to be grabbed and the region of interest and the orientation characteristic of the object to be grabbed; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped.
For step S200, the orientation feature of the object may be obtained in a similar manner to step S110, which is not described herein. The positional characteristic of the object may be the coordinates of the object in a reference coordinate system. The reference coordinate system may be a world coordinate system, i.e., a coordinate system established with a point in the real world as an origin; the camera coordinate system, namely, a coordinate system established by taking the optical center of the camera as an origin, wherein the z axis of the coordinate system points to the right front of the camera; or an image coordinate system which is established by taking the projection of the camera optical center on the imaging plane as an origin. The present invention preferably uses a camera coordinate system with a camera as an origin as a reference coordinate system. As shown in fig. 7, in the present invention, the camera takes a picture right above the middle of the deep frame, the origin of the reference coordinate system is the camera, that is, the point with coordinates (0, 0) is within the camera, under this reference coordinate system, the point located on the left of the camera has negative coordinate values of the X axis, the point located on the rear of the camera has negative coordinate values of the Y axis, the point located below the camera has negative coordinate values of the Z axis. Assume that the coordinates of the current position of an article are (X pose ,Y pose ,Z pose ) The location characteristic of the item may be (X pose ,Y pose ,Z pose )。
For step S210, although the invention is described by taking a frame as an example, in an actual industrial scenario, the articles may be placed in other containers, such as a bucket. The region of Interest (Area of Interest) refers to a region in the present embodiment in which gripping with a jig is easy, and a problem of gripping failure is less likely to occur, for example, an inner ring region of a material frame in fig. 10. The region of interest is typically the central region of the reference region, however, if a different container is used, or if the frame is also physically divided into a plurality of lattices, the region of interest may also be a region other than the central region. The reference region is a region associated with the region of interest, and the reference region may be a region of the entire container, or a partial region of the container. The specific reference area and the interest area can be determined according to the actual situation, the key point of the implementation is to judge whether the object is located in the interest area or not, control grabbing according to the judging result, not limit the method for determining the interest area, and determine the interest area in any mode, for example, take a central point as a circle center, take a fixed length as a radius as a circle, and the area in the circle is the interest area. Regardless of the orientation of the item, the region of interest itself has distinguished the difficulty of grasping the item, which, as previously described, is quite different whether the item is within or outside the region of interest. As shown in fig. 1, when the object to be grasped is located at the inner ring of the material frame, the influence of the orientation of the object on grasping is symmetrical in all directions, for example, the inner ring object is oriented at 45 degrees to the right side of the material frame or oriented at 45 degrees to the left side of the material frame, and the grasping difficulty hardly differs. The gripping difficulty of the article outside the inner ring is greatly different when the article faces the frame wall at an angle of 45 degrees and the gripping difficulty of the article faces the center of the frame at an angle of 45 degrees. Therefore, it is required to determine whether an object is located in an area of interest, and when determining whether an object is located in an area by using the existing method, an image data analysis method is generally adopted, for example, an image including the object to be grabbed and the area can be acquired, the positional relationship between the object to be grabbed and the area is identified, and whether the object to be grabbed is located in the area of interest is determined, however, because the image data analysis process is introduced in this way, the operation speed is slow. The inventor researches and invents a method for calculating whether an object to be grabbed is located in an interest area or not based on a numerical mode, and the method is not good in universality, but high in operation speed and high in accuracy, and is one of the key points of the invention.
Fig. 8 shows a flow diagram of a method of determining whether an item to be grabbed is located within a region of interest, in accordance with one embodiment of the present invention. As shown in fig. 8, the method at least comprises the following steps:
step S300, calculating a position inhibition value of an object based on the characteristics of a reference area, wherein the size of the position inhibition value is related to the size of a region of interest of the reference area;
step S310, obtaining a position characteristic value of each article in a plurality of articles;
step S320, for each article in the plurality of articles, determining whether the article is located in the interest area of the reference area based on the position characteristic value of the article, the position suppression value and the size of the reference area; the position characteristic value of the article comprises coordinate values of an X axis and coordinate values of a Y axis of the article under a reference coordinate system.
For step S300, the position suppression corresponds to suppression of the reference region based on the dimensional characteristics of the reference region, such as length, width, height, to define a numerical range of the region of interest associated with the reference region. If the location of the item is within the suppressed range of values, the item is considered to be within the region of interest. The position suppression value is a value for performing position suppression. In one embodiment, it is assumed that a plurality of objects to be grasped are placed in a material frame as shown in fig. 1, and the material frame has a length L and a width W, and the region of interest is a central region of the material frame. The position suppression value of the object can be calculated using the following formula:
Wherein X is inhibit_ratio Is the position inhibition value of X axis, Y inhibit_ratio Is the position inhibition value of the Y axis, A X For the position-inhibiting parameter of the X-axis, A Y Is a Y-axis position suppression parameter which, in this embodiment,A X and A Y The larger the value, the smaller the range considered as the center region. According to the requirement of actual grabbing scene, A X And A Y Any value can be taken. The inventor regards A in the scenario shown in FIG. 1 X And A Y Multiple tests were performed on the values of (2), wherein the best value is A X And A Y The values of (2) are all 0.1.
For step S310, the location features of the object may be acquired in a similar manner to step S200, which will not be described herein. Which position characteristic values are specifically adopted for calculation, the position characteristic values need to be determined according to a specific calculation method. If the position of an article is characterized by (X pose ,Y pose ,Z pose ) And determining whether the object is located in the region of interest using the position suppression values calculated by formulas (1) and (2), then X is required pose And Y pose These two position feature values are subjected to subsequent processing.
For step S320, for each item to be grasped, it may be determined whether the item is located within the region of interest of the reference region by determining whether the position characteristic value of the item satisfies a suppression condition, which is associated with the position suppression value. Specifically, if the position suppression value obtained by the calculation of the formulas (1) and (2) is used to determine whether an object is located in the region of interest, for a certain object to be grasped, the following formula can be used to calculate whether it is located in the region of interest:
B inner_N =X condition_N &Y condition_N (5)
Wherein X is condition_N Is a judgment result of whether the X coordinate value of the Nth object satisfies the suppression condition, and the value of 1 represents the X coordinate value of the objectIf the inhibition condition is satisfied, otherwise, the inhibition condition is not satisfied; y is Y condition_N And X is a result of judging whether the Y-coordinate value of the Nth object meets the inhibition condition condition_N Similarly, a value of 1 indicates that the Y-coordinate value of the article satisfies the suppression condition, whereas, if not, it indicates that the Y-coordinate value of the article does not satisfy the suppression condition; x is X pose_N For X coordinate value, Y of the N-th object pose_N Is the Y coordinate value of the nth object. B (B) inner_N The value of 1 is the judgment result of whether the Nth object is positioned in the interest area, and if the N th object is positioned in the interest area, the judgment result indicates that the object is positioned in the interest area, otherwise, the judgment result indicates that the object is positioned outside the interest area; n is the number of the output object pose;&representing AND operation, the operation rule is 1&1=1,1&0=0,0&1=0,0&0=0. The meaning of the formula (5) is that, for an item, the item is considered to be located in the region of interest only if both the X coordinate value and the Y coordinate value of the location of the item satisfy the corresponding suppression conditions.
In one embodiment, the X coordinate values and the Y coordinate values of a plurality of objects can be calculated after the judgment result of whether the X coordinate values and the Y coordinate values are positioned in the interest area are combined into a set condition =[X condition_1 ,X condition_2 ,…,X condition_N ],Y condition =[Y condition_1 ,Y condition_2 ,…,Y condition_N ]Then calculate B according to formula (5) inner =[B inner_1 ,B inner_2 ,…,B inner_N ]. For example, assuming that there are 5 articles or 5 positions to be grasped in one grasping task, for a combination of the 5 articles, among the five articles, the X-axis coordinates of the positions of the second article and the fifth article satisfy the suppression condition, and the Y-axis coordinates of the first, second and third articles satisfy the suppression condition, there is X condition =[0,1,0,0,1],Y condition =[1,1,1,0,0]Then B is calculated according to equation (5) to obtain this combination inner =[0,1,0,0,0]This means that only the second of these 5 items is located within the region of interest.
As for step S220, as described above, in this embodiment, the gripping difficulty of the object located in the region of interest is completely different from that of the object located outside the region of interest, so that the calculation modes of the gripping feature values are also completely different, the magnitude of the gripping feature values reflects the gripping difficulty, and the larger the gripping feature values in the present invention indicates the easier the gripping. For any object to be grasped, the following formula can be used to calculate the grasping characteristic value:
R(N)=R inner_N +R outer_N (6)
wherein R (N) is the grabbing characteristic value of the N-th object to be grabbed, R inner_N Grabbing characteristic values in the interest area of the object, R outer_N The feature values are grabbed outside the region of interest of the item. For two components of R (N), where R inner_N The calculation is performed according to the following formula:
R inner_N =B inner_N *Z vector_N (7)
in the formula (7), B inner_N The calculation mode of the judgment result is shown in a formula (5) for judging whether the N object to be grabbed is positioned in the region of interest; z is Z vector_N The characteristic value of the Z axis of the N-th object to be grabbed corresponds to the projection of the orientation of the object on the Z axis. In one embodiment, the rotation matrix of the object to be grasped is assumed to beZ is then vector_N May be the component of the third column of the 3 rd row of the rotation matrix, i.e. z 3_N 。
For another component R of R (N) outer_N . It should be understood that the object is beside the material frame, which is essentially equivalent to the object being beside a certain obstacle forming an obstacle for grabbing, the inventor finds that no one in the prior art has discussed how to judge the grabbing difficulty of the objects according to the positions and orientations of the objects and determine the grabbing sequence when performing tasks of grabbing multiple objects for a plurality of objects beside the obstacle. To solve this problem, the inventors developed a set of gripping control schemes capable of quantifying the gripping difficulty of an article according to the position and orientation of the article, dedicated to gripping capable of affectingA plurality of objects to be grasped next to the grasped obstacle is one of the important points of the present invention.
Fig. 9 shows a flow diagram of a method of determining the ease of gripping of an item to be gripped beside an obstacle according to one embodiment of the invention. As shown in fig. 9, the method at least comprises the following steps:
step S400, obtaining the position characteristics and the orientation characteristics of the object to be grabbed;
step S410, calculating an orientation inhibition value of the object to be grabbed based on the position characteristics and the orientation characteristics; the orientation inhibition value enables the grabbing characteristic value when the orientation of the object deviates from the obstacle to be larger than the grabbing characteristic value when the orientation of the object points to the obstacle;
step S420, calculating a grabbing feature value of the object to be grabbed based on the position feature of the object to be grabbed, the orientation feature and the orientation inhibition value; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped.
For step S400, the position features and the orientation features of the object to be grasped may be obtained in a similar manner to step S200, which is not described herein again;
the orientation suppression in step S410 is to suppress the orientation feature value obtained when the article is oriented in the specific direction, and the orientation suppression may be to decrease the orientation feature value obtained in the specific direction or to increase the orientation feature value in a direction different from the specific direction. The value for performing orientation suppression is referred to as an orientation suppression value. In the scenario shown in fig. 1, the orientation suppression value may be calculated using the following formula:
X inhibit_N =Max(sign[X pose_N *X vector_N ],0)*B X +C X (8)
Y inhibit_N =Max(sign[Y pose_N *Y vector_N ],0)*B Y +C Y (9)
Wherein X is inhibit_N An orientation suppression value of the X axis of the nth article; y is Y inhibit_N As the N-th itemAn orientation inhibition value of the Y axis; max () is a maximum function, and the rule of Max (a, b) is the maximum value of a and b; sign []As a sign function, and "[ only ]]The sign of the "medium value" is related to the value, and is not related to the value, specifically, when "[]The value in "is negative, the value is-1, and" when [ is ]]The value in "is positive, its value is 1; b (B) X ,B Y ,C X And C Y Are all suppression amplitude adjustment parameters; x is X vector_N The characteristic value of the orientation of the X axis of the nth object to be grabbed is equivalent to the projection of the orientation of the object on the X axis; y is Y vectorN The characteristic value of the orientation of the Y axis of the nth object to be grasped corresponds to the projection of the orientation of the object on the Y axis. For the N-th article to be grabbed, the rotation matrix is set asThen X is vector_N Can be the component of the 3 rd row and 1 st column of the rotation matrix, namely X 3_N And Y is vector_N Can be the component of the 3 rd row and the 2 nd column of the rotation matrix, namely Y 3_N . Adjusting parameter B with respect to suppression amplitude X ,B Y ,C X And C Y The purpose of setting these parameters is mainly to adjust the suppression range of the direction, and avoid that the suppression value of the direction is too large or too small, so that the calculated grabbing characteristic value is too small or too large, and the value can be obtained according to a specific scene test, for example, 1,2 or 3. In the scenario shown in FIG. 1, the inventors performed several experiments on the values of these 4 parameters to obtain the best value B X =B Y =2,C X =C Y =1。
For step S420, after the orientation suppression values of the X-axis and the Y-axis are obtained, the grasping characteristic values of the X-axis and the Y-axis may be calculated, respectively, using the orientation suppression values. The grabbing feature values comprehensively consider the position features, the orientation features and the features of the reference area, and are calculated by combining the orientation inhibition values, and in the grabbing sorting process, the grabbing priority is higher as the score is higher, so that the grabbing feature values are calculated in a mode of comprehensively considering the parameters, and the objects with better orientation can obtain higher grabbing feature values. In one embodiment, the following formulas may be used to calculate the grabbing feature values for the X-axis and the Y-axis, respectively:
wherein X (N) is a grabbing characteristic value of the X axis of the Nth object to be grabbed; y is a grabbing characteristic value of the Y axis of the Nth article to be grabbed; d (D) X And D Y For the purpose of adjusting the parameters of the grasping characteristic value, the two parameters are used for adjusting the magnitude of the finally calculated grasping characteristic value to be in a region which is easy to understand and convenient to process, similar to the normalization of the grasping characteristic value, for example, the grasping characteristic value can be controlled to be [0,2 ] by the grasping characteristic value adjusting parameters]Is within the interval of (2). In the scenario shown in FIG. 1, the inventors performed several experiments on the values of these 2 parameters to obtain the best value D X =D Y = -1. In the scenario shown in fig. 1, the material frame wall exists in the direction pointed by the X axis and the direction pointed by the Y axis, and the gripper can select any one of the two directions to grip, so that the comprehensive gripping characteristic value of the article can be Max { X (N), Y (N) }, that is, the largest value of X (N) and Y (N) is taken. It should be appreciated that the solution of the present embodiment is specific to a scene where there is an obstacle beside the object to be grasped, and does not consider a general scene.
Next, returning to step S230, as described above, the object to be grabbed may be present in the region of interest of the reference region or may be present outside the region of interest of the reference region. For a certain object to be grabbed, calculating its characteristic value based on formula (6), namely calculating the characteristic value in the area and the characteristic value outside the area of the object respectively, and adding the characteristic values to obtain a grabbed characteristic value, wherein R inner_N R is calculated by adopting a formula (7) outer_N The calculation can be performed using the following formula:
R outer_N =B outer_N *Max{X(N),Y(N)} (12)
B outer_N =~B inner_N (13)
wherein "-" is the negation operation, the operation rule is-0=1, -1=0, when B inner When being a set, B outer Also a collection. For example, assume that in a single gripping task there are 5 items or 5 poses to grip, for a combination of these 5 items, B inner =[0,1,0,0,0]Then B is obtained according to formula (13) outer =[1,0,1,1,1]This means that the 1 st, 3 rd, 4 th, 5 th of these five items are all outside the region of interest; x (N) is the grabbing characteristic value of the X axis of the N-th object to be grabbed, and Y (N) is the grabbing characteristic value of the Y axis of the N-th object to be grabbed. As shown in FIG. 1, an object to be grasped is located on either the inner ring or the outer ring, and cannot exist on both the inner ring and the outer ring, wherein the formula is B outer_N And B inner_N Mutually inverted, so that when one is 0, the other must be 1, and it is not possible to be 1 at the same time.
After the grabbing characteristic values are obtained, sorting can be carried out according to the grabbing characteristic values of each article, and the clamp is controlled to grab the articles in sequence. According to the actual situation, the sorting can be performed only according to the grabbing characteristic values, or the characteristic values obtained based on the orientation can be combined with other characteristic values, and the comprehensive sorting can be performed. If the sorting is performed in combination with the plurality of feature values, the plurality of feature values may be normalized, and a weight value may be set for each feature value, and the sorting is performed based on the normalized feature values and the corresponding weights, so as to control the jig to perform the capturing based on the sorting result.
In addition, for any of the above embodiments:
When the control clamp grabs a plurality of articles to be grabbed based on the sequence of the grabbing characteristic values, the control clamp can grab the articles to be grabbed in sequence, for example, grabbing characteristic values of three articles are obtained in one grabbing task, namely a first article 5, a second article 10 and a third article 15, then the control clamp can grab the third article in the first grabbing process, grab the second article in the second grabbing process and grab the first article in the third grabbing process; it is also possible to grasp only the article in which the grasping characteristic value is highest, and recalculate the grasping characteristic value at the next grasping, for example, the grasping characteristic values of 5 articles are obtained in one grasping task, namely, the first article 5, the second article 10, the third article 15, the fourth article 11, the fifth article 18, and since the grasping characteristic value of the fifth article is highest, the jig is controlled to grasp the fifth article at the first grasping; before the second grabbing, re-acquiring image data, calculating grabbing characteristic values of the remaining 4 articles, grabbing the characteristic value highest, and so on until grabbing is completed.
In addition, it should be noted that although each embodiment of the present invention has a specific combination of features, further combinations and cross combinations of these features between embodiments are also possible.
According to the embodiment, firstly, when the gripping scheme of the invention is used for controlling the clamp to grip, the orientation characteristics of the objects to be gripped are considered, compared with the existing scheme, the gripping difficulty of the objects with different orientations can be more accurately determined, the possibility of gripping failure is reduced, particularly when the gripping is performed in an industrial scene where a large number of objects are scattered and piled up, the existing scheme has poor robot operation effect in such a scene because the influence of the orientation characteristics of the objects is not considered, and the invention can greatly improve the robot gripping effect in such a scene; secondly, the invention also provides a grabbing control scheme which comprehensively considers the orientation characteristics of the articles and the position characteristics of the articles, when grabbing, firstly judging whether the articles are in an easy-to-grab area, and based on whether the articles are in the area or not, adopting different grabbing schemes, so that under certain scenes, for example, when a large number of articles are scattered in a container or a large number of articles are located in an area with firm barriers which can influence grabbing, compared with the scheme which only considers the orientation characteristics of the articles, the grabbing sequencing is more accurate, and the grabbing effect of a robot is further improved; thirdly, the invention provides a proposal for determining whether the object to be grabbed is positioned in a specific area based on the numerical value, and because the position inhibition value is preset, whether the object to be grabbed is positioned in the specific area can be determined only based on the position characteristic value of the object to be grabbed and the relation between the position characteristic value and the position inhibition value; fourth, the invention also provides a method for controlling the clamp and executing grabbing under the scene that a large number of articles are scattered and placed near the obstacle influencing grabbing, the method enables the articles to obtain grabbing characteristic values higher than the grabbing characteristic values when the articles face the obstacle on the basis of the orientation inhibition values in a numerical mode, and the articles which are easy to grab can be grabbed away firstly under the scene, so that grabbing effect is improved.
Fig. 10 shows a grip control device according to still another embodiment of the present invention, the device including:
an image data obtaining module 500, configured to obtain image data including at least one object to be grabbed, i.e. to implement step S100;
an orientation feature obtaining module 510, configured to process the image data to obtain an orientation feature related to an orientation of the object to be grabbed, that is, to implement step S110;
a gripping control module 520 for controlling the gripper to perform gripping of at least one object to be gripped based at least on the orientation characteristics of said object to be gripped, i.e. for implementing step S120.
Fig. 11 shows a grip control device according to still another embodiment of the present invention, the device including:
the feature obtaining module 600 is configured to obtain a position feature and an orientation feature of the object to be grabbed, that is, the feature obtaining module is configured to implement step S200;
the position relation determining module 610 is configured to determine, based on the position feature, a relation between a position of the object to be grabbed and a region of interest of a reference region where the object to be grabbed is located, that is, to implement step S210;
a grabbing feature value determining module 620, configured to determine a grabbing feature value based on a relationship between the position of the object to be grabbed and the region of interest and an orientation feature of the object to be grabbed; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped, i.e. to implement step S220.
Fig. 12 shows an article position determining apparatus according to still another embodiment of the present invention, the apparatus comprising:
the position suppression value determining module 700 is configured to calculate a position suppression value of the object based on the feature of the reference area, where the size of the position suppression value is related to the size of the region of interest of the reference area where the object is located, that is, is used to implement step S300;
a location feature value determining module 710, configured to obtain a location feature value of each of the plurality of articles, i.e. to implement step S310;
a position determining module 720, configured to determine, for each of a plurality of items, whether the item is located within a region of interest of a reference region based on a position feature value of the item, the position suppression value, and a size of the reference region; the position characteristic value of the object includes coordinate values of the object on an X axis and coordinate values of the object on a Y axis in a reference coordinate system, i.e. the position characteristic value is used for implementing step S320.
Fig. 13 shows a grip control device according to still another embodiment of the present invention, the device including:
the feature acquiring module 800 is configured to acquire a position feature and an orientation feature of the object to be grabbed, that is, to implement step S400;
an orientation suppression value determining module 810, configured to determine an orientation suppression value of the object to be grabbed based on the location feature and the orientation feature; wherein the orientation suppression value is such that the gripping feature value when the orientation of the article deviates from the obstacle is larger than the gripping feature value when the orientation of the article is directed towards the obstacle, i.e. for realizing step S410;
A gripping feature value determining module 820, configured to determine a gripping feature value of the object to be gripped based on the position feature of the object to be gripped, the orientation feature and the orientation suppression value; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped, i.e. to implement step S420.
It should be understood that in the above embodiment of the apparatus shown in fig. 10 to 13, only the main functions of the modules are described, and all the functions of each module correspond to the corresponding steps in the method embodiment, and the working principle of each module may refer to the description of the corresponding steps in the method embodiment. For example, the grabbing feature value determining module 820 is used to implement the method of step S420 in the above embodiment, and it is shown that the portion of the description that is used to describe and interpret step S420 is also used to describe and interpret the function of the grabbing feature value determining module 820. In addition, although the correspondence between functions of the functional modules and the method is defined in the above embodiments, those skilled in the art will understand that the functions of the functional modules are not limited to the correspondence, that is, a specific functional module may also implement other method steps or a part of the method steps. For example, the above embodiment describes the method for implementing step S420 by the capture feature value determining module 820, however, the capture feature value determining module 820 may be used to implement the method or a part of the method of step S400 or S410 as needed in actual situations.
The present application also provides a computer readable storage medium having stored thereon a computer program which when executed by a processor implements the method of any of the above embodiments. It should be noted that, the computer program stored in the computer readable storage medium according to the embodiment of the present application may be executed by the processor of the electronic device, and in addition, the computer readable storage medium may be a storage medium built in the electronic device or may be a storage medium capable of being plugged into the electronic device in a pluggable manner, so that the computer readable storage medium according to the embodiment of the present application has higher flexibility and reliability.
Fig. 14 is a schematic structural diagram of an electronic device according to an embodiment of the present invention, which may be a control system/electronic system configured in an automobile, a mobile terminal (e.g., a smart mobile phone, etc.), a personal computer (PC, e.g., a desktop computer or a notebook computer, etc.), a tablet computer, a server, etc., and the specific embodiment of the present invention is not limited to the specific implementation of the electronic device.
As shown in fig. 14, the electronic device may include: a processor 1202, a communication interface (Communications Interface) 1204, a memory 1206, and a communication bus 1208.
Wherein:
the processor 1202, the communication interface 1204, and the memory 1206 communicate with each other via a communication bus 1208.
A communication interface 1204 for communicating with network elements of other devices, such as clients or other servers, etc.
The processor 1202 is configured to execute the program 1210, and may specifically perform relevant steps in the method embodiments described above.
In particular, program 1210 may include program code including computer operating instructions.
The processor 1202 may be a central processing unit CPU, or a specific integrated circuit ASIC (Application Specific Integrated Circuit), or one or more integrated circuits configured to implement embodiments of the present invention. The one or more processors included in the electronic device may be the same type of processor, such as one or more CPUs; but may also be different types of processors such as one or more CPUs and one or more ASICs.
Program 1210 may be downloaded and installed from a network and/or from a removable medium via communications interface 1204. The program, when executed by the processor 1202, may cause the processor 1202 to perform the operations of the method embodiments described above.
In general terms, the invention comprises the following steps:
a grip control method comprising:
acquiring image data comprising at least one object to be grabbed;
processing the image data to acquire orientation features of the object to be grabbed, wherein the orientation features relate to the orientation;
and controlling the clamp to perform gripping of at least one article to be gripped based at least on the orientation characteristics of the article to be gripped.
Optionally, the at least one object to be grabbed includes a grabbed area of the at least one object to be grabbed.
Optionally, the controlling the clamp to perform gripping of the at least one object to be gripped includes determining a gripping order of the at least one object to be gripped, and controlling the clamp to perform gripping of the at least one object to be gripped in the gripping order.
Optionally, the image data is processed to obtain a position-related position feature of at least one object to be grabbed, and the gripper is controlled to perform grabbing of the at least one object to be grabbed based on at least the orientation feature and the position feature of the object to be grabbed.
Optionally, the orientation feature is obtained based on a rotation matrix of the object to be grabbed.
Optionally, the reference orientation of the rotation matrix is an orientation when the graspable region of the object to be grasped is perpendicular to the Z-axis.
Optionally, the rotation matrix is a euler angle-based rotation matrix.
A grip control device comprising:
the image data acquisition module is used for acquiring image data comprising at least one object to be grabbed;
the orientation feature acquisition module is used for processing the image data to acquire orientation features of the object to be grabbed, wherein the orientation features are related to the orientation;
and the grabbing control module is used for controlling the clamp at least based on the orientation characteristics of the articles to be grabbed so as to execute grabbing of at least one article to be grabbed.
Optionally, the at least one object to be grabbed includes a grabbed area of the at least one object to be grabbed.
Optionally, the grabbing control module is specifically configured to determine an grabbing order of at least one object to be grabbed, and control the fixture to execute grabbing of the at least one object to be grabbed according to the grabbing order.
Optionally, the method further comprises: the position feature acquisition module is used for processing the image data to acquire position features of at least one object to be grabbed, wherein the position features are related to positions; the gripping control module is used for controlling the clamp to perform gripping of at least one article to be gripped based on at least the orientation characteristic and the position characteristic of the article to be gripped.
Optionally, the orientation feature is obtained based on a rotation matrix of the object to be grabbed.
Optionally, the reference orientation of the rotation matrix is an orientation when the graspable region of the object to be grasped is perpendicular to the Z-axis.
Optionally, the rotation matrix is a euler angle-based rotation matrix.
A grip control method comprising:
acquiring position characteristics and orientation characteristics of an object to be grabbed;
based on the position characteristics, determining the relation between the position of the object to be grabbed and the interest area of the reference area where the object to be grabbed is located;
determining a grabbing characteristic value based on the relation between the position of the object to be grabbed and the region of interest and the orientation characteristic of the object to be grabbed; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped.
Optionally, the position feature comprises coordinates of the object to be grabbed in a reference coordinate system.
Optionally, the reference coordinate system comprises a camera coordinate system.
Optionally, the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped, including: sequencing the grabbing characteristic values of the plurality of articles to be grabbed, and controlling the clamp to execute grabbing according to the sequencing result.
Optionally, the method for determining the grabbing characteristic value of the object to be grabbed located in the region of interest is different from the method for determining the grabbing characteristic value of the object to be grabbed located outside the region of interest.
A grip control device comprising:
the characteristic acquisition module is used for acquiring the position characteristic and the orientation characteristic of the object to be grabbed;
the position relation determining module is used for determining the relation between the position of the object to be grabbed and the interest area of the reference area where the object to be grabbed is located based on the position characteristics;
the grabbing feature value determining module is used for determining grabbing feature values based on the relation between the position of the object to be grabbed and the region of interest and the orientation features of the object to be grabbed; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped.
Optionally, the position feature comprises coordinates of the object to be grabbed in a reference coordinate system.
Optionally, the reference coordinate system comprises a camera coordinate system.
Optionally, the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped, including: sequencing the grabbing characteristic values of the plurality of articles to be grabbed, and controlling the clamp to execute grabbing according to the sequencing result.
Optionally, the method for determining the grabbing characteristic value of the object to be grabbed located in the region of interest is different from the method for determining the grabbing characteristic value of the object to be grabbed located outside the region of interest.
A method of determining the location of an item, comprising:
calculating a position inhibition value of the object based on the characteristics of the reference area, wherein the size of the position inhibition value is related to the size of the interest area of the reference area where the object is located;
acquiring a position characteristic value of each article in a plurality of articles;
for each item of a plurality of items, determining whether the item is within a region of interest of a reference region based on a location characteristic value of the item, the location suppression value, and a size of the reference region; the position characteristic value of the article comprises coordinate values of an X axis and coordinate values of a Y axis of the article under a reference coordinate system.
Optionally, the range of the region of interest increases with an increase in the position suppression value.
Optionally, when the coordinate values of the X axis and the coordinate values of the Y axis of the object to be grabbed meet the suppression conditions, determining that the object to be grabbed is located in the region of interest, wherein the suppression conditions are related to the position suppression values.
Optionally, the judgment result of whether the coordinate values of the X axes of the plurality of objects to be grabbed meet the suppression condition is combined into a set, the judgment result of whether the coordinate values of the Y axes of the plurality of objects to be grabbed meet the suppression condition is combined into a set, and whether each of the plurality of objects to be grabbed is located in the region of interest is determined based on the two sets.
An article position determining apparatus comprising:
the position inhibition value determining module is used for calculating a position inhibition value of the article based on the characteristics of the reference area, and the size of the position inhibition value is related to the size of the interest area of the reference area where the article is located;
the position characteristic value determining module is used for acquiring the position characteristic value of each article in the plurality of articles;
a position determining module for determining, for each of a plurality of items, whether the item is located within a region of interest of a reference region based on a position feature value of the item, the position suppression value, and a size of the reference region; the position characteristic value of the article comprises coordinate values of an X axis and coordinate values of a Y axis of the article under a reference coordinate system.
Optionally, the range of the region of interest increases with an increase in the position suppression value.
Optionally, when the coordinate values of the X axis and the coordinate values of the Y axis of the object to be grabbed meet the suppression conditions, determining that the object to be grabbed is located in the region of interest, wherein the suppression conditions are related to the position suppression values.
Optionally, the judgment result of whether the coordinate values of the X axes of the plurality of objects to be grabbed meet the suppression condition is combined into a set, the judgment result of whether the coordinate values of the Y axes of the plurality of objects to be grabbed meet the suppression condition is combined into a set, and whether each of the plurality of objects to be grabbed is located in the region of interest is determined based on the two sets.
A grip control method comprising:
acquiring position characteristics and orientation characteristics of an object to be grabbed;
determining an orientation suppression value of the object to be grabbed based on the position feature and the orientation feature; wherein the orientation suppression value is such that the gripping feature value when the orientation of the article is away from the obstacle is greater than the gripping feature value when the orientation of the article is directed toward the obstacle;
determining a grabbing characteristic value of the object to be grabbed based on the position characteristic of the object to be grabbed, the orientation characteristic and the orientation inhibition value; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped.
Optionally, the orientation suppressing value is greater when the orientation of the item is away from the obstacle than when the orientation of the item is directed toward the obstacle.
Optionally, the determining the grabbing characteristic value of the object to be grabbed includes: and respectively calculating the grabbing characteristic value of the X axis and the grabbing characteristic value of the Y axis of the object to be grabbed, and taking the larger one as the grabbing characteristic value of the object.
Optionally, scaling is performed on the orientation suppression value and/or the grabbing feature value.
A grip control device comprising:
the characteristic acquisition module is used for acquiring the position characteristic and the orientation characteristic of the object to be grabbed;
the orientation inhibition value determining module is used for determining an orientation inhibition value of the object to be grabbed based on the position characteristics and the orientation characteristics; wherein the orientation suppression value is such that the gripping feature value when the orientation of the article is away from the obstacle is greater than the gripping feature value when the orientation of the article is directed toward the obstacle;
the grabbing feature value determining module is used for determining grabbing feature values of the to-be-grabbed objects based on the position features of the to-be-grabbed objects, the orientation features and the orientation inhibition values; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped.
Optionally, the orientation suppressing value is greater when the orientation of the item is away from the obstacle than when the orientation of the item is directed toward the obstacle.
Optionally, the determining the grabbing characteristic value of the object to be grabbed includes: and respectively calculating the grabbing characteristic value of the X axis and the grabbing characteristic value of the Y axis of the object to be grabbed, and taking the larger one as the grabbing characteristic value of the object.
Optionally, scaling is performed on the orientation suppression value and/or the grabbing feature value.
In the description of the present specification, reference to the terms "one embodiment," "some embodiments," "illustrative embodiments," "examples," "specific examples," or "some examples," etc., means that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the present application. In this specification, schematic representations of the above terms do not necessarily refer to the same embodiments or examples. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
Any process or method descriptions in flow charts or otherwise described herein may be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps of the process, and further implementations are included within the scope of the preferred embodiment of the present application in which functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those reasonably skilled in the art of the embodiments of the present application.
Logic and/or steps represented in the flowcharts or otherwise described herein, e.g., a ordered listing of executable instructions for implementing logical functions, can be embodied in any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, system that includes a processing module, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. For the purposes of this description, a "computer-readable medium" can be any means that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. More specific examples (a non-exhaustive list) of the computer-readable medium would include the following: an electrical connection (electronic device) having one or more wires, a portable computer diskette (magnetic device), a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber device, and a portable compact disc read-only memory (CDROM). In addition, the computer readable medium may even be paper or other suitable medium on which the program is printed, as the program may be electronically captured, via, for instance, optical scanning of the paper or other medium, then compiled, interpreted or otherwise processed in a suitable manner, if necessary, and then stored in a computer memory.
The processor may be a central processing unit (Central Processing Unit, CPU), but may also be other general purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), off-the-shelf programmable gate arrays (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, or the like. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
It is to be understood that portions of embodiments of the present application may be implemented in hardware, software, firmware, or a combination thereof. In the above-described embodiments, the various steps or methods may be implemented in software or firmware stored in a memory and executed by a suitable instruction execution system. For example, if implemented in hardware, as in another embodiment, may be implemented using any one or combination of the following techniques, as is well known in the art: discrete logic circuits having logic gates for implementing logic functions on data signals, application specific integrated circuits having suitable combinational logic gates, programmable Gate Arrays (PGAs), field Programmable Gate Arrays (FPGAs), and the like.
Those of ordinary skill in the art will appreciate that all or a portion of the steps carried out in the method of the above-described embodiments may be implemented by a program to instruct related hardware, where the program may be stored in a computer readable storage medium, and where the program, when executed, includes one or a combination of the steps of the method embodiments.
Furthermore, each functional unit in the embodiments of the present application may be integrated in one processing module, or each unit may exist alone physically, or two or more units may be integrated in one module. The integrated modules may be implemented in hardware or in software functional modules. The integrated modules may also be stored in a computer readable storage medium if implemented in the form of software functional modules and sold or used as a stand-alone product.
The above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, or the like.
Although the embodiments of the present application have been shown and described above, it will be understood that the above embodiments are illustrative and not to be construed as limiting the application, and that variations, modifications, alternatives and variations may be made to the embodiments described above by those of ordinary skill in the art within the scope of the application.
Claims (10)
1. A grip control method, characterized by comprising:
acquiring position characteristics and orientation characteristics of an object to be grabbed;
determining an orientation suppression value of the object to be grabbed based on the position feature and the orientation feature; wherein the orientation suppression value is such that the gripping feature value when the orientation of the article is away from the obstacle is greater than the gripping feature value when the orientation of the article is directed toward the obstacle;
determining a grabbing characteristic value of the object to be grabbed based on the position characteristic of the object to be grabbed, the orientation characteristic and the orientation inhibition value; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped.
2. The grip control method according to claim 1, characterized in that: the orientation suppression value when the orientation of the item is away from the obstacle is greater than the orientation suppression value when the orientation of the item is directed toward the obstacle.
3. The grip control method according to claim 1, wherein the determining the grip characteristic value of the article to be gripped includes: and respectively calculating the grabbing characteristic value of the X axis and the grabbing characteristic value of the Y axis of the object to be grabbed, and taking the larger one as the grabbing characteristic value of the object.
4. The grip control method according to any one of claims 1 to 3, characterized by further comprising: and scaling the orientation inhibition value and/or the grabbing characteristic value.
5. A grip control device, characterized by comprising:
the characteristic acquisition module is used for acquiring the position characteristic and the orientation characteristic of the object to be grabbed;
the orientation inhibition value determining module is used for determining an orientation inhibition value of the object to be grabbed based on the position characteristics and the orientation characteristics; wherein the orientation suppression value is such that the gripping feature value when the orientation of the article is away from the obstacle is greater than the gripping feature value when the orientation of the article is directed toward the obstacle;
the grabbing feature value determining module is used for determining grabbing feature values of the to-be-grabbed objects based on the position features of the to-be-grabbed objects, the orientation features and the orientation inhibition values; the gripping feature value can be used to control the gripper to perform gripping of the object to be gripped.
6. The grip control device according to claim 5, wherein: the orientation suppression value when the orientation of the item is away from the obstacle is greater than the orientation suppression value when the orientation of the item is directed toward the obstacle.
7. The grip control device of claim 5, wherein the determining the grip characteristic value of the object to be gripped includes: and respectively calculating the grabbing characteristic value of the X axis and the grabbing characteristic value of the Y axis of the object to be grabbed, and taking the larger one as the grabbing characteristic value of the object.
8. The grip control device of any one of claims 5-7, further comprising: and scaling the orientation inhibition value and/or the grabbing characteristic value.
9. An electronic device, comprising: memory, a processor and a computer program stored on the memory and executable on the processor, the processor implementing the grab control method of any of claims 1 to 4 when the computer program is executed.
10. A computer-readable storage medium, on which a computer program is stored, characterized in that the computer program, when executed by a processor, implements the grab control method of any of claims 1 to 4.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111429141.1A CN116175541B (en) | 2021-11-28 | 2021-11-28 | Grabbing control method, grabbing control device, electronic equipment and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111429141.1A CN116175541B (en) | 2021-11-28 | 2021-11-28 | Grabbing control method, grabbing control device, electronic equipment and storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116175541A true CN116175541A (en) | 2023-05-30 |
CN116175541B CN116175541B (en) | 2023-11-10 |
Family
ID=86449460
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111429141.1A Active CN116175541B (en) | 2021-11-28 | 2021-11-28 | Grabbing control method, grabbing control device, electronic equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116175541B (en) |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130338826A1 (en) * | 2010-12-30 | 2013-12-19 | Zenrobotics Oy | Method, computer program and apparatus for determining a gripping location |
CN110298879A (en) * | 2018-03-21 | 2019-10-01 | 北京猎户星空科技有限公司 | A kind of the determination method, apparatus and electronic equipment of object pose to be grabbed |
US20190308320A1 (en) * | 2018-04-05 | 2019-10-10 | Omron Corporation | Object recognition processing apparatus and method, and object picking apparatus and method |
CN111882610A (en) * | 2020-07-15 | 2020-11-03 | 中国科学院自动化研究所 | Method for grabbing target object by service robot based on elliptical cone artificial potential field |
JP2020197978A (en) * | 2019-06-04 | 2020-12-10 | グローリー株式会社 | Object detection device, object gripping system, object detection method, and object detection program |
US20210039257A1 (en) * | 2018-03-13 | 2021-02-11 | Omron Corporation | Workpiece picking device and workpiece picking method |
CN112947458A (en) * | 2021-02-26 | 2021-06-11 | 同济大学 | Robot accurate grabbing method based on multi-mode information and computer readable medium |
-
2021
- 2021-11-28 CN CN202111429141.1A patent/CN116175541B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130338826A1 (en) * | 2010-12-30 | 2013-12-19 | Zenrobotics Oy | Method, computer program and apparatus for determining a gripping location |
US20210039257A1 (en) * | 2018-03-13 | 2021-02-11 | Omron Corporation | Workpiece picking device and workpiece picking method |
CN110298879A (en) * | 2018-03-21 | 2019-10-01 | 北京猎户星空科技有限公司 | A kind of the determination method, apparatus and electronic equipment of object pose to be grabbed |
US20190308320A1 (en) * | 2018-04-05 | 2019-10-10 | Omron Corporation | Object recognition processing apparatus and method, and object picking apparatus and method |
JP2020197978A (en) * | 2019-06-04 | 2020-12-10 | グローリー株式会社 | Object detection device, object gripping system, object detection method, and object detection program |
CN111882610A (en) * | 2020-07-15 | 2020-11-03 | 中国科学院自动化研究所 | Method for grabbing target object by service robot based on elliptical cone artificial potential field |
CN112947458A (en) * | 2021-02-26 | 2021-06-11 | 同济大学 | Robot accurate grabbing method based on multi-mode information and computer readable medium |
Also Published As
Publication number | Publication date |
---|---|
CN116175541B (en) | 2023-11-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
DE102019130048B4 (en) | A robotic system with a sack loss management mechanism | |
US10124489B2 (en) | Locating, separating, and picking boxes with a sensor-guided robot | |
JP6697204B1 (en) | Robot system control method, non-transitory computer-readable recording medium, and robot system control device | |
WO2023092519A1 (en) | Grabbing control method and apparatus, and electronic device and storage medium | |
JP2010247959A (en) | Box-shaped work recognizing device and method | |
CN115330819B (en) | Soft package segmentation positioning method, industrial personal computer and robot grabbing system | |
CN112802107A (en) | Robot-based control method and device for clamp group | |
JP2024019690A (en) | System and method for robot system for handling object | |
JP2023154055A (en) | Robotic multi-surface gripper assemblies and methods for operating the same | |
CN110533717B (en) | Target grabbing method and device based on binocular vision | |
CN113800270B (en) | Robot control method and system for logistics unstacking | |
CN116175542B (en) | Method, device, electronic equipment and storage medium for determining clamp grabbing sequence | |
CN116175541B (en) | Grabbing control method, grabbing control device, electronic equipment and storage medium | |
CN116197888B (en) | Method and device for determining position of article, electronic equipment and storage medium | |
CN116175540B (en) | Grabbing control method, device, equipment and medium based on position and orientation | |
CN116214494A (en) | Grabbing control method, grabbing control device, electronic equipment and storage medium | |
CN116188559A (en) | Image data processing method, device, electronic equipment and storage medium | |
US20230071488A1 (en) | Robotic system with overlap processing mechanism and methods for operating the same | |
CN116197887B (en) | Image data processing method, device, electronic equipment and storage medium for generating grabbing auxiliary image | |
CN116197885B (en) | Image data filtering method, device, equipment and medium based on press-fit detection | |
CN116197886A (en) | Image data processing method, device, electronic equipment and storage medium | |
JP5332873B2 (en) | Bag-like workpiece recognition device and method | |
CN116205837A (en) | Image data processing method, device, electronic equipment and storage medium | |
CN116652974A (en) | Control unit and method for controlling robot gripping object | |
Macias | Visual and gesture guided robotic block stacking |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |