Nothing Special   »   [go: up one dir, main page]

CN106406684B - Projection process method, apparatus and projector - Google Patents

Projection process method, apparatus and projector Download PDF

Info

Publication number
CN106406684B
CN106406684B CN201610258644.XA CN201610258644A CN106406684B CN 106406684 B CN106406684 B CN 106406684B CN 201610258644 A CN201610258644 A CN 201610258644A CN 106406684 B CN106406684 B CN 106406684B
Authority
CN
China
Prior art keywords
image
view field
object run
user
run area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610258644.XA
Other languages
Chinese (zh)
Other versions
CN106406684A (en
Inventor
陈金山
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fafa Automobile China Co ltd
Original Assignee
Fafa Automobile (china) Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fafa Automobile (china) Co Ltd filed Critical Fafa Automobile (china) Co Ltd
Priority to CN201610258644.XA priority Critical patent/CN106406684B/en
Publication of CN106406684A publication Critical patent/CN106406684A/en
Application granted granted Critical
Publication of CN106406684B publication Critical patent/CN106406684B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0483Interaction with page-structured environments, e.g. book metaphor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/003Reconstruction from projections, e.g. tomography
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20068Projection on vertical or horizontal image axis

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The embodiment of the present invention provides a kind of projection process method, apparatus and projector, it is related to multimedia technology field, wherein, the projection process method comprises determining that the user object run area to be operated, wherein, object run area is located in view field, and view field is the region that projector is used for original projection data image;Operation according to user in object run area generates corresponding mark image, and mark image is merged with initial data image and generates target image;To view field's projection target image;Wherein, target image is located in the view field, and mark image is located in the object run area.The present invention generates corresponding mark image according to operation of the user in object run area, mark image is merged with initial data image and generates target image, and the target image of generation is projected to view field, with the display picture after view field shows user's operation, it realizes the flexible control to projection display picture, greatly strengthens the experience of user.

Description

Projection process method, apparatus and projector
Technical field
The embodiment of the present invention is related to multimedia technology field more particularly to a kind of projection process method, apparatus and projection Instrument.
Background technique
Projector, also known as projector are a kind of equipment that image or video are projected view field, can pass through difference Interface be connected with computer, VCD, DVD etc. and play corresponding vision signal.
Currently, projector has been widely used for family, office, school and public place of entertainment.Existing projector is all It is controlled, and then is realized to projection by the operation that the input equipments such as mouse, keyboard carry out display picture to computer connected to it The operation of the projection display picture of instrument controls.
With porjector technology continuous renewal development and daily extensive use, people to projection control convenience and Interactivity has more and more requirements.But during existing projection process, user can not leave computer directly to throwing The projection display picture of shadow instrument carries out operation control, influences user experience.
Summary of the invention
The present invention provides a kind of projection process method, apparatus and projector, can not be to solve user in the prior art The defect that computer directly carries out operation control to projection display picture is opened, the flexible control to projection display picture is realized, mentions The experience of high user.
The present invention provides a kind of projection process method, comprising:
Determine the user object run area to be operated, wherein the object run area is located in view field, the projection Region is the region of original projection data image;
Operation according to the user in the object run area generates corresponding mark image, by the mark image and The initial data image, which merges, generates target image;
The target image is projected to the view field;Wherein, the target image is located in the view field, institute Mark image is stated to be located in the object run area.
The present invention provides a kind of projection process device, comprising:
Determination unit, the object run area to be operated for determining user, wherein the object run area is located at projected area In domain, the view field is the region of original projection data image;
Generation unit generates corresponding mark image for the operation according to the user in the object run area, will The mark image merges with the initial data image generates target image;
Projecting cell, for projecting the target image that the generation unit generates to the view field;Wherein, the mesh Logo image is located in the view field, and the mark image is located in the object run area.
The present invention provides a kind of projector, including the infra-red ray detection device or image collector being arranged above view field It sets and above-mentioned projection process device;
The infra-red ray detection device is parallel to a plurality of infrared detection line of view field for transmitting vertically downward, In, the entire view field of a plurality of infrared detection beta radiation;Described image acquisition device is used for real-time acquired projections area The image data in domain.
Projection process method, apparatus provided by the invention and projector, pass through the behaviour according to user in object run area Make to generate corresponding mark image, mark image is merged into generation target image with initial data image, and by the target of generation Image is projected to view field, with the display picture after view field shows user's operation, is realized to projection display picture Flexibly control, greatly strengthens the experience of user.
Detailed description of the invention
It, below will be to embodiment or the prior art in order to illustrate more clearly of the present invention or technical solution in the prior art Attached drawing needed in description is briefly described, it should be apparent that, the accompanying drawings in the following description is of the invention one A little embodiments for those of ordinary skill in the art without creative efforts, can also be according to these Attached drawing obtains other attached drawings.
Fig. 1 is projection process embodiment of the method flow chart of the present invention;
Fig. 2 is the subdivision flow chart of step S11 in another embodiment of projection process method of the present invention;
Fig. 3 is the subdivision flow chart of step S11 in projection process method another embodiment of the present invention;
Fig. 4 is projection process Installation practice structural schematic diagram of the present invention;
Fig. 5 is a kind of entity structure schematic diagram of projection process device of the present invention.
Specific embodiment
In order to make the object, technical scheme and advantages of the embodiment of the invention clearer, below in conjunction with the embodiment of the present invention In attached drawing, technical scheme in the embodiment of the invention is clearly and completely described, it is clear that described embodiment is A part of the embodiment of the present invention, instead of all the embodiments.Based on the embodiments of the present invention, those of ordinary skill in the art Every other embodiment obtained without creative efforts, shall fall within the protection scope of the present invention.
Those skilled in the art of the present technique are appreciated that unless expressly stated, singular " one " used herein, " one It is a ", " described " and "the" may also comprise plural form.It is to be further understood that being arranged used in specification of the invention Diction " comprising " refer to that there are the feature, integer, step, operation, element and/or component, but it is not excluded that in the presence of or addition Other one or more features, integer, step, operation, element, component and/or their group.
Those skilled in the art of the present technique are appreciated that unless otherwise defined, all terms used herein (including technology art Language and scientific term), there is meaning identical with the general understanding of those of ordinary skill in fields of the present invention.Should also Understand, those terms such as defined in the general dictionary, it should be understood that have in the context of the prior art The consistent meaning of meaning, and unless otherwise will not be explained in an idealized or overly formal meaning by specific definitions.
During solving existing projection process, user can not leave computer directly to the Projection Display of projector Picture carries out the problem of operation control, and the embodiment of the present invention proposes a kind of projection process method.
Fig. 1 is projection process embodiment of the method flow chart of the present invention.
Referring to Fig.1, the projection process method that the embodiment of the present invention proposes, specifically includes the following steps:
S11, the user object run area to be operated is determined, wherein the object run area is located in view field, described View field is the region of original projection data image.
It should be noted that the object run area in the present embodiment is targeted when being user's progress gesture or motion action , and the region being located in view field.Initial data image is the data image of initial projections.
In practical applications, initial data image is projected to corresponding view field, when user is led in view field When a certain region carries out gesture or movement, determines that the region is object run area, further realize so as to subsequent to object run The operation in area identifies.
S12, the operation according to the user in the object run area generate corresponding mark image, and the mark is schemed Target image is generated as merging with the initial data image.
Wherein, mark image is the figure for the image data shown in object run area to be differently shown or identified Picture.Target image is to operate the obtained image with corresponding display effect in object run area according to user, with specific reference to The corresponding mark image of operation of the user in object run area merges to obtain with initial data image.
It should be noted that the mark image in the present embodiment is the operation generation according to the user in object run area 's.In a specific embodiment, one picture is performed to the image data in object run area in object run area such as user The operation of underscore then generates a corresponding straight line image as mark image according to the operation, and by the rectilinear of generation The image for carrying straight line image is generated as merging with initial data image, obtains target image.
S13, Xiang Suoshu view field project the target image;Wherein, the target image is located at the view field Interior, the mark image is located in the object run area.
In practical applications, after generating target image, by the target image generated to view field's projection, to realize Display picture after view field shows user's operation, specifically, target image is located at the projected area in projection process In domain, mark image is located in object run area, to guarantee that mark image accurately for showing with the region of operation, obtains Corresponding control effect, and then realize the flexible control to projection display picture.
Projection process method provided in an embodiment of the present invention, by determine the user object run area to be operated, according to Operation of the family in object run area generates corresponding mark image, and the mark image of generation is merged life with initial data image It is projected at target image, and by the target image of generation to view field, with the display after view field shows user's operation Picture realizes the flexible control to projection display picture, enhances the experience of user.
In an alternate embodiment of the present invention where, the determination user mesh to be operated in the step S11 of above-described embodiment Operating space is marked, as shown in Fig. 2, further including steps of
If S111, detecting that infrared detection line is blocked by obstructions in the view field, the shelter is obtained The currently coordinate value in the view field.
Wherein, the infrared detection line in the present embodiment is that the infra-red ray detection device being arranged in above the view field hangs down The directly a plurality of infrared ray for being parallel to the view field that emits downwards, the entire projected area of a plurality of infrared radiation Domain.
Specifically, infra-red ray detection device can be infrared word laser emitter, infrared word laser emitter is downward It shines, forms one layer of infrared smooth surface parallel with view field in front of view field.
In practical applications, what can be emitted by real-time detection infra-red ray detection device is parallel to the more of the view field Infrared ray, with the presence or absence of part is blocked by obstructions, realizes the detection in object run area, when there are quilts in the view field When a plurality of infrared ray that shelter blocks, then according to shelter, movement routine forms multiple coordinate values in the view field, To determine the user object run area to be operated.
S112, the coordinate value obtained according to step S111, determine the user object run area to be operated.
Wherein, multiple coordinate values are shelter mobile corresponding multiple seats of continuous path formed in the view field Scale value.Specifically, the path identified according to the multiple coordinate value corresponding region in view field is object run Area.The embodiment of the present invention accurately and reliably realizes the identification in object run area by using the mode of infrared induction, and Implementation is simple.
The determination user mesh to be operated in another alternative embodiment of the invention, in the step S11 of above-described embodiment Operating space is marked, as shown in figure 3, further including steps of
S111 ', the image data for obtaining the view field in real time.
In practical applications, can by image collecting device, such as camera, the image data in real-time acquired projections region, And the image data of image collecting device upload is received, realize the acquisition of the image data of the view field.Wherein, image The visual angle of acquisition device covers view field.
S112 ', image recognition is carried out to the image data of acquisition, determine that user to be operated according to recognition result described in.
In a specific embodiment, the image recognition to image data is realized by the following method.
Firstly, doing image procossing to the collected continuous picture sequence of camera, figure first is extracted with the method for background difference Movement pixel as in then filters off noise spot then with the impurity in binarization method removal part background, and by motor image Element is linked to be connected region one by one, obtains moving target, such as finger or other objects, then by the motion profile of the moving target The corresponding band of position is as object run area.
The embodiment of the present invention accurately and reliably realizes the identification in object run area by using the mode of image recognition, And implementation is simple, low cost, has good user experience.
In an alternate embodiment of the present invention where, in the step S12 of above-described embodiment according to the user described The operation in object run area generates corresponding mark image, specifically includes: identifying the user in the behaviour in the object run area Make, corresponding mark image is generated according to recognition result.In practical applications, it by multiple learning training, determines according to knowledge Other result generates the identification model of corresponding mark image, when identifying that user in the operation in object run area, utilizes identification Model generates corresponding mark image according to recognition result in real time.The embodiment of the present invention passes through the model pre-established, Neng Goushi When, accurately realize mark image generation.
In another alternative embodiment of the invention, in the step S12 of above-described embodiment according to the user described The operation in object run area generates corresponding mark image, specifically includes: identifying the user in the behaviour in the object run area Make, preset mapping relations are searched according to recognition result, instructed with obtaining mark corresponding with the recognition result, and according to obtaining The mark instruction taken generates corresponding mark image, wherein includes that the operation of user and mark instruct in the mapping relations Between corresponding relationship.
In the present embodiment, known by the track for operating and (such as touching) in view field to finger or other objects , i.e., gesture identification, gesture identification are realized by being matched with preset mapping relations, common gesture ratio if any Single finger is mobile etc. on the screen in screen scribing line, single finger, by carrying out these gestures and specific mark instruction It is corresponding, and corresponding mark image, such as completion underscore, circle etc. are generated according to the instruction of the mark of acquisition.
Preferably, the mark image includes the combination of any one or more in following image: straight line image, curve Image, circle image or rectangular image etc..
Intelligible, mark image can be used a variety of images and realize, wherein straight line image, curve image, circle image or In addition to this several frequently seen mark image by way of example only such as rectangular image further includes dotted line image, wave line image And the various irregular images etc. that can play mark action, the present invention is not especially limited this, those skilled in the art Member can be specifically arranged according to practical situations.
Further, mark image includes color.The embodiment of the present invention not only includes various informative mark image, and And identical or different mark image further includes color, further promotes the usage experience of user.
In addition, for above method embodiment, for simple description, therefore, it is stated as a series of action combinations, But those skilled in the art should understand that, the present invention is not limited by the sequence of acts described, secondly, art technology Personnel also should be aware of, and the embodiments described in the specification are all preferred embodiments, and related movement is not necessarily Necessary to the present invention.
Based on inventive concept same as method, the embodiment of the present invention also provides a kind of projection process device, and Fig. 4 is this hair Bright projection process Installation practice structural schematic diagram.
Referring to the projection process device that Fig. 4, the embodiment of the present invention are proposed, determination unit 401, generation unit are specifically included 402 and projecting cell 403, in which: determination unit 401 is for determining the user object run area to be operated, wherein target behaviour Make area to be located in view field, view field is the region of original projection data image;Generation unit 402 according to user for existing The operation in object run area generates corresponding mark image, and mark image is merged with the initial data image and generates target figure Picture;Projecting cell 403 is used for the target image generated to view field's projection generation unit;Wherein, target image is located at described In view field, mark image is located in object run area.
It should be noted that the object run area in the present embodiment is targeted when being user's progress gesture or motion action , and the region being located in view field.Initial data image is the data image of initial projections.
It should be noted that mark image in the present embodiment be for the image data shown in object run area into The image that row is distinctly displayed or identified.Target image be operated in object run area according to user obtain there is corresponding display The image of effect, the corresponding mark image of operation with specific reference to user in object run area merge with initial data image It arrives.
Projection process device provided in an embodiment of the present invention, determination unit 401 determine the user object run area to be operated, Generation unit 402 generates corresponding mark image according to operation of the user in object run area, by the mark image of generation with Initial data image, which merges, generates target image, and projecting cell 403 projects the target image of generation to view field, to throw Display picture after the domain views user's operation of shadow zone realizes the flexible control to projection display picture, enhances the experience of user.
In an alternate embodiment of the present invention where, the determination unit 401 in above-described embodiment further comprises obtaining mould Block and determining module, wherein obtain module and be used for when detecting that infrared detection line is blocked by obstructions in view field, obtain Take present coordinate values of the shelter in view field;Determining module is used for the coordinate value obtained according to module is obtained, determines and uses The family object run area to be operated.
Wherein, the infrared detection line in the present embodiment be arranged in infra-red ray detection device above view field vertically to The a plurality of infrared ray for being parallel to view field of lower transmitting, the entire view field of a plurality of infrared radiation.
Specifically, infra-red ray detection device can be realized using infrared word laser emitter, infrared word Laser emission Device shines downwards, and one layer of infrared smooth surface parallel with view field is formed in front of view field.Multiple coordinate values are shelter The mobile corresponding multiple coordinate values of continuous path formed in the view field.Specifically, according to the multiple coordinate value The path identified corresponding region in view field is object run area.The embodiment of the present invention, by using infrared sense The mode answered accurately and reliably realizes the determination in object run area, and implementation is simple.
In another alternative embodiment of the invention, determination unit 401 in above-described embodiment further comprises obtaining mould Block and determining module, wherein obtain module for obtaining the image data of view field in real time;Determining module is used for acquisition Image data carries out image recognition, determines the user object run area to be operated according to recognition result.
In practical applications, can by image collecting device, such as camera, the image data in real-time acquired projections region, And the image data of image collecting device upload is received, realize the acquisition of the image data of the view field.Wherein, image The visual angle of acquisition device covers view field.The embodiment of the present invention is accurately and reliably realized by using the mode of image recognition The identification in object run area, and implementation is simple, low cost, has good user experience.
In another alternative embodiment of the invention, generation unit 402 in above-described embodiment is specifically used for described in identification User generates corresponding mark image according to recognition result, or, searching according to recognition result in the operation in the object run area Preset mapping relations, to obtain mark instruction corresponding with the recognition result, and according to the instruction generation pair of the mark of acquisition The mark image answered, wherein include the corresponding relationship between the operation of user and mark instruction in the mapping relations.
In practical applications, generation unit 402 can realize the image recognition to image data by two schemes.The first, By multiple learning training, determine to generate the identification model of corresponding mark image according to recognition result, when identifying user In the operation in object run area, corresponding mark image is generated according to recognition result in real time using identification model.The present invention is real It applies example and passes through the model pre-established, can realize the generation of mark image precisely in real time.The second, by finger or its The track that its object operates (touch as) in view field identified, i.e. gesture identification, gesture identification be by with it is default Mapping relations matched to realize, common gesture ratio if any single finger in screen scribing line, single finger is in screen Upper movement etc., it is corresponding by carrying out these gestures with specific mark instruction, and instruct to generate according to the mark of acquisition and correspond to Mark image, for example complete underscore, circle etc..
Preferably, the mark image includes the combination of any one or more in following image: straight line image, curve Image, circle image or rectangular image etc..Intelligible, mark image can be used a variety of images and realize, wherein straight line image, song In addition to this several frequently seen mark image by way of example only such as line image, circle image or rectangular image further includes void Line image, wave line image and various irregular images that can play mark action etc., the present invention does not make this specifically It limits, those skilled in the art can be specifically arranged according to practical situations.
Further, mark image includes color.The embodiment of the present invention not only includes various informative mark image, and And identical or different mark image further includes color, further promotes the usage experience of user.
For device embodiment, since it is basically similar to the method embodiment, related so being described relatively simple Place illustrates referring to the part of embodiment of the method.
In addition, the embodiment of the present invention also provides a kind of projector, the projector includes the red of setting above view field Outer line detector or image collecting device and the projection process device as described in above-mentioned any embodiment;The infrared ray inspection Survey a plurality of infrared detection line that device is parallel to view field for transmitting vertically downward, wherein a plurality of infrared detection line Radiate the entire view field;Described image acquisition device is used for the image data in real-time acquired projections region.
Projector provided in an embodiment of the present invention is emitted vertically downward by the infra-red ray detection device of setting and is parallel to throwing The a plurality of infrared detection line in shadow zone domain;Or, the image data in the real-time acquired projections region of image collecting device, for determination unit The a plurality of infrared detection line for being parallel to view field of infra-red ray detection device transmitting is detected, or, to image collector The image data for setting the view field of acquisition identified, realizes the determination in the user object run area to be operated, and according to Operation of the family in object run area generates corresponding mark image, and the mark image of generation is merged life with initial data image At target image, projecting cell projects the target image of generation to view field, after view field shows user's operation Display picture, realize flexible control to projection display picture, enhance the experience of user.
Fig. 5 illustrates a kind of entity structure schematic diagram of projection process device, as shown in figure 5, the projection process device can To include: processor (processor) 510,520, memory communication interface (Communications Interface) (memory) 530 and bus 540, wherein processor 510, communication interface 520, memory 530 are completed mutually by bus 540 Between communication.Communication interface 540 can be used between projection process device and infra-red ray detection device or image collecting device Information transmission.Processor 510 can call the logical order in memory 530, to execute following method: determine that user is intended to operate Object run area, the object run area is located in view field, and the view field is the area of original projection data image Domain;Operation according to the user in the object run area generates corresponding mark image, by the mark image with it is described Initial data image, which merges, generates target image;The target image is projected to the view field;Wherein, the target image In the view field, the mark image is located in the object run area.
In addition, the logical order in above-mentioned memory 530 can be realized by way of SFU software functional unit and conduct Independent product when selling or using, can store in a computer readable storage medium.Based on this understanding, originally Substantially the part of the part that contributes to existing technology or the technical solution can be in other words for the technical solution of invention The form of software product embodies, which is stored in a storage medium, including some instructions to So that a computer equipment (can be personal computer, server or the network equipment etc.) executes each implementation of the present invention The all or part of the steps of example the method.And storage medium above-mentioned include: USB flash disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic or disk etc. it is various It can store the medium of program code.
To sum up, projection process method, apparatus provided in an embodiment of the present invention and projector, by being grasped according to user in target The operation made in area generates corresponding mark image, and mark image is merged with initial data image and generates target image, and will The target image of generation is projected to view field, with the display picture after view field shows user's operation, is realized to projection The flexible control for showing picture, greatly strengthens the experience of user.
The apparatus embodiments described above are merely exemplary, wherein described, unit can as illustrated by the separation member It is physically separated with being or may not be, component shown as a unit may or may not be physics list Member, it can it is in one place, or may be distributed over multiple network units.It can be selected according to the actual needs In some or all of the modules achieve the purpose of the solution of this embodiment.Those of ordinary skill in the art are not paying creativeness Labour in the case where, it can understand and implement.
Through the above description of the embodiments, those skilled in the art can be understood that each embodiment can It realizes by means of software and necessary general hardware platform, naturally it is also possible to pass through hardware.Based on this understanding, on Stating technical solution, substantially the part that contributes to existing technology can be embodied in the form of software products in other words, should Computer software product may be stored in a computer readable storage medium, such as ROM/RAM, magnetic disk, CD, including several fingers It enables and using so that a computer equipment (can be personal computer, server or the network equipment etc.) executes each implementation Method described in certain parts of example or embodiment.
Finally, it should be noted that the above embodiments are merely illustrative of the technical solutions of the present invention, rather than its limitations;Although Present invention has been described in detail with reference to the aforementioned embodiments, those skilled in the art should understand that: it still may be used To modify the technical solutions described in the foregoing embodiments or equivalent replacement of some of the technical features; And these are modified or replaceed, technical solution of various embodiments of the present invention that it does not separate the essence of the corresponding technical solution spirit and Range.

Claims (11)

1. a kind of projection process method characterized by comprising
Determine the user object run area to be operated, wherein the object run area is located in view field, the view field For the region of original projection data image;
Operation according to the user in the object run area generates corresponding mark image, by the mark image with it is described Initial data image, which merges, generates target image;Wherein, mark image is for the image data shown in object run area The image for being differently shown or identifying;
The target image is projected to the view field;Wherein, the target image is located in the view field, the mark Know image to be located in the object run area;
Wherein, the determining user object run area to be operated, comprising:
The image data of the view field is obtained in real time;
Image recognition is carried out to the image data of acquisition, the user object run area to be operated is determined according to recognition result; Wherein, the image recognition to image data is realized by the following method: firstly, doing at image to collected continuous picture sequence Reason first extracts the movement pixel in image with the method for background difference, then with miscellaneous in binarization method removal part background Matter then filters off noise spot, and movement pixel is linked to be connected region one by one, moving target is obtained, then by the movement mesh The corresponding band of position of target motion profile is as object run area;
Wherein, the operation according to the user in the object run area generates corresponding mark image, comprising:
Identify that the user in the operation in the object run area, generates corresponding mark image according to recognition result;Wherein, lead to Multiple learning training is crossed, determines to generate the identification model of corresponding mark image according to recognition result, when identifying that user exists When the operation in object run area, corresponding mark image is generated according to recognition result in real time using identification model;Or,
Identify that the user in the operation in the object run area, searches preset mapping relations according to recognition result, to obtain Mark instruction corresponding with the recognition result, and corresponding mark image is generated according to the instruction of the mark of acquisition, wherein it is described It include the corresponding relationship between the operation of user and mark instruction in mapping relations.
2. the method according to claim 1, wherein the determining user object run area to be operated, comprising:
If detecting, infrared detection line is blocked by obstructions in the view field, obtains the shelter currently described Coordinate value in view field;
According to the coordinate value, the user object run area to be operated is determined.
3. according to the method described in claim 2, it is characterized in that, the infrared detection line is to be arranged in the view field The a plurality of infrared ray for being parallel to the view field that the infra-red ray detection device of side emits vertically downward, a plurality of infrared ray Radiate the entire view field.
4. according to the method in any one of claims 1 to 3, which is characterized in that the mark image includes following image In the combination of any one or more: straight line image, curve image, circle image or rectangular image.
5. according to the method described in claim 4, it is characterized in that, the mark image includes color.
6. a kind of projection process device characterized by comprising
Determination unit, the object run area to be operated for determining user, wherein the object run area is located at view field Interior, the view field is the region of original projection data image;
Generation unit generates corresponding mark image for the operation according to the user in the object run area, will be described Mark image merges with the initial data image generates target image;Wherein, mark image is for in object run area The image that the image data of displaying is differently shown or identifies;
Projecting cell, for projecting the target image that the generation unit generates to the view field;Wherein, the target figure In in the view field, the mark image is located in the object run area image position;
Wherein, the determination unit further comprises:
Module is obtained, for obtaining the image data of the view field in real time;
Determining module determines the user institute to be operated according to recognition result for carrying out image recognition to the image data of acquisition State object run area;Wherein, the image recognition to image data is realized by the following method: firstly, to collected sequential chart Piece sequence does image procossing, first extracts the movement pixel in image with the method for background difference, is then removed with binarization method Impurity in the background of part then filters off noise spot, and movement pixel is linked to be connected region one by one, obtains moving target, Then using the corresponding band of position of the motion profile of the moving target as object run area;
Wherein, the generation unit, specifically for identifying the user in the operation in the object run area, according to recognition result Generate corresponding mark image;Wherein, it by multiple learning training, determines to generate corresponding mark image according to recognition result Identification model generated in real time according to recognition result when identifying user in the operation in object run area using identification model Corresponding mark image;Or, preset mapping relations are searched according to recognition result, to obtain mark corresponding with the recognition result Know instruction, and corresponding mark image is generated according to the instruction of the mark of acquisition, wherein includes user's in the mapping relations Corresponding relationship between operation and mark instruction.
7. device according to claim 6, which is characterized in that the determination unit further comprises:
Module is obtained, for obtaining the screening when detecting that infrared detection line is blocked by obstructions in the view field The current coordinate value in the view field of block material;
Determining module, the coordinate value for being obtained according to the acquisition module, determines the user object run area to be operated.
8. device according to claim 7, which is characterized in that the infrared detection line is to be arranged in the view field The a plurality of infrared ray for being parallel to the view field that the infra-red ray detection device of side emits vertically downward, a plurality of infrared ray Radiate the entire view field.
9. the device according to any one of claim 6 to 8, which is characterized in that the mark image includes following image In the combination of any one or more: straight line image, curve image, circle image or rectangular image.
10. device according to claim 9, which is characterized in that the mark image includes color.
11. a kind of projector, which is characterized in that the projector include the infra-red ray detection device being arranged above view field or Image collecting device and such as described in any item projection process devices of claim 6 to 10;
The infra-red ray detection device is parallel to a plurality of infrared detection line of view field for transmitting vertically downward, wherein institute State a plurality of infrared detection beta radiation entirely view field;Described image acquisition device is used for the figure in real-time acquired projections region As data.
CN201610258644.XA 2016-04-22 2016-04-22 Projection process method, apparatus and projector Active CN106406684B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610258644.XA CN106406684B (en) 2016-04-22 2016-04-22 Projection process method, apparatus and projector

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610258644.XA CN106406684B (en) 2016-04-22 2016-04-22 Projection process method, apparatus and projector

Publications (2)

Publication Number Publication Date
CN106406684A CN106406684A (en) 2017-02-15
CN106406684B true CN106406684B (en) 2019-07-02

Family

ID=58006352

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610258644.XA Active CN106406684B (en) 2016-04-22 2016-04-22 Projection process method, apparatus and projector

Country Status (1)

Country Link
CN (1) CN106406684B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107360407A (en) * 2017-08-09 2017-11-17 上海青橙实业有限公司 Picture synthesizes projection method and main control device, auxiliary device
CN108200419B (en) * 2018-03-30 2020-09-25 联想(北京)有限公司 Projection method and electronic equipment
EP3876083A4 (en) * 2018-11-01 2021-12-29 Sony Group Corporation Information processing device, information processing method, and program
CN114694545B (en) * 2020-12-30 2023-11-24 成都极米科技股份有限公司 Image display method, device, projector and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102799317A (en) * 2012-07-11 2012-11-28 联动天下科技(大连)有限公司 Smart interactive projection system
CN103365413A (en) * 2012-03-22 2013-10-23 株式会社理光 Information processing device, computer-readable storage medium and projecting system
CN103713738A (en) * 2013-12-17 2014-04-09 武汉拓宝电子系统有限公司 Man-machine interaction method based on visual tracking and gesture recognition
CN104052976A (en) * 2014-06-12 2014-09-17 海信集团有限公司 Projection method and device
CN104750243A (en) * 2013-12-27 2015-07-01 日立麦克赛尔株式会社 Image projection device
CN104853128A (en) * 2014-02-19 2015-08-19 株式会社理光 Projection system and projection method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103365413A (en) * 2012-03-22 2013-10-23 株式会社理光 Information processing device, computer-readable storage medium and projecting system
CN102799317A (en) * 2012-07-11 2012-11-28 联动天下科技(大连)有限公司 Smart interactive projection system
CN103713738A (en) * 2013-12-17 2014-04-09 武汉拓宝电子系统有限公司 Man-machine interaction method based on visual tracking and gesture recognition
CN104750243A (en) * 2013-12-27 2015-07-01 日立麦克赛尔株式会社 Image projection device
CN104853128A (en) * 2014-02-19 2015-08-19 株式会社理光 Projection system and projection method
CN104052976A (en) * 2014-06-12 2014-09-17 海信集团有限公司 Projection method and device

Also Published As

Publication number Publication date
CN106406684A (en) 2017-02-15

Similar Documents

Publication Publication Date Title
RU2678478C2 (en) Lights control in environment of eye motion tracking
Orchard et al. Converting static image datasets to spiking neuromorphic datasets using saccades
US10001845B2 (en) 3D silhouette sensing system
CN106796656B (en) Depth from time-of-flight camera
US10424341B2 (en) Dynamic video summarization
KR101292467B1 (en) Virtual controller for visual displays
CN106406684B (en) Projection process method, apparatus and projector
US20180113598A1 (en) Augmented interface authoring
KR20160146716A (en) Air and surface multitouch detection in mobile platform
JP7248177B2 (en) Information processing system, information processing method, and program
CN105229582A (en) Based on the gestures detection of Proximity Sensor and imageing sensor
US8913037B1 (en) Gesture recognition from depth and distortion analysis
US20150169082A1 (en) Method and Device for Filter-Processing Imaging Information of Emission Light Source
JP2008097172A (en) Display and display method
Dai et al. Touchscreen everywhere: On transferring a normal planar surface to a touch-sensitive display
KR20130055119A (en) Apparatus for touching a projection of 3d images on an infrared screen using single-infrared camera
JP7234502B2 (en) Method, Apparatus and Program
JP6810048B2 (en) How to simulate and control virtual balls on mobile devices
CN106845494A (en) The method and device of profile angle point in a kind of detection image
CN106796487A (en) Interacted with the user interface element for representing file
CN103761011B (en) A kind of method of virtual touch screen, system and the equipment of calculating
JP5279482B2 (en) Image processing apparatus, method, and program
CN102129332A (en) Detection method and device of touch points for image recognition
KR100977558B1 (en) Space touch apparatus using infrared rays
CN104407746A (en) Infrared photoelectric technology based multi-point touch system

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20181008

Address after: 511458 9, Nansha District Beach Road, Guangzhou, Guangdong, 9

Applicant after: Evergrande Faraday Future Smart Car (Guangdong) Co.,Ltd.

Address before: 100025 8 floor 909, 105 building 3, Yao Yuan Road, Chaoyang District, Beijing.

Applicant before: Lexus Automobile (Beijing) Co.,Ltd.

Effective date of registration: 20181008

Address after: 100025 8 floor 909, 105 building 3, Yao Yuan Road, Chaoyang District, Beijing.

Applicant after: Lexus Automobile (Beijing) Co.,Ltd.

Address before: 100025 10 floor 1102, 105 building 3, Yao Yuan Road, Chaoyang District, Beijing.

Applicant before: LE HOLDINGS (BEIJING) Co.,Ltd.

Applicant before: LEAUTO INTELLIGENT TECHNOLOGY (BEIJING) Co.,Ltd.

TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20190313

Address after: 100015 Building No. 7, 74, Jiuxianqiao North Road, Chaoyang District, Beijing, 001

Applicant after: FAFA Automobile (China) Co.,Ltd.

Address before: 511458 9, Nansha District Beach Road, Guangzhou, Guangdong, 9

Applicant before: Evergrande Faraday Future Smart Car (Guangdong) Co.,Ltd.

TA01 Transfer of patent application right
GR01 Patent grant
GR01 Patent grant