WO2018077648A1 - A method of providing information about an object - Google Patents
A method of providing information about an object Download PDFInfo
- Publication number
- WO2018077648A1 WO2018077648A1 PCT/EP2017/076327 EP2017076327W WO2018077648A1 WO 2018077648 A1 WO2018077648 A1 WO 2018077648A1 EP 2017076327 W EP2017076327 W EP 2017076327W WO 2018077648 A1 WO2018077648 A1 WO 2018077648A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- light
- light source
- information
- object information
- image
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04B—TRANSMISSION
- H04B10/00—Transmission systems employing electromagnetic waves other than radio-waves, e.g. infrared, visible or ultraviolet light, or employing corpuscular radiation, e.g. quantum communication
- H04B10/11—Arrangements specific to free-space transmission, i.e. transmission through air or vacuum
- H04B10/114—Indoor or close-range type systems
- H04B10/116—Visible light communication
Definitions
- the invention relates to a method of providing information about an object.
- the invention further relates to a system for providing information about an object.
- Smart devices such as smartphones, are becoming our shopping assistants in stores.
- shopping assistants include indoor positioning guidance systems, position or user based advertisements, augmented product information, etc.
- Retailers are more and more looking for solutions for omni-channel retailing wherein physical and digital retail channels are seamlessly integrated. For instance, retail environments may become interactive showrooms or experience centers where the products can be seen, experienced, used and ordered.
- Patent application WO 2014/170290 Al discloses a coded light device comprising a light source configured to emit coded light, an identifier detection unit configured to detect a product identifier, and a processing unit having a code generator configured to generate a product-related code on basis of the product identifier, and a light source controller configured to control the light source to emit the product-related code by coding its light output.
- a customer may use a mobile device, such as a smart phone, to receive the light by means of a built in camera, detect the product-related code and receive product related data, on basis of the product-related code, from the product information source. Once the product-related code has been received at the mobile device, product-related information can be presented to the customer on the mobile device.
- U.S. patent application 2014/0100997 Al discloses a method of enabling a user to perform a number of augmented re-ality shopping operations using a local client shopping application executable on a mobile device, for example viewing an augmented reality depiction of a shopping display using the mobile device and executing various "e- commerce" functions accessible via the augmented reality depiction.
- the object is achieved by a method of providing information about an object, the method comprising:
- controlling a first light source according to the first light setting controlling a second light source according to the second light setting, detecting light comprising an embedded code, which embedded code is representative of the object information,
- the first and the second light source provide information about the object, which information is perceivable for a user. If the user wishes to receive more information about the object, he or she may point a camera or a light detector of the mobile device towards a light source that emits light comprising the embedded code representative of the object information, and thereby receive additional object information. The additional object information is then rendered on the display, enabling a user to see the additional information. This is beneficial, because it allows a user to choose if he or she wishes to receive more object information.
- the light comprises the code representative of the object information.
- the (controller of the) light sources may be commissioned in advance such that the light sources emit light comprising the embedded code that is associated with a respective object. In other words, the association between the object and the light source may be predefined.
- a detector may detect which object is located at the light source (or which object is illuminated by the light source) in order to determine which code to emit.
- the object may for example comprise a tag (e.g. an RFID tag, a QR code, an NFC tag, etc.) which may be detected by a tag reader for identifying the object located at or nearby the light source.
- the step of controlling the first light source according to the first light setting may further comprise controlling the first light source such that it emits the embedded code.
- the step of controlling the first light source according to the first light setting may further comprise controlling the first light source such that it emits a first embedded code, which first embedded code is representative of first object information about the object
- the step of controlling the second light source according to the second light setting may further comprise controlling the second light source such that it emits a second embedded code, which second embedded code is representative of second object information about the object
- the step of detecting light comprising the embedded code may comprise detecting light comprising the first embedded code and detecting light comprising the second embedded code
- the step of retrieving the code from the light may comprise retrieving the first embedded code and the second embedded code from the respective light
- the step of retrieving the object information may comprise retrieving the first object information and the second object information from the respective codes
- the step of rendering the object information on the display may comprise rendering the first object information and the second object information on the display.
- each individual light source provides information about the object in two different ways: a first way wherein the light output of the respective light source is perceivable for a user, and a second way wherein the code comprised in the light output of the respective light source is perceivable for a light detector.
- the method may further comprise the steps of: capturing an image of the first light source and the second light source, rendering the image on a display, and rendering a the object information as an overlay on the image. Additionally, the method may comprise the steps of: determining a first light source position of the first light source and a second light source position of the second light source in the image, and determining a position for the object information on the display, wherein the position is related to at least one of the first light source position and the second light source position in the image, and rendering the the object information at the position.
- the method may further comprise the steps of: accessing a user profile of the user, selecting the object information from a set of stored object information based on the user profile and determining the first light setting and the second light setting based on the selected object information.
- This is advantageous because the object information that is provided to a user is based on the user profile. If, for example, a user is interested in price information, price-related object information may be selected, while if a user is interested in stock information, stock-related object information may be selected.
- the embedded code may comprise an identifier of the object.
- the identifier may be compared with a plurality of object identifiers stored in a memory in order to retrieve the object information related to the object associated with the identifier.
- the embedded code may comprise a link to the object information (e.g. a hyperlink, a URL, etc.) which link may provide information of a storage location of the object information in a memory.
- the embedded code may comprise the object information about the object.
- the object information is indicative of at least one of price information, material information, size information, availability information, stock information, review information, object score information and popularity information.
- the object is achieved by a lighting system for providing information about an object, the lighting system comprising:
- a controller for receiving object information about the object, for determining a first light setting and a second light setting based on the object information, wherein the difference between the first light setting and the second light setting is perceivable for a user, for controlling a first light source according to the first light setting, and for controlling a second light source according to the second light setting,
- a mobile device comprising:
- a light detector for detecting light comprising an embedded code, which embedded code is representative of the object information
- the first light source and the second light source may be comprised in a linear lighting device (such as an LED strip).
- the linear lighting device may comprise a plurality of individually addressable light sources arranged for providing the object information.
- the linear lighting device may be comprised in a supporting member (e.g. a product shelf) arranged for supporting the object.
- the controller may be arranged for controlling three or more light sources, and the controller may be arranged for determining which light sources to control according to the first light setting and which light sources to control according to the second light setting based on the object information.
- the controller may for example be arranged for controlling a linear lighting device with individually addressable light sources.
- the controller may control a first set of light sources according to the first light setting and a second set of light sources according to the second light setting.
- the controller controls the light sources according to the object information.
- the object information may, for example, be stock information, and the controller may determine how many light sources to control according to the first light setting (and how many according to the second light setting) based on how many objects there are still in stock. As a result, a user can simply see how many objects are still in stock.
- the light source may be arranged for illuminating the object.
- the light emitted by the light source and reflected from the object may be detected in order to retrieve the code, and therewith the object information, from the light.
- the object and the light sources are in each other's vicinity, or if the light source illuminates the object, because then the relation between the light sources and the object is clear.
- FIG. 1 shows schematically an embodiment of a system according to the invention for providing information about an object
- Figs. 2a and 2b show schematically embodiments of a mobile device according to the invention for providing information about an object;
- Fig. 3a and 3b show schematically embodiments of a product display according to the invention for providing light comprising an embedded code
- Fig. 4 shows schematically an embodiment of a system according to the invention for providing information about an object
- Fig. 5a shows schematically embodiments of a product display according to the invention for providing information about an object
- Fig. 6 shows schematically an embodiment of a part of a wearable mobile device according to the invention for providing information about an object
- Fig. 7 shows schematically a method according to the invention of providing information about an object
- Fig. 8 shows schematically a method according to the invention of providing information about an object
- Fig. 9 shows schematically a method according to the invention of storing object identifiers.
- Fig. 1 shows schematically an embodiment of a system 100 according to the invention for providing information about an object 120.
- the system 100 comprises a mobile device 102 for providing information 132, 134 about an object 120.
- the mobile device 102 may be any type of mobile device 120. Examples of a mobile device include but are not limited to portable smart devices, such as smartphones and tablet pes, and wearable smart devices such as smart watches and smart goggles.
- the mobile device 102 may be arranged for capturing an image 104 of the object 102 and for detecting light emitted by a light source 110, which light comprises an embedded code representative of object information about the object 120.
- the mobile device 102 may comprise a display and a processor (not shown) arranged for retrieving the code from the light.
- Figs. 1-6 illustrate examples of objects 120.
- the object is a shoe, but it should be noted that the object can be any type of object. Examples include but are not limited to consumer goods such as clothing, food, vehicles, household appliances, robots and electronic devices and animate objects such as animals.
- the processor (e.g. a microchip, circuitry, a microcontroller, etc.) of the mobile device 102 may be arranged for retrieving the code from the light.
- the mobile device 102 may comprise a light detector (such as a photodiode) for detecting the code in the light emitted by the light source 110 and a camera (not shown) for capturing the image 104 of the object.
- the mobile device 102 comprises a camera for capturing the image 104 of the object and for detecting the light emitted by the light source.
- the camera may, for example, be a rolling shutter camera or a global shutter camera. Techniques for retrieving the code from the light (either by photodiode or by camera) are well known in the art and will not be discussed further.
- the processor of the mobile device 102 may be arranged for retrieving the object information.
- the object information may be retrieved based on the code.
- the code may comprise, for example, an identifier of the object 120.
- the processor may be arranged for accessing a (remote) memory, which memory may be arranged for storing associations between object identifiers and object information.
- the processor may compare the identifier of the object 120 with the stored identifiers, and, if a stored identifier matches the object identifier, determine the object information for the identified object 120.
- the code may comprise a link to the object information.
- the link may, for example, be a URL or a hyperlink which comprises an address that points to a location where the object information is stored. This enables the processor to find the object information of the object 120 based on the link.
- the code may comprise the object information. In the latter, the processor may directly retrieve the object information from the code.
- the object information may, for example, be price information, material information, size information, stock information, color information, product age information, availability information, user review information, object score information and/or popularity information.
- the object information may comprise multiple types of object information (e.g. size and price information).
- the processor may be further arranged for identifying an object feature 122, 124, 126 of the object 120' in the image 104.
- the processor may comprise an image analysis module for identifying the object feature.
- the object feature 122, 124, 126 may be related to an outline of at least a part of the object 120, a color of at least a part of the object 120, an area of least a part of the object 120, a pattern of at least a part of an object 120 and/or a marker (e.g. a barcode or a QR-code) attached to the object 120.
- a marker e.g. a barcode or a QR-code
- the processor may, for example, identify the object feature 122, 124, 126 of the object 120' in the image 104 by accessing a memory arranged for storing object feature information about at least one object, comparing at least a part of the image 104 with stored object feature information, so as to determine if a similarity criterion has been met between a stored object feature and the at least a part of the image 104, and identifying, if the similarity criterion has been met, the object feature 122, 124, 126 in the image 104.
- any image analysis technique for identifying objects and object features from images 104 known in the art may be used for identifying the object feature. Fig.
- first object feature 122 i.e. the nose of the shoe 120 the striped pattern
- second object feature 124 i.e. the sole of the shoe 120
- third object feature 126 i.e. the logo on the shoe 120.
- the processor may determine a position of the object feature (the object feature position) in the image 104.
- the image analysis module may be arranged for determining this position.
- the position of the object feature may be an area in the image 104 with specific coordinates in the image.
- the processor may be arranged for determining a plurality of positions of different object features.
- Object features may, for example, be part of other object features (e.g. a first object feature may be the outline of the shoe 120, and a second object feature may be the logo 126 on the shoe 120, which is located in the area created by the outline of the shoe). It should be noted that any image analysis technique known in the art may be used for determining a position of an object feature in an image 104.
- the processor may be further arranged for generating a virtual representation 132, 134 of the object information.
- the virtual representationl32, 134 may, for example, be text, a color, a logo, a virtual price tag, a video, an animation, a feedback button, a virtual character (e.g. providing spoken instructions related to the object feature), or an alternative object feature (e.g. showing alternative colors for the object or object part) etc. representative of the object information.
- the processor may be further arranged for determining a position of the virtual representation 132, 134 in the image 104, wherein the position is related to the (previously determined) object feature position in the image 104.
- the processor may be arranged for determining the position of the virtual representation 132, 134 such that it is different from the object feature position.
- the position of the virtual representation may for example be an area in the image 104 with coordinates different from the coordinates of the area of the object feature position.
- the processor may be arranged for determining the position of the virtual representation 132, 134 such that it is in vicinity of the object feature position.
- the position of the virtual representation may for example be an area in the image 104 with coordinates nearby the coordinates of the area of the object feature position.
- Fig. 1 illustrates such an example.
- the processor (not shown) of the mobile device 102 may determine the position of the sole 124' of the shoe 120' in the image 104, and determine a position for the virtual representation 134 of the object information about the sole 124 of the shoe 120.
- the virtual representation is textual information about the object information about the object feature 124, for example information about the materials used for the sole 124 of the shoe 120.
- the processor may be arranged for determining a plurality of positions of different object features 122, 124, 126, and the processor may position the virtual representation of an object feature in relation to the position of that object feature and in relation to the other object features. For example, the processor may identify a first object feature (e.g. the outline of the shoe 120') and a second object feature (e.g. the striped nose 122' of the shoe 120') in the image 104 and determine to position the virtual representation 132 of the object information about the object feature 122 outside the outline of the shoe 120'.
- the virtual representation 132 is textual information about the object information about the object feature 122, for example information about the pattern on the nose 122 of the shoe 120.
- the processor may be arranged for determining the position of the virtual representation 132, 134 such that it overlaps the object feature position.
- the position of the virtual representation 132, 134 may for example be an area in the image 104 with coordinates similar to the coordinates of the area of the object feature position.
- Fig. 2a illustrates such an example.
- the processor (not shown) of the mobile device 102 may determine the position of the nose 122' of the shoe 120' in the image 104, and determine a position for the virtual representation 122" of the object information about the nose 122 of the shoe 120.
- the position of the virtual representation 122" is in this case the same position as the position of the object feature 122'.
- the virtual representation 122" is in this example an overlay on top of the 'actual' shoe 120.
- a user may provide a user input to change the virtual representation 122" to a different representation (e.g. by selecting a different color for the nose of the shoe) to see how the shoe would look like in that color.
- the user may select a different color for the logo 126' on the shoe 120'.
- the processor may be further arranged for rendering the image on a display, and for rendering the virtual representation at the position as an overlay on the image.
- the mobile device may be smart glasses 602.
- the smart glasses may comprise a camera 608 for detecting the object 120 and a display 604.
- the display may be at least one of the glasses 604 of the smart glasses 602.
- the display may be a projector arranged for projecting the virtual indicator on the glasses or on the eye of the user of the smart glasses.
- a user wearing the smart glasses may see the object 120 physically through the glasses, whereupon the virtual representation of 606 of the object information may be rendered on the glasses, projected on the glasses or on the user's eye.
- rendering the image on the smart glasses may not be required because the user can see the physical object through the glasses.
- the virtual representation is rendered as a virtual overlay on top of the physical world.
- the mobile device 102 may comprise a camera arranged for capturing a stream of images (a video) of the object 120, and the processor may be arranged for identifying an object feature of the object in the image, for determining an object feature position of the object feature in the image, for generating a virtual representation of the object information, for determining a position of the virtual representation in the image, wherein the position is related to the object feature position in the image, for rendering the image on a display and for rendering the virtual representation at the position as an overlay on the image for each image in the stream of images, or for a selected set of images in the stream of images.
- the processor may, for example, be arranged for executing the above-mentioned steps once every predetermined time period (e.g.
- the mobile device 102 may further comprise a user interface for receiving a user input related to a selection of the virtual representation.
- the user interface may be any type of interface arranged for receiving user input.
- the user interface may for example comprise a touch-sensitive device such as a touchpad, a touchscreen, one or more buttons and/or one or more sliders for receiving touch input. Additionally or alternatively, the user interface may comprise a microphone arranged for receiving voice commands from the user operating the mobile device.
- the user interface may comprise a gesture/motion detection means, such as a gyroscope and/or an accelerometer arranged for detecting gestures made with the mobile device 102, which gestures may be indicative of a selection of a virtual representation. Examples of such gestures are shaking the mobile device 102 or changing the orientation of the mobile device 102.
- the mobile device 102 may, for example, be smart glasses, and the shaking of a person's head may be the gesture.
- the user interface may comprise a camera (e.g. a front facing camera on a smartphone) arranged for detecting user gestures indicative of a selection of a virtual representation. Examples of such gestures are hand movements, eye-blinking, head shaking, etc. Additionally or alternatively, the user interface may comprise a means (e.g. a camera) for gaze detection and selection of a virtual representation may be based on a time period of gazing towards the position of the object. It should be noted that the above- mentioned user interfaces are mere examples of user input elements and illustrate rather than limit the invention, and that those skilled in the art will be able to design many alternative user interfaces without departing from the scope of the appended claims.
- the processor may be arranged for rendering additional information about the object feature on the display.
- the additional information may, for example, be audio or video content.
- the additional information may be
- a central server e.g. a server of the store wherein the product is for sale
- the processor may further be arranged for generating a message upon receiving the user input, which message is related to the object information.
- the object information may, for example, be related to a size of the shoe, and upon receiving the user input (for example by selecting the virtual representation on a touch screen), a message may be communicated to store personnel to request a different shoe size.
- the mobile device 102 may further comprise a communication unit for communicating the message, for example to a central server connected to the system 100.
- Fig. 2b illustrates an example of a virtual representation 202 which may be selected by a user.
- the virtual representation 202 may provide a link to additional information (e.g. to a video about the object or an object feature).
- the virtual representation 202 may be a feedback button which enables a user to provide feedback about the object 120.
- the feedback may be communicated from the mobile device 102 to a central server which stores the feedback from multiple users.
- the virtual representation 202 may initiate the transmission of a request message to personnel of the store wherein the object is located.
- the message may also provide an indication of where in the store the object is located, which helps the personnel to locate the user who transmitted the request.
- a user may provide input related to a question he or she has (e.g. a question about the size, colors, price, etc. which may for example be provided by the user by means of text entry or recorded spoken input).
- input related to a question he or she has e.g. a question about the size, colors, price, etc. which may for example be provided by the user by means of text entry or recorded spoken input.
- the processor of the mobile device 102 may be further arranged for communicating with a (central) controller, which controller may be connected to the light source.
- the controller may be arranged for controlling the light output of the light source 110.
- the controller may be arranged for driving the light source 110 such that it emits light comprising the code.
- the controller may be further arranged for determining which code is to be emitted by the light source 110.
- the controller may be further arranged for controlling the color, saturation and/or intensity of the light emitted by the light source 110.
- the mobile device 102 may be arranged for communicating a message to the controller when the user input has been received.
- the controller may receive the message from the mobile device 102 and change the light output of the light source 110.
- the light output may be determined based on a type of user input (e.g. a request for assistance from store personnel may result in a different light output color than a request for additional information about the object 120).
- the light source 110 may be any type of light source (e.g. an LED light source, an OLED light source, etc.) for emitting light comprising an embedded code.
- the system 100 may further comprise a driver (not shown) for driving the light source 100 such that the emitted light comprises the embedded code representative of object information about the object 120.
- the embedded code may be created by any known principle of embedding a code in light, for example by controlling a time- varying, modulated current to the one or more light sources to produce variations in the light output, by modulating the amplitude of the light, etc.
- the light source may be comprised in a supporting member 140 arranged for supporting the object 120.
- the supporting member 140 may, for example, be a product shelf.
- the light source 310 may be positioned such that it is arranged for illuminating the object 120.
- the light 312 emitted by the light source 310 may be reflected by the object 120 such that a camera or a photodiode of the mobile device 102 may detect the light 312 comprising the embedded code.
- the light source may be positioned such that a user operating the mobile device 102 and observing the object 120 is located between the light source and the object 120.
- the mobile device 102 may comprise a front-facing light detector (e.g. a camera or a photodiode) for capturing the light emitted by the light source 110 and a back-facing camera for capturing the image 104 of the object 120.
- Capturing the image 104 of the object 120 may further comprise capturing an image of the object 120 and the light source 110.
- the light source may have a predefined position in relation to the object 120.
- the light source 110 (which is comprised in a supporting member 140) has a predetermined position to the object 120.
- An object 120 such as a shoe, will always be positioned on top of the supporting member 140.
- the position of the light source 110 is predetermined (and therefore known to the processor)
- the position of the object can be inferred from the position of the light source 110.
- the system 100 may comprise a first light source 320 and a second light source 322.
- the processor may be arranged for determining the positions of the first and second light source 320, 322 in the image 104 in order to determine the position of the object 120 which, in this example, is related to the position of the light sources 320, 322. It may be beneficial to use a plurality of light sources with a predetermined position in relation to the object, because this enables determining the position of the object 120 more accurately.
- the system may further comprise a memory for storing object identifiers associated with objects 120.
- the mobile device 102 may be arranged for detecting light emitted by the light source 110, which light comprises an embedded code representative of an identifier of the object 120.
- the processor of the mobile device 102 may be arranged for retrieving the code from the light and for retrieving the identifier from the code.
- the mobile device 102 may further comprise a user interface arranged for receiving user input indicative of a selection of the object 120.
- the processor may be arranged for storing the identifier of the object 120 in a memory upon receiving the user input.
- the system may further comprise a controller for changing the color, saturation and/or the intensity of the light emitted by the light source upon receiving the user input. Changing the color, saturation and/or the intensity of the light emitted by the light source 110 provides feedback to the user that the user input has been received. The (new) color, saturation and/or intensity may be based on the type of user input.
- the user interface may be any type of user interface arranged for receiving the user input.
- the user interface may for example comprise a touch-sensitive device 104 such as a touchpad or a touchscreen.
- the processor may, for example, render a shopping cart icon on the touchscreen.
- the user may, for example, provide a touch input to select the icon and thereby add the object to the shopping cart, whereupon the mobile device 102 communicates to the control (e.g. via a wireless network) that the object has been added to a list in a memory (in this case to the shopping cart), whereupon the controller may change the light output of the light source 110.
- the user interface may be arranged for receiving a first swipe gesture in a first direction, whereupon the object 120 may be added to a favorites list, and for receiving a second swipe gesture in a second direction, whereupon the object 120 may be added to a non- favorites list.
- the controller may be arranged for controlling the light output based on which user input has been received (e.g. turn the light to green or blink the light green when the first swipe gesture has been received and turn the light to red or blink the light red when the second swipe gesture has been received).
- the user interface may comprise a microphone arranged for receiving voice commands from the user operating the mobile device.
- Example of voice commands may be: “Add this object to my shopping cart", or “I like this product”, whereupon the processor of the mobile device may store the identifier of the object in a part of the memory based on the user input command.
- the user interface may comprise a gesture/motion detection means, such as a gyroscope and/or an accelerometer arranged for detecting gestures made with the mobile device, which gestures may be indicative adding the object to a list in a memory. Examples of such gestures are shaking the mobile device 102 or changing the orientation of the mobile device 102, which may be indicative of adding the object 120 to a specific list in a memory.
- the mobile device 102 may, for example, be smart glasses, and the shaking of a person's head may be the gesture (e.g. nodding with the head may be indicative of adding a product to a first list and shaking with the head may be indicative of adding a product to a second list).
- the user interface may comprise a camera (e.g. a front facing camera on a smartphone) arranged for detecting user gestures indicative of adding the object 120 to a specific list in a memory. Examples of such gestures are hand movements, eye-blinking, head shaking, etc.
- the user interface may comprise a means (e.g. a camera) for gaze detection and adding the object 120 to a list or not may be based on a time period of gazing towards the position of the object 120.
- a means e.g. a camera
- the user input may be related to a step in a process, which process comprises a plurality of steps, and the controller may be arranged for changing the color, saturation and/or the intensity of the light based on the step.
- the process may, for example, be buying one or more objects in a store.
- the process may comprise a first step of adding one or more objects to a shopping cart, whereupon the controller may change the light output of a respective light source to orange, a second step of placing the order for one or more objects whereupon the controller may change the light output of respective light sources to blue, and a third step of paying for all the objects that are in the shopping cart, whereupon the controller may change the light output of respective light sources to green when payment has been received, or to red when payment has been cancelled.
- the controller may be arranged for changing the color, saturation and/or the intensity of the light emitted by the light source back to an original color and/or intensity after a period of time.
- the controller may, for example, be arranged for blinking the light briefly (e.g. for one second) to indicate that the user input has been received.
- the controller may be further arranged for changing the color, saturation and/or the intensity of the light emitted by the light source back to an original color and/or intensity when the code is no longer detected.
- the light output of the light source 110 may be changed back to an original color and/or intensity. If a new user approaches with a mobile device 102, he or she may add the object 120 to his or her own list, whereupon the light output is changed again accordingly.
- the controller may be further arranged for detecting a presence of a mobile device 102 and identifying the mobile device 102.
- the presence may, for example, be detected when the mobile device has been added to the same network as the controller, or, when a coded light emission of a light source 110 has been received at the mobile device 102, the mobile device 102 may communicate its presence (and an identifier) to the controller.
- the controller may change the color, saturation and/or the intensity of the light emitted by the light source 110 based on the identification of the mobile device.
- the controller may change the color of the light source associated with that object to a color, saturation and/or brightness that is associated with that list.
- the controller may be arranged for changing the color, saturation and/or the intensity of the light emitted by the light source 110 based on predefined light settings, user preferences, user profiles, etc.
- the controller may, for example, access a user profile associated with the user operating the mobile device 102 in order to retrieve user preference settings related to preferred light settings.
- a user may provide these preferences, which enables the user to personalize the light feedback that is provided upon providing a user input indicative of the selection of the object 120.
- the mobile device 102 may comprise a user interface arranged for receiving further user input related to setting these preferences. This may be beneficial, because it may help users in recognizing their own light effects,
- Fig. 4 shows schematically an embodiment of a lighting system according to the invention for providing information about an object 120.
- the lighting system may comprise a controller (not shown) for receiving object information about the object 120, for determining a first light setting and a second light setting based on the object information, wherein the difference between the first light setting and the second light setting is perceivable for a user.
- the object information may be a value indicative of any type of object information, and the controller may be arranged for determining color values for the light settings based on the value indicative of the object information.
- the controller e.g. a processor connected to a driver
- the controller may, for example, be a central controller for controlling a plurality of sets of light sources which are associated with a plurality of objects.
- the light emitted by the light source comprises the code representative of the object information.
- the controller may be commissioned in advance such that the light sources associated with (and preferably in close proximity to) an object are controlled according to the object information of that object. In other words, the association between the object and the light sources may be predefined.
- the system may further comprise a mobile device 102 comprising a light detector (e.g. a photodiode or a camera) for detecting light comprising an embedded code, which embedded code is representative of the object information.
- the mobile device 102 may further comprise a processor (not shown in Fig. 4) for retrieving the code from the light, for retrieving the object information and for rendering the object information on the display.
- the controller may be arranged for determining a first light setting and a second light setting based on the object information, whereupon the controller may control the first light source 410 and the second light source 412 according to the light settings.
- the object information about the object 120 may be related to available colors of the object 120.
- the controller may control the first light source 410 according to a green light setting, and the second light source 412 according to a red light setting. As soon as the red colored object 120 is no longer available, the controller may control the second light source 412 according to the green setting, or, alternatively, according to another color that is available.
- the processor retrieves the code, and therewith the additional object information, from the detected light.
- the processor may further render the object information on a display of the mobile device 102.
- the processor may, for example render the available colors of the object on the display.
- the object information about the object 120 may be related to reviews of the object 120.
- the controller may control the first light source 410 according to a bright green light setting (indicating that many people like this object 120), and the second light source 412 according to a dim red light setting (indicating that not many people dislike this object 120).
- the controller may dim the first light source 410 and increase the brightness of the second light source 412. If the user wishes to receive more information about what the colors of the light mean, he or she can point the light detector of the mobile device 102 to the light sources 410, 412, whereupon the processor may retrieve the code, and therewith the additional object information, from the detected light.
- the processor may further render the object information on a display of the mobile device 102.
- the processor may, for example render the popularity value (e.g. the number of positive and the number of negative reviews) on the display.
- the controller may be further arranged for controlling the first light source 410 and/or the second light source 412 such that the first light source and/or the second light source emit(s) the embedded code representative of the object information about the object 120. This requires a user to point the light detector of the mobile device 102 to the light sources 410, 412 in order to receive the additional object information.
- the processor may control a third light source such that it emits light comprising the embedded code. This may be beneficial, because a change in color, saturation and/or brightness of the first and second light source will not influence the signal strength of the code embedded in the light.
- the controller may be further arranged for controlling the first light source 410 such that it emits a first embedded code, which first embedded code is representative of first object information about the object, and for controlling the second light source 412 such that it emits a second embedded code, which second embedded code is representative of second object information about the object.
- the light detector may be arranged for detecting light comprising the first embedded code and detecting light comprising the second embedded code.
- the processor may be further arranged for retrieving the first embedded code and the second embedded code from the respective light, and for retrieving the first object information and the second object information from the respective codes, and for rendering the first object information and the second object information on the display.
- the first light source 410 may, for example, emit a code representative of the number of likes of the object
- the second light source 412 may, for example, emit a code representative of the number of dislikes of the object.
- the mobile device 102 may comprise a camera for capturing an image 104 of the first light source 410 and the second light source 412.
- the processor may be further arranged for rendering the image 104 on a display and rendering a virtual representation of the object information as an overlay on the image 104. Additionally, the processor may be arranged for determining a first light source position 410' of the first light source 410 and a second light source position 412' of the second light source 412 in the image 104.
- the processor may be further arranged for determining a position of the virtual representation on the display, wherein the position is related to at least one of the first light source position and the second light source position in the image, and for rendering the virtual representation of the object information at the position.
- This may be executed by the processor according to any one of the above-mentioned examples, wherein the position of the virtual representation is determined based on the position of the object. This enables the processor to provide the object information in the vicinity of the light sources in the image. Additionally, the processor may determine the position of the object 120 and/or object features in the image, and takes this position into account while determining the position of the virtual
- FIG. 4 illustrates an example wherein the location of the virtual
- representations 420, 422 of the object information are rendered next to the light sources 410', 412'.
- the controller may be further arranged for determining the first and second light settings based on a user profile of the user operating the mobile device 102.
- the controller may be arranged for accessing the user profile (e.g. a social media profile, a user profile stored on the mobile device, etc.).
- the controller may further select the object information from a set of stored object information based on the user profile. For example, the controller may retrieve from the user profile that a user has a certain foot size, and provide availability information about the sizes of a shoe (the object 120).
- the controller may receive information about how many friends of a user own an object 120, and determine the first light setting and the second light setting based thereon for each product of a plurality of products. This enables a user to see how many of his or her friends own an object 120 (e.g. a pair of shoes, or shoes of a specific brand).
- the first and the second light sources may be comprised in a linear lighting device, such as an LED strip.
- the linear lighting device may comprise a plurality of individually controllable light sources, and the controller may be arranged for controlling each of the plurality of light sources.
- the controller may be further arranged for determining which light sources to control according to the first light setting and which light sources to control according to the second light setting based on the object information.
- Fig. 5a illustrates an example of a linear lighting device 500 comprising a plurality of individually controllable light sources.
- the upper linear lighting device shows the individual light sources and the lower linear lighting device in Fig. 5a shows how the user may see the linear lighting device 500 when a diffuser has been attached to the linear lighting device 500.
- the controller may receive object information (for example stock information about the object 120), and determine how to control the light sources in order to communicate to a user how many shoes are still in stock. For example, the controller may receive object information indicative of that the current stock is 70% of the maximum stock. Therefore, the controller may determine to control the seven light sources 502, 512 on the left according to a first light setting (e.g. a green color) and control the three light sources 504, 514 on the right according to a second light setting (e.g. a red color).
- a first light setting e.g. a green color
- a second light setting e.g. a red color
- the processor may capture an image with the camera of the light sources, whereupon the processor of the mobile device renders a virtual representation 520 (see Fig. 5b) on the display, thereby informing the user what the (colored) light setting means.
- the virtual representation may, for example, be rendered on top of the linear lighting device 512', 514' in the image.
- the above-mentioned system may comprise different components, ranging from mobile devices to controllers, memories and light sources. It should be understood that these components may be coupled to each other, comprised in each other and/or that these components may be arranged for communicating with each other.
- the components may comprise transmitters and/or receivers arranged for communicating amongst each other.
- the transmitters and/or receivers may comprise hardware for transmitting and or receiving messages and/or control commands via any wired or wireless communication protocol.
- wired and wireless communication protocols may be used, for example Ethernet, DMX, DALI, USB, Bluetooth, Wi-Fi, Li-Fi, 3G, 4G or ZigBee.
- a specific communication technology may be selected based on the communication capabilities of the light sources, the controller and/or the mobile device, the power consumption of the communication driver for the (wireless) communication technology and/or the communication range of the signals.
- Fig. 7 shows schematically a method 700 according to the invention of providing information about an object.
- the method 700 comprises the steps of detecting 702 light emitted by a light source, which light comprises an embedded code representative of object information about the object, retrieving 704 the code from the light, retrieving 706 the object information, capturing 708 an image of the object, identifying 710 an object feature of the object in the image, determining 712 an object feature position of the object feature in the image, generating 714 a virtual representation of the object information, determining 716 a position of the virtual representation in the image, wherein the position is related to the object feature position in the image, rendering 718 the image on a display, and rendering 720 the virtual representation at the position as an overlay on the image.
- Fig. 8 shows schematically a method 800 according to the invention of providing information about an object.
- the method 800 comprises the steps of receiving 802 object information about the object, determining 804 a first light setting and a second light setting based on the object information, wherein the difference between the first light setting and the second light setting is perceivable for a user, controlling 806 a first light source according to the first light setting, controlling 808 a second light source according to the second light setting, detecting 810 light comprising an embedded code, which embedded code is representative of the object information, retrieving 812 the code from the light, retrieving 814 the object information and rendering 816 the object information on a display.
- Fig. 9 shows schematically a method 900 according to the invention of storing object identifiers.
- the method 900 comprises the steps of detecting 902 light emitted by a light source, which light comprises an embedded code representative of an identifier of an object, retrieving 904 the code from the light, retrieving 906 the identifier from the code, receiving 908 a user input indicative of a selection of the object, storing 910 the identifier of the object in a memory upon receiving the user input, changing 912 the color, saturation and/or the intensity of the light emitted by the light source upon receiving the user input.
- any reference signs placed between parentheses shall not be construed as limiting the claim.
- Use of the verb "comprise” and its conjugations does not exclude the presence of elements or steps other than those stated in a claim.
- the article “a” or “an” preceding an element does not exclude the presence of a plurality of such elements.
- the invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer or processing unit. In the device claim enumerating several means, several of these means may be embodied by one and the same item of hardware. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.
- aspects of the invention may be implemented in a computer program product, which may be a collection of computer program instructions stored on a computer readable storage device which may be executed by a computer.
- the instructions of the present invention may be in any interpretable or executable code mechanism, including but not limited to scripts, interpretable programs, dynamic link libraries (DLLs) or Java classes.
- the instructions can be provided as complete executable programs, partial executable programs, as modifications to existing programs (e.g. updates) or extensions for existing programs (e.g. plugins).
- parts of the processing of the present invention may be distributed over multiple computers or processors.
- Storage media suitable for storing computer program instructions include all forms of nonvolatile memory, including but not limited to EPROM, EEPROM and flash memory devices, magnetic disks such as the internal and external hard disk drives, removable disks and CD-ROM disks.
- the computer program product may be distributed on such a storage medium, or may be offered for download through HTTP, FTP, email or through a server connected to a network such as the Internet.
Landscapes
- Physics & Mathematics (AREA)
- Electromagnetism (AREA)
- Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
A method (800) of providing information about an object (120) is disclosed. The method (800) comprises receiving (802) object information about the object (120), determining (804) a first light setting and a second light setting based on the object information, wherein the difference between the first light setting and the second light setting is perceivable for a user, controlling (806) a first light source (410) according to the first light setting, controlling (808) a second light source (412) according to the second light setting, detecting (810) light comprising an embedded code, which embedded code is representative of the object information, retrieving (812) the code from the light, retrieving (814) the object information, and rendering (816) the object information on a display.
Description
A method of providing information about an object
FIELD OF THE INVENTION
The invention relates to a method of providing information about an object. The invention further relates to a system for providing information about an object. BACKGROUND
Smart devices, such as smartphones, are becoming our shopping assistants in stores. Examples of such shopping assistants include indoor positioning guidance systems, position or user based advertisements, augmented product information, etc. Retailers are more and more looking for solutions for omni-channel retailing wherein physical and digital retail channels are seamlessly integrated. For instance, retail environments may become interactive showrooms or experience centers where the products can be seen, experienced, used and ordered.
Patent application WO 2014/170290 Al discloses a coded light device comprising a light source configured to emit coded light, an identifier detection unit configured to detect a product identifier, and a processing unit having a code generator configured to generate a product-related code on basis of the product identifier, and a light source controller configured to control the light source to emit the product-related code by coding its light output. A customer may use a mobile device, such as a smart phone, to receive the light by means of a built in camera, detect the product-related code and receive product related data, on basis of the product-related code, from the product information source. Once the product-related code has been received at the mobile device, product-related information can be presented to the customer on the mobile device.
U.S. patent application 2014/0100997 Al discloses a method of enabling a user to perform a number of augmented re-ality shopping operations using a local client shopping application executable on a mobile device, for example viewing an augmented reality depiction of a shopping display using the mobile device and executing various "e- commerce" functions accessible via the augmented reality depiction.
SUMMARY OF THE INVENTION
It is an object of the present invention to provide a method for providing product information to a user.
According to a first aspect of the present invention, the object is achieved by a method of providing information about an object, the method comprising:
receiving object information about the object,
determining a first light setting and a second light setting based on the object information, wherein the difference between the first light setting and the second light setting is perceivable for a user,
controlling a first light source according to the first light setting, controlling a second light source according to the second light setting, detecting light comprising an embedded code, which embedded code is representative of the object information,
retrieving the code from the light,
retrieving the object information, and
rendering the object information on a display.
By controlling the first and the second light source according to object information, the first and the second light source provide information about the object, which information is perceivable for a user. If the user wishes to receive more information about the object, he or she may point a camera or a light detector of the mobile device towards a light source that emits light comprising the embedded code representative of the object information, and thereby receive additional object information. The additional object information is then rendered on the display, enabling a user to see the additional information. This is beneficial, because it allows a user to choose if he or she wishes to receive more object information.
The light comprises the code representative of the object information. The (controller of the) light sources may be commissioned in advance such that the light sources emit light comprising the embedded code that is associated with a respective object. In other words, the association between the object and the light source may be predefined.
Alternatively, a detector may detect which object is located at the light source (or which object is illuminated by the light source) in order to determine which code to emit. The object may for example comprise a tag (e.g. an RFID tag, a QR code, an NFC tag, etc.) which may be detected by a tag reader for identifying the object located at or nearby the light source.
The step of controlling the first light source according to the first light setting may further comprise controlling the first light source such that it emits the embedded code. This provides the advantage that the light source provides information about the object in two different ways: a first way wherein the light output of the light source is perceivable for a user, and a second way wherein the code comprised in the light output of the light source is perceivable for a light detector.
The step of controlling the first light source according to the first light setting may further comprise controlling the first light source such that it emits a first embedded code, which first embedded code is representative of first object information about the object, and the step of controlling the second light source according to the second light setting may further comprise controlling the second light source such that it emits a second embedded code, which second embedded code is representative of second object information about the object, and the step of detecting light comprising the embedded code may comprise detecting light comprising the first embedded code and detecting light comprising the second embedded code, and the step of retrieving the code from the light may comprise retrieving the first embedded code and the second embedded code from the respective light, and the step of retrieving the object information may comprise retrieving the first object information and the second object information from the respective codes, and the step of rendering the object information on the display may comprise rendering the first object information and the second object information on the display. This provides the advantage that each individual light source provides information about the object in two different ways: a first way wherein the light output of the respective light source is perceivable for a user, and a second way wherein the code comprised in the light output of the respective light source is perceivable for a light detector.
The method may further comprise the steps of: capturing an image of the first light source and the second light source, rendering the image on a display, and rendering a the object information as an overlay on the image. Additionally, the method may comprise the steps of: determining a first light source position of the first light source and a second light source position of the second light source in the image, and determining a position for the object information on the display, wherein the position is related to at least one of the first light source position and the second light source position in the image, and rendering the the object information at the position. By determining a position where the object information will be rendered in the image in relation to the position of the first light source and/or the
position of the second light source, the object information can be provided accurately at the light source in the image. This is advantageous, because it enables a user to see which object information is related to a light source.
The method may further comprise the steps of: accessing a user profile of the user, selecting the object information from a set of stored object information based on the user profile and determining the first light setting and the second light setting based on the selected object information. This is advantageous because the object information that is provided to a user is based on the user profile. If, for example, a user is interested in price information, price-related object information may be selected, while if a user is interested in stock information, stock-related object information may be selected.
The embedded code may comprise an identifier of the object. The identifier may be compared with a plurality of object identifiers stored in a memory in order to retrieve the object information related to the object associated with the identifier. Additionally or alternatively, the embedded code may comprise a link to the object information (e.g. a hyperlink, a URL, etc.) which link may provide information of a storage location of the object information in a memory. Alternatively, the embedded code may comprise the object information about the object.
The object information is indicative of at least one of price information, material information, size information, availability information, stock information, review information, object score information and popularity information.
According to a second aspect of the present invention, the object is achieved by a lighting system for providing information about an object, the lighting system comprising:
a controller for receiving object information about the object, for determining a first light setting and a second light setting based on the object information, wherein the difference between the first light setting and the second light setting is perceivable for a user, for controlling a first light source according to the first light setting, and for controlling a second light source according to the second light setting,
a mobile device comprising:
a display,
a light detector for detecting light comprising an embedded code, which embedded code is representative of the object information,
a processor arranged for retrieving the code from the light, for retrieving the object information, and for rendering the object information on the display.
The first light source and the second light source may be comprised in a linear lighting device (such as an LED strip). The linear lighting device may comprise a plurality of individually addressable light sources arranged for providing the object information. The linear lighting device may be comprised in a supporting member (e.g. a product shelf) arranged for supporting the object.
The controller may be arranged for controlling three or more light sources, and the controller may be arranged for determining which light sources to control according to the first light setting and which light sources to control according to the second light setting based on the object information. The controller may for example be arranged for controlling a linear lighting device with individually addressable light sources. The controller may control a first set of light sources according to the first light setting and a second set of light sources according to the second light setting. As such, the controller controls the light sources according to the object information. The object information may, for example, be stock information, and the controller may determine how many light sources to control according to the first light setting (and how many according to the second light setting) based on how many objects there are still in stock. As a result, a user can simply see how many objects are still in stock.
The light source may be arranged for illuminating the object. The light emitted by the light source and reflected from the object may be detected in order to retrieve the code, and therewith the object information, from the light.
It is beneficial if the object and the light sources are in each other's vicinity, or if the light source illuminates the object, because then the relation between the light sources and the object is clear.
It should be understood that the claimed lighting system may have similar and/or identical embodiments and advantages as the claimed method.
BRIEF DESCRIPTION OF THE DRAWINGS
The above, as well as additional objects, features and advantages of the disclosed systems, mobile devices and methods will be better understood through the following illustrative and non-limiting detailed description of embodiments of devices and methods, with reference to the appended drawings, in which:
Fig. 1 shows schematically an embodiment of a system according to the invention for providing information about an object;
Figs. 2a and 2b show schematically embodiments of a mobile device according to the invention for providing information about an object;
Fig. 3a and 3b show schematically embodiments of a product display according to the invention for providing light comprising an embedded code;
Fig. 4 shows schematically an embodiment of a system according to the invention for providing information about an object;
Fig. 5a shows schematically embodiments of a product display according to the invention for providing information about an object;
Fig. 5b shows schematically an embodiment of a mobile device according to the invention which captures the product display of fig. 5 a;
Fig. 6 shows schematically an embodiment of a part of a wearable mobile device according to the invention for providing information about an object; and
Fig. 7 shows schematically a method according to the invention of providing information about an object;
Fig. 8 shows schematically a method according to the invention of providing information about an object; and
Fig. 9 shows schematically a method according to the invention of storing object identifiers.
All the figures are schematic, not necessarily to scale, and generally only show parts which are necessary in order to elucidate the invention, wherein other parts may be omitted or merely suggested.
DETAILED DESCRIPTION OF EMBODIMENTS
Fig. 1 shows schematically an embodiment of a system 100 according to the invention for providing information about an object 120. The system 100 comprises a mobile device 102 for providing information 132, 134 about an object 120. The mobile device 102 may be any type of mobile device 120. Examples of a mobile device include but are not limited to portable smart devices, such as smartphones and tablet pes, and wearable smart devices such as smart watches and smart goggles. The mobile device 102 may be arranged for capturing an image 104 of the object 102 and for detecting light emitted by a light source 110, which light comprises an embedded code representative of object information about the object 120. The mobile device 102 may comprise a display and a processor (not shown) arranged for retrieving the code from the light. The processor may be further arranged for retrieving the object information, for identifying an object feature of the object in the image
104, for determining an object feature position of the object feature in the image 104, for generating a virtual representation of the object information, for determining a position of the virtual representation in the image 104, wherein the position is related to the object feature position in the image 104, for rendering the image 104 (and therewith the object 120') on a display, and for rendering the virtual representation at the position as an overlay on the image 104.
Figs. 1-6 illustrate examples of objects 120. In these examples the object is a shoe, but it should be noted that the object can be any type of object. Examples include but are not limited to consumer goods such as clothing, food, vehicles, household appliances, robots and electronic devices and animate objects such as animals.
The processor (e.g. a microchip, circuitry, a microcontroller, etc.) of the mobile device 102 may be arranged for retrieving the code from the light. The mobile device 102 may comprise a light detector (such as a photodiode) for detecting the code in the light emitted by the light source 110 and a camera (not shown) for capturing the image 104 of the object. Alternatively, the mobile device 102 comprises a camera for capturing the image 104 of the object and for detecting the light emitted by the light source. The camera may, for example, be a rolling shutter camera or a global shutter camera. Techniques for retrieving the code from the light (either by photodiode or by camera) are well known in the art and will not be discussed further.
The processor of the mobile device 102 may be arranged for retrieving the object information. The object information may be retrieved based on the code. The code may comprise, for example, an identifier of the object 120. The processor may be arranged for accessing a (remote) memory, which memory may be arranged for storing associations between object identifiers and object information. The processor may compare the identifier of the object 120 with the stored identifiers, and, if a stored identifier matches the object identifier, determine the object information for the identified object 120. Additionally or alternatively, the code may comprise a link to the object information. The link may, for example, be a URL or a hyperlink which comprises an address that points to a location where the object information is stored. This enables the processor to find the object information of the object 120 based on the link. Additionally or alternatively, the code may comprise the object information. In the latter, the processor may directly retrieve the object information from the code.
The object information may, for example, be price information, material information, size information, stock information, color information, product age information,
availability information, user review information, object score information and/or popularity information. The object information may comprise multiple types of object information (e.g. size and price information).
The processor may be further arranged for identifying an object feature 122, 124, 126 of the object 120' in the image 104. The processor may comprise an image analysis module for identifying the object feature. The object feature 122, 124, 126 may be related to an outline of at least a part of the object 120, a color of at least a part of the object 120, an area of least a part of the object 120, a pattern of at least a part of an object 120 and/or a marker (e.g. a barcode or a QR-code) attached to the object 120. The processor may, for example, identify the object feature 122, 124, 126 of the object 120' in the image 104 by accessing a memory arranged for storing object feature information about at least one object, comparing at least a part of the image 104 with stored object feature information, so as to determine if a similarity criterion has been met between a stored object feature and the at least a part of the image 104, and identifying, if the similarity criterion has been met, the object feature 122, 124, 126 in the image 104. It should be noted that any image analysis technique for identifying objects and object features from images 104 known in the art may be used for identifying the object feature. Fig. 1 illustrates three object features that may be identified: a first object feature 122 (i.e. the nose of the shoe 120 the striped pattern), a second object feature 124, (i.e. the sole of the shoe 120) and a third object feature 126 (i.e. the logo on the shoe 120).
Upon identifying an object feature 122, 124, 126 of the object 120 in the image 104, the processor may determine a position of the object feature (the object feature position) in the image 104. The image analysis module may be arranged for determining this position. The position of the object feature may be an area in the image 104 with specific coordinates in the image. The processor may be arranged for determining a plurality of positions of different object features. Object features may, for example, be part of other object features (e.g. a first object feature may be the outline of the shoe 120, and a second object feature may be the logo 126 on the shoe 120, which is located in the area created by the outline of the shoe). It should be noted that any image analysis technique known in the art may be used for determining a position of an object feature in an image 104.
The processor may be further arranged for generating a virtual representation 132, 134 of the object information. The virtual representationl32, 134 may, for example, be text, a color, a logo, a virtual price tag, a video, an animation, a feedback button, a virtual character (e.g. providing spoken instructions related to the object feature), or an alternative
object feature (e.g. showing alternative colors for the object or object part) etc. representative of the object information. The processor may be further arranged for determining a position of the virtual representation 132, 134 in the image 104, wherein the position is related to the (previously determined) object feature position in the image 104. The processor may be arranged for determining the position of the virtual representation 132, 134 such that it is different from the object feature position. The position of the virtual representation may for example be an area in the image 104 with coordinates different from the coordinates of the area of the object feature position. The processor may be arranged for determining the position of the virtual representation 132, 134 such that it is in vicinity of the object feature position. The position of the virtual representation may for example be an area in the image 104 with coordinates nearby the coordinates of the area of the object feature position. Fig. 1 illustrates such an example. The processor (not shown) of the mobile device 102 may determine the position of the sole 124' of the shoe 120' in the image 104, and determine a position for the virtual representation 134 of the object information about the sole 124 of the shoe 120. In this example, the virtual representation is textual information about the object information about the object feature 124, for example information about the materials used for the sole 124 of the shoe 120.
The processor may be arranged for determining a plurality of positions of different object features 122, 124, 126, and the processor may position the virtual representation of an object feature in relation to the position of that object feature and in relation to the other object features. For example, the processor may identify a first object feature (e.g. the outline of the shoe 120') and a second object feature (e.g. the striped nose 122' of the shoe 120') in the image 104 and determine to position the virtual representation 132 of the object information about the object feature 122 outside the outline of the shoe 120'. In this example, the virtual representation 132 is textual information about the object information about the object feature 122, for example information about the pattern on the nose 122 of the shoe 120.
The processor may be arranged for determining the position of the virtual representation 132, 134 such that it overlaps the object feature position. The position of the virtual representation 132, 134 may for example be an area in the image 104 with coordinates similar to the coordinates of the area of the object feature position. Fig. 2a illustrates such an example. The processor (not shown) of the mobile device 102 may determine the position of the nose 122' of the shoe 120' in the image 104, and determine a position for the virtual representation 122" of the object information about the nose 122 of the shoe 120. The
position of the virtual representation 122" is in this case the same position as the position of the object feature 122'. The virtual representation 122" is in this example an overlay on top of the 'actual' shoe 120. This enables a user operating the mobile device 102 to see how the object feature may look like. Additionally, a user may provide a user input to change the virtual representation 122" to a different representation (e.g. by selecting a different color for the nose of the shoe) to see how the shoe would look like in that color. In a similar fashion, the user may select a different color for the logo 126' on the shoe 120'.
The processor may be further arranged for rendering the image on a display, and for rendering the virtual representation at the position as an overlay on the image.
Examples of displays include but are not limited to LED displays, LCD displays and OLED displays. In embodiments, as illustrated in Fig. 6a, the mobile device may be smart glasses 602. The smart glasses may comprise a camera 608 for detecting the object 120 and a display 604. The display may be at least one of the glasses 604 of the smart glasses 602.
Alternatively, the display may be a projector arranged for projecting the virtual indicator on the glasses or on the eye of the user of the smart glasses. A user wearing the smart glasses may see the object 120 physically through the glasses, whereupon the virtual representation of 606 of the object information may be rendered on the glasses, projected on the glasses or on the user's eye. In this case, rendering the image on the smart glasses may not be required because the user can see the physical object through the glasses. As a result, the virtual representation is rendered as a virtual overlay on top of the physical world.
The mobile device 102 may comprise a camera arranged for capturing a stream of images (a video) of the object 120, and the processor may be arranged for identifying an object feature of the object in the image, for determining an object feature position of the object feature in the image, for generating a virtual representation of the object information, for determining a position of the virtual representation in the image, wherein the position is related to the object feature position in the image, for rendering the image on a display and for rendering the virtual representation at the position as an overlay on the image for each image in the stream of images, or for a selected set of images in the stream of images. The processor may, for example, be arranged for executing the above-mentioned steps once every predetermined time period (e.g. every second) when the stream of images is captured. This may reduce the required computing power for the processor. Capturing a stream of images and rendering the images and the virtual representation over time provides the advantage that the user experiences frequent updates and up-to-date positioning of the virtual representation on top of the captured physical world.
The mobile device 102 may further comprise a user interface for receiving a user input related to a selection of the virtual representation. The user interface may be any type of interface arranged for receiving user input. The user interface may for example comprise a touch-sensitive device such as a touchpad, a touchscreen, one or more buttons and/or one or more sliders for receiving touch input. Additionally or alternatively, the user interface may comprise a microphone arranged for receiving voice commands from the user operating the mobile device. Additionally or alternatively, the user interface may comprise a gesture/motion detection means, such as a gyroscope and/or an accelerometer arranged for detecting gestures made with the mobile device 102, which gestures may be indicative of a selection of a virtual representation. Examples of such gestures are shaking the mobile device 102 or changing the orientation of the mobile device 102. The mobile device 102 may, for example, be smart glasses, and the shaking of a person's head may be the gesture.
Additionally or alternatively, the user interface may comprise a camera (e.g. a front facing camera on a smartphone) arranged for detecting user gestures indicative of a selection of a virtual representation. Examples of such gestures are hand movements, eye-blinking, head shaking, etc. Additionally or alternatively, the user interface may comprise a means (e.g. a camera) for gaze detection and selection of a virtual representation may be based on a time period of gazing towards the position of the object. It should be noted that the above- mentioned user interfaces are mere examples of user input elements and illustrate rather than limit the invention, and that those skilled in the art will be able to design many alternative user interfaces without departing from the scope of the appended claims.
Upon receiving the user input, the processor may be arranged for rendering additional information about the object feature on the display. The additional information may, for example, be audio or video content. The additional information may be
communicated from a central server (e.g. a server of the store wherein the product is for sale) to the mobile device via a wireless network. This enables a user to request additional information about the object feature. The processor may further be arranged for generating a message upon receiving the user input, which message is related to the object information. The object information may, for example, be related to a size of the shoe, and upon receiving the user input (for example by selecting the virtual representation on a touch screen), a message may be communicated to store personnel to request a different shoe size. The mobile device 102 may further comprise a communication unit for communicating the message, for example to a central server connected to the system 100.
Fig. 2b illustrates an example of a virtual representation 202 which may be selected by a user. The virtual representation 202 may provide a link to additional information (e.g. to a video about the object or an object feature). Alternatively, the virtual representation 202 may be a feedback button which enables a user to provide feedback about the object 120. The feedback may be communicated from the mobile device 102 to a central server which stores the feedback from multiple users. Alternatively, the virtual representation 202 may initiate the transmission of a request message to personnel of the store wherein the object is located. The message may also provide an indication of where in the store the object is located, which helps the personnel to locate the user who transmitted the request.
Additionally, a user may provide input related to a question he or she has (e.g. a question about the size, colors, price, etc. which may for example be provided by the user by means of text entry or recorded spoken input).
The processor of the mobile device 102 may be further arranged for communicating with a (central) controller, which controller may be connected to the light source. The controller may be arranged for controlling the light output of the light source 110. The controller may be arranged for driving the light source 110 such that it emits light comprising the code. The controller may be further arranged for determining which code is to be emitted by the light source 110. The controller may be further arranged for controlling the color, saturation and/or intensity of the light emitted by the light source 110. The mobile device 102 may be arranged for communicating a message to the controller when the user input has been received. The controller may receive the message from the mobile device 102 and change the light output of the light source 110. The light output may be determined based on a type of user input (e.g. a request for assistance from store personnel may result in a different light output color than a request for additional information about the object 120).
The light source 110 may be any type of light source (e.g. an LED light source, an OLED light source, etc.) for emitting light comprising an embedded code. The system 100 may further comprise a driver (not shown) for driving the light source 100 such that the emitted light comprises the embedded code representative of object information about the object 120. The embedded code may be created by any known principle of embedding a code in light, for example by controlling a time- varying, modulated current to the one or more light sources to produce variations in the light output, by modulating the amplitude of the light, etc.
The light source may be comprised in a supporting member 140 arranged for supporting the object 120. The supporting member 140 may, for example, be a product shelf.
Alternatively, as illustrated in Fig. 3a, the light source 310 may be positioned such that it is arranged for illuminating the object 120. The light 312 emitted by the light source 310 may be reflected by the object 120 such that a camera or a photodiode of the mobile device 102 may detect the light 312 comprising the embedded code. Alternatively, the light source may be positioned such that a user operating the mobile device 102 and observing the object 120 is located between the light source and the object 120. The mobile device 102 may comprise a front-facing light detector (e.g. a camera or a photodiode) for capturing the light emitted by the light source 110 and a back-facing camera for capturing the image 104 of the object 120.
Capturing the image 104 of the object 120 may further comprise capturing an image of the object 120 and the light source 110. The light source may have a predefined position in relation to the object 120. In the example of Fig. 1, the light source 110 (which is comprised in a supporting member 140) has a predetermined position to the object 120. An object 120, such as a shoe, will always be positioned on top of the supporting member 140. Thus, if the position of the light source 110 is predetermined (and therefore known to the processor), the position of the object (and therewith the position of an object feature) can be inferred from the position of the light source 110. In a further embodiment, as illustrated in Fig. 3b, the system 100 may comprise a first light source 320 and a second light source 322. The processor may be arranged for determining the positions of the first and second light source 320, 322 in the image 104 in order to determine the position of the object 120 which, in this example, is related to the position of the light sources 320, 322. It may be beneficial to use a plurality of light sources with a predetermined position in relation to the object, because this enables determining the position of the object 120 more accurately.
The system may further comprise a memory for storing object identifiers associated with objects 120. The mobile device 102 may be arranged for detecting light emitted by the light source 110, which light comprises an embedded code representative of an identifier of the object 120. The processor of the mobile device 102 may be arranged for retrieving the code from the light and for retrieving the identifier from the code. The mobile device 102 may further comprise a user interface arranged for receiving user input indicative of a selection of the object 120. The processor may be arranged for storing the identifier of the object 120 in a memory upon receiving the user input. The system may further comprise a controller for changing the color, saturation and/or the intensity of the light emitted by the light source upon receiving the user input. Changing the color, saturation and/or the intensity of the light emitted by the light source 110 provides feedback to the user that the user input
has been received. The (new) color, saturation and/or intensity may be based on the type of user input.
The user interface may be any type of user interface arranged for receiving the user input. The user interface may for example comprise a touch-sensitive device 104 such as a touchpad or a touchscreen. Upon detecting the code, and thereby identifying the object 120, the processor may, for example, render a shopping cart icon on the touchscreen. The user may, for example, provide a touch input to select the icon and thereby add the object to the shopping cart, whereupon the mobile device 102 communicates to the control (e.g. via a wireless network) that the object has been added to a list in a memory (in this case to the shopping cart), whereupon the controller may change the light output of the light source 110. Alternatively, the user interface may be arranged for receiving a first swipe gesture in a first direction, whereupon the object 120 may be added to a favorites list, and for receiving a second swipe gesture in a second direction, whereupon the object 120 may be added to a non- favorites list. The controller may be arranged for controlling the light output based on which user input has been received (e.g. turn the light to green or blink the light green when the first swipe gesture has been received and turn the light to red or blink the light red when the second swipe gesture has been received).
Additionally or alternatively, the user interface may comprise a microphone arranged for receiving voice commands from the user operating the mobile device. Example of voice commands may be: "Add this object to my shopping cart", or "I like this product", whereupon the processor of the mobile device may store the identifier of the object in a part of the memory based on the user input command.
Additionally or alternatively, the user interface may comprise a gesture/motion detection means, such as a gyroscope and/or an accelerometer arranged for detecting gestures made with the mobile device, which gestures may be indicative adding the object to a list in a memory. Examples of such gestures are shaking the mobile device 102 or changing the orientation of the mobile device 102, which may be indicative of adding the object 120 to a specific list in a memory. The mobile device 102 may, for example, be smart glasses, and the shaking of a person's head may be the gesture (e.g. nodding with the head may be indicative of adding a product to a first list and shaking with the head may be indicative of adding a product to a second list).
Additionally or alternatively, the user interface may comprise a camera (e.g. a front facing camera on a smartphone) arranged for detecting user gestures indicative of
adding the object 120 to a specific list in a memory. Examples of such gestures are hand movements, eye-blinking, head shaking, etc.
Additionally or alternatively, the user interface may comprise a means (e.g. a camera) for gaze detection and adding the object 120 to a list or not may be based on a time period of gazing towards the position of the object 120.
It should be noted that the above-mentioned user interfaces are mere examples of user interfaces and illustrate rather than limit the invention, and that those skilled in the art will be able to design many alternative user interfaces without departing from the scope of the appended claims.
The user input may be related to a step in a process, which process comprises a plurality of steps, and the controller may be arranged for changing the color, saturation and/or the intensity of the light based on the step. The process may, for example, be buying one or more objects in a store. The process may comprise a first step of adding one or more objects to a shopping cart, whereupon the controller may change the light output of a respective light source to orange, a second step of placing the order for one or more objects whereupon the controller may change the light output of respective light sources to blue, and a third step of paying for all the objects that are in the shopping cart, whereupon the controller may change the light output of respective light sources to green when payment has been received, or to red when payment has been cancelled.
The controller may be arranged for changing the color, saturation and/or the intensity of the light emitted by the light source back to an original color and/or intensity after a period of time. The controller may, for example, be arranged for blinking the light briefly (e.g. for one second) to indicate that the user input has been received. The controller may be further arranged for changing the color, saturation and/or the intensity of the light emitted by the light source back to an original color and/or intensity when the code is no longer detected. When the user moves away from the object (and also away from the product) and the light (and therewith the code) is no longer detected by the mobile device 102, the light output of the light source 110 may be changed back to an original color and/or intensity. If a new user approaches with a mobile device 102, he or she may add the object 120 to his or her own list, whereupon the light output is changed again accordingly.
The controller may be further arranged for detecting a presence of a mobile device 102 and identifying the mobile device 102. The presence may, for example, be detected when the mobile device has been added to the same network as the controller, or, when a coded light emission of a light source 110 has been received at the mobile device 102,
the mobile device 102 may communicate its presence (and an identifier) to the controller. Upon detecting the presence of the mobile device 102, the controller may change the color, saturation and/or the intensity of the light emitted by the light source 110 based on the identification of the mobile device. If, for example, the mobile device 102 has been interacting with an object 120 before, and the object 120 has already been added to a list in a memory, the controller may change the color of the light source associated with that object to a color, saturation and/or brightness that is associated with that list.
The controller may be arranged for changing the color, saturation and/or the intensity of the light emitted by the light source 110 based on predefined light settings, user preferences, user profiles, etc. The controller may, for example, access a user profile associated with the user operating the mobile device 102 in order to retrieve user preference settings related to preferred light settings. A user may provide these preferences, which enables the user to personalize the light feedback that is provided upon providing a user input indicative of the selection of the object 120. The mobile device 102 may comprise a user interface arranged for receiving further user input related to setting these preferences. This may be beneficial, because it may help users in recognizing their own light effects,
Fig. 4 shows schematically an embodiment of a lighting system according to the invention for providing information about an object 120. The lighting system may comprise a controller (not shown) for receiving object information about the object 120, for determining a first light setting and a second light setting based on the object information, wherein the difference between the first light setting and the second light setting is perceivable for a user. The object information may be a value indicative of any type of object information, and the controller may be arranged for determining color values for the light settings based on the value indicative of the object information. The controller (e.g. a processor connected to a driver) may be further arranged for controlling a first light source 410 according to the first light setting and for controlling a second light source 412 according to the second light setting. The controller may, for example, be a central controller for controlling a plurality of sets of light sources which are associated with a plurality of objects. The light emitted by the light source comprises the code representative of the object information. The controller may be commissioned in advance such that the light sources associated with (and preferably in close proximity to) an object are controlled according to the object information of that object. In other words, the association between the object and the light sources may be predefined.
The system may further comprise a mobile device 102 comprising a light detector (e.g. a photodiode or a camera) for detecting light comprising an embedded code, which embedded code is representative of the object information. The mobile device 102 may further comprise a processor (not shown in Fig. 4) for retrieving the code from the light, for retrieving the object information and for rendering the object information on the display.
The controller may be arranged for determining a first light setting and a second light setting based on the object information, whereupon the controller may control the first light source 410 and the second light source 412 according to the light settings. This is beneficial because the object information is communicated to a user via light (i.e. via the two light sources). For example, the object information about the object 120 may be related to available colors of the object 120. The controller may control the first light source 410 according to a green light setting, and the second light source 412 according to a red light setting. As soon as the red colored object 120 is no longer available, the controller may control the second light source 412 according to the green setting, or, alternatively, according to another color that is available. If the user wishes to receive more information about what the colors of the light mean, he or she can point the light detector of the mobile device 102 to the light sources 410, 412, whereupon the processor retrieves the code, and therewith the additional object information, from the detected light. The processor may further render the object information on a display of the mobile device 102. The processor may, for example render the available colors of the object on the display. In another example, the object information about the object 120 may be related to reviews of the object 120. The controller may control the first light source 410 according to a bright green light setting (indicating that many people like this object 120), and the second light source 412 according to a dim red light setting (indicating that not many people dislike this object 120). As soon as more people start disliking the object 120, the controller may dim the first light source 410 and increase the brightness of the second light source 412. If the user wishes to receive more information about what the colors of the light mean, he or she can point the light detector of the mobile device 102 to the light sources 410, 412, whereupon the processor may retrieve the code, and therewith the additional object information, from the detected light. The processor may further render the object information on a display of the mobile device 102. The processor may, for example render the popularity value (e.g. the number of positive and the number of negative reviews) on the display.
The controller may be further arranged for controlling the first light source 410 and/or the second light source 412 such that the first light source and/or the second light
source emit(s) the embedded code representative of the object information about the object 120. This requires a user to point the light detector of the mobile device 102 to the light sources 410, 412 in order to receive the additional object information. Alternatively, the processor may control a third light source such that it emits light comprising the embedded code. This may be beneficial, because a change in color, saturation and/or brightness of the first and second light source will not influence the signal strength of the code embedded in the light.
The controller may be further arranged for controlling the first light source 410 such that it emits a first embedded code, which first embedded code is representative of first object information about the object, and for controlling the second light source 412 such that it emits a second embedded code, which second embedded code is representative of second object information about the object. In this embodiment, the light detector may be arranged for detecting light comprising the first embedded code and detecting light comprising the second embedded code. The processor may be further arranged for retrieving the first embedded code and the second embedded code from the respective light, and for retrieving the first object information and the second object information from the respective codes, and for rendering the first object information and the second object information on the display. The first light source 410 may, for example, emit a code representative of the number of likes of the object, and the second light source 412 may, for example, emit a code representative of the number of dislikes of the object.
The mobile device 102 may comprise a camera for capturing an image 104 of the first light source 410 and the second light source 412. The processor may be further arranged for rendering the image 104 on a display and rendering a virtual representation of the object information as an overlay on the image 104. Additionally, the processor may be arranged for determining a first light source position 410' of the first light source 410 and a second light source position 412' of the second light source 412 in the image 104. The processor may be further arranged for determining a position of the virtual representation on the display, wherein the position is related to at least one of the first light source position and the second light source position in the image, and for rendering the virtual representation of the object information at the position. This may be executed by the processor according to any one of the above-mentioned examples, wherein the position of the virtual representation is determined based on the position of the object. This enables the processor to provide the object information in the vicinity of the light sources in the image. Additionally, the processor may determine the position of the object 120 and/or object features in the image,
and takes this position into account while determining the position of the virtual
representation. Fig. 4 illustrates an example wherein the location of the virtual
representations 420, 422 of the object information are rendered next to the light sources 410', 412'. Thus, if a user requires additional information about what object information the light settings of the light sources convey, he or she can simply point the camera to the light sources, whereupon the additional information is provided next to the light sources.
The controller may be further arranged for determining the first and second light settings based on a user profile of the user operating the mobile device 102. The controller may be arranged for accessing the user profile (e.g. a social media profile, a user profile stored on the mobile device, etc.). The controller may further select the object information from a set of stored object information based on the user profile. For example, the controller may retrieve from the user profile that a user has a certain foot size, and provide availability information about the sizes of a shoe (the object 120). In another example, the controller may receive information about how many friends of a user own an object 120, and determine the first light setting and the second light setting based thereon for each product of a plurality of products. This enables a user to see how many of his or her friends own an object 120 (e.g. a pair of shoes, or shoes of a specific brand).
The first and the second light sources may be comprised in a linear lighting device, such as an LED strip. The linear lighting device may comprise a plurality of individually controllable light sources, and the controller may be arranged for controlling each of the plurality of light sources. The controller may be further arranged for determining which light sources to control according to the first light setting and which light sources to control according to the second light setting based on the object information. Fig. 5a illustrates an example of a linear lighting device 500 comprising a plurality of individually controllable light sources. The upper linear lighting device shows the individual light sources and the lower linear lighting device in Fig. 5a shows how the user may see the linear lighting device 500 when a diffuser has been attached to the linear lighting device 500. The controller may receive object information (for example stock information about the object 120), and determine how to control the light sources in order to communicate to a user how many shoes are still in stock. For example, the controller may receive object information indicative of that the current stock is 70% of the maximum stock. Therefore, the controller may determine to control the seven light sources 502, 512 on the left according to a first light setting (e.g. a green color) and control the three light sources 504, 514 on the right according to a second light setting (e.g. a red color). A user will be able to see the object information when he or
she looks at the linear lighting device, but the specific information that is being conveyed may not be clear to the user. The user may therefore point the light detector (e.g. the camera) towards the light sources, whereupon a code in the light (which may be emitted by one or more of the light sources comprised in the linear lighting device) is detected. The processor may capture an image with the camera of the light sources, whereupon the processor of the mobile device renders a virtual representation 520 (see Fig. 5b) on the display, thereby informing the user what the (colored) light setting means. The virtual representation may, for example, be rendered on top of the linear lighting device 512', 514' in the image.
The above-mentioned system may comprise different components, ranging from mobile devices to controllers, memories and light sources. It should be understood that these components may be coupled to each other, comprised in each other and/or that these components may be arranged for communicating with each other. The components may comprise transmitters and/or receivers arranged for communicating amongst each other. The transmitters and/or receivers may comprise hardware for transmitting and or receiving messages and/or control commands via any wired or wireless communication protocol.
Various wired and wireless communication protocols may be used, for example Ethernet, DMX, DALI, USB, Bluetooth, Wi-Fi, Li-Fi, 3G, 4G or ZigBee. A specific communication technology may be selected based on the communication capabilities of the light sources, the controller and/or the mobile device, the power consumption of the communication driver for the (wireless) communication technology and/or the communication range of the signals.
Fig. 7 shows schematically a method 700 according to the invention of providing information about an object. The method 700 comprises the steps of detecting 702 light emitted by a light source, which light comprises an embedded code representative of object information about the object, retrieving 704 the code from the light, retrieving 706 the object information, capturing 708 an image of the object, identifying 710 an object feature of the object in the image, determining 712 an object feature position of the object feature in the image, generating 714 a virtual representation of the object information, determining 716 a position of the virtual representation in the image, wherein the position is related to the object feature position in the image, rendering 718 the image on a display, and rendering 720 the virtual representation at the position as an overlay on the image.
Fig. 8 shows schematically a method 800 according to the invention of providing information about an object. The method 800 comprises the steps of receiving 802 object information about the object, determining 804 a first light setting and a second light setting based on the object information, wherein the difference between the first light setting
and the second light setting is perceivable for a user, controlling 806 a first light source according to the first light setting, controlling 808 a second light source according to the second light setting, detecting 810 light comprising an embedded code, which embedded code is representative of the object information, retrieving 812 the code from the light, retrieving 814 the object information and rendering 816 the object information on a display.
Fig. 9 shows schematically a method 900 according to the invention of storing object identifiers. The method 900 comprises the steps of detecting 902 light emitted by a light source, which light comprises an embedded code representative of an identifier of an object, retrieving 904 the code from the light, retrieving 906 the identifier from the code, receiving 908 a user input indicative of a selection of the object, storing 910 the identifier of the object in a memory upon receiving the user input, changing 912 the color, saturation and/or the intensity of the light emitted by the light source upon receiving the user input.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design many alternative embodiments without departing from the scope of the appended claims.
In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim. Use of the verb "comprise" and its conjugations does not exclude the presence of elements or steps other than those stated in a claim. The article "a" or "an" preceding an element does not exclude the presence of a plurality of such elements. The invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer or processing unit. In the device claim enumerating several means, several of these means may be embodied by one and the same item of hardware. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.
Aspects of the invention may be implemented in a computer program product, which may be a collection of computer program instructions stored on a computer readable storage device which may be executed by a computer. The instructions of the present invention may be in any interpretable or executable code mechanism, including but not limited to scripts, interpretable programs, dynamic link libraries (DLLs) or Java classes. The instructions can be provided as complete executable programs, partial executable programs, as modifications to existing programs (e.g. updates) or extensions for existing programs (e.g. plugins). Moreover, parts of the processing of the present invention may be distributed over multiple computers or processors.
Storage media suitable for storing computer program instructions include all forms of nonvolatile memory, including but not limited to EPROM, EEPROM and flash memory devices, magnetic disks such as the internal and external hard disk drives, removable disks and CD-ROM disks. The computer program product may be distributed on such a storage medium, or may be offered for download through HTTP, FTP, email or through a server connected to a network such as the Internet.
Claims
CLAIMS:
1. A method (800) of providing information about an object (120), the method
(800) comprising:
receiving (802) object information about the object (120),
determining (804) a first light setting and a second light setting based on the object information, wherein the difference between the first light setting and the second light setting is perceivable for a user,
controlling (806) a first light source (410) according to the first light setting, controlling (808) a second light source (412) according to the second light setting,
- detecting (810) light comprising an embedded code, which embedded code is representative of the object information,
retrieving (812) the code from the light,
retrieving (814) the object information,
capturing an image (104) of the first light source (410) and the second light source (412),
determining a first light source position of the first light source (410) and a second light source position of the second light source (412) in the image (104),
determining a position for the object information on the display, wherein the position is related to at least one of the first light source position and the second light source position in the image (104),
rendering the image (104) on a display, and
rendering (816) the object information as an overlay on the image (104) at the position on the display. 2. The method (800) of claim 1, wherein the step of controlling the first light source (410) according to the first light setting further comprises controlling the first light source (410) such that it emits the embedded code.
3. The method (800) of claim 2, wherein the step of controlling the first light source (410) according to the first light setting further comprises controlling the first light source (410) such that it emits a first embedded code, which first embedded code is representative of first object information about the object (120),
wherein the step of controlling the second light source (412) according to the second light setting further comprises controlling the second light source (412) such that it emits a second embedded code, which second embedded code is representative of second object information about the object (120),
wherein the step of detecting light comprising the embedded code comprises detecting light comprising the first embedded code and detecting light comprising the second embedded code,
wherein the step of retrieving the code from the light comprises retrieving the first embedded code and the second embedded code from the respective light,
wherein the step of retrieving the object information comprises retrieving the first object information and the second object information from the respective codes,
and wherein the step of rendering the object information on the display comprises rendering the first object information and the second object information on the display. 4. The method (800) of any one of the preceding claims, further comprising the steps of:
accessing a user profile of the user,
selecting the object information from a set of stored object information based on the user profile,
- determining the first light setting and the second light setting based on the selected object information.
5. The method (800) of any one of the preceding claims, wherein the embedded code comprises at least one of:
an identifier of the object (120),
a link to the object information, and
the object information.
6. The method (800) of any one of the preceding claims, wherein the object information is indicative of at least one of price information, material information, size information, availability information, stock information, review information, object score information and popularity information.
7. A lighting system for providing information about an object (120), the lighting system comprising:
a controller for receiving object information about the object (120), for determining a first light setting and a second light setting based on the object information, wherein the difference between the first light setting and the second light setting is perceivable for a user, for controlling a first light source (410) according to the first light setting and for controlling a second light source (412) according to the second light setting, a mobile device (102) comprising:
a camera for capturing an image (104) of the first light source (410) and the second light source (412),
a display,
a light detector for detecting light comprising an embedded code, which embedded code is representative of the object information,
a processor arranged for retrieving the code from the light, for retrieving the object information, for determining a first light source position of the first light source (410) and a second light source position of the second light source (412) in the image (104), for determining a position for the object information on the display, wherein the position is related to at least one of the first light source position and the second light source position in the image (104), for rendering the image (104) on the display, and for rendering the object information as an overlay on the image (104) at the position on the display.
8. The lighting system of claim 7, wherein the first light source (410) and the second light source (412) are comprised in a linear lighting device. 9. The lighting system of claim 8, wherein the linear lighting device is comprised in a supporting member arranged for supporting the object (120).
10. The lighting system of claim 8 or 9, wherein the controller is arranged for controlling three or more light sources, and wherein the controller is arranged for determining
which light sources to control according to the first light setting and which light sources to control according to the second light setting based on the object information.
11. The lighting system of claim 10, wherein the first light source (410) and/or the second light source (412) is arranged for illuminating the object (120).
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP16195906.9 | 2016-10-27 | ||
EP16195906 | 2016-10-27 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2018077648A1 true WO2018077648A1 (en) | 2018-05-03 |
Family
ID=57211330
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2017/076327 WO2018077648A1 (en) | 2016-10-27 | 2017-10-16 | A method of providing information about an object |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2018077648A1 (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8606645B1 (en) * | 2012-02-02 | 2013-12-10 | SeeMore Interactive, Inc. | Method, medium, and system for an augmented reality retail application |
US20140100997A1 (en) | 2012-10-05 | 2014-04-10 | Jochen Mayerle | Augmented-reality shopping using a networked mobile device |
WO2014170290A1 (en) | 2013-04-19 | 2014-10-23 | Koninklijke Philips N.V. | A coded light device, and a product information system comprising such a coded light device |
EP2940897A1 (en) * | 2012-12-27 | 2015-11-04 | Panasonic Intellectual Property Corporation of America | Information communication method |
-
2017
- 2017-10-16 WO PCT/EP2017/076327 patent/WO2018077648A1/en active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8606645B1 (en) * | 2012-02-02 | 2013-12-10 | SeeMore Interactive, Inc. | Method, medium, and system for an augmented reality retail application |
US20140100997A1 (en) | 2012-10-05 | 2014-04-10 | Jochen Mayerle | Augmented-reality shopping using a networked mobile device |
EP2940897A1 (en) * | 2012-12-27 | 2015-11-04 | Panasonic Intellectual Property Corporation of America | Information communication method |
WO2014170290A1 (en) | 2013-04-19 | 2014-10-23 | Koninklijke Philips N.V. | A coded light device, and a product information system comprising such a coded light device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11367130B2 (en) | Method for in-store object highlighting by a real world user interface | |
US11609607B2 (en) | Evolving docking based on detected keyboard positions | |
US10586391B2 (en) | Interactive virtual reality platforms | |
US10026116B2 (en) | Methods and devices for smart shopping | |
US10380677B2 (en) | System of sharing an augmented environment with a companion | |
US20110141011A1 (en) | Method of performing a gaze-based interaction between a user and an interactive display system | |
US20180033045A1 (en) | Method and system for personalized advertising | |
US10360613B2 (en) | System and method for monitoring display unit compliance | |
KR20200079721A (en) | Smart clothes management system and method thereof | |
EP2993638A1 (en) | Electronic garment pattern distribution | |
KR20180000007A (en) | Shoping inducing system using augmented reality | |
US20200059603A1 (en) | A method of providing information about an object | |
US10902501B2 (en) | Method of storing object identifiers | |
KR101431804B1 (en) | Apparatus for displaying show window image using transparent display, method for displaying show window image using transparent display and recording medium thereof | |
US20170300927A1 (en) | System and method for monitoring display unit compliance | |
WO2018077648A1 (en) | A method of providing information about an object | |
JP2017146939A (en) | Image display device, display control method and display control program | |
KR20180000006A (en) | Shoping inducing mehtod using augmented reality | |
JP2017146941A (en) | Image display device, display control method and display control program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 17791316 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 17791316 Country of ref document: EP Kind code of ref document: A1 |