US20220129211A1 - Converting voice command into text code blocks that support printing services - Google Patents
Converting voice command into text code blocks that support printing services Download PDFInfo
- Publication number
- US20220129211A1 US20220129211A1 US17/571,983 US202217571983A US2022129211A1 US 20220129211 A1 US20220129211 A1 US 20220129211A1 US 202217571983 A US202217571983 A US 202217571983A US 2022129211 A1 US2022129211 A1 US 2022129211A1
- Authority
- US
- United States
- Prior art keywords
- text code
- image forming
- code blocks
- forming apparatus
- voice command
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000007639 printing Methods 0.000 title claims abstract description 87
- 238000000034 method Methods 0.000 claims description 33
- 238000004891 communication Methods 0.000 claims description 15
- 230000006870 function Effects 0.000 claims description 14
- 238000013507 mapping Methods 0.000 claims description 9
- 238000010586 diagram Methods 0.000 description 22
- 230000004044 response Effects 0.000 description 10
- 230000014509 gene expression Effects 0.000 description 7
- 230000008569 process Effects 0.000 description 7
- 238000012545 processing Methods 0.000 description 5
- 230000008859 change Effects 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/32—User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/60—Protecting data
- G06F21/606—Protecting data by securing the transmission between two devices or processes
- G06F21/608—Secure printing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/12—Digital output to print unit, e.g. line printer, chain printer
- G06F3/1201—Dedicated interfaces to print systems
- G06F3/1202—Dedicated interfaces to print systems specifically adapted to achieve a particular effect
- G06F3/1203—Improving or facilitating administration, e.g. print management
- G06F3/1205—Improving or facilitating administration, e.g. print management resulting in increased flexibility in print job configuration, e.g. job settings, print requirements, job tickets
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/12—Digital output to print unit, e.g. line printer, chain printer
- G06F3/1201—Dedicated interfaces to print systems
- G06F3/1202—Dedicated interfaces to print systems specifically adapted to achieve a particular effect
- G06F3/1222—Increasing security of the print job
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/12—Digital output to print unit, e.g. line printer, chain printer
- G06F3/1201—Dedicated interfaces to print systems
- G06F3/1223—Dedicated interfaces to print systems specifically adapted to use a particular technique
- G06F3/1236—Connection management
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/12—Digital output to print unit, e.g. line printer, chain printer
- G06F3/1201—Dedicated interfaces to print systems
- G06F3/1223—Dedicated interfaces to print systems specifically adapted to use a particular technique
- G06F3/1275—Print workflow management, e.g. defining or changing a workflow, cross publishing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/12—Digital output to print unit, e.g. line printer, chain printer
- G06F3/1201—Dedicated interfaces to print systems
- G06F3/1278—Dedicated interfaces to print systems specifically adapted to adopt a particular infrastructure
- G06F3/1285—Remote printer device, e.g. being remote from client or server
- G06F3/1289—Remote printer device, e.g. being remote from client or server in server-client-printer device configuration, e.g. the server does not see the printer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/12—Digital output to print unit, e.g. line printer, chain printer
- G06F3/1201—Dedicated interfaces to print systems
- G06F3/1278—Dedicated interfaces to print systems specifically adapted to adopt a particular infrastructure
- G06F3/1292—Mobile client, e.g. wireless printing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/22—Interactive procedures; Man-machine interfaces
- G10L17/24—Interactive procedures; Man-machine interfaces the user being prompted to utter a password or a predefined phrase
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/12—Digital output to print unit, e.g. line printer, chain printer
- G06F3/1201—Dedicated interfaces to print systems
- G06F3/1202—Dedicated interfaces to print systems specifically adapted to achieve a particular effect
- G06F3/1218—Reducing or saving of used resources, e.g. avoiding waste of consumables or improving usage of hardware resources
- G06F3/122—Reducing or saving of used resources, e.g. avoiding waste of consumables or improving usage of hardware resources with regard to computing resources, e.g. memory, CPU
Definitions
- a user may remotely control an image forming apparatus, such as a printer, a copier, a facsimile machine, a multifunction apparatus, and the like, through a user terminal without having to directly manipulate the image forming apparatus.
- an image forming apparatus such as a printer, a copier, a facsimile machine, a multifunction apparatus, and the like.
- Such remote control increases the convenience for the user of the image forming apparatus.
- various devices may be connected to each other using wired/wireless communication technology, and complex operations may be processed through cloud computing.
- FIG. 1 is a diagram for explaining a printing service environment including an image forming apparatus, a user terminal, and a cloud server, according to an example;
- FIG. 2 is a diagram for explaining a configuration and an operation of each of an image forming apparatus, a user terminal, and a cloud server, according to an example;
- FIG. 3 is a diagram for explaining a process of executing a printing job using a voice command, according to an example
- FIG. 4 is a diagram for explaining an image forming apparatus and a user terminal performing user authentication using voice passwords, according to an example
- FIG. 5 is a diagram for explaining an instructions mapping table specialized for a printing service used when a cloud server converts a voice command into text code blocks, according to an example
- FIG. 6 is a diagram for explaining an example in which a cloud server converts a voice command into text code blocks based on a language habit of a user, according to an example
- FIG. 7 is a diagram for explaining text code blocks converted from voice commands and supporting a printing service, according to an example
- FIG. 8 is a diagram illustrating a user terminal providing a menu corresponding to a selected text code block in response to a user manipulation selecting the text code block, according to an example
- FIG. 9 is a diagram illustrating a user terminal changing an order of text code blocks in response to a user manipulation moving a text code block, according to an example
- FIG. 10 is a diagram illustrating a user terminal merging sets of text code blocks, according to an example.
- FIG. 11 is a diagram for explaining an example in which a user interface device of a user terminal and a user interface device of an image forming apparatus interact with each other and the user interface devices simultaneously process multiple manipulations of a user, according to an example.
- an element when an element is “connected” to another element, the elements may not only be “directly connected,” but may also be “electrically connected” or otherwise “indirectly connected” via another element there between. Also, when a region “includes” an element, the region may further include another element instead of excluding the other element, unless otherwise differently stated.
- FIG. 1 is a diagram for explaining a printing service environment including an image forming apparatus, a user terminal, and a cloud server, according to an example.
- an image forming apparatus 100 may perform image forming operations such as printing, scanning, copying, or faxing.
- the image forming apparatus 100 may form an image on a recording medium such as a printing paper by using various printing methods such as an electrophotographic method, an inkjet method, a thermal transfer method, a direct thermal method, or the like.
- a user terminal 200 may be a computing device such as a smart phone, a notebook, or the like.
- a cloud server 300 may support a printing service and provide a response to a request when requested by the image forming apparatus 100 or the user terminal 200 .
- FIG. 1 a schematic sequence in which the user terminal 200 controls the image forming apparatus 100 to connect the image forming apparatus 100 to the user terminal 200 to perform a printing job is shown in sequences ⁇ circle around ( 1 ) ⁇ , ⁇ circle around ( 2 ) ⁇ , ⁇ circle around ( 3 ) ⁇ , and ⁇ circle around ( 4 ) ⁇ .
- the user terminal 200 may access the image forming apparatus 100 supporting the printing service and execute a printing application supporting a voice command installed in the user terminal 200 . Since the printing application supporting the voice command is executed, the user terminal 200 may scan an advertising signal repeatedly transmitted from the neighboring image forming apparatus 100 and obtain device information of the image forming apparatus 100 .
- the advertising signal may be a Bluetooth low energy (BLE) beacon signal, a wireless fidelity (Wi-Fi) aware signal, or the like.
- the user terminal 200 may request a connection of the image forming apparatus 100 using the device information of the image forming apparatus 100 through communication between the user terminal 200 and the cloud server 300 and confirm that the printing service using the voice command is supported.
- the cloud server 300 may inform the image forming apparatus 100 of a connection request of the user terminal 200 and establish a connection between voice command sessions of the image forming apparatus 100 and the cloud server 300 .
- the cloud server 300 may receive a voice command from the user terminal 200 and transmit a request for executing a printing job related to the received voice command to the image forming apparatus 100 through voice recognition and natural language processing.
- FIG. 2 is a diagram for explaining a configuration and an operation of each of an image forming apparatus, a user terminal, and a cloud server, according to an example.
- the image forming apparatus 100 may include a user interface device 110 , a controller 120 , a communicator 130 , a memory 140 , and an image forming operator 150 . Further, although not shown, the image forming apparatus 100 may include a power supply for supplying power to the respective components.
- the user interface device 110 may include an input unit for receiving an input for performing an image forming operation from a user and an output unit for displaying information about a result of the image forming operation or a state of the image forming apparatus 100 .
- the user interface device 110 may include an operation panel for receiving a user input and a display panel for displaying a screen.
- the input may include devices capable of receiving various types of user input such as a keyboard, a physical button, a touch screen, a camera, a microphone, or the like.
- the output unit may include, for example, a display panel, a speaker, or the like.
- the present disclosure is not limited to this, and the user interface device 110 may include various input/output supporting devices.
- the controller 120 may control an operation of the image forming apparatus 100 and may include at least one processor such as a central processing unit (CPU).
- the controller 120 may control other components included in the image forming apparatus 100 to perform an operation corresponding to the user input received through the user interface device 110 .
- the controller 120 may include at least one specialized processor corresponding to each function, or may be a single integrated processor.
- the controller 120 may execute a program stored in the memory 140 , read data or a file stored in the memory 140 , or store a new file in the memory 140 .
- the communicator 130 may perform wired/wireless communication with another device or a network.
- the communicator 130 may include a communication module, such as a transceiver, supporting at least one of various wired/wireless communication methods.
- the communication module may be in the form of a chipset, or may be a sticker/barcode (e.g., a sticker including a near-field communication (NFC) tag) including information necessary for communication.
- NFC near-field communication
- Wireless communication may include, for example, at least one of Wi-Fi, Wi-Fi Direct, Bluetooth, BLE, ultra-wide band (UWB), NFC, or the like.
- Wired communication may include, for example, at least one of universal serial bus (USB), High-Definition Multimedia Interface (HDMI), or the like.
- USB universal serial bus
- HDMI High-Definition Multimedia Interface
- the communicator 130 may be connected to an external device located outside the image forming apparatus 100 and may transmit and receive signals or data.
- the communicator 130 may transmit signals or data received from the user terminal 200 or the cloud server 300 to the controller 120 or transmit signals or data generated by the controller 120 to the user terminal 200 or the cloud server 300 .
- the communicator 130 may be directly connected to the cloud server 300 and may transmit and receive signals or data.
- the communicator 130 may be connected to the user terminal 200 through the cloud server 300 . That is, the communicator 130 of the image forming apparatus 100 may transmit/receive signals or data to/from a communicator 230 of the user terminal 200 through the cloud server 300 .
- the memory 140 may install and store programs such as applications and various types of data such as files.
- the controller 120 may access the data stored in the memory 140 and use the data or store new data in the memory 140 . Further, the controller 120 may execute the programs installed in the memory 140 , and may install an application received from outside through the communicator 130 in the memory 140 .
- the image forming operator 150 may perform an image forming operation such as printing, scanning, copying, or faxing.
- the image forming operator 150 is shown as including a printing unit 151 , a scanning unit 152 , and a fax unit 153 .
- the image forming operator 150 may include only some of these components or may further include components for performing other kinds of image forming operations.
- the printing unit 151 may form an image on a recording medium by using various printing methods such as an electrophotographic method, an inkjet method, a thermal transfer method, a direct thermal method, or the like.
- the scanning unit 152 may irradiate light onto a document, receive the reflected light, and read an image recorded in the document.
- a charge-coupled device (CCD) or a contact image sensor (CIS) may be employed as an image sensor for reading the image from the document.
- the fax unit 153 may share a configuration for scanning an image with the scanning unit 152 , may share a configuration for printing a received file with the printing unit 151 , may transmit a scan file to a destination, or may receive a file from outside.
- the user terminal 200 may include a processor 210 , a user interface device 220 , the communicator 230 , and a memory 240 .
- the processor 210 may control an operation of the user terminal 200 and may include at least one processing unit such as a CPU.
- the processor 210 may control other components included in the user terminal 200 to perform operations corresponding to user inputs received via the user interface device 220 .
- the processor 210 may include at least one specialized processing unit corresponding to each function.
- the processor 210 may execute a program stored in the memory 240 , read data or files stored in the memory 240 , or store a new file in the memory 240 .
- the user interface device 220 may include an input unit for obtaining a user input and an output unit for providing information to the user.
- the input unit may be an operation panel capable of receiving a user manipulation or a microphone capable of acquiring a voice or noise of the user.
- the output unit may be a display panel for displaying a screen or a speaker.
- the user terminal 200 may be connected to an external device such as the image forming apparatus 100 or the cloud server 300 through the communicator 230 .
- the communicator 230 may include a transceiver and support at least one of Wi-Fi, Wi-Fi Direct, Bluetooth, BLE, UWB, 3G, 4G, and 5G.
- the memory 240 may store instructions executable by the processor 210 .
- the memory 240 may install and store programs such as applications and various types of data such as files.
- the memory 240 may store an operating system.
- the cloud server 300 may include a memory 310 , a processor 320 , and a communicator 330 .
- the memory 310 may store instructions executable by the processor 320 .
- the processor 320 may control an operation of the cloud server 300 and may include at least one processing unit.
- the communicator 330 may include a communication module, such as a transceiver, that may communicate with the communicator 130 of the image forming apparatus 100 or the communicator 230 of the user terminal 200 .
- the image forming apparatus 100 , the user terminal 200 , and the cloud server 300 described above may be used to perform a printing operation through a printing service supporting voice commands.
- the processor 210 of the user terminal 200 may execute commands stored in the memory 240 to transmit voice commands obtained from the user to the cloud server 300 providing the printing service.
- the processor 320 of the cloud server 300 may control the communicator 330 to execute the commands stored in the memory 310 , convert the voice command received from the user terminal 200 through the communicator 330 into text code blocks supporting the printing service, and transmit the text code blocks to the user terminal 200 .
- the processor 320 of the cloud server 300 may control the communicator 330 to search for text code blocks corresponding to a command that is similar to a command of the text code blocks in terms of pronunciation or function and transmit the retrieved text code blocks to the user terminal 200 .
- the processor 210 of the user terminal 200 may control the communicator 230 to receive the text code blocks supporting the printing service from the cloud server 300 and may control the user interface device 220 to display the text code blocks on the user interface device 220 .
- the printing job execution request may be transmitted to the image forming apparatus 100 through the cloud server 300 , and the printing operation may be performed.
- FIG. 3 a process of performing a printing job using a voice command will be described with reference to FIG. 3 .
- FIG. 3 is a diagram for explaining a process of executing a printing job using a voice command, according to an example.
- the cloud server 300 may receive device information about the image forming apparatus 100 from the image forming apparatus 100 in operation S 305 .
- the cloud server 300 may register the device information of the image forming apparatus 100 by storing the device information about the image forming apparatus 100 in operation S 310 .
- a user who wishes to execute the printing job may access the image forming apparatus 100 with the user terminal 200 in operation S 315 .
- the image forming apparatus 100 may repeatedly transmit an advertising signal including the device information on a space corresponding to a predetermined range in operation S 320 .
- the advertising signal may be a BLE beacon signal or a Wi-Fi aware signal.
- the user terminal 200 may receive the advertising signal repeatedly transmitted from the image forming apparatus 100 and use the device information about the image forming apparatus 100 to use a printing service.
- each of the user terminal 200 and the image forming apparatus 100 may obtain sound information generated in the image forming apparatus 100 or a keyword phrase spoken by the user as a voice password in operations S 325 and S 330 .
- the predetermined space may be a region within a range that is reached by the advertising signal repeatedly transmitted by the image forming apparatus 100 and may be a space in which each of the image forming apparatus 100 and the user terminal 200 collect the voice password together. Since the user terminal 200 and the image forming apparatus 100 collect sound generated in the predetermined space together, each of the user terminal 200 and the image forming apparatus 100 may collect similar sound as the voice password.
- the user terminal 200 and the image forming apparatus 100 may transmit the device information and the voice password to the cloud server 300 for user authentication in operations S 335 and S 340 .
- the cloud server 300 may confirm whether the voice passwords respectively received from the image forming apparatus 100 and the user terminal 200 match with each other, perform user authentication, and establish interconnection through the cloud server 300 in operation S 345 .
- the image forming apparatus 100 and the user terminal 200 may obtain information necessary for establishing a connection between the image forming apparatus 100 and the user terminal 200 from the cloud server 300 .
- a security method using both a network for receiving the advertising signal repeatedly transmitted by the image forming apparatus 100 and the voice password based on sound generated in the predetermined space for collecting the voice password together is applied, logical security and physical security may be performed together.
- FIG. 4 is a diagram for explaining an image forming apparatus and a user terminal performing user authentication using voice passwords, according to an example.
- each of the user terminal 200 and the image forming apparatus 100 may obtain sound information (in the case of ⁇ circle around ( 1 ) ⁇ ) generated in the image forming apparatus 100 or a keyword phrase (in the case of ⁇ circle around ( 2 ) ⁇ ) spoken by a user as the voice password and transmit the voice password to the cloud server 300 .
- the cloud server 300 may confirm whether the voice passwords respectively received from the image forming apparatus 100 and the user terminal 200 match with each other, perform user authentication, and establish interconnection through the cloud server 300 .
- the cloud server 300 may determine whether the voice passwords match based on unique features identified in the voice passwords. As an example, the cloud server 300 may confirm whether there is a match between the two voice passwords through a convolution operation on the received voice passwords to confirm whether the voice passwords match. When the voice passwords match, it may be proved that the user terminal 200 and the image forming apparatus 100 are in the same space in the physical sense.
- the user terminal 200 may obtain the voice command from the user who wishes to perform the printing job in operation S 350 .
- the user terminal 200 may transmit the voice command obtained from the user to the cloud server 300 providing the printing service in operation S 355 .
- the cloud server 300 may convert the voice command received from the user terminal 200 into text code blocks supporting the printing service in operation S 360 .
- the cloud server 300 may analyze a relationship between the received voice command and a previous voice command to determine a context of the printing job according to a request of the user terminal 200 and convert the received voice command into the text code blocks based on the context.
- the cloud server 300 may accurately determine a type of the printing job or a current situation that the user wishes to execute through the continuous voice commands.
- the cloud server 300 may generate the text code blocks corresponding to the voice command based on an instructions mapping table specialized for the printing service for one voice command or a language habit of the user.
- FIG. 5 is a diagram for explaining an instructions mapping table specialized for a printing service used when a cloud server converts a voice command into text code blocks, according to an example.
- the cloud server 300 may convert the received voice command into the text code blocks based on the instructions mapping table specialized for the printing service.
- the instructions mapping table specialized for a printing service is shown, but the instructions mapping table is not limited thereto.
- the cloud server 300 interprets the received voice command and what functions the cloud server 300 generates the text code blocks to perform. For example, when the voice command is “upload,” the cloud server 300 may interpret the voice command as a “scan” function and generate the text code blocks to scan a print document and then upload the scanned print document to a predetermined space. Also, the cloud server 300 may accurately identify features of the document for voice commands of similar format such as “latest document,” “recent document,” and “final document” and determine the document on which the printing job is to be performed.
- the cloud server 300 may limitedly interpret the expressions as being related to the printing job and reflect the expressions to generate the text code blocks.
- the expression “small” may be interpreted to mean lowering of a voice, but in the instructions mapping table specialized for the printing service, the cloud server 300 may interpret the expression as meaning to reduce output of the printing job and generate the text code blocks.
- FIG. 6 is a diagram for explaining an example in which a cloud server converts a voice command into text code blocks based on a language habit of a user, according to an example.
- the cloud server 300 may convert the received voice command into the text code blocks based on a learned language habit of a user.
- the cloud server 300 may learn the language habit of the user through machine learning based on a language learning model and accurately determine the intended or desired meaning of the voice command intended by the user based on the learned language habit.
- a value of a reduction option may be determined based on the language habit of the existing user with respect to the value of the reduction option.
- there may be derived commands such as “slightly small (10-20%),” “moderately small (30%),” and “very small (50%)” with respect to the expression “small.”
- the cloud server 300 may determine “small” spoken by a user A as a command to reduce “moderately small (30%)” based on the language habit of the user and a history of past printing jobs and accordingly generate the text code blocks.
- the cloud server 300 may transmit the text code blocks to the user terminal 200 in operation S 365 .
- the cloud server 300 may search for not only the text code blocks corresponding to the voice command but also text code blocks corresponding to a command that is similar to a command of the text code blocks in terms of pronunciation or function and transmit the found text code blocks to the user terminal 200 .
- the user terminal 200 may display the text code blocks supporting the printing service received from the cloud server 300 on the user interface device 220 and receive a user manipulation for the text code blocks and a request for execution of the printing job in operation S 370 . Examples of the text code blocks and the user manipulation with respect to the text code blocks will be described with reference to FIGS. 7 to 10 .
- FIG. 7 is a diagram for explaining text code blocks converted from voice commands and supporting a printing service, according to an example.
- the text code blocks converted from each voice command by the cloud server 300 are displayed on the user interface device 220 .
- a set of the text code blocks corresponding to each voice command may be displayed separately by speech bubbles as shown on the user interface device 220 in FIG. 7 .
- Each unit voice corresponding to a function of an image forming apparatus constituting the voice command and a printing job option may be separately displayed as a unit code block. For example, as shown in FIG.
- the voice command “Scan and print 2 copy” may be divided into functions of the image forming apparatus such as “Scan” and “and print” and the printing job option such as “ 2 copy” and may be displayed as a total of three unit code blocks.
- the unit code blocks may be displayed in different colors.
- the text code blocks may be stored in the user terminal 200 according to categories such as the user, time, etc., such that the user may confirm history information of the text code blocks executed in the user terminal 200 and reuse the text code blocks.
- the user may edit the text code blocks through the user interface device 220 of the user terminal 200 to compensate for incompleteness of the voice command or lack of conversion into the text code blocks.
- FIG. 8 is a diagram illustrating a user terminal providing a menu corresponding to a selected text code block in response to a user manipulation selecting the text code block, according to an example.
- the user interface device 220 of the user terminal 200 may display a menu including at least one text code block corresponding to a command that is similar to a command of a selected text code block in terms of pronunciation or function below the selected text code block in response to the user manipulation selecting one of text code blocks.
- a menu including text code blocks “Fax” and “Download” corresponding to a functionally similar command that may be performed in place of “Scan” may be displayed below the selected text code block.
- FIG. 9 is a diagram illustrating a user terminal changing an order of text code blocks in response to a user manipulation moving a text code block, according to an example.
- the user interface device 220 of the user terminal 200 may change and display the order of the text code blocks according to a position of the moved text code block in response to the user manipulation of moving one of the text code blocks.
- FIG. 9 it may be seen that when a user selects a text code block “and Print” from among the text code blocks and drags and drops the text code block “and Print” behind a text code block “and Email,” the order of the text code blocks may be changed such that text code blocks “Scan and Print and Email” may be changed to text code blocks “Scan and Email and Print.”
- the order of the text code blocks may be changed as shown in FIG. 9 .
- FIG. 10 is a diagram illustrating a user terminal merging sets of text code blocks, according to an example.
- the user interface device 220 of the user terminal 200 may display overlapped sets as one set of text code blocks in response to a user manipulation of overlapping any one of the sets of text code blocks corresponding to each of a plurality of voice commands with the other set.
- FIG. 10 it may be seen that when a user performs a user manipulation selecting a set of text code blocks “Scan and Print 2 Copy” and merging the set of text code blocks “Scan and Print 2 Copy” with a set of text code blocks “Send to Marketer By email,” one set of text code blocks “Scan and Print 2 Copy Send to Marketer By email” may be generated and displayed.
- FIG. 10 it may be seen that when a user performs a user manipulation selecting a set of text code blocks “Scan and Print 2 Copy” and merging the set of text code blocks “Scan and Print 2 Copy” with a set of text code blocks “Send to Marketer By email,” one set of text code blocks “Scan and Print 2 Copy Send to Marketer By email” may be generated and displayed.
- the two sets of text code blocks may be merged into one set through the user manipulation touching a part of a speech bubble icon corresponding to the set of text code blocks “Scan and Print 2 Copy” and linking parts of a speech bubble icon corresponding to the set of text code blocks “Send to Marketer By email.”
- the user manipulation merging the sets of text code blocks may use various methods, and thus the present disclosure is not limited thereto.
- the user terminal 200 may transmit the request for execution of the printing job to the cloud server 300 in operation S 375 .
- the replaceable information may be replaced with the corresponding information stored in the user terminal 200 at the request of execution of the printing job and transmitted to the image forming apparatus 100 through the cloud server 300 .
- the request for execution of the printing job may be transmitted after replacing “Marketer” with the corresponding email address information.
- the cloud server 300 may transmit the request for execution of the printing job to the image forming apparatus 100 corresponding to the printing job according to the request for execution of the printing job received from the user terminal 200 in operation S 380 .
- the image forming apparatus 100 may execute the printing job in response to the request for execution of the printing job in operation S 385 .
- the user when the user attempts to start the printing service supported by the cloud server 300 using the user terminal 200 or requests execution of the printing job from the image forming apparatus 100 , there may be cases in which user multi-manipulation is required. For example, there may be a case in which the user wishes to view content displayed on the user interface device 220 and perform a user manipulation on the displayed text code blocks while instructing a microphone of the user terminal 200 with a voice command. In such a case, it may be difficult for the user to simultaneously process the user multi-manipulation with one user terminal 200 .
- FIG. 11 is a diagram for explaining an example in which a user interface device of a user terminal and a user interface device of an image forming apparatus interact with each other and the user interface devices simultaneously process multiple manipulations of a user, according to an example.
- the user interface device 220 of the user terminal 200 may interact with the user interface device 110 of the image forming apparatus 100 performing a printing job according to a request for execution by the user terminal 200 in real time and simultaneously process the user multi-manipulation for progress of the printing job with the user interface device 110 of the image forming apparatus 100 .
- the user interface device 220 of the user terminal 200 may interact with the user interface device 110 of the image forming apparatus 100 and provide the same screen in real time. In this case, the user may make a voice command using the user terminal 200 and simultaneously edit text code blocks displayed on the user interface device 110 of the image forming apparatus 100 .
- the real time interaction between the user interface device 220 of the user terminal 200 and the user interface device 110 of the image forming apparatus 100 may be performed in a wireless local area network between the user terminal 200 and the image forming apparatus 100 and printing data for the printing job may be transmitted to the image forming apparatus 100 through the cloud server 300 .
- immediate feedback may be transmitted to the user terminal 200 and the image forming apparatuses 100 through a communication using the wireless local area network between the user terminal 200 and the image forming apparatuses 100 without passing through the cloud server 300 .
- the above description may be implemented in the form of a non-transitory computer-readable storage medium storing instructions or data executable by a computer or a processor.
- the above description may be implemented in a general-purpose digital computer, which may be embodied as a program executable on a computer and executes such a program using the non-transitory computer-readable storage medium.
- Such non-transitory computer-readable storage mediums may include read only memory (ROM), random access memory (RAM), flash memory, CD-ROMs, CD-Rs, CD+Rs, CD-DVD-Rs, DVD-Rs, DVD-RWs, DVD+RWs, DVD-RAMs, BD-ROMs, BD-Rs, BD-R LTHs, BD-REs, magnetic tape, floppy disk, an optical magnetic data storage device, an optical data storage device, hard disk, solid-state disk (SSD), and any device capable of storing instructions or software, associated data, data files, and data structures and providing instructions or software, associated data, data files, and data structures to the processor or the computer such that the processor or the computer may execute instructions.
- ROM read only memory
- RAM random access memory
- flash memory CD-ROMs, CD-Rs, CD+Rs, CD-DVD-Rs, DVD-Rs, DVD-RWs, DVD+RWs, DVD-RAMs, BD-ROMs, BD
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Computer Security & Cryptography (AREA)
- Multimedia (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Health & Medical Sciences (AREA)
- Acoustics & Sound (AREA)
- Computer Networks & Wireless Communication (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- Computational Linguistics (AREA)
- Bioethics (AREA)
- Facsimiles In General (AREA)
- Accessory Devices And Overall Control Thereof (AREA)
Abstract
Description
- A user may remotely control an image forming apparatus, such as a printer, a copier, a facsimile machine, a multifunction apparatus, and the like, through a user terminal without having to directly manipulate the image forming apparatus. Such remote control increases the convenience for the user of the image forming apparatus.
- To provide a printing service, various devices may be connected to each other using wired/wireless communication technology, and complex operations may be processed through cloud computing.
- The above and other aspects, features, and advantages of certain examples of the present disclosure will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
-
FIG. 1 is a diagram for explaining a printing service environment including an image forming apparatus, a user terminal, and a cloud server, according to an example; -
FIG. 2 is a diagram for explaining a configuration and an operation of each of an image forming apparatus, a user terminal, and a cloud server, according to an example; -
FIG. 3 is a diagram for explaining a process of executing a printing job using a voice command, according to an example; -
FIG. 4 is a diagram for explaining an image forming apparatus and a user terminal performing user authentication using voice passwords, according to an example; -
FIG. 5 is a diagram for explaining an instructions mapping table specialized for a printing service used when a cloud server converts a voice command into text code blocks, according to an example; -
FIG. 6 is a diagram for explaining an example in which a cloud server converts a voice command into text code blocks based on a language habit of a user, according to an example; -
FIG. 7 is a diagram for explaining text code blocks converted from voice commands and supporting a printing service, according to an example; -
FIG. 8 is a diagram illustrating a user terminal providing a menu corresponding to a selected text code block in response to a user manipulation selecting the text code block, according to an example; -
FIG. 9 is a diagram illustrating a user terminal changing an order of text code blocks in response to a user manipulation moving a text code block, according to an example; -
FIG. 10 is a diagram illustrating a user terminal merging sets of text code blocks, according to an example; and -
FIG. 11 is a diagram for explaining an example in which a user interface device of a user terminal and a user interface device of an image forming apparatus interact with each other and the user interface devices simultaneously process multiple manipulations of a user, according to an example. - Throughout the drawings, it should be noted that like reference numbers are used to depict the same or similar elements, features, parts, components, and structures.
- Hereinafter, various examples will be described with reference to the accompanying drawings. The examples described may be modified in many different forms. To more clearly describe features of examples, matters well known to one of ordinary skill in the art to which the below examples pertain will not be described in detail. In the following description and drawings, components having substantially the same function and configuration are denoted by the same reference numerals, and redundant description thereof will be omitted.
- In the specification, when an element is “connected” to another element, the elements may not only be “directly connected,” but may also be “electrically connected” or otherwise “indirectly connected” via another element there between. Also, when a region “includes” an element, the region may further include another element instead of excluding the other element, unless otherwise differently stated.
-
FIG. 1 is a diagram for explaining a printing service environment including an image forming apparatus, a user terminal, and a cloud server, according to an example. - Referring to
FIG. 1 , animage forming apparatus 100 may perform image forming operations such as printing, scanning, copying, or faxing. Theimage forming apparatus 100 may form an image on a recording medium such as a printing paper by using various printing methods such as an electrophotographic method, an inkjet method, a thermal transfer method, a direct thermal method, or the like. Auser terminal 200 may be a computing device such as a smart phone, a notebook, or the like. Acloud server 300 may support a printing service and provide a response to a request when requested by theimage forming apparatus 100 or theuser terminal 200. - In
FIG. 1 , a schematic sequence in which theuser terminal 200 controls theimage forming apparatus 100 to connect theimage forming apparatus 100 to theuser terminal 200 to perform a printing job is shown in sequences {circle around (1)}, {circle around (2)}, {circle around (3)}, and {circle around (4)}. - In the first sequence {circle around (1)}, the
user terminal 200 may access theimage forming apparatus 100 supporting the printing service and execute a printing application supporting a voice command installed in theuser terminal 200. Since the printing application supporting the voice command is executed, theuser terminal 200 may scan an advertising signal repeatedly transmitted from the neighboringimage forming apparatus 100 and obtain device information of theimage forming apparatus 100. For example, the advertising signal may be a Bluetooth low energy (BLE) beacon signal, a wireless fidelity (Wi-Fi) aware signal, or the like. - In the second sequence {circle around (2)}, the
user terminal 200 may request a connection of theimage forming apparatus 100 using the device information of theimage forming apparatus 100 through communication between theuser terminal 200 and thecloud server 300 and confirm that the printing service using the voice command is supported. - In the third sequence {circle around (3)}, the
cloud server 300 may inform theimage forming apparatus 100 of a connection request of theuser terminal 200 and establish a connection between voice command sessions of theimage forming apparatus 100 and thecloud server 300. - In the fourth sequence {circle around (4)}, the
cloud server 300 may receive a voice command from theuser terminal 200 and transmit a request for executing a printing job related to the received voice command to theimage forming apparatus 100 through voice recognition and natural language processing. - Examples of operations of the
image forming apparatus 100, theuser terminal 200, and thecloud server 300 in a printing service environment supporting the voice command will now be described in more detail with reference toFIGS. 2 to 11 . -
FIG. 2 is a diagram for explaining a configuration and an operation of each of an image forming apparatus, a user terminal, and a cloud server, according to an example. - Referring to
FIG. 2 , theimage forming apparatus 100 may include auser interface device 110, acontroller 120, acommunicator 130, amemory 140, and animage forming operator 150. Further, although not shown, theimage forming apparatus 100 may include a power supply for supplying power to the respective components. - The
user interface device 110 may include an input unit for receiving an input for performing an image forming operation from a user and an output unit for displaying information about a result of the image forming operation or a state of theimage forming apparatus 100. For example, theuser interface device 110 may include an operation panel for receiving a user input and a display panel for displaying a screen. - As an example, the input may include devices capable of receiving various types of user input such as a keyboard, a physical button, a touch screen, a camera, a microphone, or the like. Further, the output unit may include, for example, a display panel, a speaker, or the like. However, the present disclosure is not limited to this, and the
user interface device 110 may include various input/output supporting devices. - The
controller 120 may control an operation of theimage forming apparatus 100 and may include at least one processor such as a central processing unit (CPU). Thecontroller 120 may control other components included in theimage forming apparatus 100 to perform an operation corresponding to the user input received through theuser interface device 110. Thecontroller 120 may include at least one specialized processor corresponding to each function, or may be a single integrated processor. - For example, the
controller 120 may execute a program stored in thememory 140, read data or a file stored in thememory 140, or store a new file in thememory 140. - The
communicator 130 may perform wired/wireless communication with another device or a network. To this end, thecommunicator 130 may include a communication module, such as a transceiver, supporting at least one of various wired/wireless communication methods. For example, the communication module may be in the form of a chipset, or may be a sticker/barcode (e.g., a sticker including a near-field communication (NFC) tag) including information necessary for communication. - Wireless communication may include, for example, at least one of Wi-Fi, Wi-Fi Direct, Bluetooth, BLE, ultra-wide band (UWB), NFC, or the like. Wired communication may include, for example, at least one of universal serial bus (USB), High-Definition Multimedia Interface (HDMI), or the like.
- The
communicator 130 may be connected to an external device located outside theimage forming apparatus 100 and may transmit and receive signals or data. Thecommunicator 130 may transmit signals or data received from theuser terminal 200 or thecloud server 300 to thecontroller 120 or transmit signals or data generated by thecontroller 120 to theuser terminal 200 or thecloud server 300. Thecommunicator 130 may be directly connected to thecloud server 300 and may transmit and receive signals or data. Also, thecommunicator 130 may be connected to theuser terminal 200 through thecloud server 300. That is, thecommunicator 130 of theimage forming apparatus 100 may transmit/receive signals or data to/from acommunicator 230 of theuser terminal 200 through thecloud server 300. - The
memory 140 may install and store programs such as applications and various types of data such as files. Thecontroller 120 may access the data stored in thememory 140 and use the data or store new data in thememory 140. Further, thecontroller 120 may execute the programs installed in thememory 140, and may install an application received from outside through thecommunicator 130 in thememory 140. - The
image forming operator 150 may perform an image forming operation such as printing, scanning, copying, or faxing. InFIG. 2 , theimage forming operator 150 is shown as including aprinting unit 151, ascanning unit 152, and afax unit 153. However, theimage forming operator 150 may include only some of these components or may further include components for performing other kinds of image forming operations. - The
printing unit 151 may form an image on a recording medium by using various printing methods such as an electrophotographic method, an inkjet method, a thermal transfer method, a direct thermal method, or the like. - The
scanning unit 152 may irradiate light onto a document, receive the reflected light, and read an image recorded in the document. For example, a charge-coupled device (CCD) or a contact image sensor (CIS) may be employed as an image sensor for reading the image from the document. - The
fax unit 153 may share a configuration for scanning an image with thescanning unit 152, may share a configuration for printing a received file with theprinting unit 151, may transmit a scan file to a destination, or may receive a file from outside. - The
user terminal 200 may include aprocessor 210, auser interface device 220, thecommunicator 230, and amemory 240. - The
processor 210 may control an operation of theuser terminal 200 and may include at least one processing unit such as a CPU. Theprocessor 210 may control other components included in theuser terminal 200 to perform operations corresponding to user inputs received via theuser interface device 220. Theprocessor 210 may include at least one specialized processing unit corresponding to each function. Theprocessor 210 may execute a program stored in thememory 240, read data or files stored in thememory 240, or store a new file in thememory 240. - The
user interface device 220 may include an input unit for obtaining a user input and an output unit for providing information to the user. For example, the input unit may be an operation panel capable of receiving a user manipulation or a microphone capable of acquiring a voice or noise of the user. The output unit may be a display panel for displaying a screen or a speaker. - The
user terminal 200 may be connected to an external device such as theimage forming apparatus 100 or thecloud server 300 through thecommunicator 230. To this end, thecommunicator 230 may include a transceiver and support at least one of Wi-Fi, Wi-Fi Direct, Bluetooth, BLE, UWB, 3G, 4G, and 5G. - The
memory 240 may store instructions executable by theprocessor 210. Thememory 240 may install and store programs such as applications and various types of data such as files. Thememory 240 may store an operating system. - The
cloud server 300 may include amemory 310, aprocessor 320, and acommunicator 330. Thememory 310 may store instructions executable by theprocessor 320. Theprocessor 320 may control an operation of thecloud server 300 and may include at least one processing unit. Thecommunicator 330 may include a communication module, such as a transceiver, that may communicate with thecommunicator 130 of theimage forming apparatus 100 or thecommunicator 230 of theuser terminal 200. - The
image forming apparatus 100, theuser terminal 200, and thecloud server 300 described above may be used to perform a printing operation through a printing service supporting voice commands. For example, theprocessor 210 of theuser terminal 200 may execute commands stored in thememory 240 to transmit voice commands obtained from the user to thecloud server 300 providing the printing service. Theprocessor 320 of thecloud server 300 may control thecommunicator 330 to execute the commands stored in thememory 310, convert the voice command received from theuser terminal 200 through thecommunicator 330 into text code blocks supporting the printing service, and transmit the text code blocks to theuser terminal 200. Theprocessor 320 of thecloud server 300 may control thecommunicator 330 to search for text code blocks corresponding to a command that is similar to a command of the text code blocks in terms of pronunciation or function and transmit the retrieved text code blocks to theuser terminal 200. Theprocessor 210 of theuser terminal 200 may control thecommunicator 230 to receive the text code blocks supporting the printing service from thecloud server 300 and may control theuser interface device 220 to display the text code blocks on theuser interface device 220. In response to a printing job execution request of theuser terminal 200 based on the voice command, the printing job execution request may be transmitted to theimage forming apparatus 100 through thecloud server 300, and the printing operation may be performed. Hereinafter, a process of performing a printing job using a voice command will be described with reference toFIG. 3 . -
FIG. 3 is a diagram for explaining a process of executing a printing job using a voice command, according to an example. - Referring to
FIG. 3 , thecloud server 300 may receive device information about theimage forming apparatus 100 from theimage forming apparatus 100 in operation S305. - The
cloud server 300 may register the device information of theimage forming apparatus 100 by storing the device information about theimage forming apparatus 100 in operation S310. - A user who wishes to execute the printing job may access the
image forming apparatus 100 with theuser terminal 200 in operation S315. - The
image forming apparatus 100 may repeatedly transmit an advertising signal including the device information on a space corresponding to a predetermined range in operation S320. At this time, the advertising signal may be a BLE beacon signal or a Wi-Fi aware signal. Theuser terminal 200 may receive the advertising signal repeatedly transmitted from theimage forming apparatus 100 and use the device information about theimage forming apparatus 100 to use a printing service. - When the
user terminal 200 and theimage forming apparatus 100 are located in a predetermined space, each of theuser terminal 200 and theimage forming apparatus 100 may obtain sound information generated in theimage forming apparatus 100 or a keyword phrase spoken by the user as a voice password in operations S325 and S330. As an example, the predetermined space may be a region within a range that is reached by the advertising signal repeatedly transmitted by theimage forming apparatus 100 and may be a space in which each of theimage forming apparatus 100 and theuser terminal 200 collect the voice password together. Since theuser terminal 200 and theimage forming apparatus 100 collect sound generated in the predetermined space together, each of theuser terminal 200 and theimage forming apparatus 100 may collect similar sound as the voice password. - The
user terminal 200 and theimage forming apparatus 100 may transmit the device information and the voice password to thecloud server 300 for user authentication in operations S335 and S340. - The
cloud server 300 may confirm whether the voice passwords respectively received from theimage forming apparatus 100 and theuser terminal 200 match with each other, perform user authentication, and establish interconnection through thecloud server 300 in operation S345. Theimage forming apparatus 100 and theuser terminal 200 may obtain information necessary for establishing a connection between theimage forming apparatus 100 and theuser terminal 200 from thecloud server 300. When a security method using both a network for receiving the advertising signal repeatedly transmitted by theimage forming apparatus 100 and the voice password based on sound generated in the predetermined space for collecting the voice password together is applied, logical security and physical security may be performed together. -
FIG. 4 is a diagram for explaining an image forming apparatus and a user terminal performing user authentication using voice passwords, according to an example. - Referring to
FIG. 4 , a security method using voice passwords according to operations S325, S330, S335, S340, and S345 described inFIG. 3 is shown. When theuser terminal 200 and theimage forming apparatus 100 are located in a predetermined space, each of theuser terminal 200 and theimage forming apparatus 100 may obtain sound information (in the case of {circle around (1)}) generated in theimage forming apparatus 100 or a keyword phrase (in the case of {circle around (2)}) spoken by a user as the voice password and transmit the voice password to thecloud server 300. Thecloud server 300 may confirm whether the voice passwords respectively received from theimage forming apparatus 100 and theuser terminal 200 match with each other, perform user authentication, and establish interconnection through thecloud server 300. Thecloud server 300 may determine whether the voice passwords match based on unique features identified in the voice passwords. As an example, thecloud server 300 may confirm whether there is a match between the two voice passwords through a convolution operation on the received voice passwords to confirm whether the voice passwords match. When the voice passwords match, it may be proved that theuser terminal 200 and theimage forming apparatus 100 are in the same space in the physical sense. - Referring again to
FIG. 3 , theuser terminal 200 may obtain the voice command from the user who wishes to perform the printing job in operation S350. - The
user terminal 200 may transmit the voice command obtained from the user to thecloud server 300 providing the printing service in operation S355. - The
cloud server 300 may convert the voice command received from theuser terminal 200 into text code blocks supporting the printing service in operation S360. Thecloud server 300 may analyze a relationship between the received voice command and a previous voice command to determine a context of the printing job according to a request of theuser terminal 200 and convert the received voice command into the text code blocks based on the context. Thecloud server 300 may accurately determine a type of the printing job or a current situation that the user wishes to execute through the continuous voice commands. - The
cloud server 300 may generate the text code blocks corresponding to the voice command based on an instructions mapping table specialized for the printing service for one voice command or a language habit of the user. -
FIG. 5 is a diagram for explaining an instructions mapping table specialized for a printing service used when a cloud server converts a voice command into text code blocks, according to an example. - Referring to
FIG. 5 , thecloud server 300 may convert the received voice command into the text code blocks based on the instructions mapping table specialized for the printing service. InFIG. 5 , an example of an instructions mapping table specialized for a printing service is shown, but the instructions mapping table is not limited thereto. - In the example of
FIG. 5 , it may be known in what sense thecloud server 300 interprets the received voice command and what functions thecloud server 300 generates the text code blocks to perform. For example, when the voice command is “upload,” thecloud server 300 may interpret the voice command as a “scan” function and generate the text code blocks to scan a print document and then upload the scanned print document to a predetermined space. Also, thecloud server 300 may accurately identify features of the document for voice commands of similar format such as “latest document,” “recent document,” and “final document” and determine the document on which the printing job is to be performed. Also, when the voice command has expressions such as “small,” “large,” “clear,” and “easy to read,” thecloud server 300 may limitedly interpret the expressions as being related to the printing job and reflect the expressions to generate the text code blocks. For example, the expression “small” may be interpreted to mean lowering of a voice, but in the instructions mapping table specialized for the printing service, thecloud server 300 may interpret the expression as meaning to reduce output of the printing job and generate the text code blocks. -
FIG. 6 is a diagram for explaining an example in which a cloud server converts a voice command into text code blocks based on a language habit of a user, according to an example. - Referring to
FIG. 6 , thecloud server 300 may convert the received voice command into the text code blocks based on a learned language habit of a user. Thecloud server 300 may learn the language habit of the user through machine learning based on a language learning model and accurately determine the intended or desired meaning of the voice command intended by the user based on the learned language habit. - In the example of
FIG. 6 , when the voice command includes an expression “small,” a value of a reduction option may be determined based on the language habit of the existing user with respect to the value of the reduction option. In other words, there may be derived commands such as “slightly small (10-20%),” “moderately small (30%),” and “very small (50%)” with respect to the expression “small.” Thecloud server 300 may determine “small” spoken by a user A as a command to reduce “moderately small (30%)” based on the language habit of the user and a history of past printing jobs and accordingly generate the text code blocks. - Referring again to
FIG. 3 , thecloud server 300 may transmit the text code blocks to theuser terminal 200 in operation S365. Thecloud server 300 may search for not only the text code blocks corresponding to the voice command but also text code blocks corresponding to a command that is similar to a command of the text code blocks in terms of pronunciation or function and transmit the found text code blocks to theuser terminal 200. - The
user terminal 200 may display the text code blocks supporting the printing service received from thecloud server 300 on theuser interface device 220 and receive a user manipulation for the text code blocks and a request for execution of the printing job in operation S370. Examples of the text code blocks and the user manipulation with respect to the text code blocks will be described with reference toFIGS. 7 to 10 . -
FIG. 7 is a diagram for explaining text code blocks converted from voice commands and supporting a printing service, according to an example. - Referring to
FIG. 7 , with respect to voice commands “Scan andprint 2 copy” and “Send to marketer by email” of a user, the text code blocks converted from each voice command by thecloud server 300 are displayed on theuser interface device 220. A set of the text code blocks corresponding to each voice command may be displayed separately by speech bubbles as shown on theuser interface device 220 inFIG. 7 . Each unit voice corresponding to a function of an image forming apparatus constituting the voice command and a printing job option may be separately displayed as a unit code block. For example, as shown inFIG. 7 , the voice command “Scan andprint 2 copy” may be divided into functions of the image forming apparatus such as “Scan” and “and print” and the printing job option such as “2 copy” and may be displayed as a total of three unit code blocks. The unit code blocks may be displayed in different colors. - The text code blocks may be stored in the
user terminal 200 according to categories such as the user, time, etc., such that the user may confirm history information of the text code blocks executed in theuser terminal 200 and reuse the text code blocks. - The user may edit the text code blocks through the
user interface device 220 of theuser terminal 200 to compensate for incompleteness of the voice command or lack of conversion into the text code blocks. -
FIG. 8 is a diagram illustrating a user terminal providing a menu corresponding to a selected text code block in response to a user manipulation selecting the text code block, according to an example. - Referring to
FIG. 8 , theuser interface device 220 of theuser terminal 200 may display a menu including at least one text code block corresponding to a command that is similar to a command of a selected text code block in terms of pronunciation or function below the selected text code block in response to the user manipulation selecting one of text code blocks. InFIG. 8 , it may be seen that when a user selects a text code block “Scan” from among the text code blocks, a menu including text code blocks “Fax” and “Download” corresponding to a functionally similar command that may be performed in place of “Scan” may be displayed below the selected text code block. -
FIG. 9 is a diagram illustrating a user terminal changing an order of text code blocks in response to a user manipulation moving a text code block, according to an example. - Referring to
FIG. 9 , theuser interface device 220 of theuser terminal 200 may change and display the order of the text code blocks according to a position of the moved text code block in response to the user manipulation of moving one of the text code blocks. InFIG. 9 , it may be seen that when a user selects a text code block “and Print” from among the text code blocks and drags and drops the text code block “and Print” behind a text code block “and Email,” the order of the text code blocks may be changed such that text code blocks “Scan and Print and Email” may be changed to text code blocks “Scan and Email and Print.” When the user makes a voice command and determines that it is necessary to change the order that a function requiring a short time is firstly processed in consideration of time required for processing each function, the order of the text code blocks may be changed as shown inFIG. 9 . -
FIG. 10 is a diagram illustrating a user terminal merging sets of text code blocks, according to an example. - Referring to
FIG. 10 , theuser interface device 220 of theuser terminal 200 may display overlapped sets as one set of text code blocks in response to a user manipulation of overlapping any one of the sets of text code blocks corresponding to each of a plurality of voice commands with the other set. InFIG. 10 , it may be seen that when a user performs a user manipulation selecting a set of text code blocks “Scan andPrint 2 Copy” and merging the set of text code blocks “Scan andPrint 2 Copy” with a set of text code blocks “Send to Marketer By email,” one set of text code blocks “Scan andPrint 2 Copy Send to Marketer By email” may be generated and displayed. InFIG. 10 , it may be seen that the two sets of text code blocks may be merged into one set through the user manipulation touching a part of a speech bubble icon corresponding to the set of text code blocks “Scan andPrint 2 Copy” and linking parts of a speech bubble icon corresponding to the set of text code blocks “Send to Marketer By email.” However, the user manipulation merging the sets of text code blocks may use various methods, and thus the present disclosure is not limited thereto. - Referring again to
FIG. 3 , theuser terminal 200 may transmit the request for execution of the printing job to thecloud server 300 in operation S375. At this time, when replaceable information is included in the text code blocks and corresponds to information stored in theuser terminal 200, the replaceable information may be replaced with the corresponding information stored in theuser terminal 200 at the request of execution of the printing job and transmitted to theimage forming apparatus 100 through thecloud server 300. For example, when there are text code blocks “Send to Marketer By email,” and corresponding email address information of “Marketer” is stored in theuser terminal 200, the request for execution of the printing job may be transmitted after replacing “Marketer” with the corresponding email address information. - The
cloud server 300 may transmit the request for execution of the printing job to theimage forming apparatus 100 corresponding to the printing job according to the request for execution of the printing job received from theuser terminal 200 in operation S380. - The
image forming apparatus 100 may execute the printing job in response to the request for execution of the printing job in operation S385. - In an example, when the user attempts to start the printing service supported by the
cloud server 300 using theuser terminal 200 or requests execution of the printing job from theimage forming apparatus 100, there may be cases in which user multi-manipulation is required. For example, there may be a case in which the user wishes to view content displayed on theuser interface device 220 and perform a user manipulation on the displayed text code blocks while instructing a microphone of theuser terminal 200 with a voice command. In such a case, it may be difficult for the user to simultaneously process the user multi-manipulation with oneuser terminal 200. -
FIG. 11 is a diagram for explaining an example in which a user interface device of a user terminal and a user interface device of an image forming apparatus interact with each other and the user interface devices simultaneously process multiple manipulations of a user, according to an example. - Referring to
FIG. 11 , theuser interface device 220 of theuser terminal 200 may interact with theuser interface device 110 of theimage forming apparatus 100 performing a printing job according to a request for execution by theuser terminal 200 in real time and simultaneously process the user multi-manipulation for progress of the printing job with theuser interface device 110 of theimage forming apparatus 100. As shown inFIG. 11 , theuser interface device 220 of theuser terminal 200 may interact with theuser interface device 110 of theimage forming apparatus 100 and provide the same screen in real time. In this case, the user may make a voice command using theuser terminal 200 and simultaneously edit text code blocks displayed on theuser interface device 110 of theimage forming apparatus 100. To this end, the real time interaction between theuser interface device 220 of theuser terminal 200 and theuser interface device 110 of theimage forming apparatus 100 may be performed in a wireless local area network between theuser terminal 200 and theimage forming apparatus 100 and printing data for the printing job may be transmitted to theimage forming apparatus 100 through thecloud server 300. Thus, immediate feedback may be transmitted to theuser terminal 200 and theimage forming apparatuses 100 through a communication using the wireless local area network between theuser terminal 200 and theimage forming apparatuses 100 without passing through thecloud server 300. - The above description may be implemented in the form of a non-transitory computer-readable storage medium storing instructions or data executable by a computer or a processor. The above description may be implemented in a general-purpose digital computer, which may be embodied as a program executable on a computer and executes such a program using the non-transitory computer-readable storage medium. Such non-transitory computer-readable storage mediums may include read only memory (ROM), random access memory (RAM), flash memory, CD-ROMs, CD-Rs, CD+Rs, CD-DVD-Rs, DVD-Rs, DVD-RWs, DVD+RWs, DVD-RAMs, BD-ROMs, BD-Rs, BD-R LTHs, BD-REs, magnetic tape, floppy disk, an optical magnetic data storage device, an optical data storage device, hard disk, solid-state disk (SSD), and any device capable of storing instructions or software, associated data, data files, and data structures and providing instructions or software, associated data, data files, and data structures to the processor or the computer such that the processor or the computer may execute instructions.
- While one or more examples have been described with reference to the figures, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit and scope as defined by the following claims.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/571,983 US20220129211A1 (en) | 2018-05-31 | 2022-01-10 | Converting voice command into text code blocks that support printing services |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020180063014A KR20190136832A (en) | 2018-05-31 | 2018-05-31 | Converting voice command into text code blcoks that support printing services |
KR10-2018-0063014 | 2018-05-31 | ||
PCT/KR2018/011914 WO2019231055A1 (en) | 2018-05-31 | 2018-10-11 | Converting voice command into text code blocks that support printing services |
US202017049718A | 2020-10-22 | 2020-10-22 | |
US17/571,983 US20220129211A1 (en) | 2018-05-31 | 2022-01-10 | Converting voice command into text code blocks that support printing services |
Related Parent Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/049,718 Continuation US11249696B2 (en) | 2018-05-31 | 2018-10-11 | Converting voice command into text code blocks that support printing services |
PCT/KR2018/011914 Continuation WO2019231055A1 (en) | 2018-05-31 | 2018-10-11 | Converting voice command into text code blocks that support printing services |
Publications (1)
Publication Number | Publication Date |
---|---|
US20220129211A1 true US20220129211A1 (en) | 2022-04-28 |
Family
ID=68697645
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/049,718 Active US11249696B2 (en) | 2018-05-31 | 2018-10-11 | Converting voice command into text code blocks that support printing services |
US17/571,983 Abandoned US20220129211A1 (en) | 2018-05-31 | 2022-01-10 | Converting voice command into text code blocks that support printing services |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/049,718 Active US11249696B2 (en) | 2018-05-31 | 2018-10-11 | Converting voice command into text code blocks that support printing services |
Country Status (4)
Country | Link |
---|---|
US (2) | US11249696B2 (en) |
EP (1) | EP3718104A4 (en) |
KR (1) | KR20190136832A (en) |
WO (1) | WO2019231055A1 (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3910466A1 (en) * | 2020-04-27 | 2021-11-17 | Ricoh Company, Ltd. | Information processing apparatus, information processing system, method for processing information, and carrier means |
CN112286745B (en) * | 2020-10-27 | 2022-11-15 | 上海商米科技集团股份有限公司 | Automatic testing method for cloud printer |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4466572B2 (en) | 2006-01-16 | 2010-05-26 | コニカミノルタビジネステクノロジーズ株式会社 | Image forming apparatus, voice command execution program, and voice command execution method |
KR101631496B1 (en) * | 2008-06-03 | 2016-06-17 | 삼성전자주식회사 | Robot apparatus and method for registrating contracted commander thereof |
WO2012116110A1 (en) | 2011-02-22 | 2012-08-30 | Speak With Me, Inc. | Hybridized client-server speech recognition |
JP5234160B2 (en) * | 2011-03-23 | 2013-07-10 | 株式会社デンソー | Vehicle apparatus and information display system |
US9329808B2 (en) * | 2011-03-24 | 2016-05-03 | Ricoh Company, Ltd. | User interfaces for rule-based workflow generation in a print shop environment |
JP5831251B2 (en) | 2012-01-25 | 2015-12-09 | コニカミノルタ株式会社 | Image forming system, image forming apparatus, terminal device, control program for image forming apparatus, and control program for terminal device |
KR101383552B1 (en) | 2013-02-25 | 2014-04-10 | 미디어젠(주) | Speech recognition method of sentence having multiple instruction |
JP6171511B2 (en) | 2013-04-09 | 2017-08-02 | コニカミノルタ株式会社 | Control device, image forming apparatus, portable terminal device, control method, and control program |
EP2797303A1 (en) * | 2013-04-26 | 2014-10-29 | Océ-Technologies B.V. | Apparatus having a number of options for operating and/or configuring the apparatus and method and computer program product therefor |
WO2015184186A1 (en) * | 2014-05-30 | 2015-12-03 | Apple Inc. | Multi-command single utterance input method |
WO2016017978A1 (en) * | 2014-07-31 | 2016-02-04 | Samsung Electronics Co., Ltd. | Device and method for performing functions |
US20160150124A1 (en) | 2014-11-24 | 2016-05-26 | Kyocera Document Solutions Inc. | Image Forming Apparatus with User Identification Capabilities |
US9959129B2 (en) * | 2015-01-09 | 2018-05-01 | Microsoft Technology Licensing, Llc | Headless task completion within digital personal assistants |
JP2016168707A (en) | 2015-03-12 | 2016-09-23 | コニカミノルタ株式会社 | Image formation device and program |
WO2017039648A1 (en) * | 2015-09-02 | 2017-03-09 | Hewlett-Packard Development Company, L.P. | Voice control of a printer unit |
KR20170070649A (en) | 2015-12-14 | 2017-06-22 | 에스프린팅솔루션 주식회사 | Image forming apparuatus, cloud server, image forming system, and method for setting of connection with image forming apparatus |
JP6696246B2 (en) * | 2016-03-16 | 2020-05-20 | 富士ゼロックス株式会社 | Image processing device and program |
US10388282B2 (en) * | 2017-01-25 | 2019-08-20 | CliniCloud Inc. | Medical voice command device |
CN108656558A (en) * | 2017-03-29 | 2018-10-16 | 三纬国际立体列印科技股份有限公司 | Apply to the speech control system and sound control method of 3D printer |
JP7003720B2 (en) * | 2018-02-19 | 2022-01-21 | コニカミノルタ株式会社 | Image processing equipment, image processing equipment control methods, and programs |
-
2018
- 2018-05-31 KR KR1020180063014A patent/KR20190136832A/en unknown
- 2018-10-11 WO PCT/KR2018/011914 patent/WO2019231055A1/en unknown
- 2018-10-11 EP EP18921038.8A patent/EP3718104A4/en not_active Withdrawn
- 2018-10-11 US US17/049,718 patent/US11249696B2/en active Active
-
2022
- 2022-01-10 US US17/571,983 patent/US20220129211A1/en not_active Abandoned
Also Published As
Publication number | Publication date |
---|---|
US20210240403A1 (en) | 2021-08-05 |
EP3718104A4 (en) | 2021-11-10 |
WO2019231055A1 (en) | 2019-12-05 |
EP3718104A1 (en) | 2020-10-07 |
US11249696B2 (en) | 2022-02-15 |
KR20190136832A (en) | 2019-12-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7487807B2 (en) | Information processing system, server, information processing method, and information processing program | |
JP7230506B2 (en) | Information processing system, information processing device, information processing method, and information processing program | |
JP7234612B2 (en) | Information processing system and information processing method | |
JP7471379B2 (en) | Voice operation system and voice operation method | |
JP6942995B2 (en) | Information processing program, information processing device, and control method of information processing device | |
US20220129211A1 (en) | Converting voice command into text code blocks that support printing services | |
US20170054866A1 (en) | Mobile apparatus, image scan apparatus and method for processing a job | |
US11625200B2 (en) | Information processing device, mobile terminal, and non-transitory computer readable medium for parameter presentation | |
US10911611B2 (en) | Transmitting fax document by using cloud server | |
US11190652B2 (en) | Electronic apparatus for providing user interface of image forming apparatus | |
KR20220050640A (en) | Providing Shortened URL and Information related Contents corresponding to Original URL | |
US10440206B2 (en) | Fax function diagnosis method, and apparatus for performing the method | |
CN110312050A (en) | Print system, Method of printing, image forming apparatus and its control method and medium | |
JP2016096393A (en) | Information processing apparatus, system, information processing method, and program | |
US11523024B2 (en) | Embedding user specific information into user specific information input area of document | |
JP2021113899A (en) | Information processing system, information processing method, and program | |
JP2016177387A (en) | Information processing apparatus, system, information processing method, and program | |
JP2021144645A (en) | Instruction transmitter, instruction transfer system, instruction transmission method, and program | |
US11755257B2 (en) | Automatic output of document | |
US11563864B2 (en) | Information processing apparatus and non-transitory computer-readable storage medium | |
JP7388079B2 (en) | Information transmitting device, information transmitting method and program | |
US20220210291A1 (en) | Processing of image forming job based on metadata tag input | |
JP2022167157A (en) | Image processing device, image processing method, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P., TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HP PRINTING KOREA CO., LTD.;REEL/FRAME:058605/0063 Effective date: 20190620 Owner name: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P., TEXAS Free format text: CONFIRMATORY ASSIGNMENT;ASSIGNOR:HP PRINTING KOREA CO., LTD.;REEL/FRAME:058710/0507 Effective date: 20210218 Owner name: HP PRINTING KOREA CO., LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:PARK, SUNG JIN;REEL/FRAME:058604/0904 Effective date: 20190319 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |