Nothing Special   »   [go: up one dir, main page]

CN111559317B - Agent device, method for controlling agent device, and storage medium - Google Patents

Agent device, method for controlling agent device, and storage medium Download PDF

Info

Publication number
CN111559317B
CN111559317B CN202010085731.6A CN202010085731A CN111559317B CN 111559317 B CN111559317 B CN 111559317B CN 202010085731 A CN202010085731 A CN 202010085731A CN 111559317 B CN111559317 B CN 111559317B
Authority
CN
China
Prior art keywords
agent
sound
vehicle
unit
agent function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010085731.6A
Other languages
Chinese (zh)
Other versions
CN111559317A (en
Inventor
安原真也
大井裕介
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Honda Motor Co Ltd
Original Assignee
Honda Motor Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Honda Motor Co Ltd filed Critical Honda Motor Co Ltd
Publication of CN111559317A publication Critical patent/CN111559317A/en
Application granted granted Critical
Publication of CN111559317B publication Critical patent/CN111559317B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R11/00Arrangements for holding or mounting articles, not otherwise provided for
    • B60R11/02Arrangements for holding or mounting articles, not otherwise provided for for radio sets, television sets, telephones, or the like; Arrangement of controls thereof
    • B60R11/0217Arrangements for holding or mounting articles, not otherwise provided for for radio sets, television sets, telephones, or the like; Arrangement of controls thereof for loud-speakers
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R11/00Arrangements for holding or mounting articles, not otherwise provided for
    • B60R11/02Arrangements for holding or mounting articles, not otherwise provided for for radio sets, television sets, telephones, or the like; Arrangement of controls thereof
    • B60R11/0258Arrangements for holding or mounting articles, not otherwise provided for for radio sets, television sets, telephones, or the like; Arrangement of controls thereof for navigation systems
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R16/00Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for
    • B60R16/02Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for electric constitutive elements
    • B60R16/037Electric or fluid circuits specially adapted for vehicles and not otherwise provided for; Arrangement of elements of electric or fluid circuits specially adapted for vehicles and not otherwise provided for electric constitutive elements for occupant comfort, e.g. for automatic adjustment of appliances according to personal settings, e.g. seats, mirrors, steering wheel
    • B60R16/0373Voice control

Landscapes

  • Engineering & Computer Science (AREA)
  • Mechanical Engineering (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Stereophonic System (AREA)
  • Fittings On The Vehicle Exterior For Carrying Loads, And Devices For Holding Or Mounting Articles (AREA)
  • Circuit For Audible Band Transducer (AREA)

Abstract

Provided are an agent device capable of providing an agent function with a more natural sense of use, a method for controlling the agent device, and a storage medium. The agent device is provided with: a plurality of agent function sections that provide services including responses by sound in accordance with speech of occupants of the vehicle; and a sound control unit that sets positions for positioning the sound image of the sound for each of the plurality of agent function units, and controls the output of each of the plurality of speakers so that the sound image is positioned at the set positions.

Description

Agent device, method for controlling agent device, and storage medium
Technical Field
The invention relates to an agent device, a control method of the agent device, and a storage medium.
Background
Conventionally, a technology related to an agent function that provides information related to driving assistance, control of a vehicle, other applications, and the like in response to a request of an occupant while performing a dialogue with the occupant of the vehicle has been disclosed (for example, refer to patent document 1).
[ Prior Art literature ]
[ patent literature ]
[ patent document 1 ]
Japanese patent laid-open No. 2006-335231
Disclosure of Invention
Problems to be solved by the invention
In recent years, a plurality of intelligent agent functions are being put into practical use for mounting on a vehicle, but the output form of sound in this case is not sufficiently studied. Therefore, in the conventional technique, there is a case where a natural feeling of use cannot be provided to the occupant.
The present invention has been made in view of such circumstances, and an object thereof is to provide an agent device capable of providing an agent function with a more natural sense of use, a control method of the agent device, and a storage medium.
Means for solving the problems
The following configuration is adopted for the agent device, the control method of the agent device, and the storage medium of the present invention.
(1): an intelligent agent device according to an aspect of the present invention includes: a plurality of agent function sections that provide services including responses by sound in accordance with speech of occupants of the vehicle; and a sound control unit that sets positions for positioning the sound image of the sound for each of the plurality of agent function units, and controls the output of each of the plurality of speakers so that the sound image is positioned at the set positions.
(2): in the aspect of (1) above, the sound control unit positions a sound image based on sound generated by an agent function unit having a high frequency of use among the plurality of agent function units at a position where a nearest seat among seats of the vehicle is a secondary driver.
(3): in the aspect of (1) above, the plurality of agent function portions include a vehicle agent function portion to which a right to control the vehicle equipment is given, and the sound control portion positions a sound image based on a sound generated by the vehicle agent function portion at a position where a nearest seat among seats of the vehicle is a secondary driver seat.
(4): in the aspect of (1) above, the plurality of agent function portions include a vehicle agent function portion to which a right to control the vehicle equipment is given, and the sound control portion positions a sound image based on a sound generated by the vehicle agent function portion at a position where a nearest seat among seats of the vehicle is a driver seat.
(5): in the aspect of (1) above, the sound control unit positions a sound image based on sound generated by an agent function unit having a low frequency of use among the plurality of agent function units at a position where a nearest seat among seats of the vehicle is a rear seat.
(6): in the aspect of (1) above, the sound control unit sets the positions for locating the sound image of the sound for each of the plurality of agent function units based on the recognition result of the seating position recognition unit that recognizes the seating position of the occupant.
(7): in the aspect of (6) above, the sound control unit sets a position for locating a sound image based on the sound generated by each of the plurality of agent function units in a ceiling of the vehicle compartment when the seating position identification unit identifies that the occupant is seated in each seat.
(8): in the aspect of (1) above, the agent device further includes an activation agent selecting unit that activates an agent function unit that, with reference to at least a detection result of a face orientation detecting unit that detects a face orientation of the driver, positions an acoustic image at a position where the driver faces the face.
(9): another aspect of the present invention provides a method for controlling an agent device, comprising causing a computer to: enabling any one of the plurality of agent functions to be started; providing a service including a response by sound according to the speech of the occupant of the vehicle as the function of the activated agent function section; and setting positions for positioning the sound image of the sound for the plurality of agent function units, respectively, and controlling the output of the plurality of speakers so that the sound image is positioned at the set positions.
(10): a storage medium of another aspect of the present invention stores a program that causes a computer to execute: enabling any one of the plurality of agent functions to be started; providing a service including a response by sound according to the speech of the occupant of the vehicle as the function of the activated agent function section; and setting positions for positioning the sound image of the sound for the plurality of agent function units, respectively, and controlling the output of the plurality of speakers so that the sound image is positioned at the set positions.
Effects of the invention
According to the aspects (1) to (10), the agent function can be provided with a more natural sense of use.
Drawings
Fig. 1 is a block diagram of an agent system 1 including an agent device 100.
Fig. 2 is a diagram showing the structure of the agent apparatus 100 according to the first embodiment and the equipment mounted on the vehicle M.
Fig. 3 is a diagram showing a configuration example of the display-operation apparatus 20.
Fig. 4 is a diagram showing a configuration example of the speaker unit 30.
Fig. 5 is a diagram for explaining the principle of position determination for sound image localization.
Fig. 6 is a diagram showing the structure of the agent server 200 and a part of the structure of the agent device 100.
Fig. 7 is a diagram showing a first example of the sound image localization position.
Fig. 8 is a diagram showing a second example of sound image localization positions.
Fig. 9 is a diagram showing a third example of sound image localization positions.
Fig. 10 is a diagram showing a fourth example of sound image localization positions.
Fig. 11 is a diagram showing a fifth example of sound image localization positions.
Fig. 12 is a diagram showing a sixth example of sound image localization positions.
Fig. 13 is a diagram showing the configuration of an agent apparatus 100A according to the second embodiment and equipment mounted on a vehicle M.
Fig. 14 is a diagram for explaining an operation of the activation agent selecting unit 120.
Detailed Description
Embodiments of an agent device, a method for controlling an agent device, and a storage medium according to the present invention will be described below with reference to the accompanying drawings. An agent device is a device that implements a portion or all of an agent system. Hereinafter, as an example of the smart device, a smart device mounted on a vehicle (hereinafter referred to as a vehicle M) and having a plurality of types of smart functions will be described. The smart function is, for example, the following: while talking to the occupant of the vehicle M, various information is provided or a web service is interposed based on a request (instruction) included in the speech of the occupant. The functions, processing steps, control, and output forms and contents of the plurality of types of agents may be different from each other. The agent function may have a function of controlling devices in the vehicle (for example, devices related to driving control and vehicle body control).
The agent function is realized by, for example, a sound recognition function (a function of converting a sound into a text) for recognizing a sound of an occupant, a natural language processing function (a function of understanding the structure and meaning of a text), a dialogue management function, a network search function of searching other devices via a network or searching a predetermined database held by the device itself, and the like. Some or all of these functions may be implemented using AT (Artificial Intelligence) techniques. A part of the configuration for performing these functions (in particular, the voice recognition function and the natural language processing interpretation function) may be mounted on an agent server (external device) that can communicate with the in-vehicle communication device of the vehicle M or the general-purpose communication device that is carried into the vehicle M. In the following description, a part of the configuration is mounted on the agent server, and the agent device and the agent server cooperate to realize the agent system. A service providing entity (service entity) which virtually appears by cooperating an agent device with an agent server is called an agent.
< integral Structure >
Fig. 1 is a block diagram of an agent system 1 including an agent device 100. The agent system 1 includes, for example, the agent device 100 and a plurality of agent servers 200-1, 200-2, 200-3, …. The hyphen at the end of the reference numeral is set below as an identifier for distinguishing the agent. In the case where it is not discriminated which one is, it is sometimes simply referred to as an agent server 200. In fig. 1, 3 agent servers 200 are shown, but the number of agent servers 200 may be 2 or 4 or more. Each agent server 200 is operated by a provider of an agent system different from each other. Accordingly, the agent in the present invention is an agent realized by mutually different providers. Examples of the provider include an automobile manufacturer, a network service provider, an electronic commerce and trade provider, a seller of a mobile terminal, a manufacturer, and the like, and any subject (legal, group, individual, and the like) can be the provider of the intelligent agent system.
The agent device 100 communicates with the agent server 200 via the network NW. The network NW includes, for example, a part or all of the internet, a cellular network, a Wi-Fi network, WAN (Wide Area Network), LAN (Local Area Network), a public line, a telephone line, a wireless base station, and the like. Various web servers 300 are connected to the network NW, and the agent server 200 or the agent device 100 can acquire web pages from the various web servers 300 via the network NW.
The agent device 100 communicates with the occupant of the vehicle M, transmits the sound from the occupant to the agent server 200, and presents the answer obtained from the agent server 200 to the occupant in the form of sound output and image display.
< first embodiment >
[ vehicle ]
Fig. 2 is a diagram showing the structure of the agent apparatus 100 according to the first embodiment and the equipment mounted on the vehicle M. The vehicle M is mounted with one or more microphones 10, a display-operation device 20, a speaker unit 30, a navigation device 40, a vehicle device 50, an in-vehicle communication device 60, an occupant recognition device 80, and an agent device 100, for example. A general-purpose communication device 70 such as a smart phone may be brought into the vehicle interior and used as a communication device. These devices are connected to each other via a multi-way communication line such as CAN (Controller Area Network) communication line, a serial communication line, a wireless communication network, or the like. The configuration shown in fig. 2 is merely an example, and a part of the configuration may be omitted or another configuration may be added.
The microphone 10 is a sound receiving portion that collects sound emitted from the vehicle interior. The display-operation device 20 is a device (or a group of devices) that displays an image and is capable of accepting an input operation. The display-operation device 20 includes, for example, a display device configured as a touch panel. The display-operating device 20 may also include HUD (Head Up Display), mechanical input devices. The speaker unit 30 includes, for example, a plurality of speakers (sound output units) disposed at different positions in the vehicle interior. The display-operating device 20 may also be shared among the agent device 100 and the navigation device 40. Details thereof will be described later.
The navigation device 40 includes navigation HMI (Human machine Interface), GPS (Global Positioning System) and the like measuring devices, a storage device storing map information, and a control device (navigation controller) for performing route search and the like. Some or all of the microphone 10, the display-operating device 20, and the speaker unit 30 may be used as the navigation HMI. The navigation device 40 searches for a route (navigation route) for moving from the position of the vehicle M determined by the position measurement device to a destination input by the occupant, and outputs guidance information using the navigation HMI so that the vehicle M can travel along the route. The path search function may also be in a navigation server accessible via the network NW. In this case, the navigation device 40 obtains a route from the navigation server and outputs guidance information. The agent device 100 may be constructed based on a navigation controller, and in this case, the navigation controller and the agent device 100 may be hardware-integrated.
The vehicle equipment 50 includes, for example, a driving force output device such as an engine and a traveling motor, a start motor of the engine, a door lock device, a door opening/closing device, a window opening/closing device and a window opening/closing control device, a seat position control device, an interior mirror and an angular position control device thereof, an illumination device and a control device thereof for the interior and exterior of a vehicle, a wiper, a defogger and control devices thereof, a turn signal lamp and control devices thereof, an air conditioner, information on a traveling distance, air pressure of a tire, information on the remaining amount of fuel, and the like.
The in-vehicle communication device 60 is, for example, a wireless communication device capable of accessing the network NW using a cellular network or a Wi-Fi network.
The occupant recognition device 80 includes, for example, a seating sensor, an in-vehicle camera, an image recognition device, and the like. The seating sensor includes a pressure sensor provided at a lower portion of the seat, a tension sensor attached to the seat belt, and the like. The in-vehicle camera is a CCD (Charge Coupled Device) camera or CMOS (Complementary Metal Oxide Semiconductor) camera provided in the vehicle. The image recognition device analyzes an image of the camera in the vehicle interior, and recognizes the presence or absence of a passenger for each seat, the face orientation, and the like. In the present embodiment, the occupant recognition device 80 is an example of a seating position recognition unit.
Fig. 3 is a diagram showing a configuration example of the display-operation apparatus 20. The display-operation device 20 includes, for example, a first display 22, a second display 24, and an operation switch ASSY26. The display-and-operation device 20 may also include a HUD28.
The vehicle M includes, for example, a driver seat DS provided with a steering wheel SW, and a secondary driver seat AS provided in a vehicle width direction (Y direction in the drawing) with respect to the driver seat DS. The first display 22 is a horizontally long display device in the instrument panel extending from near the middle between the driver seat DS and the secondary driver seat AS to a position facing the left end portion of the secondary driver seat AS. The second display 24 is provided near the middle of the driver seat DS and the secondary driver seat AS in the vehicle width direction and below the first display. For example, the first display 22 and the second display 24 are each configured as a touch panel, and each includes a display portion LCD (Liquid Crystal Display), an organic EL (Electroluminescence), a plasma display, and the like. The operation switch ASSY26 is a combination of a dial switch, a push button switch, and the like. The display-operation device 20 outputs the content of the operation performed by the occupant to the agent device 100. The content displayed by the first display 22 or the second display 24 may be determined by the agent device 100.
Fig. 4 is a diagram showing a configuration example of the speaker unit 30. The speaker unit 30 includes, for example, speakers 30A to 30H. The speaker 30A is provided in a window pillar (so-called a pillar) on the driver seat DS side. The speaker 30B is provided at a lower portion of the door near the driver seat DS. The speaker 30C is provided in a window pillar on the side of the side driver's seat AS. The speaker 30D is provided at a lower portion of the door near the side driver seat AS. The speaker 30E is provided at a lower portion of the door near the right rear seat BS1 side. The speaker 30F is provided at a lower portion of the door near the left rear seat BS2 side. The speaker 30G is disposed near the second display 24. The speaker 30H is provided on a roof (roof) of the vehicle cabin.
In this configuration, for example, in the case where the speakers 30A and 30B are exclusively made to output sound, the sound image is positioned near the driver seat DS. When the speakers 30C and 30D are exclusively made to output sound, the sound image is positioned near the driver seat AS. In the case of exclusively letting the speaker 30E output sound, the sound image will be positioned near the right rear seat BS 1. In the case of exclusively letting the speaker 30F output sound, the sound image will be positioned near the left rear seat BS 2. When the speaker 30G is exclusively used to output sound, the sound image is positioned near the front of the vehicle interior, and when the speaker 30H is exclusively used to output sound, the sound image is positioned near the top of the vehicle interior. The speaker unit 30 is not limited to this, and may be configured to position the sound image at an arbitrary position in the vehicle interior by adjusting the distribution of the sound output from each speaker using a sound mixer and an amplifier.
[ agent device ]
Returning to fig. 2, the agent apparatus 100 includes a management unit 110, agent function units 150-1, 150-2, 150-3, and a partner application execution unit 152. The management unit 110 includes, for example, a sound processing unit 112, a person WU (Wake Up) determination unit 114, a display control unit 116, and a sound control unit 118. In the case where it is not discriminated which agent function section is, it is simply referred to as an agent function section 150. The number of the agent functions 150 is 3, which is merely an example corresponding to the number of the agent servers 200 in fig. 1, and the number of the agent functions 150 may be 2 or 4 or more. The software configuration shown in fig. 2 is simply shown for the sake of explanation, and may be changed in practice, for example, arbitrarily in such a manner that the management unit 110 exists between the agent function unit 150 and the in-vehicle communication device 60.
Each component of the agent device 100 is realized by executing a program (software) by a hardware processor such as CPU (Central Processing Unit), for example. Some or all of these components may be realized by hardware (including a circuit unit) such as LSI (Large Scale Integration), ASIC (Application Specific Integrated Circuit), FPGA (Field-Programmable Gate Array), GPU (Graphics Processing Unit), or by cooperation of software and hardware. The program may be stored in advance in a storage device (storage device including a non-transitory storage medium) such as HDD (Hard Disk Drive) or a flash memory, or may be stored in a removable storage medium (non-transitory storage medium) such as a DVD or a CD-ROM, and installed by mounting the storage medium on a drive device.
The management unit 110 functions by executing programs such as OS (Operating System) and middleware.
The sound processing unit 112 of the management unit 110 performs sound processing on the input sound so as to be in a state suitable for recognizing a wake-up word preset for each agent.
Each agent WU determination unit 114 exists in association with each agent function unit 150-1, 150-2, 150-3, and recognizes a wake-up word preset for each agent. Each agent WU determination unit 114 recognizes the meaning of a sound from the sound (sound stream) subjected to the sound processing. First, each agent WU determination unit 114 detects a sound zone based on the amplitude and zero-crossing of the sound waveform in the sound stream. Each agent WU determination unit 114 may perform section detection by voice recognition and non-voice recognition in frame units based on a mixed gaussian distribution model (GMM; gaussian mixture model).
Next, each agent WU determination unit 114 text the sound in the detected sound zone to form text information. Then, each agent WU determination unit 114 determines whether or not the textual word information belongs to a wake-up word. When the wake word is determined, each agent WU determination unit 114 causes the corresponding agent function unit 150 to start. The function corresponding to each agent WU determination unit 114 may be mounted on the agent server 200. In this case, the management unit 110 transmits the sound subjected to the sound processing by the sound processing unit 112 to the agent server 200, and when the agent server 200 determines that the sound is a wake-up word, the agent function unit 150 is activated in response to an instruction from the agent server 200. Each agent function unit 150 can always start and automatically determine the wake-up word. In this case, the management unit 110 does not need to include the individual agent WU determination unit 114.
The agent function unit 150 cooperates with the corresponding agent server 200 to cause the agent to appear, and provides a service including a response based on sound in response to the speech of the occupant of the vehicle. The agent function unit 150 may include a function unit to which a right to control the vehicle device 50 is given. The agent function unit 150 may include a function unit that communicates with the agent server 200 in cooperation with the general-purpose communication device 70 via the mating application executing unit 152. For example, the agent function portion 150-1 is given authority to control the vehicle device 50. The agent function unit 150-1 communicates with the agent server 200-1 via the in-vehicle communication device 60. The agent function unit 150-2 communicates with the agent server 200-2 via the in-vehicle communication device 60. The agent function unit 150-3 communicates with the agent server 200-3 in cooperation with the general communication device 70 via the mating application executing unit 152. The pairing application execution unit 152 pairs with the general-purpose communication device 70, for example, by Bluetooth (registered trademark), and connects the agent function unit 150-3 to the general-purpose communication device 70. The agent function unit 150-3 may be connected to the general-purpose communication device 70 by wired communication using USB (Universal Serial Bus) or the like. Hereinafter, an agent that appears by cooperation of the agent function unit 150-1 and the agent server 200-1 may be referred to as an agent 1, an agent that appears by cooperation of the agent function unit 150-2 and the agent server 200-2 may be referred to as an agent 2, and an agent that appears by cooperation of the agent function unit 150-3 and the agent server 200-3 may be referred to as an agent 3.
The display control unit 116 causes the first display 22 or the second display 24 to display an image in response to an instruction from the agent function unit 150. Hereinafter, the first display 22 is used. The display control unit 116 generates an image of an agent (hereinafter referred to as an agent image) that is personified and communicates with the occupant in the vehicle interior, for example, by control of a part of the agent function unit 150, and causes the generated agent image to be displayed on the first display 22. The agent image is, for example, an image of a form of speaking to an occupant. The agent image may include, for example, at least a face image to which a viewer (occupant) recognizes an expression or a face orientation. For example, the agent image may be a part that mimics eyes and nose in a face region, and the expression and the face orientation may be recognized based on the position of the part in the face region. The body image can also be perceived stereoscopically, the face orientation of the body can be recognized by the viewer by including the head image in the three-dimensional space, and the motion, the holding, the posture, and the like of the body can be recognized by the viewer by including the image of the body (trunk, hands, feet). The agent image may also be an animated image.
The sound control unit 118 causes some or all of the speakers included in the speaker unit 30 to output sound in response to an instruction from the agent function unit 150. The sound control unit 118 may perform control to position the sound image of the agent sound at a position corresponding to the display position of the agent image by using the plurality of speaker units 30. The position corresponding to the display position of the body image is, for example, a position predicted to be at which the occupant feels that the body image is speaking the body sound, specifically, a position near (for example, within 2 to 3 cm) the display position of the body image. The sound image localization is, for example, setting the spatial position of a sound source perceived by an occupant by adjusting the magnitude of sound transmitted to the left and right ears of the occupant.
Fig. 5 is a diagram for explaining the principle of position determination for sound image localization. In fig. 5, an example using the speakers 30B, 30D, 30G described above is shown for simplicity of explanation, but any speaker included in the speaker unit 30 may be used. The sound control unit 118 controls an Amplifier (AMP) 32 and a sound mixer 34 connected to the respective speakers to localize the sound image. For example, when the sound image is positioned at the spatial position MP1 shown in fig. 5, the sound control unit 118 controls the amplifier 32 and the sound mixer 34 to output 5% of the maximum intensity from the speaker 30B, 80% of the maximum intensity from the speaker 30D, and 15% of the maximum intensity from the speaker 30G. As a result, from the position of the occupant P, the sound image is perceived as if it were positioned at the spatial position MP1 shown in fig. 5.
When the sound image is positioned at the spatial position MP2 shown in fig. 5, the sound control unit 118 controls the amplifier 32 and the sound mixer 34 to output 45% of the maximum intensity to the speaker 30B, 45% of the maximum intensity to the speaker 30D, and 45% of the maximum intensity to the speaker 30G. As a result, from the position of the occupant P, the sound image is perceived as if it were positioned at the spatial position MP2 shown in fig. 5. In this way, by adjusting the plurality of speakers provided in the vehicle interior and the size of the sound output from each speaker, the position where the sound image is localized can be changed. More specifically, since the position of the sound image localization is determined based on the sound characteristics originally held by the sound source, the information of the interior environment of the vehicle, and the Head transfer function (HRTF; head-related transfer function), the sound control unit 118 controls the speaker unit 30 so as to localize the sound image at a predetermined position by using the optimum output distribution obtained in advance by the sensory test or the like.
[ agent Server ]
Fig. 6 is a diagram showing the structure of the agent server 200 and a part of the structure of the agent device 100. The following describes operations of the agent function unit 150 and the like, together with the configuration of the agent server 200. Here, a description of physical communication from the agent apparatus 100 to the network NW is omitted.
The agent server 200 includes a communication unit 210. The communication unit 210 is a network interface such as NIC (Network Interface Card). The agent server 200 includes, for example, a voice recognition unit 220, a natural language processing unit 222, a dialogue management unit 224, a network search unit 226, and a response document generation unit 228. These components are realized by executing a program (software) by a hardware processor such as a CPU. Some or all of these components may be realized by hardware (including a circuit unit) such as LSI, ASIC, FPGA, GPU, or by cooperation of software and hardware. The program may be stored in advance in a storage device (storage device including a non-transitory storage medium) such as an HDD or a flash memory, or may be stored in a removable storage medium (non-transitory storage medium) such as a DVD or a CD-ROM, and installed by mounting the storage medium on a drive device.
The agent server 200 includes a storage unit 250. The storage unit 250 is implemented by the various storage devices described above. The storage unit 250 stores data and programs such as personal profiles 252, dictionary DB (database) 254, knowledge base DB256, and response rule DB 258.
In the agent device 100, the agent function unit 150 transmits the audio stream or the audio stream subjected to the processing such as compression and encoding to the agent server 200. When recognizing a voice command that can be processed locally (without processing via the smart server 200), the smart function unit 150 can perform processing required by the voice command. The voice command that can be processed locally refers to a voice command that can be answered by referring to a storage unit (not shown) provided in the agent device 100, and in the case of the agent function unit 150-1, a voice command (for example, a command to turn on the air conditioner) that controls the vehicle device 50. Accordingly, the agent function unit 150 may have a part of the functions provided in the agent server 200.
When the voice stream is acquired, the voice recognition unit 220 performs voice recognition and outputs text information, and the natural language processing unit 222 performs meaning interpretation on the text information while referring to the dictionary DB 254. The dictionary DB254 has abstracted meaning information corresponding to the character information. The dictionary DB254 may contain list information of synonyms and paraphraseology. The processing by the voice recognition unit 220 and the processing by the natural language processing unit 222 may not be explicitly performed in stages, but may be performed in such a manner that the voice recognition unit 220 corrects the recognition result by receiving the processing result of the natural language processing unit 222.
For example, when it is recognized that the meaning of "weather today" is "what weather is" or the like is the recognition result, the natural language processing unit 222 generates a command to replace the standard text information "weather today". Thus, even if there is a difference in the expression of the requested sound, a dialogue corresponding to the request can be easily performed. The natural language processing unit 222 may recognize the meaning of the character information by using an artificial intelligence process such as a mechanical learning process using probability, for example, and generate an instruction based on the recognition result.
The dialogue manager 224 refers to the personal profile 252, the knowledge base DB256, and the response rule DB258, and determines the content of the speech to the occupant of the vehicle M based on the processing result (instruction) of the natural language processor 222. The personal profile 252 contains personal information of the occupant, interest preferences, histories of past conversations, and the like, which are held for each occupant. The knowledge base DB256 is information defining the relatedness of things. The response rule DB258 is information defining actions (answers, contents of device control, etc.) that the agent should perform with respect to the instruction.
The dialogue manager 224 may also identify the occupant by comparing the feature information obtained from the audio stream with the personal profile 252. In this case, in the personal profile 252, characteristic information such as sound and personal information are associated. The characteristic information of the sound is, for example, information related to a characteristic of a speaking mode such as a sound level, a intonation, a rhythm (a pattern of the sound level), or a characteristic amount based on a mel-frequency cepstrum coefficient (Mel Frequency Cepstrum Coefficients). The characteristic information of the sound is, for example, information obtained by making the occupant sound a predetermined word, sentence, or the like at the time of initial registration of the occupant, and recognizing the sound emitted.
When the instruction is an instruction requesting information that can be retrieved via the network NW, the session management unit 224 causes the network retrieval unit 226 to retrieve the information. The network search unit 226 accesses various web servers 300 via the network NW to acquire desired information. The "information that can be retrieved via the network NW" is, for example, an evaluation result by a general user of a restaurant located in the vicinity of the vehicle M, and a weather forecast corresponding to the position of the vehicle M on the next day.
The response document generation unit 228 generates a response document so that the content of the speech determined by the dialogue management unit 224 is transmitted to the occupant of the vehicle M, and transmits the response document to the agent device 100. When it is determined that the occupant is registered in the personal profile, the response document generating unit 228 may generate a response document called the name of the occupant and set to a speaking manner that mimics the speaking manner of the occupant.
When the response message is acquired, the agent function unit 150 instructs the sound control unit 118 to perform sound synthesis and output a sound. The agent function unit 150 instructs the display control unit 116 to display an image of the agent in association with the sound output. In this way, an agent function in which an agent that appears virtually responds to the occupant of the vehicle M is realized.
[ Sound image control ]
Hereinafter, sound image control performed by the agent apparatus 100 will be described. The sound control unit 118 sets positions for positioning the sound image of the sound for each of the plurality of agent function units 150, and controls the outputs of the plurality of speakers included in the speaker unit 30 so that the sound image is positioned at the set positions. Hereinafter, a rule for setting the position of sound image localization by the sound control unit 118 is exemplified. The rules shown below include non-exclusive and exclusive, but may be used in combination with non-exclusive.
(rule 1)
The sound control unit 118, for example, positions the nearest seat among the seats of the vehicle M (hereinafter, the co-driver seat vicinity position) at the co-driver seat AS based on the sound image of the sound generated by the high-frequency usage agent function unit 150 (in other words, the high-frequency usage agent) among the plurality of agent function units 150. More specifically, the sound control unit 118 positions the sound image of the sound obtained by the agent function unit 150 based on the "highest" frequency of use among the agent function units 150 at the near-to-the-driver seat position. Instead of this, the sound control unit 118 may position the sound image of the sound generated by the agent function unit 150 based on the "n upper positions from the higher side" of the use frequencies of the agent function units 150 at the position where the nearest seat among the seats of the vehicle M is the secondary driver seat AS (n is a natural number). The near-by-side driver seat position is, for example, a position facing the secondary driver seat AS on the first display 22. Fig. 7 is a diagram showing a first example of the sound image localization position. Here, the use frequency of the agent 2 is the highest, and the use frequency of the agent 1 and the agent 3 is lower than that of the agent 2. MP (2) is a position based on the sound image localization of the sound generated by the agent 2 (position near the seat of the co-driver), and MP (1, 3) is a position based on the sound image localization of the sound generated by the agent 1 and the agent 3. At this time, the first display 22 may display the agent image EG (2) as an animated image that operates in accordance with the operation state of the agent 2. In the case where the speaker is a driver, the agent image EG (2) is formed of a picture perceived as if it were directed toward the driver seat DS. By setting the position of the sound image localization in this way, it is possible for the occupant (driver) P seated in the driver seat DS to use the agent with a sense of use such AS asking what is virtually felt by the agent in the vicinity of the secondary driver seat AS (or without displaying the agent image EG (2)) to be seated in the secondary driver seat AS. As a result, the smart device 100 can provide smart functions with a natural sense of use.
The usage frequency of the agent is counted by the management unit 110 based on the number of starts or the like, and stored in the memory.
(rule 2)
The sound control unit 118 may locate the sound image based on the sound generated by the agent function unit 150-1 to which the authority to control the vehicle device 50 is given, at a position where the nearest seat among the seats of the vehicle is the driver seat DS (hereinafter, the driver seat vicinity position). The driver seat vicinity position is, for example, in the vicinity of a position where the display surface of the HUD28 exists. Fig. 8 is a diagram showing a second example of sound image localization positions. MP (1) is a position based on the sound image localization of the sound generated by the agent 1 (the position near the driver's seat), and MP (2, 3) is a position based on the sound image localization of the sound generated by the agent 2 and the agent 3. At this time, for example, the HUD28 may display the agent image EG (1) as an animated image that operates in accordance with the operation state of the agent 1. In the case where the vehicle apparatus 50 is operated by a switch or the like, the switch is generally provided in the vicinity of the driver seat, and therefore, by setting the position of sound image localization in this way, it is possible for the occupant (driver) P seated in the driver seat DS to utilize the agent function with the same sense of use as in the case where the vehicle apparatus 50 is operated without using the agent function. As a result, the smart device 100 can provide smart functions with a natural sense of use.
(rule 3)
The sound control unit 118 may be configured to locate the sound image of the sound generated by the low-frequency-use agent function unit 150 (in other words, the low-frequency-use agent) among the plurality of agent function units 150 at a position where the nearest seat among the seats of the vehicle M is the rear seat (hereinafter, the rear seat vicinity position). More specifically, the sound control unit 118 positions the sound image based on the sound generated by the "lowest frequency of use" of the plurality of agent function units 150 at a position near the rear seat. Instead of this, the sound control unit 118 may position the sound image of the sound generated by the agent function unit 150 based on the "k lower positions from the lower position" among the plurality of agent function units 150 at the rear seat vicinity position (k is a natural number). The rear seat vicinity position is, for example, an arbitrary position on a straight line (or a space region having a width) connecting the speaker 30F and the speaker 30E. Fig. 9 is a diagram showing a third example of sound image localization positions. The present figure shows an example in which rules 1, 2, and 3 are combined. MP (1) is a position based on the sound image localization of the sound generated by the agent 1, MP (2) is a position based on the sound image localization of the sound generated by the agent 2, and MP (3) is a position (rear seat vicinity position) based on the sound image localization of the sound generated by the agent 3. Here, the highest frequency of use of the agent 2 and the lowest frequency of use of the agent 3 are assumed. At this time, for example, when the agent 1 is activated, the agent image EG (1) may be displayed on the HUD28, and when the agent 2 is activated, the agent image EG (2) may be displayed on the first display 22. By setting the position of sound image localization in this way, it is possible for the occupant (driver) P seated in the driver seat DS to make sound of the agent that is only used infrequently sound to come from a position away from itself. As a result, the smart device 100 can provide smart functions with a natural sense of use.
(rule 4)
The sound control unit 118 may, for example, locate the sound image of the sound generated by the agent function unit 150-1 to which the authority to control the vehicle device 50 is given, at a position near the driver's seat. Fig. 10 is a diagram showing a fourth example of sound image localization positions. MP (1) is a position based on the sound image localization of the sound generated by the agent 1 (position near the seat of the co-driver), MP (2) is a position based on the sound image localization of the sound generated by the agent 2, and MP (3) is a position based on the sound image localization of the sound generated by the agent 3. In this case, for example, the MP (1) may be moved to the driver's seat position in response to the transition of the session with the agent 1. For example, the sound control unit 118 may set the MP (1) at the driver seat vicinity position in a stage where the agent 1 listens to the urgent request of the occupant P, and set the MP (1) at the driver seat vicinity position in a stage where the agent 1 is described in detail based on the reply from the agent server 200-1. The agent image EG (1) may be moved from a position near the driver's seat to the HUD28 on the first display 22. By setting the agent 1 having good usability capable of controlling the vehicle device 50 in the vicinity of the secondary driver seat where the conversation can be performed naturally in this way, the occupant (driver) P seated in the driver seat can start the conversation naturally. In the stage of listening to the detailed description, the sound image or the image of the agent 1 is moved to a position close to itself, so that a sense of use such as a conversation with a person can be provided. As a result, the agent apparatus 100 can provide the agent function with a natural sense of use.
(rule 5)
The sound control unit 118 may set a position for locating the sound image of the sound for each of the plurality of agent function units 150, based on the seating position of the occupant recognized by the occupant recognition device 80, for example. Fig. 11 is a diagram showing a fifth example of sound image localization positions. In the present figure, an occupant P1 is seated on a driver seat DS, an occupant P2 is seated on a passenger seat AS, and an occupant P3 is seated on a rear left seat. MP (1) is a position based on the sound image localization of the sound generated by the agent 1 (driver seat vicinity position), MP (2) is a position based on the sound image localization of the sound generated by the agent 2 (co-driver seat vicinity position), and MP (3) is a position based on the sound image localization of the sound generated by the agent 3 (position biased to the left rear seat in the rear seat vicinity position). By setting the position of sound image localization in this way, an environment in which one agent responds to each occupant can be provided.
Fig. 12 is a diagram showing a sixth example of sound image localization positions. In the present figure, all of the driver's seat DS, the passenger's seat AS, the right rear seat BS1, and the left rear seat BS2 are occupied with passengers. In this case, the sound control unit 118 sets, for example, a position for locating a sound image based on sound generated by each of the plurality of agent function units 150 in the ceiling of the vehicle interior (MP in the figure). The setting of the position for locating the sound image to the ceiling of the vehicle interior is not limited to the position for locating the sound image to the ceiling, and any operation may be used as long as the occupant feels that the sound source is located in the ceiling, and for example, the sound image may be located above the upper end of the seat. In this case, the sound control unit 118 may output only the speaker 30H, or may use another speaker 30. By doing so, an environment in which the agent responds to all the occupants simultaneously can be provided. As a result, the agent apparatus 100 can provide the agent function with a natural sense of use.
As described above, the sound control unit 118 has been described as setting the sound image localization position, but the sound control unit 118 may perform initial setting of the sound image localization position for each agent in accordance with the operation of the occupant. In this case, when the occupant performs an input operation using the first display 22 or the like, the sound control portion 118 performs initial setting according to the content of the input operation. Then, the positions of the sound image localization may be sequentially changed according to the frequency of use of the agent.
According to the agent device 100 of the first embodiment described above, the agent function can be provided with a more natural sense of use.
According to the agent device 100, the discrimination of the agent can be further facilitated. For example, if it is assumed that a plurality of agents are activated at the same time, it may be assumed that it is difficult to distinguish which agent has spoken in the case that the accents of the sounds are similar. In this case, the occupant may also forget which agent is activated, and in this case, it is conceivable that the occupant requests the agent in the activation state to be effective or not. In contrast, in the agent device 100 according to the embodiment, the position of the sound image localization of the sound of the agent is set to some extent fixedly, so that the occupant can determine the agent in the activation based on the direction in which the sound is heard.
< second embodiment >
The second embodiment will be described below. Fig. 13 is a diagram showing the configuration of an agent apparatus 100A according to the second embodiment and equipment mounted on a vehicle M. The occupant recognition device 80 in the second embodiment is an example of a face direction recognition unit. The face orientation is recognized based on, for example, arrangement of characteristic parts such as nose bridge extracted from a face region in an image. The agent apparatus 100A according to the second embodiment further includes an activation agent selecting unit 120 as compared with the first embodiment.
The mechanism for setting the position for locating the sound image based on the sound generated by the agent function unit 150 in the sound control unit 118 is the same as that in the first embodiment.
In the first embodiment, it is assumed that the wake-up word for each agent function unit 150 is specified, and the agent function unit 150 corresponding to the uttered wake-up word is activated. In the second embodiment, it is premised that free start-up of the agent is enabled without specifying the agent. In order to realize the free start, it is necessary to determine whether the occupant is speaking into each other or speaking into the agent, but in the present invention, the method of the determination is not particularly limited, and therefore, a detailed description is omitted.
The activated agent selecting unit 120 recognizes that the occupant has performed a "speech requiring an agent response without specifying an activated agent" (hereinafter, free activated speech) although not a wake-up word, based on the accent of the sound made by the occupant, the presence or absence of the response immediately after that, and the like. At this time, the agent selecting unit 120 is activated to acquire the face orientation of the speaking occupant recognized by the occupant recognition device 80. Then, the startup agent selecting unit 120 selects and starts up an agent whose sound image localization position is set to a position forward of the face orientation of the speaking occupant.
Fig. 14 is a diagram for explaining an operation of the activation agent selecting unit 120. In the upper diagram of fig. 14, the sound image of the sound of the agent 1 is set to the driver's sideIn the near position, the sound image of the sound of the agent 2 is set in the vicinity of the left rear seat BS2, and the sound image of the sound of the agent 3 is set in the vicinity of the right rear seat BS 1. In this state, when the occupant (driver) P makes the face obliquely rearward left and starts speaking freely, the startup agent selecting unit 120 starts up the agent 2 whose sound image set as sound is positioned obliquely rearward left as viewed from the occupant P, as shown in the lower diagram of fig. 14. In the figure, D P The face orientation of the occupant P recognized by the occupant recognition device 80 is shown.
With this control, in the agent device 100A according to the second embodiment, even when the occupant forgets the method of starting the agent, if the same agent can be started up when the "agent whose sound image of the past sound was located in the direction" is stored. As a result, convenience can be improved.
According to the agent device 100A of the second embodiment described above, in addition to the same effects as those of the first embodiment, convenience can be further improved.
The specific embodiments of the present invention have been described above using the embodiments, but the present invention is not limited to such embodiments, and various modifications and substitutions can be made without departing from the scope of the present invention.

Claims (10)

1. An intelligent agent apparatus, wherein,
the agent device is provided with:
a plurality of agent function sections that provide services including responses by sound in accordance with speech of occupants of the vehicle; and
a sound control section that sets positions for positioning sound images of the sounds for the plurality of agent function sections, respectively, and controls outputs of the plurality of speakers to position the sound images at the set positions, respectively, so that the occupant determines which of the plurality of agent function sections is in activation based on a direction in which the sounds are heard,
The sound control unit sequentially changes the position of sound image localization according to the frequency of use of each of the plurality of agent function units.
2. The smart device of claim 1, wherein,
the sound control unit positions a sound image based on sound generated by an agent function unit having a high frequency of use among the plurality of agent function units at a position where a nearest seat among seats of the vehicle is a secondary driver.
3. The smart device of claim 1, wherein,
the plurality of agent functions include a vehicle agent function to which a right to control the vehicle device is given,
the sound control unit positions a sound image based on the sound generated by the vehicle agent function unit at a position of a secondary driver's seat that is closest to the vehicle's seat.
4. The smart device of claim 1, wherein,
the plurality of agent functions include a vehicle agent function to which a right to control the vehicle device is given,
the sound control unit positions a sound image based on the sound generated by the vehicle agent function unit at a position where a nearest seat among seats of the vehicle is a driver seat.
5. The smart device of claim 1, wherein,
the sound control unit positions a sound image based on sound generated by an agent function unit having a low frequency of use among the plurality of agent function units at a position where a nearest seat among seats of the vehicle is a rear seat.
6. The smart device of claim 1, wherein,
the sound control unit sets, for each of the plurality of agent function units, a position at which an acoustic image of the sound is located, based on a recognition result of a seating position recognition unit that recognizes a seating position of an occupant.
7. The smart device of claim 6, wherein,
the sound control unit sets a position for locating a sound image based on the sound generated by each of the plurality of body functions to a ceiling of the vehicle interior when the seating position identification unit identifies that the occupant is seated in each seat.
8. The smart device of claim 1, wherein,
the intelligent agent device further includes an intelligent agent activation selecting unit that activates an intelligent agent function unit that makes an acoustic image be positioned at a position where the driver faces the face, by referring to at least a recognition result of the face-orientation recognizing unit that recognizes the face orientation of the driver.
9. A control method of an intelligent device, wherein,
the control method of the intelligent body device enables a computer to perform the following processing:
enabling any one of the plurality of agent functions to be started;
providing a service including a response by sound according to the speech of the occupant of the vehicle as the function of the activated agent function section;
setting positions for positioning sound images of the sounds for the plurality of agent function sections, respectively, and controlling outputs of the plurality of speakers to position the sound images at the set positions so that the occupant determines which of the plurality of agent function sections is in activation based on a direction in which the sounds are heard; and
the positions of sound image localization are sequentially changed according to the use frequency of each of the plurality of intelligent agent functional units.
10. A storage medium, wherein,
the storage medium stores a program that causes a computer to execute:
enabling any one of the plurality of agent functions to be started;
providing a service including a response by sound according to the speech of the occupant of the vehicle as the function of the activated agent function section;
Setting positions for positioning sound images of the sounds for the plurality of agent function sections, respectively, and controlling outputs of the plurality of speakers to position the sound images at the set positions so that the occupant determines which of the plurality of agent function sections is in activation based on a direction in which the sounds are heard; and
the positions of sound image localization are sequentially changed according to the use frequency of each of the plurality of intelligent agent functional units.
CN202010085731.6A 2019-02-14 2020-02-10 Agent device, method for controlling agent device, and storage medium Active CN111559317B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2019024471A JP2020135110A (en) 2019-02-14 2019-02-14 Agent device, control method of agent device, and program
JP2019-024471 2019-02-14

Publications (2)

Publication Number Publication Date
CN111559317A CN111559317A (en) 2020-08-21
CN111559317B true CN111559317B (en) 2024-03-19

Family

ID=72067736

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010085731.6A Active CN111559317B (en) 2019-02-14 2020-02-10 Agent device, method for controlling agent device, and storage medium

Country Status (2)

Country Link
JP (1) JP2020135110A (en)
CN (1) CN111559317B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024180632A1 (en) * 2023-02-27 2024-09-06 日産自動車株式会社 Sound field control method and sound field control device

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6249720B1 (en) * 1997-07-22 2001-06-19 Kabushikikaisha Equos Research Device mounted in vehicle
US6356639B1 (en) * 1997-04-11 2002-03-12 Matsushita Electric Industrial Co., Ltd. Audio decoding apparatus, signal processing device, sound image localization device, sound image control method, audio signal processing device, and audio signal high-rate reproduction method used for audio visual equipment
JP2005189301A (en) * 2003-12-24 2005-07-14 Pioneer Electronic Corp Information control device, its system, its method, its program, recording medium with the program recorded thereon and travel supporting apparatus
CN1898721A (en) * 2003-12-26 2007-01-17 株式会社建伍 Device control device, speech recognition device, agent device, on-vehicle device control device, navigation device, audio device, device control method, speech recognition method, agent processing me
JP2007298592A (en) * 2006-04-28 2007-11-15 Fujitsu Ten Ltd Speech processing apparatus
CN104442622A (en) * 2013-09-25 2015-03-25 现代自动车株式会社 Sound control system and method for vehicle
WO2017081960A1 (en) * 2015-11-09 2017-05-18 三菱自動車工業株式会社 Voice recognition control system
CN108665891A (en) * 2017-03-28 2018-10-16 卡西欧计算机株式会社 Sound detection device, sound detection method and recording medium

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB0010034D0 (en) * 2000-04-26 2000-06-14 20 20 Speech Limited Human-machine interface apparatus
JP2007308084A (en) * 2006-05-22 2007-11-29 Fujitsu Ten Ltd On-vehicle display device and acoustic control method

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6356639B1 (en) * 1997-04-11 2002-03-12 Matsushita Electric Industrial Co., Ltd. Audio decoding apparatus, signal processing device, sound image localization device, sound image control method, audio signal processing device, and audio signal high-rate reproduction method used for audio visual equipment
US6249720B1 (en) * 1997-07-22 2001-06-19 Kabushikikaisha Equos Research Device mounted in vehicle
JP2005189301A (en) * 2003-12-24 2005-07-14 Pioneer Electronic Corp Information control device, its system, its method, its program, recording medium with the program recorded thereon and travel supporting apparatus
CN1898721A (en) * 2003-12-26 2007-01-17 株式会社建伍 Device control device, speech recognition device, agent device, on-vehicle device control device, navigation device, audio device, device control method, speech recognition method, agent processing me
JP2007298592A (en) * 2006-04-28 2007-11-15 Fujitsu Ten Ltd Speech processing apparatus
CN104442622A (en) * 2013-09-25 2015-03-25 现代自动车株式会社 Sound control system and method for vehicle
WO2017081960A1 (en) * 2015-11-09 2017-05-18 三菱自動車工業株式会社 Voice recognition control system
CN108665891A (en) * 2017-03-28 2018-10-16 卡西欧计算机株式会社 Sound detection device, sound detection method and recording medium

Also Published As

Publication number Publication date
JP2020135110A (en) 2020-08-31
CN111559317A (en) 2020-08-21

Similar Documents

Publication Publication Date Title
CN111661068B (en) Agent device, method for controlling agent device, and storage medium
JP7266432B2 (en) AGENT DEVICE, CONTROL METHOD OF AGENT DEVICE, AND PROGRAM
JP2020149106A (en) Agent device, agent system, server device, method for controlling agent device, and program
JP2020144712A (en) Agent device, control method of agent device, and program
CN111559328B (en) Agent device, method for controlling agent device, and storage medium
CN111746435B (en) Information providing apparatus, information providing method, and storage medium
JP7340943B2 (en) Agent device, agent device control method, and program
CN111559317B (en) Agent device, method for controlling agent device, and storage medium
CN111717142A (en) Agent device, control method for agent device, and storage medium
CN111660966A (en) Agent device, control method for agent device, and storage medium
CN111724778B (en) In-vehicle apparatus, control method for in-vehicle apparatus, and storage medium
CN111667823B (en) Agent device, method for controlling agent device, and storage medium
JP2020154082A (en) Agent device, control method of agent device, and program
CN111754999B (en) Intelligent device, intelligent system, storage medium, and control method for intelligent device
US11437035B2 (en) Agent device, method for controlling agent device, and storage medium
CN111752235B (en) Server device, agent device, information providing method, and storage medium
CN111754288A (en) Server device, information providing system, information providing method, and storage medium
CN111731323A (en) Agent device, control method for agent device, and storage medium
CN111824174B (en) Agent device, method for controlling agent device, and storage medium
JP2020152298A (en) Agent device, control method of agent device, and program
JP2020142721A (en) Agent system, on-vehicle equipment control method, and program
CN111726772B (en) Intelligent body system, control method thereof, server device, and storage medium
JP2020160108A (en) Agent device, agent device control method and program

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant