Hereinafter, embodiments of the present invention will be described in detail with reference to the accompanying drawings, wherein like reference numerals are used to designate identical or similar elements, and redundant description thereof will be omitted. The suffix "module" and " part "for the components used in the following description are given or mixed in consideration of ease of specification, and do not have their own meaning or role. In the following description of the embodiments of the present invention, a detailed description of related arts will be omitted when it is determined that the gist of the embodiments disclosed herein may be blurred. It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory and are intended to provide further explanation of the invention as claimed. , ≪ / RTI > equivalents, and alternatives.
The vehicle described herein may be a concept including a car, a motorcycle. Hereinafter, the vehicle will be described mainly with respect to the vehicle.
The vehicle described in the present specification may be a concept including both an internal combustion engine vehicle having an engine as a power source, a hybrid vehicle having an engine and an electric motor as a power source, and an electric vehicle having an electric motor as a power source.
Fig. 1 is a view showing the appearance of a vehicle related to the present invention, and Fig. 2 is a view showing a cockpit module included in a vehicle related to the present invention.
1 and 2, the vehicle 100 includes wheels 10FR, 10FL, 10RL, ..., which are rotated by a power source, steering input means 121a for adjusting the traveling direction of the vehicle 100, A camera 122a for capturing an image of the front of the vehicle, and various electric units provided in the vehicle 100. [
The vehicle 100 includes a camera 122b for photographing an in-vehicle image, a first display unit 141 and a second display unit 141b for visually displaying various information, a mobile terminal 200 and a wearable device And an interface unit 170 electrically connected to the display unit 300.
The interface unit 170 may include a mounting unit configured to mount the mobile terminal 200 and the wearable device 300 and a connection unit connected to the mobile terminal 200 and the wearable device 200.
3 is a block diagram for explaining a vehicle related to the present invention.
3, the vehicle 100 includes a communication unit 110, an input unit 120, a sensing unit 130, an output unit 140, a vehicle driving unit 150, a memory 160, an interface unit 170, A control unit 180, and a power supply unit 190.
The communication unit 110 is provided between the vehicle 100 and the mobile terminal 200, between the vehicle 100 and the wearable device 300, between the vehicle 100 and the external server 410, 420 that are capable of communicating wirelessly with one another. In addition, the communication unit 110 may include one or more modules that connect the vehicle 100 to one or more networks.
The communication unit 110 may include a broadcast receiving module 111, a wireless Internet module 112, a short range communication module 113, a location information module 114, and an optical communication module 115.
The broadcast receiving module 111 receives broadcast signals or broadcast-related information from an external broadcast management server through a broadcast channel. Here, the broadcast includes a radio broadcast or a TV broadcast.
The wireless Internet module 112 refers to a module for wireless Internet access, and may be built in or externally mounted in the vehicle 100. The wireless Internet module 112 is configured to transmit and receive wireless signals in a communication network according to wireless Internet technologies.
Wireless Internet technologies include, for example, wireless LAN (WLAN), wireless fidelity (Wi-Fi), wireless fidelity (Wi-Fi) Direct, DLNA (Digital Living Network Alliance), WiBro Interoperability for Microwave Access, High Speed Downlink Packet Access (HSDPA), High Speed Uplink Packet Access (HSUPA), Long Term Evolution (LTE) and Long Term Evolution-Advanced (LTE-A) 112 transmit and receive data according to at least one wireless Internet technology, including Internet technologies not listed above.
The short-range communication module 113 is for short-range communication and may be a Bluetooth ™, a Radio Frequency Identification (RFID), an Infrared Data Association (IrDA), a UWB (Ultra Wideband) It is possible to support near-field communication using at least one of Near Field Communication (NFC), Wireless-Fidelity (Wi-Fi), Wi-Fi Direct and Wireless USB (Universal Serial Bus)
The short-range communication module 113 may form short-range wireless communication networks to perform short-range communication between the vehicle 100 and at least one external device.
4, the short range communication module 113 mounted on the vehicle 100 may include an NFC communication module 113_1, a Bluetooth communication module 113_2, a Wi-Fi communication module 113_3, and the like.
The NFC communication module 113_1 uses the very short-range contactless data transfer technology related to RFID (Radio Frequency IDentification) to transmit data to the devices located within a distance of 10 cm (preferably within 4 cm) And can perform data communication. In the present embodiment, the NFC communication module 113_1 may be installed in the driver's door portion of the vehicle, but is not limited thereto.
The Bluetooth communication module 113_2 can perform data communication with devices within a radius of 10 to 100 m using Bluetooth, which is one of the short-range wireless communication standards. For reference, Bluetooth is a short-range wireless networking technology jointly developed by the Bluetooth Special Group (SIG), which was formed in 1998 by five companies, including Ericsson, IBM and Toshiba.
The WiFi communication module 113_3 combines wireless technology with Hi-Fi (High Fidelity), and is a wireless LAN technology that enables high-performance wireless communication. Such a wireless LAN technology is a local area network (LAN) capable of providing high-speed Internet within a certain distance of a place where a wireless access point is installed.
The position information module 114 is a module for obtaining the position of the vehicle 100, and a representative example thereof is a Global Positioning System (GPS) module. For example, when the mobile terminal utilizes the GPS module, it can acquire the position of the mobile terminal by using a signal transmitted from the GPS satellite.
The optical communication module 115 may include a light emitting portion and a light receiving portion.
The light receiving section can convert the light signal into an electric signal and receive the information. The light receiving unit may include a photodiode (PD) for receiving light. Photodiodes can convert light into electrical signals. For example, the light receiving section can receive information of the front vehicle through light emitted from the light source included in the front vehicle.
The light emitting unit may include at least one light emitting element for converting an electric signal into an optical signal. Here, the light emitting element is preferably an LED (Light Emitting Diode). The optical transmitter converts the electrical signal into an optical signal and transmits it to the outside. For example, the optical transmitter can emit the optical signal to the outside through the blinking of the light emitting element corresponding to the predetermined frequency. According to an embodiment, the light emitting portion may include a plurality of light emitting element arrays. According to the embodiment, the light emitting portion can be integrated with the lamp provided in the vehicle 100. [ For example, the light emitting portion may be at least one of a headlight, a tail light, a brake light, a turn signal lamp, and a car light.
The input unit 120 may include a driving operation unit 121, a camera 122, a microphone 123, and a user input unit 124.
The driving operation means 121 receives a user input for driving the vehicle 100. The driving operation unit 121 may include a steering input unit 121a, a shift input unit (not shown), an acceleration input unit (not shown), and a brake input unit (not shown).
The steering input means 121a receives a forward direction input of the vehicle 100 from the user. The steering input means 121a is preferably formed in a wheel shape so that steering input can be performed by rotation. According to the embodiment, the steering input means 121a may be formed of a touch screen, a touch pad, or a button.
The shift input means receives inputs of parking (P), forward (D), neutral (N), and reverse (R) of the vehicle 100 from the user. The shift input means is preferably formed in a lever shape. According to an embodiment, the shift input means may be formed of a touch screen, a touch pad or a button.
The acceleration input means receives an input for acceleration of the vehicle 100 from the user. The brake input means receives an input for decelerating the vehicle 100 from the user. The acceleration input means and the brake input means are preferably formed in the form of a pedal. According to the embodiment, the acceleration input means or the brake input means may be formed of a touch screen, a touch pad or a button.
The camera 122 may include an image sensor and an image processing module. The camera 122 may process still images or moving images obtained by an image sensor (e.g., CMOS or CCD). The image processing module can process the still image or the moving image obtained through the image sensor, extract necessary information, and transmit the extracted information to the control unit 180. On the other hand, the vehicle 100 may include a first camera 122a for photographing an image in front of the vehicle and a second camera 122b for photographing an in-vehicle image.
The first camera 122a is constituted by a stereo camera, and can acquire a stereo image in front of the vehicle. At this time, the image processing module can provide the distance information with respect to the object detected on the stereo image through the binocular parallax information.
And the second camera 122b may acquire an image of the occupant. And the second camera 122b can acquire an image for biometrics of the passenger.
The microphone 123 can process an external acoustic signal into electrical data. The processed data can be utilized variously according to functions performed in the vehicle 100. The microphone 123 can convert the voice command of the user into electrical data. The converted electrical data may be transmitted to the control unit 180.
The camera 122 or the microphone 123 may be a component included in the sensing unit 130 and not a component included in the input unit 120. [
The user input unit 124 is for receiving information from a user. When the information is inputted through the user input unit 124, the control unit 180 can control the operation of the vehicle 100 to correspond to the input information. The user input unit 124 may include a touch input means or a mechanical input means. The user input unit 124 may be disposed in the steering input unit 121a.
The sensing unit 130 senses a signal related to the running of the vehicle 100 or the like. To this end, the sensing unit 130 may include a sensor such as a collision sensor, a wheel sensor, a velocity sensor, a tilt sensor, a weight sensor, a heading sensor, a yaw sensor, a gyro sensor, , A position module, a vehicle forward / reverse sensor, a battery sensor, a fuel sensor, a tire sensor, a steering sensor by steering wheel rotation, a vehicle internal temperature sensor, an internal humidity sensor, an ultrasonic sensor, a radar, .
Thus, the sensing unit 130 can acquire the vehicle collision information, vehicle direction information, vehicle position information (GPS information), vehicle angle information, vehicle speed information, vehicle acceleration information, vehicle tilt information, , Fuel information, tire information, vehicle lamp information, vehicle internal temperature information, vehicle interior humidity information, and the like.
The sensing unit 130 may further include an accelerator pedal sensor, a pressure sensor, an engine speed sensor, an air flow sensor AFS, an intake air temperature sensor ATS, a water temperature sensor WTS, A position sensor (TPS), a TDC sensor, a crank angle sensor (CAS), and the like.
The sensing unit 130 may include a biometric information sensing unit 131. The biometric information sensing unit 131 senses and acquires the passenger's biometric information. The biometric information may include fingerprint information, iris-scan information, retina-scan information, hand geo-metry information, facial recognition information, Voice recognition information. The biometric information sensing unit 131 may include a sensor for sensing the passenger's biometric information. Here, the camera 122 and the microphone 123 can operate as sensors. The biometric information sensing unit 131 may acquire hand shape information and facial recognition information through the second camera 122b. The biometric information sensing unit 131 can acquire the voice recognition information through the microphone 123. [
Meanwhile, the biometric information sensing unit 131 may further include a fingerprint recognition scanner, an iris recognition scanner, or a retinal recognition scanner to acquire fingerprint recognition information, iris recognition information, or retinal recognition information of a passenger.
The output unit 140 may include a display unit 141, an acoustic output unit 142, and a haptic output unit 143 for outputting information processed by the control unit 180.
The display unit 141 may display information processed by the controller 180. [ For example, the display unit 141 can display vehicle-related information. Here, the vehicle-related information may include vehicle control information for direct control of the vehicle, or vehicle driving assistance information for a driving guide to the vehicle driver.
The display unit 141 may be a liquid crystal display (LCD), a thin film transistor-liquid crystal display (TFT LCD), an organic light-emitting diode (OLED) display, a 3D display, and an e-ink display.
The display unit 141 may have a mutual layer structure with the touch sensor or may be integrally formed to realize a touch screen. This touch screen may function as a user input 124 that provides an input interface between the vehicle 100 and the user and may provide an output interface between the vehicle 100 and the user. In this case, the display unit 141 may include a touch sensor that senses a touch with respect to the display unit 141 so as to receive a control command by a touch method. When a touch is made to the display unit 141, the touch sensor senses the touch, and the control unit 180 generates a control command corresponding to the touch based on the touch. The content input by the touch method may be a letter or a number, an instruction in various modes, a menu item which can be designated, and the like.
On the other hand, two or more display units 141 may exist. For example, the first display unit 141 may be formed in the form of a cluster so that the driver can confirm the information while driving. The second display unit 141b is provided in one area of the center fascia and can operate as an AVN (Audio Video Navigation) apparatus.
Meanwhile, according to the embodiment, the display unit 141 may be implemented as a Head Up Display (HUD). When the display unit 141 is implemented as a HUD, information can be output through a transparent display provided in the windshield. Alternatively, the display unit 141 may include a projection module to output information through an image projected on the windshield.
The sound output unit 142 converts an electric signal from the control unit 180 into an audio signal and outputs the audio signal. For this purpose, the sound output unit 142 may include a speaker or the like. It is also possible for the sound output section 142 to output a sound corresponding to the operation of the user input section 124. [
The haptic output unit 143 generates a tactile output. For example, the haptic output section 143 may operate to vibrate the steering wheel, the seat belt, and the seat so that the user can recognize the output.
The vehicle drive unit 150 can control the operation of various devices of the vehicle. The vehicle driving unit 150 includes a power source driving unit 151, a steering driving unit 152, a brake driving unit 153, a lamp driving unit 154, an air conditioning driving unit 155, a window driving unit 156, an airbag driving unit 157, A driving unit 158 and a suspension driving unit 159.
The power source drive unit 151 can perform electronic control of the power source in the vehicle 100. [
For example, when the fossil fuel-based engine (not shown) is a power source, the power source drive unit 151 can perform electronic control of the engine. Thus, the output torque of the engine and the like can be controlled. When the power source driving unit 151 is an engine, the speed of the vehicle can be limited by limiting the engine output torque under the control of the control unit 180. [
As another example, when the electric-based motor (not shown) is a power source, the power source drive unit 151 can perform control on the motor. Thus, the rotation speed, torque, etc. of the motor can be controlled.
The steering driver 152 may perform electronic control of the steering apparatus in the vehicle 100. [ Thus, the traveling direction of the vehicle can be changed.
The brake driver 153 can perform electronic control of a brake apparatus (not shown) in the vehicle 100. [ For example, the speed of the vehicle 100 can be reduced by controlling the operation of the brakes disposed on the wheels. As another example, it is possible to adjust the traveling direction of the vehicle 100 to the left or right by differently operating the brakes respectively disposed on the left wheel and the right wheel.
The lamp driving unit 154 can control the turn-on / turn-off of the lamp disposed inside or outside the vehicle. Also, the intensity, direction, etc. of the light of the lamp can be controlled. For example, it is possible to perform control on a direction indicating lamp, a brake lamp, and the like.
The air conditioning driving unit 155 can perform electronic control on the air conditioner in the vehicle 100. [ For example, when the temperature inside the vehicle is high, the air conditioner can be operated to control the cooling air to be supplied into the vehicle.
The window driving unit 156 may perform electronic control of a window apparatus in the vehicle 100. [ For example, it can control the opening or closing of left and right windows on the side of the vehicle.
The airbag drive unit 157 may perform electronic control of an airbag apparatus in the vehicle 100. [ For example, at risk, the airbag can be controlled to fire.
The sunroof driving unit 158 may perform electronic control of a sunroof apparatus (not shown) in the vehicle 100. [ For example, you can control the opening or closing of the sunroof.
The suspension driving unit 159 can perform electronic control of the suspension apparatus in the vehicle 100. [ For example, when there is a curvature on the road surface, it is possible to control the suspension device so as to reduce the vibration of the vehicle 100. [
The memory 160 is electrically connected to the control unit 180. The memory 180 may store basic data for the unit, control data for controlling the operation of the unit, and input / output data. The memory 190 may be, in hardware, various storage media such as ROM, RAM, EPROM, flash drive, hard drive, and the like.
The memory 160 may store biometric information of a user corresponding to the mobile terminal and the wearable device by matching with at least one mobile terminal and a wearable device. For example, the memory 160 may store fingerprint information, iris-scan information, retina-scan information, and hand geo-metry information of a user matching the first wearable device ) Information, facial recognition information, or voice recognition information.
The interface unit 170 may serve as a pathway to various kinds of external devices connected to the vehicle 100. For example, the interface unit 170 may include a port that can be connected to the mobile terminal 200 or the wearable device 300, and may be connected to the mobile terminal 200 or the wearable device 300 through the port. have. In this case, the interface unit 170 may exchange data with the mobile terminal 200 or the wearable device 300.
Meanwhile, the interface unit 170 may serve as a path for supplying electrical energy to the connected mobile terminal 200 or the wearable device 300. When the mobile terminal 200 or the wearable device 300 is electrically connected to the interface unit 170, the interface unit 170 moves the electric energy supplied from the power supply unit 190 under the control of the controller 180 To the terminal (200) or the wearable device (300).
The control unit 180 can control the overall operation of each unit in the vehicle 100. [ Here, the controller 180 may be referred to as an ECU (Electronic Control Unit).
The controller 180 may be implemented in hardware as application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs) processors, controllers, micro-controllers, microprocessors, and other electronic units for performing other functions.
According to the control of the controller 180, the power supply 190 can supply power necessary for operation of each component. In particular, the power supply unit 190 can receive power from a battery (not shown) or the like inside the vehicle.
The mobile terminal described in this specification includes a mobile phone, a smart phone, a laptop computer, a digital broadcasting terminal, a personal digital assistant (PDA), a portable multimedia player (PMP), a navigation device, a slate PC A tablet PC, an ultrabook, a wearable device such as a smartwatch, a smart glass, and a head mounted display (HMD). have.
However, it will be appreciated by those skilled in the art that the configuration according to the embodiments described herein may be applied to fixed terminals such as a digital TV, a desktop computer, a digital signage, and the like, will be.
4A and 4C are block diagrams for explaining a mobile terminal according to the present invention, and FIGS. 4B and 4C are conceptual diagrams illustrating an example of a mobile terminal according to the present invention, from different directions.
The mobile terminal 200 includes a wireless communication unit 210, an input unit 220, a sensing unit 240, an output unit 250, an interface unit 260, a memory 270, a control unit 280, and a power supply unit 290, And the like. The components shown in FIG. 4A are not essential for implementing a mobile terminal, so that the mobile terminal described herein may have more or fewer components than the components listed above.
The wireless communication unit 210 may be connected to the wireless communication system between the mobile terminal 200 and the vehicle 100, between the mobile terminal 200 and the vehicle 100, between the mobile terminal 200 and another mobile terminal 200 ), Or one or more modules that enable wireless communication between the mobile terminal 200 and an external server. In addition, the wireless communication unit 210 may include one or more modules that connect the mobile terminal 200 to one or more networks.
The wireless communication unit 210 may include at least one of a broadcast receiving module 211, a mobile communication module 212, a wireless Internet module 213, a short distance communication module 214 and a location information module 215 .
The input unit 220 includes a camera 221 or an image input unit for inputting an image signal, a microphone 222 for inputting an audio signal, an audio input unit, a user input unit 223 for receiving information from a user A touch key, a mechanical key, and the like). The voice data or image data collected by the input unit 220 may be analyzed and processed by a user's control command.
The sensing unit 240 may include at least one sensor for sensing at least one of the information in the mobile terminal, the surrounding environment information surrounding the mobile terminal, and the user information. For example, the sensing unit 240 may include a proximity sensor 241, an illumination sensor 242, a touch sensor, an acceleration sensor, a magnetic sensor, A G-sensor, a gyroscope sensor, a motion sensor, an RGB sensor, an infrared sensor, a finger scan sensor, an ultrasonic sensor, A microphone 221, a battery gauge, an environmental sensor (such as a barometer, a hygrometer, a thermometer, a radiation detection sensor, a temperature sensor, A thermal sensor, a gas sensor, etc.), a chemical sensor (e.g., an electronic nose, a healthcare sensor, a biometric sensor, etc.). Meanwhile, the mobile terminal disclosed in the present specification can combine and utilize information sensed by at least two of the sensors.
The output unit 250 includes at least one of a display unit 251, an acoustic output unit 252, a haptrip module 253, and a light output unit 254 for generating an output related to a visual, auditory, can do. The display unit 251 may have a mutual layer structure with the touch sensor or may be integrally formed to realize a touch screen. The touch screen may function as a user input unit 223 for providing an input interface between the mobile terminal 200 and a user and may provide an output interface between the mobile terminal 200 and a user.
The interface unit 260 serves as a channel with various types of external devices connected to the mobile terminal 200. The interface unit 260 may be configured to connect a device having a wired / wireless headset port, an external charger port, a wired / wireless data port, a memory card port, And may include at least one of a port, an audio I / O port, a video I / O port, and an earphone port. In the mobile terminal 200, corresponding to the connection of the external device to the interface 260, it is possible to perform appropriate control related to the connected external device.
In addition, the memory 270 stores data supporting various functions of the mobile terminal 200. The memory 270 may store a plurality of application programs or applications driven by the mobile terminal 200, data for operation of the mobile terminal 200, and commands. At least some of these applications may be downloaded from an external server via wireless communication. Also, at least a part of these application programs may exist on the mobile terminal 200 from the time of departure for the basic functions (e.g., telephone call receiving function, message receiving function, and calling function) of the mobile terminal 200. Meanwhile, the application program may be stored in the memory 270, installed on the mobile terminal 200, and may be driven by the control unit 280 to perform the operation (or function) of the mobile terminal.
In addition to the operations related to the application program, the control unit 280 typically controls the overall operation of the mobile terminal 200. The control unit 280 may process or process signals, data, information or the like inputted or outputted through the above-mentioned components or may drive an application program stored in the memory 270 to provide or process appropriate information or functions to the user.
In addition, the controller 280 may control at least some of the components illustrated in FIG. 4A to drive an application program stored in the memory 270. FIG. Further, the control unit 280 may operate at least two or more of the components included in the mobile terminal 200 in combination with each other for driving the application program.
The power supply unit 290 receives external power and internal power under the control of the controller 280 and supplies power to the respective components included in the mobile terminal 200. The power supply unit 290 includes a battery, which may be an internal battery or a replaceable battery.
At least some of the components may operate in cooperation with one another to implement a method of operation, control, or control of a mobile terminal according to various embodiments described below. In addition, the operation, control, or control method of the mobile terminal may be implemented on the mobile terminal by driving at least one application program stored in the memory 270. [
Hereinafter, the various components of the mobile terminal 200 will be described in detail with reference to FIG. 4A.
First, referring to the wireless communication unit 210, the broadcast receiving module 211 of the wireless communication unit 210 receives broadcast signals and / or broadcast-related information from an external broadcast management server through a broadcast channel. The broadcast channel may include a satellite channel and a terrestrial channel. More than one broadcast receiving module may be provided to the mobile terminal 200 for simultaneous broadcast reception or broadcast channel switching for at least two broadcast channels.
The mobile communication module 212 may be a mobile communication module or a mobile communication module that is capable of communicating with one or more mobile communication devices in a mobile communication environment using technology standards or communication methods (e.g., Global System for Mobile communication (GSM), Code Division Multi Access (CDMA), Code Division Multi Access 2000 (Enhanced Voice-Data Optimized or Enhanced Voice-Data Only), Wideband CDMA (WCDMA), High Speed Downlink Packet Access (HSDPA), High Speed Uplink Packet Access (HSUPA), Long Term Evolution And an external terminal, or a server on a mobile communication network established according to a long term evolution (AR), a long term evolution (AR), or the like.
The wireless signal may include various types of data depending on a voice call signal, a video call signal or a text / multimedia message transmission / reception.
The wireless Internet module 213 is a module for wireless Internet access, and may be embedded in the mobile terminal 200 or externally. The wireless Internet module 213 is configured to transmit and receive wireless signals in a communication network according to wireless Internet technologies.
Wireless Internet technologies include, for example, wireless LAN (WLAN), wireless fidelity (Wi-Fi), wireless fidelity (Wi-Fi) Direct, DLNA (Digital Living Network Alliance), WiBro Interoperability for Microwave Access, High Speed Downlink Packet Access (HSDPA), High Speed Uplink Packet Access (HSUPA), Long Term Evolution (LTE) and Long Term Evolution-Advanced (LTE-A) 113 transmit and receive data according to at least one wireless Internet technology, including Internet technologies not listed above.
The wireless Internet module 113 for performing a wireless Internet connection through the mobile communication network can be used for wireless Internet access by WiBro, HSDPA, HSUPA, GSM, CDMA, WCDMA, LTE or LTE- May be understood as a kind of the mobile communication module 112.
The short-range communication module 214 is for short-range communication, and includes Bluetooth ™, Radio Frequency Identification (RFID), Infrared Data Association (IrDA), Ultra Wideband (UWB), ZigBee, NFC (Near Field Communication), Wi-Fi (Wireless-Fidelity), Wi-Fi Direct, and Wireless USB (Wireless Universal Serial Bus) technology. The local area communication module 214 is connected to the mobile terminal 200 and the wireless communication system through the wireless area networks, between the mobile terminal 200 and the vehicle 100, Or between a mobile terminal 200 and a network in which the mobile terminal 200 and another mobile terminal 200 or an external server are located. The short-range wireless communication network may be a short-range wireless personal area network.
Here, another mobile terminal 200 is a wearable device (e.g., a smartwatch, a smart glass, etc.) capable of interchanging data with the mobile terminal 200 according to the present invention (smart glass), HMD (head mounted display)). The short range communication module 214 may detect (or recognize) a wearable device capable of communicating with the mobile terminal 200 around the mobile terminal 200. If the detected wearable device is a device authenticated to communicate with the mobile terminal 200 according to the present invention, the control unit 280 may transmit at least a part of the data processed by the mobile terminal 200 to the short- 214 to the wearable device. Therefore, the user of the wearable device can use the data processed by the mobile terminal 200 through the wearable device. For example, according to this, when a phone is received in the mobile terminal 200, the user performs a phone call through the wearable device, or when a message is received in the mobile terminal 200, It is possible to check the message.
The position information module 215 is a module for obtaining the position (or current position) of the mobile terminal, and a representative example thereof is a Global Positioning System (GPS) module or a Wireless Fidelity (WiFi) module. For example, when the mobile terminal utilizes the GPS module, it can acquire the position of the mobile terminal by using a signal transmitted from the GPS satellite. As another example, when the mobile terminal utilizes the Wi-Fi module, it can acquire the position of the mobile terminal based on information of a wireless access point (AP) that transmits or receives the wireless signal with the Wi-Fi module. Optionally, the location information module 215 may replace or additionally perform any of the other modules of the wireless communication unit 210 to obtain data regarding the location of the mobile terminal. The position information module 215 is a module used for obtaining the position (or the current position) of the mobile terminal, and is not limited to the module for directly calculating or acquiring the position of the mobile terminal.
Next, the input unit 220 is for inputting image information (or signal), audio information (or signal), data, or information input from a user. For inputting image information, Or a plurality of cameras 221 may be provided. The camera 221 processes an image frame such as a still image or a moving image obtained by the image sensor in the video communication mode or the photographing mode. The processed image frame can be displayed on the display unit 251 or stored in the memory 270. [ The plurality of cameras 221 provided in the mobile terminal 200 may be arranged to have a matrix structure and various angles or foci may be provided to the mobile terminal 200 through the camera 221 having the matrix structure A plurality of pieces of image information can be input. In addition, the plurality of cameras 221 may be arranged in a stereo structure to acquire a left image and a right image for realizing the stereoscopic image.
The microphone 222 processes the external acoustic signal into electrical voice data. The processed voice data can be utilized variously according to a function (or a running application program) being executed in the mobile terminal 200. Meanwhile, the microphone 222 may be implemented with various noise reduction algorithms for eliminating noise generated in receiving an external sound signal.
The user input unit 223 is for receiving information from a user and when the information is inputted through the user input unit 223, the control unit 280 can control the operation of the mobile terminal 200 to correspond to the input information . The user input unit 223 may include a mechanical input means (or a mechanical key such as a button located on the front, rear or side of the mobile terminal 200, a dome switch, a jog wheel, Jog switches, etc.) and touch-type input means. For example, the touch-type input means may comprise a virtual key, a soft key or a visual key displayed on the touch screen through software processing, The virtual key or the visual key can be displayed on the touch screen with various forms. For example, the virtual key or the visual key can be displayed on the touch screen, ), An icon, a video, or a combination thereof.
Meanwhile, the sensing unit 240 senses at least one of information in the mobile terminal, surrounding environment information surrounding the mobile terminal, and user information, and generates a corresponding sensing signal. The control unit 280 may control the driving or operation of the mobile terminal 200 or may perform data processing, function or operation related to the application program installed in the mobile terminal 200 based on the sensing signal. Representative sensors among various sensors that may be included in the sensing unit 240 will be described in more detail.
First, the proximity sensor 241 refers to a sensor that detects the presence of an object approaching a predetermined detection surface, or the presence of an object in the vicinity of the detection surface, without mechanical contact by using electromagnetic force or infrared rays. The proximity sensor 241 may be disposed in an inner area of the mobile terminal or in proximity to the touch screen, which is covered by the touch screen.
Examples of the proximity sensor 241 include a transmission type photoelectric sensor, a direct reflection type photoelectric sensor, a mirror reflection type photoelectric sensor, a high frequency oscillation type proximity sensor, a capacitive proximity sensor, a magnetic proximity sensor, and an infrared proximity sensor. In the case where the touch screen is electrostatic, the proximity sensor 241 can be configured to detect the proximity of the object with a change in the electric field along the proximity of the object having conductivity. In this case, the touch screen (or touch sensor) itself may be classified as a proximity sensor.
On the other hand, for convenience of explanation, the act of recognizing that the object is located on the touch screen in proximity with no object touching the touch screen is referred to as "proximity touch & The act of actually touching an object on the screen is called a "contact touch. &Quot; The position at which the object is closely touched on the touch screen means a position where the object corresponds to the touch screen vertically when the object is touched. The proximity sensor 241 is capable of sensing proximity touch and a proximity touch pattern (e.g., a proximity touch distance, a proximity touch direction, a proximity touch speed, a proximity touch time, a proximity touch position, have. Meanwhile, the control unit 280 processes data (or information) corresponding to the proximity touch operation and the proximity touch pattern sensed through the proximity sensor 241 as described above, and further provides visual information corresponding to the processed data It can be output on the touch screen. Further, the control unit 280 can control the mobile terminal 200 so that different operations or data (or information) are processed depending on whether the touch to the same point on the touch screen is a proximity touch or a touch contact .
The touch sensor uses a touch (or touch input) applied to the touch screen (or the display unit 251) using at least one of various touch methods such as a resistance film type, a capacitive type, an infrared type, an ultrasonic type, Detection.
For example, the touch sensor may be configured to convert a change in a pressure applied to a specific portion of the touch screen or a capacitance generated in a specific portion to an electrical input signal. The touch sensor may be configured to detect a position, an area, a pressure at the time of touch, a capacitance at the time of touch, and the like where a touch object touching the touch screen is touched on the touch sensor. Here, the touch object may be a finger, a touch pen, a stylus pen, a pointer, or the like as an object to which a touch is applied to the touch sensor.
Thus, when there is a touch input to the touch sensor, the corresponding signal (s) is sent to the touch controller. The touch controller processes the signal (s) and transmits the corresponding data to the control unit 280. Thus, the control unit 280 can know which area of the display unit 251 is touched or the like. Here, the touch controller may be a separate component from the control unit 280, and may be the control unit 280 itself.
On the other hand, the control unit 280 may perform different controls or perform the same control according to the type of the touch object, which touches the touch screen (or a touch key provided in the touch screen). Whether to perform different controls or to perform the same control depending on the type of the touch object may be determined according to the current state of the mobile terminal 200 or an application program being executed.
On the other hand, the touch sensors and the proximity sensors discussed above can be used independently or in combination to provide a short touch (touch), a long touch, a multi touch, a drag touch ), Flick touch, pinch-in touch, pinch-out touch, swipe touch, hovering touch, and the like. Touch can be sensed.
The ultrasonic sensor can recognize the position information of the object to be sensed by using ultrasonic waves. On the other hand, the controller 280 can calculate the position of the wave generating source through the information sensed by the optical sensor and the plurality of ultrasonic sensors. The position of the wave source can be calculated using the fact that the light is much faster than the ultrasonic wave, that is, the time when the light reaches the optical sensor is much faster than the time the ultrasonic wave reaches the ultrasonic sensor. More specifically, the position of the wave generating source can be calculated using the time difference with the time when the ultrasonic wave reaches the reference signal.
The camera 221 includes at least one of a camera sensor (for example, a CCD, a CMOS, etc.), a photo sensor (or an image sensor), and a laser sensor.
The camera 221 and the laser sensor can be combined with each other to sense a touch of a sensing object with respect to a three-dimensional stereoscopic image. The photosensor can be laminated to the display element, which is adapted to scan the movement of the object to be detected proximate to the touch screen. More specifically, the photosensor mounts photo diodes and TRs (Transistors) in a row / column and scans the contents loaded on the photosensor using an electrical signal that varies according to the amount of light applied to the photo diode. That is, the photo sensor performs coordinate calculation of the object to be sensed according to the amount of change of light, and position information of the object to be sensed can be obtained through the calculation.
The display unit 251 displays (outputs) information processed by the mobile terminal 200. For example, the display unit 251 may display execution screen information of an application program driven by the mobile terminal 200 or UI (User Interface) and GUI (Graphic User Interface) information according to the execution screen information .
Also, the display unit 251 may be configured as a stereoscopic display unit for displaying a stereoscopic image.
In the stereoscopic display unit, a three-dimensional display system such as a stereoscopic system (glasses system), an autostereoscopic system (no-glasses system), and a projection system (holographic system) can be applied.
The audio output unit 252 may output audio data received from the wireless communication unit 210 or stored in the memory 270 in a call signal reception mode, a call mode or a recording mode, a voice recognition mode, a broadcast reception mode, The sound output unit 252 also outputs sound signals related to functions (e.g., call signal reception sound, message reception sound, and the like) performed by the mobile terminal 200. [ The sound output unit 252 may include a receiver, a speaker, a buzzer, and the like.
The haptic module 253 generates various tactile effects that the user can feel. A typical example of the haptic effect generated by the haptic module 253 may be vibration. The intensity and pattern of the vibration generated in the haptic module 253 can be controlled by the user's selection or the setting of the control unit. For example, the haptic module 253 may combine and output different vibrations or sequentially output the vibrations.
In addition to vibration, the haptic module 253 may be configured to perform various functions such as a pin arrangement vertically moving with respect to the contact skin surface, a spraying force or suction force of the air through the injection port or the suction port, a scratch on the skin surface, And various tactile effects such as an effect of reproducing a cold sensation using an endothermic or exothermic element can be generated.
The haptic module 253 can not only transmit the tactile effect through the direct contact but also can be implemented so that the user can feel the tactile effect through the muscular sense such as the finger or the arm. The haptic module 253 may include two or more haptic modules according to the configuration of the mobile terminal 200.
The light output unit 254 outputs a signal for notifying the occurrence of an event using the light of the light source of the mobile terminal 200. Examples of events that occur in the mobile terminal 200 may include message reception, call signal reception, missed call, alarm, schedule notification, email reception, information reception through an application, and the like.
The signal output by the optical output unit 254 is implemented as the mobile terminal emits light of a single color or a plurality of colors to the front or rear surface. The signal output may be terminated by the mobile terminal detecting the event confirmation of the user.
The interface unit 260 serves as a path for communication with all external devices connected to the mobile terminal 200. The interface unit 260 receives data from an external device or supplies power to each component in the mobile terminal 200 or allows data in the mobile terminal 200 to be transmitted to an external device. For example, a port for connecting a device equipped with a wired / wireless headset port, an external charger port, a wired / wireless data port, a memory card port, an audio input / output port, a video input / output port, an earphone port, and the like may be included in the interface unit 260.
The identification module is a chip for storing various information for authenticating the usage right of the mobile terminal 200 and includes a user identification module (UIM), a subscriber identity module (SIM) A universal subscriber identity module (USIM), and the like. Devices with identification modules (hereinafter referred to as "identification devices") can be manufactured in a smart card format. Accordingly, the identification device can be connected to the terminal 200 through the interface unit 260.
The interface unit 260 may be a path through which power from the cradle is supplied to the mobile terminal 200 when the mobile terminal 200 is connected to an external cradle, And various command signals may be transmitted to the mobile terminal 200. Various command signals or power from the cradle can be operated as a signal to recognize that the mobile terminal 200 is correctly mounted on the cradle.
The memory 270 may store a program for the operation of the controller 280 and temporarily store input / output data (e.g., phone book, message, still image, moving picture, etc.). The memory 270 may store data on vibrations and sounds of various patterns that are output upon touch input on the touch screen.
The memory 270 may be a flash memory type, a hard disk type, a solid state disk type, an SDD type (Silicon Disk Drive type), a multimedia card micro type ), Card type memory (e.g., SD or XD memory), random access memory (RAM), static random access memory (SRAM), read-only memory (ROM), electrically erasable programmable read memory, a programmable read-only memory (PROM), a magnetic memory, a magnetic disk, and / or an optical disk. The mobile terminal 200 may operate in association with a web storage that performs a storage function of the memory 270 on the Internet.
Meanwhile, as described above, the control unit 280 controls an operation related to an application program and an overall operation of the mobile terminal 200. [ For example, when the state of the mobile terminal satisfies a set condition, the control unit 280 can execute or release a lock state for restricting input of a user's control command to applications.
In addition, the control unit 280 performs control and processing related to voice communication, data communication, video call, or the like, or performs pattern recognition processing capable of recognizing handwriting input or drawing input performed on the touch screen as characters and images, respectively . Further, the control unit 280 may control any one or a plurality of the above-described components in order to implement various embodiments described below on the mobile terminal 200 according to the present invention.
The power supply unit 290 receives external power and internal power under the control of the controller 280 and supplies power required for operation of the respective components. The power supply unit 290 includes a battery, the battery may be an internal battery configured to be chargeable, and may be detachably coupled to the terminal body for charging or the like.
In addition, the power supply unit 290 may include a connection port, and the connection port may be configured as an example of an interface 260 through which an external charger for supplying power for charging the battery is electrically connected.
As another example, the power supply unit 290 may be configured to charge the battery in a wireless manner without using the connection port. In this case, the power supply unit 290 may use at least one of an inductive coupling method based on a magnetic induction phenomenon from an external wireless power transmission apparatus and a magnetic resonance coupling method based on an electromagnetic resonance phenomenon Power can be delivered.
In the following, various embodiments may be embodied in a recording medium readable by a computer or similar device using, for example, software, hardware, or a combination thereof.
In the foregoing, the configurations of the vehicle and the mobile terminal related to the present invention have been described in detail with reference to FIG. 1 to FIG. Hereinafter, a mobile terminal and a control method thereof for controlling information on the notification event to be invisible on a vehicle display unit when a notification event occurs during car mirroring will be described in detail in accordance with an embodiment of the present invention.
5 is a flowchart illustrating an operation of a mobile terminal according to an exemplary embodiment of the present invention.
Referring to FIG. 5, the control unit 280 of the mobile terminal 200 determines whether a user command for requesting a mirroring service is received (S510). That is, the control unit 280 determines whether a user input for selecting a mirroring menu is received on an environment setting screen or a menu screen.
If it is determined in operation 510 that the user command for requesting the mirroring service is received, the controller 280 searches for nearby connectable devices and displays the search result on the display unit 251.
When the vehicle 100 is selected from the devices displayed on the display unit 250, the controller 280 connects the selected vehicle 100 to the near field wireless communication at step S520. At this time, Wi-Fi direct communication may be mainly used for the short-range wireless communication. In addition, Bluetooth communication, infrared communication, UWB communication, Zigbee communication, etc. may be used. In addition, the control unit 280 can connect a wired communication with the selected vehicle 100 using a USB cable or the like.
The control unit 280 mirrors the operation screen currently displayed on the display unit 251 to be displayed on the vehicle display unit 141 (S530). That is, when the mobile terminal 200 transmits the current operation screen to the vehicle 100 using the connected short-range wireless communication, the vehicle 100 transmits the operation screen received from the mobile terminal 200 to the vehicle display unit 141). Examples of the operation screen displayed on the display unit 251 include an idle screen, a lock screen, a home screen, a main menu screen, a message screen, a call screen, an image or a video viewer screen, a broadcast screen, a map screen, A web page screen, and an execution screen for various applications.
If a predetermined notification event is generated while the operation screen is being mirrored (S540), the control unit 280 determines whether the mirroring function of the notification event is set in advance (S550). Examples of the notification event include a message reception event, an email reception event, an SNS reception event, a messenger reception event, and a schedule notification event. In addition, the controller 280 may provide a menu for presetting whether or not to mirror the notification event.
If it is determined in step 550 that the mirroring function of the notification event has been set, the controller 280 mirrors information about the notification event to the vehicle display unit 141 in step S560. At this time, the information about the notification event may include not only the type of the notification event but also the contents of the notification event.
If it is determined in step 550 that the mirroring function of the notification event is not set, the controller 280 displays information about the notification event on the terminal screen or the interlocked wearable device screen (step S570). That is, the control unit 280 does not mirror the information about the notification event to the vehicle 100, but controls only the operation screen to be displayed on the vehicle display unit 141. Further, in a further embodiment, the control unit 280 may mirror the information on the occurrence of the notification event to the vehicle display unit 141 only.
Thereafter, when a user command to terminate the mirroring service is received (S580), the controller 280 terminates the mirroring service according to the present invention. On the other hand, if the user command to terminate the mirroring service is not received (S580), the controller 280 repeatedly performs the operations of steps 530 to 570 described above.
As described above, the mobile terminal according to the present invention displays information on the notification event on the terminal screen only when a notification event occurs during mirroring, so that the notification information related to security, personal privacy, and the like is exposed to the vehicle occupant Can be prevented.
6 is a diagram for describing an operation of a mobile terminal that controls information on a notification event when the notification event occurs during car mirroring so as not to be displayed on the vehicle display unit.
Referring to FIG. 6, the mobile terminal 200 may initiate a mirroring service after connecting a short-range wireless communication with the vehicle 100 according to a user command. Accordingly, the mobile terminal 200 can mirror a home screen (not shown) or an app screen (not shown) displayed on the display unit 251 to the display unit 141 mounted on the vehicle 100. At this time, it is assumed that the home screen or the app screen includes a navigation icon.
When the navigation icon (not shown) displayed on the display unit 251 of the mobile terminal 200 is selected, the mobile terminal 200 executes the navigation application corresponding to the selected icon, The navigation screen 610) can be mirrored on the vehicle display unit 141. [
On the other hand, when a navigation icon (not shown) displayed on the display unit 141 of the vehicle 100 is selected, the vehicle 100 can transmit a control signal corresponding to the user's selection command to the mobile terminal 200. [ The mobile terminal 200 can execute the navigation application corresponding to the selected icon based on the received control signal and mirror the execution screen 610 of the application to the vehicle display unit 141. [
When the message reception event occurs while the navigation screen 610 is being mirrored, the mobile terminal 200 can confirm whether or not the mirroring function of the notification event is set in advance.
As a result of checking, if the mirroring function of the notification event is not set for security, personal privacy protection, safe driving or the like, the mobile terminal 200 generates a corresponding event as shown in FIG. 6 (b) And the message notification window 620 including the contents of the received message can be displayed only on the display unit 251 without being mirrored.
Meanwhile, if the mirroring function of the notification event is set as a result of the checking, the mobile terminal 200 displays the message notification window 620 including the notification information for notifying the occurrence of the event and the content of the received message, Can be mirrored on the vehicle display unit.
7 is a flowchart illustrating an operation of a mobile terminal according to another embodiment of the present invention.
7, when the mirroring service is requested, the controller 280 of the mobile terminal 200 searches for nearby connectable devices and displays them on the display unit 251 (S710). When the vehicle 100 is selected from among the devices displayed on the display unit 250, the controller 280 connects the selected vehicle 100 to the near field wireless communication (S720).
The control unit 280 mirrors the operation screen currently displayed on the display unit 251 to be displayed on the vehicle display unit 141 (S730). At this time, examples of the operation screen displayed on the display unit 251 include an idle screen, a lock screen, a home screen, or a main menu screen.
In the mirroring mode, when a specific application icon is selected through the operation screen displayed on the display unit 251 of the mobile terminal 200 or the display unit 141 of the vehicle 100, Executes an application corresponding to the icon, and displays an execution screen of the application on the display unit 251 (S740).
In addition, the control unit 280 determines whether the corresponding application is set as a non-mirrorable application (S750). Here, examples of the non-mirrorable application include a moving picture application, a gallery application, a broadcasting application, a game application, and the like. In addition, the control unit 280 may provide a menu for setting the non-mirrorable application in advance.
If it is determined in step 750 that the application is set as a non-mirrorable application, the control unit 280 mirrors the notification message to the vehicle display unit 141 for a predetermined period of time (S760 ).
If it is determined in step 750 that the application is not an unmirrorable application, the control unit 280 mirrors the execution screen of the application on the vehicle display unit 141 in step S770.
Thereafter, when a user command to terminate the mirroring service is received (S780), the controller 280 terminates the mirroring service according to the present invention. On the other hand, if the user command to terminate the mirroring service is not received (S780), the controller 280 repeatedly performs the operations of steps 730 to 770 described above.
As described above, according to the present invention, when selecting a specific application icon during mirroring, the mobile terminal controls the execution screen of the application corresponding to the application icon to not be mirrored on the vehicle display unit, It is possible to prevent the app screen from being exposed to the vehicle occupant, and also prevent the app screen, which is obstructed by the safe driving, from being seen by the vehicle driver.
FIG. 8 is a diagram for explaining an operation of a mobile terminal that controls an application screen corresponding to the selected icon to be displayed on a vehicle display unit when a specific application icon is selected during car mirroring.
Referring to FIG. 8, the mobile terminal 200 may initiate a mirroring service after connecting a short-range wireless communication with the vehicle 100 according to a user command. Accordingly, the mobile terminal 200 can mirror the application screen 810 displayed on the display unit 251 to the display unit 141 mounted on the vehicle 100. At this time, it is assumed that the application screen 810 includes a DMB icon.
If the DMB icon 815 displayed on the display unit 251 of the mobile terminal 200 or the display unit 141 of the vehicle 100 is selected while the app screen 810 is being mirrored, May execute a broadcast application corresponding to the selected icon 815 and display an execution screen (i.e., a broadcast screen, 820) of the application on the display unit 251. [
In addition, the mobile terminal 200 can confirm whether or not the broadcast application is preset as a non-mirrorable application. If the broadcasting application is set as a non-mirrorable application for security, personal privacy protection, safe driving, or the like, the mobile terminal 200 may mirror the application, as shown in FIG. 8 (b) It is possible to mirror the notification message 830 to the vehicle display unit 141 for a predetermined period of time.
If the broadcast application is not set as a non-mirrorable application, the mobile terminal 200 may mirror the broadcast screen 820 to the vehicle display unit 141 although it is not shown in the drawing.
9 is a flowchart illustrating an operation of a mobile terminal according to another embodiment of the present invention.
9, when the mirroring service is requested, the controller 280 of the mobile terminal 200 searches for nearby connectable devices and displays them on the display unit 251 (S910). When the vehicle 100 is selected from among the devices displayed on the display unit 250, the controller 280 connects the selected vehicle 100 to the near field wireless communication (S920).
The control unit 280 mirrors the operation screen currently displayed on the display unit 251 to be displayed on the vehicle display unit 141 (S930). Examples of the operation screen displayed on the display unit 251 include an idle screen, a lock screen, a home screen, a main menu screen, a message screen, a call screen, an image or a video viewer screen, a broadcast screen, a map screen, A web page screen, and an execution screen for various applications.
If the call reception event is generated while the operation screen is being mirrored (S940), the control unit 280 determines whether the call reception event is set to be mirrored (S950). In addition, the control unit 280 may provide a menu for presetting whether or not the call reception event is mirrored.
If it is determined in step 950 that the call reception event is set to be mirrored, the control unit 280 divides the screen displayed on the display unit 251 into a first screen and a second screen, And displays a call screen on the second screen. In addition, the control unit 280 mirrors the split screen (i.e., the operation screen and the call screen) currently displayed on the display unit 251 to be displayed on the vehicle display unit 141 (S960).
If it is determined in step 950 that the call reception event is not set to be mirrored, the control unit 280 displays a call screen corresponding to the call reception event on the display unit of the mobile terminal 200 or the wearable device 251) (S970). That is, the control unit 280 does not mirror the call screen corresponding to the call reception event to the vehicle 100, but controls only the operation screen to be displayed on the vehicle display unit 141. At this time, the control unit 280 can mirror only the information about the occurrence of the call reception event to the vehicle display unit 141. [
On the other hand, in another embodiment, the control unit 280 can confirm whether or not the divided screen providing function is set in advance when a call reception event occurs. If it is determined that the split screen providing function has been set, the control unit 280 may mirror the split screen to display the split screen on the vehicle display unit 141 after the divided screen is formed. On the other hand, if it is determined that the split screen providing function is not set, the control unit 280 may mirror the call screen so that only the call screen corresponding to the call receiving event is displayed on the vehicle display unit 141.
Accordingly, the driver of the vehicle can confirm the caller information through the call screen displayed on the display unit 251 of the mobile terminal 200 or the display unit 141 of the vehicle 100, and determine whether to talk with the caller .
Thereafter, when a user command to terminate the mirroring service is received (S980), the controller 280 terminates the mirroring service according to the present invention. On the other hand, if the user command to terminate the mirroring service is not received (S980), the controller 280 repeatedly performs the operations of steps 930 to 970 described above.
As described above, according to the present invention, when a call reception event during mirroring occurs, the mobile terminal mirrors the call screen corresponding to the call reception event together with the current operation screen on the vehicle display unit, Can be improved.
FIG. 10 is a diagram referred to explain the operation of a mobile terminal that combines a current operation screen and a call screen to mirror a vehicle display unit when a call reception event occurs during car mirroring.
Referring to FIG. 10, the mobile terminal 200 may initiate a mirroring service after connecting a short-range wireless communication with the vehicle 100 according to a user command. Accordingly, the mobile terminal 200 can mirror a home screen (not shown) or an app screen (not shown) displayed on the display unit 251 to the display unit 141 mounted on the vehicle 100. At this time, it is assumed that the home screen or the app screen includes a navigation icon.
When a navigation icon (not shown) displayed on the display unit 251 of the mobile terminal 200 or the display unit 141 of the vehicle 100 is selected, the mobile terminal 200 displays a navigation application corresponding to the selected icon And the execution screen of the application (that is, the navigation screen 1010) can be mirrored on the vehicle display unit 141. [
If the call receiving event occurs while the navigation screen 1010 is being mirrored, the mobile terminal 200 can check whether the mirroring function of the call receiving event is set.
As a result of the checking, if the mirroring function of the call reception event is set for the convenience of the driver of the vehicle, the mobile terminal 200 displays the screen displayed on the display unit 251 as shown in FIG. 10 (b) It is possible to display the navigation screen 1020 on the first screen and the call receiving screen 1030 on the second screen after dividing the screen into the first screen and the second screen.
At the same time, the mobile terminal 200 can mirror the navigation screen 1020 and the call reception screen 1030 displayed on the first and second screens to be displayed on the vehicle display unit 141. Also, the mobile terminal 200 can adjust the size of the divided screen according to the touch and drag input received through the boundary between the navigation screen 1020 and the call receiving screen 1030.
Meanwhile, although not shown in the figure, if the mirroring function of the call reception event is not set as a result of the checking, the mobile terminal 200 can display the navigation screen and the call reception screen only on the display unit 251 .
11 is a flowchart illustrating an operation of a mobile terminal according to another embodiment of the present invention.
Referring to FIG. 11, the control unit 280 of the mobile terminal 200 searches for nearby connectable devices and displays them on the display unit 251 when the mirroring service is requested (S1105). When the vehicle 100 is selected among the devices displayed on the display unit 250, the controller 280 connects the selected vehicle 100 to the near field wireless communication (S1110).
The control unit 280 mirrors the first operation screen currently displayed on the display unit 251 so that the first operation screen is displayed on the vehicle display unit 141 (S1115). At this time, examples of the first operation screen displayed on the display unit 251 include an idle screen, a lock screen, a home screen, or a main menu screen.
If at least one application is executed through the selection of the application icon in the state where the first operation screen is being mirrored (S1120), the control unit 280 checks whether the mirroring screen switching function is set in advance (S1125) . Also, the control unit 280 may provide a menu for setting the mirroring screen switching function in advance.
If it is determined in step 1125 that the mirroring screen switching function has been set, the control unit 280 continues to display the first operation screen on the display unit 251 (S1130). Then, the control unit 280 executes the application in the background and mirrors the execution screen of the application (i.e., the second operation screen) to the vehicle display unit 141 (S1135).
If the mirroring screen switching command is received while the second operation screen is being mirrored (S1140), the control unit 280 may mirror the first operation screen to the vehicle display unit 141, The application execution screen (i.e., the third operation screen) is mirrored on the vehicle display unit 141 (S1145).
On the other hand, if it is determined in step 1125 that the mirroring screen switching function is not set, the control unit 280 displays the execution screen of the application (i.e., the second operation screen) on the display unit 251, 2 operation screen to the vehicle display unit 141 (S1150, S1155).
Thereafter, when a user command to terminate the mirroring service is received (S1160), the controller 280 terminates the mirroring service according to the present invention. On the other hand, if the user command to terminate the mirroring service is not received (S1160), the controller 280 repeatedly performs the operations of steps 1140 and 1145 described above.
As described above, the mobile terminal according to the present invention can mirror the operation screen different from the operation screen currently displayed on the display unit to the vehicle display unit, and can perform mirroring You can switch screens.
12 to 14 are diagrams for explaining the operation of the mobile terminal for switching the mirroring screen displayed on the vehicle display unit when a user input event occurs during car mirroring.
Referring to FIG. 12, the mobile terminal 200 may initiate a mirroring service after connecting a short-range wireless communication with the vehicle 100 according to a user command. Accordingly, the mobile terminal 200 can mirror the application screen 1210 displayed on the display unit 251 to the display unit 141 mounted on the vehicle 100. At this time, it is assumed that the application screen 1210 includes a navigation icon 1215.
When the navigation icon 1215 displayed on the display unit 251 of the mobile terminal 200 or the display unit 141 of the vehicle 100 is selected while the application screen 1210 is being mirrored, ) Can check whether or not the mirroring screen switching function is set in advance.
As a result of the checking, if the mirroring screen switching function is set, the mobile terminal 200 executes the navigation application corresponding to the selected icon 1215 as shown in FIG. 12 (b) (That is, a navigation screen 1220) of the vehicle display unit 141 to the vehicle display unit 141. In addition, when the navigation screen 1220 is mirrored, the mobile terminal 200 can continue to display the original app screen 1210 on the display unit 251.
If the mirroring screen switching function is not set, the mobile terminal 200 displays the navigation screen on the display unit 251 and displays the navigation screen on the vehicle display unit (not shown) 141). ≪ / RTI >
13, when a flickering input 1230 having a directivity is received through the vehicle display unit 141 while the navigation screen 1220 is being mirrored, the vehicle 100 responds to the flicking input And transmits the control signal to the mobile terminal 200.
The mobile terminal 200 may mirror the application screen 1210 to the vehicle display unit 141 based on the received control signal. That is, the mobile terminal 200 can switch the mirroring screen displayed on the vehicle display unit 141 from the navigation screen 1220 to the application screen 1210 according to the user's flicking input.
When there is another application running in the background in addition to the navigation application, the mobile terminal 200 displays the mirroring screen displayed on the vehicle display unit 141 in the background in accordance with the additional flicking input of the user You can switch to the operation screen of another application. In addition, the mobile terminal 200 can continue to display the same app screen 1210 on the display unit 251 regardless of the switching of the mirroring screen.
In the present exemplary embodiment, the mirroring screen switching function is performed through the user's flicking input, but the present invention is not limited thereto. Therefore, it will be apparent to those skilled in the art that instead of the flickering input, a specific icon displayed on the vehicle display unit may be selected or a mirroring screen switching function may be performed through a voice command or the like.
14, when the app list menu 1225 displayed on the vehicle display unit 141 is selected while the navigation screen 1220 is being mirrored, the mobile terminal 200 displays an application The application list screen including the operation screens 1230, 1240, and 1250 corresponding to the application list screen may be mirrored on the vehicle display unit 141. [ The application list menu 1225 can be automatically displayed on the vehicle display unit 141 when the number of applications running in the background is two or more in the mobile terminal 200. [
The mobile terminal 200 can mirror the selected operation screen 1250 to be displayed on the vehicle display unit 141 when any one of the operation screens 1250 displayed on the app list screen is selected. That is, the mobile terminal 200 can switch the mirroring screen displayed on the vehicle display unit 141 through user input to the application list menu 1225. [
In addition, the mobile terminal 200 can continue to display the same app screen 1210 on the display unit 251 regardless of the switching of the mirroring screen.
15 is a flowchart illustrating an operation of a mobile terminal according to another embodiment of the present invention.
Referring to FIG. 15, the control unit 280 of the mobile terminal 200 searches for nearby connectable devices and displays them on the display unit 251 when a mirroring service is requested (S1505). When the vehicle 100 is selected among the devices displayed on the display unit 250, the control unit 280 connects the selected vehicle 100 to the near field wireless communication (S1510).
The control unit 280 mirrors the first operation screen currently displayed on the display unit 251 to be displayed on the vehicle display unit 141 (S1515). At this time, examples of the first operation screen displayed on the display unit 251 include an idle screen, a lock screen, a home screen, or a main menu screen.
When the first operation screen is being mirrored and two or more applications are executed simultaneously or sequentially through the selection of the application icon (S1520), the controller 280 checks whether the mirroring screen division function is set in advance (S1525). In addition, the control unit 280 may provide a menu for presetting the mirroring screen division function.
If it is determined in step 1525 that the mirroring screen division function has been set, the control unit 280 continues to display the first operation screen on the display unit 251 (S1530). The control unit 280 executes the two or more applications in the background, and mirrors the application split screen including operation screens corresponding to the applications to the vehicle display unit 141 (S1535).
If the preset user input is received through the vehicle display unit 141 while the application split screen is being mirrored in operation S1540, the control unit 280 displays the app split screen displayed on the vehicle display unit 141 , Mirroring screen) (S1545). That is, the control unit 280 may adjust the size of at least one of the operation screens displayed on the application division screen, or may remove at least one of the operation screens displayed on the application division screen.
On the other hand, if it is determined in step 1525 that the mirroring screen division function is not set, the control unit 280 displays the operation screen of the most recently executed application on the display unit 251, Mirror unit 141 (S1550, S1555).
Thereafter, when a user command to terminate the mirroring service is received (S1560), the controller 280 terminates the mirroring service according to the present invention. On the other hand, if the user command to terminate the mirroring service is not received (S1560), the controller 280 repeatedly performs the operations of steps 1540 and 1545 described above.
As described above, according to the present invention, the mobile terminal can mirror the application split screen including the application screens corresponding to two or more applications on the vehicle display unit, and display it on the vehicle display unit according to the user command You can control the mirroring screen.
FIG. 16 is a diagram for describing an operation of a mobile terminal that mirrors an application split screen in which operating screens corresponding to a plurality of applications are arranged in a mirror display mode on a vehicle display unit.
Referring to FIG. 16, the mobile terminal 200 may initiate a mirroring service after connecting a short-range wireless communication with the vehicle 100 according to a user command. Accordingly, the mobile terminal 200 can mirror the application screen 1610 displayed on the display unit 251 to the display unit 141 mounted on the vehicle 100. At this time, it is assumed that the application screen 1610 includes a navigation icon 1611 and an internet icon 1613.
When the navigation icon 1611 and the Internet icon 1614 displayed on the display unit 251 of the mobile terminal 200 or the display unit 141 of the vehicle 100 are displayed simultaneously or simultaneously with the application screen 1610 being mirrored, When sequentially selected, the mobile terminal 200 can confirm whether or not the mirroring screen division function is set in advance.
If it is determined that the mirroring screen division function is set, the mobile terminal 200 executes applications corresponding to the selected icons 1611 and 1613 in the background, as shown in FIG. 16 (b) And mirror the application split screen 1620 including operation screens corresponding to the applications to the vehicle display unit 141. [
In addition, when the application split screen 1620 is mirrored, the mobile terminal 200 can continue to display the original app screen 1610 on the display unit 251. [ In another embodiment, the mobile terminal 200 may display a mirroring manager screen on the display unit 251 to configure a mirroring screen to be displayed on the vehicle display unit 141 before the application split screen mirroring. Accordingly, the mobile terminal 200 can mirror the application split screen configured on the mirroring manager screen to the vehicle display unit 141. [
On the other hand, although not shown in the drawing, if the mirroring screen dividing function is not set as a result of the checking, the mobile terminal 200 displays an operation screen corresponding to the Internet icon selected later on the display unit 251, The operation screen can be mirrored on the vehicle display unit 141. [
FIGS. 17 and 18 are views for explaining the operation of the mobile terminal for controlling the mirroring screen displayed on the vehicle display unit according to a predetermined user input.
17, a navigation icon, an internet icon, and a messenger icon (not shown) displayed on the display unit 251 of the mobile terminal 200 or the display unit 141 of the vehicle 100, The mobile terminal 200 executes the applications corresponding to the selected icons in the background and displays the first application split screen 1710 in which the operation screens corresponding to the applications are arranged, And can be mirrored on the display unit 141.
When the delete icon 1720 displayed on the vehicle display unit 141 is selected, the mobile terminal 200 displays the messenger screen 1713 corresponding to the selected delete icon 1720 on the vehicle display unit 141, . Accordingly, the mobile terminal 200 can mirror the second app split screen 1740, in which the navigation screen 1711 and the web page screen 1712 are rearranged, to the vehicle display unit 141.
18, when the pinch-out input 1740 is received through the messenger screen 1713 displayed on the vehicle display unit 141 while the first application split screen 1710 is being mirrored, The mobile terminal 200 may enlarge and display the size of the messenger screen 1713 according to the drag distance of the pinch-out input 1740. In addition, the mobile terminal 200 may reduce the sizes of the navigation screen 1711 and the web page screen 1712 according to the enlargement ratio of the messenger screen 1713.
Conversely, when a pinch-in input (not shown) is received through the messenger screen 1713 displayed on the vehicle display unit 141 while the first application split screen 1710 is being mirrored, The size of the messenger screen 1713 can be reduced according to the drag distance of the pinch-in input. Similarly, the mobile terminal 200 may enlarge and display the size of the navigation screen 1711 and the web page screen 1712 according to the reduction ratio of the messenger screen 1713.
19 is a diagram for describing an operation of a mobile terminal providing a user interface screen for setting a mirroring related function in advance.
Referring to FIG. 19, the mobile terminal 200 may display a network setting screen 1910 on the display unit 251 according to a user command or the like.
When the connection and file exchange menu 1915 is selected in the state that the network setting screen 1910 is displayed, the mobile terminal 200 displays a connection and file exchange setting screen 1920 as shown in FIG. 19 (b) Can be displayed on the display unit 251. At this time, the connection and file exchange setting screen 1920 may include an NFC menu 1921, a smart shared beam menu 1922, a mirror cast menu 1923, and the like.
When the setting icon 1924 of the mirror cast menu 1923 is selected, the mobile terminal 200 displays the mirror cast setting screen 1930 as shown in FIG. 19C on the display unit 251 can do.
The mirror cast setting screen 1930 includes a first menu 1931 for setting whether a corresponding event is to be mirrored when a notification event occurs, a second menu 1931 for setting whether to mirror the corresponding event when a call reception event occurs, A third menu 1933 for setting a mirroring screen switching function, a fourth menu 1934 for setting a mirroring screen dividing function, a fifth menu 1935 for setting a non-mirroring app list, ), And the like. Accordingly, the user of the mobile terminal 200 can preset the mirroring related function through the selection of the menus 1931 to 1935.
The present invention described above can be embodied as computer-readable codes on a medium on which a program is recorded. The computer readable medium includes all kinds of recording devices in which data that can be read by a computer system is stored. Examples of the computer readable medium include a hard disk drive (HDD), a solid state disk (SSD), a silicon disk drive (SDD), a ROM, a RAM, a CD-ROM, a magnetic tape, a floppy disk, , And may also be implemented in the form of a carrier wave (e.g., transmission over the Internet). Also, the computer may include a control unit 180 of the terminal. Accordingly, the above description should not be construed in a limiting sense in all respects and should be considered illustrative. The scope of the present invention should be determined by rational interpretation of the appended claims, and all changes within the scope of equivalents of the present invention are included in the scope of the present invention.