WO2022072482A1 - Video teleconferencing - Google Patents
Video teleconferencing Download PDFInfo
- Publication number
- WO2022072482A1 WO2022072482A1 PCT/US2021/052632 US2021052632W WO2022072482A1 WO 2022072482 A1 WO2022072482 A1 WO 2022072482A1 US 2021052632 W US2021052632 W US 2021052632W WO 2022072482 A1 WO2022072482 A1 WO 2022072482A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user input
- video
- response
- video conferencing
- outgoing
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/15—Conference systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
- H04N7/147—Communication arrangements, e.g. identifying the communication as a video-communication, intermediate storage of the signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
Definitions
- Systems and methods disclosed herein are directed to video conferencing control systems, methods, and applications that control or adjust the local components of a video teleconference station.
- Systems, methods, and computer-readable instructions are provided that at least partially control a video conferencing station.
- the systems, methods, and instructions receive a user input, alter an outgoing audio associated with the video conferencing station in response to the user input, and alter an outgoing video associated with the video conferencing station in response to the user input.
- Some examples also reduce an output volume of a loudspeaker associated with the video conferencing station in response to the user input.
- Various examples may also post a message via the video conferencing station in response to the user input.
- Certain examples reduce an output volume of a loudspeaker associated with the video conferencing station in response to the user input and post a message via the video conferencing station in response to the user input.
- a user-defined function may be executed in response to the user input.
- altering the outgoing video comprises disabling the camera.
- altering the outgoing video comprises at least one of blurring a background, replacing the background, or replacing an entire view of the outgoing video.
- altering the outgoing audio comprises muting a microphone.
- FIG. 1 illustrates an example system 100 that provides a control input to a video conferencing system
- FIG. 2 illustrates another example system 200 that provides a control input to a video conferencing system. DESCRIPTION
- aspects of the present disclosure are directed to systems and methods for applying quick control of video conferencing gear and features such as microphone(s), camera(s), loudspeaker(s), messaging, backgrounds, etc.
- aspects and examples of systems and methods disclosed herein include physical user interface controls that could be on a remote control or conferencing accessory product that would provide easy, one-touch access to frequently used remote or video conferencing meeting controls and dedicated button(s) or other sensors to perform multiple meeting management actions at once.
- the controls, buttons, sensors, and the like may be “virtual” such as a “button” on a screen that may be selected by a user by “clicking” on it with a pointer or cursor controlled by a user input device, such as a computer mouse, touchscreen, or the like.
- FIG. 1 illustrates a system 100 that includes a docking capability to hold a personal portable device, such as a smart phone or tablet, to thereby act as, or to otherwise be coupled to, a video conferencing station.
- FIG. 2 illustrates another system 200 in the form of an accessory that couples to a video conferencing station, which may be any computing device having software to perform the functions of a video conferencing station, such as a personal computer, tablet, smart phone, etc.
- the system 200 of FIG. 2 is shown in a “puck” form factor but could be in any of numerous physical form factors, including, for example without limitation, handheld form factors, desktop form factors, etc.
- buttons and/or sensors such as touch / capacitive sensors, configured to control various functions like volume up/down, mute, answer, hang-up, camera on/off, background effect(s) on/off, etc.
- the buttons may be dedicated to their respective configured functionality.
- the buttons may be assignable to various functions, which may be user-assigned in various examples.
- Each of the system 100 and the system 200 includes a multifunction button 110 that may be coupled to a processor and configured to control or change numerous features and/or functions of the video conferencing station.
- the system 100, 200 is configured to respond to user activation of the button 110 by simultaneously (or nearly simultaneously) altering the operation of multiple functions of the video conferencing station at once, e.g., when a user needs to turn off audio and video in a hurry, such as may be desirable if something happens in the background that other participants should not see or hear.
- the system 100, 200 is configured to respond to the button 110 by muting a microphone, turning off a camera, and displaying a message, such as “be right back,” into a chat feature and/or as a background or video feed to the video conferencing system.
- the message may be a status indicator, or may be a response triggered by activity of other participant(s) in the video conference.
- the message may be user defined or user selectable in various examples.
- Such a functionality may be described as a “bio-break” feature that allows a participant (a user of the video conferencing station) to easily excuse themselves from the video conference for a brief time.
- a second activation of the button 110 may return each of the camera, microphone, loudspeaker, etc. to its prior state.
- a time delay may be enforced by the system 100, 200 to ensure that an erroneous second activation doesn’t trigger the return to prior operating state.
- activation of the button 110 may be configured to provide effortless shortcuts, which may be configurable by a user.
- each of the system 100 and the system 200 may be configured to allow a user to program multiple actions or functions to be executed upon activation of the button 110, such as a macro programming to be executed.
- the button 110 is configured to be available for discrete activation, e.g., without needing a user to make large physical movements to do so.
- the button 110 may be of a size, location, and labelling to enable quick and easy acquisition (visual sighting) and physical activation by the user, in a manner that minimizes the possibility of other participants noticing the activation in a video feed.
- system 100, 200 may be configured to respond to the button 110 by obscuring the background environment of the user, such as by activating a background blurring feature of the video conferencing station or activating a background image in lieu of a true background provided in a video feed from the video camera.
- the button 110 may be of any suitable physical or virtual type.
- the button 110 may be a physical button or switch or may be another type of user input sensor, such as a capacitive touch sensor, an optical sensor, a gesture control, etc.
- the button 110 may be a virtual button, such as an active location on a screen that may be selected by a user, such as by a mouse, touchscreen, or other computer input device.
- the button 110 may take the form of an icon on a display screen capable of being selected or activated by a user.
- Example systems 100, 200 and the like having a button 110 configured to activate or alter numerous functions with a single activation, whether pre-configured or user-definable, save time, mental effort, and potential embarrassment or social consequences by making it easier to locate and execute video conferencing meeting management functions.
- references to “or” may be construed as inclusive so that any terms described using “or” may indicate any of a single, more than one, and all of the described terms. Any references to front and back, left and right, top and bottom, upper and lower, and vertical and horizontal are intended for convenience of description, not to limit the present systems and methods or their components to any one positional or spatial orientation, unless the context reasonably implies otherwise.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Telephonic Communication Services (AREA)
Abstract
Systems, methods, and computer instructions are provided for controlling a video teleconference station. A single user interaction is received as an activating user input and in response thereto each of an outgoing audio signal and outgoing video signal are altered. Each of a camera and a microphone may be disabled or muted, for example. Additional functionality may be associated with the user input, which may be a button or other sensor in various examples.
Description
VIDEO TELECONFERENCING
CROSS REFERENCE TO RELATED APPLICATIONS
This application claims the benefit of U.S. Patent Application Serial Number 63/086,354, filed on October 1, 2020, titled “VIDEO TELECONFERENCING” the content of which is incorporated herein in its entirety for all purposes.
BACKGROUND
With the increase in working from home, such as due to the Covid- 19 pandemic, video conferencing usage has increased substantially. People are finding that at home distractions such as barking dogs, screaming and playing kids, spouses, partners, or roommates walking around in the background, or ringing doorbells, and the like, make it necessary to quickly and easily assert control over aspects of participating in a video conference, such as by controlling the microphone, speaker, and video camera to avoid the social consequences of having undesirable or distracting activities seen or heard by others on the video conference call.
SUMMARY
Systems and methods disclosed herein are directed to video conferencing control systems, methods, and applications that control or adjust the local components of a video teleconference station.
Systems, methods, and computer-readable instructions are provided that at least partially control a video conferencing station. The systems, methods, and instructions receive a user input, alter an outgoing audio associated with the video conferencing station in response to the user input, and alter an outgoing video associated with the video conferencing station in response to the user input.
Some examples also reduce an output volume of a loudspeaker associated with the video conferencing station in response to the user input.
Various examples may also post a message via the video conferencing station in response to the user input.
Certain examples reduce an output volume of a loudspeaker associated with the video conferencing station in response to the user input and post a message via the video conferencing station in response to the user input.
In some examples, a user-defined function may be executed in response to the user input.
According to various examples, altering the outgoing video comprises disabling the camera.
In certain examples altering the outgoing video comprises at least one of blurring a background, replacing the background, or replacing an entire view of the outgoing video.
According to some examples, altering the outgoing audio comprises muting a microphone.
Still other aspects, examples, and advantages of these exemplary aspects and examples are discussed in detail below. Examples disclosed herein may be combined with other examples in any manner consistent with at least one of the principles disclosed herein, and references to “an example,” “some examples,” “an alternate example,” “various examples,” “one example” or the like are not necessarily mutually exclusive and are intended to indicate that a particular feature, structure, or characteristic described may be included in at least one example. The appearances of such terms herein are not necessarily all referring to the same example.
BRIEF DESCRIPTION OF THE DRAWINGS
Various aspects of at least one example are discussed below with reference to the accompanying figures, which are not intended to be drawn to scale. The figures are included to provide illustration and a further understanding of the various aspects and examples and are incorporated in and constitute a part of this specification but are not intended as a definition of the limits of the invention(s). In the figures, identical or nearly identical components illustrated in various figures may be represented by a like reference character or numeral. For purposes of clarity, not every component may be labeled in every figure. In the figures:
FIG. 1 illustrates an example system 100 that provides a control input to a video conferencing system; and
FIG. 2 illustrates another example system 200 that provides a control input to a video conferencing system.
DESCRIPTION
Aspects of the present disclosure are directed to systems and methods for applying quick control of video conferencing gear and features such as microphone(s), camera(s), loudspeaker(s), messaging, backgrounds, etc.
Aspects and examples of systems and methods disclosed herein include physical user interface controls that could be on a remote control or conferencing accessory product that would provide easy, one-touch access to frequently used remote or video conferencing meeting controls and dedicated button(s) or other sensors to perform multiple meeting management actions at once. In various examples, the controls, buttons, sensors, and the like may be “virtual” such as a “button” on a screen that may be selected by a user by “clicking” on it with a pointer or cursor controlled by a user input device, such as a computer mouse, touchscreen, or the like.
FIG. 1 illustrates a system 100 that includes a docking capability to hold a personal portable device, such as a smart phone or tablet, to thereby act as, or to otherwise be coupled to, a video conferencing station. FIG. 2 illustrates another system 200 in the form of an accessory that couples to a video conferencing station, which may be any computing device having software to perform the functions of a video conferencing station, such as a personal computer, tablet, smart phone, etc. The system 200 of FIG. 2 is shown in a “puck” form factor but could be in any of numerous physical form factors, including, for example without limitation, handheld form factors, desktop form factors, etc. Each of the system 100 and the system 200 includes various buttons and/or sensors, such as touch / capacitive sensors, configured to control various functions like volume up/down, mute, answer, hang-up, camera on/off, background effect(s) on/off, etc. In various examples, the buttons may be dedicated to their respective configured functionality. In other examples, the buttons may be assignable to various functions, which may be user-assigned in various examples.
Each of the system 100 and the system 200 includes a multifunction button 110 that may be coupled to a processor and configured to control or change numerous features and/or functions of the video conferencing station. In certain examples, the system 100, 200 is configured to respond to user activation of the button 110 by simultaneously (or nearly simultaneously) altering the operation of multiple functions of the video conferencing station at once, e.g., when a user needs to turn off audio and video in a hurry, such as may be desirable if something happens in the background that other participants should not see or hear.
In various examples, the system 100, 200 is configured to respond to the button 110 by muting a microphone, turning off a camera, and displaying a message, such as “be right back,” into a chat feature and/or as a background or video feed to the video conferencing system. In some examples, the message may be a status indicator, or may be a response triggered by activity of other participant(s) in the video conference. The message may be user defined or user selectable in various examples. Such a functionality may be described as a “bio-break” feature that allows a participant (a user of the video conferencing station) to easily excuse themselves from the video conference for a brief time. A second activation of the button 110 may return each of the camera, microphone, loudspeaker, etc. to its prior state. In some examples, a time delay may be enforced by the system 100, 200 to ensure that an erroneous second activation doesn’t trigger the return to prior operating state.
In some examples, activation of the button 110 may be configured to provide effortless shortcuts, which may be configurable by a user. In some examples, each of the system 100 and the system 200 may be configured to allow a user to program multiple actions or functions to be executed upon activation of the button 110, such as a macro programming to be executed.
In various examples, the button 110 is configured to be available for discrete activation, e.g., without needing a user to make large physical movements to do so. For instance, the button 110 may be of a size, location, and labelling to enable quick and easy acquisition (visual sighting) and physical activation by the user, in a manner that minimizes the possibility of other participants noticing the activation in a video feed.
In various examples, the system 100, 200 may be configured to respond to the button 110 by obscuring the background environment of the user, such as by activating a background blurring feature of the video conferencing station or activating a background image in lieu of a true background provided in a video feed from the video camera.
The button 110 may be of any suitable physical or virtual type. For example, the button 110 may be a physical button or switch or may be another type of user input sensor, such as a capacitive touch sensor, an optical sensor, a gesture control, etc. In some instances the button 110 may be a virtual button, such as an active location on a screen that may be selected by a user, such as by a mouse, touchscreen, or other computer input device. Accordingly, the button 110 may take the form of an icon on a display screen capable of being selected or activated by a user.
Example systems 100, 200 and the like having a button 110 configured to activate or alter numerous functions with a single activation, whether pre-configured or user-definable, save time, mental effort, and potential embarrassment or social consequences by making it easier to locate and execute video conferencing meeting management functions.
Examples of the methods and apparatuses discussed herein are not limited in application to the details of construction and the arrangement of components set forth in the above descriptions or illustrated in the accompanying drawings. The methods and apparatuses are capable of implementation in other examples and of being practiced or of being carried out in various ways. Examples of specific implementations are provided herein for illustrative purposes only and are not intended to be limiting. In particular, functions, components, elements, and features discussed in connection with any one or more examples are not intended to be excluded from a similar role in any other examples.
Also, the phraseology and terminology used herein is for the purpose of description and should not be regarded as limiting. Any references to examples, components, elements, acts, or functions of the systems and methods herein referred to in the singular may also embrace embodiments including a plurality, and any references in plural to any example, component, element, act, or function herein may also embrace examples including only a singularity. Accordingly, references in the singular or plural form are not intended to limit the presently disclosed systems or methods, their components, acts, or elements. The use herein of “including,” “comprising,” “having,” “containing,” “involving,” and variations thereof is meant to encompass the items listed thereafter and equivalents thereof as well as additional items. References to “or” may be construed as inclusive so that any terms described using “or” may indicate any of a single, more than one, and all of the described terms. Any references to front and back, left and right, top and bottom, upper and lower, and vertical and horizontal are intended for convenience of description, not to limit the present systems and methods or their components to any one positional or spatial orientation, unless the context reasonably implies otherwise.
Having described above several aspects of at least one example, it is to be appreciated various alterations, modifications, and improvements will readily occur to those skilled in the art. Such alterations, modifications, and improvements are intended to be part of this disclosure and are intended to be within the scope of the invention. Accordingly, the foregoing description and drawings are by way of example only, and the scope of the invention should be determined from proper construction of the appended claims, and their equivalents.
Claims
1. A method of controlling a video conferencing station, the method comprising: receiving an indication of an activation of a user input; altering an outgoing audio associated with the video conferencing station in response to the user input; and altering an outgoing video associated with the video conferencing station in response to the user input.
2. The method of claim 1 further comprising reducing an output volume of a loudspeaker associated with the video conferencing station in response to the user input.
3. The method of claim 1 further comprising posting a message via the video conferencing station in response to the user input.
4. The method of claim 1 further comprising: reducing an output volume of a loudspeaker associated with the video conferencing station in response to the user input; and posting a message via the video conferencing station in response to the user input.
5. The method of claim 1 further comprising executing a user-defined function in response to the user input.
6. The method of claim 1 wherein altering the outgoing video comprises disabling the camera.
7. The method of claim 1 wherein altering the outgoing video comprises at least one of blurring a background, replacing the background, or replacing an entire view of the outgoing video.
6
8. The method of claim 1 wherein altering the outgoing audio comprises muting a microphone.
9. An apparatus comprising: an enclosure; a processor within the enclosure; and a user input sensor coupled to the processor, wherein the processor is configured to detect activation of the user input sensor and to control a video conferencing station to alter an outgoing audio and alter an outgoing video in response to the detected activation of the user input sensor.
10. The apparatus of claim 9 wherein the processor is further configured to control the video conferencing station to reduce an output volume of a loudspeaker.
11. The apparatus of claim 9 wherein the processor is further configured to control the video conferencing station to post a message viewable by remote participants.
12. The apparatus of claim 9 wherein the processor is further configured to control the video conferencing station to reduce an output volume of a loudspeaker and to post a message viewable by remote participants.
13. The apparatus of claim 9 wherein the processor is further configured to execute a user-defined function.
14. The apparatus of claim 9 wherein altering the outgoing video comprises disabling the camera.
15. The apparatus of claim 9 wherein altering the outgoing video comprises at least one of blurring a background, replacing the background, or replacing an entire view of the outgoing video.
16. The apparatus of claim 9 wherein altering the outgoing audio comprises muting a microphone.
7
17. A non-transitory machine readable storage medium having instructions encoded thereon that, when executed by a processor, cause the processor to perform a method comprising: receiving an indication of an activation of a user input; altering an outgoing audio associated with the video conferencing station in response to the user input; and altering an outgoing video associated with the video conferencing station in response to the user input.
18. The non-transitory storage medium of claim 17 further comprising instructions to cause the processor to reduce an output volume of a loudspeaker associated with the video conferencing station in response to the user input.
19. The non-transitory storage medium of claim 17 further comprising instructions to cause the processor to post a message via the video conferencing station in response to the user input.
20. The non-transitory storage medium of claim 17 wherein the method further comprises: reducing an output volume of a loudspeaker associated with the video conferencing station in response to the user input; and posting a message via the video conferencing station in response to the user input.
21. The non-transitory storage medium of claim 17 further comprising instructions to cause the processor to execute a user-defined function in response to the user input.
22. The non-transitory storage medium of claim 17 wherein altering the outgoing video comprises disabling the camera.
23. The non-transitory storage medium of claim 17 wherein altering the outgoing video comprises at least one of blurring a background, replacing the background, or replacing an entire view of the outgoing video.
8
24. The non-transitory storage medium of claim 17 wherein altering the outgoing audio comprises muting a microphone.
9
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US202063086354P | 2020-10-01 | 2020-10-01 | |
US63/086,354 | 2020-10-01 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022072482A1 true WO2022072482A1 (en) | 2022-04-07 |
Family
ID=78592908
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2021/052632 WO2022072482A1 (en) | 2020-10-01 | 2021-09-29 | Video teleconferencing |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2022072482A1 (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120154509A1 (en) * | 2010-12-16 | 2012-06-21 | Mitel Networks Corporation | Method and system for audio-video communications |
US20130002800A1 (en) * | 2011-06-28 | 2013-01-03 | Mock Wayne E | Muting a Videoconference Using Touch-Based Gestures |
EP2797279B1 (en) * | 2013-04-23 | 2016-09-14 | Gurulogic Microsystems OY | Communications using at least two different media types |
US20170374118A1 (en) * | 2016-06-23 | 2017-12-28 | Ringcentral, Inc. | Conferencing system and method implementing video quasi-muting |
US20200028883A1 (en) * | 2018-07-18 | 2020-01-23 | International Business Machines Corporation | Enhanced teleconferencing using noise filtering, amplification, and selective muting |
-
2021
- 2021-09-29 WO PCT/US2021/052632 patent/WO2022072482A1/en active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120154509A1 (en) * | 2010-12-16 | 2012-06-21 | Mitel Networks Corporation | Method and system for audio-video communications |
US20130002800A1 (en) * | 2011-06-28 | 2013-01-03 | Mock Wayne E | Muting a Videoconference Using Touch-Based Gestures |
EP2797279B1 (en) * | 2013-04-23 | 2016-09-14 | Gurulogic Microsystems OY | Communications using at least two different media types |
US20170374118A1 (en) * | 2016-06-23 | 2017-12-28 | Ringcentral, Inc. | Conferencing system and method implementing video quasi-muting |
US20200028883A1 (en) * | 2018-07-18 | 2020-01-23 | International Business Machines Corporation | Enhanced teleconferencing using noise filtering, amplification, and selective muting |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11586340B2 (en) | Terminal and method for setting menu environments in the terminal | |
US9473611B2 (en) | Use of proximity sensors for interacting with mobile devices | |
KR101892916B1 (en) | Do-not-disturb modes | |
US8090087B2 (en) | Method, system, and graphical user interface for making conference calls | |
CN107124508B (en) | Position adjusting method and device of suspension control, terminal and readable storage medium | |
CN118201133A (en) | Method, apparatus and medium for point-to-point ad hoc voice communication | |
US9953101B1 (en) | Customized home screens for electronic devices | |
EP3047356A1 (en) | Configuration of a touch screen display with conferencing | |
CN106371688A (en) | Full-screen single-hand operation method and apparatus | |
KR20130115174A (en) | Apparatus and method for providing a digital bezel | |
KR20120029412A (en) | In-call contact information display | |
JP6134870B2 (en) | Description Information Display Method, Description Information Display Device, Electronic Device, Program, and Recording Medium | |
US20180285051A1 (en) | System for controlling a display device | |
JP2017511947A (en) | Button interaction method, apparatus, program, and recording medium | |
US20200285381A1 (en) | User interface for use in computing device with sensitive display | |
EP3430802A1 (en) | Selectable interaction elements in a 360-degree video stream | |
CN110119239B (en) | Input method application display method and device | |
CN107302724A (en) | A kind of video playing control method, device and electronic equipment | |
CN110187947A (en) | A kind of message display method and terminal device | |
US20230319413A1 (en) | User interfaces for camera sharing | |
CN111610912B (en) | Application display method, application display device and storage medium | |
CN112764710A (en) | Audio playing mode switching method and device, electronic equipment and storage medium | |
US20160014577A1 (en) | System and Method for Initiating Communication from Actual, Notional, or Dissociated Previewed Images | |
WO2014045235A1 (en) | Handheld information processing device with remote control output mode | |
WO2022072482A1 (en) | Video teleconferencing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21806060 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21806060 Country of ref document: EP Kind code of ref document: A1 |