US20020027976A1 - System for recording, annotating and indexing audio data - Google Patents
System for recording, annotating and indexing audio data Download PDFInfo
- Publication number
- US20020027976A1 US20020027976A1 US09/058,283 US5828398A US2002027976A1 US 20020027976 A1 US20020027976 A1 US 20020027976A1 US 5828398 A US5828398 A US 5828398A US 2002027976 A1 US2002027976 A1 US 2002027976A1
- Authority
- US
- United States
- Prior art keywords
- audio data
- property
- user
- audio
- keyword
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 claims description 9
- 230000011218 segmentation Effects 0.000 claims description 8
- 230000009471 action Effects 0.000 claims description 6
- 230000004048 modification Effects 0.000 claims description 3
- 238000012986 modification Methods 0.000 claims description 3
- 238000012216 screening Methods 0.000 claims description 2
- 238000004891 communication Methods 0.000 claims 2
- 230000002452 interceptive effect Effects 0.000 claims 2
- 230000000977 initiatory effect Effects 0.000 description 8
- 230000003068 static effect Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000008520 organization Effects 0.000 description 2
- 238000012552 review Methods 0.000 description 2
- 238000004040 coloring Methods 0.000 description 1
- 238000013479 data entry Methods 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/64—Automatic arrangements for answering calls; Automatic arrangements for recording messages for absent subscribers; Arrangements for recording conversations
- H04M1/65—Recording arrangements for recording a message from the calling party
- H04M1/6505—Recording arrangements for recording a message from the calling party storing speech in digital form
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/34—Indicating arrangements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/57—Arrangements for indicating or recording the number of the calling subscriber at the called subscriber's set
- H04M1/575—Means for retrieving and displaying personal data about calling party
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/64—Automatic arrangements for answering calls; Automatic arrangements for recording messages for absent subscribers; Arrangements for recording conversations
- H04M1/65—Recording arrangements for recording a message from the calling party
- H04M1/656—Recording arrangements for recording a message from the calling party for recording conversations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/42221—Conversation recording systems
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/21—Disc-shaped record carriers characterised in that the disc is of read-only, rewritable, or recordable type
- G11B2220/213—Read-only discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/25—Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
- G11B2220/2508—Magnetic discs
- G11B2220/2512—Floppy disks
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/25—Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
- G11B2220/2537—Optical discs
- G11B2220/2545—CDs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/60—Solid state media
- G11B2220/61—Solid state media wherein solid state memory is used for storing A/V content
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2201/00—Electronic components, circuits, software, systems or apparatus used in telephone systems
- H04M2201/38—Displays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/42025—Calling or Called party identification service
- H04M3/42034—Calling party identification service
- H04M3/42042—Notifying the called party of information on the calling party
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/42025—Calling or Called party identification service
- H04M3/42034—Calling party identification service
- H04M3/42059—Making use of the calling party identifier
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/436—Arrangements for screening incoming calls, i.e. evaluating the characteristics of a call before deciding whether to answer it
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/50—Centralised arrangements for answering calls; Centralised arrangements for recording messages for absent or busy subscribers ; Centralised arrangements for recording messages
- H04M3/53—Centralised arrangements for recording incoming messages, i.e. mailbox systems
- H04M3/533—Voice mail systems
Definitions
- This invention is directed to a system for recording, annotating and indexing audio data.
- the invention is directed to an audio system that improves the way users record, review and retrieve audio data.
- the '311 application discloses an electronic notebook that improves the way users record, review, retrieve and reuse their personal notes.
- the electronic notebook includes an interface that is similar to a standard paper notebook. That is, the user enters freeform hand-written notes onto a screen having the appearance of a page into a digital notebook.
- the electronic notebook also has features that augment standard note-taking capabilities.
- the electronic notebook allows a property, such as “action item” or “name” to be associated with the note marks so that notes with these properties may be more easily retrieved.
- the electronic notebook allows the user to assign text keywords to a page of notes. Finally, the electronic notebook records the audio data and synchronizes the audio data with the note marks. However, only those portions of the audio data that is highlighted by the user are permanently stored.
- These features of the electronic notebook provide the ability to have multiple views on a notebook. These multiple view arc analogous to views available in a database system. For example, requesting notes from last week with the property “action item” and keywords “mobile” and “computing” causes the notebook to display only those note marks from a previous week which were marked as action items and which were on the topic of mobile computing. Once a notebook's views are defined they may be reused. This provides the user with dynamic organization of the changing content of the notebook.
- the note marks are linked to the audio data, so that the user can play the audio data that was recorded at the time a particular note was written. Audio data is displayed with an audio timeline at the top of the note page. Important audio data can be highlighted. Audio data highlights are shown as colored bands in the audio display that can be selected for playback.
- Note marks can be indexed using ink properties. Ink properties are data types that are associated with particular types of ink marks. The ink properties arc indicated to the user by ink color and/or mark size.
- pages or note marks can be indexed by keywords. Note marks can be retrieved by queries using properties and keywords. Property queries retrieve only those note marks having the specified property, while keyword queries retrieve an entire page of note marks. The electronic notebook disclosed in the '311 application, however, does allow access to the audio data without referring to the associated note marks.
- the Apple Newton® places phone calls, and provides a note page to take notes on the conversation. However, the conversation is not recorded.
- the Apple Newton® also records messages and sounds but does not associate properties, text or ink to the recording.
- This invention provides a method and apparatus that uses keywords and properties to index audio data.
- This invention further provides a method and apparatus that separates the sides of a conversation in the audio data to improve access.
- This invention additionally provides a method and apparatus for recording, indexing and retrieving voice mail messages.
- the method and apparatus of this invention applies properties to the audio data and generates indexes to the audio data. This indexing allows retrieval of specific audio data and portions of audio data.
- This invention allows users to take notes using a computer while recording the audio data of a telephone call.
- the two sides of the conversation are separated and shown in an audio display for easy access to specific comments using the associated audio data attributes, such as date and time, properties, keywords, and identity of the caller. These associated attributes are used to index the audio data for subsequent retrieval. Properties apply to the audio data as well as to note marks to provide access to specific remarks in the audio data.
- This invention allows a user to play specific comments from previous calls during a current conversation. This is useful when the tone of voice used in a call conveys information that is not available in the associated note marks or text.
- Properties for the audio data can be set without selecting or writing note marks by, for example, pressing a button for the desired property.
- the property of the audio data is displayed to the user by coloring the band for the current speaker turn in the audio bar.
- the method and apparatus of this invention assume that the extent of each speaker's turn corresponds to the extent of the selected audio data property.
- the method and apparatus of this invention route calls based upon a mode. Depending on the mode, the method and apparatus of this invention send calls to an answering machine, accept a call for conversation, or tentatively accept a call based on the identity of the caller.
- FIG. 1 is a block diagram of one embodiment of an audio data system according to this invention.
- FIG. 2 shows one embodiment of the audio data system incorporated into a telephone set in accordance with this invention
- FIG. 3 illustrates a user interface of one embodiment of the audio data system according to this invention
- FIG. 4 illustrates a sample retrieval query window of the audio data system according to this invention.
- FIG. 5 illustrates a sample table of contents of retrieved audio data files according to this invention.
- FIG. 1 is a block diagram of one embodiment of an audio data system 10 according to this invention.
- the audio data system 10 has a processor 12 communicating with a storage device 14 , a memory 16 and an input/output interface 18 .
- a user interacts with and controls the audio data system 10 through any number of conventional input/output devices, such as a pen 20 , a keyboard 22 , a mouse 24 , a microphone 26 , an audio data input device 28 , a display 30 and a speaker 32 .
- the input/output devices 20 - 32 communicate with the input/output interface 18 that, in turn, communicates with the processor 12 .
- the audio data system 10 is preferably implemented on a programmed general purpose computer. However, the system 10 can also be implemented using a specific purpose computer, a programmed microprocessor or microcontroller and any necessary peripheral integrated circuit elements, an ASIC or other integrated circuit, a hardwired electronic or logic circuit such as a discrete element circuit, a digital signal processor, a programmable logic device such as a PLD, PLA, FPGA, or PAL, or the like.
- An alternate portion of the memory 16 is preferably implemented using static or dynamic RAM, and stores executing application programs and data.
- the alterable portion of the memory 16 can also be implemented using a floppy disk and drive disk, a writable optical disk and disk drive, a hard drive, flash memory or the like.
- a fixed portion of the memory 16 is preferably implemented using a ROM and stores basic control programs for the audio data system.
- the fixed portion of the memory can also be implemented using a PROM, an EPROM, an EEPROM, or a CD-ROM and disk drive or the like.
- the storage device 14 is preferably implemented using a hard drive, and provides long-term storage for application programs and data. However, the storage device 14 can also be implemented using a floppy disk and disk drive, a writable optical disk and disk drive, flash memory, static or dynamic RAM, or the like.
- the telephone system connector 28 may be any type of connector that is capable of carrying audio data.
- the telephone system connector 28 can be a connection to a publicly or privately switched telephone network, a wired or wireless link to a data network.
- the data network can be a local area network, a wide area network, an intranet, the Internet or any other distributed processing and storage network.
- FIG. 2 shows one embodiment of the audio data system 10 of this invention that has been incorporated into a telephone 34 .
- the audio data system 10 in FIG. 2 has an audio data input device 28 that is adapted to receive audio data from a telephone system and a handset and/or a microphone 26 that have been built into the telephone for hands free operation.
- This audio data system 10 also includes a keypad 36 and a pen-based input display 20 which are used as a user interface to control the operation of the audio data system 10 .
- the audio data system 10 of this invention allows a user to indicate the type of information contained in audio data by attaching properties to the audio data.
- Properties indicate information types that are associated with the selected audio data. Properties are associated with portions of the audio data, not the entire audio data file, so that individual portions of the audio data can be easily identified. Audio data portions having a specific property from a number of audio data files can be grouped and listed together. Properties are useful for retrieval because they are a low overhead categorization that can be used to distinguish between audio data portions.
- the audio data system 10 of this invention records audio data and links the audio data to handwritten notes.
- the sides of the conversation in a telephone call are separated using a speaker segmentation algorithm.
- the resulting speaker segmentation is shown in the audio timeline bar 40 at the top of the note page 42 in FIG. 3.
- a local speaker is shown by the dark bands in the top bar 44 and another speaker is shown in the lower bar 46 .
- Speaker segmentation and display can be extended to any number of sides of a conversation. If the audio data system 10 is connected to a conventional telephone system an elaborate speaker segmentation algorithm may be avoided by relying on the fact that each party will be using separate channels. All incoming audio can be attributed to a remote speaker and all outgoing audio can be attributed to a local speaker.
- Audio data portions having different properties are distinguished from each other by the color of the timeline 44 . For example, audio data portions that do not have any assigned properties are shown in black, while audio data portions having “action item” as an assigned property are shown in green.
- the audio data system 10 allows properties to be applied to the audio data as well as to ink. Specific audio data can be selected and a property applied to that audio data.
- the property is indicated in the audio bar 40 by changing the color of the audio bar 40 to a color associated with that property. For example, the “to do” property may be indicated to the user in blue ink for the corresponding note marks and as a blue band in the audio bar 40 for the corresponding audio data.
- Properties can be assigned to the audio data portions either before or after the audio data is recorded. If the user knows beforehand that the audio data about to be recorded has a certain property, the user can select that property from a menu, assign it to the audio data and begin recording. Properties may be assigned after recording by selecting the audio and assigning a property to the selected audio. For example, toggling one of the property buttons 50 , a property corresponding to the property button is applied to the current turn in the conversation, indicating that this audio data portion has the given property.
- a property button 52 which does not assign a property
- a topic property button 54 a name property button 56 , a quote property button 58 , a phone property button 60 , a to do property button 62 , a Uniform Resource Locator (URL) property button 64 , an organization property button 66 and a user defined property button.
- Properties can be assigned to audio data and note marks that are currently being received or which have been selected by toggling the property button 50 which corresponds to the desired property.
- the audio data system 10 may also indicate an audio data property, even though there are no associated note marks with that property.
- the identity of a caller is used to index the audio data.
- This index can be obtained in several ways.
- the caller can be identified using a database linking a phone number or a caller id to the caller. If the caller is not calling from his own phone, the audio data system 10 of this invention uses a database of audio data files of known callers, and matches the voice from the incoming caller to an audio data file of a known caller. Any conventional speaker identification system may be used.
- a speaker identification system is disclosed in U.S. Pat. No. 5,655,058, herein incorporated by reference in its entirety.
- the user can also enter the caller's identity through any other conventional input method.
- the audio data can also be indexed automatically according to date and time.
- FIG. 3 shows a call initiation dialog window 70 on top of user interface 42 .
- a user can access the call initiation dialog window 70 by selecting the call initiation button 72 on the user interface 42 .
- the call initiation dialog window 70 includes a keypad 74 that enables the user to manually enter a telephone number.
- the call initiation dialog window 70 also includes a phonebook button 76 which provides access to a phonebook listing of telephone numbers (not shown).
- the call initiation dialog window 70 also has a dropdown text entry box 78 which enables a user to place a cursor in the text entry box 78 and enter text and/or numbers using a separate data input device such as a keyboard.
- the call initiation dialog window 70 has a cancel button 80 which enables a user to cancel a call initiation and also an “ok” button 82 which enables a user to initiate a call once the telephone number information has been entered.
- This invention also provides a system for call screening and/or routing based on speaker identity.
- a screen mode (not shown) of the audio data system 10 can be set so that all calls are automatically answered, all calls are automatically recorded without interaction (as in voice mail), or calls are routed in accordance with the identity of the caller.
- the audio data system 10 answers the call and sends a message to the caller to state his name and the purpose of the call. If the caller's identity matches a name on an “accept call” list, the user is notified of the call. Otherwise, the call is recorded as voice mail.
- the audio data system 10 of this invention allows keywords to be associated with audio data. This allows keyword-based retrieval of audio data.
- keywords describe the actual content of the audio data. Because keywords reflect more general information, they are assigned to the entire audio data file and not to particular portions of audio data. Keywords are displayed in the margins of a digital page (not shown) and can be entered in several ways. For example, keywords can be selected from a list of previously entered keywords. Alternately, keywords can be typed in using a pop-up keyboard (not shown). Additionally, keywords can be hand-written in a keyword entry box (not shown) and recognized using a standard handwriting recognition system.
- note marks can be selected and sent to a handwriting recognizer, and the resulting text entered as a keyword for the audio data associated with the note marks. If the user does not want to immediately interact with the handwriting recognition system, the user can associate the “keyword” property with digital ink to cause delayed recognition.
- the audio data can be retrieved using attributes associated with the audio data such as caller, time, and topic, using properties and keywords.
- the user interface 42 also has query buttons 112 that enable a user to access a context-based query (not shown) by selecting the context-based query button 114 or access a query window 84 as shown in FIG. 4 using the query button 116 .
- the query window 84 includes property check boxes 86 that enable a user to search based upon a designated property.
- the query window 84 also has a keyword entry box 88 that permits a user to search for audio data and notemarks having a specified keyword. Retrieval by keyword and name returns a list of audio data files with the identified caller and having the desired keyword.
- Retrieval by property is more specific, since it provides the audio data portions of calls and digital ink groups having the specified property. This is convenient for locating a particular comment in an audio data file. For example, if a comment is marked with the property “important”, that audio data segment containing the comment can be retrieved using a query with the “important” property designated. Segments of audio data files can also be played to both parties during a conversation. Thus, comments from previous conversations can be shared.
- Results from a search are shown in FIG. 5 in a search result window 94 .
- the search result window lists the attributes of the retrieved data files in columns. The columns indicate the creation date 96 , the date of last modification 98 , the type of property as indicated by icon 100 and the keywords 102 .
- the series of icons describe properties associated with the audio data files. For example, the telephone icon represents the property “phone number” and the check box represents the “action item” property.
- the final column displays keywords associated with the audio data files.
- the audio data system of this invention provides the user with multiple views of audio data stored in a notebook.
- the views provided by this invention are analogous to the views in a database system. If a query specifies a particular property, only those audio data portions having that property will be displayed. This view of the audio data is formed by newly generated audio data files, each having an original date and containing only the audio data portions with the specified property.
- the user can also create a view of a notebook by forming a query to request audio data files having specific keywords.
- the user can define the keywords in the query by browsing an alphabetized list of keywords that have been attached to pages, or by typing or writing the desired keywords.
- views are indicated by tabs 48 at the bottom of the notebook page.
- a name is associated with each tab 48 for easy reference.
- the user can return to the previous view by selecting the corresponding tab 48 .
- the user can decide if the view should be static or dynamic. When the view is static, the contents of the view remains the same. When the view is dynamic, new items are added to the view as they are recorded by the notebook, provided that these items satisfy the terms of the query for this view.
- a user can navigate through a notebook using the navigation buttons 104 in the user interface window 42 .
- Navigation buttons 106 and 108 allow a user to page forward and backward through pages in a notebook.
- Navigation Button 110 is a new page button that starts data entry on a fresh notebook page.
- the user interface 42 is also provided with mode select buttons 118 which enable a user to select either a page mode using a page mode select button 120 or a table of contents mode using the table of contents mode select button 122 .
- the user interface generally includes the features specified and described in co-pending U.S. patent application Ser. No. 08/821,311.
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Telephonic Communication Services (AREA)
- Management Or Editing Of Information On Record Carriers (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
- 1. Field of Invention
- This invention is directed to a system for recording, annotating and indexing audio data. In particular, the invention is directed to an audio system that improves the way users record, review and retrieve audio data.
- 2. Description of Related Art
- Many people take notes on their telephone conversations. Interviews with lawyers, managers, administrative assistants, salesmen and executives have shown that certain note-taking practices are used during phone calls. Some people record the date and time of the call, as well as the identity of the caller, some people keep a separate notebook by the phone for notes on their calls and others keep a special notebook for notes taken on voice mail messages. People take notes on information obtained during their calls, generate a list of “to do” items, and record other details such as sales orders and directions. If an administrative assistant answers a call for his boss and finds the message too complex, the call is transferred to the voice mail recording system.
- Recording a call increases the accuracy of the information that is obtained from the call. However, while voice mail systems can record phone conversations, they do not index the audio data for later access.
- U.S. patent application Ser. No. 08/821,311, filed Mar. 20, 1997, which is incorporated herein in its entirety, discloses a system that allows users to simultaneously record audio data while taking hand-written notes. The '311 application discloses an electronic notebook that improves the way users record, review, retrieve and reuse their personal notes. The electronic notebook includes an interface that is similar to a standard paper notebook. That is, the user enters freeform hand-written notes onto a screen having the appearance of a page into a digital notebook. The electronic notebook also has features that augment standard note-taking capabilities. The electronic notebook allows a property, such as “action item” or “name” to be associated with the note marks so that notes with these properties may be more easily retrieved. In order to further aid indexing, the electronic notebook allows the user to assign text keywords to a page of notes. Finally, the electronic notebook records the audio data and synchronizes the audio data with the note marks. However, only those portions of the audio data that is highlighted by the user are permanently stored.
- These features of the electronic notebook provide the ability to have multiple views on a notebook. These multiple view arc analogous to views available in a database system. For example, requesting notes from last week with the property “action item” and keywords “mobile” and “computing” causes the notebook to display only those note marks from a previous week which were marked as action items and which were on the topic of mobile computing. Once a notebook's views are defined they may be reused. This provides the user with dynamic organization of the changing content of the notebook.
- The note marks are linked to the audio data, so that the user can play the audio data that was recorded at the time a particular note was written. Audio data is displayed with an audio timeline at the top of the note page. Important audio data can be highlighted. Audio data highlights are shown as colored bands in the audio display that can be selected for playback. Note marks can be indexed using ink properties. Ink properties are data types that are associated with particular types of ink marks. The ink properties arc indicated to the user by ink color and/or mark size. In addition, pages or note marks can be indexed by keywords. Note marks can be retrieved by queries using properties and keywords. Property queries retrieve only those note marks having the specified property, while keyword queries retrieve an entire page of note marks. The electronic notebook disclosed in the '311 application, however, does allow access to the audio data without referring to the associated note marks.
- “Capturing, Structuring and Representing Ubiquitous Audio”, by D. Hindus et al., ACM Trans. Information Systems, Vol. II, No. 4, October 1993, pp. 376-400, discloses a system that records telephone calls while displaying the turn-taking structure of the conversation. During the telephone call, the user selects parts of the conversation to save. After the call, the user can add text tags to the stored parts of the conversation. However, this system does not permit the user to enter pen-based input. This system only permits typed input and then only after the telephone call.
- The Apple Newton® places phone calls, and provides a note page to take notes on the conversation. However, the conversation is not recorded. The Apple Newton® also records messages and sounds but does not associate properties, text or ink to the recording.
- This invention provides a method and apparatus that uses keywords and properties to index audio data.
- This invention further provides a method and apparatus that separates the sides of a conversation in the audio data to improve access.
- This invention additionally provides a method and apparatus for recording, indexing and retrieving voice mail messages.
- In particular, the method and apparatus of this invention, applies properties to the audio data and generates indexes to the audio data. This indexing allows retrieval of specific audio data and portions of audio data.
- This invention allows users to take notes using a computer while recording the audio data of a telephone call. The two sides of the conversation are separated and shown in an audio display for easy access to specific comments using the associated audio data attributes, such as date and time, properties, keywords, and identity of the caller. These associated attributes are used to index the audio data for subsequent retrieval. Properties apply to the audio data as well as to note marks to provide access to specific remarks in the audio data.
- This invention allows a user to play specific comments from previous calls during a current conversation. This is useful when the tone of voice used in a call conveys information that is not available in the associated note marks or text.
- Properties for the audio data can be set without selecting or writing note marks by, for example, pressing a button for the desired property. The property of the audio data is displayed to the user by coloring the band for the current speaker turn in the audio bar. The method and apparatus of this invention assume that the extent of each speaker's turn corresponds to the extent of the selected audio data property.
- The method and apparatus of this invention route calls based upon a mode. Depending on the mode, the method and apparatus of this invention send calls to an answering machine, accept a call for conversation, or tentatively accept a call based on the identity of the caller.
- These and other features and advantages of this invention are described in or are apparent from the following detailed description of the preferred embodiments.
- The preferred embodiments of this invention will be described in detail, with reference to the following figures, wherein:
- FIG. 1 is a block diagram of one embodiment of an audio data system according to this invention;
- FIG. 2 shows one embodiment of the audio data system incorporated into a telephone set in accordance with this invention;
- FIG. 3 illustrates a user interface of one embodiment of the audio data system according to this invention;
- FIG. 4 illustrates a sample retrieval query window of the audio data system according to this invention; and
- FIG. 5 illustrates a sample table of contents of retrieved audio data files according to this invention.
- These and other features and advantages of this invention are described in or are apparent from the following detailed description of the preferred embodiments.
- FIG. 1 is a block diagram of one embodiment of an
audio data system 10 according to this invention. Theaudio data system 10 has aprocessor 12 communicating with astorage device 14, amemory 16 and an input/output interface 18. A user interacts with and controls theaudio data system 10 through any number of conventional input/output devices, such as apen 20, akeyboard 22, amouse 24, amicrophone 26, an audiodata input device 28, adisplay 30 and aspeaker 32. The input/output devices 20-32 communicate with the input/output interface 18 that, in turn, communicates with theprocessor 12. - The
audio data system 10 is preferably implemented on a programmed general purpose computer. However, thesystem 10 can also be implemented using a specific purpose computer, a programmed microprocessor or microcontroller and any necessary peripheral integrated circuit elements, an ASIC or other integrated circuit, a hardwired electronic or logic circuit such as a discrete element circuit, a digital signal processor, a programmable logic device such as a PLD, PLA, FPGA, or PAL, or the like. - An alternate portion of the
memory 16 is preferably implemented using static or dynamic RAM, and stores executing application programs and data. However, the alterable portion of thememory 16 can also be implemented using a floppy disk and drive disk, a writable optical disk and disk drive, a hard drive, flash memory or the like. A fixed portion of thememory 16 is preferably implemented using a ROM and stores basic control programs for the audio data system. However, the fixed portion of the memory can also be implemented using a PROM, an EPROM, an EEPROM, or a CD-ROM and disk drive or the like. - The
storage device 14 is preferably implemented using a hard drive, and provides long-term storage for application programs and data. However, thestorage device 14 can also be implemented using a floppy disk and disk drive, a writable optical disk and disk drive, flash memory, static or dynamic RAM, or the like. - The
telephone system connector 28 may be any type of connector that is capable of carrying audio data. For example, thetelephone system connector 28 can be a connection to a publicly or privately switched telephone network, a wired or wireless link to a data network. The data network can be a local area network, a wide area network, an intranet, the Internet or any other distributed processing and storage network. - FIG. 2 shows one embodiment of the
audio data system 10 of this invention that has been incorporated into a telephone34. Theaudio data system 10 in FIG. 2 has an audiodata input device 28 that is adapted to receive audio data from a telephone system and a handset and/or amicrophone 26 that have been built into the telephone for hands free operation. Thisaudio data system 10 also includes akeypad 36 and a pen-basedinput display 20 which are used as a user interface to control the operation of theaudio data system 10. - The
audio data system 10 of this invention allows a user to indicate the type of information contained in audio data by attaching properties to the audio data. Properties indicate information types that are associated with the selected audio data. Properties are associated with portions of the audio data, not the entire audio data file, so that individual portions of the audio data can be easily identified. Audio data portions having a specific property from a number of audio data files can be grouped and listed together. Properties are useful for retrieval because they are a low overhead categorization that can be used to distinguish between audio data portions. - The
audio data system 10 of this invention records audio data and links the audio data to handwritten notes. The sides of the conversation in a telephone call are separated using a speaker segmentation algorithm. The resulting speaker segmentation is shown in theaudio timeline bar 40 at the top of thenote page 42 in FIG. 3. A local speaker is shown by the dark bands in thetop bar 44 and another speaker is shown in thelower bar 46. Speaker segmentation and display can be extended to any number of sides of a conversation. If theaudio data system 10 is connected to a conventional telephone system an elaborate speaker segmentation algorithm may be avoided by relying on the fact that each party will be using separate channels. All incoming audio can be attributed to a remote speaker and all outgoing audio can be attributed to a local speaker. - Audio data portions having different properties are distinguished from each other by the color of the
timeline 44. For example, audio data portions that do not have any assigned properties are shown in black, while audio data portions having “action item” as an assigned property are shown in green. - The
audio data system 10 allows properties to be applied to the audio data as well as to ink. Specific audio data can be selected and a property applied to that audio data. The property is indicated in theaudio bar 40 by changing the color of theaudio bar 40 to a color associated with that property. For example, the “to do” property may be indicated to the user in blue ink for the corresponding note marks and as a blue band in theaudio bar 40 for the corresponding audio data. - Properties can be assigned to the audio data portions either before or after the audio data is recorded. If the user knows beforehand that the audio data about to be recorded has a certain property, the user can select that property from a menu, assign it to the audio data and begin recording. Properties may be assigned after recording by selecting the audio and assigning a property to the selected audio. For example, toggling one of the
property buttons 50, a property corresponding to the property button is applied to the current turn in the conversation, indicating that this audio data portion has the given property. Theproperty buttons 50 shown in theuser interface 42 of FIG. 3 include aproperty button 52 which does not assign a property, atopic property button 54, aname property button 56, aquote property button 58, aphone property button 60, a to doproperty button 62, a Uniform Resource Locator (URL)property button 64, anorganization property button 66 and a user defined property button. Properties can be assigned to audio data and note marks that are currently being received or which have been selected by toggling theproperty button 50 which corresponds to the desired property. Theaudio data system 10 may also indicate an audio data property, even though there are no associated note marks with that property. - While the extent of an ink property is defined by the note marks, the extent of an audio property is not automatically defined. However, since there is a natural segmentation of the audio according to the turn-taking in the conversation, turns can be used to define the extent of the audio property. Audio properties replace the notion of the audio highlighting used in previous systems, since a property such as “important” can be used instead. Further, the user does not have to specify the extent of the property.
- In one embodiment of the
audio data system 10 of this invention, the identity of a caller is used to index the audio data. This index can be obtained in several ways. The caller can be identified using a database linking a phone number or a caller id to the caller. If the caller is not calling from his own phone, theaudio data system 10 of this invention uses a database of audio data files of known callers, and matches the voice from the incoming caller to an audio data file of a known caller. Any conventional speaker identification system may be used. One example of a speaker identification system is disclosed in U.S. Pat. No. 5,655,058, herein incorporated by reference in its entirety. The user can also enter the caller's identity through any other conventional input method. The audio data can also be indexed automatically according to date and time. - The user can also use the
audio data system 10 of this invention to initiate calls using a database of phone numbers and callers, by specifying a person, either directly, by name, or indirectly, by referring to a note page associated with the person or to a recorded call from the person. FIG. 3 shows a call initiation dialog window 70 on top ofuser interface 42. A user can access the call initiation dialog window 70 by selecting thecall initiation button 72 on theuser interface 42. The call initiation dialog window 70 includes akeypad 74 that enables the user to manually enter a telephone number. The call initiation dialog window 70 also includes aphonebook button 76 which provides access to a phonebook listing of telephone numbers (not shown). The call initiation dialog window 70 also has a dropdowntext entry box 78 which enables a user to place a cursor in thetext entry box 78 and enter text and/or numbers using a separate data input device such as a keyboard. Lastly, the call initiation dialog window 70 has a cancelbutton 80 which enables a user to cancel a call initiation and also an “ok”button 82 which enables a user to initiate a call once the telephone number information has been entered. - This invention also provides a system for call screening and/or routing based on speaker identity. A screen mode (not shown) of the
audio data system 10 can be set so that all calls are automatically answered, all calls are automatically recorded without interaction (as in voice mail), or calls are routed in accordance with the identity of the caller. In one embodiment of the screen mode, theaudio data system 10 answers the call and sends a message to the caller to state his name and the purpose of the call. If the caller's identity matches a name on an “accept call” list, the user is notified of the call. Otherwise, the call is recorded as voice mail. - In addition to properties, the
audio data system 10 of this invention allows keywords to be associated with audio data. This allows keyword-based retrieval of audio data. In contrast to properties, which describe the type of information contained in the audio data portions having those properties, keywords describe the actual content of the audio data. Because keywords reflect more general information, they are assigned to the entire audio data file and not to particular portions of audio data. Keywords are displayed in the margins of a digital page (not shown) and can be entered in several ways. For example, keywords can be selected from a list of previously entered keywords. Alternately, keywords can be typed in using a pop-up keyboard (not shown). Additionally, keywords can be hand-written in a keyword entry box (not shown) and recognized using a standard handwriting recognition system. In the user-interface shown in FIG. 3, there is a drop downtext entry box 124 for entering the keywords. Also, note marks can be selected and sent to a handwriting recognizer, and the resulting text entered as a keyword for the audio data associated with the note marks. If the user does not want to immediately interact with the handwriting recognition system, the user can associate the “keyword” property with digital ink to cause delayed recognition. - The audio data can be retrieved using attributes associated with the audio data such as caller, time, and topic, using properties and keywords. The
user interface 42 also hasquery buttons 112 that enable a user to access a context-based query (not shown) by selecting the context-basedquery button 114 or access aquery window 84 as shown in FIG. 4 using thequery button 116. Thequery window 84 includesproperty check boxes 86 that enable a user to search based upon a designated property. Thequery window 84 also has akeyword entry box 88 that permits a user to search for audio data and notemarks having a specified keyword. Retrieval by keyword and name returns a list of audio data files with the identified caller and having the desired keyword. Retrieval by property is more specific, since it provides the audio data portions of calls and digital ink groups having the specified property. This is convenient for locating a particular comment in an audio data file. For example, if a comment is marked with the property “important”, that audio data segment containing the comment can be retrieved using a query with the “important” property designated. Segments of audio data files can also be played to both parties during a conversation. Thus, comments from previous conversations can be shared. - Results from a search are shown in FIG. 5 in a
search result window 94. The search result window lists the attributes of the retrieved data files in columns. The columns indicate thecreation date 96, the date oflast modification 98, the type of property as indicated byicon 100 and thekeywords 102. The series of icons describe properties associated with the audio data files. For example, the telephone icon represents the property “phone number” and the check box represents the “action item” property. The final column displays keywords associated with the audio data files. - Using the indexing mechanisms provided by the digital properties, time stamping and keywords, the audio data system of this invention provides the user with multiple views of audio data stored in a notebook. The views provided by this invention are analogous to the views in a database system. If a query specifies a particular property, only those audio data portions having that property will be displayed. This view of the audio data is formed by newly generated audio data files, each having an original date and containing only the audio data portions with the specified property. The user can also create a view of a notebook by forming a query to request audio data files having specific keywords. The user can define the keywords in the query by browsing an alphabetized list of keywords that have been attached to pages, or by typing or writing the desired keywords.
- As shown in FIG. 4, views are indicated by
tabs 48 at the bottom of the notebook page. A name is associated with eachtab 48 for easy reference. The user can return to the previous view by selecting the correspondingtab 48. The user can decide if the view should be static or dynamic. When the view is static, the contents of the view remains the same. When the view is dynamic, new items are added to the view as they are recorded by the notebook, provided that these items satisfy the terms of the query for this view. - A user can navigate through a notebook using the
navigation buttons 104 in theuser interface window 42.Navigation buttons - The
user interface 42 is also provided with modeselect buttons 118 which enable a user to select either a page mode using a page modeselect button 120 or a table of contents mode using the table of contents modeselect button 122. The user interface generally includes the features specified and described in co-pending U.S. patent application Ser. No. 08/821,311. - While this invention has been described with the specific embodiments outlined above, many alternatives, modifications and variations are apparent to those skilled in the art. Accordingly, the preferred embodiments described above are illustrative and not limiting. Various changes may be made without departing from the spirit and scope of the invention as defined in the following claims.
Claims (20)
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/058,283 US6404856B1 (en) | 1998-04-10 | 1998-04-10 | System for recording, annotating and indexing audio data |
JP11089695A JPH11313173A (en) | 1998-04-10 | 1999-03-30 | Digital sound recording system |
EP99302805A EP0949621A3 (en) | 1998-04-10 | 1999-04-12 | System for recording, annotating and indexing audio data |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/058,283 US6404856B1 (en) | 1998-04-10 | 1998-04-10 | System for recording, annotating and indexing audio data |
Publications (2)
Publication Number | Publication Date |
---|---|
US20020027976A1 true US20020027976A1 (en) | 2002-03-07 |
US6404856B1 US6404856B1 (en) | 2002-06-11 |
Family
ID=22015844
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/058,283 Expired - Lifetime US6404856B1 (en) | 1998-04-10 | 1998-04-10 | System for recording, annotating and indexing audio data |
Country Status (3)
Country | Link |
---|---|
US (1) | US6404856B1 (en) |
EP (1) | EP0949621A3 (en) |
JP (1) | JPH11313173A (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6741680B2 (en) * | 2000-12-22 | 2004-05-25 | International Business Machines Corporation | Telephone message system with flexible presentation capability |
US20050005016A1 (en) * | 2003-03-13 | 2005-01-06 | Fuji Xerox Co., Ltd. | User-data relating apparatus with respect to continuous data |
US20050243979A1 (en) * | 2004-04-30 | 2005-11-03 | Starbuck Bryan T | Integrated messaging user interface with message-based logging |
US20070297786A1 (en) * | 2006-06-22 | 2007-12-27 | Eli Pozniansky | Labeling and Sorting Items of Digital Data by Use of Attached Annotations |
US20090111437A1 (en) * | 2001-08-17 | 2009-04-30 | Elman Boris S | Automated conversation recording device and service |
US20090228279A1 (en) * | 2008-03-07 | 2009-09-10 | Tandem Readers, Llc | Recording of an audio performance of media in segments over a communication network |
US20090248645A1 (en) * | 2008-03-28 | 2009-10-01 | Brother Kogyo Kabushiki Kaisha | Device, method and computer readable medium for management of time-series data |
US20090254828A1 (en) * | 2004-10-26 | 2009-10-08 | Fuji Xerox Co., Ltd. | System and method for acquisition and storage of presentations |
US20090285379A1 (en) * | 2006-03-24 | 2009-11-19 | Realnetworks Asia Pacific Co., Ltd. | Method and system for providing ring back tone played at a point selected by user |
US20130311178A1 (en) * | 2012-05-21 | 2013-11-21 | Lg Electronics Inc. | Method and electronic device for easily searching for voice record |
US20140365514A1 (en) * | 1999-11-12 | 2014-12-11 | Sony Corporation | Apparatus and associated methodology for automatically selecting content using stored criteria |
US20180129800A1 (en) * | 2016-11-10 | 2018-05-10 | Ricoh Company, Ltd. | Information processing apparatus, information processing system, and method of processing information |
Families Citing this family (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7562392B1 (en) * | 1999-05-19 | 2009-07-14 | Digimarc Corporation | Methods of interacting with audio and ambient music |
GB2359918A (en) | 2000-03-01 | 2001-09-05 | Sony Uk Ltd | Audio and/or video generation apparatus having a metadata generator |
US6697796B2 (en) * | 2000-01-13 | 2004-02-24 | Agere Systems Inc. | Voice clip search |
US7263484B1 (en) | 2000-03-04 | 2007-08-28 | Georgia Tech Research Corporation | Phonetic searching |
US7065182B1 (en) * | 2000-08-10 | 2006-06-20 | Glenayre Electronics, Inc. | Voice mail message repositioning device |
KR100853334B1 (en) * | 2000-09-19 | 2008-08-21 | 후지쯔 가부시끼가이샤 | Store system |
US20020118798A1 (en) * | 2001-02-27 | 2002-08-29 | Christopher Langhart | System and method for recording telephone conversations |
US7065187B2 (en) * | 2002-01-10 | 2006-06-20 | International Business Machines Corporation | System and method for annotating voice messages |
US7035807B1 (en) * | 2002-02-19 | 2006-04-25 | Brittain John W | Sound on sound-annotations |
US6853710B2 (en) * | 2002-07-17 | 2005-02-08 | Timothy M Harris | Telephone call messaging device |
US8055503B2 (en) | 2002-10-18 | 2011-11-08 | Siemens Enterprise Communications, Inc. | Methods and apparatus for audio data analysis and data mining using speech recognition |
US20040196306A1 (en) * | 2003-04-04 | 2004-10-07 | Microsoft Corporation | Method and system for flagging digital ink note content |
WO2005020209A2 (en) * | 2003-08-22 | 2005-03-03 | Ser Solutions, Ic. | System for and method of automated quality monitoring |
US8209185B2 (en) * | 2003-09-05 | 2012-06-26 | Emc Corporation | Interface for management of auditory communications |
US7499531B2 (en) * | 2003-09-05 | 2009-03-03 | Emc Corporation | Method and system for information lifecycle management |
US8103873B2 (en) * | 2003-09-05 | 2012-01-24 | Emc Corporation | Method and system for processing auditory communications |
US7457396B2 (en) * | 2003-09-05 | 2008-11-25 | Emc Corporation | Automated call management |
US20050069095A1 (en) * | 2003-09-25 | 2005-03-31 | International Business Machines Corporation | Search capabilities for voicemail messages |
US8229904B2 (en) * | 2004-07-01 | 2012-07-24 | Emc Corporation | Storage pools for information management |
US8180742B2 (en) * | 2004-07-01 | 2012-05-15 | Emc Corporation | Policy-based information management |
US8244542B2 (en) * | 2004-07-01 | 2012-08-14 | Emc Corporation | Video surveillance |
US7444287B2 (en) * | 2004-07-01 | 2008-10-28 | Emc Corporation | Efficient monitoring system and method |
US8180743B2 (en) | 2004-07-01 | 2012-05-15 | Emc Corporation | Information management |
US7707037B2 (en) * | 2004-07-01 | 2010-04-27 | Emc Corporation | Archiving of surveillance data |
US9268780B2 (en) * | 2004-07-01 | 2016-02-23 | Emc Corporation | Content-driven information lifecycle management |
US20060004579A1 (en) * | 2004-07-01 | 2006-01-05 | Claudatos Christopher H | Flexible video surveillance |
US8626514B2 (en) * | 2004-08-31 | 2014-01-07 | Emc Corporation | Interface for management of multiple auditory communications |
US20060075347A1 (en) * | 2004-10-05 | 2006-04-06 | Rehm Peter H | Computerized notetaking system and method |
US8204192B1 (en) * | 2004-12-30 | 2012-06-19 | Alcatel Lucent | Documented call log |
GB2424791A (en) * | 2005-03-31 | 2006-10-04 | Digits Ltd | Integrated telephone handset, computer caller display, voice call recording and automatic redial system |
US7620551B2 (en) * | 2006-07-20 | 2009-11-17 | Mspot, Inc. | Method and apparatus for providing search capability and targeted advertising for audio, image, and video content over the internet |
US8239480B2 (en) * | 2006-08-31 | 2012-08-07 | Sony Ericsson Mobile Communications Ab | Methods of searching using captured portions of digital audio content and additional information separate therefrom and related systems and computer program products |
US8311823B2 (en) * | 2006-08-31 | 2012-11-13 | Sony Mobile Communications Ab | System and method for searching based on audio search criteria |
BRPI0622225A8 (en) * | 2006-12-22 | 2016-11-08 | Motorola Inc | METHOD AND DEVICE FOR DATA CAPTURE TO PRESS TO SPEAK ON PHONE |
WO2008150924A1 (en) * | 2007-05-29 | 2008-12-11 | Livescribe, Inc. | Animation of audio ink |
US10002192B2 (en) * | 2009-09-21 | 2018-06-19 | Voicebase, Inc. | Systems and methods for organizing and analyzing audio content derived from media files |
US9142216B1 (en) * | 2012-01-30 | 2015-09-22 | Jan Jannink | Systems and methods for organizing and analyzing audio content derived from media files |
US20110218994A1 (en) * | 2010-03-05 | 2011-09-08 | International Business Machines Corporation | Keyword automation of video content |
KR101844903B1 (en) | 2011-08-31 | 2018-04-04 | 삼성전자 주식회사 | Providing Method for Data Complex Recording And Portable Device thereof |
JP6128146B2 (en) * | 2015-02-24 | 2017-05-17 | カシオ計算機株式会社 | Voice search device, voice search method and program |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5276731A (en) | 1991-04-26 | 1994-01-04 | Rolm Company | Method and apparatus for handling incoming telephone calls |
WO1993007562A1 (en) | 1991-09-30 | 1993-04-15 | Riverrun Technology | Method and apparatus for managing information |
US5390236A (en) * | 1992-03-31 | 1995-02-14 | Klausner Patent Technologies | Telephone answering device linking displayed data with recorded audio message |
JPH06178296A (en) * | 1992-12-09 | 1994-06-24 | Hitachi Ltd | Electronic conference system |
JP3138370B2 (en) * | 1993-09-09 | 2001-02-26 | 株式会社日立製作所 | Information processing device |
JP3185505B2 (en) * | 1993-12-24 | 2001-07-11 | 株式会社日立製作所 | Meeting record creation support device |
US5655058A (en) | 1994-04-12 | 1997-08-05 | Xerox Corporation | Segmentation of audio data for indexing of conversational speech for real-time or postprocessing applications |
GB9408042D0 (en) | 1994-04-22 | 1994-06-15 | Hewlett Packard Co | Device for managing voice data |
US5651055A (en) * | 1995-06-29 | 1997-07-22 | Lucent Technologies Inc. | Digital secretary |
US5559875A (en) * | 1995-07-31 | 1996-09-24 | Latitude Communications | Method and apparatus for recording and retrieval of audio conferences |
US5717879A (en) | 1995-11-03 | 1998-02-10 | Xerox Corporation | System for the capture and replay of temporal data representing collaborative activities |
US5835577A (en) * | 1996-04-25 | 1998-11-10 | Copytele, Inc. | Multi-functional personal telecommunications apparatus |
US6034970A (en) * | 1996-05-31 | 2000-03-07 | Adaptive Micro Systems, Inc. | Intelligent messaging system and method for providing and updating a message using a communication device, such as a large character display |
US5970455A (en) | 1997-03-20 | 1999-10-19 | Xerox Corporation | System for capturing and retrieving audio data and corresponding hand-written notes |
-
1998
- 1998-04-10 US US09/058,283 patent/US6404856B1/en not_active Expired - Lifetime
-
1999
- 1999-03-30 JP JP11089695A patent/JPH11313173A/en active Pending
- 1999-04-12 EP EP99302805A patent/EP0949621A3/en not_active Ceased
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140365514A1 (en) * | 1999-11-12 | 2014-12-11 | Sony Corporation | Apparatus and associated methodology for automatically selecting content using stored criteria |
US6741680B2 (en) * | 2000-12-22 | 2004-05-25 | International Business Machines Corporation | Telephone message system with flexible presentation capability |
US8284907B2 (en) * | 2001-08-17 | 2012-10-09 | Verizon Laboratories Inc. | Automated conversation recording device and service |
US20090111437A1 (en) * | 2001-08-17 | 2009-04-30 | Elman Boris S | Automated conversation recording device and service |
US20050005016A1 (en) * | 2003-03-13 | 2005-01-06 | Fuji Xerox Co., Ltd. | User-data relating apparatus with respect to continuous data |
US7350140B2 (en) * | 2003-03-13 | 2008-03-25 | Fuji Xerox Co., Ltd. | User-data relating apparatus with respect to continuous data |
US7512400B2 (en) * | 2004-04-30 | 2009-03-31 | Microsoft Corporation | Integrated messaging user interface with message-based logging |
AU2004279177B2 (en) * | 2004-04-30 | 2009-11-12 | Microsoft Technology Licensing, Llc | Integrated messaging user interface with message-based logging |
US20050243979A1 (en) * | 2004-04-30 | 2005-11-03 | Starbuck Bryan T | Integrated messaging user interface with message-based logging |
WO2005112409A3 (en) * | 2004-04-30 | 2006-05-18 | Microsoft Corp | Integrated messaging user interface with message-based logging |
US20090254828A1 (en) * | 2004-10-26 | 2009-10-08 | Fuji Xerox Co., Ltd. | System and method for acquisition and storage of presentations |
US9875222B2 (en) * | 2004-10-26 | 2018-01-23 | Fuji Xerox Co., Ltd. | Capturing and storing elements from a video presentation for later retrieval in response to queries |
US9596354B2 (en) * | 2006-03-24 | 2017-03-14 | Realnetworks Asia Pacific Co., Ltd. | Method and system for providing ring back tone played at a point selected by user |
US20090285379A1 (en) * | 2006-03-24 | 2009-11-19 | Realnetworks Asia Pacific Co., Ltd. | Method and system for providing ring back tone played at a point selected by user |
US8718256B2 (en) * | 2006-03-24 | 2014-05-06 | Realnetworks Asia Pacific Co., Ltd. | Method and system for providing ring back tone played at a point selected by user |
US20140321332A1 (en) * | 2006-03-24 | 2014-10-30 | Realnetworks Asia Pacific Co., Ltd. | Method and system for providing ring back tone played at a point selected by user |
US8301995B2 (en) * | 2006-06-22 | 2012-10-30 | Csr Technology Inc. | Labeling and sorting items of digital data by use of attached annotations |
US20070297786A1 (en) * | 2006-06-22 | 2007-12-27 | Eli Pozniansky | Labeling and Sorting Items of Digital Data by Use of Attached Annotations |
US20090228279A1 (en) * | 2008-03-07 | 2009-09-10 | Tandem Readers, Llc | Recording of an audio performance of media in segments over a communication network |
US20090248645A1 (en) * | 2008-03-28 | 2009-10-01 | Brother Kogyo Kabushiki Kaisha | Device, method and computer readable medium for management of time-series data |
US20130311178A1 (en) * | 2012-05-21 | 2013-11-21 | Lg Electronics Inc. | Method and electronic device for easily searching for voice record |
US9224397B2 (en) * | 2012-05-21 | 2015-12-29 | Lg Electronics Inc. | Method and electronic device for easily searching for voice record |
US20180129800A1 (en) * | 2016-11-10 | 2018-05-10 | Ricoh Company, Ltd. | Information processing apparatus, information processing system, and method of processing information |
US10558795B2 (en) * | 2016-11-10 | 2020-02-11 | Ricoh Company, Ltd. | Information processing apparatus, information processing system, and method of processing information |
US11017073B2 (en) | 2016-11-10 | 2021-05-25 | Ricoh Company, Ltd. | Information processing apparatus, information processing system, and method of processing information |
Also Published As
Publication number | Publication date |
---|---|
US6404856B1 (en) | 2002-06-11 |
EP0949621A3 (en) | 2001-03-14 |
JPH11313173A (en) | 1999-11-09 |
EP0949621A2 (en) | 1999-10-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6404856B1 (en) | System for recording, annotating and indexing audio data | |
US5970455A (en) | System for capturing and retrieving audio data and corresponding hand-written notes | |
US5457738A (en) | Method and system for searching an on-line directory at a telephone station | |
JP3175399B2 (en) | Card data management device | |
CN102483917B (en) | For the order of display text | |
Whittaker et al. | SCANMail: a voicemail interface that makes speech browsable, readable and searchable | |
US8781080B2 (en) | Systems and methods for presenting audio messages | |
US5943055A (en) | Computer interface method and system | |
US5742736A (en) | Device for managing voice data automatically linking marked message segments to corresponding applications | |
US6411822B1 (en) | Communication terminal | |
US20030026392A1 (en) | Method and apparatus for extracting information from voice messages | |
US20030128820A1 (en) | System and method for gisting, browsing and searching voicemail using automatic speech recognition | |
CN104679727A (en) | Search Based User Interface | |
US7684552B2 (en) | Phone batch calling task management system | |
JP2009522693A (en) | Applying metadata to documents and document objects through the software application user interface | |
US20090164942A1 (en) | User interface and communication terminal | |
US20020077832A1 (en) | Computer based integrated text/graphic document analysis | |
US8792863B2 (en) | Method for collecting and storing annotations associated to a voice audio data | |
JP2009522695A (en) | How to attach metadata to documents and document objects using the operating system user interface | |
JPH0962740A (en) | Consultation system | |
EP1595208A2 (en) | Automatic running and ranking of suitable applications based on given user input | |
US20070136433A1 (en) | Intelligent buffer and related applications | |
KR101139448B1 (en) | Mobile communication terminal with memorial day searching and its operating method | |
US20030120787A1 (en) | Portable radio telephone | |
KR100673448B1 (en) | Mobile communication terminal searching memo and its operating method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: FUJI XEROX CO., LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WILCOX, LYNN D.;CHIU, PATRICK R.;GWIZDKA, JACEK;REEL/FRAME:009411/0319;SIGNING DATES FROM 19980626 TO 19980807 Owner name: XEROX CORPORATION, CONNECTICUT Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WILCOX, LYNN D.;CHIU, PATRICK R.;GWIZDKA, JACEK;REEL/FRAME:009411/0319;SIGNING DATES FROM 19980626 TO 19980807 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: BANK ONE, NA, AS ADMINISTRATIVE AGENT, ILLINOIS Free format text: SECURITY AGREEMENT;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:013111/0001 Effective date: 20020621 Owner name: BANK ONE, NA, AS ADMINISTRATIVE AGENT,ILLINOIS Free format text: SECURITY AGREEMENT;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:013111/0001 Effective date: 20020621 |
|
AS | Assignment |
Owner name: JPMORGAN CHASE BANK, AS COLLATERAL AGENT, TEXAS Free format text: SECURITY AGREEMENT;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:015134/0476 Effective date: 20030625 Owner name: JPMORGAN CHASE BANK, AS COLLATERAL AGENT,TEXAS Free format text: SECURITY AGREEMENT;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:015134/0476 Effective date: 20030625 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
FPAY | Fee payment |
Year of fee payment: 12 |
|
AS | Assignment |
Owner name: XEROX CORPORATION, NEW YORK Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK ONE, NA;REEL/FRAME:034781/0055 Effective date: 20030625 |
|
AS | Assignment |
Owner name: XEROX CORPORATION, NEW YORK Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:034781/0444 Effective date: 20061204 |
|
AS | Assignment |
Owner name: XEROX CORPORATION, CONNECTICUT Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE'S NAME PREVIOUSLY RECORDED AT REEL: 009411 FRAME: 0319. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNORS:WILCOX, LYNN D;CHIU, PATRICK R;GWIZDKA, JACEK;SIGNING DATES FROM 19980626 TO 19980807;REEL/FRAME:035226/0537 |
|
AS | Assignment |
Owner name: MAJANDRO LLC, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:053258/0949 Effective date: 20200413 |
|
AS | Assignment |
Owner name: XEROX CORPORATION, CONNECTICUT Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:JPMORGAN CHASE BANK, N.A. AS SUCCESSOR-IN-INTEREST ADMINISTRATIVE AGENT AND COLLATERAL AGENT TO BANK ONE, N.A.;REEL/FRAME:061388/0388 Effective date: 20220822 Owner name: XEROX CORPORATION, CONNECTICUT Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:JPMORGAN CHASE BANK, N.A. AS SUCCESSOR-IN-INTEREST ADMINISTRATIVE AGENT AND COLLATERAL AGENT TO JPMORGAN CHASE BANK;REEL/FRAME:066728/0193 Effective date: 20220822 |