There is disclosed a method and apparatus for providing continuous speech-to-subtitles translation utilizing a video-based communications device but without speech synthesis at the output. Instead, a translation of each user's speech is displayed continuously in text form on the other user's screen. In the preferred embodiment, the sending party speaks into a conventional videophone. speech recognition and translation of the transmitted signal are performed by a remote device at the receiving party's location. The audio portion of the signal is sent both to a speaker for audio output and to a speech recognizer and text-based translation system, the output of which is text translated into the target language. The video portion of the signal and the translated text are combined in a subtitle generator and sent to a display device for viewing by the receiving party.
|
5. A method of providing continuous speech-to-subtitle translation of a signal containing a video portion and an audio portion, comprising the steps of:
(e) converting said audio portion to a corresponding first textual signal at a sending party's location; (f) translating said corresponding first textual signal to a second textual signal at a location remote from said sending party's location, wherein said second textual signal is in a target language; and (g) combining said video portion with said second textual signal to form a display signal, wherein said display signal displays said second textual signal as subtitles.
1. An apparatus for providing continuous speech-to-subtitle translation of a signal containing a video portion, and an audio portion comprising:
(a) means for converting said audio portion to a corresponding first textual signal, wherein said converting means is located at a sending party's location; (b) means for translating said corresponding first textual signal to a second textual signal wherein said second textual signal is in a target language and wherein said translating means is located remotely from said sending party's location; (c) means for combining said video portion with said second textual signal to form a display signal, wherein said display signal displays said second textual signal as subtitles; and (d) means for simultaneously displaying said display signal and outputting said audio portion.
3. An apparatus according to
4. An apparatus according to
6. A method according to
8. A method according to
9. A method according to
|
This invention relates to a method and apparatus for providing continuous speech-to-subtitles translation for communication between people speaking different languages.
As the world moves closer and closer to a true global economy, the need for individuals who speak different languages to be able to easily communicate has increased. Efforts have been made to facilitate communication between people speaking different languages using current speech-to-speech translation technology wherein the translated speech is synthesized in the target language.
Current speech-to-speech translation technology operates such that language is translated and synthesized sentence by sentence or phrase by phrase. Typically in such systems, a user speaks an entire sentence or phrase and presses a button or flips a switch when completed. The device then translates the entire sentence or phrase and synthesizes and outputs the translation in the target language. Thus, the other party must wait for the speech synthesizer to stop before responding. Such systems are currently preferred because they include in the synthesized translation the intonations, often related to emotion, contained in the original speech. Such intonations are generally thought to increase the quality of the communication. The result, however, is a delay between one party speaking and the other party hearing the synthesized translation which can make communication awkward and unnatural. A system which instead translates and synthesizes one word at a time would most likely also sound awkward and unnatural and would lack the normal intonations of speech. Thus, although research and development efforts are aimed at eliminating the current limitations of speech-to-speech translation technology, it is unlikely that any resulting systems will be capable of perfect simultaneous translation and speech synthesis for many years.
The problems and limitations associated with speech-to-speech translation are avoided, in accordance with the principles of the present invention, by using a video-based communication device for speech translation, but without speech synthesis of the output. Instead, a translation of each user's speech is displayed continuously in text form on the other user's screen. At the same time, the original, untranslated speech is played over a speaker.
In the preferred embodiment, the sending party speaks into a conventional prior art videophone. The output of the videophone, a signal consisting of both the audio and video portions of the communication, is transmitted to the receiving party's location. Speech recognition and translation of the transmitted signal are performed by a remote device at the receiving party's location. The audio portion of the signal is sent both to a speaker for audio output and to a speech recognizer and text-based machine translation system, the output of which is translated text. The video portion of the signal and the translated text are then combined in a subtitle generator and sent to a display device for viewing by the receiving party.
Because users hear the actual voice of the other party, the communication is more personal and is likely to be perceived to be of higher quality. Hearing the original speech can also reduce misunderstanding because emotional clues are available to the listener. Also, in the event that an imperfect translation takes place, users can look over the stream of translated words and make use of their knowledge of the other language to try to reconstruct the intended meaning. Moreover, because the original untranslated speech is provided audibly and the translated text is provided visually, users can employ any knowledge they may have of the original language. Finally, according to the preferred embodiment, the sending party need not be aware that the other party has translation subtitles displayed on their screen, a feature that would be appreciated by users embarrassed about their foreign language skills.
FIG. 1 is a block diagram of an embodiment of the present invention wherein recognition, translation and subtitle generation are performed remotely.
FIG. 2 is a flow diagram illustrating the method for providing continuous speech-to-subtitles translation.
FIG. 3 is a block diagram of an embodiment of the present invention wherein recognition, translation and subtitle generation are performed by a telephone service provider network.
Referring to FIG. 1, a diagram of the presently preferred embodiment of the system is shown. Conventional prior art videophone 5, such as the AT&T VT2500, is located at the sending party's location. Remote receiving device 8 is located at the receiving party's location. The sending party speaks into videophone 5 which contains camera 9 and microphone 11. Camera 9 outputs video signal 10, which represents the visual component of the communication, and microphone 11 outputs audio signal 12, which represents the speech component of the communication. Video signal 10 and audio signal 12 are fed into audio/video encoder 13. Audio/video encoder 13 is a conventional digital signal processing device which can be found in the transmitting end of the AT&T VT2500. Audio/Video encoder 13 converts video signal 10 and audio signal 12 into a single encoded digital signal 14. Encoded digital signal 14 is placed on conventional telephone line 15 for transmission to remote receiving device 8.
At remote receiving device 8, encoded digital signal 14 is fed into audio/video decoder 16. Similar to audio/video encoder 13, audio/video decoder 16 is a conventional digital signal processing device which can be found in the receiving end of the AT&T VT2500. Audio/video decoder 16 converts encoded digital signal 14 back into two separate signals, video signal 23 and audio signal 17. Audio signal 17, which is in the original language of the sending party, is simultaneously fed into speaker 18 for audio output to the receiving party and into recognizer 19. Recognizer 19 is a conventional speech recognizer which converts human speech to text. Speech recognizers are well known in the art and are described, for example, in L. R. Rabiner and B. H. Juang, Fundamentals of Speech Recognition, Prentice-Hall (1993). Using a prior art statistical pattern recognition technique, recognizer 19 converts audio signal 17 into recognition hypothesis 20, which consists of one or more possible sequences of words in text format corresponding to audio signal 17. Essentially, recognition hypothesis 20 is a signal representing the most likely textual counterpart to the spoken language represented by audio signal 17. Recognition hypothesis 20, however, is still in the original language of the sending party, and thus needs to be translated into the target language.
Recognition hypothesis 20 is sent to translator 21. Translator 21 is a conventional text-based machine translation system which converts text in one natural language to text in another natural language. Text-based machine translation systems are well known in the art and are described, for example, in W. J. Hutchins and H. L. Somers, An Introduction to Machine Translation, Academic Press (1992). Translator 21 takes recognition hypothesis 20 and translates it into the target language. If recognition hypothesis 20 consists of a set of possible sequences of words, translator 21 applies a language model which chooses the most likely grammatical version for translation. The output of translator 21 is text signal 22.
Text signal 22 and video signal 23 are sent to subtitle generator 24 where the two signals are overlaid onto one another to create display signal 25. In display signal 25, text signal 22 appears as subtitles to video signal 23. Subtitle generator 24 is common in the prior art, especially in the film industry.
Subtitle generator 24 outputs display signal 25 which in turn is sent to video display device 26, such as a monitor, for display to the receiving party. Thus, the receiving party can simultaneously hear the original speech of the sending party and view the video of the sending party overlaid with subtitles translating the sending party's speech.
Because the recognition and translation functions are performed at remote receiving device 8, standard existing videophone transmission signals and protocols, such as ITU-H.261 and IT-TH.263, can be used. As such, the only party that needs to have any special equipment is the receiving party. The sending party need only use a standard videophone and need not know that the translation and subtitling are taking place.
Referring now to FIG. 2, a flow diagram illustrating the method for providing continuous speech-to-subtitles translation is shown. Block 30 shows that after a sending party speaks into his or her videophone, the encoded signal output thereby is sent to the receiving party's location. As shown in Block 32, the encoded signal is then decoded into an audio signal and a video signal. Block 34 shows that the audio signal is then converted into a corresponding textual signal in the sending party's language. Block 36 shows that the textual signal is then translated into a textual signal in the receiving party's language, also known as the target language. As shown in block 38, the textual signal in the target language is then overlaid onto the video signal as subtitles. Finally, as shown in block 40, the video signal with subtitles overlaid thereon and the audio signal are simultaneously output to the receiving party.
Referring now to FIG. 3, an alternate embodiment of the system is shown wherein the recognition, translation and subtitle generation functions are performed by telephone service provider network 60 rather than at the receiving party's location. Conventional videophone 62 located at the sending party's location 64 outputs signal 66. Signal 66 is a standard videophone signal. Signal 66 is sent to central processing unit, or CPU, 68 which is attached to telephone network switch 70. CPU 68 and telephone network switch 70 are part of telephone service provider network 60. CPU 68 contains algorithms which perform the recognition, translation and subtitle generation functions on signal 66. CPU 68 outputs signal 72. Signal 72 consists of an audio portion, which contains the sending party's original speech, and a subtitled video portion. Signal 72 is sent to conventional videophone 74 located at the receiving party's location 76 where it can be viewed by the receiving party.
In this embodiment, the continuous speech-to-subtitles translation would be provided as a service by the telephone service provider wherein the user is charged a fee for each use. As such, a person desiring speech-to-subtitle translation could access the service as needed, using a conventional videophone. This embodiment would also allow the use of standard videophone transmission signals and protocols. Finally, like the preferred embodiment, this embodiment would allow the receiving party to use the service without the knowledge of the sending party.
Still further alternate embodiments of the system are possible. One such embodiment would entail performing the recognition and translation functions at the sending party's locations. Another such embodiment would entail performing the recognition function at the sending party's location and the translation function at the receiving party's location. Both such alternate embodiments would require the transmission of data, i.e., the translated text in the case of the former and the recognition hypothesis in the case of the latter, in addition to the transmission of audio and video signals. Thus, these alternate embodiments would require the use of modified videophone equipment. These alternate embodiments, however, would have the advantage of providing more accurate recognition than in the preferred embodiment because the recognition function is performed locally rather than at the receiving party's location. Local recognition is more accurate because the audio signal does not have to be transmitted over telephone lines before recognition takes place. Conversely, in the preferred embodiment, when encoded digital signal 14 is sent to the receiving party's location, the quality of audio signal 17 is somewhat diminished due to the limited bandwidth of conventional telephone line 15.
It is to be understood that the above description comprises only a few of the possible embodiments of the present invention. Numerous other arrangements may be devised by one skilled in the art without departing from the spirit and scope of the invention. The invention is thus limited only as defined in the accompanying claims.
Patent | Priority | Assignee | Title |
10015311, | Jun 29 2005 | Ultratec, Inc. | Device independent text captioned telephone service |
10051207, | Nov 12 2015 | SORENSON IP HOLDINGS, LLC | Captioning communication systems |
10140433, | Aug 03 2001 | Comcast IP Holdings I, LLC | Video and digital multimedia aggregator |
10349096, | Aug 03 2001 | Comcast IP Holdings I, LLC | Video and digital multimedia aggregator content coding and formatting |
10372831, | Dec 12 2011 | GOOGLE LLC | Auto-translation for multi user audio and video |
10389876, | Feb 28 2014 | ULTRATEC, INC | Semiautomated relay method and apparatus |
10469660, | Jun 29 2005 | Ultratec, Inc. | Device independent text captioned telephone service |
10491746, | Feb 18 2004 | Ultratec, Inc. | Captioned telephone service |
10542141, | Feb 28 2014 | ULTRATEC, INC | Semiautomated relay method and apparatus |
10587751, | Feb 18 2004 | Ultratec, Inc. | Captioned telephone service |
10614173, | Dec 12 2011 | GOOGLE LLC | Auto-translation for multi user audio and video |
10643036, | Aug 18 2016 | HYPERCONNECT INC | Language translation device and language translation method |
10742805, | Feb 28 2014 | ULTRATEC, INC | Semiautomated relay method and apparatus |
10748523, | Feb 28 2014 | ULTRATEC, INC | Semiautomated relay method and apparatus |
10878721, | Feb 28 2014 | ULTRATEC, INC | Semiautomated relay method and apparatus |
10917519, | Feb 28 2014 | Ultratec, Inc. | Semiautomated relay method and apparatus |
10972604, | Jun 29 2005 | Ultratec, Inc. | Device independent text captioned telephone service |
10972683, | Nov 12 2015 | SORENSON IP HOLDINGS, LLC | Captioning communication systems |
11005991, | Feb 18 2004 | Ultratec, Inc. | Captioned telephone service |
11190637, | Feb 18 2004 | Ultratec, Inc. | Captioned telephone service |
11227129, | Aug 18 2016 | HYPERCONNECT INC | Language translation device and language translation method |
11258900, | Jun 29 2005 | Ultratec, Inc. | Device independent text captioned telephone service |
11368581, | Feb 28 2014 | Ultratec, Inc. | Semiautomated relay method and apparatus |
11509838, | Nov 12 2015 | SORENSON IP HOLDINGS, LLC | Captioning communication systems |
11539900, | Feb 21 2020 | ULTRATEC, INC | Caption modification and augmentation systems and methods for use by hearing assisted user |
11627221, | Feb 28 2014 | Ultratec, Inc. | Semiautomated relay method and apparatus |
11664029, | Feb 28 2014 | ULTRATEC, INC | Semiautomated relay method and apparatus |
11741963, | Feb 28 2014 | ULTRATEC, INC | Semiautomated relay method and apparatus |
5978014, | Sep 19 1997 | 8x8, Inc | Video TTY device and method for videoconferencing |
5991711, | Feb 26 1996 | Fuji Xerox Co., Ltd. | Language information processing apparatus and method |
5991723, | Jan 17 1998 | Genesys Telecommunications Laboratories, Inc | Method and apparatus for translating text and speech transferred over a telephony or similar network |
6195631, | Apr 15 1998 | Nuance Communications, Inc | Method and apparatus for automatic construction of hierarchical transduction models for language translation |
6323892, | Aug 04 1998 | Olympus Corporation | Display and camera device for videophone and videophone apparatus |
6513003, | Feb 03 2000 | REFINITIV US ORGANIZATION LLC | System and method for integrated delivery of media and synchronized transcription |
6542200, | Aug 14 2001 | CHELDAN TECHNOLOGIES, INC | Television/radio speech-to-text translating processor |
6599130, | Feb 02 2001 | TAMABO INC | Iterative video teaching aid with recordable commentary and indexing |
6754619, | Nov 15 1999 | Sony Corporation | Digital recording and playback system with voice recognition capability for concurrent text generation |
7013273, | Mar 29 2001 | Matsushita Electric Industrial Co., Ltd. | Speech recognition based captioning system |
7047191, | Mar 06 2000 | ENCO SYSTEMS, INC | Method and system for providing automated captioning for AV signals |
7110946, | Nov 12 2002 | United States of America as represented by the Secretary of the Navy | Speech to visual aid translator assembly and method |
7191117, | Jun 09 2000 | British Broadcasting Corporation | Generation of subtitles or captions for moving pictures |
7295969, | Nov 15 1999 | Sony Corporation; Sony Electronics, Inc. | Digital recording and playback system with voice recognition capability for concurrent text generation |
7310605, | Nov 25 2003 | Microsoft Technology Licensing, LLC | Method and apparatus to transliterate text using a portable device |
7359849, | Dec 17 2003 | SPEECHGEAR, INC | Translation techniques for acronyms and ambiguities |
7532229, | Nov 14 2003 | NEC Corporation | Data processor, data processing method and electronic equipment |
7552053, | Aug 22 2005 | International Business Machines Corporation | Techniques for aiding speech-to-speech translation |
7734467, | Aug 22 2005 | International Business Machines Corporation | Techniques for aiding speech-to-speech translation |
8112270, | Nov 15 1999 | Sony Corporation; Sony Electronics, Inc. | Digital recording and playback system with voice recognition capability for concurrent text generation |
8319819, | Mar 26 2008 | Cisco Technology, Inc.; Cisco Technology, Inc | Virtual round-table videoconference |
8326596, | May 28 2009 | Method and apparatus for translating speech during a call | |
8355041, | Feb 14 2008 | Cisco Technology, Inc. | Telepresence system for 360 degree video conferencing |
8390667, | Apr 15 2008 | Cisco Technology, Inc. | Pop-up PIP for people not in picture |
8416925, | Jun 29 2005 | Ultratec, Inc. | Device independent text captioned telephone service |
8472415, | Mar 06 2006 | Cisco Technology, Inc. | Performance optimization with integrated mobility and MPLS |
8477175, | Mar 09 2009 | Cisco Technology, Inc. | System and method for providing three dimensional imaging in a network environment |
8515024, | Jan 13 2010 | Ultratec, Inc. | Captioned telephone service |
8542264, | Nov 18 2010 | Cisco Technology, Inc. | System and method for managing optics in a video environment |
8570373, | Jun 08 2007 | Cisco Technology, Inc. | Tracking an object utilizing location information associated with a wireless device |
8599865, | Oct 26 2010 | Cisco Technology, Inc. | System and method for provisioning flows in a mobile network environment |
8599934, | Sep 08 2010 | Cisco Technology, Inc. | System and method for skip coding during video conferencing in a network environment |
8610755, | Feb 18 2011 | SORENSON IP HOLDINGS LLC | Methods and apparatuses for multi-lingual support for hearing impaired communication |
8635070, | Sep 29 2010 | Kabushiki Kaisha Toshiba | Speech translation apparatus, method and program that generates insertion sentence explaining recognized emotion types |
8659637, | Mar 09 2009 | Cisco Technology, Inc. | System and method for providing three dimensional video conferencing in a network environment |
8659639, | May 29 2009 | Cisco Technology, Inc. | System and method for extending communications between participants in a conferencing environment |
8670019, | Apr 28 2011 | Cisco Technology, Inc. | System and method for providing enhanced eye gaze in a video conferencing environment |
8682087, | Dec 19 2011 | Cisco Technology, Inc. | System and method for depth-guided image filtering in a video conference environment |
8692862, | Feb 28 2011 | Cisco Technology, Inc. | System and method for selection of video data in a video conference environment |
8694658, | Sep 19 2008 | Cisco Technology, Inc. | System and method for enabling communication sessions in a network environment |
8699457, | Nov 03 2010 | Cisco Technology, Inc. | System and method for managing flows in a mobile network environment |
8701020, | Feb 01 2011 | GOOGLE LLC | Text chat overlay for video chat |
8723914, | Nov 19 2010 | Cisco Technology, Inc.; Cisco Technology, Inc | System and method for providing enhanced video processing in a network environment |
8730297, | Nov 15 2010 | Cisco Technology, Inc. | System and method for providing camera functions in a video environment |
8768699, | Aug 22 2005 | International Business Machines Corporation | Techniques for aiding speech-to-speech translation |
8786631, | Apr 30 2011 | Cisco Technology, Inc. | System and method for transferring transparency information in a video environment |
8797377, | Feb 14 2008 | Cisco Technology, Inc. | Method and system for videoconference configuration |
8896655, | Aug 31 2010 | Cisco Technology, Inc.; University of North Carolina at Chapel Hill | System and method for providing depth adaptive video conferencing |
8902244, | Nov 15 2010 | Cisco Technology, Inc. | System and method for providing enhanced graphics in a video environment |
8908838, | Aug 23 2001 | ULTRATEC, INC | System for text assisted telephony |
8917821, | Jun 29 2005 | Ultratec, Inc. | Device independent text captioned telephone service |
8917822, | Aug 23 2001 | ULTRATEC, INC | System for text assisted telephony |
8934026, | May 12 2011 | Cisco Technology, Inc. | System and method for video coding in a dynamic environment |
8947493, | Nov 16 2011 | Cisco Technology, Inc. | System and method for alerting a participant in a video conference |
9031849, | Sep 30 2006 | Huawei Technologies Co., Ltd. | System, method and multipoint control unit for providing multi-language conference |
9082297, | Aug 11 2009 | Cisco Technology, Inc. | System and method for verifying parameters in an audiovisual environment |
9110891, | Dec 12 2011 | GOOGLE LLC | Auto-translation for multi user audio and video |
9111138, | Nov 30 2010 | Cisco Technology, Inc. | System and method for gesture interface control |
9131045, | Aug 23 2001 | ULTRATEC, INC | System for text assisted telephony |
9143725, | Nov 15 2010 | Cisco Technology, Inc. | System and method for providing enhanced graphics in a video environment |
9204096, | May 29 2009 | Cisco Technology, Inc. | System and method for extending communications between participants in a conferencing environment |
9225916, | Mar 18 2010 | Cisco Technology, Inc. | System and method for enhancing video images in a conferencing environment |
9301057, | Jan 17 2014 | HEARGLASS, INC | Hearing assistance system |
9313452, | May 17 2010 | Cisco Technology, Inc. | System and method for providing retracting optics in a video conferencing environment |
9331948, | Oct 26 2010 | Cisco Technology, Inc. | System and method for provisioning flows in a mobile network environment |
9338394, | Nov 15 2010 | Cisco Technology, Inc. | System and method for providing enhanced audio in a video environment |
9374536, | Nov 12 2015 | SORENSON IP HOLDINGS, LLC | Video captioning communication system, devices and related methods for captioning during a real-time video communication session |
9380374, | Jan 17 2014 | HEARGLASS, INC | Hearing assistance systems configured to detect and provide protection to the user from harmful conditions |
9525830, | Nov 12 2015 | SORENSON IP HOLDINGS, LLC | Captioning communication systems |
9681154, | Dec 06 2012 | PATENT CAPITAL GROUP | System and method for depth-guided filtering in a video conference environment |
9683862, | Aug 24 2015 | International Business Machines Corporation | Internationalization during navigation |
9689699, | Aug 24 2015 | International Business Machines Corporation | Internationalization during navigation |
9720909, | Dec 12 2011 | GOOGLE LLC | Auto-translation for multi user audio and video |
9813641, | Jun 19 2000 | Comcast IP Holdings I, LLC | Method and apparatus for targeting of interactive virtual objects |
9843621, | May 17 2013 | Cisco Technology, Inc. | Calendaring activities based on communication processing |
9934219, | Aug 24 2015 | International Business Machines Corporation | Internationalization during navigation |
9959872, | Dec 14 2015 | International Business Machines Corporation | Multimodal speech recognition for real-time video audio-based display indicia application |
9961196, | Aug 23 2001 | Ultratec, Inc. | System for text assisted telephony |
9967380, | Aug 23 2001 | Ultratec, Inc. | System for text assisted telephony |
9998686, | Nov 12 2015 | SORENSON IP HOLDINGS, LLC | Transcribing video communication sessions |
D636359, | Mar 21 2010 | Cisco Technology, Inc. | Video unit with integrated features |
D636747, | Mar 21 2010 | Cisco Technology, Inc. | Video unit with integrated features |
D637568, | Mar 21 2010 | Cisco Technology, Inc. | Free-standing video unit |
D637569, | Mar 21 2010 | Cisco Technology, Inc. | Mounted video unit |
D637570, | Mar 21 2010 | Cisco Technology, Inc. | Mounted video unit |
D653245, | Mar 21 2010 | Cisco Technology, Inc. | Video unit with integrated features |
D655279, | Mar 21 2010 | Cisco Technology, Inc. | Video unit with integrated features |
D678307, | Dec 16 2010 | Cisco Technology, Inc. | Display screen with graphical user interface |
D678308, | Dec 16 2010 | Cisco Technology, Inc. | Display screen with graphical user interface |
D678320, | Dec 16 2010 | Cisco Technology, Inc. | Display screen with graphical user interface |
D678894, | Dec 16 2010 | Cisco Technology, Inc. | Display screen with graphical user interface |
D682293, | Dec 16 2010 | Cisco Technology, Inc. | Display screen with graphical user interface |
D682294, | Dec 16 2010 | Cisco Technology, Inc. | Display screen with graphical user interface |
D682854, | Dec 16 2010 | Cisco Technology, Inc. | Display screen for graphical user interface |
D682864, | Dec 16 2010 | Cisco Technology, Inc. | Display screen with graphical user interface |
Patent | Priority | Assignee | Title |
5369704, | Mar 24 1993 | Engate LLC | Down-line transcription system for manipulating real-time testimony |
5512938, | Apr 06 1994 | Matsushita Electric Industrial Co., Ltd. | Teleconference terminal |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Dec 29 1995 | Lucent Technologies Inc. | (assignment on the face of the patent) | / | |||
Mar 21 1996 | ALSHAWI, HIYAN | Lucent Technologies Inc | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 007885 | /0466 | |
Feb 22 2001 | LUCENT TECHNOLOGIES INC DE CORPORATION | THE CHASE MANHATTAN BANK, AS COLLATERAL AGENT | CONDITIONAL ASSIGNMENT OF AND SECURITY INTEREST IN PATENT RIGHTS | 011722 | /0048 | |
Nov 30 2006 | JPMORGAN CHASE BANK, N A FORMERLY KNOWN AS THE CHASE MANHATTAN BANK , AS ADMINISTRATIVE AGENT | Lucent Technologies Inc | TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENT RIGHTS | 018590 | /0047 | |
Nov 01 2008 | Lucent Technologies Inc | Alcatel-Lucent USA Inc | MERGER SEE DOCUMENT FOR DETAILS | 033053 | /0885 | |
Jun 30 2014 | Alcatel Lucent | Sound View Innovations, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 033416 | /0763 | |
Jan 03 2018 | Alcatel-Lucent USA Inc | Nokia of America Corporation | CHANGE OF NAME SEE DOCUMENT FOR DETAILS | 050476 | /0085 | |
Sep 27 2019 | Nokia of America Corporation | Alcatel Lucent | NUNC PRO TUNC ASSIGNMENT SEE DOCUMENT FOR DETAILS | 050668 | /0829 |
Date | Maintenance Fee Events |
Feb 27 2002 | M183: Payment of Maintenance Fee, 4th Year, Large Entity. |
Mar 09 2002 | ASPN: Payor Number Assigned. |
Mar 06 2006 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Jun 19 2007 | ASPN: Payor Number Assigned. |
Jun 19 2007 | RMPN: Payer Number De-assigned. |
Mar 24 2010 | M1553: Payment of Maintenance Fee, 12th Year, Large Entity. |
Date | Maintenance Schedule |
Sep 29 2001 | 4 years fee payment window open |
Mar 29 2002 | 6 months grace period start (w surcharge) |
Sep 29 2002 | patent expiry (for year 4) |
Sep 29 2004 | 2 years to revive unintentionally abandoned end. (for year 4) |
Sep 29 2005 | 8 years fee payment window open |
Mar 29 2006 | 6 months grace period start (w surcharge) |
Sep 29 2006 | patent expiry (for year 8) |
Sep 29 2008 | 2 years to revive unintentionally abandoned end. (for year 8) |
Sep 29 2009 | 12 years fee payment window open |
Mar 29 2010 | 6 months grace period start (w surcharge) |
Sep 29 2010 | patent expiry (for year 12) |
Sep 29 2012 | 2 years to revive unintentionally abandoned end. (for year 12) |