There is disclosed a method and apparatus for providing continuous speech-to-subtitles translation utilizing a video-based communications device but without speech synthesis at the output. Instead, a translation of each user's speech is displayed continuously in text form on the other user's screen. In the preferred embodiment, the sending party speaks into a conventional videophone. speech recognition and translation of the transmitted signal are performed by a remote device at the receiving party's location. The audio portion of the signal is sent both to a speaker for audio output and to a speech recognizer and text-based translation system, the output of which is text translated into the target language. The video portion of the signal and the translated text are combined in a subtitle generator and sent to a display device for viewing by the receiving party.

Patent
   5815196
Priority
Dec 29 1995
Filed
Dec 29 1995
Issued
Sep 29 1998
Expiry
Dec 29 2015
Assg.orig
Entity
Large
123
2
all paid
5. A method of providing continuous speech-to-subtitle translation of a signal containing a video portion and an audio portion, comprising the steps of:
(e) converting said audio portion to a corresponding first textual signal at a sending party's location;
(f) translating said corresponding first textual signal to a second textual signal at a location remote from said sending party's location, wherein said second textual signal is in a target language; and
(g) combining said video portion with said second textual signal to form a display signal, wherein said display signal displays said second textual signal as subtitles.
1. An apparatus for providing continuous speech-to-subtitle translation of a signal containing a video portion, and an audio portion comprising:
(a) means for converting said audio portion to a corresponding first textual signal, wherein said converting means is located at a sending party's location;
(b) means for translating said corresponding first textual signal to a second textual signal wherein said second textual signal is in a target language and wherein said translating means is located remotely from said sending party's location;
(c) means for combining said video portion with said second textual signal to form a display signal, wherein said display signal displays said second textual signal as subtitles; and
(d) means for simultaneously displaying said display signal and outputting said audio portion.
2. An apparatus according to claim 1, wherein said converting means comprises a speech recognizer.
3. An apparatus according to claim 1, wherein said translating means comprises a text-based machine translation system.
4. An apparatus according to claim 1, wherein said combining means comprises a subtitle generator.
6. A method according to claim 5, further comprising the step of displaying said display signal and outputting said audio portion simultaneously.
7. A method according to claim 5, wherein said converting step is performed by a speech recognizer.
8. A method according to claim 5, wherein said translating step is performed by a text-based machine translation system.
9. A method according to claim 5, wherein said combining step is performed by a subtitle generator.

This invention relates to a method and apparatus for providing continuous speech-to-subtitles translation for communication between people speaking different languages.

As the world moves closer and closer to a true global economy, the need for individuals who speak different languages to be able to easily communicate has increased. Efforts have been made to facilitate communication between people speaking different languages using current speech-to-speech translation technology wherein the translated speech is synthesized in the target language.

Current speech-to-speech translation technology operates such that language is translated and synthesized sentence by sentence or phrase by phrase. Typically in such systems, a user speaks an entire sentence or phrase and presses a button or flips a switch when completed. The device then translates the entire sentence or phrase and synthesizes and outputs the translation in the target language. Thus, the other party must wait for the speech synthesizer to stop before responding. Such systems are currently preferred because they include in the synthesized translation the intonations, often related to emotion, contained in the original speech. Such intonations are generally thought to increase the quality of the communication. The result, however, is a delay between one party speaking and the other party hearing the synthesized translation which can make communication awkward and unnatural. A system which instead translates and synthesizes one word at a time would most likely also sound awkward and unnatural and would lack the normal intonations of speech. Thus, although research and development efforts are aimed at eliminating the current limitations of speech-to-speech translation technology, it is unlikely that any resulting systems will be capable of perfect simultaneous translation and speech synthesis for many years.

The problems and limitations associated with speech-to-speech translation are avoided, in accordance with the principles of the present invention, by using a video-based communication device for speech translation, but without speech synthesis of the output. Instead, a translation of each user's speech is displayed continuously in text form on the other user's screen. At the same time, the original, untranslated speech is played over a speaker.

In the preferred embodiment, the sending party speaks into a conventional prior art videophone. The output of the videophone, a signal consisting of both the audio and video portions of the communication, is transmitted to the receiving party's location. Speech recognition and translation of the transmitted signal are performed by a remote device at the receiving party's location. The audio portion of the signal is sent both to a speaker for audio output and to a speech recognizer and text-based machine translation system, the output of which is translated text. The video portion of the signal and the translated text are then combined in a subtitle generator and sent to a display device for viewing by the receiving party.

Because users hear the actual voice of the other party, the communication is more personal and is likely to be perceived to be of higher quality. Hearing the original speech can also reduce misunderstanding because emotional clues are available to the listener. Also, in the event that an imperfect translation takes place, users can look over the stream of translated words and make use of their knowledge of the other language to try to reconstruct the intended meaning. Moreover, because the original untranslated speech is provided audibly and the translated text is provided visually, users can employ any knowledge they may have of the original language. Finally, according to the preferred embodiment, the sending party need not be aware that the other party has translation subtitles displayed on their screen, a feature that would be appreciated by users embarrassed about their foreign language skills.

FIG. 1 is a block diagram of an embodiment of the present invention wherein recognition, translation and subtitle generation are performed remotely.

FIG. 2 is a flow diagram illustrating the method for providing continuous speech-to-subtitles translation.

FIG. 3 is a block diagram of an embodiment of the present invention wherein recognition, translation and subtitle generation are performed by a telephone service provider network.

Referring to FIG. 1, a diagram of the presently preferred embodiment of the system is shown. Conventional prior art videophone 5, such as the AT&T VT2500, is located at the sending party's location. Remote receiving device 8 is located at the receiving party's location. The sending party speaks into videophone 5 which contains camera 9 and microphone 11. Camera 9 outputs video signal 10, which represents the visual component of the communication, and microphone 11 outputs audio signal 12, which represents the speech component of the communication. Video signal 10 and audio signal 12 are fed into audio/video encoder 13. Audio/video encoder 13 is a conventional digital signal processing device which can be found in the transmitting end of the AT&T VT2500. Audio/Video encoder 13 converts video signal 10 and audio signal 12 into a single encoded digital signal 14. Encoded digital signal 14 is placed on conventional telephone line 15 for transmission to remote receiving device 8.

At remote receiving device 8, encoded digital signal 14 is fed into audio/video decoder 16. Similar to audio/video encoder 13, audio/video decoder 16 is a conventional digital signal processing device which can be found in the receiving end of the AT&T VT2500. Audio/video decoder 16 converts encoded digital signal 14 back into two separate signals, video signal 23 and audio signal 17. Audio signal 17, which is in the original language of the sending party, is simultaneously fed into speaker 18 for audio output to the receiving party and into recognizer 19. Recognizer 19 is a conventional speech recognizer which converts human speech to text. Speech recognizers are well known in the art and are described, for example, in L. R. Rabiner and B. H. Juang, Fundamentals of Speech Recognition, Prentice-Hall (1993). Using a prior art statistical pattern recognition technique, recognizer 19 converts audio signal 17 into recognition hypothesis 20, which consists of one or more possible sequences of words in text format corresponding to audio signal 17. Essentially, recognition hypothesis 20 is a signal representing the most likely textual counterpart to the spoken language represented by audio signal 17. Recognition hypothesis 20, however, is still in the original language of the sending party, and thus needs to be translated into the target language.

Recognition hypothesis 20 is sent to translator 21. Translator 21 is a conventional text-based machine translation system which converts text in one natural language to text in another natural language. Text-based machine translation systems are well known in the art and are described, for example, in W. J. Hutchins and H. L. Somers, An Introduction to Machine Translation, Academic Press (1992). Translator 21 takes recognition hypothesis 20 and translates it into the target language. If recognition hypothesis 20 consists of a set of possible sequences of words, translator 21 applies a language model which chooses the most likely grammatical version for translation. The output of translator 21 is text signal 22.

Text signal 22 and video signal 23 are sent to subtitle generator 24 where the two signals are overlaid onto one another to create display signal 25. In display signal 25, text signal 22 appears as subtitles to video signal 23. Subtitle generator 24 is common in the prior art, especially in the film industry.

Subtitle generator 24 outputs display signal 25 which in turn is sent to video display device 26, such as a monitor, for display to the receiving party. Thus, the receiving party can simultaneously hear the original speech of the sending party and view the video of the sending party overlaid with subtitles translating the sending party's speech.

Because the recognition and translation functions are performed at remote receiving device 8, standard existing videophone transmission signals and protocols, such as ITU-H.261 and IT-TH.263, can be used. As such, the only party that needs to have any special equipment is the receiving party. The sending party need only use a standard videophone and need not know that the translation and subtitling are taking place.

Referring now to FIG. 2, a flow diagram illustrating the method for providing continuous speech-to-subtitles translation is shown. Block 30 shows that after a sending party speaks into his or her videophone, the encoded signal output thereby is sent to the receiving party's location. As shown in Block 32, the encoded signal is then decoded into an audio signal and a video signal. Block 34 shows that the audio signal is then converted into a corresponding textual signal in the sending party's language. Block 36 shows that the textual signal is then translated into a textual signal in the receiving party's language, also known as the target language. As shown in block 38, the textual signal in the target language is then overlaid onto the video signal as subtitles. Finally, as shown in block 40, the video signal with subtitles overlaid thereon and the audio signal are simultaneously output to the receiving party.

Referring now to FIG. 3, an alternate embodiment of the system is shown wherein the recognition, translation and subtitle generation functions are performed by telephone service provider network 60 rather than at the receiving party's location. Conventional videophone 62 located at the sending party's location 64 outputs signal 66. Signal 66 is a standard videophone signal. Signal 66 is sent to central processing unit, or CPU, 68 which is attached to telephone network switch 70. CPU 68 and telephone network switch 70 are part of telephone service provider network 60. CPU 68 contains algorithms which perform the recognition, translation and subtitle generation functions on signal 66. CPU 68 outputs signal 72. Signal 72 consists of an audio portion, which contains the sending party's original speech, and a subtitled video portion. Signal 72 is sent to conventional videophone 74 located at the receiving party's location 76 where it can be viewed by the receiving party.

In this embodiment, the continuous speech-to-subtitles translation would be provided as a service by the telephone service provider wherein the user is charged a fee for each use. As such, a person desiring speech-to-subtitle translation could access the service as needed, using a conventional videophone. This embodiment would also allow the use of standard videophone transmission signals and protocols. Finally, like the preferred embodiment, this embodiment would allow the receiving party to use the service without the knowledge of the sending party.

Still further alternate embodiments of the system are possible. One such embodiment would entail performing the recognition and translation functions at the sending party's locations. Another such embodiment would entail performing the recognition function at the sending party's location and the translation function at the receiving party's location. Both such alternate embodiments would require the transmission of data, i.e., the translated text in the case of the former and the recognition hypothesis in the case of the latter, in addition to the transmission of audio and video signals. Thus, these alternate embodiments would require the use of modified videophone equipment. These alternate embodiments, however, would have the advantage of providing more accurate recognition than in the preferred embodiment because the recognition function is performed locally rather than at the receiving party's location. Local recognition is more accurate because the audio signal does not have to be transmitted over telephone lines before recognition takes place. Conversely, in the preferred embodiment, when encoded digital signal 14 is sent to the receiving party's location, the quality of audio signal 17 is somewhat diminished due to the limited bandwidth of conventional telephone line 15.

It is to be understood that the above description comprises only a few of the possible embodiments of the present invention. Numerous other arrangements may be devised by one skilled in the art without departing from the spirit and scope of the invention. The invention is thus limited only as defined in the accompanying claims.

Alshawi, Hiyan

Patent Priority Assignee Title
10015311, Jun 29 2005 Ultratec, Inc. Device independent text captioned telephone service
10051207, Nov 12 2015 SORENSON IP HOLDINGS, LLC Captioning communication systems
10140433, Aug 03 2001 Comcast IP Holdings I, LLC Video and digital multimedia aggregator
10349096, Aug 03 2001 Comcast IP Holdings I, LLC Video and digital multimedia aggregator content coding and formatting
10372831, Dec 12 2011 GOOGLE LLC Auto-translation for multi user audio and video
10389876, Feb 28 2014 ULTRATEC, INC Semiautomated relay method and apparatus
10469660, Jun 29 2005 Ultratec, Inc. Device independent text captioned telephone service
10491746, Feb 18 2004 Ultratec, Inc. Captioned telephone service
10542141, Feb 28 2014 ULTRATEC, INC Semiautomated relay method and apparatus
10587751, Feb 18 2004 Ultratec, Inc. Captioned telephone service
10614173, Dec 12 2011 GOOGLE LLC Auto-translation for multi user audio and video
10643036, Aug 18 2016 HYPERCONNECT INC Language translation device and language translation method
10742805, Feb 28 2014 ULTRATEC, INC Semiautomated relay method and apparatus
10748523, Feb 28 2014 ULTRATEC, INC Semiautomated relay method and apparatus
10878721, Feb 28 2014 ULTRATEC, INC Semiautomated relay method and apparatus
10917519, Feb 28 2014 Ultratec, Inc. Semiautomated relay method and apparatus
10972604, Jun 29 2005 Ultratec, Inc. Device independent text captioned telephone service
10972683, Nov 12 2015 SORENSON IP HOLDINGS, LLC Captioning communication systems
11005991, Feb 18 2004 Ultratec, Inc. Captioned telephone service
11190637, Feb 18 2004 Ultratec, Inc. Captioned telephone service
11227129, Aug 18 2016 HYPERCONNECT INC Language translation device and language translation method
11258900, Jun 29 2005 Ultratec, Inc. Device independent text captioned telephone service
11368581, Feb 28 2014 Ultratec, Inc. Semiautomated relay method and apparatus
11509838, Nov 12 2015 SORENSON IP HOLDINGS, LLC Captioning communication systems
11539900, Feb 21 2020 ULTRATEC, INC Caption modification and augmentation systems and methods for use by hearing assisted user
11627221, Feb 28 2014 Ultratec, Inc. Semiautomated relay method and apparatus
11664029, Feb 28 2014 ULTRATEC, INC Semiautomated relay method and apparatus
11741963, Feb 28 2014 ULTRATEC, INC Semiautomated relay method and apparatus
5978014, Sep 19 1997 8x8, Inc Video TTY device and method for videoconferencing
5991711, Feb 26 1996 Fuji Xerox Co., Ltd. Language information processing apparatus and method
5991723, Jan 17 1998 Genesys Telecommunications Laboratories, Inc Method and apparatus for translating text and speech transferred over a telephony or similar network
6195631, Apr 15 1998 Nuance Communications, Inc Method and apparatus for automatic construction of hierarchical transduction models for language translation
6323892, Aug 04 1998 Olympus Corporation Display and camera device for videophone and videophone apparatus
6513003, Feb 03 2000 REFINITIV US ORGANIZATION LLC System and method for integrated delivery of media and synchronized transcription
6542200, Aug 14 2001 CHELDAN TECHNOLOGIES, INC Television/radio speech-to-text translating processor
6599130, Feb 02 2001 TAMABO INC Iterative video teaching aid with recordable commentary and indexing
6754619, Nov 15 1999 Sony Corporation Digital recording and playback system with voice recognition capability for concurrent text generation
7013273, Mar 29 2001 Matsushita Electric Industrial Co., Ltd. Speech recognition based captioning system
7047191, Mar 06 2000 ENCO SYSTEMS, INC Method and system for providing automated captioning for AV signals
7110946, Nov 12 2002 United States of America as represented by the Secretary of the Navy Speech to visual aid translator assembly and method
7191117, Jun 09 2000 British Broadcasting Corporation Generation of subtitles or captions for moving pictures
7295969, Nov 15 1999 Sony Corporation; Sony Electronics, Inc. Digital recording and playback system with voice recognition capability for concurrent text generation
7310605, Nov 25 2003 Microsoft Technology Licensing, LLC Method and apparatus to transliterate text using a portable device
7359849, Dec 17 2003 SPEECHGEAR, INC Translation techniques for acronyms and ambiguities
7532229, Nov 14 2003 NEC Corporation Data processor, data processing method and electronic equipment
7552053, Aug 22 2005 International Business Machines Corporation Techniques for aiding speech-to-speech translation
7734467, Aug 22 2005 International Business Machines Corporation Techniques for aiding speech-to-speech translation
8112270, Nov 15 1999 Sony Corporation; Sony Electronics, Inc. Digital recording and playback system with voice recognition capability for concurrent text generation
8319819, Mar 26 2008 Cisco Technology, Inc.; Cisco Technology, Inc Virtual round-table videoconference
8326596, May 28 2009 Method and apparatus for translating speech during a call
8355041, Feb 14 2008 Cisco Technology, Inc. Telepresence system for 360 degree video conferencing
8390667, Apr 15 2008 Cisco Technology, Inc. Pop-up PIP for people not in picture
8416925, Jun 29 2005 Ultratec, Inc. Device independent text captioned telephone service
8472415, Mar 06 2006 Cisco Technology, Inc. Performance optimization with integrated mobility and MPLS
8477175, Mar 09 2009 Cisco Technology, Inc. System and method for providing three dimensional imaging in a network environment
8515024, Jan 13 2010 Ultratec, Inc. Captioned telephone service
8542264, Nov 18 2010 Cisco Technology, Inc. System and method for managing optics in a video environment
8570373, Jun 08 2007 Cisco Technology, Inc. Tracking an object utilizing location information associated with a wireless device
8599865, Oct 26 2010 Cisco Technology, Inc. System and method for provisioning flows in a mobile network environment
8599934, Sep 08 2010 Cisco Technology, Inc. System and method for skip coding during video conferencing in a network environment
8610755, Feb 18 2011 SORENSON IP HOLDINGS LLC Methods and apparatuses for multi-lingual support for hearing impaired communication
8635070, Sep 29 2010 Kabushiki Kaisha Toshiba Speech translation apparatus, method and program that generates insertion sentence explaining recognized emotion types
8659637, Mar 09 2009 Cisco Technology, Inc. System and method for providing three dimensional video conferencing in a network environment
8659639, May 29 2009 Cisco Technology, Inc. System and method for extending communications between participants in a conferencing environment
8670019, Apr 28 2011 Cisco Technology, Inc. System and method for providing enhanced eye gaze in a video conferencing environment
8682087, Dec 19 2011 Cisco Technology, Inc. System and method for depth-guided image filtering in a video conference environment
8692862, Feb 28 2011 Cisco Technology, Inc. System and method for selection of video data in a video conference environment
8694658, Sep 19 2008 Cisco Technology, Inc. System and method for enabling communication sessions in a network environment
8699457, Nov 03 2010 Cisco Technology, Inc. System and method for managing flows in a mobile network environment
8701020, Feb 01 2011 GOOGLE LLC Text chat overlay for video chat
8723914, Nov 19 2010 Cisco Technology, Inc.; Cisco Technology, Inc System and method for providing enhanced video processing in a network environment
8730297, Nov 15 2010 Cisco Technology, Inc. System and method for providing camera functions in a video environment
8768699, Aug 22 2005 International Business Machines Corporation Techniques for aiding speech-to-speech translation
8786631, Apr 30 2011 Cisco Technology, Inc. System and method for transferring transparency information in a video environment
8797377, Feb 14 2008 Cisco Technology, Inc. Method and system for videoconference configuration
8896655, Aug 31 2010 Cisco Technology, Inc.; University of North Carolina at Chapel Hill System and method for providing depth adaptive video conferencing
8902244, Nov 15 2010 Cisco Technology, Inc. System and method for providing enhanced graphics in a video environment
8908838, Aug 23 2001 ULTRATEC, INC System for text assisted telephony
8917821, Jun 29 2005 Ultratec, Inc. Device independent text captioned telephone service
8917822, Aug 23 2001 ULTRATEC, INC System for text assisted telephony
8934026, May 12 2011 Cisco Technology, Inc. System and method for video coding in a dynamic environment
8947493, Nov 16 2011 Cisco Technology, Inc. System and method for alerting a participant in a video conference
9031849, Sep 30 2006 Huawei Technologies Co., Ltd. System, method and multipoint control unit for providing multi-language conference
9082297, Aug 11 2009 Cisco Technology, Inc. System and method for verifying parameters in an audiovisual environment
9110891, Dec 12 2011 GOOGLE LLC Auto-translation for multi user audio and video
9111138, Nov 30 2010 Cisco Technology, Inc. System and method for gesture interface control
9131045, Aug 23 2001 ULTRATEC, INC System for text assisted telephony
9143725, Nov 15 2010 Cisco Technology, Inc. System and method for providing enhanced graphics in a video environment
9204096, May 29 2009 Cisco Technology, Inc. System and method for extending communications between participants in a conferencing environment
9225916, Mar 18 2010 Cisco Technology, Inc. System and method for enhancing video images in a conferencing environment
9301057, Jan 17 2014 HEARGLASS, INC Hearing assistance system
9313452, May 17 2010 Cisco Technology, Inc. System and method for providing retracting optics in a video conferencing environment
9331948, Oct 26 2010 Cisco Technology, Inc. System and method for provisioning flows in a mobile network environment
9338394, Nov 15 2010 Cisco Technology, Inc. System and method for providing enhanced audio in a video environment
9374536, Nov 12 2015 SORENSON IP HOLDINGS, LLC Video captioning communication system, devices and related methods for captioning during a real-time video communication session
9380374, Jan 17 2014 HEARGLASS, INC Hearing assistance systems configured to detect and provide protection to the user from harmful conditions
9525830, Nov 12 2015 SORENSON IP HOLDINGS, LLC Captioning communication systems
9681154, Dec 06 2012 PATENT CAPITAL GROUP System and method for depth-guided filtering in a video conference environment
9683862, Aug 24 2015 International Business Machines Corporation Internationalization during navigation
9689699, Aug 24 2015 International Business Machines Corporation Internationalization during navigation
9720909, Dec 12 2011 GOOGLE LLC Auto-translation for multi user audio and video
9813641, Jun 19 2000 Comcast IP Holdings I, LLC Method and apparatus for targeting of interactive virtual objects
9843621, May 17 2013 Cisco Technology, Inc. Calendaring activities based on communication processing
9934219, Aug 24 2015 International Business Machines Corporation Internationalization during navigation
9959872, Dec 14 2015 International Business Machines Corporation Multimodal speech recognition for real-time video audio-based display indicia application
9961196, Aug 23 2001 Ultratec, Inc. System for text assisted telephony
9967380, Aug 23 2001 Ultratec, Inc. System for text assisted telephony
9998686, Nov 12 2015 SORENSON IP HOLDINGS, LLC Transcribing video communication sessions
D636359, Mar 21 2010 Cisco Technology, Inc. Video unit with integrated features
D636747, Mar 21 2010 Cisco Technology, Inc. Video unit with integrated features
D637568, Mar 21 2010 Cisco Technology, Inc. Free-standing video unit
D637569, Mar 21 2010 Cisco Technology, Inc. Mounted video unit
D637570, Mar 21 2010 Cisco Technology, Inc. Mounted video unit
D653245, Mar 21 2010 Cisco Technology, Inc. Video unit with integrated features
D655279, Mar 21 2010 Cisco Technology, Inc. Video unit with integrated features
D678307, Dec 16 2010 Cisco Technology, Inc. Display screen with graphical user interface
D678308, Dec 16 2010 Cisco Technology, Inc. Display screen with graphical user interface
D678320, Dec 16 2010 Cisco Technology, Inc. Display screen with graphical user interface
D678894, Dec 16 2010 Cisco Technology, Inc. Display screen with graphical user interface
D682293, Dec 16 2010 Cisco Technology, Inc. Display screen with graphical user interface
D682294, Dec 16 2010 Cisco Technology, Inc. Display screen with graphical user interface
D682854, Dec 16 2010 Cisco Technology, Inc. Display screen for graphical user interface
D682864, Dec 16 2010 Cisco Technology, Inc. Display screen with graphical user interface
Patent Priority Assignee Title
5369704, Mar 24 1993 Engate LLC Down-line transcription system for manipulating real-time testimony
5512938, Apr 06 1994 Matsushita Electric Industrial Co., Ltd. Teleconference terminal
////////
Executed onAssignorAssigneeConveyanceFrameReelDoc
Dec 29 1995Lucent Technologies Inc.(assignment on the face of the patent)
Mar 21 1996ALSHAWI, HIYANLucent Technologies IncASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0078850466 pdf
Feb 22 2001LUCENT TECHNOLOGIES INC DE CORPORATION THE CHASE MANHATTAN BANK, AS COLLATERAL AGENTCONDITIONAL ASSIGNMENT OF AND SECURITY INTEREST IN PATENT RIGHTS0117220048 pdf
Nov 30 2006JPMORGAN CHASE BANK, N A FORMERLY KNOWN AS THE CHASE MANHATTAN BANK , AS ADMINISTRATIVE AGENTLucent Technologies IncTERMINATION AND RELEASE OF SECURITY INTEREST IN PATENT RIGHTS0185900047 pdf
Nov 01 2008Lucent Technologies IncAlcatel-Lucent USA IncMERGER SEE DOCUMENT FOR DETAILS 0330530885 pdf
Jun 30 2014Alcatel LucentSound View Innovations, LLCASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0334160763 pdf
Jan 03 2018Alcatel-Lucent USA IncNokia of America CorporationCHANGE OF NAME SEE DOCUMENT FOR DETAILS 0504760085 pdf
Sep 27 2019Nokia of America CorporationAlcatel LucentNUNC PRO TUNC ASSIGNMENT SEE DOCUMENT FOR DETAILS 0506680829 pdf
Date Maintenance Fee Events
Feb 27 2002M183: Payment of Maintenance Fee, 4th Year, Large Entity.
Mar 09 2002ASPN: Payor Number Assigned.
Mar 06 2006M1552: Payment of Maintenance Fee, 8th Year, Large Entity.
Jun 19 2007ASPN: Payor Number Assigned.
Jun 19 2007RMPN: Payer Number De-assigned.
Mar 24 2010M1553: Payment of Maintenance Fee, 12th Year, Large Entity.


Date Maintenance Schedule
Sep 29 20014 years fee payment window open
Mar 29 20026 months grace period start (w surcharge)
Sep 29 2002patent expiry (for year 4)
Sep 29 20042 years to revive unintentionally abandoned end. (for year 4)
Sep 29 20058 years fee payment window open
Mar 29 20066 months grace period start (w surcharge)
Sep 29 2006patent expiry (for year 8)
Sep 29 20082 years to revive unintentionally abandoned end. (for year 8)
Sep 29 200912 years fee payment window open
Mar 29 20106 months grace period start (w surcharge)
Sep 29 2010patent expiry (for year 12)
Sep 29 20122 years to revive unintentionally abandoned end. (for year 12)