(71) Applicant: SPEAKWRITE, LLC,Austin, TX (US)



Similar documents
(12) Patent Application Publication (10) Pub. No.: US 2003/ A1 Wu et al. (43) Pub. Date: Feb. 20, 2003

(Us) (73) Assignee: Avaya Technology Corp. Je?' McElroy, Columbia, SC (US); (21) Appl. No.: 10/413,024. (22) Filed: Apr. 14, 2003 (57) ABSTRACT

Hay (43) Pub. Date: Oct. 17, 2002

US Al (19) United States (12) Patent Application Publication (10) Pub. No.: US 2009/ A1 Sanvido (43) Pub. Date: Jun.

/ \33 40 \ / \\ \ \ M / f 1. (19) United States (12) Patent Application Publication Lawser et al. NETWORK \ 36. SERVlCE 'NTERNET SERVICE

(12) Patent Application Publication (10) Pub. No.: US 2013/ A1 Kim et al. (43) Pub. Date: Dec. 5, 2013

\ \ \ connection connection connection interface interface interface

software, and perform automatic dialing according to the /*~102

(72) Inventors: Juergen RIEDL, Koenigsbrunn (DE); USPC ( 267/285)

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2002/ A1 STRANDBERG (43) Pub. Date: Oct.

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2013/ A1 Chen (57)

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2007/ A1 Du et al. (43) Pub. Date: Aug.

US A1 (19) United States (12) Patent Application Publication (10) Pub. N0.: US 2013/ A1 DANG (43) Pub. Date: Jul.

US A1 (19) United States (12) Patent Application Publication (10) Pub. N0.: US 2002/ A1. Mannarsamy (43) Pub. Date: NOV.

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2005/ A1 Owhadi et al. (43) Pub. Date: Feb.

(21) (22) (57) ABSTRACT. Appl. No.: 10/752,736

(12) United States Patent (16) Patent N6.= US 6,198,814 B1 Gill (45) Date of Patent: Mar. 6, 2001

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2007/ A1 Ollis et al. HOME PROCESSOR /\ J\ NETWORK

(30) Foreign Application Priority Data

(12) United States Patent (10) Patent N0.: US 8,282,471 B1 Korner (45) Date of Patent: Oct. 9, 2012

United States Patent [191

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2002/ A1 Fukuzato (43) Pub. Date: Jun.

(12> Ulllted States Patent (16) Patent N6.= US 6,320,621 B1 Fu (45) Date of Patent: Nov. 20, 2001

NJ (US) (51) Int. Cl. H04L 9/00 ( ) Correspondence Address: (52) US. Cl /278; 713/ 150 ALFRED C. ROTH (57) ABSTRACT

(12) United States Patent Wen et a].

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2013/ A1 Warren (43) Pub. Date: Jan.

US Al (19) United States (12) Patent Application Publication (10) Pub. No.: US 2013/ A1 White et al. TROUBLE TICKET SYSTEM

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2005/ A1 Reuveni (43) Pub. Date: NOV.

1,5 F., n M 3 My MM, 3 2. M5, ' (21) App1.N0.: 13/789,334 M/WMWW W ~ 3> ( I INTERNET < 114. (71) ApplicantszRobert Monster, Sammamish, WA

Vignet (43) Pub. Date: Nov. 24, 2005

TEPZZ 6_Z76 A_T EP A1 (19) (11) EP A1 (12) EUROPEAN PATENT APPLICATION. (51) Int Cl.:

Filetto et al. [45] Date of Patent: Feb. 15, 2000

Place your bets please.

(54) METHODS AND SYSTEMS FOR FINDING Publication Classi?cation CONNECTIONS AMONG SUBSCRIBERS TO AN CAMPAIGN (51) Int- Cl

(12) United States Patent (10) Patent N0.: US 8,326,445 B2 Baak et al. (45) Date of Patent: Dec. 4, 2012

(54) (71) (72) Vedelago (TV) (IT) (73) (21) (22) (30) Chirignago (VE) (IT); Alberto Al?er, Foreign Application Priority Data

(12) United States Patent (10) Patent No.: US 8,253,226 B2 Oguri (45) Date of Patent: Aug. 28, 2012

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2012/ A1 Chung (43) Pub. Date: Aug.

Levy Processing System 1_Q

110-\ CALLER TERMINAL

United States Patent [191 Romo et al.

US A1 (19) United States (12) Patent Application Publication (10) Pub. N0.: US 2007/ A1 Makowka (43) Pub. Date: Feb.

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2006/ A1 Clegg (43) Pub. Date: Sep.

US Al (19) United States (12) Patent Application Publication (10) Pub. No.: US 2012/ A1 Kuehl (43) Pub. Date: Aug.

US A1 (19) United States (12) Patent Application Publication (10) Pub. No.: US 2013/ A1 Fan et al.

(54) Applicant: (71) (72) Assignee: (73) (21) (22) (60)

(12) United States Patent (16) Patent N6.= US 6,611,861 B1 Schairer et al. (45) Date of Patent: Aug. 26, 2003

g 0 No 17 Personalize Message 26 > Fig. 5 '2 i 22 2 Approve Message 12 > Fig. 2

USOO A United States Patent [191 [11] Patent Number: 5,469,362. Hunt et al. [45] Date of Patent: Nov. 21, 1995

(54) SYSTEM AND METHOD FOR RING DELAY (52) US. Cl /252 ON A USER AGENT

US B1 (12) United States Patent. (10) Patent N0.: US 6,282,278 B1 D0ganata et al. (45) Date 0f Patent: Aug. 28, 2001

Transcription:

US 20130304465Al (19) United States (12) Patent Application Publication (10) Pub. No.: US 2013/0304465 A1 Henry et al. (43) Pub. Date: NOV. 14, 2013 (54) METHOD AND SYSTEM FOR AUDIO-VIDEO (52) US. Cl. INTEGRATION CPC..... G10L 15/265 (2013.01) (71) Applicant: SPEAKWRITE, LLC,Austin, TX (US) USPC..... 704/235 (72) Inventors: Paul Henry, Austin, TX (U S); Richard Jackson, Austin, TX (US) (57) ABSTRACT 73 ( ) A' :S kw't LLCAt'TXUS sslgnee Pea n e us In ( ) A method of dictation is described Which, after transcription, (21) App1_ NO; 13/889,990 integrates video into text at locations designated by the user. A user collects audio and visual information using an appli (22) Filed: May 8, 2013 cation that has been installed on the user s mobile device. The user designates, using the application, the desired location of Related U-s- Application Data the video?les Within the audio?le. Both the audio and video (60) Provisional application No 61/644,044?led on May 8, 2012 information are uploaded to a transcription provider. The transcription provider uses transcription software to tran Publication Classi?cation scribe the audio?les into text, the transcription software being able to identify the transcriptionist as to the location Within the audio?le each video is to be inserted. The tran (51) Int. Cl. scribed document With integrated audio and video is then G10L 15/26 (2006.01) delivered back to the user. User collects Both the audio and Transcription provider uses Document is audio and video video are upioaded transcription software to deiivered back to with the to the transcription transcribe audio and insert user application provider video into document

Patent Application Publication Nov. 14, 2013 Sheet 1 0f 7 US 2013/0304465 A1 FIGURE 1 User coiiects audio and video with the application Both the audio and video are uploaded to the transcription provider Transcription provider uses transcription software to transcribe audio and insert video into document Document is delivered back to user

Patent Application Publication Nov. 14, 2013 Sheet 2 0f 7 US 2013/0304465 A1 FIGURE 2 204 201 203 202 205 206 220 207 210 209 208 211 212 FIGURE 3

Patent Application Publication Nov. 14, 2013 Sheet 3 0f 7 US 2013/0304465 A1 FIGURE 4A FIGURE 43

Patent Application Publication Nov. 14, 2013 Sheet 4 0f 7 US 2013/0304465 A1 FIGURE 5

Patent Application Publication Nov. 14, 2013 Sheet 5 0f 7 US 2013/0304465 A1 FIGURE 6

Patent Application Publication Nov. 14, 2013 Sheet 6 0f 7 US 2013/0304465 A1 FIGURE 7,,,,,,,,,,,,,,,,,, oftware The transc_r'mmn WIH pause Typist submits V the V i_ Typist plays aud=o nd a a bee a Typlat Inserts the transcnbed L ~)- and transcnhes. the --- -~} p v p f- } image into the,1 I " back to lcmre was audia located. _ ezectmmc dacumem. the transcnptmn the current audlo wider paaitlon FIGURE 8

Patent Application Publication Nov. 14, 2013 Sheet 7 0f 7 US 2013/0304465 A1 FIGURE 9 a mmgsiie an e559:3z'sniaéammemwith gm? ' 9592c; fmssi a: arsm i 951mm. 15mm) mom taxi}

US 2013/0304465 A1 Nov. 14, 2013 METHOD AND SYSTEM FOR AUDIO-VIDEO INTEGRATION CROSS-REFERENCE TO RELATED APPLICATIONS [0001] This non-provisional application claims priority based upon prioru.s. Provisional PatentApplication Ser. No. 61/644,044?led May 8, 2012 in the names ofpaul Henry and Richard Jackson entitled Smartphone Photo-Transcription Integration, the disclosure of Which is incorporated herein in its entirety by reference as if fully set forth herein. BACKGROUND OF THE INVENTION [0002] Dictation and transcription systems known in the art have a number of shortcomings. For example, existing sys tems do not allow for a user to dictate audio information into a mobile device, add video in speci?c locations, upload that resulting recorded audio and video to an offsite transcription ist to be transcribed With the video inserted in the proper locations so that, after transcription, the completed, tran scribed audio and video?le is delivered to the user, or to a location designated by the user. [0003] There is a need, therefore, for a dictation and tran scription system that allows for the e?icient capture of audio and video?les through an application on a mobile device, With the corresponding ability to send the?les to a transcrip tion provider Who can effectively integrate the video?les With the audio?les as intended by the user. SUMMARY OF THE INVENTION [0004] This invention includes the ability to capture audio and video from a mobile device, to provide the audio and video to an offsite transcription provider, and to include the videos in the transcription provided by the transcription pro vider at the locations designated by the user. More speci? cally a user collects audio and visual information using an application that has been installed on the user s mobile device. The user designates, using the application, the desired location of the video?les Within the audio?le. Both the audio and video information are uploaded to the transcription pro vider. The transcription provider uses transcription software to transcribe the audio?les to text, the transcription software being able to identify the transcriptionist as to the location Within the audio?le each video is to be inserted. The tran scribed document With integrated text and video is then deliv ered back to the user. [0005] The foregoing has outlines rather broadly certain aspects of the present invention in order that the detailed description of the invention that follows may better be under stood. Additional features and advantages of the invention Will be described hereinafter Which form the subject of the claims of the invention. It should be appreciated by those skilled in the art that the conception and speci?c embodiment disclosed may be readily utilized as a basis for modifying or designing other structures or processes for carrying out the same purposes of the present invention. It should also be realized by those skilled in the art that such equivalent con structions do not depart from the spirit and scope of the invention as set forth in the appended claims. BRIEF DESCRIPTION OF THE DRAWINGS [0006] For a more complete understanding of the present invention, and the advantages thereof, reference is now made to the following descriptions taken in conjunction With the accompanying drawings, in Which: [0007] FIG. 1 depicts an exemplary audio-video integration process according to one embodiment of the present inven tion; [0008] FIG. 2 is a screen shot of the main screen of one embodiment of the application of the present invention; [0009] FIG. 3 is a screen shot of the camera roll of one embodiment of the application of the present invention; [0010] FIGS. 4A and 4B are screen shots of the camera view?nder of one embodiment of the application of the present invention; [0011] FIG. 5 shows screen shots of the record and pause screens of one embodiment of the application of the present invention; [0012] FIG. 6 is a screen shot of the uploading of one embodiment of the application of the present invention; [0013] FIG. 7 depicts an exemplary transcription process according to the one embodiment of the present invention; [0014] FIG. 8 shows the top half of a screen shot showing transcription software functionality in one embodiment of the present invention; and [0015] FIG. 9 shows a transcribed document integrating text and video. DETAILED DESCRIPTION OF THE INVENTION [0016] The present invention is directed to improved meth ods and systems for, among other things, the integration of video into a transcription during dictation. The con?guration and use of the presently preferred embodiments are discussed in detail below. It should be appreciated, however, that the present invention provides many applicable inventive con cepts that can be embodied in a Wide variety of contexts other than photo -transcription integration. Accordingly, the spe ci?c embodiments discussed are merely illustrative of spe ci?c Ways to make and use the invention, and do not limit the scope of the invention. In addition, the following terms shall have the associated meaning When used herein: [0017] application includes a software application Which is installed on, or downloaded to, a mobile device, including an application that provides instruction to the mobile device on the collection, storage and transmission of audio and video?les; [0018] mobile device can include a personal computer, a computer terminal, a personal digital assistant (PDA), smart phone, and/or other types of devices generally known to those skilled in the art that can be communicatively connected to a Wired or Wireless network, including a Wi-? network or a cellular telephone network; [0019] transcription provider includes any person or entity Who provides a service Wherein audio and video con tent is collected from a user and made available for electronic or manual transcription; [0020] transcription software includes software that allows the transcriptionist to combine the audio and video into an electronic document; and [0021] video includes pictures, drawings, graphics, video and any other visual content. [0022] Before a user commences dictating, the user invokes a particular application preinstalled on a mobile device. The application allows the user to record audio and add pictures at locations Within the audio. Then the user then submits the audio and video to the transcription provider.

US 2013/0304465 A1 Nov. 14, 2013 [0023] Referring now to FIG. 1 Which depicts an exemplary integration process according to one embodiment of the present invention in Which a user collects audio and visual information using the application 101. Both the audio and visual information are uploaded to the transcription provid er s servers 102. The transcription provider uses transcription software to transcribe the audio?les and to insert the corre sponding video?les When and Where applicable into an elec tronic document 103. The electronic document is then deliv ered back to the user 104. [0024] Referring now to FIG. 2, Wherein one embodiment of a main screen of an application of the present invention is depicted. The device may have a touch sensitive display, an audio conversion module, a microphone, and network con nectivity technology. The touch-sensitive display may dis play graphical user interfaces Which include interactive con tent, such as the buttons described in more detail below, and non-interactive content. The microphone may include acous tic-to-electric transducers that convert sound Waves into one or more electrical signals. For example, the microphone may convert an electrical signal corresponding to a sound Wave into a digital representation that may be stored as an audio?le on a computer. [0025] A user using the application is presented With the play 208 and record 209 buttons found in other dictation applications known in the art. HoWever, the user is also pre sented With a camera roll button 201 and a camera button 202 Which allow the integration of video into the audio stream. When the user activates the camera roll 201 button, the user is presented With the camera roll resident on the mobile device and the user is allowed to select an existing video using methods known in the art. An example of how the camera roll screen looks in some embodiments is shown on FIG. 3. Once a video is selected by the user, the video is inserted at the current location Within the audio?le and the user is returned to the main screen. The location of the video Within the audio?le is designated by a hash mark 220 on the audio slide bar 206. [0026] Similarly, When the user selects the camera button 202, the user is allowed to take a picture or video With the mobile device. Once a picture or video is taken by the user, the video is inserted at the current location Within the audio?le and the user is returned to the main screen. Once again, the location of the picture or video Within the audio?le is desig nated by a hash mark 220 on the audio slide bar 206. [0027] Referring now back to FIG. 2, showing a audio slide bar 206 and a picture preview WindoW 204. The user can move the audio slide bar 206 to jump to different positions Within the audio?le. The hash mark indicates the location Within the audio?le in Which the video Was placed. As the user moves along the audio slide bar the picture associated With the cor responding portion of the audio is displayed in the picture preview WindoW 204. The picture or video in the picture preview WindoW 204 Will change to the picture at the current audio location. Thus, the user can easily scrub back and forth along the audio slide bar 206 While Watching the video pre view WindoW 204 to see Which video is associated With the corresponding portion of audio. [0028] If the user Wants to edit the picture associated With any portion of the audio, the user can click the edit button in the picture preview WindoW 204. This allows the user to change the current picture With a new picture from the camera roll, or from a new photo taken from the mobile device s camera. [0029] When the user Wants to record additional audio, the user activates the record button 209, Whereupon audio Will be recorded commencing at the current location on the audio slide bar 206. [0030] The main screen of the application also includes features that Would be useful in navigating the recording and storage of audio and video?les, including a picture back button 203 that allows the user to jump to the previous picture and changes the audio position of the audio slide bar 206 to the location at Which the picture shown in the picture preview WindoW Was inserted and a picture next button 205 that allows the user to jump to the next picture and changes the audio position of the audio slide bar 206 to the location at Which the picture shown in the picture preview WindoW Was inserted. [0031] Also, the application provides the user With tradi tional audio recording controls including a back 5 button 207, play button 208, record button 209, erase to end button 210 (Which erases any audio from the current position to the end of the recording), and a save button 211. In addition, the main screen includes a submit button 212 Which, When activated, uploads the audio and video?les to the transcription provider as shown in FIG. 6 [0032] In some embodiments of the invention, after the audio and video?les have been uploaded to the transcription provider, the?les are provided to a transcriptionist for tran scription. The transcription software allows the transcription ist to listen to the audio and insert pictures into an electronic document. As generally shown in FIG. 7, a transcriptionist receives a job, the transcriptionist listens to the audio?le and transcribes the job. At the point in the audio Where the user inserted a video, the transcription software plays a beep, thereby alerting the transcriptionist to insert the video. After inserting the video, the transcriptioni st continues transcribing the audio?le until the entire integrated audio/video?le is completed. [0033] FIG. 8 shows the upper half of a screen shot of one embodiment of the transcription software. Audio bar 1 indi cates the current location in the audio?le to the transcription ist and allows the transcriptionist to jump to different sections Within the audio. Hash marks 7 show the transcriptionist Where videos are to be inserted Within the audio?le. Audio controls 2 allow the transcriptioni st to play and pause the audio as Well as navigate forwards or backwards Within the audio?le. [0034] The transcription software also may include an image dropdown feature 3 Which identi?es all the pictures associated With the audio?le, an insert image button 4 Which inserts the image at the current location, and a skip feature 5 Which allows the transcriptioni st to listen to the audio Without the transcription software pausing at picture locations. [0035] As shown in FIG. 9, after the transcriptionist has uploaded the electronic document, it is delivered back to the user. In some embodiments, clicking on the picture Within the completed document Will navigate to a Web page showing the location Where the picture Was taken. [0036] While the present system and method has been dis closed according to the preferred embodiment of the inven tion, those of ordinary skill in the art Will understand that other embodiments have also been enabled. Even though the foregoing discussion has focused on particular embodiments, it is understood that other con?gurations are contemplated. In particular, even though the expressions in one embodiment or in another embodiment are used herein, these phrases are meant to generally reference embodiment possibilities and

US 2013/0304465 A1 Nov. 14, 2013 are not intended to limit the invention to those particular embodiment con?gurations. These terms may reference the same or different embodiments, and unless indicated other Wise, are combinable into aggregate embodiments. The terms a, an and the mean one or more unless expressly speci?ed otherwise. The term connected means commu nicatively connected unless otherwise de?ned. [0037] When a single embodiment is described herein, it Will be readily apparent that more than one embodiment may be used in place of a single embodiment. Similarly, Where more than one embodiment is described herein, it Will be readily apparent that a single embodiment may be substituted for that one device. [0038] In light of the Wide variety of transcription methods known in the art, the detailed embodiments are intended to be illustrative only and should not be taken as limiting the scope of the invention. Rather, What is claimed as the invention is all such modi?cations as may come Within the spirit and scope of the following claims and equivalents thereto. [0039] None of the description in this speci?cation should be read as implying that any particular element, step or func tion is an essential element Which must be included in the claim scope. The scope of the patented subject matter is de?ned only by the allowed claims and their equivalents. Unless explicitly recited, other aspects of the present inven tion as described in this speci?cation do not limit the scope of the claims. What is claimed is: 1. A method for integrating video into a transcribed docu ment, comprising; providing an application for use on a mobile device; Wherein said application is con?gured so that a user records audio to an audio?le, selects video, and desig nates a location Within said audio?le at Which said video is associated; receiving said audio?le and said video from said user; transcribing, using transcription software, said audio?le to text, Wherein said transcription software displays for a transcriptionist said location Within said audio?le at Which said video is associated; and generating one or more transcriptions in Which said video is integrated into said text at said location designated by said user. 2. The method of claim 1, Wherein said application is downloaded and installed by said user prior to use. 3. The method of claim 1, Wherein said recording utilizes said mobile device s integral microphone. 4. The method of claim 1, Wherein said video is a photo graph. 5. The method of claim 1, Wherein said application includes a graphical user interface con?gured to designate said location Within said audio?le at Which said video is associated by presenting hash marks along an audio slide bar. 6. A system for integrating video into a transcribed docu ment, comprising; an application resident on a mobile device; Wherein said application is con?gured so that a user records audio to an audio?le, selects video, and designates a location Within said audio?le at Which said video is associated; a server, Wherein said audio?le and said video are uploaded from said user; transcription software con?gured to transcribe said audio?le to text, Wherein said transcription software displays for a transcriptionist said location Within said audio?le at Which said video is associated; and one or more transcriptions in Which said video is integrated into said text at said location designated by said user. 7. The system of claim 6, Wherein said application is down loaded and installed by said user prior to use. 8. The system of claim 6, Wherein said recording utilizes said mobile device s integral microphone. 9. The system of claim 6, Wherein said video is a photo graph. 10. The system of claim 6, Wherein said application includes a graphical user interface con?gured to designate said location Within said audio?le at Which said video is associated by presenting hash marks along an audio slide bar. 11. A method for integrating video into a transcribed docu ment, comprising; providing, on a mobile device, means for a user to record audio to an audio?le, select one or more videos, and designate a location Within said audio?le at Which said video is associated; receiving said audio?le and said one or more videos from said user; means for transcribing said audio?le to text, Wherein said means for transcribing displays for a transcriptionist said location Within said audio?le at Which said one or more videos is associated; and means for generating one or more transcriptions in Which said one or more videos are integrated into said text at said locations designated by said user. 12. The method of claim 11, Wherein said means for a user to record audio to an audio?le, select one or more videos, and designate a location Within said audio?le at Which said video is associated is downloaded and installed by said user prior to use. 13. The method of claim 1 1, Wherein said recording utilizes said mobile device s integral microphone. 14. The method of claim 11, Wherein said one or more videos is one or more photographs. 15. The method of claim 11, Wherein said means for a user to record audio to an audio?le, select one or more videos, and designate a location Within said audio?le at Which said video is associated includes a graphical user interface con?gured to designate said location Within said audio?le at Which said one or more videos is associated by presenting hash marks along an audio slide bar.