throbber
(12) United States Patent
`Srinivasan et al.
`
`USOO6357042B2
`US 6,357,042 B2
`(10) Patent No.:
`Mar. 12, 2002
`(45) Date of Patent:
`
`(54) METHOD AND APPARATUS FOR
`MULTIPLEXING SEPARATELY-AUTHORED
`METADATA FOR INSERTION INTO AVIDEO
`DATA STREAM
`
`(76) Inventors: Anand Srinivasan, 12718 Torrey Bluff
`Dr. #155; Mehul Y Shah; Indranil
`Chakraborty, both of 12633 El
`Camino Real #3408; Mohan Mardikar,
`12640 Torrey Bluff Dr. #7; P Venkat
`Rangan, 13011 Callcott Way; Kamal
`Bhadada, 12782 Torrey Bluff Dr. #103,
`all of San Diego, CA (US) 92130
`Subject to any disclaimer, the term of this
`patent is extended or adjusted under 35
`U.S.C. 154(b) by 0 days.
`
`(*) Notice:
`
`(21) Appl. No.: 09/235,781
`(22) Filed:
`Jan. 22, 1999
`Related U.S. Application Data
`(63) Continuation-in-part of application No. 09/225,250, filed on
`Jan. 4, 1999, which is a continuation-in-part of application
`No. 09/218,738, filed on Dec. 22, 1998, which is a continu
`ation-in-part of application No. 09/168,315, filed on Oct. 7,
`1998, which is a continuation-in-part of application No.
`09/160,365, filed on Sep. 24, 1998, which is a continuation
`in-part of application No. 09/154,532, filed on Sep. 16,
`1998.
`(51) Int. Cl. .......................... H04N 7/025; H04N 7/10;
`H04N 7/173; G06F 3/00; G06F 13/00
`(52) U.S. Cl. .......................... 725/32; 72.5/119; 345/328
`(58) Field of Search .......................... 709/217; 345/327,
`345/328; 348/12, 13, 10, 6, 7; 455/6.1,
`6.2, 6.3, 3.1, 3.2; 370/503–519; 725/119,
`143, 32,36; G06F 15/167; HO4N 1/00,
`1/17, 7/025, 7/11
`
`(56)
`
`References Cited
`U.S. PATENT DOCUMENTS
`
`
`
`5,424,770 A * 6/1995 Schmelzer et al. ............ 34.8/9
`5,793.980 A * 8/1998 Glaser et al. ............... 709/231
`5,889,950 A
`3/1999 Kuzma .......................... 34.8/6
`5,894,328 A
`4/1999 Tahara et al. ............... 348/423
`5,917,830 A * 6/1999 Chen et al. .................. 455/6.3
`6,002,393 A * 12/1999 Hite et al. .................. 345/327
`6,018,768 A
`1/2000 Ullman et al. ................ 348/12
`6,034,746 A
`3/2000 Desai et al. ................ 348/423
`6,035,332 A * 3/2000 Ingrassia, Jr. et al. ...... 709/218
`* cited by examiner
`Primary Examiner Andrew Faile
`ASSistant Examiner Vivek Srivastava
`(74) Attorney, Agent, or Firm Donald R. Boys; Central
`Coast Patent Agency, Inc.
`(57)
`ABSTRACT
`An authoring System for interactive Video has two or more
`authoring Stations for providing authored metadata to be
`related to a main Video data Stream and a multiplexer for
`relating authored metadata from the authoring Sources to the
`main Video data Stream. The authoring Stations annotate
`created metadata with presentation time stamps (PTS) from
`the main Video Stream, and the multiplexer relates the
`metadata to the main video stream by the PTS signatures. In
`analog streams PTS may be created and integrated. In Some
`embodiments there may be multiple and cascaded Systems,
`and Some Sources may be stored Sources. Various methods
`are disclosed for monitoring and compensating time differ
`ences among Sources to ensure time coordination in end
`product. In different embodiments transport of metadata to
`an end user Station is provided by Internet Streaming, VBI
`insertion or by Internet downloading. User equipment is
`enhanced with hardware and Software to coordinate and
`present authored material with the main data Stream.
`21 Claims, 18 Drawing Sheets
`
`12
`
`
`
`Video Head End
`
`-
`
`
`
`
`
`Object Tracking
`Module
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 1 of 18
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 1 0f 18
`
`US 6,357,042 B2
`
`I `81-I
`Fig.I
`
`11
`
`
`
`ObjectTrackingModule
`
`
`
`
`
`
`
`K.VideoHeadEnd
`
`12
`
`ZI
`
`15
`
`S I
`
`TTI-1028, Page 2
`
`

`

`U.S. Patent
`
`Mar. 12, 2002.
`
`Sheet 2 of 18
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 2 0f 18
`
`US 6,357,042 B2
`
`
`
`Fig.2
`
`s
`
`TTI-1028, Page 3
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 3 of 18
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 3 0f 18
`
`US 6,357,042 132
`
`
`
`‘————_—__
`
`hm
`
`TTI-1028, Page 4
`
`
`

`

`U.S. Patent
`
`Mar. 12, 2002.
`
`Sheet 4 of 18
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 4 0f 18
`
`US 6,357,042 B2
`
`33
`
`
`
`
`
`
`
`
`
`
`
`
`
`Fig.4
`
`III”
`|-|--|--
`II”
`:
`
`TTI-1028, Page 5
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 5 of 18
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 5 0f 18
`
`US 6,357,042 B2
`
`
`
`ç (81-I
`Fig.5
`
`._1
`
`IEV)
`O
`D4
`"O5.)H
`U
`.5.)
`3H
`D4V
`CL
`
`AE0
`
`TTI-1028, Page 6
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 6 of 18
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 6 0f 18
`
`US 6,357,042 B2
`
`69
`
`9 I
`
`13
`
`
`
`
`
`me
`
`:1.
`.9:
`2‘55
`:30
`"0°08
`oxm
`2M0
`35
`SW”
`4.853
`On“
`
`S9
`
`lg.6
`
`TTI-1028, Page 7
`
`

`

`U.S. Patent
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 7 0f 18
`
`US 6,357,042 132
`
`Gwv
`
`mm.
`
`«E»
`
`$3985
`
`Eupmxm
`
`Du
`
`9»
`
`mm
`
`NNE
`
`
`
`«mm
`
`
`
`839mmwEhofiz<
`
`Mm
`
`mm
`
`SS
`
`L
`m
`m
`
`Q99
`
`hm
`
`L
`S
`
`93E83>
`
`tam
`
`mw
`
`TTI-1028, Page 8
`
`
`
`

`

`U.S. Patent
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 8 0f 18
`
`US 6,357,042 132
`
`mm
`
`mm.
`
`_—m—mmmm$—
`
`30
`
`03
`
`Momfio
`
`Stowwccofis<
`
`mo
`
`[FF,,
`
`TTI-1028, Page 9
`
`
`
`
`
`
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 9 of 18
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 9 0f 18
`
`US 6,357,042 132
`
`
`
`m.uE
`
` uofiCBEEmD
`
`IIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIII
`
`:
`
`TTI-1028, Page 10
`
`
`

`

`U.S. Patent
`
`Mar. 12, 2002.
`
`Sheet 10 of 18
`
`US 6,357,042 B2
`
`
`
`Reader/Frame Counter
`
`Number/Marker
`Generator
`
`Writer/Data Inserter
`
`Pixel Command
`
`To Brodcast System
`
`TO WAN
`
`Fig. 10
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 11 of 18
`
`US 6,357,042 B2
`
`103
`
`105
`
`107
`
`109
`
`1 11
`
`Monitor and identify
`frames in streams
`as needed.
`
`
`
`
`
`Generate binary
`signatures
`
`
`
`Write binary data
`into VBI/HBI
`
`
`
`
`
`
`
`Select pixels for
`alteration and
`activate
`
`Insert markers
`as needed
`
`Fig. 11
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 12 of 18
`
`US 6,357,042 B2
`
`
`
`©InpOWN
`
`KeIds[GI
`
`LÕI1nduI
`
`qnduILÕI
`
`6 I I
`
`/~ || A,
`JLdIO
`
`====
`
`
`
`
`
`
`
`
`
`
`
`
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 13 0f 18
`
`US 6,357,042 B2
`
`
`
`Capture and load
`Separate data streams
`into pipeline
`
`Search for markers
`in Video broadcast
`data Stream
`
`Search for frame
`numbers in
`annotation Stream
`
`Determine timing
`error factor between
`StreamS
`
`Buffer streams to
`reduce error factor
`
`Repeate steps 145
`-151 for each
`marker found
`
`Fig. 13
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 14 of 18
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 14 0f 18
`
`US 6,357,042 132
`
`5-,.
`
`
`
`fig. 4
`
`
`
`TTI-1028, Page 15
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 15 of 18
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 15 0f 18
`
`US 6,357,042 B2
`
`
`
`i
`
`01:01:52
`
`38:3:38
`
`8:33:33
`
`“(‘9 5:»: the firm mme-n '
`
`00:20:03? .
`
`‘
`
`,
`
`I:
`
`fifig.£§
`ig.
`
`TTL1028,Page16
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 16 of 18
`
`US 6,357,042 B2
`
`US. Patent
`
`Mar. 12, 2002
`
`Sheet 16 0f 18
`
`US 6,357,042 132
`
`
`
`“2,50pr
`
`3a
`
`W78258Z358me
`
`3ME
`
`IIIIIIIIIIIIIIIIIIIIIIIIIIIIIIJ
`
`TTI-1028, Page 17
`
`
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 17 of 18
`
`US 6,357,042 B2
`
`243
`
`245
`
`
`
`User Logs On to Service
`
`Ad Server Notified and
`Accesses User-Profile Data
`
`Ad Server Executes Ad
`Schedule to Integrate Ads
`With Video, Based on
`Profile Data
`
`Fig. 17
`
`

`

`U.S. Patent
`
`Mar. 12, 2002
`
`Sheet 18 of 18
`
`US 6,357,042 B2
`
`/ 90
`
`Smom
`
`m
`
`001
`
`
`
`BmmmMI
`aAl'-
`
`US. Patent
`
`20mn,
`
`r.maflofisdcNQ4watofi=<m02.3-5th
`
`mvm
`
`
`
`6S
`
`«W,
`
`m
`
`m.3.me
`
`?I '81-I
`
`
`
`
`
`NaBRO—Eon.BEBE@22585%::5
`
`
`
`UuEmubmBEBE
`
`Umommom
`
`SN
`
`
`
`Emfimnooztongs;33302
`
`
`
`
`
`th
`LSZ
`
`TTL1028,Page19
`
`
`

`

`US 6,357,042 B2
`
`1
`METHOD AND APPARATUS FOR
`MULTIPLEXING SEPARATELY-AUTHORED
`METADATA FOR INSERTION INTO AVIDEO
`DATA STREAM
`
`CROSS-REFERENCE TO RELATED
`DOCUMENTS
`The present invention is a continuation-in-part (CIP) of
`copending patent application Ser. No. 09/225,250, filed Jan.
`4, 1999 which is a CIP of Ser. No. 09/218,738, filed Dec. 22,
`1998, which is a CIP of Ser. No. 09/168,315, filed Oct. 7,
`1998, which is a CIP of copending application Ser. No.
`09/160,365, filed Sep. 24, 1998 which is a CIP of Ser. No.
`09/154,532, filed Sep. 16, 1998.
`FIELD OF THE INVENTION
`The present invention is in the field of Video broadcasting,
`and pertains more particularly to methods and apparatus for
`multiplexing Separately-authored metadata for coordination
`with a main Video data Stream.
`
`15
`
`2
`Set-top boxes and computer-integrated TVS adapted as
`described above typically have inputs for such as a TV
`antenna (analog), cable TV (analog or digital), more recently
`direct-satellite TV (digital), and may also connect to Video
`cassette recorders and to mass Storage devices Such as hard
`disk drives and CD-ROM drives to provide a capability for
`uploading Video data from Such devices and presenting the
`dynamic result as a display on the TV Screen.
`The present inventors have noted that with the coupling of
`computer technology with TV, many capabilities familiar to
`computer users have been made available to TV users. For
`example, ability to provide text annotation for TV presen
`tations is considerably enhanced. Computer techniques Such
`a Pix-on-Pix are now available, wherein separate TV pre
`Sentations may be made in Separate windows, or overlaid
`windows on the display Screen. Separate windows may also
`Support display from Separate Sources, Such as an analog TV
`program in one window, a computer game in another, and a
`Video conference in a third.
`With the technologies described above becoming more
`available in the market place, it has become desirable to
`further integrate the technologies described So that a user
`Viewing a Video presentation might be enabled to gather
`additional information about a specific image entity or
`entities portrayed in a Video through interactive method. An
`ultimate goal is to provide a means for advertisers to
`promote and Sell products through user interaction in a way
`that minimizes Steps required by Such a user to access
`additional information regarding traditionally advertised
`products Such as through commercials and the like.
`In typical prior art Video authoring Systems, end users
`receive a Single video stream that contains the Video data and
`any added annotated data Such as Subtitling, sponsor logos,
`information blocks, and the like. However, it is desirable to
`build upon the goal Stated in the preceding paragraph above,
`by having Separate Streams, one containing video data, and
`the other containing annotative data, that may arrive at and
`end user's location via different delivery media and be
`displayed Synchronously on a Suitable display Screen.
`An authoring System, known to the inventor, may provide
`image tracking coordinates along with various further
`annotation, and may deliver Separate Streams via Separate
`carriers to an end user. Also known to the inventor is a
`System for providing a means of applying a signature and
`asSociative frame identification to the Separate Streams
`respectively before broadcast So that both Streams may later
`be re-synchronized at the user's end. Such a System is
`likewise described under the cross-referencing Section.
`In current art commercial programming, various compa
`nies may purchase advertising blockS or time slots from a
`content provider. The content provider then edits-in Such
`commercials to the appropriate slots before broadcasting.
`Typically, Such commercial ads may be local to an area of
`broadcast and are limited in profiling to those general
`demographics associated with a range or geography of local
`Viewers. For example, in a broadcast football game, com
`mercials may be geared to appealing to a general profile of
`a Sports fan. For a cable channel carrying exclusively
`Women's programming, advertisements would be geared
`more toward women in general. The profiling or focusing of
`advertisement a company can do is thus quite limited.
`A System known to the inventors and disclosed in this
`patent application under the sub-heading below titled “Per
`sonalized and Interactive Ad System/Network” provides in
`one embodiment an Internet-connected Subscription Server
`running an ad-engine in the form of a Software application
`
`25
`
`35
`
`40
`
`BACKGROUND OF THE INVENTION
`With continuing development of new and better ways of
`delivering television and other Video presentations to end
`users, and parallel development of computerized informa
`tion Systems, Such as the Internet and the associated World
`Wide Web (WWW), there have been concerted efforts to
`integrate various Systems to provide enhanced information
`delivery and entertainment Systems. For example, develop
`erS are introducing integrated Systems combining TVs with
`computer subsystems, so a TV may be used as a WEB
`browser, or a PC may be used for enhanced TV viewing.
`In Some Systems computer elements, Such as a CPU,
`memory, and the like, are built into the familiar chassis of a
`TV set. In such a system, the TV screen becomes the display
`monitor in the computer mode. In Such a System, conven
`tional TV elements and circuitry are incorporated along with
`the computer elements, and capability is provided for a user
`to Switch modes, or to view recorded or broadcast video with
`added computer interaction. One may thus, with a properly
`equipped System, Select to view analog TV programs, digital
`TV programs, conventional cable TV, satellite TV, pay TV
`from various sources, and browse the WWW as well,
`displaying WEB pages and interacting with on-screen fields
`and relational Systems for jumping to related information,
`databases, and other WEB pages. The capabilities are often
`integrated into a Single display, that is, one may view a
`broadcast presentation and also have a window on the
`display for WEB interaction.
`In Some other Systems, computer elements are provided in
`an enclosure separate from the TV, often referred to in the art
`as a Set-top box. Set-top box Systems have an advantage for
`providers in that they may be connected to conventional
`television sets, so end users don’t have to buy a new TV
`55
`along with the computer elements.
`In Such integrated Systems, whether in a single enclosure
`or as Set-top box Systems, user input is typically through a
`hand-held device quite Similar to a familiar remote
`controller, usually having infra-red communication with the
`Set-top box or a receiver in the integrated TV. For computer
`modes, Such as WEB browsing, a cursor is displayed on the
`TV Screen, and cursor manipulation is provided by buttons
`or other familiar pointer apparatus on the remote. Select
`buttons are also provided in the remote to perform the
`familiar function of Such buttons on a pointer device, like a
`mouse or trackball more familiar to computer users.
`
`45
`
`50
`
`60
`
`65
`
`

`

`US 6,357,042 B2
`
`3
`that has ability to Select Video ads according to user profile
`and to Stream Such ads to a user along with a main Video data
`Stream. In Some cases the ads are interactive. In Systems
`wherein the main Video and Such video ads are Sent by a
`common carrier, Such as an Internet connection, the ads are
`inserted in the main video stream in the form of video
`metadata.
`It is desirable that that more than one authoring Station or
`System may be used when creating metadata for delivery to
`an end user, because there are a variety of functions that may
`be implemented through metadata. For example, it is desir
`able that Separate authoring Stations will be used in hyper
`Video authoring, Such as in providing object tracking
`coordinates, creating hot spots (hyperlinks) in a video,
`providing interactive regions for tracked objects, inserting
`URLs, providing review markers by Scene authoring, and
`So on. Scene authoring based on Scene-change-detection
`technology (SCDT) has several purposes, Such as providing
`thumbnails as bookmarks for users to Select and review
`particular portions of Video presentations, and for markers
`for ad insertion or insertion of other information. In addition,
`Separate ad information may be authored by yet additional
`authors and provided as metadata for delivery to an end user.
`While combination of live video and live annotation
`Streams is treated herein concerning hyper-video authoring
`and delivery, it is generally understood that in the live case,
`annotation Streams may be timed to run in Sync alongside or
`over a main video Steam. This process is performed at the
`provider's end. However, the presence of possibly two or
`more Separately-authored annotation data-Sets wherein the
`method of delivery is not necessarily in real time, requires
`a more comprehensive approach.
`What is clearly needed is a method and apparatus for
`merging Separately-authored Sets of metadata Such that the
`metadata is associated appropriately to a correct frame
`location in a main Video. Such a method and apparatus
`would serve to ad flexibility to the authoring process and to
`simplify delivery methods.
`
`15
`
`25
`
`35
`
`40
`
`45
`
`4
`embodiments the PTS-enhanced metadata is stored to be
`downloaded as needed by a user.
`In various embodiments of the invention the authoring
`Stations may include one or more of Scene authoring, hyper
`Video authoring, and ad authoring Stations. At the user end
`the user System is enhanced with Software for displaying the
`main Video data Stream and the authored metadata according
`to the PTS.
`In another aspect of the invention a method for coordi
`nating authored video metadata with a main Video data
`Stream is provided, comprising steps of (a) ensuring the
`main video data stream has a presentation time Stamp (PTS);
`(b) feeding the digital main video data Stream to authoring
`Stations; (c) authoring matadata at the authoring Stations;
`and (d) marking the metadata with presentation time stamps
`(PTS) from the main video data stream.
`This method may further comprise a step for multiplexing
`authored metadata with the main Video data Stream, wherein
`the multiplexer places the metadata in relation to the main
`video data stream according to the PTS. There may also be
`multiple Sources of Video fed to the multiplexer as well as
`multiple metadata Streams for a Video, and a step as well for
`compensating for real-time differences between the multiple
`Sources. In the compensating Step, presentation time Stamps
`(PTS) may be amended according to source time differences.
`In Some cases the PTS-enhanced metadata is streamed
`over the Internet to an end user. In other cases the PTS
`enhanced metadata is inserted into Video blanking intervals
`(VBI) of an analog stream according to the PTS. In still other
`cases the PTS-enhanced metadata is stored to be down
`loaded as needed by a user. The authoring Stations may
`include one or more of Scene authoring, hyper-video
`authoring, and ad authoring Stations, and analog streams
`may be accommodated in Some embodiments by conversion
`to a digital format before authoring and multiplexing, and in
`others by integrating a PTS with the analog stream. Also, at
`the final user's end, there is Software for rendering the main
`Video data Stream and authored metadata according to PTS.
`In yet another aspect of the invention a digital Video
`multiplexing System is provided comprising metadata inputs
`from Video authoring Stations, an input for a main digital
`Video data Stream; and an output to a Video transport
`interface. The multiplexer notes presentation time Stamps
`asSociated with authored metadata, and places the authored
`metadata relative to the main Video data Stream for transport
`to end users. The multiplexing System may have multiple
`Video data Stream inputs, and one or more of the inputs may
`be from a stored source. There may also be multiple video
`data Stream inputs from multiple Sources, and the multi
`plexer monitorS real time clocks of the Sources and uses the
`information to compensate one or both of the multiple
`Streams. In case of real-time differences the multiplexer
`compensates incoming Streams by buffering one or more of
`the Streams. The multiplexer may also compensate incoming
`Streams by amending the presentation time Stamps of one or
`more of the Streams.
`In embodiments of the invention as taught in enabling
`detail below, for the first time it is possible to annotate one
`or more main Video streams, either analog or digital Streams,
`and to enhance the Streams with authored metadata in a
`manner that multiple inputs may be made and fully coordi
`nated to be completely useful when finally delivered to the
`end user, and many interactive functions not previously
`known in the art are provided.
`BRIEF DESCRIPTION OF THE DRAWING
`FIGURES
`FIG. 1 is a block diagram illustrating an authoring System
`incorporating an image tracking module according to an
`embodiment of the present invention.
`
`SUMMARY OF THE INVENTION
`In a preferred embodiment of the present invention an
`authoring System for interactive video, comprising a Video
`feed providing a main Video presentation Stream; two or
`more authoring Stations coupled to the Video feed providing
`authoring functions creating metadata for enhancing the
`main Video stream; and a multiplexer for coordinating
`authored metadata with the main Video Stream. The author
`ing stations may note a presentation time Stamp (PTS) of
`Video frames or any other time Stamp and incorporate it in
`the authored metadata for matching the metadata with the
`main Video presentation Stream.
`In various embodiments there is a multiplexer for com
`bining authored metadata with the main Video data Stream,
`and the multiplexer places the metadata in relation to the
`main video data stream according to the PTS. The multi
`plexer in Some cases receives multiple Video Streams as well
`as the authored metadata, and time clocks are monitored for
`Separate Stream Sources and clocks are adjusted to compen
`Sate for real-time differences in Sources. One or more of the
`Stream Sources may be from a stored Source. In Some
`embodiments PTS values are rewritten in one or more
`Streams to compensate for perceived time differences. Also
`in Some embodiments PTS-enhanced metadata is streamed
`over the Internet to an end user, and in others the PTS
`65
`enhanced metadata is inserted into Video blanking intervals
`(VBI) of an analog stream according to the PTS. In still other
`
`50
`
`55
`
`60
`
`

`

`S
`FIG. 2 is an illustration of a display Screen depicting a
`tracked image entity according to an embodiment of the
`present invention.
`FIG. 3 is an enlarged view of the image entity of FIG. 2
`wherein key pixels are visible.
`FIG. 4 is a exemplary view of a value table containing
`values of Signature pixels according to an embodiment of the
`present invention.
`FIG. 5 is a motion diagram illustrating an image tracking
`process according to an embodiment of the present inven
`tion.
`FIG. 6 is a block diagram illustrating a data-stream
`generator and Set-up function of the tracking module of FIG.
`1 according to an embodiment of the present invention.
`FIG. 7 is a block diagram illustrating a real-time video
`authoring and delivery System according to an embodiment
`of the present invention.
`FIG. 8 is a block diagram illustrating multiple authoring
`Station architecture according to an embodiment of the
`present invention.
`FIG. 9 is a block diagram illustrating an exemplary
`modular architecture of a single authoring Station according
`to an embodiment of the present invention.
`FIG. 10 is a block diagram illustrating a signature appli
`cation apparatus at the authoring end according to an
`embodiment of the present invention.
`FIG. 11 is a process flow chart illustrating logical Steps of
`providing a signature at the authoring end according to an
`embodiment of the present invention.
`FIG. 12 is a block diagram illustrating a data capture and
`Synchronization device at the user's end according to an
`embodiment of the present invention.
`FIG. 13 is a Process flow chart illustrating logical steps
`for capturing and Synchronizing Separate Video Streams for
`user display and interaction according to an embodiment of
`the present invention.
`FIGS. 14 and 15 are illustrations of a graphic user
`interface of a Video editor according to an embodiment of
`the present invention.
`FIG. 16 is an overview of an exemplary ISP-based video
`delivery network Supporting personalized and interactive
`Video advertising according to an embodiment of the present
`invention.
`FIG. 17 is a process flow chart illustrating video-ad server
`function and interaction according to an embodiment of the
`present invention.
`FIG. 18 is a diagram of a multiplexing System according
`to an embodiment of the present invention.
`DESCRIPTION OF THE PREFERRED
`EMBODIMENTS
`According to a preferred embodiment of the present
`invention, a method and apparatus is provided which allows
`a programmer functioning in a Video editing mode to initiate
`tracking of any image entity or entities in a Video Stream,
`after which initiation tracking may be automatic, wherein
`appropriate coordinate-tracking data associated with the
`image entity or entities is provided Synchronized with the
`original data Stream So that Such entities may later be
`rendered identifiable and interaction-capable to an end user.
`By image entity is meant any perSon or thing depicted in a
`Video display, Such as a player in a Sports game, and actor
`in a play, a car in a car race, and So on.
`FIG. 1 is a block diagram illustrating an authoring Station
`11 adapted for accepting a Video data Stream 16 from a Video
`
`15
`
`25
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`US 6,357,042 B2
`
`6
`head end 15 as is known in the art. The original video data
`Source 12 may be any recorded or live Source and is not
`limited as to protocol. Authoring Station 11 includes a
`display 14 and an object tracking module 13 which is
`adapted to automatically track image entities in a Video
`presentation, and to Synchronize the tracking result with the
`Video data Stream.
`The Overall purpose of the authoring Station is addition of
`innovative material to the Video data Stream, Such as text
`overlay, graphic icons and logos for advertisement, Some of
`which may be associated with identity and address data to
`allow a viewer at a computerized end Station to access
`advertisements and other data which may be associated with
`individual entities in the video presentation. Advertisements
`may, for example, be associated with a tracked object. Also
`the text annotations could either be set to track along with an
`object, or appear in a fixed position anywhere on the Screen,
`as they are typical in broadcasts today.
`In FIG. 1 input data stream 15 to tracking module 13 is a
`Stream of Successive bitmapped frames in a normalized
`resolution, required by the tracking module. The authoring
`Station can be based on Virtually any Sort of computer
`platform and operating System, and in a preferred
`embodiment, a PC station running MS Windows is used, in
`which case the input Stream 16, regardless of protocol, is
`converted to bitmapped frames displayed typically at 30
`frames per second with a resolution of 352 by 240 pixels.
`The use and operation of the tracking module is disclosed
`below in enabling detail, and outputs both the original data
`Stream 19 and an annotation Stream, which in this case
`comprises the data of one or more tracks for moving entities
`in the video display. These two streams may be combined or
`Separately provided to downstream processes and
`equipment, and other annotations and alterations may be
`made to the data Stream in the authoring Station, as described
`above. Output 21 from authoring station 11 is meant to be a
`general indication of data output, and not to indicate that
`there is a single data Stream. There may be multiple Streams
`of various protocol.
`An authoring Station Such as Station 11 comprises Soft
`ware that may process various media (analog or digital)
`including Video presentations or Streams, both live and
`recorded. For example, included Video Stream types includ
`ing but not limited to common formats Such as Audio Video
`Interleave (AVI) and Moving Pictures Experts Group
`(MPEG). Video source 12 may embody any video source
`that may be known in the art such as a CD-ROM, Satellite
`TV, cable TV, VCR, Internet Server, and so on. Video source
`12 may provide prerecorded video or live broadcast video.
`Also, future new formats of Video Streams shall be consid
`ered equivalent in the Sense of this invention.
`As described briefly above, the inventor has found that a
`Suitable computer platform for performing the tracking
`process enabled by tracking module 13 is a PC/VDU run
`ning Windows with a central processing unit (CPU) oper
`ating at at least 300 megahertz and having at least 64
`megabytes of random access memory (RAM). Video fre
`quency or a frame speed of 30 frames per second (FPS) is
`utilized in this example as a Suitable and compatible stan
`dard for processing. The display size of the Video picture in
`this instance is 352x240 pixels, which is a reduced resolu
`tion from that normally displayed in a Video presentation on
`most end use Stations, but provides a Suitable format for the
`present tracking process. It will, however, be apparent to the
`skilled artisan that the example parameters presented herein
`are not limiting to the Success of practicing the present
`
`

`

`US 6,357,042 B2
`
`15
`
`35
`
`40
`
`7
`invention. Other Video frame Speeds, Video display sizes,
`computer platform types and Speeds, and Software types
`may be employed in other embodiments of the present
`invention. The only requirement is that the processor be
`powerful enough to enable the tracking process at the
`necessary frame Speed.
`Tracking module 13 is a Software module provided as a
`component of a Windows driver in this particular embodi
`ment. In other embodiments, module 13 may be provided as
`a driver or as a component of a driver for another platform
`such as may be used as an alternative to WindowsTM.
`Module 13 interfaces with other software components
`executing on Station 11 for the purpose of adding the
`functionality of the present invention which is the ability to
`track a moving image entity within a presented Video Stream.
`Tracking module 13 is adapted to determine a pixel
`Signature associated with a visual entity to be tracked in a
`Video display, and then to determine the path of the pixel
`Signature in the display as the path of the image entity to be
`tracked.
`Once the Video stream is in the domain of tracking module
`13, an image entity in a fist bitmap image is Selected to be
`tracked, and further frame-by-frame movement of the image
`entity is automatically tracked. Tracking data associated
`25
`with a tracked image entity is associated with the bitmap
`Video Stream via a Second Synchronous data-stream com
`prising the frame-by-frame coordinates of the tracked image
`entity. This output video stream comprises the original
`Stream plus the Synchronous data Stream that contains the
`tracking data and is illustrated (via directional arrow) as
`leaving module.
`The end result of tracking process of the present invention
`is a Series of coordinates of an assumed center point of a
`tracking object associated with the image entity Selected to
`be tracked. Through additional editing processes a moving
`region associated with the image entity in a display may be
`made to be interactive and identifiable to an end user. User
`interaction with Such an image entity during viewing of a
`Video can be programmed to provide additional network
`Stored information about that entity to Suitable customer
`premises equipment (CPE) adapted to receive and display
`the information. Such further information may be displayed,
`for example, as an overlay on the display of the dynamic
`Video containing the Subject image entity. In this way,
`advertisers, product promoters, or the like may present
`information to end users based on user interaction with an
`asSociated entity in a dynamic Video display.
`FIG. 2 is an illustration of a display Screen depicting an
`image entity 27 to be tracked according to an embodiment
`of the present invention. PC/VDU display 14 shows a
`bitmap image Stream during the tracking process. Screen 14
`is a display Screen as viewed by a programmer or editor
`performing the tracking process.
`To begin, an editor Selects, typically by well-known
`cursor technique, a tracking box 29 of pre-Selected shape
`and size, and places (drag-and-drop, click-on etc.) the
`Selected box over an image entity to be tracked, with the
`center point of the box Substantially centered on the

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket