throbber
Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 1 of 34
`
`
`
`6:20cv693
`
`EXHIBIT A
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 2 of 34
`111111
`1111111111111111111111111111111111111111111111111111111111111
`US009691429B2
`
`(12) United States Patent
`Leiberman et al.
`
`(10) Patent No.:
`(45) Date of Patent:
`
`US 9,691,429 B2
`Jun. 27,2017
`
`(54) SYSTEMS AND METHODS FOR CREATING
`MUSIC VIDEOS SYNCHRONIZED WITH AN
`AUDIO TRACK
`
`(71) Applicant: Mibblio, Inc., Brooklyn, NY (US)
`
`(72)
`
`Inventors: David Leiberman, Brooklyn, NY (US);
`Samuel Rubin, Brooklyn, NY (US)
`
`(73) Assignee: MIBBLIO, INC., Brooklyn, NY (US)
`
`( *) Notice:
`
`Subject to any disclaimer, the term of this
`patent is extended or adjusted under 35
`U.S.c. 154(b) by 43 days.
`
`(21) Appl. No.: 141708,805
`
`(22) Filed:
`
`May 11, 2015
`
`(65)
`
`(51)
`
`(52)
`
`(58)
`
`Prior Publication Data
`
`US 2016/0336039 Al
`
`Nov. 17,2016
`
`(2006.01)
`(2006.01)
`(2006.01)
`(2006.01)
`(2011.01 )
`(2011.01 )
`
`Int. Cl.
`H04N 5/93
`G11B 27/031
`H04N9/82
`G11B 27/34
`H04N 21/43
`H04N 21/8547
`U.S. Cl.
`CPC ............ G11B 27/031 (2013.01); G11B 27/34
`(2013.01); H04N 9/8211 (2013.01); H04N
`21/4307 (2013.01); H04N 21/8547 (2013.01)
`Field of Classification Search
`CPC ... GliB 2220/90; GliB 27/034; GliB 27/34;
`GliB 27/024; GliB 27/036
`USPC .......................................................... 386/285
`See application file for complete search history.
`
`(56)
`
`References Cited
`
`U.S. PATENT DOCUMENTS
`
`5,265,248 A
`7,026,536 B2
`
`1111993 Moulios et al.
`4/2006 Lu et al.
`
`7,027,124 B2
`7,301,092 Bl
`7,500,176 B2
`8,046,688 B2
`8,244,103 Bl *
`
`4/2006 Foote et al.
`1112007 McNally et al.
`3/2009 Thomson et al.
`1012011 Adams et al.
`8/2012 Shore.
`
`912012 Salvucci
`8,271,872 B2
`8,704,069 B2 * 4/2014 Naik
`
`8,896,609 B2
`
`1112014 Xu et al.
`(Continued)
`
`OTHER PUBLICATIONS
`
`G11B 27/34
`386/278
`
`GI0H 1140
`84/612
`
`Musicstory: A personalized Music Video Creator; David A.
`Shamma, Bryan Paredo, Kristian J. Hammond; Proceedings of the
`13th Annual ACEM International Conference on Multimedia;
`ACM,2005.
`
`(Continued)
`
`Primary Examiner - William Tran
`(74) Attorney, Agent, or Firm - Robert W. Morris;
`Eckert Seamans Cherin & Mellott, LLC
`
`(57)
`
`ABSTRACT
`
`Systems and methods for creating music videos synchro(cid:173)
`nized with an audio track are provided. In some embodi(cid:173)
`ments' an audio track may be selected and one or more video
`takes may be captured while the selected audio track plays.
`The video takes may be analyzed while they are captured to
`determine, for example, a video intensity level and/or a
`number of faces recognized within each take. By capturing
`the video takes with the audio track, the video takes may be
`synchronized to the audio tracks so that they are in time with
`one another. Portions or subsets of the video takes may be
`paired or matched with certain sections of the audio track
`based on, for example, the audio characteristics for a par(cid:173)
`ticular section and video characteristics of a particular take.
`
`19 Claims, 17 Drawing Sheets
`
`TAKE: 1 TAKE: 2 TAKE: 1
`TAKE: 3
`TIME: 00:00 TIME:
`TIME:
`TIME: 01:00
`·00:30
`01:30
`00:30
`00:50
`00:50
`01:00
`722
`I32
`
`742
`
`712
`
`/1/700
`
`TAKE: 1 TAKE: 3
`TAKE: 2 TAKE: 3
`TIME:
`TIME: 02:00 TIME:
`TIME: 03:00
`02:40
`03:36
`02:40
`01:30
`02:00
`03:00
`~ m
`Z6Z
`
`Z6Z
`
`760
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 3 of 34
`
`US 9,691,429 B2
`Page 2
`
`(56)
`
`References Cited
`
`OTHER PUBLICATIONS
`
`U.S. PATENT DOCUMENTS
`
`Creating Music Videos Using Automatic Media Analysis; Jonathan
`
`200210035475 Al * 3/2002 Yoda ....................... GI0L 15/24
`704/270
`
`3/2004 Mizushima
`2004/0060070 Al
`2005/0143915 Al * 6/2005 Odagawa.
`
`912005 Brown et al.
`2005/0190199 Al
`2006/0288849 Al * 1212006 Peeters
`
`2008/0037953 Al * 212008 Kawamura
`
`2008/0055469 Al * 3/2008 Miyasaka
`
`6/2009 Cohen et al.
`200910164034 Al
`1112010 Xu et al.
`201010290538 Al
`1212012 Luo et al.
`201210316660 Al
`2013/0330062 Al * 1212013 Meikle.
`
`2014/0160250 Al * 6/2014 Pomerantz.
`
`1012014 Chau et al.
`2014/0317480 Al
`1012014 Lanuners et al.
`2014/0320697 Al
`2015/0050009 Al * 212015 Svendsen.
`
`G08G 110962
`7011443
`
`GI0H 110008
`84/616
`H04N 5/783
`386/343
`GlOH 11368
`348/521
`
`H04N9/87
`386/285
`H04N 5/23229
`348/47
`
`Foote, Matthew Cooper, and Andreas Girgensohn; Proceedsings of
`
`the 10th International Conference on Multimedia; ACM, 2002.
`MuViSync: Realtime Music Video Alignment; R. Macrae, X.
`Anguera, N. Oliver; 2010 IEEE International Conference on Mul(cid:173)
`timedia and Expo (ICME), 534-9, 2010; ISBN-13: 978-1-4244-
`7491-2; DOl: 10.1109/ICME.2010.5583863; Conference: 2010
`IEEE International Conference on Multimedia and Expo (ICME),
`Jul. 19-23, 2010, Suntec City, Singapore; Publisher:
`IEEE,
`Piscataway, NJ, US.
`Cati Dance: Self-Edited, Self-Synchronized Music Video; Tristan
`Jehan, Michael Lew, and Cati Vaucelle; ACM SIGGRAPH 2003
`Sketches & Applications. ACM, 2003.
`Dubsmash by Mobile Motion GmbH; https:llitunes.apple.comiapp/
`dubsmashlid918820076 retrieved on May 11, 2015.
`Video Star by Frontier Design Group; https:llitunes.apple.comius/
`app/video-starlid438596432?mt~8 retrieved on May 11, 2015.
`
`G11B 27/036
`386/280
`
`* cited by examiner
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 4 of 34
`
`u.s. Patent
`
`Juo.27,2017
`
`Sheet 1 of 17
`
`US 9,691,429 B2
`
`PROCESSOR(S)
`
`MEMORY
`
`STORAGE
`
`[
`
`I
`
`I
`
`r~102
`
`t--104
`
`~106
`
`COMMUNICATIONS
`CIRCUITRY
`
`------108
`
`OUTPUT INTERFACE
`I
`I
`
`DISPLAY
`
`SPEAKERS
`
`110 ........... ---l
`112 -
`- ""-
`
`114
`
`-
`
`INPUT INTERFACE
`
`CAMERA(S)
`
`I
`
`MICROPHONE(S)
`
`EXTERNAL DEVICE
`INPUT
`
`116
`
`FIG. 1
`
`r-r--120
`~ r- ~ 122
`
`r------
`
`118
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 5 of 34
`
`u.s. Patent
`
`Juo.27,2017
`
`Sheet 2 of 17
`
`US 9,691,429 B2
`
`202
`
`208---
`
`212
`
`Projects
`
`/1/'200
`
`206
`
`SONG 1
`
`210a
`
`210
`
`[ SONG 7
`
`]lillJ
`
`210b
`
`FIG.2
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 6 of 34
`
`u.s. Patent
`
`Jun. 27, 2017
`
`Sheet 3 of 17
`
`US 9,691,429 B2
`
`/1-/300
`
`Songs
`
`Cancel
`
`302a-----", ~@ B
`-I
`l®-t
`I@
`. [ SONG3
`302d~ ~@ M
`l/
`I@ Q
`-j
`R s
`I@ u
`302f ----..,. -[
`V w
`I@ X
`~I
`y
`Z
`JJ.
`it
`
`@1
`A
`
`C
`D
`E
`
`I
`J
`K
`L
`
`N
`0
`p
`
`T
`
`304
`~
`
`306
`
`i. ~
`
`~ ';
`
`'-Ii'
`
`SONG2
`
`302c-~_
`
`"--'II
`
`SONGS
`
`SONGS
`
`302g--,_
`
`SONG?
`
`FIG.3A
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 7 of 34
`
`u.s. Patent
`
`Jun. 27, 2017
`
`Sheet 4 of 17
`
`US 9,691,429 B2
`
`SELECT SONG
`DURATION
`
`'" , , , , ,
`
`~50
`
`'" , , , , ,
`
`,
`: 3:35
`, ,
`
`364
`
`360
`
`362
`
`FIG. 38
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 8 of 34
`
`u.s. Patent
`
`Juo.27,2017
`
`Sheet 5 of 17
`
`US 9,691,429 B2
`
`404-
`
`402
`
`SONG 2
`
`/V400
`
`406
`
`408
`
`FIG.4
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 9 of 34
`
`u.s. Patent
`
`Juo.27,2017
`
`Sheet 6 of 17
`
`US 9,691,429 B2
`
`Normal
`
`Start
`
`-
`
`/'1./500
`
`502
`
`508
`
`.-------1---506
`
`FIG.5A
`
`550
`~
`
`504
`
`502
`
`518
`
`~
`\
`
`FIG. 58
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 10 of 34
`
`600~
`
`620~
`
`640~
`
`628
`
`648
`
`CAPTURE VIDEO
`
`CAPTURE VIDEO
`
`e •
`
`7Jl
`•
`~
`~
`~
`
`~ = ~
`
`2-
`?
`N
`~-....l
`
`N o ....
`
`-....l
`
`('D
`('D
`
`rFJ =(cid:173)
`.....
`-....l
`o ....
`....
`
`-....l
`
`660
`
`606
`
`624
`
`660
`
`606
`
`614
`
`00:10
`
`602
`
`FIG. 6A
`
`612
`
`602
`
`FIG. 68
`
`626
`
`602
`
`FIG.6C
`
`646
`
`d
`rJl
`,,'..c
`0'1
`'..c
`"""'" ~
`N
`
`'..c = N
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 11 of 34
`
`/(/700
`
`710
`
`7~
`TAKE: 1 TAKE: 3
`TAKE: 2 TAKE: 3
`TAKE: 1 TAKE: 2 TAKE: 1
`TAKE: 3
`TIME: 00:00 TIME:
`TIME:
`TIME: 01 :00 TIME:
`TIME: 02:00 TIME:
`TIME: 03:00
`03:36 I
`02:40
`-00:30
`01 :30
`00:30
`00:50
`01:30
`02:40
`00:50
`01 :00
`02:00
`03:00
`
`712
`
`1722
`
`1732
`
`1742
`
`1752
`
`1772
`
`1782
`
`1792
`
`726
`
`1736
`
`1746
`
`1756
`
`1776 A~
`
`~
`~~~-::=::I I
`I
`•
`" \ \
`TAKE: 3
`TAKE: 1 TAKE: 2
`
`\ I WI
`
`A \~ I 1((
`
`TAKE: 2
`
`I l~l'~I~~l
`TAK~ 1
`
`TAKE: 3
`
`TAKE: 3
`
`TAKE: 3
`
`e •
`
`7Jl
`•
`~
`~
`~
`
`~ = ~
`
`~ = ?
`
`N
`~-....l
`N
`
`0 ....
`
`-....l
`
`('D
`
`QO
`
`rFJ =-('D
`.....
`0 ....
`....
`
`-....l
`
`760
`
`FIG.7A
`
`d
`rJl
`,,'..c
`0'1
`'..c
`"""'" ~
`N
`
`'..c = N
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 12 of 34
`
`u.s. Patent
`
`Juo.27,2017
`
`Sheet 9 of 17
`
`US 9,691,429 B2
`
`a
`a
`0
`C0
`C0
`
`N
`a
`~~
`
`0
`0
`Q) a
`E a
`f- a
`
`IIlIlllIlll
`I:IjttJilru!
`ttU.UII
`IIlIIUII
`IliIIUIi
`III III III
`IliIIHII
`
`Ilillllllll
`III III III
`IliIIHII
`III III III
`lUI III
`IIlIlllIlll
`~~ilru:t
`!['U1Hr:u
`III III III
`Ililllllll
`IIlIllIlll
`Ilillllll
`IIlIlllIlll
`III III III
`Ililllllll
`III III III
`Ililllllll
`IIlIllIlll
`III III III
`IIlIlllIlll
`~~ru:t
`!['U1Hr:u
`III III III
`Ililllllll
`IIlIllIlll
`Ilillllll
`ttU.UII
`I:IjttJilru!
`!['U1Hr:u
`III III III
`Ilillllll
`IIlIllIlll
`III III III
`IIlIlllIlll
`l:Ij~ill1:l't
`!['U1Hr:u
`~~ru:t
`Ilillllll
`IIlIllIlll
`III III III
`IIlIllIlll
`l:Ij~ill1:l't
`
`~tpJElI
`III III III
`Ilillllll
`III III III
`Ilillllll
`IIlIlllIlll
`III III III
`Ilillllll
`IIlIllIlll
`Ilillllll
`IIlIlllIlll
`III III III
`IIlIlllIlll
`~~ru:t
`
`IIlIlllIlll
`III III III
`!['U1Hr:u
`~~ru:t
`Ilillllll
`III III III
`Ilillllll
`IIlIllIlll
`l:Ij~ill1:l't
`
`CO
`f'..
`0
`LL
`
`"'0
`Q)
`fj')
`:.1
`C
`:::>
`I!
`
`"'0
`Q)
`(f)
`=>
`II
`
`Ilillllll D
`
`'!""""
`Q)
`.Y
`
`(\j
`Q)
`.Y
`
`C0
`Q)
`.Y
`
`~ ~ ~ :::J «
`
`0
`""0
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 13 of 34
`
`u.s. Patent
`
`Jun. 27, 2017
`
`Sheet 10 of 17
`
`US 9,691,429 B2
`
`SELECT AN AUDIO TRACK
`
`SELECT A DURATION FOR
`THE AUDIO TRACK
`
`802
`
`804
`
`CAPTURE VIDEO TAKES WHILE
`PLAYING AUDIO TRACK
`
`FINISHED
`CAPTU RI NG
`VIDEO
`TAKES?
`
`)-N_O __ -----!
`
`YES
`
`810
`
`CREATE MUSIC VIDEO
`FEATURING AUDIO TRACKS AND
`AT LEASE A SUBSET OF
`THE VIDEO TAKES
`
`FIG. SA
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 14 of 34
`
`u.s. Patent
`
`Juo.27,2017
`
`Sheet 11 of 17
`
`US 9,691,429 B2
`
`ANALYZE AUDIO TRACK
`
`~852
`
`850
`
`/
`
`856
`;
`ANALYZE VIDEO TAKES
`
`DETERMINE
`AUDIO INTENSITY LEVELS
`OF AUDIO TRACK
`
`~854
`
`DETERM!NE
`VIDEO INTENS!TY LEVELS
`OF VIDEO TAKES
`
`860 ;
`
`PAIR SECTIONS OF AUDIO
`TRACK WTIH PORTIONS OF
`VIDEO TAKES BASED ON
`DETERMINED AUDIO
`INTENS!TY LEVELS AND
`DETERMINED
`VIDEO INTENSITY LEVELS
`
`FIG. 88
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 15 of 34
`
`u.s. Patent
`
`Jun. 27, 2017
`
`Sheet 12 of 17
`
`US 9,691,429 B2
`
`908b
`902b
`908e
`902e
`
`900
`~
`
`908a
`
`906
`
`904
`
`AKE 1
`
`FIG. 9A
`
`920
`~
`
`922a-h
`
`928b------~
`
`928a'-!'--"'8
`
`928d
`
`9289
`
`928e 928h
`FIG. 98
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 16 of 34
`
`u.s. Patent
`
`Juo.27,2017
`
`Sheet 13 of 17
`
`US 9,691,429 B2
`
`940
`/L-/
`CAPTURE VIDEO
`
`948
`
`.----
`
`942
`
`AKE3
`
`FIG.9C
`
`946
`
`944
`
`970b
`
`970e
`970a
`
`970b 970a
`
`,
`
`\ )
`
`944
`
`904
`
`I
`I
`
`,
`
`t:\~'(\~
`944
`904
`924
`\
`\
`944
`904 944
`
`FIG.9D
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 17 of 34
`
`u.s. Patent
`
`Jun. 27, 2017
`
`Sheet 14 of 17
`
`US 9,691,429 B2
`
`980
`
`/
`SELECT AN AUDIO TRACK ~ 982
`t
`CAPTURE A PLURALITY OF ~ 84
`VIDEO TAKES
`
`+
`DETERMINE A NUMBER OF FACES
`WITHIN EACH VIDEO TAKE WHILE
`CAPTURING THE VIDEO TAKES
`
`986
`
`~'
`
`+
`PAIR AT LEAST A SUBSET OF THE
`PLURALITY OF CAPTURED
`VIDEO TAKES TO THE SELECTED ~ 988
`AUDIO TRACK BASED ON THE
`DETERMINED NUMBER OF FACES
`WITHIN THE VIDEO TAKES
`
`+
`CREATE MUSIC VIDEO INCLUDING
`THE SUBSET OF VIDEO TAKES
`SYNCHRONIZED TO THE
`SELECTED AUDIO TRACK
`
`~: 990
`
`FIG.9E
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 18 of 34
`
`u.s. Patent
`
`Juo.27,2017
`
`Sheet 15 of 17
`
`US 9,691,429 B2
`
`i00\).. d1-Z d1 k:i002
`
`1 ooo~1--1 -----"'-,V~--7~"..--~r~-------l
`1002
`FIG.10A
`
`1024
`
`1024
`
`1040~
`
`t1
`
`t2
`
`iII»!.
`
`t1
`
`t2
`iII»!.
`
`-110-1 ...
`
`t1
`
`-110-1
`
`FIG. 100
`
`1062
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 19 of 34
`
`u.s. Patent
`
`Jun. 27, 2017
`
`Sheet 16 of 17
`
`US 9,691,429 B2
`
`1082 ~
`
`1084 ~
`
`~086 ~
`
`RECEIVE AUDIO INPUT VIA
`AN AUDiO INPUT iNTERFACE
`+
`RECORD THE RECEIVED AUDIO ON
`A USER DEVICE
`
`t
`
`DETERMINE THAT THE RECORDED
`AUDIO INCLUDES AT LEAST ONE OF
`A VOCAL PHRASE AND A
`MELODIC PHRASE
`t
`PAIR AT LEAST A PORTION OF AT
`LEAST ONE OF A PLURALITY OF
`CAPTURED VIDEO TAKES TO THE ~ 1088
`RECORDED AUDIO BASED ON THE
`DETERMINED VOCALAND/OR
`MELODIC PHRASE WHILE THE
`VIDEO TAKES ARE CAPTURED
`t
`GENERATE A MUSIC VIDEO INCLUDING
`THE RECORDED AUDIO TRACK AND
`AT LEAST THE PORTION OF THE
`AT LEAST ONE OF THE PLURALITY
`OF CAPTURED VIDEO TAKES
`SYNCRONIZED TO THE
`RECORDED AUDIO TRACK
`FIG.10E
`
`1090 ~
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 20 of 34
`
`u.s. Patent
`
`Jun. 27, 2017
`
`Sheet 17 of 17
`
`US 9,691,429 B2
`
`/l/1100
`
`RECORD AN
`AUDiO TRACK
`
`1102
`
`......... _------
`11 04 ""'+-__
`
`EMPO: 90Bpm
`
`ey: A minor
`
`1106
`
`1108
`
`00 :00 ..... ,"""', ........ ,"""', ......... ,"" ,00 :56
`
`FIG. 11
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 21 of 34
`
`US 9,691,429 B2
`
`1
`SYSTEMS AND METHODS FOR CREATING
`MUSIC VIDEOS SYNCHRONIZED WITH AN
`AUDIO TRACK
`
`FIELD OF THE INVENTION
`
`Various embodiments described herein generally relate to
`systems and methods for creating music videos. In particu(cid:173)
`lar, music videos may be created including portions of one
`or more video takes that are automatically synchronized to
`an audio track.
`
`BACKGROUND OF THE INVENTION
`
`Music videos, whether they involve famous musicians or
`independent artists, are fun and creative mediums for shar(cid:173)
`ing one's music and musical style with the world. While
`most music videos, at least historically, were created in
`support of established musicians for marketing purposes, the
`enhanced capabilities of mobile devices allows almost any
`individual to record and edit music, as well as capture video,
`all using one device. Individuals, whether an experienced
`musician or a novice, young or old, now have the ability to
`create their own music videos using such devices.
`Although music videos often include an individual's own
`music, it is also possible for music videos to be created based
`on an individual's favorite or a popular song. While there are
`presently some applications of this concept, most of these
`applications have several inherent drawbacks.
`In one instance, music videos have been created where an
`audio track plays in the background while a single video is
`captured or recorded. This, however, creates an extremely
`poor quality music video as there is no visual transition
`between various parts of the audio track. For example, a 35
`single video take may be used for the audio track's verse and
`chorus. This leads to extremely unprofessional looking
`music videos that, while potentially entertaining, are aes(cid:173)
`thetically inferior to professional quality music videos,
`which may use multiple video takes captured at one or more 40
`locations.
`Another instance of music videos being created focuses
`on a linear application of a video and audio track. For
`example, a single video take may be captured and multiple
`end points may be applied to that video. However, this
`application is extremely limited in that it does not allow a
`user to use multiple video, and, as such, does not allow the
`user to apply multiple end points to the multiple videos.
`Furthermore, in order to have different locations in the music
`video, a user would need to visit multiple locations in
`chronological order and capture video at each location. The
`editing of the video takes captured at each location would
`then only present the locations in the order that they were
`visited. This, as mentioned previously, creates a music video
`that is unprofessional in appearance, as professional music
`videos may have varying locations throughout the music
`video.
`In another, somewhat similar, instance, music videos have
`been created where a user is required to capture video of
`another individual within a silhouette. A generic silhouette is
`overlaid on the user's display and, while the user records
`video, the individual being recorded must stay generally
`within the bounds of the silhouette. While this may expand
`the manipulative aspects of the video, the user is unfortu(cid:173)
`nately constrained to be within one particular area, and does
`not have a free range of motion to fully capture any desired
`action. Furthermore, as the user is bound by the silhouette,
`
`2
`the ability to transition to different video takes for different
`portions of an audio track is limited, if at all possible.
`In yet another instance, music videos have been created
`that include only a small portion or snippet of the audio track
`with an individual capturing a single video for that portion
`of the audio track. For example, a user may select a song to
`form a music video for, and create a music video based on
`the song's verse or chorus. This may lead to short, dull, and
`unprofessional music videos as the music video may become
`10 nothing more than a video clip for a small tidbit of a song.
`Thus, in light of some of the aforementioned problems, it
`would be beneficial for there to be systems, methods, and
`non-transitory computer readable mediums that allow a user
`to create a professional style music video using portions of
`15 multiple videos takes taken at different times and at different
`locations that are automatically synchronized to a selected
`audio track. Furthermore, it would be beneficial for there to
`be systems, methods, and non-transitory computer readable
`medium that allow video takes to be paired with an audio
`20 track such that sections of the audio track having various
`intensity levels or dynamics are matched with suitable
`portions of the video takes.
`
`SUMMARY OF THE INVENTION
`
`25
`
`This generally relates to systems, methods, and non(cid:173)
`transitory computer readable mediums for creating music
`videos that are synchronized to an audio track.
`In one exemplary embodiment, a method for creating a
`30 music video where an audio track is synchronized with a
`plurality of video takes is described. An audio track, such as
`a song, may be selected. For example, a user may select a
`song stored on their user device, in a music library on an
`external device, or on a music server. A plurality of video
`takes may be captured also using the user device. While the
`plurality of video takes are being captured, they may also be
`synchronized with the selected audio track. The synchroni(cid:173)
`zation allows for the captured video takes to be aesthetically
`and/or musically synchronized with an appropriate section
`or sections of the audio track. A music video may then be
`created including the audio track and at least a subset of the
`plurality of video takes that are already synchronized to the
`selected audio tracks. For example, portions of one or more
`captured video takes may be matched to certain sections of
`45 the audio track based on the audio track's audio intensity
`levels and/or a particular video take's video intensity.
`In another exemplary embodiment, a user device includ(cid:173)
`ing at least one audio input component, at least image
`capturing component, memory, and at least one processor is
`50 described. The memory may store an audio track recorded
`using the at least one audio input component and a plurality
`of video takes captured by the at least one image capturing
`component. For example, a user may record an audio track
`using their user device's microphone as well as record video
`55 takes using one or more cameras resident on the user device.
`The at least one processor of the user device may then
`determine a vocal and/or melodic phrase within the recorded
`audio track, and synchronize at least a portion of one or more
`captured video takes to the recorded audio track based on the
`60 determined vocal and/or melodic phrase while the at least
`one of the plurality of video takes is captured.
`In yet another exemplary embodiment, another method
`for creating a music video is described. An audio track may
`be selected and a plurality of video takes may be captured.
`65 A number of faces within each video take of the captured
`video takes may be determined while the plurality of video
`takes are being captured. Also while the plurality of video
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 22 of 34
`
`US 9,691,429 B2
`
`3
`takes are being captured, at least a subset of the plurality of
`captured video takes may be synchronized to the selected
`audio track based on the number of faces determined to be
`within each video take. A music video may then be created
`including the selected audio track and at least the subset of
`the plurality of captured video takes synchronized to the
`selected audio track.
`In still yet another exemplary embodiment, a user device
`including memory for storing a plurality of video takes, at
`least one image capturing component, and at least one 10
`processor is described. The at least one processor is operable
`to receive a selection of an audio track from the plurality of
`audio tracks stored in memory. The audio track may then
`play and, while playing, at least one video take may be 15
`captured using the at least one image capturing component.
`The at least one captured video take may be synchronized to
`the selected audio track while the selected audio track plays.
`A music video may then be generated that includes the
`selected audio track and at least a subset of the at least one 20
`video take that is already synchronized to the selected audio
`track.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`The above and other features of the present invention, its
`nature and various advantages will be more apparent upon
`consideration of the following detailed description, taken in
`conjunction with the accompanying drawings in which:
`FIG. 1 is an illustrative block diagram of a user device in
`accordance with various embodiments;
`FIG. 2 is an illustrative diagram of a user interface
`displayed on a user device in accordance with various
`embodiments;
`FIGS. 3A and 3B are illustrative diagrams of user inter(cid:173)
`faces displayed on a user device for selecting an audio track
`in accordance with various embodiments;
`FIG. 4 is an illustrative diagram of a user interface
`displayed on a user device presenting a selected audio track
`for a music video to be created in accordance with various
`embodiments;
`FIGS. 5A and 5B are illustrative diagrams of various user
`interfaces displayed on a user device for capturing video
`takes for a music video to be created in accordance with
`various embodiments;
`FIGS. 6A-C are illustrative diagrams of various user
`interfaces displaying video takes being captured by a user
`device for a music video in accordance with various embodi(cid:173)
`ments;
`FIGS. 7A and 7B are illustrative diagrams of a created
`music video including a plurality of captured video takes
`synchronized to selected audio track in accordance with
`various embodiments;
`FIGS. 8A and 8B is an illustrative flowchart of a process
`for creating music videos in accordance with various
`embodiments;
`FIGS. 9A-C are illustrative diagrams of user interfaces
`including various video takes having a number of faces or
`images determined to be within the video takes in accor(cid:173)
`dance with various embodiments;
`FIG. 9D is an illustrative diagram of various sections of
`a selected audio track's waveform synchronized with por(cid:173)
`tions of video takes based on the number of faces determined
`to be within each video take in accordance with various
`embodiments;
`FIG. 9E is an illustrative flowchart of a process for
`synchronizing video takes to an audio track based on a
`
`4
`number of faces determined to be within the video takes in
`accordance with various embodiments;
`FIGS. 10A-D are illustrative diagrams of various vocal
`and melodic indicators and phrases for use as transition
`points between video takes for a music video synchronized
`to an audio track in accordance with various embodiments;
`FIG. 10E is an illustrative flowchart of a process for
`creating a music video including various video takes syn(cid:173)
`chronized to a recorded audio track based on a determined
`vocal and/or melodic phrase within the audio track in
`accordance with various embodiments; and
`FIG. 11 is an illustrative diagram of a user interface for
`recording an audio track to be used for creating a music
`video in accordance with various embodiments.
`
`DETAILED DESCRIPTION OF THE
`INVENTION
`
`The present invention may take form in various compo(cid:173)
`nents and arrangements of components, and in various
`techniques, methods, or procedures and arrangements of
`steps. The referenced drawings are only for the purpose of
`illustrated embodiments, and are not to be construed as
`limiting the present invention. Various inventive features are
`25 described below that can each be used independently of one
`another or in combination with other features. Furthermore,
`in at least some embodiments, like referenced numerals refer
`to like parts throughout.
`FIG. 1 is an illustrative block diagram of a user device in
`30 accordance with various embodiments. User device 100, in
`some embodiments, may correspond to any electronic
`device or system. Various types of user devices include, but
`are not limited to, portable media players, cellular tele(cid:173)
`phones or smart phones, pocket-sized personal computers,
`35 personal digital assistants ("PDAs"), desktop computers,
`laptop computers, tablet computers, and/or electronic acces(cid:173)
`sory devices such as smart watches and bracelets. User
`device 100 may communicate with one or more additional
`user devices, networks, and/or servers. For example, user
`40 device 100 may send text messages to other user devices
`across a network, or user device 1 00 may access one or more
`websites located on a server.
`User device 100, in some embodiments, may include one
`or more processors 102, memory 104, storage 106, commu-
`45 nications circuitry 108, an input interface 110, and an output
`interface 118. In some embodiments, input interface 110
`may include one or more cameras 110 or other image
`capturing components, one or more microphones 112 or
`other audio capturing components, and one or more external
`50 device inputs 116. Further, in some embodiments, output
`interface 118 may include display 120 and one or more
`speakers 122 or other audio output components. Persons of
`ordinary skill in the art will recognize that user device 100
`may include any number of components, and one or more
`55 additional components or modules may be added or omitted
`without deviating from the scope of the present disclosure.
`Additionally, one or more components may be combined or
`separated, and multiple instances of various components are
`also possible, however only one of each component is shown
`60 within user device 100 for simplicity.
`Processor(s) 102 may include any suitable processing
`circuitry, such as one or more processors, capable of con(cid:173)
`trolling the operations and functionality of user device 100.
`In some embodiments, processor(s) 102 may facilitate com-
`65 munications between various components within user
`device 100. For example, processor(s) 102 may cause output
`interface 118 to perform an associated output in response to
`
`

`

`Case 6:20-cv-00693-ADA Document 1-2 Filed 07/29/20 Page 23 of 34
`
`US 9,691,429 B2
`
`5
`one or more inputs being detected by input interface 110.
`Processor(s) 102 may run an operating system for user
`device 100, applications resident on user device 100, firm(cid:173)
`ware applications, media application, and/or any other type
`of application, or any combination thereof function on, or in
`conjunction with, user device 100.
`Memory 104 may include any suitable form of memory,
`such as cache memory, semi-permanent memory (e.g.,
`RAM), or any other memory type, or any combination of. In
`some embodiments, memory 104 may be used in place of
`and/or in addition to an external memory or storage unit or
`device for storing data on user device 100.
`Storage 106 may include one or more storage mediums.
`Various types of storage mediums include, but are not
`limited to, hard drives, solid state drives, flash memory, 15
`permanent memory (e.g., ROM), or any other storage type,
`or any combination thereof. Any form of data or content may
`be stored within storage 106, such as photographs, music
`files, videos, contact information, applications, documents,
`or any other file type, or any combination thereof.
`In some embodiments, memory 104 and storage 106 may
`be combined into a single component. For example, a single
`memory component may include memory and storage func(cid:173)
`tions. In other embodiments, multiple instances of memory
`104 and/or storage 106 may be present, however it is also 25
`possible for memory 104 and/or storage 106 to be external
`to user device 100. For example, one or more files may be
`stored remotely on an external hard drive or on a cloud
`storage provider. However, persons of ordinary skill in the
`art will recognize that the aforementioned scenarios are 30
`merely examples.
`Communications circuitry 108 may include any circuitry
`capable of connecting user device 100 to one or more
`additional devices (e.g., laptop computers, smartphones,
`etc.), one or more networks (e.g., local area networks 35
`("LAN"), wide area networks ("WAN"), point-to-point net(cid:173)
`works, etc.), and/or one or more servers (e.g., file manage(cid:173)
`ment systems, music directories, etc.). Communications
`circuitry may support any suitable communications protocol
`including, but not limited to, Wi-Fi (e.g., 802.11 protocol), 40
`Bluetooth®, radio frequency systems (e.g., 900 MHz, 1.4
`GHz, and 5.6 GHz communications systems), infrared,
`GSM, GSM plus EDGE, CDMA, quadband, LTE, VOIP, or
`any other communications protocol, or any combination
`thereof.
`Input interface 110 may include any suitable mechanism
`and/or component for receiving inputs from a user operating
`user device 100. For example, input interface 110, in one
`embodiment, includes one or more cameras 112. Cameras
`112 may correspond to any suitable image capturing com(cid:173)
`ponent capable of capturing images and/or video. For
`example, camera 112 may capture photographs, sequences
`of photographs, rapid shots, videos, or any other type of
`image, or any combination thereof. In some embodiments,
`cameras 112 may be capable of capturing high-definition
`("HD"), 3-D, and/or panoramic images and/or videos. In
`some embodiments, cameras 112 may include one or more
`filters or settings for images and/or video that may be
`captured by cameras 112 (e.g., black and white, monochro(cid:173)
`matic, fades, slow-motion, etc.). In some embodiments, user
`device 100 may include multiple instances of camera 112.
`For example, user device 100 may include a front-facing
`camera and a rear-facing camera. In some embodiments, one
`or more additional image capturing components, such as a
`zoom or add on filter, may be used in connection with, or 65
`instead of, camera 112 to aid in capturing images and/or
`videos.
`
`6
`Microphone(s) 114 may be any component capable of
`detecting and/or receiving audio signals. For example,
`microphone(s) 114 may include one or more sensors for
`generating electrical signals and circuitry capable of pro(cid:173)
`cessing the generated electrical signals. In some embodi(cid:173)
`ments, user device 100 may include multiple instances of
`microphone 114, such as a first microphone and a second
`microphone. In some embodiments, user device 100 may
`include multiple microphones capable of detecting various
`10 frequency levels (e.g., high/low-frequency microphones).
`Furthermore, in some embodiments, one or more external
`microphones may be connected to user device 100 and may
`be used in conjunction with, or instead of, microphone(s)
`114.
`External device input 116 may correspond to any input
`interface or set of input interfaces capable of receiving
`inputs from an external device. For example, one or more
`external microphones, as described above, may be coupled
`to user device 100 through external device input 116. As
`20 another example, a user may couple an electric guitar,
`drums, and/or keyboard to user device 100 via external
`device input 116. However, it is also possible for a user to
`couple one or more external devices, suc

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket