`(10) Patent N0.:
`US 8,949,245 B2
`Mok et al.
`Feb. 3, 2015(45) Date of Patent:
`
`
`USOOS949245B2
`
`5/2012 Li
`2012/0130781 A1
`7/2012 IIart et al.
`2012/0185456 A1
`8/2012 Arriola et al.
`2012/0216117 Al
`O'I'HHR PUB] ,ICA'I'IONS
`
`(54)
`
`(71)
`
`(72)
`
`(73)
`
`ENHANCED VIEWS OF TAGGED ITEMS IN A
`GALLERY
`
`Applicant: Microsoft Corporation, Redmond, WA
`(US)
`
`Inventors: Kin IIong Mok, Richmond, TX (US);
`Avijit Sinha, Newcastle, WA (US)
`
`Assignee: Microsoft Corporation, Redmond, WA
`(US)
`
`(’l)
`
`Notice:
`
`Subject to any disclaimer, the term of this
`patent is extended or adjusted under 35
`U.S.C. 154(b) by 105 days.
`
`13/661,847
`
`Oct. 26, 2012
`
`Prior Publication Data
`
`(21)
`
`(22)
`
`(65)
`
`(51)
`
`(52)
`
`(58)
`
`(56)
`
`Primary Eranrziner 7 Etienne Leroux
`(74) Attorney, Agent, or Firm 7 Steve Crocker; Leonard
`Smith; Micky Minhas
`
`ABSTRACT
`(57)
`Disclosed herein are systems, methods, and software for
`facilitating gallery environments and views. In at least one
`implementation an initial View is presented of tagged items
`arranged in tag groups. The tag groups correspond to tags and
`the tagged items are arranged in the tag groups based on with
`which ofthe tags each of the tagged items is associated. One
`of the groups may be identified for enhanced viewing.
`Accordingly, tagged items associated with the identified
`group, by way oftheir association with a tag corresponding to
`the group, are identified. In some implementations they may
`be referred to as enhanced ta gged items. An enhanced View
`may then be presented of at least the enhanced taggcd items,
`
`20 Claims, 5 Drawng Sheets
`
`Rainer Simon. et al.; "Tag Clouds and Old Maps: Annotations as
`Linked Spatiotemporal Data in the Cultural Heritage Domain,”
`Workshop on Linked Spatiotemporal Data 2010, held in conjunction
`with the 6th International Conference on Geographic Information
`Science (GIScience 2010); Sep. 14, 2010; pp. [-12; http://ceur-ws.
`org/V0169 1/paper2.pdf.
`Stefanie Lindstaedt, et al.; “Recommending Tags for Pictures Based
`on Text, Visual Content and User Contex ,” In Proceedings of the
`2008, Third International Conference on Internet and Web Applica-
`tions and Services; Jun. 8, 2008; pp. 1-6; http://know-center.tugraz.
`anp-contenUuploads/2010/ 10/200 8,slind,ICIVv’,Sasu.pdl'.
`James Lockman; “Nested Overlay Conundrum and Solution for
`DPS;" Feb. 8, 2012; pp. 1-6; Adobe Systems Inc; http://blogsadobe.
`com/j lockrnam’ZO 12 /02/08/nested—overlay-solution-for-dps/'.
`(Continued)
`
`ENHANCED VIEWfl
`
`May 1, 2014
`
`(2006.01)
`
`US 2014/0122498 A1
`Int. Cl.
`G06F 17/30
`US. Cl.
`USPC ........................................................ .. 707/741
`Field of Classification Search
`USPC .......... .. 707/2, 3. 741; 715/220; 709/2047206
`See application fi e for complete search history.
`References Cited
`
`
`
`U.S. PATENT DOCUMENTS
`
`2009/0299725 A1
`2010/0058163 A1*
`2011/0029873 A1
`2011/0137548 A1
`2011/0246942 A1
`
`12/2009 Grigsby et al.
`3/2010 Garcia-Molina et al.
`2/2011 Eseanu et al.
`6/2011 Shen et al.
`10/2011 Misawa
`
`715/220
`
`
`
`INITIALVIEWfl
`
`MICROSOFT CORP. EX. 1013
`Page 1 of 14
`
`
`
`US 8,949,245 B2
`Page 2
`
`(56)
`
`References Cited
`OTHER PUBLICATIONS
`
`Daniele Panza. et a1; “Semantic Zoomable Interface for Multimedia
`Content;” 12th International Workshop on Image Analysis for Mul-
`timedia Interactive Services. Delft, Netherlands, Apr. 13. 2011; pp.
`1-4, http://wwwcs.unibo.it/~marfia/pubblicazi(mi/€023 ,pdf.
`Stephanie Hertrich; “A Tag Cloud control for Windows Phone 7;”
`Tagged Content List; Mar, 28. 2011; pp, 1-7; http://blogsrnsdncom/
`b/stephe/archive/201 1/03/2 8/a-tag-cloud-control-for-Windows-
`phone-7 aspx.
`
`Kirupa; “Deep Zoom Composer Updated: Seadragon Ajax Support,
`New API, and More;” Expression Blend and Design; Nov. 24. 2008;
`pp.
`1-6; http://blogs.msdn.corIr/b/expressiort/archive/2008/l1/24/
`deep-Zoorn-cornposer-updated-seag ax-support-new-api-and-rnore.
`aspx.
`International Search Report & Written Opinion for PCT Patent
`Application No. PCT/US2013/O66470, Mailed Date: Feb. 21. 2014,
`Filed Date: Oct. 24. 2013, 13 Pages.
`
`* cited by examiner
`
`MICROSOFT CORP. EX. 1013
`Page 2 of 14
`
`
`
`SNw§E_om_oo<Eozm_>._<Ezemmmmn_
`
`\N_‘
`
`
`
`w_>_m_._._Dm®®<._.
`
`N$50.“—
`
`.0.mm:0.“—
`
`
`
`|—>>m__>Dmoz<Izm
`
`fiE/MSIZEZ
`
`Fwy—30.n—
`
`
`
`mwlw_>_m_._._
`
`asmE
`
`V _
`
`
`
`
`
`QDOmODmcbmjmwZODmm<m>>m__>
`
`
`
`Dm__n=._.2m_n__n_O>>m__>Dm.OZ<IZm_Hzmwmma
`
`
`
`
`
`
`
`OZ_>>m__>DmOZ<IZm
`
`
`
`
`
`
`
`DmOZ<Izmm0“.w_>_m_._._DmOOdP>n=._.Zm_D_
`
`m0;Om.._.0m_._m_wQDOIOO<._.>n__._.Zm_D_
`
`US. Patent
`
`b.eF
`
`5l021,
`
`Sheet 1 of 5
`
`US 8,949,245 B2
`
`manomo032_
`
`
`
`
`
`
`
`MICROSOFT CORP. EX. 1013
`Page 3 of 14
`
`
`
`Feb. 3, 2015
`
`Sheet 2 of 5
`
`US 8,949,245 B2
`
`@EflIIIEgg
`
`fififiEEE
`
`m0<._.
`
`35§
`
`US. Patent
`
`
`
`Q
`
`.vE30."—
`
`MICROSOFT CORP. EX. 1013
`Page 4 of 14
`
`
`
`b.eF
`
`5l021,
`
`Sheet 3 of 5
`
`US 8,949,245 B2
`
`US. Patent
`\W‘M4_g._.rmg<o<h8w@<o<hQ<o<h
`
`V.._.7i.
`
`mE30."—
`
`MICROSOFT CORP. EX. 1013
`Page 5 of 14
`
`
`
`Feb. 3, 2015
`
`Sheet 4 of 5
`
`US 8,949,245 B2
`
`US. Patent
`
`Emma
`
`oE30."—
`
`<06....
`
`\\\\\\\\\\&mofim
`\\\\\\\\\\\\
`
`EEE5m
`
`MICROSOFT CORP. EX. 1013
`Page 6 of 14
`
`
`
`Feb.3,2015
`
`5«I0.3teBhS
`
`<0(...
`
`US 8,949,245 B2
`
`NMMHEHm
`
`a“0.004....me
`
`US. Patent
`
`MICROSOFT CORP. EX. 1013
`Page 7 of 14
`
`
`
`US 8,949,245 B2
`
`2
`disclosed herein. On the contrary, the intent is to cover all
`altematives, modifications, and equivalents.
`FIG. 1 illustrates a gallery environment in an implementa-
`tion.
`FIG. 2 illustrates a gallery process in an implementation.
`FIG. 3 illustrates a computing system in an implementa—
`tion.
`FIG. 4 illustra es a gallery scenario in an implementation.
`FIG. 5 illustra es a gallery scenario in an implementation,
`FIG. 6 illustrates a gallery scenario in an implementation.
`FIG. 7 illustra es a gallery scenario in an implementation.
`
` TECHNICAL DISCLOSUR:
`
`1
`ENHANCED VIEWS OF TAGGED ITEMS IN A
`GALLERY
`
`TECHNICAL FIELD
`
`Aspects o l‘the disclosure are related to computer hardware
`and software technology, and in particular, to galleries for
`viewing content items.
`
`
`
`TECHNICAL BACKGROUND
`
`Gallery applications allow users to manage content, such
`as photos, videos, documents and other items. A variety of
`computing enviromnents support gallery applications, such
`as mobile, desktop, laptop, and tablet computing environ-
`ments. In fact, gallery applications or similar Functionality
`can also be found on websites. gaming systems. or provided
`by any other suitable environment.
`Items Viewed within a gallery can be browsed, shared, and
`edited, for example. To aid with these tasks, many gallery
`applications arrange items in a grid chronologically accord—
`ing to when the items were created. Navigating a gallery can
`involve swiping from one sheet ofitems in the grid to the next.
`In addition, it is possible to arrange items in a gallery other
`than chronologically. For example. some gallery applications
`provide menu options that allow for items to be grouped
`according to time and date,
`location, or people. Custom
`groups can also be created and items tagged or allocated to the
`groups on a custom basis. For instance, photos of a particular
`person may be tagged and viewed together.
`In some scenarios, a gallery application presents a main
`gallery view in which items are arranged on a grid chrono-
`logically. Menu options may be available that allow the items
`to be viewed in a group fashion. By selecting the appropriate
`menu option, a user can select how to group the items, for
`example by time, location, or people. Upon selecting the
`desired group, the items associated with that group are pre-
`sented. To view a different group, the menu is again navigated
`and the different group selected for viewing.
`
`OVERVIEW
`
`Provided herein are systems, methods, and software for
`facilitating gallery environments and views. In at least one
`implementation an initial view is presented of tagged items
`arranged in tag groups. The tag groups correspond to tags and
`the tagged items are arranged in the tag groups based on with
`which of the tags each of the tagged items is associated. One
`of the groups may be identified for enhanced viewing.
`Accordingly,
`tagged items associated with the identified
`group, by way o l‘their association with a tag corresponding to
`the group, are identified. An enhanced view may then be
`presented of at least the enhanced tagged items.
`This Overview is provided to introduce a selection of con-
`cepts in a simplified form that are further described below in
`the Technical Disclosure. It should be understood that this
`Overview is not intended to identify key features or essential
`features ofthe claimed subject matter, nor is it intended to be
`used to limit the scope of the claimed subject matter.
`
`»
`
`
`
`Implementations described herein provide for improved
`gallery environments and views. In at least one implementa—
`tion, an initial view may be presented of tagged items
`arranged in several tag groups. Each tag group may include
`multiple tagged items associated with that group and may
`include some information indicative of its associated tag. For
`instance, a tag group may involve the presentation ofmultiple
`items in the group along with a visual or graphical represen-
`tation of its associated tag, such as a title bar. The items may
`be presented within each view in a variety of formats, for
`example as thlunbnail graphics, reduced size images, bitmap
`images, icons, text representations, or any other suitable for-
`mat.
`In operation, a user may desire to View an enhanced view of
`one of the groups in a gallery, such as a photo gallery, docu-
`ment gallery, or any other kind ofgallery. Accordingly, tagged
`items associated with the group of interest, sometimes
`referred to herein as enhanced tagged items, are identified,
`These enhanced tagged items can be presented by way of an
`enhanced View, thereby providing the user with an improved
`view of at least some of the tagged items relative to that
`provided by way of the initial View. In some scenarios, the
`enhanced view may be an entirely new sheet or grid of items
`that replace the initial view. In other scenarios, the enhanced
`view may be presented simultaneously with at least a portion
`of the initial View. In a situation where groups in an initial
`view are presented in rows, items associated with each group
`may occupy the row corresponding to the group. The
`enhanced view may involve scrolling horizontally through
`one ofthe rows while the remaining rows remain stationary in
`order to reveal or otherwise render visible additional items in
`the group.
`Tagging items may occur in a variety ol'ways and can be
`accomplished automatically, manually, or otherwise. For
`example, a collection of photos may be tagged based on a
`location, in which case the items may be arranged in groups
`according to their respective locations. Other examples of
`tags include the identity of people in photos, time and date
`associated with photo or productivity documents, topics asso-
`ciated with productivity documents, author identifies, time
`and date 0 [document creation, or any other identi lying infor—
`mation that may be useful to apply as a tag.
`A gallery may be arranged in an initial view based one or
`more of the tags. For instance. a gallery may be arranged in
`location groups, but each location group may include one or
`more sub-groups pertaining again to location but possibly to
`some other category or other identifying information. In some
`cases the initial view may be considered a main view of items
`in a gallery, such as the main View presented upon launching
`a gallery application, although it may be possible to navigate
`to the main view at other times during operation of a gallery
`application.
`
`
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`
`
`Many aspects of the disclosure can be better understood
`with reference to the following drawings. While several
`implementations are described in connection with these
`drawings, the disclosure is not limited to the implementations
`
`MICROSOFT CORP. EX. 1013
`Page 8 of 14
`
`
`
`impractical or even impossible to display the dozens or hun-
`
`3
`Upon selecting one of the groups, such as by touching a
`Visual or graphical representation ofa tag that defines a group,
`swiping an area associated with a group, or providing some
`other suitable user input, items are identified for an enhanced
`view. The graphical representation of a tag may be presented
`in a variety of ways, such as above, below, next to, or over-
`layed upon a group of items with which the tag is associated.
`Referring to the example above involving photos tagged by a
`location, a group of photos associated with a particular city
`may be presented in an initial view, along with other groups of
`photos associated with other cities. By selecting the group
`associated with the particular city, an enhanced View of the
`group may be launched. The enhanced view may include
`another item or items in addition to those presented in the
`initial view. For instance, if the group as initially presented
`included a view of three photos, the enhanced view may
`include a view of four or more photos associated with that
`particular city.
`Some aspects may be apparent from the implementations
`disclosed herein. By arranging items in tagged groups and
`allowing for navigation via the tagged groups, it may be easier
`to find items among a large collection of items. For example,
`a user can zoom into a particular group of sub-group to locate
`a file or photo rather than having to scroll through an entire
`collection of files or photos. In another aspect, additional
`context may be provided with respect to any given tag group
`by presenting at least a portion of the items in the group in an
`initial view, possibly in addition to a tag. These and other
`aspects may improve the ease of use of a gallery, as well has
`possibly improve the operating of an underlying device.
`Referring now to the drawings, FIG. 1 illustrates a gallery
`environment that may be rendered when implementing a gal—
`lery process, as illustrated in FIG. 2. FIG. 3 illustrates a
`computing system suitable for implementing the gallery pro-
`cess and rendering the gallery environment. FIGS. 4-7 illus-
`trate various operational scenarios involving gallery environ-
`ments in several implementations.
`Turning to FIG. 1, gallery environment 100 is illustrated in
`an implementation. Gallery environment 100 may be ren—
`dered when gallery process 200, described in more detail with
`respect to FIG. 2, is implemented by computing system 300
`when executing various software applications or compo-
`nents, such as application program or operating system com-
`ponents. Gallery environment 100 may itself also include
`various software elements or components to facilitate the
`rendering of such views.
`Gallery environment 100 may be any environment within
`which views of items may be rendered and displayed. For
`example, gallery environment 100 includes an initial view
`102 of several groups of tagged items arranged in the groups ,
`based on with which tag of several tags each of the items are
`associated. Gallery environment 100 also includes an
`enhanced view 104 of one of the groups presented in the
`initial view 104 that may be selected for enhanced viewing.
`The items in the enhanced group are associated with the tag
`corresponding to the group selected for enhanced viewing.
`Note that the presentation of initial view 102 may occur prior
`to the presentation of enhanced view 104. In some scenarios,
`enhanced View 104 may replace initial 102, although it may
`be possible to present all or portions of both initial view 102
`and enhanced view 104 at the same time.
`Inparticular, initial view 102 includes tag groups 101, 111,
`and 121. Tag group 101 includes item 103 and item 105. Item
`103 and item 105 may be included in tag group 101 because
`both are tagged with the same or a similar tag. As an example,
`item 103 and item 105 may be tagged as being photos of the
`same person or location or documents pertaining to the same
`
`US 8,949,245 B2
`
`4
`topic. Tag group 111 includes item 113 and item 115. Item
`113 and item 115 may be included in tag group 111 because
`both are tagged with the same or a similar tag. As an example,
`item 113 and item 115 may be tagged as being photos ofthe
`same person or location or documents pertaining to the same
`topic. Tag group 121 includes item 123 and item 125. Item
`123 and item 125 may be included in tag group 121 because
`both are tagged with the same or a similar tag. As an example,
`item 123 and item 125 may be tagged as being photos o l‘ the
`same person or location or documents pertaining to the same
`topic. While location, people, and topics are some examples
`oftags, a variety oftags are possible and are not limited to just
`those disclosed herein.
`111 operation, gallery process 200 may be employed when
`rendering gallery environment 100. To begin, initial View 102
`may be presented for viewing to a user, including a view of
`tagged items arranged in groups ba sed on their respective ta gs
`(step 201). As discussed above, initial View 102 includes tag
`groups 101, 111, and 121, and items 103,105,113,115,123,
`and 125. The items included in the tag groups in initial view
`102 are presented merely for exemplary purposes. More or
`fewer items may be presented. For example, it may be pos-
`sible to display three or four or more items in each group.
`Note also that more items may belong to each group than just
`those presented in initial view 102.
`A tag group may then be identified for enhanced viewing
`(step 203). Identifying the tag group may occur in response to
`a user input, such as a gesture, mouse click, or some other user
`input. Input information corresponding to the user input can
`be processed to identify which group was selected from tag
`groups 101, 111, and 121. It is assumed for exemplary pur—
`poses that lag group 101 was been selected, as indicated by its
`shaded background.
`Next, tagged items are identified for presentation in an
`enhanced view based on which tagged group was selected
`(step 205). As mentioned above, each tagged group may
`include more tagged items thanjust those presented via initial
`view 102. Thus, upon selection of one of tagged groups 101,
`111, and 121, it can be determined Which tag (not shown)
`corresponds to the selected group. From this determination, it
`can then be determined which tagged items are associated
`with the corresponding tag. Those tagged items. or at least a
`portion of them, may then be presented via enhanced view
`104. In some scenarios, they may be referred to as enhanced
`tagged items since they will be presented in the enhanced
`view 104. In some scenarios, the tagged items presented in
`initial view 102 may be referred to as initial tagged items.
`In this example,
`tagged group 101 is selected. Thus, a
`look—up or other similar determination is performed to deter—
`mine which items are associated with the tag corresponding
`to tagged group 101. Those items. or at least a portion ofthem,
`may then be presented in enhanced view 104. Having identi-
`fied the relevant items, enhanced view 104 may be presented
`(step 207).
`111 this example,
`the enhanced tagged items
`included in enhanced view 104 include item 103, item 105,
`and item 107. Item 107 was notpresented in tagged group 101
`within initial view 102, demonstrating that an enhanced
`group oftagged items may include more items then those in a
`corresponding initial group of items.
`111 fact, such a result may occur frequently when groups
`include a sufficient number ofitems such that presenting all of
`them within an initial view may not be practical or otherwise
`desired. For example, several tagged groups may correspond
`to various locations of where photos were taken. One group
`, corresponding to one location may include dozens or hun-
`dreds of photos, or even more. In such a scenario, it may be
`
`MICROSOFT CORP. EX. 1013
`Page 9 of 14
`
`
`
`may be performed with respect to gallery enviromnent 100.
`
`5
`dreds of items in a tagged group all within an initial view,
`Rather, by selecting that particular tagged group, an enhanced
`view of the tagged group may be presented showing addi-
`tional ones of the items in the group not shown in the initial
`view.
`While gallery process 200 generally refers to presenting an
`initial View, identifying a tagged group selected for enhanced
`viewing, identifying enhanced tagged items, and presenting
`an enhanced view, variations 0 f gallery process 200 are pos—
`sible. For example, a variation of gallery process 200 may
`involve initiating each of the aforementioned steps, in addi-
`tion to or in place of actually performing each of the afore-
`mentioned steps. For example, a variation of gallery process
`200 may include some or all of initiating the presentation of
`an initial View, initiating the identification of a tagged group,
`initiating the identification of enhanced tagged items, and
`initiating the presentation of an enhanced view.
`Computing system 300, referred to in FIG. 3, may be
`representative of any computing apparatus, system, or sys-
`tems 011 which gallery process 200 or variations thereof may
`be suitably implemented. Optionally, computing system 300
`may also be suitable for implementing all of or portions of
`gallery enviromnent 100. Implementing gallery process 200
`or variations thereof may include local execution, streaming
`execution, or hosted execution of gallery process 200, or any
`hybrid combination or variation thereof. Examples of com-
`
`puting system 300 include mobile compu ing devices, desk-
`top computers, cell phones, tablet computers, laptop comput-
`ers, notebook computers, media devices, and gaming devices,
`as well as any other type ofmobile computing devices and any
`combination or variation thereof. Examples of computing
`system 300 also include server computers, cloud computing
`platforms, and Virtual machines, as well as any other type of
`computing system, variation, or combination thereof.
`Computing system 300 includes processing system 301,
`storage system 303, software 305, communication interface
`307, and user interface 309. In some implementations, user
`interface 309 may be omitted. Processing system 301 is
`operatively coupled with storage system 3 03, conmiunication
`interface 307, and user interface 309. Processing system 30]
`loads and executes software 305 from storage system 303,
`When executed by computing system 300 in general, and
`processing system 301 in particular, software 305 directs
`computing system 300 to operate as described herein for
`gallery process 200 or variations thereof. Computing system
`300 may optionally include additional devices, features, or
`functionality not discussed here for purposes of brevity.
`Referring still to FIG. 3, processing system 301 may com—
`prise a microprocessor and other circuitry that retrieves and
`executes software 305 from storage system 303. Processing .
`system 301 may be implemented within a single processing
`device but may also be distributed across multiple processing
`devices or sub-systems that cooperate in executing program
`instructions. Examples of processing system 301 include
`general purpose central processing units, application specific
`processors, and logic devices, as well as any other type of
`processing device, combinations, or variations thereof,
`Storage system 303 may comprise any computer readable
`storage media readable by processing system 301 and capable
`of storing software 305. Storage system 303 may include
`volatile and nonvolatile,
`removable and non-removable
`media implemented in any method or technology for storage
`of information, such as computer readable instructions, data
`structures, program modules, or other data. Examples ot'stor—
`age media include random access memory,
`read only
`memory, magnetic disks, optical disks, flash memory, Virtual
`memory and non-virtual memory, magnetic cassettes, mag-
`
`US 8,949,245 B2
`
`6
`netic tape, magnetic disk storage or other magnetic storage
`devices, or any other suitable storage media. In no case is the
`storage media a propagated signal. In addition to storage
`media, in some implementations storage system 303 may also
`include communication media over which software 305 may
`be communicated. Storage system 303 may be implemented
`as a single storage device but may also be implemented across
`multiple storage devices or sub—systems co—located or distrib—
`uted relative to each other. Storage system 303 may comprise
`additional elements, such as a controller, capable ofcommu-
`nicating with processing system 301.
`Software 305 may be implemented in program instructions
`and among other functions may, when executed by comput-
`ing system 300 in general or processing system 301 in par-
`ticular, direct computing system 300 or processing system
`301 to present an initial view of at least a plurality of tagged
`items arranged in a plurality of ta g groups, corresponding to
`a plurality oftags, based on with which ofthe plurality oftags
`each ofthe plurality oftagged items is associated, identify an
`enhanced plurality of tagged items to present based at least in
`part on to which ofthe plurality of tags a tag group, identified
`from the plurality of tag groups for enhanced viewing, corre-
`sponds, and present an enhanced view ofat least the enhanced
`plurality of tagged items. Software 305 may include addi—
`tional processes, programs, or components, such as operating
`system software or other application software. Software 305
`may also comprise firmware or some other form of machine-
`readable processing instructions executable by processing
`system 301.
`In general, software 305 may, when loaded into processing
`system 301 and executed, transform computing system 300
`overall from a general—purpose computing system into a spe—
`cial—purpose computing system customized to facilitate gal—
`lery views as described herein for each implementation,
`Indeed, encoding software 305 on storage system 303 may
`transform the physical structure of storage system 303. The
`specific transfonnation of the physical structure may depend
`on various factors in different
`implementations of this
`description. Examples of such factors may include, but are
`not limited to the technology used to implement the storage
`media of storage system 303 and whether the computer—
`storagc media are characterized as primary or secondary stor-
`age.
`For example, if the computer-storage media are imple-
`mented as semiconductor-based memory, software 305 may
`transform the physical state of the semiconductor memory
`when the program is encoded therein. For example, software
`305 may transform the state oftransistors, capacitors, or other
`discrete circuit elements constituting the semiconductor
`memory. A similar transformation may occur with respect to
`magnetic or optical media. Other transformations ofphysical
`media are possible without departing from the scope of the
`present description, with the foregoing examples provided
`only to facilitate this discussion.
`It should be understood that computing system 300 is gen—
`erally intended to represent a computing system with which
`software 305 is deployed and executed in order to implement
`gallery process 200 (and variations thereof) and optionally all
`or portions of gallery environment 100. However, computing
`system 300 may also represent any computing system on
`which software 305 may be staged and fioni where software
`305 may be distributed, transported, downloaded, or other—
`wise provided to yet another computing system for deploy—
`ment and execution, or yet additional distribution.
`Referring again to FIG. 1, through the operation of com-
`puting system 300 employing software 305, transformations
`
`MICROSOFT CORP. EX. 1013
`Page 10 of 14
`
`
`
`US 8,949,245 B2
`
`computer, or the like.
`
`7
`As an example, gallery environment 100 could be considered
`transformed from one state to another when subject to gallery
`process 200 or variations thereof. In a first state, computing
`system 300 may render initial view 102 in gallery view 102.
`Upon a selection of one of the groups 101, 111, and 121 in
`initial view 102, gallery environment 100 may in a second
`state include a presentation of enhanced view 104.
`Referring again to FIG. 3, communication interface 307
`may include communication connections and devices that
`allow for commtlnicationbetwcen computing system 300 and
`other computing systems (not shown) over a communication
`network or collection of networks (not shown). Examples of
`connections and devices that together allow for inter-system
`comimmication may include network interface cards, anten-
`nas, power amplifiers, RF circuitry, transceivers, and other
`communication circuitry. The cormections and devices may
`communicate over communication media to exchange com—
`munications with other computing systems or networks of
`systems, such as metal, glass, air, or any other suitable com-
`munication media. The aforementioned communication
`media, network, connections, and devices are well known and
`need not be discussed at length here.
`User interface 309 may include a mouse, a voice input
`device, a touch input device for receiving a touch gesture from
`a user, a motion input device for detecting non—tou ch gestures
`and other motions by a user, and other comparable input
`devices and associated processing elements capable of
`receiving user input from a user. Output devices such as a
`display, speakers, haptic devices, and other types of output
`devices may also be included in user interface 309. The afore-
`mentioned user input devices are well known in the art and
`need not be discussed at length here. User interface 309 may
`also include associated user interface software executable by
`processing system 301 in support ofthe various user input and
`output devices discussed above. Separately or in conjunction
`with each other and other hardware and software elements,
`the user interface software and devices may provide a graphi-
`cal user interface, a natural user interface, or any other kind of
`user interface.
`FIG. 4 illustrates gallery scenario 400 in an implementa—
`tion involving mobile phone 401, Gallery scenario 400 may
`apply as well to any other type of computing device, such as
`a tablet device, laptop computer, desktop computer, or the
`like. In this scenario, mobile phone 401 includes control
`element 403 and display 405. Mobile phone 401 may include
`additional elements that are not illustrated here for purposes
`of clarity. In gallery scenario 400, mobile phone 401 may
`execute a gallery application or interface with a remote gal—
`lery application (such as one hosted on a website) capable of
`producing gallery views, such as initial view 407 and .
`enhanced View 409.
`Initial view 407 is presented by way of display 405 and
`includes Views of several groups associated with several tags.
`The items in presented in enhanced View 409 are associated
`with the tag corresponding to a group selected from initial
`View 407 for enhanced viewing. The presentation ofinitial
`view 407 may occur prior to the presentation of enhanced
`view 409. In some scenarios, enhanced View 407 may replace
`initial 409, although it may be possible to present all or
`portions of initial View 407 and enhanced view 409 at the
`same time.
`In initial viwv 407, items 411, 413, and 415 associated with
`tag A are presented as a group; items 421, 423, and 425
`associated with tag B are presented as another group; and
`items 431, 433, and 435 associated with tag C are presented as
`yet another group. The items presented in groups along with
`tags A, B, and C are done so because they are each associated
`
`8
`with at least one or another of the tags, The tags may be
`automatically associated with the items or a manual associa—
`tion may be made. The order in which the groups are pre-
`sented may be determined randomly, alphabetically based on
`the spelling of each tag, by the quantity of items included in
`each group, or by any other suitable criteria.
`As an example, items 411, 413, and 415 may be photos
`corresponding to the same or a similar location, or documents
`pertaining to the same or similar topic, generically repre—
`sented by tag A. Likewise, items 421, 423, and 425 may be
`photos corresponding to the same or a similar location, or
`documents pertaining to the same or similar topic, generically
`represented by tag B. Items 431, 433, and 435 may be photos
`corresponding to the same or a similar location, or documents
`pertaining to the same topic, as represented by tag C.
`While location, people, and topics are some examples of
`tags, a variety of tags are possible and are not limited to just
`those disclosed herein. Moreover, the items included in the
`tag groups in initial view 407 are presented merely for exem-
`plary purposes. More or fewer items may be presented. For
`example, it may be possible to display two or four items in
`each group. More items may belong to each group than just
`those presented herein.
`Continuing with gallery scenario 400, in operation a user
`may desire to view an enha