US20140059477A1 - Localized thumbnail preview of related content during spatial browsing - Google Patents
Localized thumbnail preview of related content during spatial browsing Download PDFInfo
- Publication number
- US20140059477A1 US20140059477A1 US14/066,469 US201314066469A US2014059477A1 US 20140059477 A1 US20140059477 A1 US 20140059477A1 US 201314066469 A US201314066469 A US 201314066469A US 2014059477 A1 US2014059477 A1 US 2014059477A1
- Authority
- US
- United States
- Prior art keywords
- view
- field
- presented
- image
- context bar
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04806—Zoom, i.e. interaction techniques or interactors for controlling the zooming operation
Definitions
- Conventional spatial exploration tools typically provide for the browsing of spatial environments and present information related to objects in a field of view of the spatial environment, such as names, identifiers, and other related information.
- information related to objects in a field of view of the spatial environment such as names, identifiers, and other related information.
- Such in-the-field-of-view presentation tends to clutter the field of view and obscure features of the environment being viewed.
- conventional special exploration tools can become so cluttered with information that seeing an un-obscured image of the spatial environment becomes nearly impossible.
- the present examples provide technologies, systems, and methods for exploring a virtual space without cluttering or obscuring a field of view of the virtual space with overlaid names, labels, symbols, markers, and the like. Instead, a user may select objects of interest presented in a context bar to locate them in the field of view. Further, the user can make use of various forms of selection to zoom in on objects or points in the field of view and access information regarding the virtual space and objects without adding persistent clutter to the field of view.
- FIG. 1 is an example display from a conventional spatial exploration tool.
- field of view (“FOV”) shows a satellite image of a surface of the earth.
- FIG. 2 is another example display from a conventional spatial exploration tool.
- FIG. 3 is an example display from an example improved spatial exploration tool.
- FIG. 4 is a block diagram of an example improved spatial exploration system.
- FIG. 5 is a block diagram showing the example improved spatial exploration system of HG, 4 including one method of how to locate an object in the field of view.
- FIG. 6 is a block diagram showing the example improved spatial exploration system of FIG. 4 including one method of how to identify an object in the field of view.
- FIG. 7 is a block diagram showing an example method for exploring a virtual space without presenting a persistently cluttered field of view.
- FIG. 8 is a block diagram showing an example computing environment in which the technologies described herein may be implemented.
- FIG. 1 is an example display from a conventional spatial exploration tool.
- field of view (“FOV”) 110 shows a satellite image of a surface of the earth.
- FOV field of view
- Conventional spatial exploration tools tend to use such an overlay display method as a matter of course such that it may not be possible for a user to view an unobscured virtual space image.
- virtual space generally refers to a representation of some special environment, actual or imaginary, from a particular point of reference, such as outer space (the Earth, for example, being a point of reference) or some other space from a particular point of reference (some point on the Earth such as a landscape or aerial view, for example).
- spatial environment generally refers to some space a real space, virtual space, and/or imaginary space. Such spaces may, for example, be galactic, landscape, aerial, subatomic, or the like.
- cluttered typically refers to hindering, obscuring, covering, blocking, and/or the like a user's view of an image of the virtual space, and/or representations of objects within the virtual space, in the field of view with names, labels, symbols, markers, and the like overlaid on the image.
- FIG. 2 is another example display from a conventional spatial exploration tool.
- FOV 210 shows a representation of outer space, the area of the Andromeda constellation in particular.
- the image in this example does not take on the appearance of the actual space and is cluttered and obscured by overlaid names, labels, symbols, markers, and the like.
- FIG. 3 is an example display from an example improved spatial exploration tool.
- FOV 310 shows a representation of outer space, the area of the Andromeda constellation in particular, as in FIG. 2 .
- the virtual space image takes on the appearance of the actual sky view and is generally not cluttered or obscured by overlaid names, labels, symbols, markers, and the like.
- example object bar 320 typically presents thumbnail images of objects within the current FOV, such as FOV 310 . A user may select a thumbnail to zoom in on the corresponding object and/or access information associated with the object.
- a user may use a mouse control or the like to hover over a thumbnail (or otherwise indicate a desired thumbnail) causing the corresponding object in FOV 310 to be indicated, such as by noticeably marking it or highlighting it or the like.
- object refers to anything identifiable by the spatial exploration tool; for example, a road, landmark, galaxy, constellation, building, or anything else typically within the virtual space considered by the tool.
- FIG. 4 is a block diagram of an example improved spatial exploration system 400 .
- System 400 includes field of view (“FOV”) 410 and context bar 420 .
- FOV 410 typically presents images of a virtual space including objects in the virtual space (examples represented by circles such as circle 412 ).
- Not shown in FIG. 4 are one or more data stores that provide virtual space and object images, information, and data to system 400 .
- a user may generally browse or explore a virtual space by moving the FOV to a desired location in the virtual space via suitable user interface mechanisms. Further, the user may typically zoom in or out of the virtual space as desired, thus narrowing or widening the FOV respectively.
- Context bar 420 typically includes representations of objects in the current FOV (examples represented by rectangles such as rectangle 422 ). Typically (unlike shown in FIG. 4 ) there is at least one object representation in context bar 420 for each object in FOV 410 . As the FOV changes, the objects represented in context bar typically change accordingly.
- representation 422 includes a thumbnail image of the object, as shown in FIG. 3 . Such a thumbnail image may also include identification information such as a name or title of the object and/or other information related to the object.
- FOV 410 of example system 400 is not cluttered or obscured by overlaid names, labels, symbols, markers, or the like as is common with conventional spatial exploration tools.
- FIG. 5 is a block diagram showing the example improved spatial exploration system 400 of FIG. 4 including one method of how to locate an object in the field of view.
- a user may locate an identified object in FOV 410 by selecting the corresponding thumbnail on context bar 420 .
- a user may hover mouse cursor 530 over thumbnail 524 (as indicated by the bold rectangle) causing system 400 to highlight corresponding object 514 in FOV 410 (as indicated by the bold circle and the presentation of the example object name, “Mu Andomedea”).
- system 400 removes the object highlighting.
- a user may select any of the thumbnails presented in context bar 420 , such selection causing system 400 to highlight the corresponding object in the FOV.
- De-selecting the thumbnails generally causes system 400 to un-highlight the corresponding object in the FOV.
- the highlighting is not persistent and the FOV remains uncluttered showing only min mal additional information in response to a user selection and removing the information upon de-selection.
- Selection may be made via mouse cursor hovering over a thumbnail (as in the example above) or via any other suitable means or method or the like. Other forms of selection may be used to access other information regarding the virtual space and/or objects in the virtual space and/or FOV.
- FIG. 6 is a block diagram showing the example improved spatial exploration system 400 of FIG. 4 including one method of how to identify an object in the field of view.
- a user selects object 610 presented in FOV 410 by pointing to it with example mouse cursor 630 .
- a user may hover mouse cursor 630 over object 614 causing system 400 to highlight object 614 in FOV 410 (as indicated by the bold circle and the presentation of the example object name, “Mu Andomedea”).
- system 400 removes the object highlighting.
- a user may select any object presented in FOV 410 , such selection causing system 400 to highlight the corresponding object in the FOV if it is identified by system 400 . If an object at cursor 630 is not identified, or if there is no object at cursor 630 , then typically a nearby object is highlighted. De-selecting the object (moving cursor 630 away) generally causes system 400 to un-highlight the object in the FOV. In this manner, the highlighting is not persistent and the FOV remains uncluttered showing only minimal additional information in response to a user selection and removing the information upon de-selection. Selection may be made via mouse cursor hovering over a thumbnail (as in the example above) or via any other suitable means or method or the like. Other forms of selection may be used to access other information regarding the virtual space and/or objects in the virtual space and/or FOV.
- FIG. 7 is a block diagram showing an example method 700 for exploring a virtual space without presenting a persistently cluttered field of view.
- method 700 is used with the example improved spatial exploration system 400 of FIG. 4 .
- Method 700 provides for an uncluttered FOV while allowing a user to explore a virtual space, identify objects in the virtual space, and obtain additional information related to the virtual space and objects.
- Block 710 typically indicates selecting a particular FOV within a virtual space.
- a user may select a particular FOV as described in connection with FIG. 4 .
- method 700 typically continues at block 71
- Block 711 typically indicates updating a context bar, such as context bar 420 of FIG. 4 .
- the context bar is updated by the improved spatial exploration system as described in connection with FIG. 4 to include thumbnail images of objects identified by the system within the current FOV.
- Block 712 typically indicates detecting a change in the FOV.
- the user may change the FOV to another portion of the virtual space. Such changing can include moving position within the virtual space and/or zoom in our out of the virtual space. If the FOV is changes, method 700 typically continues at block 710 . Otherwise, method 700 typically continues at block 713 .
- Block 713 typically indicates detecting an identification selection of a thumbnail (“TN”).
- identification selection of a thumbnail is made by selecting a thumbnail in a context bar as described in connection with FIG. 5 . If a thumbnail is identification selected, method 700 typically continues at block 714 . Otherwise, method 700 typically continues at block 715 .
- Block 714 typically indicates highlighting an object in response to a thumbnail selection.
- an object is highlighted in response to a thumbnail selection, and un-highlighted in response to a de-selection, as described in connection with FIG. 5 .
- method 700 typically continues at block 715 .
- Block 715 typically indicates detecting a smooth zoom (“szoom”) selection of a thumbnail.
- smooth zoom selection of a thumbnail is made by left mouse clicking a thumbnail in a context bar. Other methods of smooth zoom selection may alternatively be used. If a thumbnail is smooth zoom selected, method 700 typically continues at block 716 . Otherwise, method 700 typically continues at block 717 .
- Block 716 typically indicates zoom ir to a close up view of the object in the FOV corresponding to the smooth zoom selected thumbnail. Once the zoom is complete, method 700 typically continues at block 712 .
- Block 717 typically indicates detecting a warp zoom (“wzoom”) selection of a thumbnail.
- warp zoom selection of a thumbnail is made by double mouse clicking a thumbnail in a context bar.
- Other methods of warp zoom selection may alternatively be used.
- a warp zoom is typically a cross-fade—a blend transition from one image to another (e.g., one image fades out while the other image fades in).
- a warp zoom may be an immediate transition from one image to another. If a thumbnail is warp zoom selected, method 700 typically continues at block 718 . Otherwise, method 700 typically continues at block 719 .
- Block 718 typically indicates warp zooming or cross-fading in to a close up view of the object in the FOV corresponding to the warp zoom selected thumbnail.
- Block 719 typically indicates detecting a point selection in the FOV.
- a point selection in the FOV is made by moving a mouse cursor to point to a particular point in the FOV and hovering the mouse cursor over the same point for a conventional period of time. Other methods of point selection may alternatively be used. If a point in the FOV is selected, method 700 typically continues at block 720 . Otherwise, method 700 typically continues at block 721 .
- Block 720 typically indicates highlighting the object that is point selected or an object close to the point in the FOV that is point selected.
- the identified object in the virtual space closest to the point selected is highlighted.
- the highlighting and un-highlighting is performed as described in connection with FIG. 6 .
- Block 721 typically indicates detecting an information selection of a point in the FOV.
- information selection of a point in the FOV is made by moving a mouse cursor to point to a particular point or object in the FOV and keeping the mouse cursor pointing to the same point or object for a conventional period of time. Other methods of information selection of a point may alternatively be used. If point information in the FOV is selected, method 700 typically continues at block 722 . Otherwise, method 700 typically continues at block 723 .
- Block 722 typically indicates presenting information about the object that is selected or an object close to the point in the FOV that is information selected.
- information is presented in a pop-up dialog box. In this example, once the point is de-selected, the information is no longer presented. Once the information is presented, method 700 typically continues at block 723 .
- Block 723 typically indicates detecting a point zoom selection in the FOV.
- point zoom selection in the FOV is made by double mouse clicking a point in the FOV. Other methods of point zoom selection may alternatively be used. If a point zoom in the FOV is selected, method 700 typically continues at block 724 . Otherwise, method 700 typically continues at block 712 .
- Block 724 typically indicates zooming in to a close up view of the point in the FOV corresponding to the point zoom selection. Once the zoom is complete, method 700 typically continues at block 712 .
- FIG. 8 is a block diagram showing an example computing environment 800 in which the technologies described herein may be implemented.
- a suitable computing environment may be implemented with numerous general purpose or special purpose systems. Examples of well known systems may include, hut are not limited to, cell phones, personal digital assistants (“PDA”), personal computers (“PC”), hand-held or laptop devices, microprocessor-based systems, multiprocessor systems, servers, workstations, consumer electronic devices, set-top boxes, and the like.
- Computing environment 800 typically includes a general-purpose computing system in the form of a computing device 801 coupled to various components, such as peripheral devices 802 , 803 , 804 and the like.
- System 800 may couple to various other components, such as input devices 803 , including voice recognition, touch pads, buttons, keyboards and/or pointing devices, such as a mouse or trackball, via one or more input/output (“I/O”) interfaces 812 .
- the components of computing device 801 may include one or more processors (including central processing units (“CPU”), graphics processing units (“CPU”), microprocessors (“ ⁇ P”), and the like) 807 , system memory 809 , and a system bus 808 that typically couples the various components.
- processors including central processing units (“CPU”), graphics processing units (“CPU”), microprocessors (“ ⁇ P”), and the like
- System bus 808 represents any number of several types of bus structures, including a memory bus or memory controller, a peripheral bus, a serial bus, an accelerated graphics port, a processor or local bus using any of a variety of bus architectures, and the like.
- System memory 809 may include computer readable media in the form of volatile memory, such as random access memory (“RAM”), and/or non-volatile memory, such as read only memory (“ROM”) or flash memory (“FLASH”).
- RAM random access memory
- ROM read only memory
- FLASH flash memory
- a basic input/output system (“BIOS”) may be stored in non-volatile or the like.
- System memory 809 typically stores data, computer-executable instructions and/or program modules comprising computer-executable instructions that are immediately accessible to and/or presently operated on by one or more of the processors 807 .
- Mass storage devices 804 and 810 may be coupled to computing device 801 or incorporated into computing device 801 via coupling to the system bus.
- Such mass storage devices 804 and 810 may include non-volatile RAM, a magnetic disk drive which reads from and/or writes to a removable, non-volatile magnetic disk (e.g., a “floppy disk”) 805 , and/or an optical disk drive that reads from and/or writes to a non-volatile optical disk such as a CD ROM, DVD ROM 806 .
- a mass storage device, such as hard disk 810 may include non-removable storage medium.
- Other mass storage devices may include memory cards, memory sticks, tape storage devices, and the like.
- Any number of computer programs, files, data structures, and the like may be stored in mass storage 810 , other storage devices 804 , 805 , 806 and system memory 809 (typically limited by available space) including, by way of example and not limitation, operating systems, application programs, data files, directory structures, computer-executable instructions, and the like.
- Output components or devices may be coupled to computing device 801 , typically via an interface such as a display adapter 811 .
- Output device 802 may be a liquid crystal display (“LCD”).
- Other example output devices may include printers, audio outputs, voice outputs, cathode ray tube (“CRT”) displays, tactile devices or other sensory output mechanisms, or the like.
- Output devices may enable computing device 801 to interact with human operators or other machines, systems, computing environments, or the like.
- a user may interface with computing environment 800 via any number of different I/O devices 803 such as a touch pad, buttons, keyboard, mouse, joystick, game pad, data port, and the like.
- I/O devices may be coupled to processor 807 via I/O interfaces 812 which may be coupled to system bus 808 , and/or may be coupled by other interfaces and bus structures, such as a parallel port, game port, universal serial bus (“USB”), fire wire, infrared (“IR”) port, and the like.
- I/O interfaces 812 may be coupled to system bus 808 , and/or may be coupled by other interfaces and bus structures, such as a parallel port, game port, universal serial bus (“USB”), fire wire, infrared (“IR”) port, and the like.
- USB universal serial bus
- IR infrared
- Computing device 801 may operate in a networked environment via communications connections to one or more remote computing devices through one or more cellular networks, wireless networks, local area networks (“LAN”), wide area networks (“WAN”), storage area networks (“SAN”), the Internet, radio links, optical links and the like.
- Computing device 801 may be coupled to a network via network adapter 813 or the like, or, alternatively, via a modem, digital subscriber line (“DSL”) link, integrated services digital network (“ISDN”) link, Internet link, wireless link, or the like.
- DSL digital subscriber line
- ISDN integrated services digital network
- Communications connection 814 typically provides a coupling to communications media, such as a network.
- Communications media typically provide computer-readable and computer-executable instructions, data structures, files, program modules and other data using a modulated data signal, such as a carrier wave or other transport mechanism.
- modulated data signal typically means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
- communications media may include wired media, such as a wired network or direct-wired connection or the like, and wireless media, such as acoustic, radio frequency, infrared, or other wireless communications mechanisms.
- Power source 890 such as a battery or a power supply, typically provides power for portions or all of computing environment 800 .
- power source 890 may be a battery.
- power source 890 may be a power supply designed to connect to an alternating current (“AC”) source, such as via a wall outlet.
- AC alternating current
- an electronic badge may be comprised of a coil of wire along with a simple processing unit 807 or the like, the coil configured to act as power source 890 when in proximity to a card reader device or the like.
- a coil may also be configure to act as an antenna coupled to the processing unit 807 or the like, the coil antenna capable of providing a form of communication between the electronic badge and the card reader device.
- Such communication may not involve networking, but may alternatively be general or special purpose communications via telemetry, point-to-point, RF, IR, audio, or other means.
- An electronic card may not include display 802 , I/O device 803 , or many of the other components described in connection with FIG. 8 .
- Other mobile devices that may not include many of the components described in connection with FIG. 8 , by way of example and not limitation, include electronic bracelets, electronic tags, implantable devices, and the like.
- a remote computer or storage device may store computer-readable and computer-executable instructions in the form of software applications and data.
- a local computer may access the remote computer or storage device via the network and download part or all of a software application or data and may execute any computer-executable instructions.
- the local computer may download pieces of the software or data as needed, or distributively process the software by executing some of the instructions at the local computer and some at remote computers and/or devices.
- DSP digital signal processor
- PLA programmable logic array
- discrete circuits and the like.
- DSP digital signal processor
- electronic apparatus may include computing devices or consumer electronic devices comprising any software, firmware or the like, or electronic devices or circuits comprising no software, firmware or the like.
- firmware typically refers to executable instructions, code, data, applications, programs, or the like maintained in an electronic device such as a ROM.
- software generally refers to executable instructions, code, data, applications, programs, or the like maintained in or on any form of computer-readable media.
- computer-readable media typically refers to system memory, storage devices and their associated media, and the like.
Abstract
Technologies, systems, and methods for exploring a virtual space without cluttering or obscuring a field of view of the virtual space with overlaid names, labels, symbols, markers, and the like. Instead, a user may select objects of interest presented in a context bar to locate them in the field of view. Further, the user can make use of various forms of selection to zoom in on objects or points in the field of view and access information regarding the virtual space and objects without adding persistent clutter to the field of view.
Description
- This application is a Continuation of and claims benefit from U.S. patent application Ser. No. 11/941,098 that was filed on Nov. 16, 2007, and that is incorporated herein by reference in its entirety.
- Conventional spatial exploration tools typically provide for the browsing of spatial environments and present information related to objects in a field of view of the spatial environment, such as names, identifiers, and other related information. Such in-the-field-of-view presentation tends to clutter the field of view and obscure features of the environment being viewed. For example, conventional special exploration tools can become so cluttered with information that seeing an un-obscured image of the spatial environment becomes nearly impossible.
- The following presents a simplified summary of the disclosure in order to provide a basic understanding to the reader. This summary is not an extensive overview of the disclosure and it does not identify key/critical elements of the invention or delineate the scope of the invention. Its sole purpose is to present some concepts disclosed herein in a simplified form as a prelude to the more detailed description that is presented later.
- The present examples provide technologies, systems, and methods for exploring a virtual space without cluttering or obscuring a field of view of the virtual space with overlaid names, labels, symbols, markers, and the like. Instead, a user may select objects of interest presented in a context bar to locate them in the field of view. Further, the user can make use of various forms of selection to zoom in on objects or points in the field of view and access information regarding the virtual space and objects without adding persistent clutter to the field of view.
- Many of the attendant features will be more readily appreciated as the some become better understood by reference to the following detailed description considered in connection with the accompanying drawings.
- The present description will be better understood from the following detailed description considered in connection with the accompanying drawings, wherein:
-
FIG. 1 is an example display from a conventional spatial exploration tool. In this example, field of view (“FOV”) shows a satellite image of a surface of the earth. -
FIG. 2 is another example display from a conventional spatial exploration tool. -
FIG. 3 is an example display from an example improved spatial exploration tool. -
FIG. 4 is a block diagram of an example improved spatial exploration system. -
FIG. 5 is a block diagram showing the example improved spatial exploration system of HG, 4 including one method of how to locate an object in the field of view. -
FIG. 6 is a block diagram showing the example improved spatial exploration system ofFIG. 4 including one method of how to identify an object in the field of view. -
FIG. 7 is a block diagram showing an example method for exploring a virtual space without presenting a persistently cluttered field of view. -
FIG. 8 is a block diagram showing an example computing environment in which the technologies described herein may be implemented. - Like reference numerals are used to designate like parts in the accompanying drawings.
- The detailed description provided below in connection with the accompanying drawings is intended as a description of the present examples and is not intended to represent the only forms in which the present examples may be constructed or utilized. The description sets forth at least some of the functions of the examples and/or the sequence of steps for constructing and operatiing examples. However, the same or equivalent functions and sequences may be accomplished by different examples.
- Although the present examples are described and illustrated herein as being implemented in a computing environment, the environment described is provided as an example and not a limitation. As those skilled in the art will appreciate, the present examples are suitable for application in a variety of different types of computing environments.
-
FIG. 1 is an example display from a conventional spatial exploration tool. In this example, field of view (“FOV”) 110 shows a satellite image of a surface of the earth. Unfortunately, the image in this example is cluttered and obscured by overlaid names, labels, symbols, markers, and the like. Conventional spatial exploration tools tend to use such an overlay display method as a matter of course such that it may not be possible for a user to view an unobscured virtual space image. The term “virtual space” as used herein generally refers to a representation of some special environment, actual or imaginary, from a particular point of reference, such as outer space (the Earth, for example, being a point of reference) or some other space from a particular point of reference (some point on the Earth such as a landscape or aerial view, for example). The term “spatial environment” as used herein generally refers to some space a real space, virtual space, and/or imaginary space. Such spaces may, for example, be galactic, landscape, aerial, subatomic, or the like. The term “cluttered” as used herein with respect to a field of view typically refers to hindering, obscuring, covering, blocking, and/or the like a user's view of an image of the virtual space, and/or representations of objects within the virtual space, in the field of view with names, labels, symbols, markers, and the like overlaid on the image. -
FIG. 2 is another example display from a conventional spatial exploration tool. In this example, FOV 210 shows a representation of outer space, the area of the Andromeda constellation in particular. Unfortunately, the image in this example does not take on the appearance of the actual space and is cluttered and obscured by overlaid names, labels, symbols, markers, and the like. -
FIG. 3 is an example display from an example improved spatial exploration tool. In this example, FOV 310 shows a representation of outer space, the area of the Andromeda constellation in particular, as inFIG. 2 . In this example, the virtual space image takes on the appearance of the actual sky view and is generally not cluttered or obscured by overlaid names, labels, symbols, markers, and the like. Further,example object bar 320 typically presents thumbnail images of objects within the current FOV, such asFOV 310. A user may select a thumbnail to zoom in on the corresponding object and/or access information associated with the object. Further, a user may use a mouse control or the like to hover over a thumbnail (or otherwise indicate a desired thumbnail) causing the corresponding object inFOV 310 to be indicated, such as by noticeably marking it or highlighting it or the like. As used herein, the term “object” refers to anything identifiable by the spatial exploration tool; for example, a road, landmark, galaxy, constellation, building, or anything else typically within the virtual space considered by the tool. -
FIG. 4 is a block diagram of an example improvedspatial exploration system 400.System 400 includes field of view (“FOV”) 410 andcontext bar 420. FOV 410 typically presents images of a virtual space including objects in the virtual space (examples represented by circles such as circle 412). Not shown inFIG. 4 are one or more data stores that provide virtual space and object images, information, and data tosystem 400. Usingsystem 400, a user may generally browse or explore a virtual space by moving the FOV to a desired location in the virtual space via suitable user interface mechanisms. Further, the user may typically zoom in or out of the virtual space as desired, thus narrowing or widening the FOV respectively.Context bar 420 typically includes representations of objects in the current FOV (examples represented by rectangles such as rectangle 422). Typically (unlike shown inFIG. 4 ) there is at least one object representation incontext bar 420 for each object inFOV 410. As the FOV changes, the objects represented in context bar typically change accordingly. In one example,representation 422 includes a thumbnail image of the object, as shown inFIG. 3 . Such a thumbnail image may also include identification information such as a name or title of the object and/or other information related to the object. In general, FOV 410 ofexample system 400 is not cluttered or obscured by overlaid names, labels, symbols, markers, or the like as is common with conventional spatial exploration tools. -
FIG. 5 is a block diagram showing the example improvedspatial exploration system 400 ofFIG. 4 including one method of how to locate an object in the field of view. In this example, a user may locate an identified object inFOV 410 by selecting the corresponding thumbnail oncontext bar 420. In one example, a user may hovermouse cursor 530 over thumbnail 524 (as indicated by the bold rectangle) causingsystem 400 to highlightcorresponding object 514 in FOV 410 (as indicated by the bold circle and the presentation of the example object name, “Mu Andomedea”). In this example, oncemouse cursor 530 is removed from hovering overthumbnail 524 thensystem 400 removes the object highlighting. In general, a user may select any of the thumbnails presented incontext bar 420, suchselection causing system 400 to highlight the corresponding object in the FOV. De-selecting the thumbnails generally causessystem 400 to un-highlight the corresponding object in the FOV. In this manner, the highlighting is not persistent and the FOV remains uncluttered showing only min mal additional information in response to a user selection and removing the information upon de-selection. Selection may be made via mouse cursor hovering over a thumbnail (as in the example above) or via any other suitable means or method or the like. Other forms of selection may be used to access other information regarding the virtual space and/or objects in the virtual space and/or FOV. -
FIG. 6 is a block diagram showing the example improvedspatial exploration system 400 ofFIG. 4 including one method of how to identify an object in the field of view. In this example, a user selects object 610 presented inFOV 410 by pointing to it withexample mouse cursor 630. In one example, a user may hovermouse cursor 630 overobject 614 causingsystem 400 to highlightobject 614 in FOV 410 (as indicated by the bold circle and the presentation of the example object name, “Mu Andomedea”). In this example, oncemouse cursor 630 is removed from hovering overobject 614 thensystem 400 removes the object highlighting. In general, a user may select any object presented inFOV 410, suchselection causing system 400 to highlight the corresponding object in the FOV if it is identified bysystem 400. If an object atcursor 630 is not identified, or if there is no object atcursor 630, then typically a nearby object is highlighted. De-selecting the object (movingcursor 630 away) generally causessystem 400 to un-highlight the object in the FOV. In this manner, the highlighting is not persistent and the FOV remains uncluttered showing only minimal additional information in response to a user selection and removing the information upon de-selection. Selection may be made via mouse cursor hovering over a thumbnail (as in the example above) or via any other suitable means or method or the like. Other forms of selection may be used to access other information regarding the virtual space and/or objects in the virtual space and/or FOV. -
FIG. 7 is a block diagram showing anexample method 700 for exploring a virtual space without presenting a persistently cluttered field of view. In one example,method 700 is used with the example improvedspatial exploration system 400 ofFIG. 4 .Method 700 provides for an uncluttered FOV while allowing a user to explore a virtual space, identify objects in the virtual space, and obtain additional information related to the virtual space and objects. - Block 710 typically indicates selecting a particular FOV within a virtual space. In one example, a user may select a particular FOV as described in connection with
FIG. 4 . Once a particular FOV has been selected,method 700 typically continues at block 71 -
Block 711 typically indicates updating a context bar, such as context bar 420 ofFIG. 4 . In one example, the context bar is updated by the improved spatial exploration system as described in connection withFIG. 4 to include thumbnail images of objects identified by the system within the current FOV. Once the context bar has been updated,method 700 typically continues atblock 712. - Block 712 typically indicates detecting a change in the FOV. In one example, the user may change the FOV to another portion of the virtual space. Such changing can include moving position within the virtual space and/or zoom in our out of the virtual space. If the FOV is changes,
method 700 typically continues atblock 710. Otherwise,method 700 typically continues atblock 713. - Block 713 typically indicates detecting an identification selection of a thumbnail (“TN”). In one example, identification selection of a thumbnail is made by selecting a thumbnail in a context bar as described in connection with
FIG. 5 . If a thumbnail is identification selected,method 700 typically continues atblock 714. Otherwise,method 700 typically continues atblock 715. - Block 714 typically indicates highlighting an object in response to a thumbnail selection. In one example, an object is highlighted in response to a thumbnail selection, and un-highlighted in response to a de-selection, as described in connection with
FIG. 5 . Once the object is highlighted,method 700 typically continues atblock 715. - Block 715 typically indicates detecting a smooth zoom (“szoom”) selection of a thumbnail. In one example, smooth zoom selection of a thumbnail is made by left mouse clicking a thumbnail in a context bar. Other methods of smooth zoom selection may alternatively be used. If a thumbnail is smooth zoom selected,
method 700 typically continues atblock 716. Otherwise,method 700 typically continues atblock 717. - Block 716 typically indicates zoom ir to a close up view of the object in the FOV corresponding to the smooth zoom selected thumbnail. Once the zoom is complete,
method 700 typically continues atblock 712. - Block 717 typically indicates detecting a warp zoom (“wzoom”) selection of a thumbnail. In one example, warp zoom selection of a thumbnail is made by double mouse clicking a thumbnail in a context bar. Other methods of warp zoom selection may alternatively be used. A warp zoom is typically a cross-fade—a blend transition from one image to another (e.g., one image fades out while the other image fades in). Alternatively, a warp zoom may be an immediate transition from one image to another. If a thumbnail is warp zoom selected,
method 700 typically continues atblock 718. Otherwise,method 700 typically continues atblock 719. - Block 718 typically indicates warp zooming or cross-fading in to a close up view of the object in the FOV corresponding to the warp zoom selected thumbnail. Once the cross-fade is complete,
method 700 typically continues atblock 712. - Block 719 typically indicates detecting a point selection in the FOV. In one example, a point selection in the FOV is made by moving a mouse cursor to point to a particular point in the FOV and hovering the mouse cursor over the same point for a conventional period of time. Other methods of point selection may alternatively be used. If a point in the FOV is selected,
method 700 typically continues atblock 720. Otherwise,method 700 typically continues atblock 721. - Block 720 typically indicates highlighting the object that is point selected or an object close to the point in the FOV that is point selected. In one example, the identified object in the virtual space closest to the point selected is highlighted. In one example, the highlighting and un-highlighting is performed as described in connection with
FIG. 6 . Once the highlighting is complete,method 700 typically continues atblock 721. - Block 721 typically indicates detecting an information selection of a point in the FOV. In one example, information selection of a point in the FOV is made by moving a mouse cursor to point to a particular point or object in the FOV and keeping the mouse cursor pointing to the same point or object for a conventional period of time. Other methods of information selection of a point may alternatively be used. If point information in the FOV is selected,
method 700 typically continues atblock 722. Otherwise,method 700 typically continues atblock 723. - Block 722 typically indicates presenting information about the object that is selected or an object close to the point in the FOV that is information selected. In one example, information is presented in a pop-up dialog box. In this example, once the point is de-selected, the information is no longer presented. Once the information is presented,
method 700 typically continues atblock 723. - Block 723 typically indicates detecting a point zoom selection in the FOV. In one example, point zoom selection in the FOV is made by double mouse clicking a point in the FOV. Other methods of point zoom selection may alternatively be used. If a point zoom in the FOV is selected,
method 700 typically continues atblock 724. Otherwise,method 700 typically continues atblock 712. - Block 724 typically indicates zooming in to a close up view of the point in the FOV corresponding to the point zoom selection. Once the zoom is complete,
method 700 typically continues atblock 712. -
FIG. 8 is a block diagram showing anexample computing environment 800 in which the technologies described herein may be implemented. A suitable computing environment may be implemented with numerous general purpose or special purpose systems. Examples of well known systems may include, hut are not limited to, cell phones, personal digital assistants (“PDA”), personal computers (“PC”), hand-held or laptop devices, microprocessor-based systems, multiprocessor systems, servers, workstations, consumer electronic devices, set-top boxes, and the like. -
Computing environment 800 typically includes a general-purpose computing system in the form of acomputing device 801 coupled to various components, such asperipheral devices System 800 may couple to various other components, such asinput devices 803, including voice recognition, touch pads, buttons, keyboards and/or pointing devices, such as a mouse or trackball, via one or more input/output (“I/O”) interfaces 812. The components ofcomputing device 801 may include one or more processors (including central processing units (“CPU”), graphics processing units (“CPU”), microprocessors (“μP”), and the like) 807,system memory 809, and asystem bus 808 that typically couples the various components.Processor 807 typically processes or executes various computer-executable instructions to control the operation ofcomputing device 801 and to communicate with other electronic and/or computing devices, systems or environment (not shown) via various communications connections such as anetwork connection 814 or the like.System bus 808 represents any number of several types of bus structures, including a memory bus or memory controller, a peripheral bus, a serial bus, an accelerated graphics port, a processor or local bus using any of a variety of bus architectures, and the like. -
System memory 809 may include computer readable media in the form of volatile memory, such as random access memory (“RAM”), and/or non-volatile memory, such as read only memory (“ROM”) or flash memory (“FLASH”). A basic input/output system (“BIOS”) may be stored in non-volatile or the like.System memory 809 typically stores data, computer-executable instructions and/or program modules comprising computer-executable instructions that are immediately accessible to and/or presently operated on by one or more of theprocessors 807. -
Mass storage devices computing device 801 or incorporated intocomputing device 801 via coupling to the system bus. Suchmass storage devices DVD ROM 806. Alternatively, a mass storage device, such ashard disk 810, may include non-removable storage medium. Other mass storage devices may include memory cards, memory sticks, tape storage devices, and the like. - Any number of computer programs, files, data structures, and the like may be stored in
mass storage 810,other storage devices - Output components or devices, such as
display device 802, may be coupled tocomputing device 801, typically via an interface such as adisplay adapter 811.Output device 802 may be a liquid crystal display (“LCD”). Other example output devices may include printers, audio outputs, voice outputs, cathode ray tube (“CRT”) displays, tactile devices or other sensory output mechanisms, or the like. Output devices may enablecomputing device 801 to interact with human operators or other machines, systems, computing environments, or the like. A user may interface withcomputing environment 800 via any number of different I/O devices 803 such as a touch pad, buttons, keyboard, mouse, joystick, game pad, data port, and the like. These and other I/O devices may be coupled toprocessor 807 via I/O interfaces 812 which may be coupled tosystem bus 808, and/or may be coupled by other interfaces and bus structures, such as a parallel port, game port, universal serial bus (“USB”), fire wire, infrared (“IR”) port, and the like. -
Computing device 801 may operate in a networked environment via communications connections to one or more remote computing devices through one or more cellular networks, wireless networks, local area networks (“LAN”), wide area networks (“WAN”), storage area networks (“SAN”), the Internet, radio links, optical links and the like.Computing device 801 may be coupled to a network vianetwork adapter 813 or the like, or, alternatively, via a modem, digital subscriber line (“DSL”) link, integrated services digital network (“ISDN”) link, Internet link, wireless link, or the like. -
Communications connection 814, such as a network connection, typically provides a coupling to communications media, such as a network. Communications media typically provide computer-readable and computer-executable instructions, data structures, files, program modules and other data using a modulated data signal, such as a carrier wave or other transport mechanism. The term “modulated data signal” typically means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communications media may include wired media, such as a wired network or direct-wired connection or the like, and wireless media, such as acoustic, radio frequency, infrared, or other wireless communications mechanisms. -
Power source 890, such as a battery or a power supply, typically provides power for portions or all ofcomputing environment 800. In the case of thecomputing environment 800 being a mobile device or portable device or the like,power source 890 may be a battery. Alternatively, in thecase computing environment 800 is a desktop computer or server or the like,power source 890 may be a power supply designed to connect to an alternating current (“AC”) source, such as via a wall outlet. - Some mobile devices may not include many of the components described in connection with
FIG. 8 . For example, an electronic badge may be comprised of a coil of wire along with asimple processing unit 807 or the like, the coil configured to act aspower source 890 when in proximity to a card reader device or the like. Such a coil may also be configure to act as an antenna coupled to theprocessing unit 807 or the like, the coil antenna capable of providing a form of communication between the electronic badge and the card reader device. Such communication may not involve networking, but may alternatively be general or special purpose communications via telemetry, point-to-point, RF, IR, audio, or other means. An electronic card may not includedisplay 802, I/O device 803, or many of the other components described in connection withFIG. 8 . Other mobile devices that may not include many of the components described in connection withFIG. 8 , by way of example and not limitation, include electronic bracelets, electronic tags, implantable devices, and the like. - Those skilled in the art will realize that storage devices utilized to provide computer-readable and computer-executable instructions and data can be distributed over a network. For example, a remote computer or storage device may store computer-readable and computer-executable instructions in the form of software applications and data. A local computer may access the remote computer or storage device via the network and download part or all of a software application or data and may execute any computer-executable instructions. Alternatively, the local computer may download pieces of the software or data as needed, or distributively process the software by executing some of the instructions at the local computer and some at remote computers and/or devices.
- Those skilled in the art will also realize that, by utilizing conventional techniques, all or portions of the software's computer-executable instructions may be carried out by a dedicated electronic circuit such as a digital signal processor (“DSP”), programmable logic array (“PLA”), discrete circuits, and the like. The term “electronic apparatus” may include computing devices or consumer electronic devices comprising any software, firmware or the like, or electronic devices or circuits comprising no software, firmware or the like.
- The term “firmware” typically refers to executable instructions, code, data, applications, programs, or the like maintained in an electronic device such as a ROM. The term “software” generally refers to executable instructions, code, data, applications, programs, or the like maintained in or on any form of computer-readable media. The term “computer-readable media” typically refers to system memory, storage devices and their associated media, and the like.
- In view of the many possible embodiments to which the principles of the present invention and the forgoing examples may be applied, it should be recognized that the examples described herein are meant to be illustrative only and should not be taken as limiting the scope of the present invention. Therefore, the invention as described herein contemplates all such embodiments as may come within the scope of the following claims and any equivalents thereto.
Claims (20)
1. A method performed on a computing device, the method comprising presenting an image in a context bar that is located outside of a field of view, where the image is a first representation of an object, and where the image is distinct from a second representation of the object presented in the field of view.
2. The method of claim 1 where the presenting the image in the context bar comprises presenting a plurality of images in the context bar, where each of the presented plurality of images is a representation of a corresponding one of a plurality of objects, and where the each of the presented plurality of images is distinct from a different representation of the corresponding one of a plurality of objects presented in the field of view.
3. The method of claim 2 where the presented plurality of images changes according to the corresponding plurality of objects presented in the field of view changing.
4. The method of claim 2 further comprising locating an object of interest in the field of view in response to selecting the corresponding image presented in the context bar.
5. The method of claim 4 where the locating comprises highlighting the object of interest in the field of view.
6. The method of claim 5 where the highlighting comprises presenting additional information about the highlighted object.
7. The method of claim 5 further comprising un-highlighting the object of interest in the field of view in response to de-selecting the corresponding image presented in the context bar.
8. A system comprising a computing device and at least one program module together configured for performing actions comprising presenting an image in a context bar that is located outside of a field of view, where the image is a first representation of an object, and where the image is distinct from a second representation of the object presented in the field of view.
9. The system of claim 8 where the presenting the image in the context bar comprises presenting a plurality of images in the context bar, where each of the presented plurality of images is a representation of a corresponding one of a plurality of objects, and where the each of the presented plurality of images is distinct from a different representation of the corresponding one of a plurality of objects presented in the field of view.
10. The system of claim 9 where the presented plurality of images changes according to the corresponding plurality of objects presented in the field of view changing.
11. The system of claim 9 , the actions further comprising locating an object of interest in the field of view in response to selecting the corresponding image presented in the context bar.
12. The system of claim 11 where the locating comprises highlighting the object of interest in the field of view.
13. The system of claim 12 where the highlighting comprises presenting additional information about the highlighted object.
14. The system of claim 12 , the actions further comprising un-highlighting the object of interest in the field of view in response to de-selecting the corresponding image presented in the context bar.
15. At least one computer storage device storing computer-executable instructions that, when executed by a computing device, cause the computing device to perform actions comprising presenting an image in a context bar that is located outside of a field of view, where the image is a first representation of an object, and where the image is distinct from a second representation of the object presented in the field of view.
16. The at least one computer storage device of claim 15 where the presenting the image in the context bar comprises presenting a plurality of images in the context bar, where each of the presented plurality of images is a representation of a corresponding one of a plurality of objects, and where the each of the presented plurality of images is distinct from a different representation of the corresponding one of a plurality of objects presented in the field of view.
17. The at least one computer storage device of claim 16 where the presented plurality of images changes according to the corresponding plurality of objects presented in the field of view changing.
18. The at least one computer storage device of claim 16 , the actions further comprising locating an object of interest in the field of view in response to selecting the corresponding image presented in the context bar.
19. The at least one computer storage device of claim 18 where the locating comprises highlighting the object of interest in the field of view, and where the actions further comprise un-highlighting the object of interest in the field of view in response to de-selecting the corresponding image presented in the context bar.
20. The at least one computer storage device of claim 19 where the highlighting comprises presenting additional information about the highlighted object.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/066,469 US20140059477A1 (en) | 2007-11-16 | 2013-10-29 | Localized thumbnail preview of related content during spatial browsing |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/941,098 US8584044B2 (en) | 2007-11-16 | 2007-11-16 | Localized thumbnail preview of related content during spatial browsing |
US14/066,469 US20140059477A1 (en) | 2007-11-16 | 2013-10-29 | Localized thumbnail preview of related content during spatial browsing |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/941,098 Continuation US8584044B2 (en) | 2007-11-16 | 2007-11-16 | Localized thumbnail preview of related content during spatial browsing |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140059477A1 true US20140059477A1 (en) | 2014-02-27 |
Family
ID=40643280
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/941,098 Expired - Fee Related US8584044B2 (en) | 2007-11-16 | 2007-11-16 | Localized thumbnail preview of related content during spatial browsing |
US14/066,469 Abandoned US20140059477A1 (en) | 2007-11-16 | 2013-10-29 | Localized thumbnail preview of related content during spatial browsing |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/941,098 Expired - Fee Related US8584044B2 (en) | 2007-11-16 | 2007-11-16 | Localized thumbnail preview of related content during spatial browsing |
Country Status (1)
Country | Link |
---|---|
US (2) | US8584044B2 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160189430A1 (en) * | 2013-08-16 | 2016-06-30 | Audi Ag | Method for operating electronic data glasses, and electronic data glasses |
US10564820B1 (en) * | 2014-08-08 | 2020-02-18 | Amazon Technologies, Inc. | Active content in digital media within a media universe |
US10621228B2 (en) | 2011-06-09 | 2020-04-14 | Ncm Ip Holdings, Llc | Method and apparatus for managing digital files |
CN111399735A (en) * | 2020-04-16 | 2020-07-10 | Oppo广东移动通信有限公司 | Screen capturing method, screen capturing device, electronic equipment and storage medium |
US11209968B2 (en) | 2019-01-07 | 2021-12-28 | MemoryWeb, LLC | Systems and methods for analyzing and organizing digital photos and videos |
Families Citing this family (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CA2601154C (en) | 2007-07-07 | 2016-09-13 | Mathieu Audet | Method and system for distinguising elements of information along a plurality of axes on a basis of a commonality |
US8601392B2 (en) | 2007-08-22 | 2013-12-03 | 9224-5489 Quebec Inc. | Timeline for presenting information |
US20090132967A1 (en) * | 2007-11-16 | 2009-05-21 | Microsoft Corporation | Linked-media narrative learning system |
TW200923758A (en) * | 2007-11-27 | 2009-06-01 | Wistron Corp | A key-in method and a content display method of an electronic device, and the application thereof |
US9165302B2 (en) * | 2008-09-29 | 2015-10-20 | Apple Inc. | System and method for scaling up an image of an article displayed on a sales promotion web page |
US9064023B2 (en) * | 2008-12-29 | 2015-06-23 | Avaya Inc. | Providing web content in the context of a virtual environment |
US8707200B2 (en) * | 2009-06-25 | 2014-04-22 | Autodesk, Inc. | Object browser with proximity sorting |
US9430118B2 (en) * | 2011-04-19 | 2016-08-30 | International Business Machines Corporation | Spatially-oriented traversal animations for network address transitions |
US10353566B2 (en) * | 2011-09-09 | 2019-07-16 | Microsoft Technology Licensing, Llc | Semantic zoom animations |
US10289657B2 (en) | 2011-09-25 | 2019-05-14 | 9224-5489 Quebec Inc. | Method of retrieving information elements on an undisplayed portion of an axis of information elements |
US9354763B2 (en) * | 2011-09-26 | 2016-05-31 | The University Of North Carolina At Charlotte | Multi-modal collaborative web-based video annotation system |
US10079039B2 (en) * | 2011-09-26 | 2018-09-18 | The University Of North Carolina At Charlotte | Multi-modal collaborative web-based video annotation system |
US9519693B2 (en) | 2012-06-11 | 2016-12-13 | 9224-5489 Quebec Inc. | Method and apparatus for displaying data element axes |
US20130346888A1 (en) * | 2012-06-22 | 2013-12-26 | Microsoft Corporation | Exposing user interface elements on search engine homepages |
US20140046923A1 (en) | 2012-08-10 | 2014-02-13 | Microsoft Corporation | Generating queries based upon data points in a spreadsheet application |
JP2014112302A (en) * | 2012-12-05 | 2014-06-19 | Ricoh Co Ltd | Prescribed area management system, communication method, and program |
US10262462B2 (en) | 2014-04-18 | 2019-04-16 | Magic Leap, Inc. | Systems and methods for augmented and virtual reality |
US10671266B2 (en) | 2017-06-05 | 2020-06-02 | 9224-5489 Quebec Inc. | Method and apparatus of aligning information element axes |
US10721510B2 (en) | 2018-05-17 | 2020-07-21 | At&T Intellectual Property I, L.P. | Directing user focus in 360 video consumption |
US10482653B1 (en) | 2018-05-22 | 2019-11-19 | At&T Intellectual Property I, L.P. | System for active-focus prediction in 360 video |
US10827225B2 (en) | 2018-06-01 | 2020-11-03 | AT&T Intellectual Propety I, L.P. | Navigation for 360-degree video streaming |
Citations (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4574395A (en) * | 1983-01-14 | 1986-03-04 | Fuji Xerox Co., Ltd. | Picture image filing apparatus |
US5821945A (en) * | 1995-02-03 | 1998-10-13 | The Trustees Of Princeton University | Method and apparatus for video browsing based on content and structure |
US6026389A (en) * | 1996-08-23 | 2000-02-15 | Kokusai, Denshin, Denwa, Kabushiki Kaisha | Video query and editing system |
US6208344B1 (en) * | 1997-07-31 | 2001-03-27 | Ncr Corporation | System and process for manipulating and viewing hierarchical iconic containers |
US20010009420A1 (en) * | 2000-01-26 | 2001-07-26 | Toru Kamiwada | Display interface method and apparatus |
US20010030667A1 (en) * | 2000-04-10 | 2001-10-18 | Kelts Brett R. | Interactive display interface for information objects |
US6335745B1 (en) * | 1999-02-24 | 2002-01-01 | International Business Machines Corporation | Method and system for invoking a function of a graphical object in a graphical user interface |
US20020011990A1 (en) * | 2000-04-14 | 2002-01-31 | Majid Anwar | User interface systems and methods for manipulating and viewing digital documents |
US20020113816A1 (en) * | 1998-12-09 | 2002-08-22 | Frederick H. Mitchell | Method and apparatus providing a graphical user interface for representing and navigating hierarchical networks |
US20040225635A1 (en) * | 2003-05-09 | 2004-11-11 | Microsoft Corporation | Browsing user interface for a geo-coded media database |
US20050138570A1 (en) * | 2003-12-22 | 2005-06-23 | Palo Alto Research Center, Incorporated | Methods and systems for supporting presentation tools using zoomable user interface |
US20050192924A1 (en) * | 2004-02-17 | 2005-09-01 | Microsoft Corporation | Rapid visual sorting of digital files and data |
US20050195154A1 (en) * | 2004-03-02 | 2005-09-08 | Robbins Daniel C. | Advanced navigation techniques for portable devices |
US20050207672A1 (en) * | 2000-10-06 | 2005-09-22 | Bernardo Enrico D | System and method for creating, storing, and utilizing composite images of a geographic location |
US6956590B1 (en) * | 2001-02-28 | 2005-10-18 | Navteq North America, Llc | Method of providing visual continuity when panning and zooming with a map display |
US20060136406A1 (en) * | 2004-12-17 | 2006-06-22 | Erika Reponen | Spatial search and selection feature |
US20060174209A1 (en) * | 1999-07-22 | 2006-08-03 | Barros Barbara L | Graphic-information flow method and system for visually analyzing patterns and relationships |
US20060224997A1 (en) * | 2005-03-31 | 2006-10-05 | Microsoft Corporation | Graphical web browser history toolbar |
US20060241860A1 (en) * | 2005-04-21 | 2006-10-26 | Microsoft Corporation | Virtual earth mapping |
US7143362B2 (en) * | 2001-12-28 | 2006-11-28 | International Business Machines Corporation | System and method for visualizing and navigating content in a graphical user interface |
US20060271280A1 (en) * | 2005-05-27 | 2006-11-30 | O'clair Brian | Using boundaries associated with a map view for business location searching |
US20060279630A1 (en) * | 2004-07-28 | 2006-12-14 | Manoj Aggarwal | Method and apparatus for total situational awareness and monitoring |
US20070064018A1 (en) * | 2005-06-24 | 2007-03-22 | Idelix Software Inc. | Detail-in-context lenses for online maps |
US20070110338A1 (en) * | 2005-11-17 | 2007-05-17 | Microsoft Corporation | Navigating images using image based geometric alignment and object based controls |
US7251790B1 (en) * | 2002-01-23 | 2007-07-31 | Microsoft Corporation | Media browsing system displaying thumbnails images based on preferences of plurality of users and placing the thumbnails images at a scene change |
US20080069480A1 (en) * | 2006-09-14 | 2008-03-20 | Parham Aarabi | Method, system and computer program for interactive spatial link-based image searching, sorting and/or displaying |
US7509581B1 (en) * | 1999-08-26 | 2009-03-24 | Lg Electronics, Inc. | Video browser based on character relation |
US20090089701A1 (en) * | 2007-09-27 | 2009-04-02 | Rockwell Automation Technologies, Inc. | Distance-wise presentation of industrial automation data as a function of relevance to user |
US20090128565A1 (en) * | 2007-11-16 | 2009-05-21 | Microsoft Corporation | Spatial exploration field of view preview mechanism |
US20100034451A1 (en) * | 2007-06-21 | 2010-02-11 | Hughes Ronald J | Systems and Methods for Improving Directed People Screening |
US7966301B2 (en) * | 2003-05-09 | 2011-06-21 | Planeteye Company Ulc | System and method for employing a grid index for location and precision encoding |
US8368722B1 (en) * | 2006-04-18 | 2013-02-05 | Google Inc. | Cartographic display of content through dynamic, interactive user interface elements |
US8819569B2 (en) * | 2005-02-18 | 2014-08-26 | Zumobi, Inc | Single-handed approach for navigation of application tiles using panning and zooming |
Family Cites Families (86)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4763280A (en) | 1985-04-29 | 1988-08-09 | Evans & Sutherland Computer Corp. | Curvilinear dynamic image generation system |
US5617332A (en) | 1988-08-10 | 1997-04-01 | Fressola; Alfred A. | Method and system for producing stereographic images of celestial objects |
DE4293175T1 (en) | 1991-10-08 | 1993-10-07 | Citizen Watch Co Ltd | Clock with constellation display |
US5675746A (en) | 1992-09-30 | 1997-10-07 | Marshall; Paul S. | Virtual reality generator for use with financial information |
US5396583A (en) | 1992-10-13 | 1995-03-07 | Apple Computer, Inc. | Cylindrical to planar image mapping using scanline coherence |
US5684937A (en) | 1992-12-14 | 1997-11-04 | Oxaal; Ford | Method and apparatus for performing perspective transformation on visible stimuli |
US5555354A (en) | 1993-03-23 | 1996-09-10 | Silicon Graphics Inc. | Method and apparatus for navigation within three-dimensional information landscape |
US5473746A (en) | 1993-04-01 | 1995-12-05 | Loral Federal Systems, Company | Interactive graphics computer system for planning star-sensor-based satellite attitude maneuvers |
US6321158B1 (en) | 1994-06-24 | 2001-11-20 | Delorme Publishing Company | Integrated routing/mapping information |
AUPM701394A0 (en) | 1994-07-22 | 1994-08-18 | Monash University | A graphical display system |
FR2725291B1 (en) | 1994-09-30 | 1996-11-29 | Thomson Broadband Systems | METHOD FOR GENERATING SYNTHESIS IMAGES BY SPHERICAL BUFFER |
USRE37356E1 (en) | 1994-10-07 | 2001-09-04 | Vista Medical Technologies, Inc. | Endoscope with position display for zoom lens unit and imaging device |
US5850352A (en) | 1995-03-31 | 1998-12-15 | The Regents Of The University Of California | Immersive video, including video hypermosaicing to generate from multiple video views of a scene a three-dimensional video mosaic from which diverse virtual video scene images are synthesized, including panoramic, scene interactive and stereoscopic images |
CN100501768C (en) | 1995-05-19 | 2009-06-17 | 世嘉股份有限公司 | Image processing device, image processing method, and equipment |
JPH10504127A (en) | 1995-06-09 | 1998-04-14 | フィリップス エレクトロニクス ネムローゼ フェンノートシャップ | Method for enabling a user to derive a particular information item from a set of information items, and a system for performing such a method |
US6020885A (en) | 1995-07-11 | 2000-02-01 | Sony Corporation | Three-dimensional virtual reality space sharing method and system using local and global object identification codes |
DE19549306A1 (en) | 1995-12-22 | 1997-07-03 | Art & Com Medientechnologie Un | Method and device for the visual representation of spatial data |
US5987363A (en) | 1996-03-26 | 1999-11-16 | California Institute Of Technology | Three-dimensional representation of a spacecraft's trajectory |
JPH1079045A (en) | 1996-07-23 | 1998-03-24 | Internatl Business Mach Corp <Ibm> | Rendering method, device therefor, luminance value smoothing method and device therefor |
US6057856A (en) | 1996-09-30 | 2000-05-02 | Sony Corporation | 3D virtual reality multi-user interaction with superimposed positional information display for each user |
US6545687B2 (en) | 1997-01-09 | 2003-04-08 | Canon Kabushiki Kaisha | Thumbnail manipulation using fast and aspect ratio zooming, compressing and scaling |
US6776618B1 (en) | 1997-03-12 | 2004-08-17 | D'zmura David Andrew | Method of determining zodiac signs |
US6154213A (en) | 1997-05-30 | 2000-11-28 | Rennison; Earl F. | Immersive movement-based interaction with large complex information structures |
US6094196A (en) | 1997-07-03 | 2000-07-25 | International Business Machines Corporation | Interaction spheres of three-dimensional objects in three-dimensional workspace displays |
US5864337A (en) | 1997-07-22 | 1999-01-26 | Microsoft Corporation | Mehtod for automatically associating multimedia features with map views displayed by a computer-implemented atlas program |
US6121969A (en) | 1997-07-29 | 2000-09-19 | The Regents Of The University Of California | Visual navigation in perceptual databases |
US6301586B1 (en) | 1997-10-06 | 2001-10-09 | Canon Kabushiki Kaisha | System for managing multimedia objects |
JP3928229B2 (en) | 1997-11-28 | 2007-06-13 | ソニー株式会社 | Display control apparatus, display control method, and recording medium |
JP2000076271A (en) | 1998-08-31 | 2000-03-14 | Sony Corp | Information processor, information processing method and distribution medium |
US6388688B1 (en) | 1999-04-06 | 2002-05-14 | Vergics Corporation | Graph-based visual navigation through spatial environments |
US6346938B1 (en) | 1999-04-27 | 2002-02-12 | Harris Corporation | Computer-resident mechanism for manipulating, navigating through and mensurating displayed image of three-dimensional geometric model |
EP1087323A1 (en) | 1999-09-24 | 2001-03-28 | Nokia Corporation | A wireless system for interacting with a virtual space |
JP3306397B2 (en) | 1999-11-25 | 2002-07-24 | 株式会社ソニー・コンピュータエンタテインメント | Entertainment device, image generation method, and storage medium |
US6687387B1 (en) | 1999-12-27 | 2004-02-03 | Internet Pictures Corporation | Velocity-dependent dewarping of images |
US20020075311A1 (en) | 2000-02-14 | 2002-06-20 | Julian Orbanes | Method for viewing information in virtual space |
EP1256046A2 (en) | 2000-02-14 | 2002-11-13 | Geophoenix, Inc. | Methods and apparatus for viewing information in virtual space |
US6525732B1 (en) | 2000-02-17 | 2003-02-25 | Wisconsin Alumni Research Foundation | Network-based viewing of images of three-dimensional objects |
US20020054134A1 (en) | 2000-04-10 | 2002-05-09 | Kelts Brett R. | Method and apparatus for providing streaming media in a communication network |
US7375728B2 (en) | 2001-10-01 | 2008-05-20 | University Of Minnesota | Virtual mirror |
US20020029226A1 (en) | 2000-09-05 | 2002-03-07 | Gang Li | Method for combining data with maps |
US6754400B2 (en) | 2001-02-06 | 2004-06-22 | Richard Wilson, Jr. | System and method for creation, processing and visualization of omni-directional images |
US7213214B2 (en) | 2001-06-12 | 2007-05-01 | Idelix Software Inc. | Graphical user interface with zoom for detail-in-context presentations |
US6795972B2 (en) | 2001-06-29 | 2004-09-21 | Scientific-Atlanta, Inc. | Subscriber television system user interface with a virtual reality media space |
US7069506B2 (en) | 2001-08-08 | 2006-06-27 | Xerox Corporation | Methods and systems for generating enhanced thumbnails |
US7337396B2 (en) | 2001-08-08 | 2008-02-26 | Xerox Corporation | Methods and systems for transitioning between thumbnails and documents based upon thumbnail appearance |
US7096428B2 (en) | 2001-09-28 | 2006-08-22 | Fuji Xerox Co., Ltd. | Systems and methods for providing a spatially indexed panoramic video |
KR100468967B1 (en) | 2001-12-28 | 2005-01-29 | 엘지전자 주식회사 | Thumbnail image generating system |
US7739601B1 (en) * | 2002-01-23 | 2010-06-15 | Microsoft Corporation | Media authoring and presentation |
EP1333376A1 (en) | 2002-02-05 | 2003-08-06 | Fulvio Dominici | Encoding method for efficient storage, transmission and sharing of multidimensional virtual worlds |
US7228507B2 (en) | 2002-02-21 | 2007-06-05 | Xerox Corporation | Methods and systems for navigating a workspace |
US20030210281A1 (en) | 2002-05-07 | 2003-11-13 | Troy Ellis | Magnifying a thumbnail image of a document |
US20030222901A1 (en) | 2002-05-28 | 2003-12-04 | Todd Houck | uPrime uClient environment |
US7292243B1 (en) * | 2002-07-02 | 2007-11-06 | James Burke | Layered and vectored graphical user interface to a knowledge and relationship rich data source |
JP4300767B2 (en) | 2002-08-05 | 2009-07-22 | ソニー株式会社 | Guide system, content server, portable device, information processing method, information processing program, and storage medium |
US7117453B2 (en) | 2003-01-21 | 2006-10-03 | Microsoft Corporation | Media frame object visualization system |
JP2004348241A (en) | 2003-05-20 | 2004-12-09 | Hitachi Ltd | Information providing method, server, and program |
WO2004107012A1 (en) | 2003-05-30 | 2004-12-09 | Vixen Co., Ltd. | Automatic introduction system of celestial body |
US7467356B2 (en) | 2003-07-25 | 2008-12-16 | Three-B International Limited | Graphical user interface for 3d virtual display browser using virtual display windows |
US7667700B1 (en) | 2004-03-05 | 2010-02-23 | Hrl Laboratories, Llc | System and method for navigating operating in a virtual environment |
US7441207B2 (en) | 2004-03-18 | 2008-10-21 | Microsoft Corporation | Method and system for improved viewing and navigation of content |
CA2820249C (en) | 2004-03-23 | 2016-07-19 | Google Inc. | A digital mapping system |
GB2414309B (en) | 2004-05-18 | 2009-02-25 | Simon Richard Daniel | Spherical display and control device |
US20080049012A1 (en) | 2004-06-13 | 2008-02-28 | Ittai Bar-Joseph | 3D Line-of-Sight (Los) Visualization in User Interactive 3D Virtual Reality Environments |
KR100521808B1 (en) | 2004-10-14 | 2005-10-14 | (주) 아모센스 | A constellation searching apparatus, a constellation searching method, and storage medium recorded the constellation searching program |
JP2006195882A (en) | 2005-01-17 | 2006-07-27 | Namco Bandai Games Inc | Program, information storage medium and image generation system |
JP4533791B2 (en) | 2005-04-19 | 2010-09-01 | 株式会社日立製作所 | Information browsing device |
US7725820B2 (en) | 2005-05-16 | 2010-05-25 | Planview, Inc. | Method of generating a display for a directed graph and a system for use with the method |
US20070011617A1 (en) | 2005-07-06 | 2007-01-11 | Mitsunori Akagawa | Three-dimensional graphical user interface |
WO2007033354A2 (en) | 2005-09-13 | 2007-03-22 | Spacetime3D, Inc. | System and method for providing three-dimensional graphical user interface |
GB2431804B (en) | 2005-10-31 | 2011-04-13 | Hewlett Packard Development Co | Image capture device and method of capturing an image |
US7600198B2 (en) | 2005-11-23 | 2009-10-06 | Bluebeam Software, Inc. | Method of tracking data objects using related thumbnails in a palette window |
US8396657B2 (en) | 2005-12-22 | 2013-03-12 | Hewlett-Packard Development Company, L.P. | Techniques to improve location accuracy for a map |
JP2007233996A (en) | 2006-02-06 | 2007-09-13 | Olympus Imaging Corp | Image compositing apparatus, image compositing method, image compositing program and recording medium |
US20080086696A1 (en) | 2006-03-03 | 2008-04-10 | Cadcorporation.Com Inc. | System and Method for Using Virtual Environments |
US7616217B2 (en) | 2006-04-26 | 2009-11-10 | Google Inc. | Dynamic exploration of electronic maps |
JP4804256B2 (en) | 2006-07-27 | 2011-11-02 | キヤノン株式会社 | Information processing method |
US7889212B2 (en) | 2006-09-07 | 2011-02-15 | Apple Inc. | Magnifying visual information using a center-based loupe |
US20090300528A1 (en) | 2006-09-29 | 2009-12-03 | Stambaugh Thomas M | Browser event tracking for distributed web-based processing, spatial organization and display of information |
US20080109761A1 (en) | 2006-09-29 | 2008-05-08 | Stambaugh Thomas M | Spatial organization and display of travel and entertainment information |
US20080231643A1 (en) * | 2007-03-21 | 2008-09-25 | Nick Fletcher | Method and apparatus for controlling the size or opacity of map elements rendered in an interactive map view |
US20080263460A1 (en) | 2007-04-20 | 2008-10-23 | Utbk, Inc. | Methods and Systems to Connect People for Virtual Meeting in Virtual Reality |
US20090132967A1 (en) | 2007-11-16 | 2009-05-21 | Microsoft Corporation | Linked-media narrative learning system |
US20090307618A1 (en) | 2008-06-05 | 2009-12-10 | Microsoft Corporation | Annotate at multiple levels |
US8194102B2 (en) | 2008-10-06 | 2012-06-05 | Microsoft Corporation | Rendering annotations for images |
US20100306696A1 (en) | 2008-11-26 | 2010-12-02 | Lila Aps (Ahead.) | Dynamic network browser |
US20120042282A1 (en) | 2010-08-12 | 2012-02-16 | Microsoft Corporation | Presenting Suggested Items for Use in Navigating within a Virtual Space |
-
2007
- 2007-11-16 US US11/941,098 patent/US8584044B2/en not_active Expired - Fee Related
-
2013
- 2013-10-29 US US14/066,469 patent/US20140059477A1/en not_active Abandoned
Patent Citations (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4574395A (en) * | 1983-01-14 | 1986-03-04 | Fuji Xerox Co., Ltd. | Picture image filing apparatus |
US5821945A (en) * | 1995-02-03 | 1998-10-13 | The Trustees Of Princeton University | Method and apparatus for video browsing based on content and structure |
US6026389A (en) * | 1996-08-23 | 2000-02-15 | Kokusai, Denshin, Denwa, Kabushiki Kaisha | Video query and editing system |
US6208344B1 (en) * | 1997-07-31 | 2001-03-27 | Ncr Corporation | System and process for manipulating and viewing hierarchical iconic containers |
US20020113816A1 (en) * | 1998-12-09 | 2002-08-22 | Frederick H. Mitchell | Method and apparatus providing a graphical user interface for representing and navigating hierarchical networks |
US6335745B1 (en) * | 1999-02-24 | 2002-01-01 | International Business Machines Corporation | Method and system for invoking a function of a graphical object in a graphical user interface |
US20060174209A1 (en) * | 1999-07-22 | 2006-08-03 | Barros Barbara L | Graphic-information flow method and system for visually analyzing patterns and relationships |
US7509581B1 (en) * | 1999-08-26 | 2009-03-24 | Lg Electronics, Inc. | Video browser based on character relation |
US20010009420A1 (en) * | 2000-01-26 | 2001-07-26 | Toru Kamiwada | Display interface method and apparatus |
US20010030667A1 (en) * | 2000-04-10 | 2001-10-18 | Kelts Brett R. | Interactive display interface for information objects |
US20020011990A1 (en) * | 2000-04-14 | 2002-01-31 | Majid Anwar | User interface systems and methods for manipulating and viewing digital documents |
US20050207672A1 (en) * | 2000-10-06 | 2005-09-22 | Bernardo Enrico D | System and method for creating, storing, and utilizing composite images of a geographic location |
US6956590B1 (en) * | 2001-02-28 | 2005-10-18 | Navteq North America, Llc | Method of providing visual continuity when panning and zooming with a map display |
US7143362B2 (en) * | 2001-12-28 | 2006-11-28 | International Business Machines Corporation | System and method for visualizing and navigating content in a graphical user interface |
US7251790B1 (en) * | 2002-01-23 | 2007-07-31 | Microsoft Corporation | Media browsing system displaying thumbnails images based on preferences of plurality of users and placing the thumbnails images at a scene change |
US7966301B2 (en) * | 2003-05-09 | 2011-06-21 | Planeteye Company Ulc | System and method for employing a grid index for location and precision encoding |
US20040225635A1 (en) * | 2003-05-09 | 2004-11-11 | Microsoft Corporation | Browsing user interface for a geo-coded media database |
US7475060B2 (en) * | 2003-05-09 | 2009-01-06 | Planeteye Company Ulc | Browsing user interface for a geo-coded media database |
US20050138570A1 (en) * | 2003-12-22 | 2005-06-23 | Palo Alto Research Center, Incorporated | Methods and systems for supporting presentation tools using zoomable user interface |
US20050192924A1 (en) * | 2004-02-17 | 2005-09-01 | Microsoft Corporation | Rapid visual sorting of digital files and data |
US20050195154A1 (en) * | 2004-03-02 | 2005-09-08 | Robbins Daniel C. | Advanced navigation techniques for portable devices |
US20060279630A1 (en) * | 2004-07-28 | 2006-12-14 | Manoj Aggarwal | Method and apparatus for total situational awareness and monitoring |
US20060136406A1 (en) * | 2004-12-17 | 2006-06-22 | Erika Reponen | Spatial search and selection feature |
US8819569B2 (en) * | 2005-02-18 | 2014-08-26 | Zumobi, Inc | Single-handed approach for navigation of application tiles using panning and zooming |
US20060224997A1 (en) * | 2005-03-31 | 2006-10-05 | Microsoft Corporation | Graphical web browser history toolbar |
US20060241860A1 (en) * | 2005-04-21 | 2006-10-26 | Microsoft Corporation | Virtual earth mapping |
US20060271280A1 (en) * | 2005-05-27 | 2006-11-30 | O'clair Brian | Using boundaries associated with a map view for business location searching |
US20070064018A1 (en) * | 2005-06-24 | 2007-03-22 | Idelix Software Inc. | Detail-in-context lenses for online maps |
US20070110338A1 (en) * | 2005-11-17 | 2007-05-17 | Microsoft Corporation | Navigating images using image based geometric alignment and object based controls |
US8368722B1 (en) * | 2006-04-18 | 2013-02-05 | Google Inc. | Cartographic display of content through dynamic, interactive user interface elements |
US20080069480A1 (en) * | 2006-09-14 | 2008-03-20 | Parham Aarabi | Method, system and computer program for interactive spatial link-based image searching, sorting and/or displaying |
US20100034451A1 (en) * | 2007-06-21 | 2010-02-11 | Hughes Ronald J | Systems and Methods for Improving Directed People Screening |
US20090089701A1 (en) * | 2007-09-27 | 2009-04-02 | Rockwell Automation Technologies, Inc. | Distance-wise presentation of industrial automation data as a function of relevance to user |
US20090128565A1 (en) * | 2007-11-16 | 2009-05-21 | Microsoft Corporation | Spatial exploration field of view preview mechanism |
Non-Patent Citations (8)
Title |
---|
Federov et al. Automatic Registration and Mosaicking System for Remotely Sensed Imagery. Image and Signal Processing for Remote Sensing VIII, Sebastiano B. Serpico, Editor. Proceedings of SPIE Vol. 4885 (2003) © 2003 SPIE · 0277/786X/03. 8 pages. * |
Rosenbaum et al. Grid-based Interaction for Effective Image Browsing on Mobile Devices. Multimedia on Mobile Devices, edited by Reiner Creutzburg, Jarmo H. Takala, Proc. of SPIE-IS&T Electronic Imaging,Vol. 5684 © 2005 SPIE and IS&T · 0277-786X/05. 11 pages. * |
SDSS SkyServer. Retrieved from [http://research.microsoft.com/en-us/um/people/gray/SDSS/default.htm] on 4 Nov 16. 3 pages. * |
Snavely et al. Photo Tourism: Exploring Photo Collections in 3D. © 2006 ACM 0730-0301/06/0700-0835. 12 pages * |
Strauss, Michael. User Interfaces [for SDSS EDR]. 30 Jan 2003. Retrieved from [http://classic.sdss.org/dr2/products/general/edr_html/node13.html] on [4 Nov 2016] * |
Szalay et al. The World-Wide Telescope. MSR-TR-2001-77. Science V. 293 pp. 2037-2040 14 Sept 2001 © ACM. 6 pages. * |
Toyama et al. Geographic Location Tags on Digital Images. MM’03, November 2-8, 2003, Berkeley, California, USA. Copyright 2003 ACM 1-58113-722-2/03/0011. 11 pages. * |
Yousef et al. Sunrise: Towards Location Based Clustering For Assisted Photo Management. ICMI ’07 Workshop on Tagging, Mining and Retrieval of Human-Related Activity Information, November 15, 2007, Nagoya, Japan. ACM 978-1-59593-870-1/07/11. 8 pages. * |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11599573B1 (en) | 2011-06-09 | 2023-03-07 | MemoryWeb, LLC | Method and apparatus for managing digital files |
US11481433B2 (en) | 2011-06-09 | 2022-10-25 | MemoryWeb, LLC | Method and apparatus for managing digital files |
US10621228B2 (en) | 2011-06-09 | 2020-04-14 | Ncm Ip Holdings, Llc | Method and apparatus for managing digital files |
US11899726B2 (en) | 2011-06-09 | 2024-02-13 | MemoryWeb, LLC | Method and apparatus for managing digital files |
US11017020B2 (en) | 2011-06-09 | 2021-05-25 | MemoryWeb, LLC | Method and apparatus for managing digital files |
US11163823B2 (en) | 2011-06-09 | 2021-11-02 | MemoryWeb, LLC | Method and apparatus for managing digital files |
US11768882B2 (en) | 2011-06-09 | 2023-09-26 | MemoryWeb, LLC | Method and apparatus for managing digital files |
US11636149B1 (en) | 2011-06-09 | 2023-04-25 | MemoryWeb, LLC | Method and apparatus for managing digital files |
US11170042B1 (en) | 2011-06-09 | 2021-11-09 | MemoryWeb, LLC | Method and apparatus for managing digital files |
US11636150B2 (en) | 2011-06-09 | 2023-04-25 | MemoryWeb, LLC | Method and apparatus for managing digital files |
US20160189430A1 (en) * | 2013-08-16 | 2016-06-30 | Audi Ag | Method for operating electronic data glasses, and electronic data glasses |
US10564820B1 (en) * | 2014-08-08 | 2020-02-18 | Amazon Technologies, Inc. | Active content in digital media within a media universe |
US11209968B2 (en) | 2019-01-07 | 2021-12-28 | MemoryWeb, LLC | Systems and methods for analyzing and organizing digital photos and videos |
US11954301B2 (en) | 2019-01-07 | 2024-04-09 | MemoryWeb. LLC | Systems and methods for analyzing and organizing digital photos and videos |
CN111399735A (en) * | 2020-04-16 | 2020-07-10 | Oppo广东移动通信有限公司 | Screen capturing method, screen capturing device, electronic equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
US8584044B2 (en) | 2013-11-12 |
US20090132952A1 (en) | 2009-05-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8584044B2 (en) | Localized thumbnail preview of related content during spatial browsing | |
US7493570B2 (en) | User interface options of a data lineage tool | |
US7475359B2 (en) | User interface for a tabbed pane | |
US8081186B2 (en) | Spatial exploration field of view preview mechanism | |
US10410606B2 (en) | Rendering graphical assets on electronic devices | |
CN107766106B (en) | The method and apparatus for generating configuration file | |
CN112558851A (en) | Object processing method, device, equipment and readable storage medium | |
CN113407075A (en) | Icon sorting method and device and electronic equipment | |
US6307562B1 (en) | Graphical interface with event horizon | |
CN112698771B (en) | Display control method, device, electronic equipment and storage medium | |
CN111694627B (en) | Desktop editing method and device | |
CN111796733B (en) | Image display method, image display device and electronic equipment | |
CN111610917B (en) | Display method, display device and electronic equipment | |
CN113779288A (en) | Photo storage method and device | |
CN111796736A (en) | Application sharing method and device and electronic equipment | |
US8949245B2 (en) | Enhanced views of tagged items in a gallery | |
CN113126863A (en) | Object selection implementation method and device, storage medium and electronic equipment | |
CN111752428A (en) | Icon arrangement method and device, electronic equipment and medium | |
US9032331B2 (en) | Visual widget search | |
CN117193919A (en) | Display method, display device, electronic equipment and readable storage medium | |
CN115686285A (en) | Page display method and device, electronic equipment and readable storage medium | |
CN116627291A (en) | Applet management method, device, electronic equipment and readable storage medium | |
CN117519884A (en) | Application remarking method and device, electronic equipment and storage medium | |
CN111859202A (en) | Information collection method and device and electronic equipment | |
CN113835815A (en) | Image previewing method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034544/0541 Effective date: 20141014 |
|
AS | Assignment |
Owner name: MICROSOFT CORPORATION, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:WONG, CURTIS GLENN;REEL/FRAME:041602/0217 Effective date: 20071112 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |