US20120320060A1 - System and method for presenting user generated digital information - Google Patents

System and method for presenting user generated digital information Download PDF

Info

Publication number
US20120320060A1
US20120320060A1 US13/599,884 US201213599884A US2012320060A1 US 20120320060 A1 US20120320060 A1 US 20120320060A1 US 201213599884 A US201213599884 A US 201213599884A US 2012320060 A1 US2012320060 A1 US 2012320060A1
Authority
US
United States
Prior art keywords
location
globs
glob
recited
gds
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/599,884
Inventor
Edward Lee Koch
Daniel Allan Hennage
Wesley Clay Collier
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Geopeg Inc
Original Assignee
Geopeg Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US11/645,001 external-priority patent/US7904483B2/en
Application filed by Geopeg Inc filed Critical Geopeg Inc
Priority to US13/599,884 priority Critical patent/US20120320060A1/en
Assigned to Geopeg, Inc. reassignment Geopeg, Inc. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: COLLIER, WESLEY CLAY, HENNAGE, DANIEL ALLAN, KOCH, EDWARD LEE
Publication of US20120320060A1 publication Critical patent/US20120320060A1/en
Priority to US14/539,263 priority patent/US20150142806A1/en
Priority to US15/175,374 priority patent/US10488860B1/en
Priority to US16/694,957 priority patent/US11415986B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/285Clustering or classification
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/38Electronic maps specially adapted for navigation; Updating thereof
    • G01C21/3804Creation or updating of map data
    • G01C21/3833Creation or updating of map data characterised by the source of data
    • G01C21/3856Data obtained from user input
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/38Electronic maps specially adapted for navigation; Updating thereof
    • G01C21/3863Structures of map data
    • G01C21/387Organisation of map data, e.g. version management or database structures
    • G01C21/3874Structures specially adapted for data searching and retrieval
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/44Browsing; Visualisation therefor
    • G06F16/444Spatial browsing, e.g. 2D maps, 3D or virtual spaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9537Spatial or temporal dependent retrieval, e.g. spatiotemporal queries
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99941Database schema or data structure
    • Y10S707/99942Manipulating data structure, e.g. compression, compaction, compilation
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99941Database schema or data structure
    • Y10S707/99943Generating database or data structure, e.g. via user interface
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99941Database schema or data structure
    • Y10S707/99944Object-oriented database structure
    • Y10S707/99945Object-oriented database structure processing
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99941Database schema or data structure
    • Y10S707/99948Application of database or data structure, e.g. distributed, multimedia, or image
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99951File or database maintenance
    • Y10S707/99952Coherency, e.g. same view to multiple users
    • Y10S707/99953Recoverability

Definitions

  • non-provisional patent application Ser. No. 13/014,600 is a continuation of non-provisional patent application Ser. No. 12/571,547 now U.S. Pat. No. 7,917,543 filed on Oct. 1, 2009, which is a continuation of non-provisional patent application Ser. No. 11/645,001 now U.S. Pat. No. 7,904,483 filed on Dec. 23, 2006, and which claims the benefit of provisional patent applications 60/774,931 filed on Feb. 21, 2006, 60/792,211 filed on Apr. 17, 2006, and 60/793,928 filed on Apr. 24, 2006, all of which are hereby incorporated by reference.
  • the invention relates to a system and method for geo-coding user generated content. More particularly, the invention relates to a system and method for geo-coding a plurality of user generated content.
  • Geocoding is the process of assigning geographic identifiers such as latitude, longitude, or Universal Transverse Mercator (UTM) to information.
  • the information having the geographic coordinates can be mapped and entered into a Geographic Information System (GIS).
  • GIS Geographic Information System
  • geocoded information is a geocoded photo.
  • the geocoded photo is an image that is associated with a geographical location.
  • a geocoded image can be associated to geographic coordinates such as latitude and longitude or a physical address.
  • every part of a picture can be tied to geographic location, but in the most typical application, only the position of the photograph is associated with the entire digital image.
  • a location aware device such as a GPS receiver
  • the location of the camera used to take a particular photo is recorded.
  • an external GPS device can be used to keep track of the camera location and the data can be synchronized using specialized software.
  • the process of geo-coding user generated content is relatively challenging and time consuming task.
  • the currently available location aware devices are specialized devices that do not integrate with other electronic devices that may be used to gather or capture user generated content.
  • a system and method for generating a virtual tour on a display device comprises providing at least one map.
  • the method further comprises providing a plurality of sequenced images, wherein each of the images is associated with at least one location by a geo-coding module configured to generate a geo-location object data sheet that associates sequential images with a corresponding location.
  • the sequenced images are organized based on the location of each of the sequenced images and displayed on the map.
  • the method is implemented by the system.
  • FIG. 1 shows illustrative examples of information converted to a geo-location object (“GLOB”).
  • GLOB geo-location object
  • FIG. 2 shows illustrative location descriptors.
  • FIG. 3 shows illustrative attributes of a GLOB data sheet.
  • FIG. 4 shows an illustrative system configured to query GLOB data sheets.
  • FIG. 5 shows an illustrative diagram of software modules that generate GLOB data sheets.
  • FIG. 6 shows an illustrative flowchart of a geocoding process.
  • FIG. 7 shows an illustrative example of a Chronicle of GLOBs.
  • FIG. 8 shows an image with more detail of a particular GLOB.
  • FIG. 9 shows a flowchart for generating an illustrative Chronicle or Narrative that integrates third party information.
  • FIG. 10 shows illustrative system architecture for generating a Chronicle or Narrative.
  • FIG. 11 shows the various components of the Chronicle or Narrative presentation systems.
  • FIG. 12 presents an illustrative system architecture with fat servers and thin clients.
  • FIG. 13 presents an illustrative system architecture with no remote servers.
  • FIG. 14 presents an illustrative system architecture with remote servers that generate skinny Chronicles or Narratives.
  • FIG. 15 presents an illustrative system architecture with remote servers and clients generating the Chronicle or Narratives.
  • FIG. 16 provides an illustrative depiction of a time stamped photographs.
  • FIG. 17 illustrates the explicit and implicit geocoding of photographs.
  • FIG. 18 further illustrates explicit and implicit geocoding of photographs.
  • FIG. 19 provides an illustrative example of implicit geocoding based on different path and distribution rules.
  • FIG. 20 provides an illustrative example of implicit geocoding of a user defined photograph sequence with no time stamps.
  • FIG. 21 provides an illustrative example of implicit geocoding of a user defined photograph sequence with no time stamps and additional relationship rules.
  • FIG. 22 provides an illustrative Virtual Tour.
  • FIG. 23 shows an illustrative display with detail for a selected Point of Interest.
  • FIG. 24 shows an illustrative display of a selected geo-located image.
  • FIG. 25 presents an illustrative system architecture for an illustrative virtual tour server.
  • FIG. 26 presents an illustrative system architecture for a more detailed virtual tour server architecture.
  • geocoding an object refers to associating location or geographic information, i.e. a geocode, with the particular object.
  • Geocodes are also referred to as geo-locations.
  • an object that has an associated geocode is referred to as a geo-located object, hereinafter “GLOB.”
  • the GLOB is a piece of information tagged with the geographic location of the content of the information.
  • FIG. 1 there is shown illustrative examples of information that may be converted to a GLOB. More particularly, FIG. 1 shows an illustrative example of a plurality of sources of information that can be turned into a GLOB such as video, audio, sensors/data, maps, text/document, location, i.e. GPS logs, mobile telemetry, and pictures. Additionally, the attributes associated with each GLOB are shown: location (where); time (when); content description (what); and source of data (who).
  • a description of a system that can be used to present one or more GLOBs is described. Additionally, a system to animate GLOBs that are indexed by time and that include geographic information is also described. Additionally, GLOB information may be presented in a synopsized form that allows the user to obtain detail about a particular object. Additionally, a plurality GLOBs can be generated from a plurality of different sources of information.
  • the systems and methods described herein may also be used to geocode multiple objects and to organize the GLOBs. For example, geocoding of multiple objects that are sequenced in relation to one another is described. The systems and methods take advantage of the objects belonging to the same group. Finally, the systems and methods can be used to generate virtual tours where the virtual tour is a visual and/or audio presentation of what one might experience when visiting a particular location.
  • a GLOB could be a digital photograph that is tagged with the time and location at which the photograph was taken.
  • the content of the photograph is the photographic image itself. There may also be tags which describe who took the photograph and the subject matter of the photograph.
  • illustrative information that can be converted to a GLOB includes, but is not limited to, maps, videos, recorded location logs (e.g. GPS logs), audio recordings, weather reports, news items, sensor measurements, automotive telemetry such as location, speed, camera images, and any other means for generating content.
  • the examples of information presented above are referred to as “generated content” or sometimes just “content” for short.
  • the content described can be grouped together into sets of related objects. For example, in the case of photographs, a user might decide to group together a collection of photographs that were taken around the same time frame as a particular trip or event.
  • GLOBs may be tagged with other state information such as altitude, heading, and azimuth. GLOBs may also be tagged with motion state parameters such as speed, acceleration, and bearing.
  • the location may have a further qualifier that specifies the accuracy and resolution of the location.
  • FIG. 2 there is shown illustrative location descriptors, which shows the topological forms that a location attribute of a GLOB can take.
  • the different forms can range from being a single point to some defined area.
  • a location area represents the bound within which the GLOB may exist.
  • the location was derived from a GPS fix then it will have a high degree of resolution, but if the location was derived from some annotation that simply says “Montana” then it will only have a resolution that corresponds to the state of Montana. Either of these situations may be modeled using one or the topological forms shown in FIG. 2 .
  • Information that is geocoded may be stored or associated with a particular digital file.
  • digital recording formats such as JPEG have facilities built into their specification that allow for the addition of location information.
  • Digital formats which do not may require some sort of auxiliary data file that can be stored with the content.
  • the first and most straightforward method is for the device that generates the content to geocode it when the content is generated. This requires that the device have access to location information such as an embedded GPS receiver or perhaps some sort of remote server that can locate the device and transmit its location information to the device. Examples of the latter approach include cell phones which have so called E-911 capabilities.
  • the device that generates the content is not able to geocode the content when it is generated then it is necessary to perform some sort of processing step after the content is generated in order to geocode it.
  • Most digital recording devices in use today such as digital cameras have a means to take the content that it generates with the time in which the content was generated.
  • the geocoding step can be automated if another collection of GLOBs exists that were collected at the same time and within the same area as the content that is not geocoded. It is then possible to geocode the content by time correlating with the collection of GLOBs that have already been geocoded. For example, a user could carry around a GPS logger that periodically recorded location and geocode a set of pictures that were taken at the same time.
  • Another method of geocoding is simply for the user to manually specify the location of the content. This could be done when the content was generated by allowing the use to enter some location data with the content. Examples of this could be text or even voice recordings.
  • Another way the user may specify the location of the content is by displaying maps on a computer screen and allowing the user to simply point and select where the was generated on the map. Additionally, it is not necessary to specify where all the content was generated, but only a select few that provide location landmarks. If the content has a time code that specifies when it was generated then it is possible to interpolate where the content was generated from the landmarks that the user has created.
  • a GLOB may represents a historical event and includes a time code for a future event such as a predictive model of traffic flow where data is generated that represents what is to be expected at a certain location and at a certain time.
  • GDS GLOB Data Sheet
  • the GDS supports two main functions, namely, a data structure that holds auxiliary information about the content such as location, and a data structure that facilitates the searching for content based upon certain attributes such as location.
  • the illustrative GDS is a data record that contains the following pieces of information: location, time when the content was generated, owner of content, access specification for content, type of content, description of content, rating of content, and access permissions.
  • location information is provided as longitude and latitude.
  • An illustrative time period associated with the content may also be provided, e.g. time when was a photograph taken.
  • An illustrative data field may also identify who or what is responsible for generating and/or owns the content, e.g. an individual, an organization, or perhaps even a Web site from which the content originates.
  • Another piece of information that is contained in the GDS may comprise a specification that allows the GLOB content to be accessed, which can be in the form of an identifier that specifies where the content is located as well as a specification that dictates how to access the content from the specified location. Examples of specifications that support this are Uniform Resource Identifiers (URI's) which are well known methods for doing this on the World Wide Web.
  • URI's Uniform Resource Identifiers
  • the GDS may comprise a data record having a description of the type of content, i.e. photograph, video, audio, etc.
  • the GDS may comprise some sort of annotation that describes the content where the illustrative annotation is in the form of text, but can take many forms including audio, video, or images.
  • the GDS may comprise a rating of the content so that higher rated GLOBs are more likely to be included in Chronicles, Narratives, or Virtual Tours, which are described in further detail below; the rating may be derived from a variety of sources including third parties who have viewed the content.
  • GDS records are placed into a geo-spatial database and searched according to any of the GDS attributes mentioned above. Thus, it should be possible to create queries such as “give me all the photographs taken by John Smith along a route specified by the coordinates X, Y . . . Z.” Such queries would return all the GDS records that satisfy the parameters and the GDS records could be then be used to access the content.
  • the illustrative GLOB content is stored in some location 401 , which may be a database or a file system that is managed by servers 402 . Furthermore, the illustrative GLOB data sheets' are stored in servers 403 . A GDS may be stored in server 403 via a registration request 405 . Application computers 404 , which need to locate and use specific GLOB content, proceed to submit a GLOB query 408 to GDS server 403 . The GDS server 403 would respond with a set of GLOB data sheets that satisfy the query 407 .
  • the application computer 404 may then use the GDS to access the GLOBs 406 from the servers 402 that manage the GLOB content.
  • the accessing of the GLOB content may be performed using a variety of technologies including FTP, SQL, and Web Services, to name a few.
  • GDS control location information.
  • the process of transforming content into a GLOB with a GDS takes advantage of multiple sources of information that determine the content's location.
  • the sources of location information can be categorized as explicit location information, annotations, implicit location information, and association with location logs.
  • Explicit location information is embedded in the content when it is generated and is easily extracted from the content without any processing or transformation of the information.
  • the explicit location information is stored with the content when the content is generated.
  • the device that generated the content has access to location information such as an embedded GPS receiver.
  • a digital camera has an embedded GPS receiver that can be used to store latitude and longitude coordinates in the EXIF header of the images taken and stored by the digital camera.
  • Another example of explicit location information would include a camera phone with E-911 capabilities that receives location from a remote server.
  • the remote server can track the location of the phone and wirelessly send that information to the camera phone so that it can use it to tag any photographs that it is taking. It could use the location information to store latitude and longitude coordinates in the EXIF header of the images that it generates and saves.
  • Annotations are added to the content after it is generated and typically require some sort of data entry step that is separate from the actual generation of the content itself. Annotations are created as a separate step, whereas the explicit location information is captured simultaneously with the content. Examples of annotations include text entry and voice recordings. Annotations are typically stored separate from the content, but it may be embedded in the content itself. For example the JPEG image standard allows for some simple annotations to be stored with the JPEG file itself.
  • Annotations may contain information that can be used to determine the location of the content. Examples are text annotations that contain addresses or simply location key words such as “Eiffel Tower” or “Montana.” These location references can be compared to a database of address or location references in order to resolve them to a specific set of geo-spatial coordinates. Databases that contain addresses or well known points of interest that can be used to translate descriptions or references into geo-spatial coordinates are in common use today.
  • annotations are only for the purposes of location and are explicitly added for that purpose. For example it is possible to create map and text based tools for the user to type in addresses and display maps that allow the user to point and click on the map to specify where it was located. The locations that are specified by using the map based tools become location specific annotations to the content. The annotations in this case would be a set of coordinates that were added by the user.
  • Implicit location information is embedded in the content itself and requires some amount of processing to extract. Implicit location information includes information within the content that may be processed to identify where the content was generated. Additionally, any information that can be uniquely matched to something that does have a known location or can be translated into another form such as text can be used to implicitly encode location information by comparing it to a database of known locations.
  • the location of the Eiffel Tower may be part of a database that can be used to resolve well known references or landmarks to a specific set of coordinates.
  • Another example of this technique is where the user takes pictures of signs in the area that give information such as city and/or street names. It is possible to perform optical character recognition on such images to extract the location information in the form of text. These location references can then be compared to a database of address or location references in order to resolve them to a specific set of geo-spatial coordinates.
  • a location log is a history of locations that are specific to a person or object.
  • a location log contains a time sequenced series of records with at least the following information: a specification for a location; an identifier for whom or what the location refers to; and a time stamp. Therefore, if there exists a log of locations that specify where some individual as located at particular times, then it is possible to correlate the times that the individual generates content to the locations inside the location log to geocode the content.
  • the specification for a location may comprise a variety of units including latitude/longitude or a description of the location such as an address.
  • the location specification is used to determine where on the earth something is located.
  • the location may have a further qualifier that specifies the accuracy and resolution of the location. For example some records may be derived from GPS locations which have a high degree of resolution and accuracy while other records my be derived from auxiliary documents such as travel itineraries which may only have resolution down to the city, but not where in the city.
  • the identifier comprises information related to whom or what the location refers to. Typically, this is an individual where the identifier can simply be the individual's name.
  • the time stamp designates when the above mentioned person or thing was located at the above mentioned location.
  • the time can be specified in a variety of units as long as it can be translated or converted into a date and time. Additionally, the time stamp may have varying degrees of accuracy.
  • the location log takes of advantage of the capabilities of digital recording devices in use today, e.g. digital cameras, which have a means to tag the content with the time.
  • the geocoding step can be automated if a set of location logs exist that specify where the person that is generating the content is located at times that correlate to the time when the content is being generated. In the illustrative instance where the time the content is generated does not match a record in the location log, it is still possible to estimate the location by extrapolating the location from one or two locations in the location log that are close in time.
  • location logs are stored in a geo-spatial database that supports the querying of entries either by their location or by their time and identifiers.
  • location log records there are a variety of technologies and information sources that can be used to generate location log records.
  • cell phones with embedded GPS receivers or that adhere to the E911 directive may be used, and it is possible for enterprise servers to query and log the location of Cell phones.
  • location trackers such as GPS loggers that record the latitude/longitude of the person carrying the device, and TV-GPS and/or Wi-Fi location technologies may be used; these location logs can later be uploaded and added to the user's overall location log.
  • the explicit location information embedded in content as described above may be used where the location information can be extracted from content and added to the user's location log. Further still, the implicit location information derived from content as described above can be extracted from content and added to the user's location log. Further yet, content annotations as described above can be converted to locations that can be added to the user's location log.
  • Vehicle navigation systems are used to track a vehicle's location and give directions and they use a variety of methods including GPS to track the location of the vehicle; these systems could be used to provide location information that could be loaded into a user's location log.
  • RFID technology such as that used in toll tags in automobiles may also be used.
  • Financial transactions of the user could be mined for information that specifies where the user was located when the transaction occurred. Schedules from a user's calendar or personal information management system can be mined to determine if a location was specified as part of a meeting appointment.
  • IP address location may be used since it is possible to locate where a particular IP address is physically located based upon whom or what organization was assigned that address, so that when a person accesses on line services from some remote location it may be possible to determine where they were when they accessed those services.
  • the general concepts of a GLOB were introduced above.
  • a discussion of GDS followed that described the two GDS functions, namely, a data structure that holds auxiliary information, and a data structure that facilitates searching.
  • the GDS information comprises one or more sources of location information such as explicit location information, annotations, implicit location information, and location logs.
  • FIG. 5 there is shown an illustrative diagram of software modules and processes that generate GLOB data sheets. More particularly, FIG. 5 shows the various functions and modules involved in the process of geocoding content to generate at least one GDS.
  • sources of content 508 including digital cameras.
  • This content is collected and stored within repository 506 and managed using a variety of third party and vendor specific tools 507 .
  • the management of the content includes downloading it from the sources and storing 515 the content in some repository 506 .
  • the repository 506 may be local and stored on the user's computer.
  • the content is uploaded to some third party server on the internet.
  • the repository 506 may be simply a file system on some computer or perhaps a database and access to the repository 506 may be via some server (not shown).
  • the repository 503 is a geo-spatial database and typically resides on a network that is publicly accessible.
  • the location log database server is able to handle queries 512 in order to select specific log entries that correspond to specific users and time spans.
  • the geocoding process 504 uses all available sources of location information in order to create a GDS for each entry in the content repository 506 .
  • Each GDS that is created is stored 513 in a GDS repository 505 .
  • the GDS contains a reference to the content in the content repository 506 as well as location information about where the content is located.
  • the geocoding process 504 uses multiple stages to resolve the content's location as shown in FIG. 6 .
  • the first step, 601 is to fetch some content from the content repository 506 .
  • Decision diamond 602 is then executed to determine if there is any explicit location information stored with the content. If YES then process 603 is used to extract the explicit location information from the content. Extracting the location information is dependent upon the type of content and how the location information is stored with the content. A variety of formats and content types can be supported.
  • Decision diamond 604 is then executed to determine if the location information that was extracted is adequate. Locations are deemed to be inadequate if their resolution and accuracy do not meet certain minimum standards or if there simply is no location record. If decision diamond 604 is YES (i.e. location data is adequate) then it is not necessary to do any more geocode processing to resolve the content's location and process 616 is executed. Process 616 adds location information to the location log 503 if it does not already exist. This is performed so that the location information can be used in the future to geocoding other content. Process 617 is then executed and uses the location information to generate a GDS and add it to the GDS repository 513 and geocoding is complete.
  • YES i.e. location data is adequate
  • decision diamond 602 was NO (i.e. no explicit location information in content) or decision diamond 604 is NO (i.e. locations inadequate) then further geocoding needs to take place and decision diamond 605 is executed to see if there are any content annotations. If YES then the annotations are processed by process 606 to extract any location information from the annotations and decision diamond 607 is executed to see if these locations are adequate. Decision diamond 607 is then executed to see if the locations are adequate (analogous to 604 as described above) and if the answer is YES then process blocks 616 and 617 are executed as described above and geocoding is complete.
  • decision diamond 605 was NO (i.e. no content annotations) or decision diamond 607 is NO (i.e. locations inadequate) then further geocoding needs to take place and decision diamond 608 is executed to determine if there are any locations in the log that correspond to the content. This decision is based upon the owner of the content and the time period that the content was generated. If there are records in the location log that correspond to the content (i.e. have the same owner and similar time stamps) then the answer to decision diamond 608 is YES and process 609 is executed. Process 609 uses the records from the location logs to calculate locations for the content. This may require extrapolation of the locations from the location log depending upon how closely the time stamp of the content matches the time stamps of the location log records. Decision diamond 610 is then executed to see if the locations are adequate (analogous to 604 as described above) and if the answer is YES then process' 616 and 617 are executed as described above and geocoding is complete.
  • decision diamond 608 was NO (i.e. no corresponding records in the location log) or decision diamond 610 is NO (i.e. locations inadequate) then further geocoding needs to take place and process 611 is executed to extract any implicit location information from the content itself. Decision diamond 612 is then executed to see if the locations are adequate (analogous to 604 as described above) and if the answer is YES then process' 616 and 617 are executed as described above and geocoding is complete.
  • decision diamond 612 is NO (i.e. locations inadequate) then a further processing step 613 is performed to try and combine all the locations from the previous stages together to see if the location can be further resolved.
  • This step would be to simply take the intersection of all the previous locations as see if it results in a more refined location.
  • Decision diamond 614 is then executed to see if the locations are adequate (analogous to 604 as described above) and if the answer is YES then process' 616 and 617 are executed as described above and geocoding is complete.
  • Process 615 is a final stage in which various graphical map and text based tools are used to allow the user to specify addresses and point and click on a map to manually specify the contents location. After this step it is assumed that the location has been specified to the level of accuracy and resolution necessary and process' 616 and 617 are executed to generate the GDS and place it in the GDS repository 505 and geocoding is complete.
  • the algorithm associated with FIG. 6 describes six distinct location resolution stages: explicit locations in content, i.e. process blocks 602 , 603 , and 604 ; content annotations, i.e. process blocks 605 , 606 , and 607 ; location log association, i.e. process blocks 608 , 609 , and 610 ; implicit location information, i.e. process blocks 611 , and 612 ; combine all previous results, i.e. process blocks 613 , and 614 ; and tools to allow the user to manually annotate the content with location, i.e. process block 615 .
  • stage 3 i.e. check the location logs
  • stage 2 i.e. extract location information from the content annotations
  • process blocks 610 , 611 , and 612 may be performed before process blocks 605 , 606 and 607 .
  • This algorithm is in fact designed so that stages 1 - 4 can be re-ordered or even omitted to increase efficiency.
  • GDS GLOB Data Sheets
  • a Chronicle or Narrative is a collection of GLOBs that are organized according to time and location, and may then be animated.
  • the Chronicle is a multi-resolutional depiction of the GLOB data with respect to time and location.
  • the illustrative Chronicle comprises a map display where the time sequenced GLOB content is displayed in order to show the geographic location that the content refers to.
  • a synopsis of the GLOB content as it changes over time can also be shown in the illustrative Chronicle.
  • the Chronicle has a notion of time that is displayed to the user 709 .
  • This notion of time is in relation to the GLOB content that is being animated, i.e. the pictures 701 .
  • the time being displayed corresponds to the time at which the pictures were taken.
  • the illustrative images or pictures are animated in the form of a slide show that is time sequenced.
  • the pictures are being animated by thumbnails that are scrolling smoothly across the screen.
  • the illustrative maps include maps generated using aerial imagery, or computer generated street maps, or any combination thereof.
  • the illustrative maps show the location of the images that are currently being displayed in the slide show 701 .
  • the location of the GLOB pictures are shown by both markers on the map 703 and pointers 702 which point to those markers. This allows the user to easily see where the GLOBs are geographically located.
  • a route is simply a time sequenced collection of geo-locations and can be animated by drawing the route on the map as the Chronicle's time elapses.
  • the route could represent a route that was taken when the pictures were generated or could represent a route that is to be followed in the future and the pictures represent what is expected to be seen along the way. Therefore, a Chronicle can either represent something that is either currently happening, has happened in the past, or is expected to happen.
  • the Chronicle may comprise a set of player controls 708 that allow the user to control various aspects of the Chronicle such as playing/pausing, forward/reverse and a method to quickly move to certain time indices of the Blog that is similar to a videotape player.
  • the illustrative Chronicle may animate a synopsis of GLOB sequences in order to allow the user to quickly and easily view the information represented by the GLOB. Additionally, the user may “drill down” and view more detail associated with a particular GLOB, which may entail getting a more detailed view of a particular picture. The user could do this by simply selecting the thumbnail of the picture and then the Chronicle would pause and a more detailed view of the picture may be displayed. An illustrative example of this is shown in FIG. 8 , which displays a more detailed view of the selected GLOB.
  • the Chronicle is generated using the attributes and features associated with GLOB Data Sheets, which were described above. However, the method for generating Chronicles may require processing third party content and including this third party content in the Blog.
  • FIG. 9 there is shown the process of generating a Chronicle or Narrative that integrates third party information.
  • the process for generating a Chronicle involves combining GLOB data that has been generated with maps and other third party content. Initially there exist a set of data content that is generated for the purpose of using as a GLOB 901 . Note, when the data is generated it may or may not be geocoded. There also exists various third party content such as Maps, documents, etc., which may be used in conjunction with the generated GLOB data to create a Chronicle.
  • the first process step 903 is for the generated date to be collected and turned into GLOBs by adding geo-location information if necessary.
  • Process step 903 may also entail annotating the GLOBs with additional information such as textual description of the content of the GLOB.
  • additional information such as textual description of the content of the GLOB.
  • the information is indexed and archived in some fashion so that it may be retrieved later. This may entail the use of well known database technologies.
  • the result of this is a set of GLOB data sets 905 that can bused for the Chronicle generation.
  • the actual Chronicle generation process 907 entails using a Chronicle specification 909 to collect all the data, namely, both GLOBs 905 and third party date 906 .
  • the Chronicle specification may comprise the type of GLOB data desired, the location, the time frame, and source of the GLOB data.
  • the collection of the third party data may interface with various map servers and other databases and possibly Internet based information servers 904 to collect all the information that is relevant to each GLOB and its associated content 906 .
  • the information that is retrieved is dependent upon the specification of the Chronicle.
  • a Chronicle specification may refer to the past, present or future time frames.
  • a graphical representation is generated where the GLOB data is animated in time and shown geographically, as described above. The manner in which the Chronicle is generated and displayed is discussed in more detail below.
  • the Chronicle or Narrative specifications 909 can specify various types of information such as the following: show traffic and roadside images for a specific route taken at a specific time; show a photographic travel log of a trip; shows the states of various sensors located in a particular region; show additional third party information such as weather; or any combination thereof.
  • a Chronicle Once a Chronicle is generated it may be stored in a variety of formats including the following: source GLOB data and Chronicle specifications that allows the Chronicle either be regenerated or viewed; video files on servers or computers that can be played in a variety of ways; video files on local store media such as DVD's, hard disks, flash memory, or other such storage media.
  • Chronicles can be stored with only synopsis information, e.g. skinny Chronicle as described blow, where the detailed GLOB information is stored separately and accessed on an as needed basis.
  • Chronicles which have been generated can be viewed in a number of different ways including the following: using a computer like PC or possible even a mobile device that runs a special Chronicle viewer application and uses the specification and GLOB information in the Chronicle to generate and display it; using any sort of digital video capable device such as computer or mobile device that can play video files that are stored either locally or remotely on the computer; using a television and DVD player where there is an interactive video of the Chronicle stored on some type of local media such as DVD or video recorder; using a network, e.g. Internet, capable device that is capable of receiving and displaying streaming video.
  • Such devices include, but are not limited to, desktop computers, mobile PDAs, navigation devices, cell phones, cable television, and similar network capable devices.
  • FIG. 10 there is shown a general overall architecture for Chronicle generation.
  • the generation of Chronicles requires gathering GLOB information from various sources 1001 , 1003 , 1005 , and 1007 that are served up from a variety of servers 1002 , 1004 , and 1006 that may be distributed and interfaced via some sort of wide area network 1012 such as the Internet and the World Wide Web.
  • the vehicle telemetry server 1009 is an example of a server that continuously collects data from a potentially large number of sources (vehicle probes 1007 ) and stores the information in a database 1008 . It shall be appreciated by those skilled in the art that servers such as 1008 and 1009 do not necessarily apply only to vehicles, but could apply to any data sources which may be numerous and continuously generate information such as sensors that are part of some large sensor network.
  • Chronicles or Narrative may be viewed by a number of different devices 1013 , 1014 , 1015 , 1016 .
  • These devices may be normal desktop systems 1014 , mobile devices 1615 such as PDA's or cell phones, special embedded systems such as navigation or vehicle information systems inside vehicles 1013 or even video display systems such as televisions 1016 .
  • mobile devices 1615 such as PDA's or cell phones
  • special embedded systems such as navigation or vehicle information systems inside vehicles 1013 or even video display systems such as televisions 1016 .
  • these devices are network via some sort of WAN 1012 that includes such systems as cable television.
  • FIG. 10 also shows a server 1010 which aggregates information from the various GLOB sources and both generates the Chronicle and serves them to the devices used to view the Chronicles.
  • the Chronicle server 1010 will store the generated Chronicles in a database 1011 .
  • FIG. 11-15 there is shown a plurality of different system configurations for the various servers, GLOB sources, and Blog viewing devices that are depicted in FIG. 10 .
  • These system configurations are provided as illustrative examples and those skilled in the art shall appreciate that there may be different combinations that perform the same of similar functions.
  • FIG. 11 there is shown the various components of the Chronicle or Narrative presentation systems. More particularly, FIG. 11 shows the various components that are used to describe the various scenarios.
  • third party content datasets 1101 that contain GLOB information such as maps and other readily available information such as traffic reports, weather, news, and other such information that may come from various third party vendors or sources. This information may be stored in traditional databases or simple files and made available through standard interfaces such as web services.
  • This data may also be stored either in files or in a database of some sort.
  • user devices 1107 which represent the various platforms that the end user will use to view the generated Chronicles. In some scenarios, these same devices may also store the GLOB content and generate the Blogs.
  • remote servers 1105 that are used to either serve up GLOB content or generate and serve Chronicles.
  • Fat Chronicles There exist so called “Fat Chronicles” 1103 that provide fully self-contained Chronicles in which all the GLOB content of the Chronicle is stored with the Chronicle. Fat chronicles do not require additional information, but are very large and thus more difficult to move around over WAN's.
  • Skinny Chronicles 1104 that contain only enough content for the previewing of the GLOB data, but not the detailed GLOB information. Skinny Chronicles are much smaller and easier to move around, but require that detailed content be served up on an as needed basis.
  • FIGS. 12-15 describe a number of system architectures that may be used to manage the GLOB content and generate the Chronicles. Each of these scenarios may be used depending on system design and limitations. The main differences between the various scenarios revolve around the following design parameters: how much GLOB content is managed and stored locally versus remotely; and how much of Chronicle is generated locally versus remotely.
  • FIG. 12 there is shown an illustrative Chronicle architecture with fat servers and thin clients.
  • this architecture all the GLOB content is managed remotely. This may entail users to transfer any GLOB content that they have generated to the remote servers.
  • the Chronicles are generated by remote servers and served up to the user's computers.
  • the Chronicles may be served up as simple web based application or there may exist as a custom application on the user's computers that allow the Chronicle to be viewed.
  • the Chronicles are accessible from anywhere and thus easier to share among various users.
  • the Chronicle may be served up as a web application thus the user utilizes a browser based interface, thus no custom program is needed to view the Chronicles.
  • FIG. 13 there is shown an illustrative Chronicle architecture with no remote servers.
  • all the GLOB content and Blogs are generated and managed locally on the users computers so that everything is self-contained and so there is no Wide Area Network (WAN) required and no remote servers are required.
  • WAN Wide Area Network
  • FIG. 14 there is shown an illustrative Chronicle architecture with remote servers having generated skinny chronicles.
  • the remote server serves up skinny Chronicles while the locally generated GLOB content is stored and managed on the user's local computer, and there is less of a burden to upload all locally generated GLOB content. All that needs to be uploaded are the GLOB synopsis. There is less storage required on the remote servers.
  • FIG. 15 there is shown an illustrative chronicle architecture where the server serves remote content and clients generate the chronicle.
  • the remote servers serve up remote GLOB content and user's computers manage locally generated GLOB content and generate the Chronicles. This scenario does not require transferring all locally generated GLOB content to a remote server, and there is less storage required on the remote servers.
  • GLOBs and GLOB Data Sheets have been provided. Additionally, attention has been directed to the software modules that collect geocodes and content from multiple sources, and combines the content and geocodes to produce GLOB Data Sheets (GDS), which may be described as “multi-stage” geocoding and GDS generation. Subsequently, an illustrative example of a “Chronicle” or “Narrative” of GLOBs that are displayed to a user was discussed. Furthermore, an illustrative process for generating a Chronicle or Narrative, which integrates third party information, was also described. A plurality of different system architecture for Chronicle generation was also presented.
  • GDS GLOB Data Sheets
  • This method is also referred to as geocoding groups of sequenced object wherein location information is used, regardless of its source, to geocode groups of sequenced objects.
  • the method requires that there exist enough location information to individually geocode each object in the group and does not rely on the fact that the objects in the group are sequenced.
  • the method describes explicitly geocoding a subset of the objects in a group and then implicitly calculating the locations of the other related objects in the group by exploiting their relationship to the explicit location information for one or more GLOBs.
  • the first and simplest way to geo-locate a group of related objects is to group them all together and assign the same geocode to them all. Therefore, when one of the objects is geo-located all the objects can be assigned the same geocode. For example, there may be a collection of photographs that all have the same annotation of “May field trip” and are thus related in this fashion. If any of the photographs in that collection can be geo-located, then by using the associated annotation of “May field trip” all the photographs can be located in the same place. Although this may be adequate for many applications, it does not allow for the differentiation by location of the individual objects in the group.
  • FIG. 16 there is shown a group of related objects such as photographs 2301 , 2302 , 2303 , 2304 , and 1205 that have specific time stamps.
  • FIG. 16 it is possible to order the pictures into a sequence based upon their time stamps. In FIG. 16 , they are ordered from left to right according to their time stamps as indicated.
  • FIG. 17 shows the display of an example graphical based tool that some user can use to geocode the photographs from FIG. 16 .
  • Such a tool is a computer program that could be a stand alone program or a web based application that runs within an internet browser. It contains a map display 2406 that the user can use to explicitly geocode photographs.
  • FIG. 17 shows the location on the map where photographs 2301 and 2305 were explicitly located as described above, and this is depicted by locations 2401 and 2405 , respectively.
  • an implicit linear path 2407 that can be used to implicitly geocode the remaining photographs 2302 , 2303 , and 2304 . Assuming the remaining photographs lie somewhere on this path we can use the time stamps of the photographs and their relationships to the explicitly located photographs to determine their location on the path 2407 .
  • FIG. 17 shows implicit locations 2402 , 2403 , and 2404 that lie on path 2407 and whose relative locations are a linear interpolation along the path based upon time. For example, implicit location 2403 lies on the mid point of path 2407 because the time stamp of photograph 2303 is midway between the time stamp of 2301 and 2305 which form the endpoints of path 2407 .
  • FIG. 18 shows a case where photographs 2301 , 2303 , and 2305 have been explicitly located to locations 2401 , 2503 , and 2405 , respectively. In this instance, there are two paths generated. There exists path 2506 between 2401 and 2503 and path 2507 between 2503 and 2405 . In FIG. 18 , location 2502 is implicitly calculated as a linear interpolation based upon time along path 2506 and likewise location 2504 is implicitly calculated as a linear interpolation along path 2507 based upon time.
  • each successive pair of explicitly located objects in the list determine a path along which the other objects may be implicitly located by extrapolating their locations along the paths formed by the two objects that are greater than and less than they are according to time.
  • FIG. 19 shows examples of using non-linear paths and non-linear time interpolation along those paths.
  • photographs 2301 and 2305 have been explicitly located at locations 2601 and 2605 .
  • the implicit locations 2602 , 2603 , and 2604 are extrapolated along path 2607 using the rule that all photographs taken within 10 minutes of an explicitly located photograph are assumed to be in the same location as the explicitly located photographs.
  • the general method described above relies on the fact that the objects are sequenced and not upon the fact that they have time stamps.
  • the time stamps of the objects were used to create the sequence and to also extrapolate locations along paths, but it is possible to perform similar operations even if the objects did not contain time stamps.
  • FIG. 20 shows an example of geocoding pictures that do not have time stamps, but are still part of a user defined sequence.
  • the pictures are related by the sequence, but they do not have any additional information (such as time) that allows the location to be extrapolated other than by distance along the path.
  • the implicitly located photographs 2702 , 2703 , and 2704 are evenly spread between the explicitly located photographs 2701 and 2705 based upon their distance along the path 2707 from 2701 to 2705 .
  • FIG. 21 provides an illustrative example where the user has specified that photographs 2801 , 2802 , and 2803 were taken in the same place.
  • photographs 2801 and 2805 are explicitly located and photographs 2802 , 2803 , and 2804 are implicitly located along path 2807 . Since photographs 2802 and 2803 were related to 2801 to be in the same place they are located in the same place as 2801 . Since 2804 does not contain any additional relation information it is located midway between 2801 and 2805 along the path 2807 .
  • a “Virtual Tour” is a visual and/or audio presentation of what one might experience if they were to visit some location. This presentation is typically rendered on some sort of electronic display or screen that is connected to a computing device such as a desktop or handheld computer.
  • the Virtual Tour shows one or more maps at various levels of scale 3105 , 3106 , 3107 , and 3110 that clearly show to the user the geographic location of the tour. Although multiple scale levels are shown, only one scale level is required.
  • the maps may be from aerial imagery, computer generated street maps, or any combination thereof.
  • the Virtual Tour contains images in the form of pictures and/or videos that show what the user might see and/or hear if they were to actually go to the locations indicated in the map.
  • the images are being animated in the form of a slide show that is time sequenced.
  • the pictures are being animated by thumbnails that are scrolling smoothly across the screen.
  • the maps show the location of the images that are currently being displayed in the slide show 3101 .
  • the location of the images are shown by both markers on the map 3103 and pointers 3102 , which point to those markers. This allows the user to easily see where the images are geographically located and what they might expect to see if they were at those locations.
  • a route is simply a time sequenced collection of geo-locations and can be animated by drawing the route on the map as the images scroll by (i.e. time elapses). It is important to note that the route could represent a specific tour and correspond to the sequence of images being shown.
  • the Virtual Tour may have a notion of time that is displayed to the user 3109 .
  • This notion of time is in relation to the content that is being animated, i.e. the images 3101 .
  • the time being displayed corresponds to the time at which a proposed tour may be taken.
  • the virtual tour may contain a set of player controls 3108 that allow the user to control various aspects of the tour presentation such as playing/pausing, forward/reverse and a method to quickly move to certain time indices of the Virtual Tour. This is much like one might find on a Video player.
  • keys or controls which allow the user to pan and zoom the various maps.
  • the Virtual Tour may contain icons or images superimposed on the map 3111 and 3112 , which represent “Points of Interest” (POI). These points of interest may be businesses such as hotels and gas stations or they may be places which have some attraction such as a museum or a tennis club. Businesses such as hotels or restaurants may be sponsored by the business establishment and any images or information may be provided by the business itself. This might even include promotional items such as discounts or coupons.
  • POI Points of Interest
  • Virtual Tours animate a synopsis of image sequences in order to allow the user to quickly and easily view the images and their locations. If the user so desired they could “drill down” and view more detail associated with a particular image or location. In the illustrative example provided in FIG. 22 this might entail getting a more detailed view of a picture. The user could do this by simply selecting the thumbnail of the picture and then the Virtual Tour would pause and a more detailed view of the picture displayed, as shown in FIG. 24 .
  • the illustrative Virtual Tour is generated automatically by a remotely located Virtual Tour Server in response to a user specifying the type of tour they are interested in seeing.
  • FIG. 25 shows a Virtual Tour Server 3404 that creates a Virtual Tour that can be viewed on a variety of user viewing devices 3405 including mobile devices, desktop computers, video monitors (TV's), and in-vehicle systems such as navigation systems.
  • the Virtual Tour Server 3404 is accessed by the user via some sort of wide area network such as the Internet 3403 .
  • the user specifies the type of Virtual Tour they want to view by indicating one or more factors that they are interested in such as: location of the tour such as a city or country; route to be taken; type of activities desired such as surfing, and other well-known activities; type of attractions, e.g. museums; specific events; date and time of the tour; and new tour based upon previous Virtual Tour.
  • a Virtual Tour is generated by the Virtual Tour Server by bringing together and displaying GLOBs such as images, maps and POI data 3401 from a variety of sources.
  • This GLOB content is accessed via GLOB servers 3402 that are accessible via a wide area network 3403 such as the internet.
  • the Virtual Tour Server combines all the GLOB content together and presents it to the user in a manner as described previously.
  • FIG. 26 shows is a more detailed depiction of the full architecture where there exists a Virtual Tour Server that generates the Virtual Tours. Users access the Virtual Tour Server from a variety of devices such as vehicle information systems 3813 , desktop computers 3814 , mobile devices 3815 , and streaming video displays 3816 .
  • the Virtual Tour Server may be accessed over a variety of public networks 3812 such as the Internet.
  • the Virtual Tour Server Upon receiving a request from a user for a Virtual Tour the Virtual Tour Server collects a variety of GLOB content from various sources and dynamically generates the Virtual Tour. As previously discussed there exists a variety of GLOB content such as maps 3801 , third party content 3803 , user generated content 3805 , and POI data and advertisements 3817 . This content is managed by a variety of servers 3802 , 3804 , 3806 , and 3807 and accessed via a network 3812 .
  • the Virtual Tour Server 3810 locates the GLOB content by submitting requests to the GLOB Resource Server 3809 and 3808 .
  • the Virtual Tour Server Upon getting a collection of GLOB Data Sheets from the GLOB Resource Server, the Virtual Tour Server accesses the GLOB content and uses it to produce a Virtual Tour. The Virtual Tour is then served up to the user that made the request for the Virtual Tour.
  • Virtual Tour Server need only access and serve up the Virtual Tour without generating it.
  • Such pre-produced Virtual Tours may exist as the result of users generating and storing their own Virtual Tours or perhaps as the result of a business related to a POI generating one to help promote their business. For example, a hotel may produce a Virtual Tour that can be shown to potential customers that are interested in staying at their hotel.

Abstract

A system and method for generating a virtual tour on a display device is described. The method comprises providing at least one map. The method further comprises providing a plurality of sequenced images, wherein each of the images is associated with at least one location by a geo-coding module configured to generate a geo-location object data sheet that associates sequential images with a corresponding location. The sequenced images are organized based on the location of each of the sequenced images and displayed on the map. The method is implemented by the system.

Description

    CROSS REFERENCE
  • This patent application is a continuation of non-provisional patent application Ser. No. 13/014,600 filed on Jan. 26, 2011, which is a continuation of non-provisional patent application Ser. No. 11/645,001 now U.S. Pat. No. 7,904,483 filed on Dec. 23, 2006, which claims the benefit of provisional patent applications 60/774,931 filed on Feb. 21, 2006, 60/792,211 filed on Apr. 17, 2006, and 60/793,928 filed on Apr. 24, 2006, all of which are hereby incorporated by reference.
  • Additionally, non-provisional patent application Ser. No. 13/014,600 is a continuation of non-provisional patent application Ser. No. 12/571,547 now U.S. Pat. No. 7,917,543 filed on Oct. 1, 2009, which is a continuation of non-provisional patent application Ser. No. 11/645,001 now U.S. Pat. No. 7,904,483 filed on Dec. 23, 2006, and which claims the benefit of provisional patent applications 60/774,931 filed on Feb. 21, 2006, 60/792,211 filed on Apr. 17, 2006, and 60/793,928 filed on Apr. 24, 2006, all of which are hereby incorporated by reference.
  • BACKGROUND
  • 1. Field of the Invention
  • The invention relates to a system and method for geo-coding user generated content. More particularly, the invention relates to a system and method for geo-coding a plurality of user generated content.
  • 2. Description of Related Art
  • With the widespread use of Global Position System (GPS) sensors, geographic identifiers are commonly used to track vehicles and to help drivers find a particular location. Geocoding is the process of assigning geographic identifiers such as latitude, longitude, or Universal Transverse Mercator (UTM) to information. The information having the geographic coordinates can be mapped and entered into a Geographic Information System (GIS). These location tags may also take the form of an actual address which can be translated into latitude and longitude.
  • An illustrative example of geocoded information is a geocoded photo. The geocoded photo is an image that is associated with a geographical location. A geocoded image can be associated to geographic coordinates such as latitude and longitude or a physical address. In theory, every part of a picture can be tied to geographic location, but in the most typical application, only the position of the photograph is associated with the entire digital image.
  • There are various methods for geocoding images. For example, using a location aware device such as a GPS receiver, the location of the camera used to take a particular photo is recorded. Although most digital cameras presently available do not contain a built in GPS receiver, an external GPS device can be used to keep track of the camera location and the data can be synchronized using specialized software.
  • Regretfully, the process of geo-coding user generated content is relatively challenging and time consuming task. For example, the currently available location aware devices are specialized devices that do not integrate with other electronic devices that may be used to gather or capture user generated content. Thus, it becomes quite challenging to geo-code user generated content that is generated from a first device with location specific content that is generated by another device that has different capabilities.
  • It would therefore be beneficial to provide a simplified system and method for geo-coding user generated content.
  • SUMMARY
  • A system and method for generating a virtual tour on a display device is described. The method comprises providing at least one map. The method further comprises providing a plurality of sequenced images, wherein each of the images is associated with at least one location by a geo-coding module configured to generate a geo-location object data sheet that associates sequential images with a corresponding location. The sequenced images are organized based on the location of each of the sequenced images and displayed on the map. The method is implemented by the system.
  • DRAWINGS
  • References to the following drawings are for illustrative purposes and not for limiting purposes. The following drawings describe one or more illustrative embodiments.
  • FIG. 1 shows illustrative examples of information converted to a geo-location object (“GLOB”).
  • FIG. 2 shows illustrative location descriptors.
  • FIG. 3 shows illustrative attributes of a GLOB data sheet.
  • FIG. 4 shows an illustrative system configured to query GLOB data sheets.
  • FIG. 5 shows an illustrative diagram of software modules that generate GLOB data sheets.
  • FIG. 6 shows an illustrative flowchart of a geocoding process.
  • FIG. 7 shows an illustrative example of a Chronicle of GLOBs.
  • FIG. 8 shows an image with more detail of a particular GLOB.
  • FIG. 9 shows a flowchart for generating an illustrative Chronicle or Narrative that integrates third party information.
  • FIG. 10 shows illustrative system architecture for generating a Chronicle or Narrative.
  • FIG. 11 shows the various components of the Chronicle or Narrative presentation systems.
  • FIG. 12 presents an illustrative system architecture with fat servers and thin clients.
  • FIG. 13 presents an illustrative system architecture with no remote servers.
  • FIG. 14 presents an illustrative system architecture with remote servers that generate skinny Chronicles or Narratives.
  • FIG. 15 presents an illustrative system architecture with remote servers and clients generating the Chronicle or Narratives.
  • FIG. 16 provides an illustrative depiction of a time stamped photographs.
  • FIG. 17 illustrates the explicit and implicit geocoding of photographs.
  • FIG. 18 further illustrates explicit and implicit geocoding of photographs.
  • FIG. 19 provides an illustrative example of implicit geocoding based on different path and distribution rules.
  • FIG. 20 provides an illustrative example of implicit geocoding of a user defined photograph sequence with no time stamps.
  • FIG. 21 provides an illustrative example of implicit geocoding of a user defined photograph sequence with no time stamps and additional relationship rules.
  • FIG. 22 provides an illustrative Virtual Tour.
  • FIG. 23 shows an illustrative display with detail for a selected Point of Interest.
  • FIG. 24 shows an illustrative display of a selected geo-located image.
  • FIG. 25 presents an illustrative system architecture for an illustrative virtual tour server.
  • FIG. 26 presents an illustrative system architecture for a more detailed virtual tour server architecture.
  • DESCRIPTION
  • Persons of ordinary skill in the art will realize that the following description is illustrative and not in any way limiting. Other embodiments of the claimed subject matter will readily suggest themselves to such skilled persons having the benefit of this disclosure. It shall be appreciated by those of ordinary skill in the art that the systems described hereinafter may vary as to configuration, as to details, and as to the functionality of system elements. Additionally, the method may vary as to details, partitioning, repetition, act inclusion, order of the actions, or other variations without departing from the illustrative method disclosed herein.
  • As described above, geocoding an object refers to associating location or geographic information, i.e. a geocode, with the particular object. Geocodes are also referred to as geo-locations. For purposes of this patent application, an object that has an associated geocode is referred to as a geo-located object, hereinafter “GLOB.” The GLOB is a piece of information tagged with the geographic location of the content of the information.
  • Referring to FIG. 1 there is shown illustrative examples of information that may be converted to a GLOB. More particularly, FIG. 1 shows an illustrative example of a plurality of sources of information that can be turned into a GLOB such as video, audio, sensors/data, maps, text/document, location, i.e. GPS logs, mobile telemetry, and pictures. Additionally, the attributes associated with each GLOB are shown: location (where); time (when); content description (what); and source of data (who).
  • A description of a system that can be used to present one or more GLOBs is described. Additionally, a system to animate GLOBs that are indexed by time and that include geographic information is also described. Additionally, GLOB information may be presented in a synopsized form that allows the user to obtain detail about a particular object. Additionally, a plurality GLOBs can be generated from a plurality of different sources of information. The systems and methods described herein may also be used to geocode multiple objects and to organize the GLOBs. For example, geocoding of multiple objects that are sequenced in relation to one another is described. The systems and methods take advantage of the objects belonging to the same group. Finally, the systems and methods can be used to generate virtual tours where the virtual tour is a visual and/or audio presentation of what one might experience when visiting a particular location.
  • By way of example and not of limitation, a GLOB could be a digital photograph that is tagged with the time and location at which the photograph was taken. The content of the photograph is the photographic image itself. There may also be tags which describe who took the photograph and the subject matter of the photograph.
  • Other illustrative information that can be converted to a GLOB includes, but is not limited to, maps, videos, recorded location logs (e.g. GPS logs), audio recordings, weather reports, news items, sensor measurements, automotive telemetry such as location, speed, camera images, and any other means for generating content. The examples of information presented above are referred to as “generated content” or sometimes just “content” for short. The content described can be grouped together into sets of related objects. For example, in the case of photographs, a user might decide to group together a collection of photographs that were taken around the same time frame as a particular trip or event.
  • In addition to location information and possibly temporal information, GLOBs may be tagged with other state information such as altitude, heading, and azimuth. GLOBs may also be tagged with motion state parameters such as speed, acceleration, and bearing. The location may have a further qualifier that specifies the accuracy and resolution of the location.
  • Referring to FIG. 2 there is shown illustrative location descriptors, which shows the topological forms that a location attribute of a GLOB can take. The different forms can range from being a single point to some defined area. A location area represents the bound within which the GLOB may exist. There may also be some statistical distribution which can be used to determine the probability of the GLOB existing in a particular point. It is also important to note that the resolution and accuracy of the location can vary widely and is dependent upon the resolution and accuracy of the location information source.
  • For example if the location was derived from a GPS fix then it will have a high degree of resolution, but if the location was derived from some annotation that simply says “Montana” then it will only have a resolution that corresponds to the state of Montana. Either of these situations may be modeled using one or the topological forms shown in FIG. 2.
  • Information that is geocoded may be stored or associated with a particular digital file. For example, many digital recording formats such as JPEG have facilities built into their specification that allow for the addition of location information. Digital formats which do not may require some sort of auxiliary data file that can be stored with the content.
  • There are a number of ways in which content can be geocoded in order to be turned into a GLOB. The first and most straightforward method is for the device that generates the content to geocode it when the content is generated. This requires that the device have access to location information such as an embedded GPS receiver or perhaps some sort of remote server that can locate the device and transmit its location information to the device. Examples of the latter approach include cell phones which have so called E-911 capabilities.
  • If the device that generates the content is not able to geocode the content when it is generated then it is necessary to perform some sort of processing step after the content is generated in order to geocode it. Most digital recording devices in use today such as digital cameras have a means to take the content that it generates with the time in which the content was generated. The geocoding step can be automated if another collection of GLOBs exists that were collected at the same time and within the same area as the content that is not geocoded. It is then possible to geocode the content by time correlating with the collection of GLOBs that have already been geocoded. For example, a user could carry around a GPS logger that periodically recorded location and geocode a set of pictures that were taken at the same time.
  • Another method of geocoding is simply for the user to manually specify the location of the content. This could be done when the content was generated by allowing the use to enter some location data with the content. Examples of this could be text or even voice recordings. Another way the user may specify the location of the content is by displaying maps on a computer screen and allowing the user to simply point and select where the was generated on the map. Additionally, it is not necessary to specify where all the content was generated, but only a select few that provide location landmarks. If the content has a time code that specifies when it was generated then it is possible to interpolate where the content was generated from the landmarks that the user has created. Furthermore, a GLOB may represents a historical event and includes a time code for a future event such as a predictive model of traffic flow where data is generated that represents what is to be expected at a certain location and at a certain time.
  • In the illustrative embodiment, all content that is geo-located will use an additional data structure referred to as a GLOB Data Sheet (hereinafter “GDS”). The GDS supports two main functions, namely, a data structure that holds auxiliary information about the content such as location, and a data structure that facilitates the searching for content based upon certain attributes such as location.
  • Referring to FIG. 3 there is shown illustrative attributes of a GDS. The illustrative GDS is a data record that contains the following pieces of information: location, time when the content was generated, owner of content, access specification for content, type of content, description of content, rating of content, and access permissions. By way of example, the location information is provided as longitude and latitude. An illustrative time period associated with the content may also be provided, e.g. time when was a photograph taken. An illustrative data field may also identify who or what is responsible for generating and/or owns the content, e.g. an individual, an organization, or perhaps even a Web site from which the content originates.
  • Another piece of information that is contained in the GDS may comprise a specification that allows the GLOB content to be accessed, which can be in the form of an identifier that specifies where the content is located as well as a specification that dictates how to access the content from the specified location. Examples of specifications that support this are Uniform Resource Identifiers (URI's) which are well known methods for doing this on the World Wide Web.
  • Additionally, the GDS may comprise a data record having a description of the type of content, i.e. photograph, video, audio, etc. Furthermore, the GDS may comprise some sort of annotation that describes the content where the illustrative annotation is in the form of text, but can take many forms including audio, video, or images. Further still, the GDS may comprise a rating of the content so that higher rated GLOBs are more likely to be included in Chronicles, Narratives, or Virtual Tours, which are described in further detail below; the rating may be derived from a variety of sources including third parties who have viewed the content.
  • GDS records are placed into a geo-spatial database and searched according to any of the GDS attributes mentioned above. Thus, it should be possible to create queries such as “give me all the photographs taken by John Smith along a route specified by the coordinates X, Y . . . Z.” Such queries would return all the GDS records that satisfy the parameters and the GDS records could be then be used to access the content.
  • Referring to FIG. 4 there is shown an illustrative system configured to query GLOB data sheets. The illustrative GLOB content is stored in some location 401, which may be a database or a file system that is managed by servers 402. Furthermore, the illustrative GLOB data sheets' are stored in servers 403. A GDS may be stored in server 403 via a registration request 405. Application computers 404, which need to locate and use specific GLOB content, proceed to submit a GLOB query 408 to GDS server 403. The GDS server 403 would respond with a set of GLOB data sheets that satisfy the query 407. The application computer 404 may then use the GDS to access the GLOBs 406 from the servers 402 that manage the GLOB content. The accessing of the GLOB content may be performed using a variety of technologies including FTP, SQL, and Web Services, to name a few.
  • There are many sources of information that can be used to geocode content to create GLOBs and the corresponding GDS that are used to search and access the GLOB content. A “multi-modal” approach to exploit as many sources of location information as possible is described.
  • As previously discussed, one of the two functions of GDS is for the GDS to control location information. The process of transforming content into a GLOB with a GDS takes advantage of multiple sources of information that determine the content's location. The sources of location information can be categorized as explicit location information, annotations, implicit location information, and association with location logs.
  • Explicit location information is embedded in the content when it is generated and is easily extracted from the content without any processing or transformation of the information. Thus, the explicit location information is stored with the content when the content is generated. By way of example, the device that generated the content has access to location information such as an embedded GPS receiver. In one illustrative embodiment, a digital camera has an embedded GPS receiver that can be used to store latitude and longitude coordinates in the EXIF header of the images taken and stored by the digital camera.
  • Another example of explicit location information would include a camera phone with E-911 capabilities that receives location from a remote server. The remote server can track the location of the phone and wirelessly send that information to the camera phone so that it can use it to tag any photographs that it is taking. It could use the location information to store latitude and longitude coordinates in the EXIF header of the images that it generates and saves.
  • Annotations are added to the content after it is generated and typically require some sort of data entry step that is separate from the actual generation of the content itself. Annotations are created as a separate step, whereas the explicit location information is captured simultaneously with the content. Examples of annotations include text entry and voice recordings. Annotations are typically stored separate from the content, but it may be embedded in the content itself. For example the JPEG image standard allows for some simple annotations to be stored with the JPEG file itself.
  • Annotations may contain information that can be used to determine the location of the content. Examples are text annotations that contain addresses or simply location key words such as “Eiffel Tower” or “Montana.” These location references can be compared to a database of address or location references in order to resolve them to a specific set of geo-spatial coordinates. Databases that contain addresses or well known points of interest that can be used to translate descriptions or references into geo-spatial coordinates are in common use today.
  • It is also possible to perform voice recognition on audio annotations to extract location key words and then cross reference them to a locations database in much the same way that text annotations are resolved.
  • In some cases the annotations are only for the purposes of location and are explicitly added for that purpose. For example it is possible to create map and text based tools for the user to type in addresses and display maps that allow the user to point and click on the map to specify where it was located. The locations that are specified by using the map based tools become location specific annotations to the content. The annotations in this case would be a set of coordinates that were added by the user.
  • Implicit location information is embedded in the content itself and requires some amount of processing to extract. Implicit location information includes information within the content that may be processed to identify where the content was generated. Additionally, any information that can be uniquely matched to something that does have a known location or can be translated into another form such as text can be used to implicitly encode location information by comparing it to a database of known locations.
  • For example if someone takes a picture of the Eiffel tower it is possible to perform image recognition that identifies the Eiffel tower which by implication means that the picture was taken near the Eiffel tower. Thus if the location of the Eiffel Tower is known then the picture's location is also known. The location of the Eiffel Tower may be part of a database that can be used to resolve well known references or landmarks to a specific set of coordinates.
  • Another example of this technique is where the user takes pictures of signs in the area that give information such as city and/or street names. It is possible to perform optical character recognition on such images to extract the location information in the form of text. These location references can then be compared to a database of address or location references in order to resolve them to a specific set of geo-spatial coordinates.
  • A location log is a history of locations that are specific to a person or object. A location log contains a time sequenced series of records with at least the following information: a specification for a location; an identifier for whom or what the location refers to; and a time stamp. Therefore, if there exists a log of locations that specify where some individual as located at particular times, then it is possible to correlate the times that the individual generates content to the locations inside the location log to geocode the content.
  • In a location log, the specification for a location may comprise a variety of units including latitude/longitude or a description of the location such as an address. In short, the location specification is used to determine where on the earth something is located. The location may have a further qualifier that specifies the accuracy and resolution of the location. For example some records may be derived from GPS locations which have a high degree of resolution and accuracy while other records my be derived from auxiliary documents such as travel itineraries which may only have resolution down to the city, but not where in the city.
  • The identifier comprises information related to whom or what the location refers to. Typically, this is an individual where the identifier can simply be the individual's name.
  • The time stamp designates when the above mentioned person or thing was located at the above mentioned location. The time can be specified in a variety of units as long as it can be translated or converted into a date and time. Additionally, the time stamp may have varying degrees of accuracy.
  • The location log takes of advantage of the capabilities of digital recording devices in use today, e.g. digital cameras, which have a means to tag the content with the time. The geocoding step can be automated if a set of location logs exist that specify where the person that is generating the content is located at times that correlate to the time when the content is being generated. In the illustrative instance where the time the content is generated does not match a record in the location log, it is still possible to estimate the location by extrapolating the location from one or two locations in the location log that are close in time.
  • As previously mentioned, location logs are stored in a geo-spatial database that supports the querying of entries either by their location or by their time and identifiers. By way of example and not of limitation, there are a variety of technologies and information sources that can be used to generate location log records. For example, cell phones with embedded GPS receivers or that adhere to the E911 directive may be used, and it is possible for enterprise servers to query and log the location of Cell phones. Additionally, location trackers such as GPS loggers that record the latitude/longitude of the person carrying the device, and TV-GPS and/or Wi-Fi location technologies may be used; these location logs can later be uploaded and added to the user's overall location log. Furthermore, the explicit location information embedded in content as described above may be used where the location information can be extracted from content and added to the user's location log. Further still, the implicit location information derived from content as described above can be extracted from content and added to the user's location log. Further yet, content annotations as described above can be converted to locations that can be added to the user's location log.
  • Other illustrative examples include, but are not limited to, vehicle navigation systems, RFID technology, financial transactions, user schedules, travel itineraries, and Internet Protocol (IP) addresses. Vehicle navigation systems are used to track a vehicle's location and give directions and they use a variety of methods including GPS to track the location of the vehicle; these systems could be used to provide location information that could be loaded into a user's location log. RFID technology such as that used in toll tags in automobiles may also be used. Financial transactions of the user could be mined for information that specifies where the user was located when the transaction occurred. Schedules from a user's calendar or personal information management system can be mined to determine if a location was specified as part of a meeting appointment. Travel itineraries such as those stored on online travel services contain information about where a user was at certain times. Finally, Internet protocol (IP) address location may be used since it is possible to locate where a particular IP address is physically located based upon whom or what organization was assigned that address, so that when a person accesses on line services from some remote location it may be possible to determine where they were when they accessed those services.
  • The general concepts of a GLOB were introduced above. A discussion of GDS followed that described the two GDS functions, namely, a data structure that holds auxiliary information, and a data structure that facilitates searching. The GDS information comprises one or more sources of location information such as explicit location information, annotations, implicit location information, and location logs.
  • Now attention is directed to the software modules that collect geocodes and content from multiple sources, and combines the content and geocodes to produce GLOB data sheets. This process may be described as “multi-stage” geocoding and GDS generation.
  • Referring to FIG. 5 there is shown an illustrative diagram of software modules and processes that generate GLOB data sheets. More particularly, FIG. 5 shows the various functions and modules involved in the process of geocoding content to generate at least one GDS. There exist numerous sources of content 508 including digital cameras. This content is collected and stored within repository 506 and managed using a variety of third party and vendor specific tools 507. The management of the content includes downloading it from the sources and storing 515 the content in some repository 506. In one embodiment, the repository 506 may be local and stored on the user's computer. In another embodiment, the content is uploaded to some third party server on the internet. The repository 506 may be simply a file system on some computer or perhaps a database and access to the repository 506 may be via some server (not shown).
  • Additionally, there exist a number of sources of user location information 501 such as GPS loggers, Cell phones, and personal information as described above within the context of the location logs. Each of these sources is processed at arrow 510. Additionally, the multi-modal location mining function 502 is performed where the location information that is extracted or mined is stored 511 in a location log repository 503. The repository 503 is a geo-spatial database and typically resides on a network that is publicly accessible. The location log database server is able to handle queries 512 in order to select specific log entries that correspond to specific users and time spans.
  • The geocoding process 504 uses all available sources of location information in order to create a GDS for each entry in the content repository 506. Each GDS that is created is stored 513 in a GDS repository 505. As previously described, the GDS contains a reference to the content in the content repository 506 as well as location information about where the content is located. The geocoding process 504 uses multiple stages to resolve the content's location as shown in FIG. 6.
  • Referring to FIG. 6, there is shown an illustrative geocoding process. The first step, 601, is to fetch some content from the content repository 506. Decision diamond 602 is then executed to determine if there is any explicit location information stored with the content. If YES then process 603 is used to extract the explicit location information from the content. Extracting the location information is dependent upon the type of content and how the location information is stored with the content. A variety of formats and content types can be supported.
  • Decision diamond 604 is then executed to determine if the location information that was extracted is adequate. Locations are deemed to be inadequate if their resolution and accuracy do not meet certain minimum standards or if there simply is no location record. If decision diamond 604 is YES (i.e. location data is adequate) then it is not necessary to do any more geocode processing to resolve the content's location and process 616 is executed. Process 616 adds location information to the location log 503 if it does not already exist. This is performed so that the location information can be used in the future to geocoding other content. Process 617 is then executed and uses the location information to generate a GDS and add it to the GDS repository 513 and geocoding is complete.
  • If decision diamond 602 was NO (i.e. no explicit location information in content) or decision diamond 604 is NO (i.e. locations inadequate) then further geocoding needs to take place and decision diamond 605 is executed to see if there are any content annotations. If YES then the annotations are processed by process 606 to extract any location information from the annotations and decision diamond 607 is executed to see if these locations are adequate. Decision diamond 607 is then executed to see if the locations are adequate (analogous to 604 as described above) and if the answer is YES then process blocks 616 and 617 are executed as described above and geocoding is complete.
  • If decision diamond 605 was NO (i.e. no content annotations) or decision diamond 607 is NO (i.e. locations inadequate) then further geocoding needs to take place and decision diamond 608 is executed to determine if there are any locations in the log that correspond to the content. This decision is based upon the owner of the content and the time period that the content was generated. If there are records in the location log that correspond to the content (i.e. have the same owner and similar time stamps) then the answer to decision diamond 608 is YES and process 609 is executed. Process 609 uses the records from the location logs to calculate locations for the content. This may require extrapolation of the locations from the location log depending upon how closely the time stamp of the content matches the time stamps of the location log records. Decision diamond 610 is then executed to see if the locations are adequate (analogous to 604 as described above) and if the answer is YES then process' 616 and 617 are executed as described above and geocoding is complete.
  • If decision diamond 608 was NO (i.e. no corresponding records in the location log) or decision diamond 610 is NO (i.e. locations inadequate) then further geocoding needs to take place and process 611 is executed to extract any implicit location information from the content itself. Decision diamond 612 is then executed to see if the locations are adequate (analogous to 604 as described above) and if the answer is YES then process' 616 and 617 are executed as described above and geocoding is complete.
  • If decision diamond 612 is NO (i.e. locations inadequate) then a further processing step 613 is performed to try and combine all the locations from the previous stages together to see if the location can be further resolved. One example of this step would be to simply take the intersection of all the previous locations as see if it results in a more refined location. Decision diamond 614 is then executed to see if the locations are adequate (analogous to 604 as described above) and if the answer is YES then process' 616 and 617 are executed as described above and geocoding is complete.
  • If decision diamond 614 is NO (i.e. locations inadequate) then process 615 is executed. Process 615 is a final stage in which various graphical map and text based tools are used to allow the user to specify addresses and point and click on a map to manually specify the contents location. After this step it is assumed that the location has been specified to the level of accuracy and resolution necessary and process' 616 and 617 are executed to generate the GDS and place it in the GDS repository 505 and geocoding is complete.
  • In summary, the algorithm associated with FIG. 6 describes six distinct location resolution stages: explicit locations in content, i.e. process blocks 602, 603, and 604; content annotations, i.e. process blocks 605, 606, and 607; location log association, i.e. process blocks 608, 609, and 610; implicit location information, i.e. process blocks 611, and 612; combine all previous results, i.e. process blocks 613, and 614; and tools to allow the user to manually annotate the content with location, i.e. process block 615.
  • It shall be appreciated by those of ordinary skill in the art that the last stage is the only one which is not completely automated and may require the user to perform some manual task, and is only required if none of the previous stages are successful in generating an adequate location. Those skilled in the art shall also appreciate that the algorithm is not dependent upon the order in which stages 1 through 4 are executed. Thus, if it is found that it is more efficient to perform stage 3, i.e. check the location logs, before performing stage 2, i.e. extract location information from the content annotations, then process blocks 610, 611, and 612 may be performed before process blocks 605, 606 and 607. This algorithm is in fact designed so that stages 1-4 can be re-ordered or even omitted to increase efficiency.
  • Up to this point a description of GLOBs, GLOB Data Sheets (GDS), and the process for geo-coding content to generate at least one GDS has been described. The resulting GDS, which is described above, may be used to organize the GLOB content as shown in FIG. 7.
  • Referring to FIG. 7 there is shown an illustrative example of a “Chronicle” or “Narrative” of GLOBs that are displayed to a user. Note, for purposes of this patent the terms Chronicle and Narrative are used interchangeably. A Chronicle or Narrative is a collection of GLOBs that are organized according to time and location, and may then be animated. The Chronicle is a multi-resolutional depiction of the GLOB data with respect to time and location. The illustrative Chronicle comprises a map display where the time sequenced GLOB content is displayed in order to show the geographic location that the content refers to. A synopsis of the GLOB content as it changes over time can also be shown in the illustrative Chronicle.
  • The Chronicle has a notion of time that is displayed to the user 709. This notion of time is in relation to the GLOB content that is being animated, i.e. the pictures 701. In the illustrative example, the time being displayed corresponds to the time at which the pictures were taken. The illustrative images or pictures are animated in the form of a slide show that is time sequenced. By way of example, the pictures are being animated by thumbnails that are scrolling smoothly across the screen. For example, there are multiple levels of map display 705, 706, 707 and 710. Each map shows a different level of detail ranging from the entire planet 705 to the local area 710. It shall be appreciated by those of ordinary skill in the art that the multiple levels are not required as long as there is at least one level. It shall also be appreciated that the illustrative maps include maps generated using aerial imagery, or computer generated street maps, or any combination thereof.
  • The illustrative maps show the location of the images that are currently being displayed in the slide show 701. In the case of the highest resolution map 710, the location of the GLOB pictures are shown by both markers on the map 703 and pointers 702 which point to those markers. This allows the user to easily see where the GLOBs are geographically located.
  • In addition to the picture GLOB animation there is a route GLOB animation being drawn on the map 704. A route is simply a time sequenced collection of geo-locations and can be animated by drawing the route on the map as the Chronicle's time elapses. Note, the route could represent a route that was taken when the pictures were generated or could represent a route that is to be followed in the future and the pictures represent what is expected to be seen along the way. Therefore, a Chronicle can either represent something that is either currently happening, has happened in the past, or is expected to happen.
  • The Chronicle may comprise a set of player controls 708 that allow the user to control various aspects of the Chronicle such as playing/pausing, forward/reverse and a method to quickly move to certain time indices of the Chronicle that is similar to a videotape player.
  • The illustrative Chronicle may animate a synopsis of GLOB sequences in order to allow the user to quickly and easily view the information represented by the GLOB. Additionally, the user may “drill down” and view more detail associated with a particular GLOB, which may entail getting a more detailed view of a particular picture. The user could do this by simply selecting the thumbnail of the picture and then the Chronicle would pause and a more detailed view of the picture may be displayed. An illustrative example of this is shown in FIG. 8, which displays a more detailed view of the selected GLOB.
  • The Chronicle is generated using the attributes and features associated with GLOB Data Sheets, which were described above. However, the method for generating Chronicles may require processing third party content and including this third party content in the Chronicle.
  • Referring to FIG. 9 there is shown the process of generating a Chronicle or Narrative that integrates third party information. The process for generating a Chronicle involves combining GLOB data that has been generated with maps and other third party content. Initially there exist a set of data content that is generated for the purpose of using as a GLOB 901. Note, when the data is generated it may or may not be geocoded. There also exists various third party content such as Maps, documents, etc., which may be used in conjunction with the generated GLOB data to create a Chronicle.
  • The first process step 903 is for the generated date to be collected and turned into GLOBs by adding geo-location information if necessary. Process step 903 may also entail annotating the GLOBs with additional information such as textual description of the content of the GLOB. Finally the information is indexed and archived in some fashion so that it may be retrieved later. This may entail the use of well known database technologies. The result of this is a set of GLOB data sets 905 that can bused for the Chronicle generation.
  • The actual Chronicle generation process 907 entails using a Chronicle specification 909 to collect all the data, namely, both GLOBs 905 and third party date 906. The Chronicle specification may comprise the type of GLOB data desired, the location, the time frame, and source of the GLOB data. The collection of the third party data may interface with various map servers and other databases and possibly Internet based information servers 904 to collect all the information that is relevant to each GLOB and its associated content 906. The information that is retrieved is dependent upon the specification of the Chronicle. Note, a Chronicle specification may refer to the past, present or future time frames.
  • Once the data sets 905 and 906 have been collected, a graphical representation is generated where the GLOB data is animated in time and shown geographically, as described above. The manner in which the Chronicle is generated and displayed is discussed in more detail below.
  • Note, the Chronicle or Narrative specifications 909 can specify various types of information such as the following: show traffic and roadside images for a specific route taken at a specific time; show a photographic travel log of a trip; shows the states of various sensors located in a particular region; show additional third party information such as weather; or any combination thereof.
  • Once a Chronicle is generated it may be stored in a variety of formats including the following: source GLOB data and Chronicle specifications that allows the Chronicle either be regenerated or viewed; video files on servers or computers that can be played in a variety of ways; video files on local store media such as DVD's, hard disks, flash memory, or other such storage media.
  • Additionally, Chronicles can be stored with only synopsis information, e.g. skinny Chronicle as described blow, where the detailed GLOB information is stored separately and accessed on an as needed basis. Chronicles which have been generated can be viewed in a number of different ways including the following: using a computer like PC or possible even a mobile device that runs a special Chronicle viewer application and uses the specification and GLOB information in the Chronicle to generate and display it; using any sort of digital video capable device such as computer or mobile device that can play video files that are stored either locally or remotely on the computer; using a television and DVD player where there is an interactive video of the Chronicle stored on some type of local media such as DVD or video recorder; using a network, e.g. Internet, capable device that is capable of receiving and displaying streaming video. Illustrative examples of such devices include, but are not limited to, desktop computers, mobile PDAs, navigation devices, cell phones, cable television, and similar network capable devices.
  • Referring to FIG. 10 there is shown a general overall architecture for Chronicle generation. As shown in FIG. 10, the generation of Chronicles requires gathering GLOB information from various sources 1001, 1003, 1005, and 1007 that are served up from a variety of servers 1002, 1004, and 1006 that may be distributed and interfaced via some sort of wide area network 1012 such as the Internet and the World Wide Web. The vehicle telemetry server 1009 is an example of a server that continuously collects data from a potentially large number of sources (vehicle probes 1007) and stores the information in a database 1008. It shall be appreciated by those skilled in the art that servers such as 1008 and 1009 do not necessarily apply only to vehicles, but could apply to any data sources which may be numerous and continuously generate information such as sensors that are part of some large sensor network.
  • Chronicles or Narrative may be viewed by a number of different devices 1013, 1014, 1015, 1016. These devices may be normal desktop systems 1014, mobile devices 1615 such as PDA's or cell phones, special embedded systems such as navigation or vehicle information systems inside vehicles 1013 or even video display systems such as televisions 1016. Like the GLOB sources, these devices are network via some sort of WAN 1012 that includes such systems as cable television.
  • FIG. 10 also shows a server 1010 which aggregates information from the various GLOB sources and both generates the Chronicle and serves them to the devices used to view the Chronicles. Typically, the Chronicle server 1010 will store the generated Chronicles in a database 1011.
  • Referring to FIG. 11-15, there is shown a plurality of different system configurations for the various servers, GLOB sources, and Chronicle viewing devices that are depicted in FIG. 10. These system configurations are provided as illustrative examples and those skilled in the art shall appreciate that there may be different combinations that perform the same of similar functions.
  • Referring to FIG. 11 there is shown the various components of the Chronicle or Narrative presentation systems. More particularly, FIG. 11 shows the various components that are used to describe the various scenarios. There exist third party content datasets 1101 that contain GLOB information such as maps and other readily available information such as traffic reports, weather, news, and other such information that may come from various third party vendors or sources. This information may be stored in traditional databases or simple files and made available through standard interfaces such as web services.
  • There also exist user generated GLOB content and potentially location logs 1102. This data may also be stored either in files or in a database of some sort.
  • There exist a variety of user devices 1107 which represent the various platforms that the end user will use to view the generated Chronicles. In some scenarios, these same devices may also store the GLOB content and generate the Chronicles.
  • There also exist a number of remote servers 1105 that are used to either serve up GLOB content or generate and serve Chronicles.
  • There exist so called “Fat Chronicles” 1103 that provide fully self-contained Chronicles in which all the GLOB content of the Chronicle is stored with the Chronicle. Fat chronicles do not require additional information, but are very large and thus more difficult to move around over WAN's.
  • There also exist so called “Skinny Chronicles” 1104 that contain only enough content for the previewing of the GLOB data, but not the detailed GLOB information. Skinny Chronicles are much smaller and easier to move around, but require that detailed content be served up on an as needed basis.
  • FIGS. 12-15 describe a number of system architectures that may be used to manage the GLOB content and generate the Chronicles. Each of these scenarios may be used depending on system design and limitations. The main differences between the various scenarios revolve around the following design parameters: how much GLOB content is managed and stored locally versus remotely; and how much of Chronicle is generated locally versus remotely.
  • Referring to FIG. 12 there is shown an illustrative Chronicle architecture with fat servers and thin clients. In this architecture all the GLOB content is managed remotely. This may entail users to transfer any GLOB content that they have generated to the remote servers. Additionally, the Chronicles are generated by remote servers and served up to the user's computers. The Chronicles may be served up as simple web based application or there may exist as a custom application on the user's computers that allow the Chronicle to be viewed. In FIG. 12, the Chronicles are accessible from anywhere and thus easier to share among various users. Additionally, the Chronicle may be served up as a web application thus the user utilizes a browser based interface, thus no custom program is needed to view the Chronicles.
  • Referring to FIG. 13 there is shown an illustrative Chronicle architecture with no remote servers. In FIG. 13 all the GLOB content and Chronicles are generated and managed locally on the users computers so that everything is self-contained and so there is no Wide Area Network (WAN) required and no remote servers are required.
  • Referring to FIG. 14 there is shown an illustrative Chronicle architecture with remote servers having generated skinny chronicles. In FIG. 14, the remote server serves up skinny Chronicles while the locally generated GLOB content is stored and managed on the user's local computer, and there is less of a burden to upload all locally generated GLOB content. All that needs to be uploaded are the GLOB synopsis. There is less storage required on the remote servers.
  • Referring to FIG. 15 there is shown an illustrative chronicle architecture where the server serves remote content and clients generate the chronicle. In FIG. 15, the remote servers serve up remote GLOB content and user's computers manage locally generated GLOB content and generate the Chronicles. This scenario does not require transferring all locally generated GLOB content to a remote server, and there is less storage required on the remote servers.
  • Up to this point, a description of GLOBs and GLOB Data Sheets has been provided. Additionally, attention has been directed to the software modules that collect geocodes and content from multiple sources, and combines the content and geocodes to produce GLOB Data Sheets (GDS), which may be described as “multi-stage” geocoding and GDS generation. Subsequently, an illustrative example of a “Chronicle” or “Narrative” of GLOBs that are displayed to a user was discussed. Furthermore, an illustrative process for generating a Chronicle or Narrative, which integrates third party information, was also described. A plurality of different system architecture for Chronicle generation was also presented.
  • Since there are many sources of information that can be used to geocode content to create GLOBs, a method for implicitly calculating location based on limited explicit location information is described below. This method is also referred to as geocoding groups of sequenced object wherein location information is used, regardless of its source, to geocode groups of sequenced objects. The method requires that there exist enough location information to individually geocode each object in the group and does not rely on the fact that the objects in the group are sequenced. In operation, the method describes explicitly geocoding a subset of the objects in a group and then implicitly calculating the locations of the other related objects in the group by exploiting their relationship to the explicit location information for one or more GLOBs.
  • In one illustrative embodiment, the first and simplest way to geo-locate a group of related objects is to group them all together and assign the same geocode to them all. Therefore, when one of the objects is geo-located all the objects can be assigned the same geocode. For example, there may be a collection of photographs that all have the same annotation of “May field trip” and are thus related in this fashion. If any of the photographs in that collection can be geo-located, then by using the associated annotation of “May field trip” all the photographs can be located in the same place. Although this may be adequate for many applications, it does not allow for the differentiation by location of the individual objects in the group.
  • Additionally, when the objects in the group are sequenced it is possible to exploit their sequential relationship to extrapolate their locations when one or more objects in the group are explicitly located.
  • For example, in FIG. 16 there is shown a group of related objects such as photographs 2301, 2302, 2303, 2304, and 1205 that have specific time stamps. As shown in FIG. 16, it is possible to order the pictures into a sequence based upon their time stamps. In FIG. 16, they are ordered from left to right according to their time stamps as indicated. FIG. 17 shows the display of an example graphical based tool that some user can use to geocode the photographs from FIG. 16. Such a tool is a computer program that could be a stand alone program or a web based application that runs within an internet browser. It contains a map display 2406 that the user can use to explicitly geocode photographs. An example of how this could be done would be for the user to select a photograph and then click on the map where the photograph is located. Another illustrative method would be to enter a textual address which gets resolved to a location on the map. These are just two examples of how to explicitly geocode a photograph, but in fact any of the location sources mentioned above may be used for this purpose.
  • FIG. 17 shows the location on the map where photographs 2301 and 2305 were explicitly located as described above, and this is depicted by locations 2401 and 2405, respectively. Between points 2401 and 2405 there exists an implicit linear path 2407 that can be used to implicitly geocode the remaining photographs 2302, 2303, and 2304. Assuming the remaining photographs lie somewhere on this path we can use the time stamps of the photographs and their relationships to the explicitly located photographs to determine their location on the path 2407. FIG. 17 shows implicit locations 2402, 2403, and 2404 that lie on path 2407 and whose relative locations are a linear interpolation along the path based upon time. For example, implicit location 2403 lies on the mid point of path 2407 because the time stamp of photograph 2303 is midway between the time stamp of 2301 and 2305 which form the endpoints of path 2407.
  • To further illustrate this process, FIG. 18 shows a case where photographs 2301, 2303, and 2305 have been explicitly located to locations 2401, 2503, and 2405, respectively. In this instance, there are two paths generated. There exists path 2506 between 2401 and 2503 and path 2507 between 2503 and 2405. In FIG. 18, location 2502 is implicitly calculated as a linear interpolation based upon time along path 2506 and likewise location 2504 is implicitly calculated as a linear interpolation along path 2507 based upon time.
  • In general, sorting the list of related objects according to time then each successive pair of explicitly located objects in the list determine a path along which the other objects may be implicitly located by extrapolating their locations along the paths formed by the two objects that are greater than and less than they are according to time.
  • The examples above used a linear path between successive explicitly located objects, but it is possible to use other types of paths and extrapolate along them. For example, if the actual path taken by the person generating the content (i.e. taking the pictures) is known it is possible to extrapolate along that path instead of using a linear path. For example, it might be known that the person taking the pictures took a certain route through a street network in which case that route should be used to extrapolate the implicit locations. Such a route might be defined by a sorted list of line segments.
  • Furthermore the examples above did a linear interpolation along the path based upon time, but it is possible to use alternative distributions and/or rules that are not linear. For example, it is possible to use a rule wherein all pictures that are taken within 10 minutes of an explicitly located photograph are assumed to be in the same location.
  • FIG. 19 shows examples of using non-linear paths and non-linear time interpolation along those paths. In FIG. 6 photographs 2301 and 2305 have been explicitly located at locations 2601 and 2605. In addition, it is known that during this time the person traveled along route 2606. Furthermore the implicit locations 2602, 2603, and 2604 are extrapolated along path 2607 using the rule that all photographs taken within 10 minutes of an explicitly located photograph are assumed to be in the same location as the explicitly located photographs.
  • It is important to note that the general method described above relies on the fact that the objects are sequenced and not upon the fact that they have time stamps. In the examples above the time stamps of the objects were used to create the sequence and to also extrapolate locations along paths, but it is possible to perform similar operations even if the objects did not contain time stamps. For example, it is possible for a user to create a sequence of pictures that are based upon personal choice and not upon any time stamps that the picture has. In this case it is still possible to explicitly locate a subset of the pictures and implicitly generate locations for the other objects.
  • FIG. 20 shows an example of geocoding pictures that do not have time stamps, but are still part of a user defined sequence. In the illustrative example of FIG. 20, the pictures are related by the sequence, but they do not have any additional information (such as time) that allows the location to be extrapolated other than by distance along the path. Thus the implicitly located photographs 2702, 2703, and 2704 are evenly spread between the explicitly located photographs 2701 and 2705 based upon their distance along the path 2707 from 2701 to 2705.
  • If there does exist additional information that allows the photographs to be further related then it is possible to use this information to further refine their location extrapolation. FIG. 21 provides an illustrative example where the user has specified that photographs 2801, 2802, and 2803 were taken in the same place. In FIG. 21, photographs 2801 and 2805 are explicitly located and photographs 2802, 2803, and 2804 are implicitly located along path 2807. Since photographs 2802 and 2803 were related to 2801 to be in the same place they are located in the same place as 2801. Since 2804 does not contain any additional relation information it is located midway between 2801 and 2805 along the path 2807.
  • It shall be appreciated by those of ordinary skill in the art having the benefit of this disclosure that a number of different paths and extrapolations rules and/or equations could be used to generate implicit locations. It shall also be appreciated by those of ordinary skill in the art that the process of explicitly geocoding some subset of the photographs could use any of the sources of location information described above and that in principle the implicit geocoding process would still be the same. Thus as long as there is some means to explicitly geocode some subset of related objects the remainder of the objects can be implicitly geocoded by generating paths between the explicitly geocoded objects and extrapolating locations along those paths based upon time.
  • In the illustrative example of a Chronicle or Narrative having a plurality of GLOBs that are displayed to a user, which is described above, the process of data collection and animation is not automated. An animated system and method for viewing a Chronicle, which is referred to as a “Virtual Tour”, is described below.
  • A “Virtual Tour” is a visual and/or audio presentation of what one might experience if they were to visit some location. This presentation is typically rendered on some sort of electronic display or screen that is connected to a computing device such as a desktop or handheld computer.
  • Referring to FIG. 22 there is shown an illustrative Virtual Tour. The Virtual Tour shows one or more maps at various levels of scale 3105, 3106, 3107, and 3110 that clearly show to the user the geographic location of the tour. Although multiple scale levels are shown, only one scale level is required. The maps may be from aerial imagery, computer generated street maps, or any combination thereof.
  • The Virtual Tour contains images in the form of pictures and/or videos that show what the user might see and/or hear if they were to actually go to the locations indicated in the map. The images are being animated in the form of a slide show that is time sequenced. The pictures are being animated by thumbnails that are scrolling smoothly across the screen.
  • The maps show the location of the images that are currently being displayed in the slide show 3101. In the case of the highest resolution map 3110 the location of the images are shown by both markers on the map 3103 and pointers 3102, which point to those markers. This allows the user to easily see where the images are geographically located and what they might expect to see if they were at those locations.
  • In addition to the image animations there may be a route being drawn on the map 3104. A route is simply a time sequenced collection of geo-locations and can be animated by drawing the route on the map as the images scroll by (i.e. time elapses). It is important to note that the route could represent a specific tour and correspond to the sequence of images being shown.
  • The Virtual Tour may have a notion of time that is displayed to the user 3109. This notion of time is in relation to the content that is being animated, i.e. the images 3101. In this example the time being displayed corresponds to the time at which a proposed tour may be taken.
  • The virtual tour may contain a set of player controls 3108 that allow the user to control various aspects of the tour presentation such as playing/pausing, forward/reverse and a method to quickly move to certain time indices of the Virtual Tour. This is much like one might find on a Video player. In addition there may be keys or controls which allow the user to pan and zoom the various maps.
  • The Virtual Tour may contain icons or images superimposed on the map 3111 and 3112, which represent “Points of Interest” (POI). These points of interest may be businesses such as hotels and gas stations or they may be places which have some attraction such as a museum or a tennis club. Businesses such as hotels or restaurants may be sponsored by the business establishment and any images or information may be provided by the business itself. This might even include promotional items such as discounts or coupons.
  • As shown in FIG. 23, if the user selects one of the POI icons it may bring up additional information or even another Virtual Tour that is associated with the business or location represented by the icon.
  • Virtual Tours animate a synopsis of image sequences in order to allow the user to quickly and easily view the images and their locations. If the user so desired they could “drill down” and view more detail associated with a particular image or location. In the illustrative example provided in FIG. 22 this might entail getting a more detailed view of a picture. The user could do this by simply selecting the thumbnail of the picture and then the Virtual Tour would pause and a more detailed view of the picture displayed, as shown in FIG. 24.
  • The illustrative Virtual Tour is generated automatically by a remotely located Virtual Tour Server in response to a user specifying the type of tour they are interested in seeing.
  • FIG. 25 shows a Virtual Tour Server 3404 that creates a Virtual Tour that can be viewed on a variety of user viewing devices 3405 including mobile devices, desktop computers, video monitors (TV's), and in-vehicle systems such as navigation systems. The Virtual Tour Server 3404 is accessed by the user via some sort of wide area network such as the Internet 3403. The user specifies the type of Virtual Tour they want to view by indicating one or more factors that they are interested in such as: location of the tour such as a city or country; route to be taken; type of activities desired such as surfing, and other well-known activities; type of attractions, e.g. museums; specific events; date and time of the tour; and new tour based upon previous Virtual Tour.
  • A Virtual Tour is generated by the Virtual Tour Server by bringing together and displaying GLOBs such as images, maps and POI data 3401 from a variety of sources. This GLOB content is accessed via GLOB servers 3402 that are accessible via a wide area network 3403 such as the internet. The Virtual Tour Server combines all the GLOB content together and presents it to the user in a manner as described previously.
  • FIG. 26 shows is a more detailed depiction of the full architecture where there exists a Virtual Tour Server that generates the Virtual Tours. Users access the Virtual Tour Server from a variety of devices such as vehicle information systems 3813, desktop computers 3814, mobile devices 3815, and streaming video displays 3816. The Virtual Tour Server may be accessed over a variety of public networks 3812 such as the Internet.
  • Upon receiving a request from a user for a Virtual Tour the Virtual Tour Server collects a variety of GLOB content from various sources and dynamically generates the Virtual Tour. As previously discussed there exists a variety of GLOB content such as maps 3801, third party content 3803, user generated content 3805, and POI data and advertisements 3817. This content is managed by a variety of servers 3802, 3804, 3806, and 3807 and accessed via a network 3812. The Virtual Tour Server 3810 locates the GLOB content by submitting requests to the GLOB Resource Server 3809 and 3808. Upon getting a collection of GLOB Data Sheets from the GLOB Resource Server, the Virtual Tour Server accesses the GLOB content and uses it to produce a Virtual Tour. The Virtual Tour is then served up to the user that made the request for the Virtual Tour.
  • It is important to note that in some case it may not be necessary to dynamically generate a Virtual Tour if one satisfying the request has previously been generated and is stored in a Virtual Tour database 3811. In this case the Virtual Tour Server need only access and serve up the Virtual Tour without generating it. Such pre-produced Virtual Tours may exist as the result of users generating and storing their own Virtual Tours or perhaps as the result of a business related to a POI generating one to help promote their business. For example, a hotel may produce a Virtual Tour that can be shown to potential customers that are interested in staying at their hotel.
  • It is to be understood that the foregoing is a detailed description of illustrative embodiments. The scope of the claims is not limited to these specific embodiments. Various elements, details, execution of any methods, and uses can differ from those just described, or be expanded on or implemented using technologies not yet commercially viable, and yet still be within the inventive concepts of the present disclosure. The scope of the invention is determined by the following claims and their legal equivalents.

Claims (29)

1-36. (canceled)
37. A system for presenting user generated digital information, comprising:
a plurality of geo-located objects (GLOBs), wherein each GLOB among the plurality of GLOBs includes a location component associated with a digital information object;
a plurality of GLOB Data Sheets (GDSs), wherein each GDS among the plurality of GDSs is associated with a GLOB among the plurality of GLOBs, wherein each GDS includes one or more data fields storing auxiliary information associated with the GLOB;
a grouping of GLOBs organized as a function of at least one of time and location;
an implicit local module organizing the grouping of GLOBs based on the location of at least one GLOB that has a location defined; and
a display device configured to present the organized grouping of GLOBs.
38. The system as recited in claim 37, wherein a type of the digital information object includes at least one of a picture, a video, an audio, a sensor measurement, a map, a document, a recorded location log, an automotive telemetry, a weather report, and a news item.
39. The system as recited in claim 37, further comprising an input system enabling the user to specify the location of one or more GLOBs among the plurality of GLOBs.
40. The system as recited in claim 37, wherein the display device is further configured to present a map, wherein the input system enables the user to click on the map to specify the location of the one or more GLOBs among the plurality of GLOBs.
41. The system as recited in claim 37, further comprising a storage unit receiving a plurality of third party content from a remote server, wherein each GLOB, each GDS, and the grouping of GLOBs are generated using the plurality of third party content.
42. The system as recited in claim 37, further comprising a storage unit that is remote from the system, wherein at least one of the GDS and the plurality of GLOBs are stored in the storage unit.
43. The system as recited in claim 37, wherein the one or more auxiliary data fields are searchable.
44. The system as recited in claim 37, wherein the display device is further configured to display the auxiliary information extracted from the GDS.
45. The system as recited in claim 37 further comprising,
an input system enabling the user to specify the location of one or more GLOBs among the plurality of GLOBs;
wherein the display device is further configured to present a map, wherein the input system enables the user to click on the map to specify the location of the one or more GLOBs among the plurality of GLOBs;
wherein the one or more auxiliary data fields are searchable; and
wherein the display device is further configured to display the auxiliary information extracted from the GDS.
46. The system as recited in claim 45, wherein a type of the digital information object includes at least one of a picture, a video, an audio, a sensor measurement, a map, a document, a recorded location log, an automotive telemetry, a weather report, and a news item.
47. A system for presenting user generated digital information, comprising:
a plurality of geo-located objects (GLOBs), wherein each GLOB among the plurality of GLOBs includes a location component associated with a digital information object;
a plurality of GLOB Data Sheets (GDS), wherein each GDS among the plurality of GDSs is associated with a GLOB among the plurality of GLOBs, each GDS configured to include one or more data fields storing auxiliary information associated with the GLOB;
a means for generating a grouping of GLOBs organized as a function of at least one of time and location;
a means for organizing the grouping of GLOBs based on the location of at least one GLOB that has a location defined; and
a means for displaying the organized grouping of GLOBs on a display device.
48. The system as recited in claim 47, wherein a type of the digital information object includes at least one of a picture, a video, an audio, a sensor measurement, a map, a document, a recorded location log, an automotive telemetry, a weather report, and a news item.
49. The system as recited in claim 47, further comprising input means for enabling the user to specify the location of one or more GLOBs among the plurality of GLOBs.
50. The system as recited in claim 49, further comprising means for displaying a map on the display device, wherein the input means enables the user to click on the map to specify the location of the one or more GLOBs among the plurality of GLOBs.
51. The system as recited in claim 47, further comprising a storage unit receiving a plurality of third party content from a remote server, wherein each GLOB, each GDS, and the grouping of GLOBs are generated using the plurality of third party content.
52. The system as recited in claim 47, further comprising a storage unit that is remote from the system, wherein at least one of the GDS and the plurality of GLOBs are stored in the storage unit.
53. The system as recited in claim 47, wherein the one or more auxiliary data fields are searchable.
54. The system as recited in claim 47, wherein the means for displaying includes means for displaying the auxiliary information extracted from a GDS among the plurality of GDSs.
55. A method for presenting user generated digital information, comprising the steps of:
receiving a plurality of geo-located objects (GLOBs), wherein each GLOB includes a location associated with a digital information object;
generating a plurality of GLOB Data Sheets (GDSs), wherein each GDS among the plurality of GDSs is associated with a GLOB among the plurality of GLOBs, each GDS including one or more auxiliary data fields storing auxiliary information associated with the GLOB;
organizing a grouping of GLOBs based on at least one of time and location, wherein organizing based on location includes an implicit local module organizing one or more GLOBs with missing location based on the location of at least one GLOB with a location defined; and
displaying the organized grouping of GLOBs on a display device.
56. The method as recited in claim 55, wherein a type of the digital information object includes at least one of a picture, a video, an audio, a sensor measurement, a map, a document, a recorded location log, an automotive telemetry, a weather report, and a news item.
57. The method as recited in claim 55, further comprising the step of enabling a user to specify the location of one or more GLOBs among the plurality of GLOBs.
58. The method as recited in claim 55, further comprising the steps of:
displaying a map on the display device; and
enabling a user to click on the map to specify the location of one or more GLOBs among the plurality of GLOBs.
59. The method as recited in claim 55, further comprising the step of receiving a plurality of third party content, wherein the plurality of GDSs and the grouping of GLOBs are generated using the plurality of third party content.
60. The method as recited in claim 55, wherein the one or more auxiliary data fields are searchable.
61. The method as recited in claim 55, further comprising the step of displaying on the display device the auxiliary information extracted from a GDS among the plurality of GDSs.
62. The method as recited in claim 55, wherein the step of receiving the plurality of GLOBs includes the steps of:
collecting a plurality of digital information objects; and
associating one or more objects among the plurality of digital information objects with one or more locations among a plurality of locations, resulting in the plurality of GLOBs.
63. The method as recited in claim 55 further comprising,
enabling a user to specify the location of one or more GLOBs among the plurality of GLOBs;
displaying a map on the display device;
enabling a user to click on the map to specify the location of one or more GLOBs among the plurality of GLOBs;
enabling at least one auxiliary data fields to be searchable;
displaying on the display device the auxiliary information extracted from a GDS among the plurality of GDSs.
64. The method as recited in claim 63, wherein a type of the digital information object includes at least one of a picture, a video, an audio, a sensor measurement, a map, a document, a recorded location log, an automotive telemetry, a weather report, and a news item.
US13/599,884 2005-12-23 2012-08-30 System and method for presenting user generated digital information Abandoned US20120320060A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US13/599,884 US20120320060A1 (en) 2006-02-21 2012-08-30 System and method for presenting user generated digital information
US14/539,263 US20150142806A1 (en) 2006-02-21 2014-11-12 System and method for presenting user generated digital information
US15/175,374 US10488860B1 (en) 2006-02-21 2016-06-07 Geocoding data for an automated vehicle
US16/694,957 US11415986B2 (en) 2005-12-23 2019-11-25 Geocoding data for an automated vehicle

Applications Claiming Priority (8)

Application Number Priority Date Filing Date Title
US77493106P 2006-02-21 2006-02-21
US79221106P 2006-04-17 2006-04-17
US79392806P 2006-04-24 2006-04-24
US11/645,001 US7904483B2 (en) 2005-12-23 2006-12-23 System and method for presenting geo-located objects
US12/571,547 US7917543B2 (en) 2006-02-21 2009-10-01 System and method for geo-coding user generated content
US13/014,600 US8200712B2 (en) 2006-02-21 2011-01-26 System and method for generating a virtual tour on a display device
US13/489,076 US8447787B2 (en) 2006-02-21 2012-06-05 System and method for geocoding content
US13/599,884 US20120320060A1 (en) 2006-02-21 2012-08-30 System and method for presenting user generated digital information

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/489,076 Continuation US8447787B2 (en) 2005-12-23 2012-06-05 System and method for geocoding content

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/539,263 Continuation US20150142806A1 (en) 2005-12-23 2014-11-12 System and method for presenting user generated digital information

Publications (1)

Publication Number Publication Date
US20120320060A1 true US20120320060A1 (en) 2012-12-20

Family

ID=38605873

Family Applications (7)

Application Number Title Priority Date Filing Date
US11/709,330 Active 2027-01-16 US7617246B2 (en) 2005-12-23 2007-02-20 System and method for geo-coding user generated content
US12/571,547 Active US7917543B2 (en) 2005-12-23 2009-10-01 System and method for geo-coding user generated content
US13/014,600 Active US8200712B2 (en) 2005-12-23 2011-01-26 System and method for generating a virtual tour on a display device
US13/489,076 Active US8447787B2 (en) 2005-12-23 2012-06-05 System and method for geocoding content
US13/599,884 Abandoned US20120320060A1 (en) 2005-12-23 2012-08-30 System and method for presenting user generated digital information
US13/599,827 Active US8447792B2 (en) 2006-02-21 2012-08-30 System and method for presenting user generated geo-located objects
US14/539,263 Abandoned US20150142806A1 (en) 2005-12-23 2014-11-12 System and method for presenting user generated digital information

Family Applications Before (4)

Application Number Title Priority Date Filing Date
US11/709,330 Active 2027-01-16 US7617246B2 (en) 2005-12-23 2007-02-20 System and method for geo-coding user generated content
US12/571,547 Active US7917543B2 (en) 2005-12-23 2009-10-01 System and method for geo-coding user generated content
US13/014,600 Active US8200712B2 (en) 2005-12-23 2011-01-26 System and method for generating a virtual tour on a display device
US13/489,076 Active US8447787B2 (en) 2005-12-23 2012-06-05 System and method for geocoding content

Family Applications After (2)

Application Number Title Priority Date Filing Date
US13/599,827 Active US8447792B2 (en) 2006-02-21 2012-08-30 System and method for presenting user generated geo-located objects
US14/539,263 Abandoned US20150142806A1 (en) 2005-12-23 2014-11-12 System and method for presenting user generated digital information

Country Status (1)

Country Link
US (7) US7617246B2 (en)

Families Citing this family (121)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8942483B2 (en) 2009-09-14 2015-01-27 Trimble Navigation Limited Image-based georeferencing
US8712192B2 (en) * 2006-04-20 2014-04-29 Microsoft Corporation Geo-coding images
US7945852B1 (en) * 2006-05-19 2011-05-17 Washington State University Research Foundation Strategies for annotating digital maps
US7945582B2 (en) 2006-09-23 2011-05-17 Gis Planning, Inc. Web-based interactive geographic information systems mapping analysis and methods of using thereof
JP2008108246A (en) * 2006-10-23 2008-05-08 Internatl Business Mach Corp <Ibm> Method, system and computer program for generating virtual image according to position of browsing person
US8914729B2 (en) * 2006-10-30 2014-12-16 Yahoo! Inc. Methods and systems for providing a customizable guide for navigating a corpus of content
US8032427B1 (en) * 2007-04-03 2011-10-04 Local.com System for providing localized shopping information
US8788334B2 (en) * 2007-06-15 2014-07-22 Social Mecca, Inc. Online marketing platform
US8788335B2 (en) * 2007-06-15 2014-07-22 Social Mecca, Inc. Content distribution system including cost-per-engagement based advertising
US8024113B2 (en) * 2007-08-02 2011-09-20 Volkswagen Ag Navigation system
US8874364B2 (en) * 2007-08-02 2014-10-28 Volkswagen Ag Navigation system
KR101424807B1 (en) * 2007-08-14 2014-07-31 엘지전자 주식회사 An image display system and method of controlling the same
US20100094849A1 (en) * 2007-08-17 2010-04-15 Robert Rose Systems and methods for creating user generated content incorporating content from a content catalog
WO2009062320A1 (en) * 2007-11-13 2009-05-22 Route 66 Switzerland Gmbh Automatically linking geographic terms to geographic information
US8656298B2 (en) 2007-11-30 2014-02-18 Social Mecca, Inc. System and method for conducting online campaigns
JP4470992B2 (en) * 2007-12-05 2010-06-02 セイコーエプソン株式会社 Video management system
KR20090063610A (en) * 2007-12-14 2009-06-18 삼성디지털이미징 주식회사 Method for displaying image and for image display system applying the method
US7847729B2 (en) 2008-01-28 2010-12-07 Research In Motion Limited GPS pre-acquisition for geotagging digital photos
US20090248300A1 (en) * 2008-03-31 2009-10-01 Sony Ericsson Mobile Communications Ab Methods and Apparatus for Viewing Previously-Recorded Multimedia Content from Original Perspective
US7921114B2 (en) * 2008-04-10 2011-04-05 Microsoft Corporation Capturing and combining media data and geodata in a composite file
US8522143B2 (en) 2008-05-05 2013-08-27 Microsoft Corporation Scene-granular geographical-based video footage visualizations
CN101582876A (en) * 2008-05-12 2009-11-18 华为技术有限公司 Method, device and system for registering user generated content (UGC)
KR20090123227A (en) * 2008-05-27 2009-12-02 삼성전자주식회사 Offering apparatus of searching service, method and program thereof
US20090324211A1 (en) * 2008-06-25 2009-12-31 Nokia Corporation Method and Device for Geo-Tagging an Object Before or After Creation
US8199251B2 (en) 2008-07-07 2012-06-12 Woodman Labs, Inc. Camera housing with integrated expansion module
DK2324460T3 (en) * 2008-08-12 2013-09-30 Google Inc TOUR OF A GEOGRAPHICAL INFORMATION SYSTEM
JP4582428B2 (en) * 2008-08-29 2010-11-17 ソニー株式会社 Place name registration device and place name registration method
US20100076976A1 (en) * 2008-09-06 2010-03-25 Zlatko Manolov Sotirov Method of Automatically Tagging Image Data
EP2327003B1 (en) * 2008-09-17 2017-03-29 Nokia Technologies Oy User interface for augmented reality
US20100082237A1 (en) * 2008-09-30 2010-04-01 Microsoft Corporation Journal service
US8060582B2 (en) 2008-10-22 2011-11-15 Google Inc. Geocoding personal information
US20110320495A1 (en) * 2008-10-27 2011-12-29 Guy Levy-Yurista Personal information tracking, recording, reporting and sharing system and method
TWI455585B (en) * 2009-02-17 2014-10-01 Mstar Semiconductor Inc Digital television system and receiver and associated method for providing information of point of interest
US20130278832A1 (en) * 2009-02-17 2013-10-24 Mstar Semiconductor, Inc. Digital Television System and Receiver for Providing Information of Point of Interest
US8769396B2 (en) * 2009-06-05 2014-07-01 Microsoft Corporation Calibration and annotation of video content
US9324003B2 (en) 2009-09-14 2016-04-26 Trimble Navigation Limited Location of image capture device and object features in a captured image
US8897541B2 (en) 2009-09-14 2014-11-25 Trimble Navigation Limited Accurate digitization of a georeferenced image
DE102009045021A1 (en) * 2009-09-25 2011-03-31 Robert Bosch Gmbh Navigation device for use by e.g. driver for representing, planning route or guiding target, has selected auxiliary datasets associated to selected datasets that represent auxiliary datasets in card representation
US20110074811A1 (en) * 2009-09-25 2011-03-31 Apple Inc. Map Layout for Print Production
US8306729B2 (en) * 2009-10-13 2012-11-06 Telenav, Inc. Navigation system with user generated content mechanism and method of operation thereof
KR101662595B1 (en) * 2009-11-03 2016-10-06 삼성전자주식회사 User terminal, route guide system and route guide method thereof
JP5494346B2 (en) * 2009-11-26 2014-05-14 株式会社Jvcケンウッド Information display device, information display device control method, and program
KR101302134B1 (en) * 2009-12-18 2013-08-30 한국전자통신연구원 Apparatus and method for providing hybrid sensor information
US20110211737A1 (en) * 2010-03-01 2011-09-01 Microsoft Corporation Event Matching in Social Networks
US9465993B2 (en) 2010-03-01 2016-10-11 Microsoft Technology Licensing, Llc Ranking clusters based on facial image analysis
US20110213549A1 (en) * 2010-03-01 2011-09-01 Hallas Maria E Location based virtual tour
US20110246066A1 (en) * 2010-04-01 2011-10-06 Epsillion Media Technologies Ltd. Method and System for Managing Media Items
AU2011240621B2 (en) * 2010-04-12 2015-04-16 Smule, Inc. Continuous score-coded pitch correction and harmony generation techniques for geographically distributed glee club
US8780741B2 (en) 2010-06-21 2014-07-15 International Business Machines Corporation On-demand information retrieval using wireless communication devices
KR20110139375A (en) * 2010-06-23 2011-12-29 삼성전자주식회사 Method for displaying image including a position information and apparatus of enabling the method
US8768667B2 (en) 2010-10-25 2014-07-01 Trimble Navigation Limited Water erosion management incorporating topography, soil type, and weather statistics
US9213905B2 (en) 2010-10-25 2015-12-15 Trimble Navigation Limited Automatic obstacle location mapping
US8855937B2 (en) 2010-10-25 2014-10-07 Trimble Navigation Limited Crop characteristic estimation
US10115158B2 (en) 2010-10-25 2018-10-30 Trimble Inc. Generating a crop recommendation
US9846848B2 (en) 2010-10-25 2017-12-19 Trimble Inc. Exchanging water allocation credits
US20120101784A1 (en) 2010-10-25 2012-04-26 Trimble Navigation Limited Wide-area agricultural monitoring and prediction
US9058633B2 (en) 2010-10-25 2015-06-16 Trimble Navigation Limited Wide-area agricultural monitoring and prediction
US20120101762A1 (en) * 2010-10-26 2012-04-26 Research In Motion Limited System and Method for Calibrating a Magnetometer Using Partial and Full Calibrations
US20120158850A1 (en) * 2010-12-21 2012-06-21 Harrison Edward R Method and apparatus for automatically creating an experiential narrative
JP5195986B2 (en) * 2010-12-27 2013-05-15 カシオ計算機株式会社 Image output apparatus and program
EP2697767A4 (en) * 2011-04-12 2014-09-24 Google Inc Spreading user activities to interested users of a community
US20130060710A1 (en) * 2011-05-04 2013-03-07 Matthew Preuss System, Method and Apparatus for Managing and Conducting Property Inspections
US8638375B2 (en) 2011-05-17 2014-01-28 Trimble Navigation Limited Recording data with an integrated field-portable device
CN102790944A (en) * 2011-05-19 2012-11-21 昆达电脑科技(昆山)有限公司 Method and electronic system for searching coordinators
US8891832B2 (en) * 2011-06-03 2014-11-18 Facebook, Inc. Computer-vision-assisted location check-in
US8671741B2 (en) 2011-06-29 2014-03-18 Trimble Navigation Limited Extendable moisture content sensing system
WO2013032955A1 (en) * 2011-08-26 2013-03-07 Reincloud Corporation Equipment, systems and methods for navigating through multiple reality models
US11068532B2 (en) 2011-09-21 2021-07-20 Horsetooth Ventures, LLC Interactive image display and selection system
US9734167B2 (en) 2011-09-21 2017-08-15 Horsetooth Ventures, LLC Interactive image display and selection system
US9183273B2 (en) * 2011-09-23 2015-11-10 Omnitracs, Llc Systems and methods for processing location-and entity-based workflow data
JP2013098879A (en) * 2011-11-04 2013-05-20 Sony Corp Imaging control device, imaging device, and control method for imaging control device
EP2800083A4 (en) * 2011-12-27 2015-08-19 Sony Corp Information processing device, information processing method, and program
US8996036B2 (en) 2012-02-09 2015-03-31 Southwest Research Institute Autonomous location of objects in a mobile reference frame
US20130251344A1 (en) * 2012-03-23 2013-09-26 Microsoft Corporation Manipulation of User Experience State
US9595015B2 (en) 2012-04-05 2017-03-14 Nokia Technologies Oy Electronic journal link comprising time-stamped user event image content
US20130268848A1 (en) * 2012-04-05 2013-10-10 Nokia Corporation User event content, associated apparatus and methods
CN103383688B (en) * 2012-05-02 2018-11-02 Sap欧洲公司 Memory headroom database for geocoding/geography processing
US8725413B2 (en) 2012-06-29 2014-05-13 Southwest Research Institute Location and motion estimation using ground imaging sensor
US9218529B2 (en) 2012-09-11 2015-12-22 Southwest Research Institute 3-D imaging sensor based location estimation
US9488489B2 (en) 2012-09-28 2016-11-08 Google Inc. Personalized mapping with photo tours
WO2014059374A1 (en) * 2012-10-11 2014-04-17 Imsi Design, Llc Method for fine-tuning the physical position and orientation on an electronic device
WO2014059377A1 (en) * 2012-10-11 2014-04-17 Imsi Design, Llc Method for calibrrating the physical position and orientation of an electronic device using device sensors
WO2014059386A1 (en) * 2012-10-11 2014-04-17 Imsi Design, Llc Method for calibrating the physical position and orientation of an electronic device
WO2014086357A1 (en) * 2012-12-05 2014-06-12 Aspekt R&D A/S Photo survey
US20140229253A1 (en) * 2013-02-08 2014-08-14 Robert L. Farrar, JR. System and Method for Promoting Social Interaction and Efficient Information Exchange
US9218420B1 (en) 2013-02-26 2015-12-22 Google Inc. Detecting new businesses with unrecognized query terms
USD737842S1 (en) * 2013-03-14 2015-09-01 Microsoft Corporation Display screen with graphical user interface
US10331733B2 (en) 2013-04-25 2019-06-25 Google Llc System and method for presenting condition-specific geographic imagery
US9672223B2 (en) * 2013-04-25 2017-06-06 Google Inc. Geo photo searching based on current conditions at a location
KR20150000030A (en) * 2013-06-20 2015-01-02 삼성전자주식회사 Contents sharing service
US10430015B2 (en) * 2013-08-09 2019-10-01 International Business Machines Corporation Image analysis
US9244940B1 (en) * 2013-09-27 2016-01-26 Google Inc. Navigation paths for panorama
US9805057B2 (en) 2013-10-15 2017-10-31 Google Inc. Automatic generation of geographic imagery tours
US20150106403A1 (en) * 2013-10-15 2015-04-16 Indooratlas Oy Generating search database based on sensor measurements
US9152806B2 (en) 2013-12-06 2015-10-06 Sony Corporation Computer ecosystem providing privacy and tracking in sharing user-generated content by encrypting the UGC at the imaging source
US9866534B2 (en) 2013-12-06 2018-01-09 Sony Corporation Computer ecosystem providing privacy and tracking in sharing user-generated content
US9473745B2 (en) 2014-01-30 2016-10-18 Google Inc. System and method for providing live imagery associated with map locations
US9189839B1 (en) 2014-04-24 2015-11-17 Google Inc. Automatically generating panorama tours
US9002647B1 (en) 2014-06-27 2015-04-07 Google Inc. Generating turn-by-turn direction previews
US9418472B2 (en) 2014-07-17 2016-08-16 Google Inc. Blending between street view and earth view
US10089785B2 (en) * 2014-07-25 2018-10-02 mindHIVE Inc. Real-time immersive mediated reality experiences
DE102014014372A1 (en) * 2014-10-03 2016-04-07 Iunera GmbH & Co. KG System, method and arrangements for locating and storing documents
US9471695B1 (en) * 2014-12-02 2016-10-18 Google Inc. Semantic image navigation experiences
US9618344B2 (en) * 2014-12-09 2017-04-11 Brett Harrison Digital map tracking apparatus and methods
WO2016135971A1 (en) * 2015-02-27 2016-09-01 楽天株式会社 Route search system, route search device, route search method, program, and information storage medium
US9652486B2 (en) * 2015-06-07 2017-05-16 Apple Inc. Power saving techniques for a navigation application
JPWO2017002505A1 (en) * 2015-06-30 2018-04-19 ソニー株式会社 Information processing apparatus, information processing method, and program
CN105357636A (en) * 2015-10-22 2016-02-24 努比亚技术有限公司 Method, device and system for informing nearby users, and terminals
US9460616B1 (en) * 2015-12-16 2016-10-04 International Business Machines Corporation Management of mobile objects and service platform for mobile objects
US11153355B2 (en) * 2016-07-29 2021-10-19 Smarter Systems, Inc. Systems and methods for providing individual and/or synchronized virtual tours through a realm for a group of users
JP2019091203A (en) * 2017-11-14 2019-06-13 株式会社ほぼ日 Program, information processor, and method for processing information
JP6413035B1 (en) * 2018-01-18 2018-10-24 株式会社ほぼ日 Program, information processing apparatus and information processing method
CN108647189B (en) * 2018-05-15 2022-03-15 浙江斑智科技有限公司 Method and device for identifying user crowd attributes
WO2020033389A1 (en) 2018-08-07 2020-02-13 Gopro, Inc. Camera and camera mount
USD905786S1 (en) 2018-08-31 2020-12-22 Gopro, Inc. Camera mount
USD894256S1 (en) 2018-08-31 2020-08-25 Gopro, Inc. Camera mount
US11003330B1 (en) * 2018-11-30 2021-05-11 BlueOwl, LLC Vehicular telematic systems and methods for generating interactive animated guided user interfaces
US10832449B1 (en) 2018-11-30 2020-11-10 BlueOwl, LLC Vehicular telematic systems and methods for generating interactive animated guided user interfaces
USD920419S1 (en) 2019-09-17 2021-05-25 Gopro, Inc. Camera
USD946074S1 (en) 2020-08-14 2022-03-15 Gopro, Inc. Camera
US11137976B1 (en) 2020-09-11 2021-10-05 Google Llc Immersive audio tours

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006125271A1 (en) * 2005-05-27 2006-11-30 Damit Australia Pty Ltd A digital asset management system
US7277126B2 (en) * 2003-07-11 2007-10-02 Seiko Epson Corporation Image data quality adjustment
US7603334B2 (en) * 2006-04-06 2009-10-13 Microsoft Corporation Code coverage of declarative objects

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6321158B1 (en) 1994-06-24 2001-11-20 Delorme Publishing Company Integrated routing/mapping information
US6118404A (en) * 1998-01-21 2000-09-12 Navigation Technologies Corporation Method and system for representation of overlapping features in geographic databases
US6681231B1 (en) 1999-07-26 2004-01-20 The Real Estate Cable Network, Inc. Integrated information processing system for geospatial media
AU2001230955A1 (en) 2000-01-18 2001-07-31 Richard Liming System and method providing a spatial location context
EP1263626A2 (en) * 2000-03-02 2002-12-11 Donnelly Corporation Video mirror systems incorporating an accessory module
JP2002015215A (en) * 2000-06-30 2002-01-18 Hitachi Ltd Multimedia information distribution system and portable information terminal device
US6985588B1 (en) 2000-10-30 2006-01-10 Geocodex Llc System and method for using location identity to control access to digital information
GB0117541D0 (en) * 2001-07-19 2003-08-06 Bae Systems Plc Automatic registration of images in digital terrain elevation data
US6989765B2 (en) 2002-03-05 2006-01-24 Triangle Software Llc Personalized traveler information dissemination system
WO2003081483A1 (en) 2002-03-18 2003-10-02 Daniel Rex Greening Community directory
US7006829B2 (en) * 2003-05-14 2006-02-28 Bellsouth Intellectual Property Corporation Method and system for routing a telephone call
FI20030881A (en) * 2003-06-12 2004-12-13 Nokia Corp Method, Device Arrangement, Cellular Network Terminal, and Software Application within the Terminal to Provide a Location-Based Alarm
US20050063563A1 (en) * 2003-09-23 2005-03-24 Soliman Samir S. System and method for geolocation using imaging techniques
US20050209815A1 (en) * 2004-03-02 2005-09-22 Russon Virgil K Method, system, and computer-readable medium for user-assignment of geographic data to an image file
US20050228860A1 (en) * 2004-04-12 2005-10-13 Kimmo Hamynen Methods and apparatus for geographically based Web services
US8850011B2 (en) 2005-04-21 2014-09-30 Microsoft Corporation Obtaining and displaying virtual earth images
US20070032244A1 (en) * 2005-08-08 2007-02-08 Microsoft Corporation Group-centric location tagging for mobile devices

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7277126B2 (en) * 2003-07-11 2007-10-02 Seiko Epson Corporation Image data quality adjustment
WO2006125271A1 (en) * 2005-05-27 2006-11-30 Damit Australia Pty Ltd A digital asset management system
US7603334B2 (en) * 2006-04-06 2009-10-13 Microsoft Corporation Code coverage of declarative objects

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
EXIF data date unknown [Captured by Archive.org on 31 Mar 09, Digicamhelp, http://web.archive.org/web/20090331182525/http://www.digicamhelp.com/glossary/exif-data/ *
TsuruZoh Tachibanaya, Description of Exif file format 19 Dec 99, rev 1.1, http://www.media.mit.edu/pia/Research/deepview/exif.html *

Also Published As

Publication number Publication date
US7917543B2 (en) 2011-03-29
US20110196897A1 (en) 2011-08-11
US8447787B2 (en) 2013-05-21
US20070244634A1 (en) 2007-10-18
US8447792B2 (en) 2013-05-21
US20150142806A1 (en) 2015-05-21
US8200712B2 (en) 2012-06-12
US20100146397A1 (en) 2010-06-10
US20120246198A1 (en) 2012-09-27
US20120330956A1 (en) 2012-12-27
US7617246B2 (en) 2009-11-10

Similar Documents

Publication Publication Date Title
US8447792B2 (en) System and method for presenting user generated geo-located objects
US7904483B2 (en) System and method for presenting geo-located objects
US10473465B2 (en) System and method for creating, storing and utilizing images of a geographical location
AU2007290815B2 (en) Panoramic ring user interface
US7272501B2 (en) System and method for automatically collecting images of objects at geographic locations and displaying same in online directories
US10318110B2 (en) Location-based visualization of geo-referenced context
US8532916B1 (en) Switching between best views of a place
US20120278171A1 (en) System and method of providing information based on street address
CN102289520A (en) Traffic video retrieval system and realization method thereof
US9870572B2 (en) System and method of providing information based on street address
KR100563085B1 (en) Method for compositively displaying digital map and photo image
US20150261858A1 (en) System and method of providing information based on street address
US20230044871A1 (en) Search Results With Result-Relevant Highlighting
US20110246066A1 (en) Method and System for Managing Media Items
Chang et al. Real-Time Recording and Updating of Personal Journeys via Ubiquitous Computing Technologies
KR20100066343A (en) Apparatus and method for updating spatial information digital contents

Legal Events

Date Code Title Description
AS Assignment

Owner name: GEOPEG, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KOCH, EDWARD LEE;HENNAGE, DANIEL ALLAN;COLLIER, WESLEY CLAY;REEL/FRAME:029222/0023

Effective date: 20121019

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION