Abstract:
Methods, systems, computer-readable media, and apparatuses for hierarchical clustering for view management in augmented reality are presented. For example one disclosed method includes the steps of accessing point of interest (POI) metadata for a plurality of points of interest associated with a scene; generating a hierarchical cluster tree for at least a portion of the POIs; establishing a plurality of subdivisions associated with the scene; selecting a plurality of POIs from the hierarchical cluster tree for display based on an augmented reality (AR) viewpoint of the scene, the plurality of subdivisions, and a traversal of at least a portion of the hierarchical cluster tree; and displaying labels comprising POI metadata associated with the selected plurality of POIs, the displaying based on placements determined using image-based saliency.
Abstract:
A mobile device uses an image-driven view management approach for annotating images in real-time. An image-based layout process used by the mobile device computes a saliency map and generates an edge map from a frame of a video stream. The saliency map may be further processed by applying thresholds to reduce the number of saliency levels. The saliency map and edge map are used together to determine a layout position of labels to be rendered over the video stream. The labels are displayed in the layout position until a change of orientation of the camera that exceeds a threshold is detected. Additionally, the representation of the label may be adjusted, e.g., based on a plurality of pixels bounded by an area that is coincident with a layout position for a label in the video frame.
Abstract:
A mobile device uses an image-driven view management approach for annotating images in real-time. An image-based layout process used by the mobile device computes a saliency map and generates an edge map from a frame of a video stream. The saliency map may be further processed by applying thresholds to reduce the number of saliency levels. The saliency map and edge map are used together to determine a layout position of labels to be rendered over the video stream. The labels are displayed in the layout position until a change of orientation of the camera that exceeds a threshold is detected. Additionally, the representation of the label may be adjusted, e.g., based on a plurality of pixels bounded by an area that is coincident with a layout position for a label in the video frame.