Abstract:
This invention integrates model simplification and bounding volume hierarchy construction for collision detection in interactive 3D graphics. In particular, it provides general framework and a preferred method to construct bounding volume hierarchy using outputs of model simplification. Simplified models, besides their application to multi-resolution rendering, can provide clues to the shape of the input object. These clues help in the partitioning of the object's model into components that may be more tightly bounded by simple bounding volumes. The framework and method naturally employ both the bottom-up and the top-down approaches of hierarchy building, and thus can have the advantages of both approaches. The framework and method includes the steps of simplified models generation, component derivation, component tree generation, and bounding volume hierarchy generation. The operation of the method includes the steps of interactively computing, displaying and recording simplified models and bounding volume hierarchy in response to user commands. Ray tracing and collision detection may be efficiently performed using the bounding volume hierarchy generated by the invention.
Abstract:
A method for producing an animation of an object includes displaying an object on a display of a computer system. A perspective system that has a predefined relationship with the object is created on the display. The perspective system defines different planes, and different portions of the object are associated with the different planes. Each of these different portions is stored as a different image file. The perspective system is rotated about the object to establish different frames of the animation. For each frame, the different portions of the object are manipulated to orient the object to establish the predefined relationship with the rotated perspective system. For each frame, script is created describing the manipulation of the different portions of the object.
Abstract:
A nonlinear editing system for audio, video and/or other source materials comprises a system to define at least one time line representing the edit, the time line being represented in the editing system as an operator graph. The operator graph includes operators, representing sources and time-based effects, and time varying connections therebetween. For each period of the defined time line wherein the connections do not change, a expression tree comprising a simplified representation of the operators and the connections required to render that portion of the time line is extracted. Each expression tree can be examined to ensure that all of the required sources and/or connections are available to render that portion of the time line in real time and, if this is not the case, a determination can be made of which portions of the expression tree can be pre-rendered to an intermediate result, the intermediate result being substituted into expression tree to provide real time rendering.
Abstract:
A method and apparatus for interfacing locomotive 3D movements of a user to a reference in a virtual or remote environment are provided. Initially, a 3D motion of a body portion of a user is sensed as the user takes a gestural pace. This sensing includes the determining of a beginning and an end of the gestural pace taken by the user, the determining of a 3D direction characteristic of the body portion motion during the gestural pace, and the determining of a 3D extent characteristic of the body portion motion during the gestural pace. Next, a 3D direction and extent of motion in the environment corresponding to the determined direction and extent characteristics of the gestural pace is computed. Finally, the computed 3D motion is used to move the reference in the environment.
Abstract:
An image representing a motion of an object is obtained with a small processing volume. As shown in FIGS. 5A through 5G, positions in a three-dimensional space associated with sparks of a firework ball are sequentially calculated in accordance with motion characteristics of the sparks. Positions on a camera screen are obtained by projecting the calculated positions of the sparks at the current point in time in FIG. 5C, and a spark texture D is drawn in those positions. An image G representing a locus of each of the sparks is obtained by drawing a line segment connecting points that form a history of the points on the camera screen which are projections of points associated with each of the spark (points on a same radial line in FIG. 5F).
Abstract:
A technique for high-resolution motion capture according to the present invention includes a high-resolution makeup scheme and calibration object to optimize data capture. Data reduction and processing techniques reduce noise in the raw data and result in sufficient captured data to permit the use of several different image-processing techniques alone or in combination. The processed image data may be used to drive an animated character. It is emphasized that this abstract is provided to comply with the rules requiring an abstract that will allow a searcher or other reader to quickly ascertain the subject matter of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims.
Abstract:
A system and process for generating a 3D video animation of an object referred to as a 3D Video Texture is presented. The 3D Video Texture is constructed by first simultaneously videotaping an object from two or more different cameras positioned at different locations. Video from, one of the cameras is used to extract, analyze and synthesize a video sprite of the object of interest. In addition, the first, contemporaneous, frames captured by at least two of the cameras are used to estimate a 3D depth map of the scene. The background of the scene contained within the depth map is then masked out, and a clear shot of the scene background taken before filming of the object began, leaving just the object. To generate each new frame in the 3D video animation, the extracted region making up a “frame” of the video sprite is mapped onto the previously generated 3D surface. The-resulting image is rendered from a novel viewpoint, and then combined with a flat image of the background which has been warped to the correct location. In cases where it is anticipated that the subject could move frequently, the foregoing part of the procedure associated with estimating a 3D depth map of the scene and extracting the 3D surface representation of the object is performed for each subsequent set of contemporaneous frames captured by at least two of the cameras.
Abstract:
A method using computer software for automatic audio visual dubbing (5) using an efficient computerized automatic method for audio visual dubbing of movies by computerized image copying of the characteristic features of the lip movements of the dubber onto the mouth area of the original speaker. The invention uses a method of vicinity-searching, three-dimensional head modeling of the original speaker (3), and texture mapping (10) technique to produce new images which correspond to the dubbed sound track: The invention thus overcomes the well known disadvantage of the correlation problems between lip movement in an original movie and the sound track of the dubbed movie.
Abstract:
An apparatus and method for transmitting graphical representations is disclosed. The present invention has an image generator (112) that generates an image (114) of a head (106) with facial features, including an eye portion (202) that correlates to an eye (240) in the head (106). The apparatus also has a non-updating tracker (116) that searches for an eye target image of the eye (240) in the head (106). The eye portion (202) correlating to the eye (240) is positioned at the target image located by the non-updating tracker (116). The apparatus also has an updating tracker (118) that searches for an updated eye target image when the non-updating tracker (116) is unable to locate the eye target image. The updating tracker (118) also positions the eye portion (202) at the updated eye target image and updates the updated eye target image with the eye portion (202). The updating tracker further determines the location of the eye portion (202). The apparatus also has an image differentiator (120) coupled to the trackers for providing image differential signals (124), including eye differential signals representing positional difference from the location of the eye portion (202). The apparatus also has a display device (130) that receives the image differential signals (124), including the eye differential signals, and animates a model (136) with the received image differential signals (134).
Abstract:
Conveying a relationship between two objects in a scene is accomplished by selecting source and destination ends of an imaginary line between the objects in the scene, a source object being located at the source end, a destination object being located at the destination end. An image representing the destination object may be displayed near the source end of the imaginary line. The image may be animated to the location of the destination object along the imaginary line. A relationship indicator, such as a line, may also be displayed between the source object and the image as the image is animated to the destination end.