Training text recognition systems
    71.
    发明授权

    公开(公告)号:US10997463B2

    公开(公告)日:2021-05-04

    申请号:US16184779

    申请日:2018-11-08

    Applicant: Adobe Inc.

    Abstract: In implementations of recognizing text in images, text recognition systems are trained using noisy images that have nuisance factors applied, and corresponding clean images (e.g., without nuisance factors). Clean images serve as supervision at both feature and pixel levels, so that text recognition systems are trained to be feature invariant (e.g., by requiring features extracted from a noisy image to match features extracted from a clean image), and feature complete (e.g., by requiring that features extracted from a noisy image be sufficient to generate a clean image). Accordingly, text recognition systems generalize to text not included in training images, and are robust to nuisance factors. Furthermore, since clean images are provided as supervision at feature and pixel levels, training requires fewer training images than text recognition systems that are not trained with a supervisory clean image, thus saving time and resources.

    HIERARCHICAL SCALE MATCHING AND PATCH ESTIMATION FOR IMAGE STYLE TRANSFER WITH ARBITRARY RESOLUTION

    公开(公告)号:US20200349688A1

    公开(公告)日:2020-11-05

    申请号:US16930736

    申请日:2020-07-16

    Applicant: Adobe Inc.

    Abstract: A style of a digital image is transferred to another digital image of arbitrary resolution. A high-resolution (HR) content image is segmented into several low-resolution (LR) patches. The resolution of a style image is matched to have the same resolution as the LR content image patches. Style transfer is then performed on a patch-by-patch basis using, for example, a pair of feature transforms—whitening and coloring. The patch-by-patch style transfer process is then repeated at several increasing resolutions, or scale levels, of both the content and style images. The results of the style transfer at each scale level are incorporated into successive scale levels up to and including the original HR scale. As a result, style transfer can be performed with images having arbitrary resolutions to produce visually pleasing results with good spatial consistency.

    HIERARCHICAL SCALE MATCHING AND PATCH ESTIMATION FOR IMAGE STYLE TRANSFER WITH ARBITRARY RESOLUTION

    公开(公告)号:US20200258204A1

    公开(公告)日:2020-08-13

    申请号:US16271058

    申请日:2019-02-08

    Applicant: Adobe Inc.

    Abstract: A style of a digital image is transferred to another digital image of arbitrary resolution. A high-resolution (HR) content image is segmented into several low-resolution (LR) patches. The resolution of a style image is matched to have the same resolution as the LR content image patches. Style transfer is then performed on a patch-by-patch basis using, for example, a pair of feature transforms—whitening and coloring. The patch-by-patch style transfer process is then repeated at several increasing resolutions, or scale levels, of both the content and style images. The results of the style transfer at each scale level are incorporated into successive scale levels up to and including the original HR scale. As a result, style transfer can be performed with images having arbitrary resolutions to produce visually pleasing results with good spatial consistency.

    Font attributes for font recognition and similarity

    公开(公告)号:US10699166B2

    公开(公告)日:2020-06-30

    申请号:US15853120

    申请日:2017-12-22

    Applicant: Adobe Inc.

    Abstract: Font recognition and similarity determination techniques and systems are described. In a first example, localization techniques are described to train a model using machine learning (e.g., a convolutional neural network) using training images. The model is then used to localize text in a subsequently received image, and may do so automatically and without user intervention, e.g., without specifying any of the edges of a bounding box. In a second example, a deep neural network is directly learned as an embedding function of a model that is usable to determine font similarity. In a third example, techniques are described that leverage attributes described in metadata associated with fonts as part of font recognition and similarity determinations.

    Recognizing unseen fonts based on visual similarity

    公开(公告)号:US10528649B2

    公开(公告)日:2020-01-07

    申请号:US15280505

    申请日:2016-09-29

    Applicant: Adobe Inc.

    Abstract: Font recognition and similarity determination techniques and systems are described. For example, a computing device receives an image including a font and extracts font features corresponding to the font. The computing device computes font feature distances between the font and fonts from a set of training fonts. The computing device calculates, based on the font feature distances, similarity scores for the font and the training fonts used for calculating features distances. The computing device determines, based on the similarity scores, final similarity scores for the font relative to the training fonts.

    Generating a compact video feature representation in a digital medium environment

    公开(公告)号:US10430661B2

    公开(公告)日:2019-10-01

    申请号:US15384831

    申请日:2016-12-20

    Applicant: Adobe Inc.

    Abstract: Techniques and systems are described to generate a compact video feature representation for sequences of frames in a video. In one example, values of features are extracted from each frame of a plurality of frames of a video using machine learning, e.g., through use of a convolutional neural network. A video feature representation is generated of temporal order dynamics of the video, e.g., through use of a recurrent neural network. For example, a maximum value is maintained of each feature of the plurality of features that has been reached for the plurality of frames in the video. A timestamp is also maintained as indicative of when the maximum value is reached for each feature of the plurality of features. The video feature representation is then output as a basis to determine similarity of the video with at least one other video based on the video feature representation.

    OIL PAINTING STROKE SIMULATION USING NEURAL NETWORK

    公开(公告)号:US20190147627A1

    公开(公告)日:2019-05-16

    申请号:US15814751

    申请日:2017-11-16

    Applicant: Adobe Inc.

    Abstract: Oil painting simulation techniques are disclosed which simulate painting brush strokes using a trained neural network. In some examples, a method may include inferring a new height map of existing paint on a canvas after a new painting brush stroke is applied based on a bristle trajectory map that represents the new painting brush stroke and a height map of existing paint on the canvas prior to the application of the new painting brush stroke, and generating a rendering of the new painting brush stroke based on the new height map of existing paint on the canvas after the new painting brush stroke is applied to the canvas and a color map.

    Multi-style texture synthesis
    79.
    发明授权

    公开(公告)号:US10192321B2

    公开(公告)日:2019-01-29

    申请号:US15409321

    申请日:2017-01-18

    Applicant: ADOBE INC.

    Abstract: Systems and techniques that synthesize an image with similar texture to a selected style image. A generator network is trained to synthesize texture images depending on a selection unit input. The training configures the generator network to synthesize texture images that are similar to individual style images of multiple style images based on which is selected by the selection unit input. The generator network can be configured to minimize a covariance matrix-based style loss and/or a diversity loss in synthesizing the texture images. After training the generator network, the generator network is used to synthesize texture images for selected style images. For example, this can involve receiving user input selecting a selected style image, determining the selection unit input based on the selected style image, and synthesizing texture images using the generator network with the selection unit input and noise input.

Patent Agency Ranking