THREE-DIMENSIONAL (3D) CONVOLUTION WITH 3D BATCH NORMALIZATION
    141.
    发明申请
    THREE-DIMENSIONAL (3D) CONVOLUTION WITH 3D BATCH NORMALIZATION 审中-公开
    三维(3D)三维拼接正则化的解决方案

    公开(公告)号:US20170046616A1

    公开(公告)日:2017-02-16

    申请号:US15237575

    申请日:2016-08-15

    Abstract: The technology disclosed uses a 3D deep convolutional neural network architecture (DCNNA) equipped with so-called subnetwork modules which perform dimensionality reduction operations on 3D radiological volume before the 3D radiological volume is subjected to computationally expensive operations. Also, the subnetworks convolve 3D data at multiple scales by subjecting the 3D data to parallel processing by different 3D convolutional layer paths. Such multi-scale operations are computationally cheaper than the traditional CNNs that perform serial convolutions. In addition, performance of the subnetworks is further improved through 3D batch normalization (BN) that normalizes the 3D input fed to the subnetworks, which in turn increases learning rates of the 3D DCNNA. After several layers of 3D convolution and 3D sub-sampling with 3D across a series of subnetwork modules, a feature map with reduced vertical dimensionality is generated from the 3D radiological volume and fed into one or more fully connected layers.

    Abstract translation: 所公开的技术使用配备有所谓的子网模块的3D深卷积神经网络架构(DCNNA),其在3D放射体积经受计算上昂贵的操作之前对3D放射体积进行降维操作。 此外,子网络通过对3D数据进行不同的3D卷积层路径的并行处理,在多个尺度上卷积3D数据。 这种多尺度操作在计算上比执行串行卷积的传统CNN便宜。 此外,通过对馈送到子网络的3D输入进行归一化的3D批量归一化(BN),进一步提高了子网络的性能,从而提高了3D DCNNA的学习速率。 在通过一系列子网模块进行三维3D卷积和三维子采样与三维子采样之后,从3D放射体积产生具有降低的垂直维数的特征图,并将其馈送到一个或多个完全连接的层。

    Parameter utilization for language pre-training

    公开(公告)号:US12072955B2

    公开(公告)日:2024-08-27

    申请号:US17532851

    申请日:2021-11-22

    CPC classification number: G06F18/2148 G06F18/2163 G06F40/00

    Abstract: Embodiments are directed to pre-training a transformer model using more parameters for sophisticated patterns (PSP++). The transformer model is divided into a held-out model and a main model. A forward pass and a backward pass are performed on the held-out model, where the forward pass determines self-attention hidden states of the held-out model and the backward pass determines loss of the held-out model. A forward pass on the main model is performed to determine a self-attention hidden states of the main model. The self-attention hidden states of the main model are concatenated with the self-attention hidden states of the held-out model. A backward pass is performed on the main model to determine a loss of the main model. The parameters of the held-out model are updated to reflect the loss of the held-out model and parameters of the main model are updated to reflect the loss of the main model.

    Systems and methods for structured text translation with tag alignment

    公开(公告)号:US11822897B2

    公开(公告)日:2023-11-21

    申请号:US17463227

    申请日:2021-08-31

    CPC classification number: G06F40/58 G06N3/08

    Abstract: Approaches for the translation of structured text include an embedding module for encoding and embedding source text in a first language, an encoder for encoding output of the embedding module, a decoder for iteratively decoding output of the encoder based on generated tokens in translated text from previous iterations, a beam module for constraining output of the decoder with respect to possible embedded tags to include in the translated text for a current iteration using a beam search, and a layer for selecting a token to be included in the translated text for the current iteration. The translated text is in a second language different from the first language. In some embodiments, the approach further includes scoring and pointer modules for selecting the token based on the output of the beam module or copied from the source text or reference text from a training pair best matching the source text.

    Robustness evaluation via natural typos

    公开(公告)号:US11669712B2

    公开(公告)日:2023-06-06

    申请号:US16559196

    申请日:2019-09-03

    CPC classification number: G06N3/008 G06F40/232 G06N3/044 G06N3/045 G06N3/08

    Abstract: A method for evaluating robustness of one or more target neural network models using natural typos. The method includes receiving one or more natural typo generation rules associated with a first task associated with a first input document type, receiving a first target neural network model, and receiving a first document and corresponding its ground truth labels. The method further includes generating one or more natural typos for the first document based on the one or more natural typo generation rules, and providing, to the first target neural network model, a test document generated based on the first document and the one or more natural typos as an input document to generate a first output. A robustness evaluation result of the first target neural network model is generated based on a comparison between the output and the ground truth labels.

    REINFORCEMENT LEARNING BASED GROUP TESTING

    公开(公告)号:US20230113750A1

    公开(公告)日:2023-04-13

    申请号:US17498155

    申请日:2021-10-11

    Abstract: A system performs group testing on a population of items. The group testing identifies items satisfying particular criteria from a population of items, for example, defective items from the population. The group testing may be performed for software or hardware testing, for testing a human population, for training of deep learning applications, and so on. The system trains a machine learning based model, for example, a reinforcement learning based model to evaluate groups. The model may further determine system dynamics that may represent priors of items. An agent treats the population and groups of items being tested as the environment and performs actions, for example, adjusting the groups. The system also performs a non-adaptive strategy based on monte carlo simulation of tests based on a simulation results.

Patent Agency Ranking