Providing Fairness in Fine-Tuning of Pre-Trained Language Models

    公开(公告)号:US20230409969A1

    公开(公告)日:2023-12-21

    申请号:US18176374

    申请日:2023-02-28

    CPC classification number: G06N20/00

    Abstract: Bias in a language model generated through fine tuning of a pre-trained language model may be mitigated, whether the bias may be incorporated in the pre-trained language model or in fine-tuning data. A pre-trained language model may be fine-tuned using downstream training data. Prior to tuning, elements within the downstream data may be identified that either match or serve as proxies for one or more identity elements associated with training bias sensitivity. Proxy elements may be identified using an analysis of distributions of the downstream elements and distributions of identity elements. Once the elements are identified, instances of the identified elements may be replaced in the downstream data with one or more masking element to generate masked downstream data. A fine-tuned language model with reduced bias may then be generated from the pre-trained language model by tuning the pre-trained language model using the masked downstream data.

    Enforcing Fairness on Unlabeled Data to Improve Modeling Performance

    公开(公告)号:US20230394371A1

    公开(公告)日:2023-12-07

    申请号:US18453929

    申请日:2023-08-22

    CPC classification number: G06N20/00 G06N3/088

    Abstract: Fairness of a trained classifier may be ensured by generating a data set for training, the data set generated using input data points of a feature space including multiple dimensions and according to different parameters including an amount of label bias, a control for discrepancy between rarity of features, and an amount of selection bias. Unlabeled data points of the input data comprising unobserved ground truths are labeled according to the amount of label bias and the input data sampled according to the amount of selection bias and the control for the discrepancy between the rarity of features. The classifier is then trained using the sampled and labeled data points as well as additional unlabeled data points. The trained classifier is then usable to determine unbiased classifications of one or more labels for one or more other data sets.

    Evaluating language models using negative data

    公开(公告)号:US11488579B2

    公开(公告)日:2022-11-01

    申请号:US16890263

    申请日:2020-06-02

    Abstract: A method of evaluating a language model using negative data may include accessing a first language model that is trained using a first training corpus, and accessing a second language model. The second language model may be configured to generate outputs that are less grammatical than outputs generated by the first language model. The method may also include training the second language model using a second training corpus, and generating output text from the second language model. The method may further include testing the first language model using the output text from the second language model.

    Similarity Analysis Using Enhanced MinHash
    19.
    发明公开

    公开(公告)号:US20240168934A1

    公开(公告)日:2024-05-23

    申请号:US18426100

    申请日:2024-01-29

    CPC classification number: G06F16/2228 G06F17/18 G06F18/22 G06F18/231

    Abstract: A first set and a second set are identified as operands for a set operation of a similarity analysis task iteration. Using respective minimum hash information arrays and contributor count arrays of the two sets, a minimum hash information array and contributor count array of a derived set resulting from the set operation is generated. An entry in the contributor count array of the derived set indicates the number of child sets of the derived set that meet a criterion with respect to a corresponding entry in the minimum hash information array of the derived set. The generated minimum hash information array and the contributor count array are stored as part of input for a subsequent iteration. After a termination criterion of the task is met, output of the task is stored.

Patent Agency Ranking