Description
Method introduces a novel unlabeled debiasing technique which works on classification task to reduce the bias of the transformer based language models on downstream classification task. In their method authors use the classes as metric for regularization and punish the network if the embedding produced by the model are far from each other. by doing so the authors claim to be able to reduce the domain shift caused by any unwanted attribute information hence results in fair embedding.
Papers Using This Method
Loss-Versus-Rebalancing under Deterministic and Generalized block-times2025-05-08Impermanent loss and Loss-vs-Rebalancing II2025-02-06Impermanent loss and loss-vs-rebalancing I: some statistical properties2024-10-01Unlabeled Debiasing in Downstream Tasks via Class-wise Low Variance Regularization2024-09-29