Taiyi-Roberta-124M-D-v2

我要开发同款
匿名用户2024年07月31日
31阅读
所属分类ai、roberta、pytorch、exbert、mutlimodal、roberta
开源地址https://modelscope.cn/models/Fengshenbang/Taiyi-Roberta-124M-D-v2
授权协议Apache License 2.0

作品详情

Taiyi-Roberta-124M-D-v2

简介 Brief Introduction

使用了1M图文对进行特殊预训练的,英文版的MAP(名称暂定)的文本端RoBERTa-base。

Special pre-training on 1M image-text pairs, the textual encoder for MAP (temporary) in English, RoBERTa-base.

模型分类 Model Taxonomy

需求 Demand 任务 Task 系列 Series 模型 Model 参数 Parameter 额外 Extra
特殊 Special 多模态 Multimodal 太乙 Taiyi 待定 TBD 124M 特殊预训练方法-第二版-英文 D-v2-English

模型信息 Model Information

基于Roberta-base,我们使用特殊的训练任务引入一些多模态信息。"D"表示这是一种新的预训练方法。对于特殊的多模态表征,在论文中我们设计了集中不同的训练目标。预训练数据集为MSCOCO,VG和SBU。我们的代码和预训练任务的细节将在论文接受后公开。

Based on pre-trained Roberta-base, we apply some multimodal information with special pre-training tasks. "D" implies a special training method. For special multimodal representations, we design several special training objectives in our paper. The pre-training datasets are MSCOCO, VG and SBU. Our code and details of pre-training tasks will be made publicly available upon paper acceptance.

下游效果 Performance

GLUE

Task MNLI QQP QNLI SST-2 CoLA STS-B MRPC RTE
Robert-base (official) 87.6 91.9 92.8 94.8 63.6 91.2 90.2 78.7
Roberta-base (local) 87.0 91.3 92.5 94.2 62.8 90.6 92.9 78.0
Taiyi-Roberta-124M-D (local) 87.1 91.8 92.3 94.5 62.6 90.4 92.4 78.7
Taiyi-Roberta-124M-D-v2 (local) 87.1 91.9 92.4 94.5 65.5 91.0 93.0 79.8

The local test settings are: Sequence length: 128, Batch size: 32, Learning rate: 3e-5

使用 Usage

from transformers import RobertaTokenizer, RobertaModel

tokenizer = RobertaTokenizer.from_pretrained("IDEA-CCNL/Taiyi-Roberta-124M-D-v2")
model = RobertaModel.from_pretrained("IDEA-CCNL/Taiyi-Roberta-124M-D-v2")

引用 Citation

如果您在您的工作中使用了我们的模型,可以引用我们的论文

If you are using the resource for your work, please cite the our paper:

@article{fengshenbang,
  author    = {Junjie Wang and Yuxiang Zhang and Lin Zhang and Ping Yang and Xinyu Gao and Ziwei Wu and Xiaoqun Dong and Junqing He and Jianheng Zhuo and Qi Yang and Yongfeng Huang and Xiayu Li and Yanghan Wu and Junyu Lu and Xinyu Zhu and Weifeng Chen and Ting Han and Kunhao Pan and Rui Wang and Hao Wang and Xiaojun Wu and Zhongshen Zeng and Chongpei Chen and Ruyi Gan and Jiaxing Zhang},
  title     = {Fengshenbang 1.0: Being the Foundation of Chinese Cognitive Intelligence},
  journal   = {CoRR},
  volume    = {abs/2209.02970},
  year      = {2022}
}

也可以引用我们的网站:

You can also cite our website:

@misc{Fengshenbang-LM,
  title={Fengshenbang-LM},
  author={IDEA-CCNL},
  year={2021},
  howpublished={\url{https://github.com/IDEA-CCNL/Fengshenbang-LM}},
}
声明:本文仅代表作者观点,不代表本站立场。如果侵犯到您的合法权益,请联系我们删除侵权资源!如果遇到资源链接失效,请您通过评论或工单的方式通知管理员。未经允许,不得转载,本站所有资源文章禁止商业使用运营!
下载安装【程序员客栈】APP
实时对接需求、及时收发消息、丰富的开放项目需求、随时随地查看项目状态

评论