Transformers: bert-large config file

Created on 15 Jul 2019  路  1Comment  路  Source: huggingface/transformers

Here is the config file I download from path in modelling for bert large,
{
"attention_probs_dropout_prob": 0.1,
"directionality": "bidi",
"hidden_act": "gelu",
"hidden_dropout_prob": 0.1,
"hidden_size": 1024,
"initializer_range": 0.02,
"intermediate_size": 4096,
"max_position_embeddings": 512,
"num_attention_heads": 16,
"num_hidden_layers": 24,
"pooler_fc_size": 768,
"pooler_num_attention_heads": 12,
"pooler_num_fc_layers": 3,
"pooler_size_per_head": 128,
"pooler_type": "first_token_transform",
"type_vocab_size": 2,
"vocab_size": 28996
}

I am wondering what are the following params for? I can't find them in the modelling file and the checkpoint I download.

"pooler_fc_size": 768,
"pooler_num_attention_heads": 12,
"pooler_num_fc_layers": 3,
"pooler_size_per_head": 128,
"pooler_type": "first_token_transform",

wontfix

>All comments

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

Was this page helpful?
0 / 5 - 0 ratings

Related issues

iedmrc picture iedmrc  路  3Comments

lemonhu picture lemonhu  路  3Comments

lcswillems picture lcswillems  路  3Comments

HanGuo97 picture HanGuo97  路  3Comments

0x01h picture 0x01h  路  3Comments