Updates
This commit is contained in:
@@ -53,11 +53,13 @@ class TestSuperAttention(unittest.TestCase):
|
||||
@parameterized.expand([[6], [12], [24], [48]])
|
||||
def test_transformer_encoder(self, input_dim):
|
||||
output_dim = spaces.Categorical(12, 24, 36)
|
||||
model = super_core.SuperTransformerEncoderLayer(
|
||||
input_dim,
|
||||
output_dim=output_dim,
|
||||
num_heads=spaces.Categorical(2, 4, 6),
|
||||
mlp_hidden_multiplier=spaces.Categorical(1, 2, 4),
|
||||
model = super_core.SuperSequential(
|
||||
super_core.SuperLinear(input_dim, output_dim),
|
||||
super_core.SuperTransformerEncoderLayer(
|
||||
output_dim,
|
||||
num_heads=spaces.Categorical(2, 4, 6),
|
||||
mlp_hidden_multiplier=spaces.Categorical(1, 2, 4),
|
||||
),
|
||||
)
|
||||
print(model)
|
||||
model.apply_verbose(True)
|
||||
|
Reference in New Issue
Block a user