Fix CUDA memory issues

This commit is contained in:
D-X-Y
2021-03-30 09:02:41 +00:00
parent 5fb900bcb1
commit 9fc2c991f5
7 changed files with 72 additions and 60 deletions

View File

@@ -15,8 +15,8 @@
# python exps/trading/baselines.py --alg TabNet #
# #
# python exps/trading/baselines.py --alg Transformer#
# python exps/trading/baselines.py --alg TSF #
# python exps/trading/baselines.py --alg TSF-4x64-d0
# python exps/trading/baselines.py --alg TSF
# python exps/trading/baselines.py --alg TSF-4x64-drop0_0
#####################################################
import sys
import copy
@@ -40,10 +40,11 @@ from qlib.workflow import R
from qlib.utils import flatten_dict
def to_pos_drop(config, value):
def to_drop(config, pos_drop, other_drop):
config = copy.deepcopy(config)
net = config["task"]["model"]["kwargs"]["net_config"]
net["pos_drop"] = value
net["pos_drop"] = pos_drop
net["other_drop"] = other_drop
return config
@@ -59,11 +60,12 @@ def to_layer(config, embed_dim, depth):
def extend_transformer_settings(alg2configs, name):
config = copy.deepcopy(alg2configs[name])
for i in range(1, 7):
for j in [6, 12, 24, 32, 48, 64]:
for k in [0, 0.1]:
alg2configs[name + "-{:}x{:}-d{:}".format(i, j, k)] = to_layer(
to_pos_drop(config, k), j, i
)
for j in (6, 12, 24, 32, 48, 64):
for k1 in (0, 0.1, 0.2):
for k2 in (0, 0.1):
alg2configs[
name + "-{:}x{:}-drop{:}_{:}".format(i, j, k1, k2)
] = to_layer(to_drop(config, k1, k2), j, i)
return alg2configs