/dports/graphics/vapoursynth-waifu2x-ncnn-vulkan/vapoursynth-waifu2x-ncnn-vulkan-r4/deps/ncnn/tests/ |
H A D | test_multiheadattention.cpp | 18 static int test_multiheadattention(const ncnn::Mat& a, int num_heads) in test_multiheadattention() 51 static int test_multiheadattention_sameqkv(const ncnn::Mat& a, int num_heads) in test_multiheadattention_sameqkv()
|
/dports/graphics/waifu2x-ncnn-vulkan/waifu2x-ncnn-vulkan-20210521/src/ncnn/tests/ |
H A D | test_multiheadattention.cpp | 18 static int test_multiheadattention(const ncnn::Mat& a, int num_heads) in test_multiheadattention() 51 static int test_multiheadattention_sameqkv(const ncnn::Mat& a, int num_heads) in test_multiheadattention_sameqkv()
|
/dports/benchmarks/vkpeak/vkpeak-20210430/ncnn/tests/ |
H A D | test_multiheadattention.cpp | 18 static int test_multiheadattention(const ncnn::Mat& a, int num_heads) in test_multiheadattention() 51 static int test_multiheadattention_sameqkv(const ncnn::Mat& a, int num_heads) in test_multiheadattention_sameqkv()
|
/dports/misc/ncnn/ncnn-20211208/tests/ |
H A D | test_multiheadattention.cpp | 18 static int test_multiheadattention(const ncnn::Mat& a, int num_heads) in test_multiheadattention() 51 static int test_multiheadattention_sameqkv(const ncnn::Mat& a, int num_heads) in test_multiheadattention_sameqkv()
|
/dports/misc/py-gluonnlp/gluon-nlp-0.10.0/tests/unittest/ |
H A D | test_attention_cell.py | 9 num_heads=None): argument
|
/dports/misc/py-gluonnlp/gluon-nlp-0.10.0/src/gluonnlp/model/ |
H A D | transformer.py | 199 def __init__(self, *, attention_cell='multi_head', units=128, hidden_size=512, num_heads=4, argument 315 max_length=50, num_heads=4, scaled=True, scale_embed=True, norm_inputs=True, argument 491 hidden_size=512, num_heads=4, scaled=True, argument 585 max_length=50, num_heads=4, scaled=True, scale_embed=True, norm_inputs=True, argument 880 num_heads=8, scaled=True, argument
|
H A D | attention_cell.py | 214 def __init__(self, base_cell, query_units, key_units, value_units, num_heads, use_bias=True, argument 509 scaled=True, num_heads=None, argument
|
H A D | bert.py | 84 def __init__(self, units, num_heads, dropout=0.0, use_bias=True, argument 201 def __init__(self, units=128, hidden_size=512, num_heads=4, argument 311 max_length=50, num_heads=4, dropout=0.0, argument
|
/dports/misc/py-gluonnlp/gluon-nlp-0.10.0/scripts/text_generation/model/ |
H A D | gpt.py | 70 def __init__(self, units, num_heads, dropout=0.0, argument 215 def __init__(self, units, vocab_size, max_length, num_layers, num_heads, dropout=0.0, argument
|
/dports/misc/py-gluonnlp/gluon-nlp-0.10.0/scripts/machine_translation/ |
H A D | hyperparameters.py | 31 num_heads = 8 variable
|
/dports/misc/py-gluonnlp/gluon-nlp-0.10.0/docs/examples/machine_translation/ |
H A D | hyperparameters.py | 33 num_heads = 8 variable
|
/dports/misc/py-gluonnlp/gluon-nlp-0.10.0/scripts/tests/ |
H A D | test_transformer_xl.py | 39 def test_positional_embedding_multihead_attention_cell(d_head, num_heads, query_length, argument
|
/dports/misc/py-gluonnlp/gluon-nlp-0.10.0/scripts/language_model/transformer/ |
H A D | transformer.py | 107 hidden_size=512, num_heads=4, activation='relu', scaled=True, dropout=0.0, argument 173 units=128, hidden_size=2048, num_heads=4, scaled=True, dropout=0.0, argument 531 def __init__(self, vocab_size, num_layers=2, units=128, hidden_size=2048, num_heads=4, argument
|
/dports/x11/workrave/workrave-1.10.44/frontend/gtkmm/src/ |
H A D | GUI.hh | 257 int num_heads; member in GUI
|
H A D | MainWindow.cc | 807 int num_heads = gui->get_number_of_heads(); in relocate_window() local
|
/dports/sysutils/fatback/fatback-1.3/ |
H A D | vbr.h | 28 unsigned num_heads; /* Number of Heads */ member
|
/dports/math/py-flax/flax-0.3.3/flax/core/nn/ |
H A D | attention.py | 170 num_heads, argument
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/tegra/ |
H A D | hub.h | 48 unsigned int num_heads; member
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/tegra/ |
H A D | hub.h | 48 unsigned int num_heads; member
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/tegra/ |
H A D | hub.h | 48 unsigned int num_heads; member
|
/dports/math/py-flax/flax-0.3.3/flax/nn/ |
H A D | attention.py | 215 num_heads, argument
|
/dports/games/stonesoup/crawl-0.27.1/crawl-ref/source/ |
H A D | monster.h | 88 int num_heads; ///< Hydra-like head number member
|
/dports/misc/rump/buildrump.sh-b914579/src/sys/external/bsd/drm2/dist/drm/radeon/ |
H A D | evergreen.c | 1981 u32 num_heads; /* number of active crtcs */ member 2200 u32 lb_size, u32 num_heads) in evergreen_program_watermarks() 2365 u32 num_heads = 0, lb_size; in evergreen_bandwidth_update() local
|
/dports/math/py-jax/jax-0.2.9/jax/experimental/ |
H A D | doubledouble.py | 101 def screen_nones(num_heads, in_tree_def, *heads_and_tails): argument
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/radeon/ |
H A D | evergreen.c | 1936 u32 num_heads; /* number of active crtcs */ member 2149 u32 lb_size, u32 num_heads) in evergreen_program_watermarks() 2320 u32 num_heads = 0, lb_size; in evergreen_bandwidth_update() local
|