File size: 2,280 Bytes
abafd4e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
model_class: STPatch   # NDT2 is a sub-class of STPatch


encoder:

  stitching: false
  from_pt: null
  embed_region: false

  masker:
    force_active: true         
    mode: random_token  
    ratio: 0.3                # ratio of data to predict
    zero_ratio: 1.0           # of the data to predict, ratio of zero-ed out
    random_ratio: 1.0         # of the not zero-ed, ratio of randomly replaced
    expand_prob: 0.0          # probability of expanding the mask in "temporal" mode
    max_timespan: 1           # max span of mask if expanded
    channels: null            # neurons to mask in "co-smoothing" mode
    timesteps: null           # time steps to mask in "forward-pred" mode
    mask_regions: ['all']     # brain regions to mask in "inter-region" mode
    target_regions: ['all']   # brain regions to predict in "intra-region" mode
    n_mask_regions: 1         # num of regions to choose from the list of mask_regions or target_regions
    
  patcher:
    active: true       
    time_stride: 0

  # context available for each timestep
  context:
    forward: -1
    backward: -1

  embedder:
    n_neurons: 1280 
    n_timesteps: 100
    max_time_F: 1
    max_space_F: 128
    max_spikes: 0         # max number of spikes in a single time bin
    mode: linear          # linear/embed/identity
    mult: 2               # embedding multiplier. hiddden_sizd = n_channels * mult
    act: softsign         # activation for the embedding layers
    scale: 1              # scale the embedding multiplying by this number
    bias: true            # use bias in the embedding layer
    dropout: 0.2          # dropout in embedding layer
    use_prompt: false
    use_session: true

  transformer:
    n_layers: 5           # number of transformer layers
    hidden_size: 128      # hidden space of the transformer
    n_heads: 8            # number of attentiomn heads
    attention_bias: true  # learn bias in the attention layers
    act: gelu             # activiation function in mlp layers
    inter_size: 512       # intermediate dimension in the mlp layers
    mlp_bias: true        # learn bias in the mlp layers
    dropout: 0.4          # dropout in transformer layers
    fixup_init: true      # modify weight initialization


decoder:
  from_pt: null