1
0
Fork 0
mirror of synced 2024-07-06 07:01:23 +12:00
waifu2x/models/vgg_7/art_y/noise1_model.t7

1651 lines
4.2 MiB
Plaintext
Raw Normal View History

2015-05-16 17:48:05 +12:00
4
1
3
V 1
13
nn.Sequential
3
2
9
2016-03-27 21:15:14 +13:00
2
13
w2nn_channels
1
1
2016-03-27 21:15:14 +13:00
2
17
w2nn_scale_factor
1
1
2016-03-27 21:15:14 +13:00
2
6
output
4
2015-05-16 17:48:05 +12:00
3
2016-03-27 21:15:14 +13:00
3
V 1
16
torch.CudaTensor
0
1
0
2015-05-16 17:48:05 +12:00
2
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-03-27 21:15:14 +13:00
4
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
7
modules
3
2016-03-27 21:15:14 +13:00
5
2015-05-16 17:48:05 +12:00
14
1
1
4
2016-03-27 21:15:14 +13:00
6
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
2016-03-27 21:15:14 +13:00
7
16
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
2
dH
1
1
2
2
dW
1
1
2
11
nInputPlane
1
1
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-03-27 21:15:14 +13:00
8
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2
2016-03-27 21:15:14 +13:00
2
kH
2015-05-16 17:48:05 +12:00
1
2016-03-27 21:15:14 +13:00
3
2
2016-01-23 13:36:54 +13:00
9
2016-03-27 21:15:14 +13:00
gradInput
4
4
2016-01-23 13:36:54 +13:00
2
2
2016-03-27 21:15:14 +13:00
kW
2015-05-16 17:48:05 +12:00
1
2016-03-27 21:15:14 +13:00
3
2015-05-16 17:48:05 +12:00
2
12
nOutputPlane
2015-05-16 17:48:05 +12:00
1
32
2015-05-16 17:48:05 +12:00
2
4
2016-03-27 21:15:14 +13:00
padW
2015-05-16 17:48:05 +12:00
1
0
2
2016-01-23 13:36:54 +13:00
6
weight
2015-05-16 17:48:05 +12:00
4
2016-03-27 21:15:14 +13:00
9
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
2
32 9
9 1
2015-05-16 17:48:05 +12:00
1
4
2016-03-27 21:15:14 +13:00
10
3
V 1
17
torch.CudaStorage
2016-01-23 13:36:54 +13:00
288
2016-03-27 21:15:14 +13:00
0.0236385372 0.161608517 0.100228615 -0.0498797521 -0.0256561618 0.0797657892 -0.0295766201 -0.216625899 -0.0382243544 -0.00966789573 -0.048580993 0.0490917601 0.0672969297 0.0176148359 -0.0884531438 -0.0608262643 0.0429916531 0.0198188853 0.0644578785 -0.0986441001 0.0384854786 -0.101563461 0.14195253 -0.0483320095 0.0482366942 -0.0606093407 0.018800674 -0.0658788532 0.0791210979 0.0949304104 -0.0528751351 -0.06710089 0.209388062 0.0264986958 0.184565231 0.13765049 0.0021111134 0.159141898 0.0475657471 0.00227641477 -0.0560809486 -0.133702293 -1.71132579e-05 -0.0104781631 -0.016501762 -0.00294497283 0.0125072757 -0.00172730873 0.00619779062 0.170819774 -0.0119940052 -0.00349967019 0.00370716327 -0.172641397 -0.00557552418 -0.110617429 0.120372228 0.023376869 -0.118951552 0.0900648758 -0.00135315699 -0.012033972 0.0156980548 -0.0235765353 -0.128287315 -0.0606651716 0.0625041947 0.00657891482 -0.0718166903 0.0101898108 0.16934742 0.0385997146 -0.0540178865 0.0398611426 -0.00806593429 -0.0227660779 -0.011346993 0.00351928687 -0.0205458924 -0.0166823491 -0.00763849542 0.00737901777 -0.00589831825 -0.00158130086 0.155859023 0.00524113141 -0.0111329025 -0.0101837805 0.00247843936 0.00078520301 -0.0381705314 0.00984696392 -0.0247839727 0.0325654931 0.0167124253 0.00939460471 0.00315193948 -0.0188349467 -0.00245267316 0.00759131368 -0.0158620384 -0.0010624032 -0.024625862 0.0434264094 -0.0291855801 0.0239853095 -0.0776515082 0.0583680756 -0.000528637145 -0.165776521 0.134456947 -0.00119506312 -0.0607389063 -0.0636905804 0.000569747761 -0.066347219 -0.0625761673 -0.0030676818 -0.100392133 -0.0322485305 0.00406643748 -0.0170747638 0.00637601549 0.00405062921 -0.00073939556 0.0113485139 -0.00167617423 0.0228640996 -0.00232793577 0.0123241153 -0.0208963081 -0.00947349519 -0.00439497875 -0.0021232022 -0.01082562 0.0221626647 -0.0275627617 0.00536771677 0.0625763312 -0.082505472 -0.0379155204 0.0793289915 -0.000932817115 -0.0266423803 0.00118642719 -0.015199854 0.00949841645 -0.0194953717 -0.208703145 -0.0425602533 0.0186593886 0.224720627 0.0314429067 -0.00549812475 0.0141942278 -5.28863129e-05 0.081404902 0.0644556135 0.0105421059 -0.06764476 -0.0839853138 -0.0154159497 0.00552086718 -0.0754293725 -0.00941077713 -0.063940689 -0.0960616693 0.0798697695 -0.00282597495 0.0862777233 0.0700693056 0.161970228 -0.029953979 -0.135672361 0.107565455 -0.0362037309 -0.075346306 0.00971181411 -0.00428676512 0.00157421059 -0.00677057682 0.00364455138 0.00531580625 -0.00122944766 0.0113157677 -0.0173879936 0.017068699 -0.0199526604 0.00572623825 0.0563422889 -0.0674937889 -0.0814125016 0.0452018268 0.0568280518 -0.179675624 -0.0227521174 -0.157801971 -0.118883669 -0.191528618 -0.0519427769 -0.00340013416 0.0693743154 0.160693973 0.00328320381 0.00770053966 0.0139692761 -0.00126577355 0.155041292 -0.127151564 0.00334492349 0.0515887104 0.0600706153 -0.00304401084 0.0625304058 0.0569789335 0.00244790036 0.00859138835 0.0222355463 0.0022261315 -0.00418943819 0.170989737 0.00447220402 -0.00777340261 -0.187078655 -0.00966135971 0.00116462144 1.64600115e-05 0.001089887 -0.00103845389 -0.330238581 0.00099725218 0.00425210316 0.000930653478 0.00261735055 0.0668927655 -0.0559243709 -0.0359675996 0.0177956894 -0.0398254171 0.0528545231 -0.0462526642 -0.021336019 0.0628727525 -0.0300407894 -0.0703735575 -0.149280161 -0.0282354876 -0.0458415598 -0.0050180126 0.0054902276 -0.0118584512 0.00442240434 -0.0553331859 0.110387072 -0.0654396042 -0.0380402505 0.0197635051 0.018787507 0.0948799923 -0.132325098 0.0439289808 -0.0413912013 0.0962318182 0.0368312225 -0.0401573181 -0.0952333137 -0.0332083404 0.0402752161 0.0334511176 -0.0112788873 -0.0027501788 -0.00014320889 0.00640216563 -0.142596468 0.0238603335 0.116472609 -0.110974126 0.0295457114 0.0807825476 -0.00060687674 0.00436778506 -0.022026768 -0.0205131117 0.0689378753 -0.0411637351 -0.0542537086 0.0376105867 0.00509550422
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
10
gradWeight
2015-05-16 17:48:05 +12:00
4
2016-03-27 21:15:14 +13:00
11
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
32 9
9 1
2015-05-16 17:48:05 +12:00
1
4
2016-03-27 21:15:14 +13:00
12
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
288
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
5
_type
2
16
torch.CudaTensor
2
2015-05-16 17:48:05 +12:00
4
2016-03-27 21:15:14 +13:00
bias
4
13
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-03-27 21:15:14 +13:00
32
1
1
4
14
3
V 1
17
torch.CudaStorage
32
0.0083896108 -0.00638516014 -0.00306427013 -0.135742351 -0.00821015425 -0.00892026257 -0.0074065919 0.00272147474 0.0192636736 -0.135550618 -0.00504514435 -0.00972002372 0.0544545576 0.0151337022 -0.00136777037 -0.0137777328 -0.00810726359 -0.0248421077 -0.0103313522 -0.00529162074 -0.00584792951 0.116672158 0.00990944263 -0.0497390702 -0.00796338823 0.00858508516 -0.0103212344 0.0124345412 -0.000902330212 -0.0024440398 -0.00441889279 -0.00251973979
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
8
gradBias
2016-01-23 13:36:54 +13:00
4
2016-03-27 21:15:14 +13:00
15
3
V 1
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
1
2016-03-27 21:15:14 +13:00
32
2015-05-16 17:48:05 +12:00
1
1
2016-03-27 21:15:14 +13:00
4
16
3
V 1
17
torch.CudaStorage
32
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
4
padH
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
0
2015-05-16 17:48:05 +12:00
1
2
4
2016-03-27 21:15:14 +13:00
17
2015-05-16 17:48:05 +12:00
3
V 1
12
nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
2016-03-27 21:15:14 +13:00
18
6
2015-05-16 17:48:05 +12:00
2
7
inplace
2016-01-23 13:36:54 +13:00
5
1
2
2016-01-23 13:36:54 +13:00
5
_type
2016-01-23 13:36:54 +13:00
2
2015-05-16 17:48:05 +12:00
16
torch.CudaTensor
2
6
negval
2015-05-16 17:48:05 +12:00
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
19
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
6
output
2015-05-16 17:48:05 +12:00
4
20
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
5
train
5
0
2015-05-16 17:48:05 +12:00
1
3
4
21
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
22
2016-03-27 21:15:14 +13:00
16
2
2016-03-27 21:15:14 +13:00
2
dH
1
1
2
2
dW
1
2015-05-16 17:48:05 +12:00
1
2
11
nInputPlane
1
32
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
23
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
2
kH
1
3
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
24
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
kW
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
3
2015-05-16 17:48:05 +12:00
2
12
nOutputPlane
1
2015-05-16 17:48:05 +12:00
32
2
4
2016-03-27 21:15:14 +13:00
padW
1
0
2
6
weight
2016-01-23 13:36:54 +13:00
4
25
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-03-27 21:15:14 +13:00
2
32 288
288 1
2015-05-16 17:48:05 +12:00
1
4
26
2015-05-16 17:48:05 +12:00
3
V 1
17
torch.CudaStorage
2016-03-27 21:15:14 +13:00
9216
-0.00925297942 -0.108442836 -0.0211179014 0.0550649948 -0.0530612618 -0.0806138292 -0.0339739136 0.101304874 0.00290906988 0.0667883307 -0.0677550137 0.00404634327 0.0845462382 -0.0393010862 -0.0539485849 -0.0739791915 -0.0223166347 -0.0600319318 0.0114345029 -0.0728272721 -0.0663402751 0.105131097 -5.95501369e-05 -0.00266773906 -0.00628210651 0.0938234329 0.082727775 0.10234458 0.176190883 0.0553822294 0.0244597942 -0.389565855 -0.0424597934 -0.098841697 0.0721115246 0.0256554298 -0.0579242632 -0.00257363962 0.0141921164 -0.102411836 0.00699009141 0.00507201161 -0.122524649 -0.0122531848 0.0397480577 0.0309157036 0.00975911878 -0.0121958014 -0.0776088387 0.00461126538 0.0421374142 0.0278787818 0.0571378358 0.0592891388 0.0096510062 0.0304295793 0.0387894362 -0.0015252853 0.0534478463 -0.0368924923 0.00499951933 0.0189000983 -0.0117024137 0.00571231544 0.0931284502 0.0503981002 0.0284631327 -0.0193199068 0.0942417011 0.058855325 -0.0456241444 -0.0146354213 -0.0238932632 0.0203088131 0.0284663457 -0.0333782695 0.0304804668 0.00868880562 -0.100040309 0.0330113024 -0.00695317192 0.0046831877 0.0741874129 -0.0223444607 0.0966489539 0.0160755888 0.0536299683 -0.0154670989 -0.0617870577 -0.169237703 0.0151324933 -0.0196400769 0.0640675947 -0.010834516 0.0100330301 -0.0645365268 -0.111274473 -0.0773977041 0.0709909499 0.00144628016 -0.0209650807 0.0382298641 -0.0441485383 -0.0331865437 0.0873150304 -0.0732038915 0.00812170003 0.065356046 -0.134642005 -0.0629936382 -0.0111347837 0.0428658016 0.0398798883 0.0286165681 0.378294826 -0.202860236 -0.0210488066 0.00501132896 0.0700390935 -0.00477288617 -0.00885622762 0.0448588878 0.0397694521 0.0290242881 -0.240672886 0.0662621781 -0.0735511184 -0.0281270649 0.0436584391 -0.0765103698 0.0109947771 -0.0323700458 0.00219170633 0.0177802853 0.0354105718 -0.089780353 -0.00311417994 -0.0532211438 -0.16023168 0.0102210185 -0.0912483186 -0.0739716291 0.0505074747 -0.0453357771 0.0490933396 -0.0612553507 0.012580324 0.110213749 0.111052565 -0.0228032824 0.0638903305 0.0796732008 -0.0970304832 0.00291406619 0.0858561322 0.038337376 0.00186546962 -0.146178812 -0.145837292 -0.00175029878 -0.06911055 0.0144591117 -0.0548633188 -0.0507476293 -0.0484798588 -0.0495647006 0.0405452922 -0.0139209963 0.0413265117 0.0203671344 -0.0914059952 0.0702807233 0.0331546701 0.0110129062 0.0377618708 0.0347986706 -0.0380962566 0.130508646 -0.0681781396 0.0280755386 -0.0486369692 0.0371532962 0.0269680712 -0.0625050738 -0.066247955 -0.0195791945 0.0291847959 -0.00185126695 0.0258534122 -0.14782463 -0.226612717 -0.0531931333 -0.0563896485 0.485351503 0.0523832366 0.0794355497 -0.0815159008 -0.0318473764 -0.0166521426 0.0690802708 0.0214235112 -0.0251269955 0.104281567 -0.0470357835 -0.00529167289 0.0456712656 -0.143375322 0.00418825587 0.121541314 0.110364966 -0.0210277401 0.00146849232 -0.0710501894 0.0442799218 -0.365519971 0.277394295 -0.142500058 0.0584126264 -0.0338332653 -0.13261801 -0.0500108786 -0.015615846 -0.136744395 -0.166052938 0.0265302621 -0.0204569306 -0.099251695 -0.115823373 -0.0102943294 -0.33611846 -0.0405174159 0.238302603 0.0294494256 -0.122846194 -0.000569970754 0.00653304299 0.0325520635 0.0399029963 -0.0806222185 -0.0268758014 0.0137116844 0.0357208215 -0.0446989201 0.101663925 -0.155037418 0.15481323 0.0920236707 -0.0203882046 -0.0223031957 -0.207311168 0.0299781989 -0.025802305 -0.0148988022 0.0638434514 0.00620582933 0.0474899374 0.0742334872 0.0698142126 0.0770208836 0.0671139732 0.0712592602 -0.0104929302 -0.00591508951 -0.00255784672 -0.0738764703 -0.0553255491 -0.029572593 -0.0615333244 -0.0357224196 0.0179582983 0.0830319971 0.0895719156 0.10687498 -0.00571668614 0.00689404458 0.0462640077 0.07701464 0.00537824724 0.0231056958 0.0642113984 -0.0161306504 0.0346609764 -0.00119924161 0.0337541439 0.0144789573 -0.0303051267 -0.132072628 -0.0616089888 0.000352514733 -0.192770779 -0.105199195 -0.0428788401 -0.150706217 -0.237319574 0.056610696 -0.109073006 -0.0375075005 0.0529380031 0.153346628 0.0121154897 -0.185275063 -0.266600043 -0.171582744 0.195368975 0.156708762 0.0316773616 0.07
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
5
train
5
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
10
gradWeight
2015-05-16 17:48:05 +12:00
4
27
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
32 288
288 1
2016-01-23 13:36:54 +13:00
1
4
28
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
9216
2016-03-27 21:15:14 +13:00
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2015-05-16 17:48:05 +12:00
2
5
2016-03-27 21:15:14 +13:00
_type
2
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
2
4
2016-03-27 21:15:14 +13:00
bias
4
29
2016-01-23 13:36:54 +13:00
3
V 1
16
torch.CudaTensor
2016-03-27 21:15:14 +13:00
1
32
1
2016-01-23 13:36:54 +13:00
1
4
30
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
2016-03-27 21:15:14 +13:00
32
-0.00132384081 -0.00344423368 -0.0394922122 -0.00964759756 -0.033902958 -0.0116256075 -0.0460234247 0.0164736081 0.10529957 -0.0162672233 0.00237166998 -0.000373113202 0.0236097649 0.00496797124 0.00630606851 0.0243535694 -0.247873276 0.0029400147 -0.00447470322 -0.0199754257 -0.052990716 -0.0326747634 -0.177702963 0.00835684501 -0.0943474695 -0.0172563363 -0.0138601838 0.00209668418 -0.0123849213 -0.0456500351 0.0171501823 -0.02814354
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
8
gradBias
2016-01-23 13:36:54 +13:00
4
31
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-03-27 21:15:14 +13:00
32
1
1
4
32
2016-03-27 21:15:14 +13:00
3
V 1
17
torch.CudaStorage
32
0 0 0 0 -1.52919712e-12 1.28913685e-10 -3.71987065e-16 8.71783375e-12 2.56229564e-23 2.5783727e-14 -1.26235578e-09 -5.53569237e-25 1.25325522e-17 -2.26395264e-13 -1.91967509e-13 -8.70900913e-18 1.58536854e-16 3.63058151e-12 1.10780846e-18 1.35604232e-14 6.29832075e-10 2.44191594e-15 2.14760765e-14 -1.82017118e-12 -1.1374771e-16 -9.64144418e-25 -1.36045419e-09 -3.67763925e-13 2.33650026e-05 8.49650385e-12 -8.19205374e-15 3.3510794e-12
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
4
padH
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
4
4
33
2015-05-16 17:48:05 +12:00
3
V 1
12
nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
34
2016-03-27 21:15:14 +13:00
6
2016-01-23 13:36:54 +13:00
2
7
inplace
2016-01-23 13:36:54 +13:00
5
1
2
2015-05-16 17:48:05 +12:00
5
_type
2015-05-16 17:48:05 +12:00
2
16
torch.CudaTensor
2
6
negval
2015-05-16 17:48:05 +12:00
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
35
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
6
output
2015-05-16 17:48:05 +12:00
4
36
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
5
train
5
0
2015-05-16 17:48:05 +12:00
1
5
4
37
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
38
2016-03-27 21:15:14 +13:00
16
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
2
dH
1
1
2
2
dW
1
1
2
11
nInputPlane
1
32
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
39
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2
2016-03-27 21:15:14 +13:00
2
kH
1
3
2
2016-01-23 13:36:54 +13:00
9
gradInput
4
40
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
kW
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
3
2015-05-16 17:48:05 +12:00
2
12
nOutputPlane
2015-05-16 17:48:05 +12:00
1
64
2015-05-16 17:48:05 +12:00
2
4
2016-03-27 21:15:14 +13:00
padW
2015-05-16 17:48:05 +12:00
1
0
2
2016-01-23 13:36:54 +13:00
6
weight
2015-05-16 17:48:05 +12:00
4
41
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
2
64 288
288 1
2015-05-16 17:48:05 +12:00
1
4
42
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
18432
2016-03-27 21:15:14 +13:00
-0.132789627 -0.152469471 0.0625797212 0.0376783311 0.226782009 0.272415191 -0.00494563906 0.00103140215 -0.146276638 -0.0335593373 -0.0201606639 -0.147524163 0.0284295958 -0.0516806655 -0.0530493669 0.0251825191 -0.047394637 0.095182091 0.0790583268 -0.150137916 -0.0701677799 0.184083119 0.267225623 -0.138477713 0.0662729368 0.290367246 -0.185078368 -0.0240302775 0.00188864325 -0.00893039629 0.0623296611 -0.0248939786 -0.00366679044 0.0127566978 0.127870813 -0.033232443 -0.00707248552 0.0687894747 0.00202841288 -0.147007048 0.352625072 0.190973133 0.156140119 0.176588923 -0.0254433211 0.0645569265 0.0287655368 0.0213664509 -0.028513262 0.260846019 0.180930525 -0.0299606938 -0.101599291 -0.0477329902 0.118297666 -0.381276667 -0.259746492 0.161068901 0.170469955 -0.360756665 0.413942307 0.43720746 0.120546721 -0.0844969675 -0.256984115 -0.108457685 -0.146833777 -0.0326621383 -0.112511434 -0.118573911 -0.0227109585 -0.21597968 -0.0933912471 -0.104717255 0.0212090518 -0.0654146597 -0.045512218 0.0710904673 0.0422748849 0.0819015279 0.103943586 -0.0008381019 0.00739931874 0.0355693586 -0.0197960716 0.0398308635 0.103176318 0.0359701328 0.0607980415 0.0984802842 -0.0663833916 -0.119135536 -0.000623426982 -0.0520625748 -0.201290607 0.110270783 -0.0256964136 0.0227817744 0.0967105851 -0.0106940893 0.0661169291 -0.0096158376 0.0222065989 -0.0779137462 -0.00169565517 0.0242957827 0.115305118 0.19064635 0.00339599559 0.0288791191 0.0926911831 -0.00855293963 0.0807923973 0.00274419435 -0.0648520514 -0.0505325496 0.0332128853 0.0505264066 0.394112706 -0.19518882 0.199080616 -0.0697017983 0.0392520502 -0.0824396983 -0.621216595 -0.219179094 -0.127979875 0.0924842805 -0.000940995233 0.0045370874 0.0182977188 -0.057815358 -0.0273403376 0.0327448174 0.091602549 -0.0647193566 0.171424806 0.0858488902 -0.0360876136 0.0192167722 -0.336725801 0.0176319666 -0.0550212823 -0.518423915 -0.204985365 0.102084957 -0.0261338484 0.316655636 -0.0620861724 0.0266590007 0.041426681 -0.232372075 -0.0320427716 -0.0620680749 -0.0273458883 -0.0301268138 -0.0361710601 0.06249623 0.0842787474 0.0346760824 -3.79912744e-05 0.026592087 0.174068868 0.155483559 -0.034967348 0.0877109915 -0.071272172 -0.104005545 -0.116993114 -0.135277927 -0.161374956 0.0108203003 -0.0925972834 -0.0734174326 -0.400227606 -0.104538523 -0.403188169 -0.106477827 -0.883409023 -0.53658396 -0.212605804 -0.0422150157 0.0104993964 -0.0508152582 -0.192294091 0.092531532 -0.0191679727 0.121287301 0.297855109 -0.245576158 -0.0114672901 0.316056371 -0.0501550697 0.0503108278 0.30502066 -0.311887413 -0.0104956245 -0.117064357 -0.0193519797 0.00778887607 0.173550352 -0.032343924 -0.113963641 0.150270805 -0.127385393 0.357393831 0.0653399229 -0.184386954 -0.225319356 0.0148257259 -0.0143789286 0.101126313 -0.0682300106 -0.0964291021 0.0726310685 0.0241635088 0.101521425 0.105350479 -0.0287619103 0.0627783462 0.0692849159 -0.0938497409 -0.0481913425 -0.108856857 -0.0668901205 -0.193137795 -0.128553018 -0.148857743 -0.0165282004 -0.163208649 -0.443788022 0.080164969 0.070148848 -0.0970593393 -0.0830162466 0.00438823923 -0.0340323634 -0.019354187 -0.0679862872 -0.0572691932 0.0660815164 -0.00626125699 -0.0195680242 0.0229762327 0.0783642977 0.0514519066 0.106541924 0.165877759 0.141418591 0.0512529425 0.0204501022 0.0615573153 -0.115061343 -0.0596411675 -0.0349350385 0.00358516583 -0.194586262 0.0282285027 0.101479016 -0.00744010275 0.134286687 -0.153817877 -0.0683223829 0.184652314 -0.12298739 0.50530386 0.0792980194 -0.124049984 0.623762071 0.0507702492 -0.0255035181 0.0373866782 -0.00286946516 -0.0544758439 0.190692231 -0.00526557909 -0.0566102602 0.0586653724 -0.415963858 0.104555123 -0.220193252 0.0505010746 0.0597594529 -0.199338928 0.180315688 -0.264750957 -0.0253576003 0.0807771087 0.051847674 0.114006281 0.0492769293 -0.0720537603 -0.0169353373 0.0982961357 -0.0066196206 0.0218829736 0.130754113 -0.0736880228 -0.122502796 0.103395663 -0.00262493663 -0.235024273 0.0191434156 0.147284389 -0.135388106 0.00460766675 0.0533839837 -0.00286635943 -0.131213889 -0.0891767442 0.0144882649 0.1567594
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
10
gradWeight
4
43
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
2
64 288
288 1
2016-01-23 13:36:54 +13:00
1
4
44
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
18432
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2016-03-27 21:15:14 +13:00
2
5
_type
2
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
2
4
2016-03-27 21:15:14 +13:00
bias
4
45
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-03-27 21:15:14 +13:00
64
1
1
4
46
2016-03-27 21:15:14 +13:00
3
V 1
17
torch.CudaStorage
64
0.0264734644 0.00816035923 0.0206836835 -0.000252836442 0.0174133461 -0.00765001075 -0.000555174658 -0.0169738196 0.0313083045 -0.0237838533 -0.0110224411 -0.00145795441 -0.0190424304 -0.00769775128 0.0165005513 0.00627836958 8.88656359e-05 0.00200100313 0.00520492159 -0.00574251218 0.0196575187 0.000387709239 -0.0570803098 0.00787345599 -0.00661598705 0.00900525786 -0.0359034128 0.0120987156 -0.00079422415 -0.0375892632 0.00657900795 -0.00904147048 -0.0528435521 0.011615782 -0.000647827226 -0.0137551324 -0.0116028506 -0.000317042373 -0.0223500095 -0.00905874185 -0.0184780229 0.0169718843 -0.0530708991 -0.00847028196 0.00684394455 -0.018807061 -0.0018402799 0.0129027469 -0.000490908977 -0.000347654917 0.0045023174 0.00151954964 0.00741293421 -0.0193045642 -0.0321075246 0.00651028519 -0.0272354372 -0.0394220576 -0.013170029 -0.00661093369 -0.0130162332 0.0140172439 0.00629138295 0.0130806835
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
8
gradBias
2016-01-23 13:36:54 +13:00
4
47
2016-03-27 21:15:14 +13:00
3
V 1
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
1
2016-03-27 21:15:14 +13:00
64
2015-05-16 17:48:05 +12:00
1
1
2016-03-27 21:15:14 +13:00
4
48
2016-03-27 21:15:14 +13:00
3
V 1
17
torch.CudaStorage
64
0 0 0 0 1.44865431e-09 2.14426348e-15 -1.40916337e-12 -1.19082139e-15 9.36346334e-10 -1.28824267e-22 -5.78590317e-14 -7.05066139e-09 8.94607479e-17 -3.59801381e-12 1.40104844e-10 9.74763889e-13 1.83353649e-07 6.79487109e-12 1.29352014e-08 3.27402211e-12 7.91648756e-13 -5.30167663e-12 -2.45423369e-11 2.14107378e-11 -2.82955499e-11 -2.25216303e-13 4.79471801e-17 4.73082526e-17 -2.25550103e-08 1.7384989e-09 -1.36712384e-08 -4.65001388e-12 1.15986678e-14 -1.71374982e-17 -3.14808875e-14 3.18820653e-13 1.01117315e-09 -4.22824531e-09 -7.9789586e-10 -8.8503857e-11 -2.21740883e-08 -2.32779821e-23 -9.2055252e-06 3.99038067e-20 2.95097323e-07 -2.78381656e-08 1.89515115e-09 8.76508088e-10 -2.15495781e-11 -5.98454186e-12 -2.27958014e-10 4.53940772e-17 -4.02973476e-17 5.62322833e-16 2.82318922e-07 1.82210306e-14 -1.16918519e-12 3.18877886e-14 1.84092436e-10 -9.15497985e-16 1.25714877e-07 -6.24641033e-11 3.69972733e-11 2.10144879e-12
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
4
padH
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
0
2015-05-16 17:48:05 +12:00
1
6
4
49
2015-05-16 17:48:05 +12:00
3
V 1
12
nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
50
2016-03-27 21:15:14 +13:00
6
2016-01-23 13:36:54 +13:00
2
7
inplace
2016-01-23 13:36:54 +13:00
5
1
2
2015-05-16 17:48:05 +12:00
5
_type
2015-05-16 17:48:05 +12:00
2
16
torch.CudaTensor
2
6
negval
2015-05-16 17:48:05 +12:00
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
51
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
6
output
2015-05-16 17:48:05 +12:00
4
52
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
5
train
5
0
2015-05-16 17:48:05 +12:00
1
7
4
53
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
54
2016-03-27 21:15:14 +13:00
16
2
2016-03-27 21:15:14 +13:00
2
dH
1
1
2
2
dW
1
2015-05-16 17:48:05 +12:00
1
2
11
nInputPlane
1
64
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
55
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
2
kH
1
3
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
56
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
kW
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
3
2015-05-16 17:48:05 +12:00
2
12
nOutputPlane
1
64
2
2015-05-16 17:48:05 +12:00
4
2016-03-27 21:15:14 +13:00
padW
2015-05-16 17:48:05 +12:00
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
6
weight
2015-05-16 17:48:05 +12:00
4
57
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
2
64 576
576 1
2015-05-16 17:48:05 +12:00
1
4
58
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
36864
2016-03-27 21:15:14 +13:00
0.026572153 0.0668337643 0.00143410789 0.0446672291 -0.0798001587 0.0225247499 0.00012538569 0.0222574137 0.0634558946 -0.167554066 0.00415156875 -0.045605626 0.164322019 -0.0352405719 0.0377729833 0.0389885604 0.0588522255 -0.0713945553 0.237841874 -0.134581566 -0.0707350001 -0.0370001048 -0.202631027 0.0333517008 0.104553491 -0.254804254 0.0518143475 -0.0391699485 -0.160347745 0.0612191223 0.0565335229 -0.177545711 0.0280775912 -0.0341862813 0.0449945852 0.0396317542 0.0785646439 -0.266556859 0.0364719033 0.037824247 -0.0485591143 0.0858581141 0.0166305881 0.0334401689 0.0935866088 -0.0585394613 0.628153443 -0.0431787968 -0.00294269994 0.130802885 0.0422540233 0.0184009615 -0.0786825493 0.0330247357 0.0816508159 0.0116929384 -0.218191847 -0.00789721776 -0.0677130371 -0.0955521241 0.117342018 0.00713981921 0.0991603509 -0.0314819664 0.0578101464 0.120377138 -0.0458955131 -0.00827180687 -0.017562246 0.0949566588 -0.150726274 0.0718940869 0.0285980683 0.000568096351 -0.181377649 0.0754467621 -0.130265266 -0.278389603 0.0729194432 -0.118941277 0.0440650284 -0.0990331098 -0.00699980557 -0.170117348 0.0787599459 0.0126095172 0.00518887863 0.0512943044 0.0783670247 0.0423074737 -0.049327977 -0.108350575 -0.13370268 0.123788573 0.0747396573 0.0675733015 0.0334069543 0.0663995072 0.0337890238 0.145680472 0.156791031 0.108089603 -0.0645771995 0.169008106 -0.0435730219 0.233210176 0.135586202 -0.0494523384 0.0902898759 0.0620796122 0.15202719 0.0187730957 -0.046306096 0.297412515 -0.279087484 -0.138245404 0.0256173462 -0.106136277 0.109047391 -0.0390800089 -0.143983737 -0.457743168 -0.390403271 -0.161181822 -0.0500333346 0.0128127802 0.212326184 0.00291213393 -0.0761030614 -0.0187634416 0.049197875 -0.0966437757 0.0174203478 0.0852090269 -0.0321750604 0.0231645964 0.111872576 0.00952902809 0.0243364535 -0.0723303631 0.113260284 -0.0657749027 -0.0345375575 0.0685364231 0.107330509 0.0399852507 0.0602416098 0.0250396803 0.00698762015 0.0368477777 0.0162753575 -0.00256931572 0.0236688908 0.0504302494 0.149495184 0.134568721 0.0834962577 0.101737157 -0.0233655684 0.03657021 -0.075868383 -0.0522464365 -0.0649615452 0.00158606027 -0.0990866795 0.0516435243 0.0126567464 0.0246065669 0.0274915639 -0.0225343332 -0.0465514213 -0.0898829624 0.037333481 0.0647752509 -0.0602997504 0.0776325762 -0.100089781 -0.0690935552 0.282758325 -0.012597478 -0.0632798597 -0.167973891 -0.159062833 0.0209548064 0.0439979844 -0.0742889121 -0.00337092672 -0.0460451581 0.0190123897 0.136711314 0.229017749 0.0777349919 0.11855109 0.160740674 0.0758700967 -0.0969780162 0.0819298923 0.0123395724 -0.113164164 -0.278326392 0.269849867 -0.230152696 -0.478301615 0.0352588333 -0.143010199 0.0157325454 0.00236052321 -0.00738645485 -0.0767659917 0.021192221 0.0723463669 0.0804442391 -0.0731127858 -0.0322861075 -0.0402708501 0.0515112206 0.0297320895 -0.0613654442 -0.025389418 0.0305022467 0.000636811892 0.0154660856 0.0576577 0.0432463475 0.0276193246 -0.0952891186 -0.095630087 -0.268186748 0.06743595 -0.0865482315 -0.081024833 -0.10079968 0.152748466 -0.0245968625 -0.202726543 0.162156045 -0.253882498 -0.407480419 0.186004519 0.0517856255 0.00279204105 -0.207884938 0.158742309 -0.0338365696 0.0656281412 0.036160484 0.0584277324 0.0485869907 0.0544758625 -0.0365806706 0.0294319391 0.10367474 0.0850614235 -0.00295654335 -0.0592859499 -0.0418932065 0.0754439831 -0.0346371457 0.0503024794 0.0325369723 -0.0306878332 -0.0858135521 0.00526320329 0.125554889 -0.0498659238 0.0383515917 -0.0213963129 -0.0398781486 -0.0335535146 0.0148168961 -0.06135194 -0.031997323 -0.215131968 0.0545991063 0.144416094 0.166907534 -0.0514015369 0.0634916946 -0.0485892296 -0.039684236 0.211439952 -0.450027496 0.102439173 -0.125636995 -0.106847785 0.0811403245 0.0603663065 0.000380427227 0.0384341553 -0.341255963 -0.596962929 0.0420665592 0.195786461 -0.0921786129 0.00585907232 -0.039244663 0.0287756566 0.046097666 0.00888908003 0.131450891 -0.0432409346 0.0262408275 -0.0104507217 -0.04090764 -0.0982973501 -0.0242316704 0.148609102 -0.32503289 0.143030167 0.216676995 -0.173545942 0.12410222 -0.04
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
10
gradWeight
2015-05-16 17:48:05 +12:00
4
59
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
64 576
576 1
2016-01-23 13:36:54 +13:00
1
4
60
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
36864
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2016-03-27 21:15:14 +13:00
2
5
_type
2
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
2
4
2016-03-27 21:15:14 +13:00
bias
4
61
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-03-27 21:15:14 +13:00
64
2016-01-23 13:36:54 +13:00
1
2015-05-16 17:48:05 +12:00
1
4
62
2015-05-16 17:48:05 +12:00
3
V 1
2016-03-27 21:15:14 +13:00
17
torch.CudaStorage
64
0.00287283817 0.00410895934 -0.0136568602 0.000926387846 -0.000510914775 0.00321400817 0.000803606352 -0.00513593853 0.00586551335 -0.000226130927 -0.00411162106 0.00168238487 -0.00853737071 0.00494164787 -0.00494083809 -0.00215931819 0.00572809996 0.00372767937 -0.0124567291 -0.00145524542 -0.0281030461 0.00273273559 -0.0134136248 0.00221312325 -0.00995087158 0.00256598601 0.000468523882 -0.00670486968 -0.012019597 -0.0127669917 -0.000665903848 0.000718555588 0.00213586376 -0.00873502716 0.00662876293 0.00831089914 -0.00389193324 0.00254328828 0.00046858337 -0.0120671308 -0.00316181034 -0.00294889021 -0.00302230031 -0.00726673566 -0.0165238064 0.00283488794 -0.0150566874 0.00467786193 -0.0112500526 -0.0025048505 -0.000823690905 -0.0138116879 -0.162377387 -0.00668258592 -0.00111048657 0.00591891492 -0.00469855452 -0.00298102899 0.00651084259 -0.0174795277 -0.00283351773 -0.0114791263 -0.0204909593 -0.0232721549
2016-01-23 13:36:54 +13:00
2
2015-05-16 17:48:05 +12:00
8
2016-03-27 21:15:14 +13:00
gradBias
2015-05-16 17:48:05 +12:00
4
63
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-03-27 21:15:14 +13:00
64
1
1
2015-05-16 17:48:05 +12:00
4
64
2015-05-16 17:48:05 +12:00
3
V 1
2016-03-27 21:15:14 +13:00
17
torch.CudaStorage
64
0 0 0 0 5.18902366e-10 -2.71916406e-06 2.73334377e-09 1.79332375e-07 1.29414081e-14 -1.98752694e-08 2.22577389e-14 -7.09905063e-17 1.60504526e-10 1.1057996e-19 1.68907184e-12 -9.25186039e-10 1.44532379e-18 -1.0382405e-11 9.0506258e-10 2.80303523e-11 1.65003189e-09 -7.00884554e-13 2.70003082e-11 -1.45120651e-07 1.67857533e-10 5.88339299e-06 2.89825065e-12 3.34834868e-11 -7.71764121e-07 1.03144995e-08 -7.29800113e-18 1.85474782e-08 2.39320137e-08 -1.81157787e-07 6.3113148e-10 3.23103469e-13 -5.10223752e-10 4.38576963e-13 -4.79679514e-08 8.90640561e-10 9.62105916e-12 -8.05439804e-09 -6.58231358e-10 4.00555369e-16 1.36715439e-08 4.89031298e-12 1.84966015e-10 -1.97938783e-11 -1.19932553e-11 1.82181837e-15 2.00619326e-07 -1.52726616e-11 1.51231973e-13 -1.64865455e-15 -1.1258716e-18 -5.95539706e-09 2.81545655e-11 6.20516878e-19 -1.20570832e-07 8.19314216e-09 -5.61108035e-11 -2.42737338e-26 6.75641042e-18 -8.11072431e-10
2015-05-16 17:48:05 +12:00
2
4
2016-03-27 21:15:14 +13:00
padH
2015-05-16 17:48:05 +12:00
1
0
1
2016-03-27 21:15:14 +13:00
8
2015-05-16 17:48:05 +12:00
4
65
2015-05-16 17:48:05 +12:00
3
V 1
12
nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
66
2016-03-27 21:15:14 +13:00
6
2
7
inplace
2016-03-27 21:15:14 +13:00
5
1
2
2016-03-27 21:15:14 +13:00
5
_type
2015-05-16 17:48:05 +12:00
2
16
torch.CudaTensor
2
6
negval
2015-05-16 17:48:05 +12:00
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
67
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2
6
output
2015-05-16 17:48:05 +12:00
4
68
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-03-27 21:15:14 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-03-27 21:15:14 +13:00
0
2
5
train
5
0
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
9
2016-01-23 13:36:54 +13:00
4
69
2016-01-23 13:36:54 +13:00
3
V 1
2016-03-27 21:15:14 +13:00
23
nn.SpatialConvolutionMM
3
70
2016-03-27 21:15:14 +13:00
16
2
2
dH
1
1
2016-01-23 13:36:54 +13:00
2
2
dW
1
1
2
2016-03-27 21:15:14 +13:00
11
nInputPlane
2015-05-16 17:48:05 +12:00
1
2016-03-27 21:15:14 +13:00
64
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
71
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-03-27 21:15:14 +13:00
0
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
0
2016-01-23 13:36:54 +13:00
2
2
kH
1
3
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
72
3
V 1
16
torch.CudaTensor
0
2015-05-16 17:48:05 +12:00
1
0
2
2016-03-27 21:15:14 +13:00
2
kW
1
3
2
12
nOutputPlane
1
128
2
4
padW
1
0
2
2016-01-23 13:36:54 +13:00
6
weight
2015-05-16 17:48:05 +12:00
4
73
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
2
128 576
576 1
2015-05-16 17:48:05 +12:00
1
4
74
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
73728
2016-03-27 21:15:14 +13:00
0.0136647951 -0.0716358349 -0.0288348235 0.0317118689 0.0660200566 0.00940165669 -0.00185399107 0.0909731686 0.064246282 0.0104914661 0.098336488 0.0587775186 -0.0906640217 0.113871992 -0.0544623137 0.0249879807 0.0398510657 -0.0384023413 -0.191337034 -0.249538094 2.08761739e-05 -0.0941151381 0.0119282231 -0.0463649891 0.0587516911 -0.0502966307 0.0220767856 -0.0473187491 0.00252326461 -0.110454202 -0.0580095388 -0.0298291072 -0.0503852181 -0.0413115844 -0.109072804 -0.171224147 0.116376109 -0.00844355766 -0.0750791803 0.0661038458 -0.111096278 -0.0102290241 0.0417858697 -0.010916302 0.0587222427 0.0116869155 -0.0346367881 0.0455410182 -0.0208585896 -0.0101586236 0.0146371722 0.0190281253 0.00729728816 -0.00495655183 0.0723022521 -0.0599194095 -0.00660447543 0.0879389718 -0.0774955526 0.045780506 0.114691399 0.0818340257 -0.0513675548 0.0167149715 -0.0253158696 0.0437394902 0.087211594 0.077252239 -0.00352729205 0.0625508726 0.0700638369 0.0331025943 -0.126825616 -0.0415726155 -0.0543128662 -0.12688525 -0.000688834523 -0.0304036867 -0.194727138 -0.0956733227 -0.0270933956 0.0429943129 -0.138676167 0.135387897 0.0590893999 0.0886129215 -0.0125925671 -0.0453875326 -0.114471748 -0.0252545271 -0.102009386 0.0611154735 -0.0277574342 -0.0181986727 0.00779044302 -0.0388392396 0.21984981 0.158445492 -0.00747636938 -0.0499174222 0.0554419979 -0.119788483 -0.0307305027 0.0769117102 -0.0953646153 0.0846082866 0.0570672341 0.0196515173 -0.00401271181 -0.00342400884 0.0381392017 -0.0666364729 0.0159733202 -0.109485038 -0.00702800881 0.00493968045 -0.0703506097 -0.0607146546 -0.00120412558 -0.070826456 0.117941156 -0.0423465706 0.0961770937 0.0786962658 0.109681875 0.0955849588 -0.0545046292 -0.0486753806 -0.0423044935 -0.0863459632 0.00309452997 0.184918061 -0.0287742615 0.0281175189 0.0587917194 -0.14021109 -0.0699648485 -0.0189733598 -0.0853689313 -0.0215774085 -0.051059287 -0.209663242 -0.137086943 -0.101769194 0.0095329117 0.140271664 -0.0936140642 -0.0116330488 0.00849843677 0.0494200885 0.0550309718 -0.0421724841 -0.028131526 0.00462025777 -0.0150060384 0.0225213412 -0.0222815629 0.0134227164 -0.0791984573 -0.129961267 -0.104680479 0.042522572 -0.101738036 0.0207173601 -0.0163061451 -0.0553694256 0.0603660084 -0.000239756177 0.0389653668 -0.0192172006 -0.0369227231 -0.0101703368 -0.00918281171 0.0552964099 0.0453204513 -0.00740995863 0.0300890747 0.0276021697 0.0479245782 -0.0193119813 0.00402164552 0.0235502608 0.136199415 -0.0346992388 -0.0347502306 0.21992439 -0.0314712897 0.0281627849 0.155656219 0.108293764 0.0896248072 -0.0752153546 -0.00463236915 -0.0230905619 -0.0620365143 -0.15060094 0.0215908214 0.0187315047 -0.096235998 0.0691977069 0.0106838513 -0.0848643631 0.000736205431 0.0600826591 -0.0717485249 -0.0523550063 -0.0310533755 0.00206340011 -0.04259561 -0.0866293907 -0.0382597968 0.108855173 -0.0451730378 0.130112723 0.0730466619 0.0213081613 0.0112667102 0.030730851 -0.04019868 -0.0368275605 -0.014019263 -0.0174233429 0.00455610594 -0.0183935799 -0.0233897045 -0.0109843398 -0.00179808075 -0.0723103732 0.054168392 0.0678915083 -0.00748898555 -0.099275358 0.0248559117 -0.0978201181 0.0150101781 0.0830738097 0.0358866043 -0.021164719 -0.0746106133 0.0367861167 -0.105962932 -0.0635562763 0.00616766 0.0880765617 0.021577742 0.00539289741 0.0426731966 -0.104045019 0.0395238735 0.0273667518 0.0199601054 0.12280871 0.035065461 0.171942204 -0.00271325861 -0.0648721084 0.0480639264 -0.00308718812 -0.0108268037 0.220838889 0.0815603361 -0.105586067 0.183552936 -0.124751993 -0.0251015946 0.073286362 0.0866710618 0.0585659407 0.116252705 -0.0579895265 -0.0406101719 0.0489366576 0.0620254017 -0.0773374066 0.0129663339 -0.0924913958 -0.00927953701 -0.0609708391 -0.0598672107 0.00699465256 -0.0226449221 -0.0131474314 0.11579188 -0.0297447406 0.0267202873 0.034310028 -0.0246088523 -0.0338712111 0.0443330035 -0.0714647248 -0.0476263911 -0.100119755 -0.114521012 0.012346291 -0.0298403576 -0.0509607382 -0.0682035908 0.0325669348 -0.187000573 -0.00735622039 0.0372700356 -0.163860425 -0.155266896 -0.0415716693 -0.133120224 0.0598003715
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
10
gradWeight
2015-05-16 17:48:05 +12:00
4
75
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
128 576
576 1
2016-01-23 13:36:54 +13:00
1
2015-05-16 17:48:05 +12:00
4
76
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
73728
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
5
_type
2
16
torch.CudaTensor
2
4
2016-03-27 21:15:14 +13:00
bias
4
77
3
V 1
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
1
2016-03-27 21:15:14 +13:00
128
1
1
4
78
2016-03-27 21:15:14 +13:00
3
V 1
17
torch.CudaStorage
128
0.00341775105 -0.0123526324 0.0021114354 0.0042190617 -0.00703899702 -0.000429547159 0.0224375464 -0.0124805896 -0.016585499 -0.0205210447 -0.0263521727 0.00237781811 -0.00917536858 0.00581391528 -0.000472862797 -0.0252343155 -0.0118678333 -0.0188612249 0.00331203337 -0.0122704701 -0.00812184531 -0.0127454568 0.00119134947 -0.00636684243 -0.0209472831 0.00105610199 -0.00837289169 0.00107208837 -0.0122553017 -0.00600566389 -0.00764326379 -0.012898894 -0.00968901441 -0.000126728613 -0.00188319746 0.0145658171 -0.0030793685 -0.00340710883 -0.0130291069 0.0060741147 -0.00160745531 -0.000769623672 -0.0199531391 0.00832263473 -0.0084215682 -0.00433288375 0.000305014575 0.00221228041 -0.000843319751 0.00796727464 -0.0181250274 -0.00883137528 0.000608171453 0.00494095962 0.00153386651 -0.0297539625 -0.00820906833 -0.0129946824 -0.023935793 -0.00512591051 0.0105631165 0.00121986982 -0.0043469728 -0.00517952535 -0.0100395652 -0.0109183285 0.000481845665 -0.00431947224 -0.0203338973 1.99596452e-05 -0.012179913 -0.00808515493 -0.0234239176 -0.0160897728 -0.00232322561 4.15048598e-05 -0.00577304 -0.0130171962 -0.000972081674 -0.0107003152 -0.0115221618 0.000518861925 0.00231671124 0.000744350487 -0.00664923852 0.00673086988 -0.00536536612 -0.00726034818 -0.000897893507 -0.00292207696 -0.0206724275 -0.00822472293 -0.00595955784 0.00970744994 -0.0379348584 0.00448895106 -0.00464991713 -0.0185567625 0.0040590805 0.00721422024 -0.000189296989 -0.00573624484 0.00578719517 0.00329980068 -0.0116418377 -0.000238796798 0.000874469231 -0.00189570303 -0.0111084506 -0.000740644871 -0.00628482923 0.00497214217 -0.0187645368 0.00152426551 -0.00789819844 -0.0171666835 -0.0478483178 -0.00671117334 -0.00622730097 2.65463314e-06 -0.0207043607 -0.0103657236 -0.0116332229 -0.00871771574 0.00492599187 -0.00910943374 -0.0240263045 -0.0116440933
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
8
gradBias
2016-01-23 13:36:54 +13:00
4
79
2016-03-27 21:15:14 +13:00
3
V 1
16
torch.CudaTensor
1
2016-03-27 21:15:14 +13:00
128
1
2015-05-16 17:48:05 +12:00
1
2016-03-27 21:15:14 +13:00
4
80
2016-03-27 21:15:14 +13:00
3
V 1
17
torch.CudaStorage
128
0 0 0 0 -2.37406145e-12 -1.29588414e-17 -4.65815275e-10 1.37061705e-14 -5.77166023e-11 8.58124302e-19 6.26106532e-16 1.83281133e-11 -5.73961434e-10 9.32609719e-12 -4.61609848e-11 2.12792254e-18 2.29486458e-10 -4.08945933e-10 5.27794646e-22 -6.99532527e-14 -2.3864412e-13 3.81697007e-13 -8.9694118e-21 7.23637696e-15 2.13450901e-09 1.57233493e-09 2.37173525e-09 -8.43846884e-14 -1.02911345e-21 -6.83362672e-11 1.74151338e-09 1.64217841e-11 -5.16842178e-08 8.02540846e-12 6.50627437e-15 3.61776277e-12 2.32987186e-13 -5.72156797e-13 -3.04818835e-16 1.94902716e-09 7.15945619e-12 6.17836737e-16 -1.30367497e-15 1.00266488e-13 -1.06912024e-13 1.32298766e-07 -4.68137248e-12 -3.22156581e-12 3.45256324e-10 2.14590272e-17 7.14166837e-10 4.57000017e-17 -1.66923435e-08 4.03751358e-07 -2.0468935e-08 -4.17429719e-12 -1.66447654e-11 -5.75865844e-09 8.2468421e-10 -1.12590867e-10 -2.15218199e-09 -6.31261796e-12 -5.21961407e-09 -1.61474668e-06 -2.07661284e-11 -6.84319019e-08 3.05396194e-11 -1.12811871e-09 -8.15573581e-11 5.92450893e-13 -1.26684739e-07 1.74422354e-07 2.37507733e-14 2.14878199e-10 -3.7593253e-07 -2.21458024e-13 -4.21356435e-06 7.52104906e-11 -3.67905522e-13 -1.29795717e-05 -3.07550962e-08 2.54714827e-08 1.69717053e-16 9.11436082e-09 2.08685016e-08 -2.89953483e-08 1.32546595e-11 -4.52486518e-13 -0.000123120932 -1.09481838e-14 -7.26602223e-10 6.32974711e-14 -7.05101233e-13 -2.77095846e-10 -3.76643315e-20 5.39840864e-14 -1.92918614e-09 -4.75258464e-14 -9.11991619e-13 1.32898743e-21 -4.03767663e-14 6.69620744e-14 -9.31631483e-09 -5.67686309e-15 4.93656574e-11 -2.30401264e-21 -1.78003736e-16 1.76486428e-17 1.38306155e-11 1.51597596e-10 8.45391782e-24 -1.59049302e-10 -2.96879654e-13 8.27383384e-10 2.00967671e-12 -3.4819031e-16 -8.39674857e-11 -9.69726743e-10 5.05780875e-08 -5.92689902e-19 -3.681589e-13 -9.02006768e-12 -5.76924012e-12 -9.41802886e-11 -1.45977489e-14 7.70959055e-15 -9.01868979e-09 -2.98369995e-09
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
4
padH
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
0
2015-05-16 17:48:05 +12:00
1
10
4
81
2015-05-16 17:48:05 +12:00
3
V 1
12
nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
82
2016-03-27 21:15:14 +13:00
6
2016-01-23 13:36:54 +13:00
2
7
inplace
2015-05-16 17:48:05 +12:00
5
1
2
2016-01-23 13:36:54 +13:00
5
_type
2015-05-16 17:48:05 +12:00
2
16
torch.CudaTensor
2
6
negval
2015-05-16 17:48:05 +12:00
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
83
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
6
output
2015-05-16 17:48:05 +12:00
4
84
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
5
train
5
0
2015-05-16 17:48:05 +12:00
1
11
4
85
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
86
2016-03-27 21:15:14 +13:00
16
2
2016-03-27 21:15:14 +13:00
2
dH
1
1
2
2
dW
1
2015-05-16 17:48:05 +12:00
1
2
11
nInputPlane
1
128
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
87
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
2
kH
1
3
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
88
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
kW
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
3
2015-05-16 17:48:05 +12:00
2
12
nOutputPlane
1
128
2
2016-03-27 21:15:14 +13:00
4
padW
2016-01-23 13:36:54 +13:00
1
0
2015-05-16 17:48:05 +12:00
2
6
weight
2015-05-16 17:48:05 +12:00
4
89
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
128 1152
1152 1
2016-01-23 13:36:54 +13:00
1
4
90
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
147456
2016-03-27 21:15:14 +13:00
-0.0285263527 -0.0352377258 -0.0068745357 0.0385964178 0.0681571737 0.00163682119 -0.0229605064 0.0299100652 -0.0398273505 0.0252056718 -0.0302461702 -0.0368794687 0.0965413451 -0.0489096828 0.0338300355 -0.0431947149 -0.0453068763 -0.0289970972 -0.0562364571 -0.0292611551 0.00826352183 -0.0571052656 -0.00349123031 -0.0319858268 0.0237773191 0.0519115031 0.0271041933 -0.0287844446 0.0502938963 0.024093762 -0.0490158722 -0.0421848334 -0.0493305326 -0.0126610911 0.0347886942 -0.0405277833 0.00139986095 -0.0484467968 -0.00973371044 -0.0509137139 -0.00403692899 0.0236946419 0.032898441 -0.0607111864 0.0154694784 -0.00867180713 -0.0312726237 -0.0834830776 -0.0320407264 -0.0196592342 -0.0108755268 0.053546682 -0.0320858955 0.0743539929 0.0113033419 -0.0352136604 0.0207757652 0.0260963161 -0.00698771607 0.0503740534 -0.0499158166 0.0347664915 -0.0777802914 0.0611919053 0.0260557607 -0.0159973651 0.0163975563 0.0605180897 -0.0880150571 -0.0587118827 0.0528045148 -0.00597707741 -0.0226793047 0.0138008324 0.0778757855 0.0269968323 -0.121922404 -0.0118294191 0.0644617975 -0.0558213405 0.00735422783 0.0133092413 -0.0525780581 -0.0272896737 -0.0217465349 0.0311494097 0.0308163743 -0.00317765307 -0.0185680892 0.0220640227 -0.0860864148 0.0346564762 0.208350748 -0.0306521766 0.0232481938 0.0405101664 0.0155287357 0.0472926758 0.0304432567 0.0243266635 0.0805681497 0.0315761305 -0.0475201458 -0.0209628977 -0.0394456163 -0.0386636741 -0.095523335 -0.0614358149 -0.0469568744 0.0502205715 -0.0288700722 0.0125019001 0.0784243867 0.0784693733 0.0203716177 0.074906759 -0.0140406471 -0.00156721182 0.0684082285 0.0433491468 0.0301453006 0.000719074975 -0.0241397917 0.0246584173 -0.0268965457 0.0854623765 0.0132983178 0.10786216 0.00213228585 0.0580387861 0.0452570021 0.0491325036 -0.0446618311 -0.0524139144 -0.0362909772 0.0603588596 0.0922028944 -0.0257350709 0.0235825609 0.0241773222 0.120573722 0.0648113266 0.0413780659 0.0970383212 -0.0887024775 0.0318525806 0.0626707301 -0.0113241468 0.0385477543 -0.0238354895 0.0590190664 -0.00914060604 -0.00780702662 0.0812001824 -0.023945462 -0.00185401365 0.0288114753 0.0432457067 -0.031319458 -0.0634160936 -0.0402003005 0.0360022932 0.031021934 -0.014506436 -0.0264474787 0.00864196196 0.004944182 -0.0306590628 0.0746803209 -0.000384552375 0.0373826027 -0.0365746729 -0.00962374546 -0.0373223312 -0.0365908034 0.0138256149 -0.0475915223 -0.0138955638 -0.0414929092 -0.0527108833 0.0125509398 -0.115708411 0.0149747841 -0.0373532772 -0.00752768992 0.0519992337 0.0337878205 0.0382461101 -0.00113613636 -0.02057763 -0.0142610418 0.00984007213 -0.0123999594 0.000783644093 0.0871442109 -0.0177232325 -0.0381582044 0.0198077299 -0.0285516027 0.0197982602 -0.0525319092 0.0785147697 -0.0366660133 -0.0319276154 -0.0225257296 0.043784637 0.0730065256 -0.0561159514 -0.0704459772 -0.00355171156 0.0392293334 0.00769583089 0.0360441655 0.00735548604 0.0363528989 0.00487199705 -0.0236793384 0.00560006313 0.067259483 -0.00201475853 0.0136870528 0.0440514423 0.0608639009 0.0567733869 -0.0233592708 -0.0351841301 -0.0370223373 0.0183653403 0.0481178649 0.0520636439 -0.0160252117 0.0331580266 -0.0134595484 0.0125184637 0.0590300411 0.0755384564 -0.0310056545 0.0062929485 0.075667724 0.0135932649 -0.00185340841 -0.0318986773 0.0685023963 0.0236716606 0.0195223354 -0.0102081439 -0.0135867419 -0.0525863133 0.0511434972 -0.00632989453 0.0393941887 -0.00683528557 -0.0426177643 -0.0360269062 -0.0630999506 -0.0215002745 -0.0517054647 -0.0351411104 -0.0413899161 -0.021916613 0.0204399228 0.0116974134 0.0205492303 -0.101180814 0.029893415 -0.0367939733 0.11847569 -0.0425230078 0.0342017785 -0.00399051374 0.0172425061 -0.0095185684 0.0266459379 0.0252227709 0.0392601527 -0.0123270424 0.05970487 0.0472708903 0.0435393006 0.0130320555 0.0225883797 -0.00889914576 -0.0723556951 0.0204004999 0.0334834829 0.0461083725 -0.0309751816 -0.0119258612 0.0121268667 -0.0295275282 -0.000421135977 -0.0718739107 -0.068510823 -0.010323382 -0.0116322795 -0.000683317077 0.0577269569 0.0612117648 -0.0375193208 0.0754567981 0.0222664159 0.0614478737 0.031419
2015-05-16 17:48:05 +12:00
2
5
train
5
2015-05-16 17:48:05 +12:00
0
2
2016-01-23 13:36:54 +13:00
10
gradWeight
2015-05-16 17:48:05 +12:00
4
91
2016-01-23 13:36:54 +13:00
3
V 1
16
torch.CudaTensor
2
128 1152
1152 1
1
4
92
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
147456
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2
2016-03-27 21:15:14 +13:00
5
_type
2
16
torch.CudaTensor
2
2016-01-23 13:36:54 +13:00
4
2016-03-27 21:15:14 +13:00
bias
4
93
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-03-27 21:15:14 +13:00
128
1
1
4
94
2016-03-27 21:15:14 +13:00
3
V 1
17
torch.CudaStorage
128
0.0113355163 0.0149319796 0.016723061 0.0127985878 -0.00966645684 -0.000958230579 0.0217821859 -0.0136382589 -0.0922811255 0.00473489566 0.0196744427 0.0152079863 0.0169325806 -0.00586016523 0.00963418465 0.0232133623 0.0160201024 0.0168120284 0.0054396824 0.0170651227 0.00952291675 -0.0830982178 0.0231156889 0.0193503778 -0.00398460962 0.0204227176 0.0191447102 0.0165922679 0.0103171496 0.017503975 0.016213987 -0.0967383534 0.0177563597 0.00482405629 -0.0905043855 0.0159615967 -0.0740483776 -0.10803704 -0.0950520635 0.0119545935 -0.0010901331 0.017805893 0.00450871466 0.0177251417 0.00554739172 0.0164679419 0.0166474078 0.0202358663 0.0157794617 0.0192348696 0.00465277769 0.0170862488 0.0183421373 0.0152489683 0.000345825159 -0.00388412946 0.0134792691 -0.0980254933 0.0214170478 -0.087236762 0.019653799 0.00398037303 0.0213143583 -0.102951244 -0.00106289284 -0.0848783925 0.0219972581 0.0158293676 0.0141868377 0.0207565837 -0.000124416882 0.0231930222 -0.113905549 0.0143944873 0.0182515495 0.0270305574 -0.0122920815 0.0171432234 0.00687775901 0.0239335056 0.016548967 -0.0912528113 0.00848879758 -0.00484039262 0.00737872673 0.0153160375 0.0216792095 0.0106822914 0.0094965687 0.0136672137 0.0153552154 0.015789751 0.0201339535 0.00476962281 0.00882571749 -0.0916571543 -0.0659663007 -0.0108889677 0.0128333131 0.0105054108 0.0165859908 0.0127577642 0.0160462204 0.0087761078 0.019877838 -0.0593221299 -0.105887406 0.0183754154 0.00501813227 0.000465275574 0.0175207853 -0.00500382716 -0.0657817349 0.00688560354 0.0831545293 0.00724622188 0.0247853715 0.015072736 0.0192698408 0.0213011988 0.0529705696 -0.000217877619 0.0123948399 0.0172202084 0.0228580628 0.0134469559 -0.0120413471 0.0144215524
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
8
gradBias
2016-01-23 13:36:54 +13:00
4
95
2016-03-27 21:15:14 +13:00
3
V 1
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
1
2016-03-27 21:15:14 +13:00
128
2015-05-16 17:48:05 +12:00
1
1
2016-03-27 21:15:14 +13:00
4
96
2016-03-27 21:15:14 +13:00
3
V 1
17
torch.CudaStorage
128
0 0 0 0 -2.26336385e-14 7.2452816e-16 -8.92990265e-11 -4.33400604e-09 7.94610376e-18 6.34353767e-12 1.05146326e-11 7.07717711e-14 1.8333276e-13 1.42457799e-11 3.18870418e-13 -1.52048208e-14 -4.87914824e-23 5.37023293e-14 1.62900214e-11 3.57633853e-11 -5.80339509e-12 -8.87204199e-10 8.39869702e-11 1.45407262e-19 -4.02176296e-12 2.73625424e-17 2.52485146e-19 -2.21661474e-11 -5.69256073e-08 5.24619279e-16 8.22538475e-17 -1.96280576e-19 -1.1108483e-14 -3.33950469e-31 -7.07999008e-20 -4.97984815e-15 1.22784404e-11 -3.99249245e-10 -1.91283205e-11 -4.31743739e-14 3.82796184e-10 4.29927087e-29 -7.94208267e-16 -6.59400031e-21 2.02428325e-13 -1.47196677e-09 2.84315005e-18 6.40010354e-12 -3.21910264e-14 7.01720078e-17 2.26635741e-08 5.63783353e-10 9.18966792e-10 9.47204953e-11 9.22156934e-11 4.48605725e-13 1.66586367e-16 -1.4448108e-13 9.16673667e-16 -4.9544804e-12 -4.50185624e-13 -2.07369203e-16 -5.47288384e-14 -3.82121038e-13 9.58781277e-10 1.44806073e-17 2.05364697e-10 -3.1676586e-14 4.63854554e-14 1.97355739e-16 -1.25132326e-17 -1.57625145e-15 -9.89632664e-17 -5.03676165e-17 3.58569215e-22 4.98498279e-17 5.51862653e-11 -1.50219111e-13 -8.98254471e-10 -3.33998517e-15 2.66865152e-09 5.77465986e-09 4.09348277e-10 4.94765381e-22 1.34442611e-14 3.5331174e-14 1.99515855e-11 -5.67688804e-21 4.09070012e-18 3.02665042e-16 4.82707454e-16 1.10174356e-16 -3.15707367e-13 4.56630556e-09 -1.12258657e-11 1.97270188e-16 3.28373057e-14 4.9608888e-16 6.82609899e-22 -5.36085323e-14 -1.82316761e-14 -7.88797712e-17 3.89209406e-18 4.98619109e-20 -7.69878938e-10 4.26777191e-08 2.54690678e-11 -1.00940744e-16 8.82378044e-16 -1.73055875e-10 -1.38048097e-08 -1.23582851e-07 -3.14568103e-21 -1.91134892e-19 -1.05721452e-14 -5.29040076e-11 -3.3360599e-11 -5.66751575e-14 4.25504619e-16 5.416666e-10 -1.06748269e-12 -2.84090607e-31 -1.92148264e-06 2.77518231e-09 -9.18536601e-12 5.74747125e-11 -2.3908755e-09 -1.37087475e-09
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
4
padH
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
0
2015-05-16 17:48:05 +12:00
1
12
4
97
2015-05-16 17:48:05 +12:00
3
V 1
12
nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
98
2016-03-27 21:15:14 +13:00
6
2016-01-23 13:36:54 +13:00
2
7
inplace
2016-01-23 13:36:54 +13:00
5
1
2
2015-05-16 17:48:05 +12:00
5
_type
2015-05-16 17:48:05 +12:00
2
16
torch.CudaTensor
2
6
negval
2015-05-16 17:48:05 +12:00
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
99
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
6
output
2015-05-16 17:48:05 +12:00
4
100
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
5
train
5
0
2015-05-16 17:48:05 +12:00
1
13
4
101
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
102
2016-03-27 21:15:14 +13:00
16
2
2016-03-27 21:15:14 +13:00
2
dH
1
1
2
2
dW
1
2015-05-16 17:48:05 +12:00
1
2
11
nInputPlane
1
128
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
103
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
2
kH
1
3
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
104
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
kW
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
3
2016-01-23 13:36:54 +13:00
2
12
nOutputPlane
2015-05-16 17:48:05 +12:00
1
1
2
2015-05-16 17:48:05 +12:00
4
2016-03-27 21:15:14 +13:00
padW
2015-05-16 17:48:05 +12:00
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
6
weight
2015-05-16 17:48:05 +12:00
4
105
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
2
1 1152
1152 1
2015-05-16 17:48:05 +12:00
1
4
106
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
1152
2016-03-27 21:15:14 +13:00
-0.705288231 -0.176139399 -0.230661288 0.0737797394 -0.0434918441 0.763007343 0.340610594 0.488269806 -0.510180473 -0.549092054 0.829579771 -0.00874371827 0.489260912 -0.143722609 -0.842903793 -0.223025933 -0.166321412 0.615390897 -0.00182754814 -0.0213809721 0.131075844 -0.0311710332 -0.418366104 0.407666355 -0.383920997 0.635586321 -0.317353129 -0.0300769694 0.190373808 0.215056852 -0.371580392 -0.102174036 -0.914889634 -0.108370744 0.658928812 0.46405226 -0.308554441 -0.491540074 0.370190382 0.708659649 -0.176172778 -0.499563217 -0.303258806 0.741280675 -0.0423225984 -0.0364141688 -0.000545244315 0.0243488643 0.027292503 0.182680488 -0.0361422002 0.0362400115 -0.148568347 0.00271723559 0.357525378 0.383969903 -0.568440676 -0.078203626 -0.628643513 -0.233228594 0.196525425 0.40105933 0.171155006 0.377210021 -0.586953938 -0.394255638 -0.0257141646 0.253035843 0.20637165 0.364240736 0.213957459 0.576112747 -0.00513384771 0.57103622 -0.188746735 0.226900309 -0.0870934576 -0.83857584 -0.140491456 -0.453070521 0.0102122566 -0.2668567 0.0794951096 0.00537947938 0.108108737 0.00278966385 -0.00427468074 0.0144217378 -0.0122328755 -0.00907345768 -0.381083608 -0.00311636506 -0.0639888868 0.694425821 0.071965538 -0.678311408 0.0226987265 0.136258915 0.20313786 -0.699146628 -0.0517932996 -0.116386734 -0.265014082 0.142272845 0.499881893 0.290608674 0.157713234 0.0409522243 -0.286060065 0.636932969 0.093390882 -0.0542578846 -0.270048648 0.694800556 -0.168586731 -0.341200143 -0.304083198 -0.0955201015 1.05171895 -0.534018934 -0.293765694 0.263226718 -0.35009104 -0.0899156183 0.175189868 0.345579237 -0.610615253 0.129298925 -0.811712325 0.16083087 0.135165915 0.185081393 0.164895251 0.302183568 0.343839765 0.414180845 -0.491862386 -0.0975044817 0.323645324 0.543570399 0.152435288 -0.75291878 0.488694668 -0.580587864 0.743074179 0.236715868 -0.543653905 -1.31449163 0.705594718 -0.147727847 -0.28391093 0.333164603 0.271454006 0.0296253636 0.950831115 -0.154211447 0.624448061 0.20655112 -0.560937881 -0.243060008 -0.331595182 -0.521006227 -0.140175939 0.0879472271 -0.0064292429 0.15029487 -0.172808021 0.164835542 -0.0508487523 0.130507216 -0.162636489 0.953130662 -0.589786828 0.0572487488 0.311873376 -0.468761027 0.435113311 -0.54240942 -0.401878715 0.245484635 -0.349451512 0.487702221 -0.088120915 -0.0935252383 -0.0351684727 -0.35053736 0.065566279 0.213171989 0.141251311 0.483506113 -0.430074632 0.307190627 0.287759662 0.0919076577 -0.422199458 -0.500360072 0.174193725 -0.645987153 -0.585743964 -0.113562278 0.155164942 0.280149639 -0.0987197086 0.114896879 -0.216134548 0.244717687 0.218515366 0.210138142 -0.101516202 -0.334398061 -0.503038347 -0.175137028 0.133638456 0.479521573 0.331821918 -0.0310739651 -0.0216664225 -0.0730358809 0.106801145 0.158594757 -0.232402623 0.0825801119 -0.0334543027 0.120064832 -0.0859518275 -0.202230871 -0.289442301 0.208790749 0.0258999579 0.142757803 0.672086895 -0.26308006 0.178256705 -0.472888499 -0.0878362358 0.164279789 -0.385304838 0.37204963 -0.193298027 0.180352241 0.0688182563 -0.363958627 0.247128367 0.573479354 0.169487014 -0.465178609 0.333552271 -0.443890542 -0.136766881 0.110169031 -0.0510392636 -0.0860522911 -0.0420373194 0.0245226696 0.103655137 0.163462698 -0.226235256 -0.0545466952 -0.0820046216 0.18697685 -0.0531870872 -0.0601455681 0.0833969116 -0.0836152881 -0.286875576 -0.184907898 0.269118577 -0.147541597 0.0317004807 0.387466192 -0.108878441 -0.473036468 -0.227580994 0.452530831 -0.150453135 0.303688943 -0.629297972 0.420024902 0.429743022 0.509008586 -0.00149220985 0.0620030649 -0.0747806206 -0.406498998 0.0503911413 -0.84285605 -0.0750690028 0.00669580419 0.0733298436 0.217050523 0.948121011 -0.612897635 -0.219671831 0.0889303312 0.044291988 0.207284287 -0.746042013 0.0141930236 -0.0965286791 0.318573892 -0.0226868428 -0.102985613 -0.0445756726 -0.00270250021 -0.0236582719 -0.00600237725 0.239842534 -1.40624762 -0.158020064 0.0326886475 0.545041263 -0.117012225 0.675329685 -0.0219220668 -0.472720742 -0.320463747 0.203882992 0.416217446 0.452992857 0.127933115 -0.448022395 -0.204
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
10
gradWeight
2015-05-16 17:48:05 +12:00
4
107
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
1 1152
1152 1
2016-01-23 13:36:54 +13:00
1
4
108
2016-01-23 13:36:54 +13:00
3
V 1
17
torch.CudaStorage
1152
1.81751597e-12 8.4346452e-10 1.07176924e-11 4.49963711e-09 1.25000298e-13 -1.516498e-07 9.62582236e-10 1.2756314e-07 -9.77631061e-08 9.18223386e-10 -2.79210717e-11 4.11515101e-14 -9.23025371e-12 -7.61970376e-10 -5.39167555e-09 -1.37230834e-11 5.1245328e-09 -3.76475855e-14 -6.23083918e-10 2.56599043e-14 -2.16190964e-21 -3.71045208e-11 1.00920675e-10 -1.42629161e-18 1.8256046e-11 6.79341913e-12 6.67707045e-09 1.8575335e-13 -6.94376556e-09 -5.28714103e-14 -2.10070812e-13 -9.2593086e-11 9.33741001e-11 1.99105637e-20 -1.49736388e-15 -2.66286604e-10 -7.02266578e-10 -5.61195423e-10 -1.33729909e-17 4.17034528e-07 -6.79524254e-13 -1.0203681e-10 1.38700307e-14 -2.12130742e-13 -6.19339513e-09 -2.34018664e-12 1.55274078e-15 -6.19450162e-08 -3.13096793e-15 3.28738304e-16 8.05508438e-10 -7.4417601e-09 -3.11942789e-12 4.24269856e-16 2.26969412e-08 3.03376746e-09 -4.488581e-06 -6.60944693e-07 -4.51485889e-08 -6.97071179e-10 -7.79253238e-17 1.3922909e-08 -6.94015623e-10 1.13162297e-08 3.36222995e-15 6.9230394e-12 -3.29379556e-18 -2.7599782e-07 -8.19622564e-11 8.21309031e-14 9.76693905e-14 -1.00370212e-19 -1.34894846e-15 -3.81747422e-13 -4.15413315e-10 -3.1193906e-12 6.92477145e-18 -7.88155131e-12 1.83832567e-11 -3.85894929e-13 -3.197892e-15 1.35668959e-11 -8.64211756e-07 -9.35059105e-13 -2.49872302e-12 -3.49074369e-08 2.06616129e-08 -1.42937728e-09 -2.87242062e-12 -5.28322897e-09 8.53889293e-10 -4.91746435e-11 -5.85263366e-11 -2.29293313e-08 -8.19295423e-08 3.23624989e-15 2.7258249e-08 2.47279919e-10 2.78638094e-12 2.05637807e-12 -2.14915064e-09 3.46287625e-12 5.44083871e-17 -7.78600184e-10 -3.65410742e-08 -5.94155747e-09 -2.82438428e-08 -1.44425553e-21 1.05633589e-07 9.10953674e-11 -3.52083163e-12 -8.11803846e-10 -7.78936784e-14 -6.37725589e-06 -1.35566444e-13 2.29732628e-13 7.13625655e-17 3.5260384e-12 -8.05605693e-10 1.77236668e-15 4.39867517e-11 -5.85941848e-07 -1.16709598e-09 -2.06037375e-11 -4.06894892e-15 -8.44699599e-08 6.28624597e-10 -2.91960513e-23 7.52671347e-09 7.63870411e-10 2.77690683e-11 -9.12737081e-16 -6.77005587e-12 9.71909863e-14 4.95768912e-11 -3.12798236e-13 -3.47106743e-16 7.78161389e-12 -3.96127304e-13 1.71409587e-18 -1.05322882e-13 5.63117442e-10 -2.07404405e-08 -3.43332518e-09 -1.80573605e-19 1.4440608e-12 -2.57406271e-07 -1.33061967e-16 7.81312376e-11 1.11027914e-20 -8.45077897e-10 -4.11931166e-12 -2.13245784e-07 5.88369943e-16 2.83523595e-12 1.61215685e-09 5.38983932e-06 7.47719413e-21 -1.27963435e-20 -2.2452238e-11 -1.34291067e-08 -3.87494254e-07 -1.02534992e-13 2.00289074e-09 7.44724248e-12 -8.47297343e-08 2.00491831e-13 2.72945966e-14 6.07315738e-13 6.70018085e-09 2.21581142e-14 -5.27207339e-24 -1.80333193e-09 -1.42167233e-11 -3.87454391e-12 5.0615046e-20 1.76705737e-13 2.09793154e-14 -2.23651609e-12 5.55880872e-15 -6.61415627e-12 4.06120016e-12 -1.97747994e-07 -1.29020994e-09 1.43967366e-10 -3.40326855e-12 -2.73857932e-26 -1.15992742e-20 -7.73129893e-10 -3.20139755e-16 2.55561838e-09 2.30257174e-10 1.10207312e-11 3.38334775e-13 2.23981227e-14 -2.87327874e-14 -7.7430895e-10 -5.98075478e-10 -2.1074032e-20 1.76456766e-10 2.95789004e-10 3.40409826e-23 1.39194213e-07 -2.48300636e-18 9.76348558e-10 -1.46183623e-07 7.52345582e-12 -4.6885315e-15 -3.45707735e-29 6.04516177e-12 1.62454961e-09 4.01123135e-09 9.60088432e-12 -5.01911527e-22 -3.50539693e-12 -1.99727985e-08 -1.546072e-31 -4.44654745e-07 1.33139558e-11 -7.46625956e-11 8.44130674e-14 1.62493557e-16 -2.22291652e-09 1.34480871e-07 3.96431599e-09 9.91506444e-15 -1.32198788e-12 4.97983946e-11 -3.19003879e-10 -9.41376663e-15 -9.70281321e-12 9.34240596e-09 6.87677276e-11 6.05577033e-10 1.71915776e-15 5.62291438e-14 -5.46629297e-10 1.79489908e-12 -1.3318055e-10 -5.1255254e-17 -1.83708728e-08 -8.58693006e-10 -5.70384517e-10 6.01963684e-25 8.24318818e-07 4.06509992e-10 9.69874354e-12 -2.41466179e-12 -5.62113193e-16 -3.54483864e-09 -7.32104308e-18 -4.62446401e-11 -2.88612224e-11 2.55954946e-09 -1.32166861e-10 -2.40724856e-22 -3.52018558e-14 -1.28905095e-15 -3.10993028e-12 1.50068602e-09 -1.38419029e-12 -5.02904651e-09 -1.97651367e-10 1.24773369e-09 -4.44405346e-26 5
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
5
_type
2
16
torch.CudaTensor
2
4
bias
2015-05-16 17:48:05 +12:00
4
109
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-03-27 21:15:14 +13:00
1
1
1
2016-01-23 13:36:54 +13:00
4
110
2016-03-27 21:15:14 +13:00
3
V 1
17
torch.CudaStorage
2015-05-16 17:48:05 +12:00
1
2016-03-27 21:15:14 +13:00
0.0685224086
2015-05-16 17:48:05 +12:00
2
2016-03-27 21:15:14 +13:00
8
gradBias
4
111
2016-03-27 21:15:14 +13:00
3
V 1
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
1
1
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
1
4
112
2016-01-23 13:36:54 +13:00
3
2016-03-27 21:15:14 +13:00
V 1
17
torch.CudaStorage
1
0
2
4
padH
1
0
2015-05-16 17:48:05 +12:00
1
14
4
113
2015-05-16 17:48:05 +12:00
3
V 1
7
nn.View
3
114
2016-03-27 21:15:14 +13:00
7
2016-01-23 13:36:54 +13:00
2
2016-03-27 21:15:14 +13:00
11
numElements
1
2016-01-23 13:36:54 +13:00
1
2016-03-27 21:15:14 +13:00
2
5
_type
2
16
torch.CudaTensor
2
6
output
4
3
2
9
gradInput
4
115
2016-01-23 13:36:54 +13:00
3
2016-03-27 21:15:14 +13:00
V 1
16
torch.CudaTensor
0
1
0
2015-05-16 17:48:05 +12:00
2
4
size
4
116
3
V 1
17
torch.LongStorage
1
-1
2
2016-03-27 21:15:14 +13:00
12
numInputDims
1
3
2
2016-01-23 13:36:54 +13:00
5
train
5
2015-05-16 17:48:05 +12:00
0
2
14
w2nn_arch_name
2
5
vgg_7
2
2015-05-16 17:48:05 +12:00
5
train
5
0
2
11
w2nn_offset
1
7
2
5
_type
2
16
torch.CudaTensor