1
0
Fork 0
mirror of synced 2024-10-05 04:33:29 +13:00
waifu2x/models/anime_style_art/noise1_model.t7

1728 lines
5 MiB
Text
Raw Normal View History

2015-05-16 17:48:05 +12:00
4
1
3
V 1
13
nn.Sequential
3
2
3
2
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
3
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
7
modules
3
2016-01-23 13:36:54 +13:00
4
2015-05-16 17:48:05 +12:00
14
1
1
4
2016-01-23 13:36:54 +13:00
5
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
6
17
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
4
padW
1
0
2
11
nInputPlane
1
1
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
7
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2
2015-05-16 17:48:05 +12:00
9
2016-01-23 13:36:54 +13:00
gradInput
4
8
2015-05-16 17:48:05 +12:00
3
V 1
2016-01-23 13:36:54 +13:00
16
torch.CudaTensor
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2
2016-01-23 13:36:54 +13:00
8
gradBias
4
2016-01-23 13:36:54 +13:00
9
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
1
32
1
1
4
2016-01-23 13:36:54 +13:00
10
2015-05-16 17:48:05 +12:00
3
V 1
2016-01-23 13:36:54 +13:00
17
torch.CudaStorage
32
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2
2
dW
1
2015-05-16 17:48:05 +12:00
1
2
12
nOutputPlane
2015-05-16 17:48:05 +12:00
1
32
2015-05-16 17:48:05 +12:00
2
4
2016-01-23 13:36:54 +13:00
bias
4
11
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
1
32
1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
12
2015-05-16 17:48:05 +12:00
3
V 1
17
torch.CudaStorage
2016-01-23 13:36:54 +13:00
32
-0.00684573269 -0.00637226226 -0.00277734967 -0.0294353068 0.0144880181 0.00427451916 -0.0049711368 0.00475662993 0.00178989174 -0.0423850566 -0.00898060296 2.49422628e-05 0.0068934015 0.0295247398 -0.0112056024 -0.0089507103 -0.0114619434 -0.0067871646 -0.0894550309 -0.00312883221 -0.00602424983 -0.00384070165 -0.0059074834 -0.0275169816 -0.00575044425 0.00181247387 -0.00546376593 0.0298137274 -0.00599276973 0.00873152819 -0.0113751665 0.00866923016
2
2
kH
1
3
2015-05-16 17:48:05 +12:00
2
6
finput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
13
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
6
weight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
14
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
2
32 9
9 1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
15
3
V 1
17
torch.CudaStorage
2016-01-23 13:36:54 +13:00
288
0.00341350655 0.0458156839 0.113944672 -0.0341889188 -0.045758646 0.0656073242 -0.00961190742 -0.1336395 -0.00490449648 0.0123231197 -0.0885350928 -0.03395449 0.169168547 0.0242489744 -0.1424319 0.00636893604 0.0521279462 0.00187571568 0.0404166877 -0.162858054 -0.0183527339 -0.145589232 0.0661874413 0.0487699285 0.11572019 -0.0153444409 0.0684367195 -0.0849501342 0.0614063591 0.105136879 -0.0814591721 -0.0948814452 0.208156824 0.027181508 0.179122508 0.1546323 -0.0523519553 0.0949050039 -0.00741603924 0.00588348974 -0.00954097323 -0.186365828 0.0302568786 -0.0716165528 -0.0317036957 0.0233615749 0.0186686572 0.039104607 0.0676270127 0.0898721665 0.00700242259 -0.0158134867 -0.0861557052 -0.140708849 -0.0688599497 -0.0374103338 0.109915555 -0.0237225723 -0.0174742043 0.0345251523 -0.0546832606 0.0363595262 0.0211156625 -0.0257789697 -0.00750720128 -0.132565573 0.00293769874 0.0101883337 -0.0278122928 -0.0207320657 0.0541261062 -0.0117061753 0.0780502483 -0.0251062047 -0.0608989708 0.0592535883 -0.0182384122 0.0365770012 0.0096066026 -0.0703002885 -0.00656775339 0.00223788689 -0.0299549494 0.00569941243 0.0729200616 0.0167557076 -0.00754519692 -0.00415435852 -0.00712406449 -0.00178886484 -0.0339846648 0.0843394473 -0.0209923889 -0.0366859213 0.176744536 -0.121812701 0.0386844948 -0.0244777557 -0.0606756508 -0.0166573767 -0.0224696696 -0.0216382109 -0.030700231 0.103083126 0.00452999352 0.0427792817 -0.167893589 0.0460000299 -0.00945626851 -0.0413753577 0.224247977 0.0112845311 0.0206025485 0.00157605542 -0.018590888 -0.119760722 -0.0684947297 0.0367834829 -0.0787921026 -0.198773578 -0.0125441449 -0.0914982632 0.0612413622 -0.0242638011 -0.107072346 0.0464910492 -0.0178682897 0.0506577566 -0.0182655901 -0.00425863219 -0.00848899316 0.00462822057 0.0219419543 -0.0439640284 0.0207389519 0.0235506613 -0.0054835584 -0.0203756634 0.051769603 -0.0410722643 -0.0147879105 0.0483195931 0.00781725999 -0.0565207638 0.0132371094 0.021063976 0.0184462406 0.0089556193 -0.0798342898 -0.0335134715 -0.0121612325 0.0351015925 0.0296302661 0.00685549527 0.0181911048 0.0280332733 0.0777351633 0.0505635962 0.00955720618 -0.0988985598 -0.0807330906 -0.0114262905 -0.000165991602 -0.00211372622 -0.00855380297 -0.000811162696 -0.00284903683 0.0131837875 -0.0100800497 0.0371007435 0.0714738742 0.1191824 0.042583216 -0.134629145 0.0511905141 -0.0189949051 -0.0624319836 0.00379655696 0.0721147135 -0.0756103098 -0.00563253835 -0.000587762974 -0.0176298972 -0.0387935899 0.0379211493 0.0410523228 0.0439282209 -0.0380360708 -0.018449707 0.0117725004 -0.0308121294 0.0615751334 0.00792680029 0.0777379721 0.0200924519 -0.00566490088 -0.0539949313 -0.0899161249 -0.100355178 -0.0787737966 -0.0420944989 0.0986464769 0.066283755 0.107892863 0.0768600032 -0.109832875 -0.0169925038 0.167733118 -0.130198866 0.062951237 0.0494440831 0.0363337994 -0.0164697208 0.0776759535 0.0518691503 0.0327665247 -0.0115807978 0.0356472805 -0.0140839545 -0.0549529903 0.124713309 0.0196047481 -0.0934365764 -0.0172733497 0.0125425709 0.0611372255 -0.0201489609 -0.0474918559 -0.0441291146 -0.0123957964 0.0487338006 -0.0185255259 -0.0560659654 -0.017791383 0.0948804319 -0.109817423 -0.0754501596 0.079133004 -0.0130955866 0.105537809 -0.134932071 -0.00723044714 0.0621199496 -0.0194802769 -0.0321444646 -0.186288312 -0.149300262 0.0152965616 -0.0426707529 0.0254763328 -0.0457975119 -0.0645391345 -0.0704244301 0.0325125828 -0.0602877289 -0.0397022218 0.0476760566 -0.0132862534 0.096661374 -0.0568884835 0.0360813253 -0.057780955 0.110247307 0.121268883 -0.0667895526 -0.0531002171 -0.0844902322 0.0561725684 0.00994271319 -0.0355214737 -0.0260043293 0.0394788757 0.0476637855 -0.0327913985 -0.00412955787 0.0340651982 -0.101748966 0.0368920714 0.0604003966 0.0236345083 -0.0357628576 -0.00350684626 0.0222181641 0.00423535379 -0.131580785 -0.0105107715 -0.000120736819 0.00721027423
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
10
gradWeight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
16
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
32 9
9 1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
17
3
V 1
17
torch.CudaStorage
288
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
10
fgradInput
2015-05-16 17:48:05 +12:00
4
18
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
2015-05-16 17:48:05 +12:00
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
4
padH
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2
dH
2015-05-16 17:48:05 +12:00
1
1
2016-01-23 13:36:54 +13:00
2
2
kW
1
3
2015-05-16 17:48:05 +12:00
1
2
4
19
2015-05-16 17:48:05 +12:00
3
V 1
2016-01-23 13:36:54 +13:00
14
w2nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
20
2015-05-16 17:48:05 +12:00
5
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
2015-05-16 17:48:05 +12:00
8
negative
4
21
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
22
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
23
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
14
negative_scale
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
1
3
4
24
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
25
2016-01-23 13:36:54 +13:00
17
2
2016-01-23 13:36:54 +13:00
4
padW
2015-05-16 17:48:05 +12:00
1
0
2015-05-16 17:48:05 +12:00
2
11
nInputPlane
1
32
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
26
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
27
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2
2016-01-23 13:36:54 +13:00
8
gradBias
2015-05-16 17:48:05 +12:00
4
28
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-01-23 13:36:54 +13:00
32
1
1
4
29
3
V 1
17
torch.CudaStorage
32
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2
2
dW
1
1
2015-05-16 17:48:05 +12:00
2
12
nOutputPlane
1
2015-05-16 17:48:05 +12:00
32
2
4
2016-01-23 13:36:54 +13:00
bias
4
30
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
1
32
1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
31
2015-05-16 17:48:05 +12:00
3
V 1
17
torch.CudaStorage
2016-01-23 13:36:54 +13:00
32
-0.00519692199 0.00607845699 -0.0284984559 -0.0018347142 -0.00864386 -0.00445417454 0.0242372602 0.0101757934 0.0527517125 -0.0155116254 0.0204261169 -0.0089069847 -0.00148167962 0.00154970924 0.0383142121 -0.00111893995 -0.0242180973 -0.0169460382 -0.0138717163 0.00906137284 0.00811559148 -0.0149479927 0.0133768776 -0.00112005603 -0.0262185149 0.0113706933 0.00780665549 -0.0136464909 0.0104716709 -0.00434859004 0.00144677714 -0.00569095463
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
2
kH
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
3
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
6
finput
4
32
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
6
weight
2015-05-16 17:48:05 +12:00
4
33
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
32 288
288 1
2016-01-23 13:36:54 +13:00
1
4
2016-01-23 13:36:54 +13:00
34
3
V 1
17
torch.CudaStorage
9216
-0.0297243688 -0.167205527 0.00319871888 0.0918795317 0.0184864737 -0.0269257817 0.11173965 0.105263136 0.0400212556 0.103009671 -0.00695425132 -0.0128535675 0.0222104155 -0.0734103993 0.13890104 -0.0454708897 -0.120365769 -0.0252791811 0.106929876 0.0397200398 0.0262829922 0.0994812995 -0.00657472294 -0.0564616248 0.0268094763 0.0476735793 0.0956888422 0.0360837802 0.0157480594 0.127995118 -0.0369663462 -0.080371432 0.00122216926 -0.010907311 0.0111564584 -0.0203015916 0.00391260395 -0.00753835309 -0.00915352069 -0.0277704336 0.0832540691 0.0385344252 0.160805374 0.0679753944 -0.0199624412 0.0579191484 -0.106996447 -0.00694738235 -0.103883043 -0.0138578797 -0.0289484914 -0.0167461354 0.0352699831 0.163915962 -0.0742107853 -0.0288184155 -0.107793882 -0.00196221611 -0.0882532299 -0.0502811521 0.0983926579 -0.096287027 -0.152442366 -0.0699883103 0.0315520056 0.0377745517 -0.0496957935 0.108410932 0.102085873 -0.0298132319 0.00822764449 -0.129032761 0.134599119 0.108416602 0.0895896405 -0.00281068007 -0.0105478531 -0.0637042299 0.109189659 -0.0432775766 0.0235373266 -0.0372076072 -0.0786526799 0.0558785051 0.0350631066 -0.00573299639 -0.00913363229 -0.0304793436 -0.0628005713 0.108696096 0.0341638774 8.41962974e-05 0.0396542475 0.0772089362 0.0782170743 -0.0116186244 -0.0325709097 -0.04329798 -0.0850394592 0.0763644129 0.00514547713 -0.0340823941 0.109999493 0.105990373 0.0219444335 0.0761221573 0.01157982 0.132143229 0.147554055 0.092903927 -0.0369535424 -0.0279010963 0.0485079139 -0.0103837335 -0.13947241 0.00601474103 -0.0296726748 0.0147988424 -0.0614679903 -0.0346321538 -0.0193558726 -0.0799846128 -0.0498493873 -0.108313181 3.39053477e-05 0.0149662979 -0.0679272339 -0.0312783308 0.0403006114 -0.0645123869 -0.0491221696 0.0817224607 0.0390471071 0.133014828 0.0145682488 -0.063889876 0.0365782268 -0.0255028177 0.0135227069 0.0581658892 0.00546887144 -0.0782824382 -0.0215588212 -0.0790985152 -0.179151565 -0.0101478538 0.147943094 0.00137343048 -0.0589264892 -0.0301892906 -0.00216747122 -0.00130496407 0.145915255 -0.0630459264 -0.0501909479 -0.0138029326 0.140363485 0.237820745 0.0989925191 -0.00162920519 -0.0796386376 0.0278356094 -0.428479105 -0.0100097796 0.0459175929 -0.0294131041 -0.172931463 -0.0161244869 -0.0252063181 0.126444131 0.101688556 0.0284556318 0.0181334559 0.116076447 -0.0377914459 0.0518539473 0.0247683916 0.0256443135 -0.0177989788 0.0301504005 0.0188892651 0.0878992528 0.173306718 0.041996289 0.105305426 -0.0721640438 0.068457827 -0.0211384706 0.061805293 -0.0473216996 -0.0960456654 0.0736141801 0.214122936 0.0843477398 -0.0614502989 0.133096308 -0.11260733 -0.117749333 -0.00459959684 8.4866133e-06 -0.0525478534 0.0454142466 0.151630312 0.126849726 0.0537472852 0.14650929 0.0543469675 -0.144710094 -0.00632335711 0.0602139086 -0.0231627319 0.0106023634 -0.0926488638 0.118971355 -0.126964763 0.180722222 0.0819944963 -0.0486255921 -0.228827447 0.0935036689 0.208936796 0.0542235784 0.0733317882 -0.131115317 0.0181155857 0.0223736688 0.0492203459 -0.198351637 -0.0400263295 -0.031944491 -0.047954876 0.147656962 0.0750021264 -0.0328446105 0.0664606988 0.0631166846 -0.0951436982 0.172713414 -0.0474855602 0.0331263207 0.133333147 -0.0583443604 0.051117152 0.0672874674 -0.149242193 0.190112695 -0.13005738 -0.0200548563 -0.0837037116 -0.105943471 0.187296242 0.119229294 -0.0377589203 -0.0627509952 -0.0610843785 -0.071780175 -0.0934936702 -0.0037201487 -0.000832385034 -0.0230694525 -0.0852470025 0.0157072451 0.0453531742 -0.0548163801 0.0753384456 0.00689572236 0.042114988 -0.0487973802 0.076537393 0.0314364061 -0.0659659058 -0.0737456307 0.00123758905 -0.061309021 0.0623601153 -0.102109276 -0.00743963476 0.135723829 0.0520469584 0.145839438 0.0736116171 0.0585359521 -0.168183193 0.12483687 -0.0404708609 -0.0209810659 0.064345032 -0.0800463408 0.112784341 -0.0416960828 -0.00701820711 0.0730911717 -0.106534496 -0.084773168 -0.0938478038 -0.141610205 0.0686162189 0.0496935584 0.104709491 -0.00114628579 -0.00983795524 -0.0180585831 -0.0284101926 -0.00828895904 -0.0975634083 -0.079073064 -0.017616801 -0.0482534021 -0
2015-05-16 17:48:05 +12:00
2
5
train
5
2015-05-16 17:48:05 +12:00
0
2
2016-01-23 13:36:54 +13:00
10
gradWeight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
35
3
V 1
16
torch.CudaTensor
2
32 288
288 1
1
4
36
3
V 1
17
torch.CudaStorage
9216
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2
10
fgradInput
4
37
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
4
padH
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2
dH
2015-05-16 17:48:05 +12:00
1
1
2016-01-23 13:36:54 +13:00
2
2
kW
1
3
2015-05-16 17:48:05 +12:00
1
4
4
2016-01-23 13:36:54 +13:00
38
2015-05-16 17:48:05 +12:00
3
V 1
2016-01-23 13:36:54 +13:00
14
w2nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
39
5
2
5
train
2015-05-16 17:48:05 +12:00
5
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
8
negative
4
2016-01-23 13:36:54 +13:00
40
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
41
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
42
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
14
negative_scale
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
1
5
4
2016-01-23 13:36:54 +13:00
43
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
44
17
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
4
padW
1
0
2
11
nInputPlane
1
32
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
45
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2
2016-01-23 13:36:54 +13:00
9
gradInput
4
2016-01-23 13:36:54 +13:00
46
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
8
gradBias
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
47
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-01-23 13:36:54 +13:00
64
1
1
4
48
3
V 1
17
torch.CudaStorage
64
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2
2
dW
1
1
2015-05-16 17:48:05 +12:00
2
12
nOutputPlane
2015-05-16 17:48:05 +12:00
1
64
2015-05-16 17:48:05 +12:00
2
4
2016-01-23 13:36:54 +13:00
bias
4
49
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
1
64
1
1
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
50
3
V 1
17
torch.CudaStorage
2016-01-23 13:36:54 +13:00
64
-0.00688071316 0.000252442987 0.00607155124 -0.0013304454 0.00780095253 -0.00939233974 -0.00322324783 0.00617225468 -0.0014559218 -0.0102953725 -0.00534518203 -0.00517421123 -0.00898248237 0.0111770695 -0.00419491855 -0.000944293919 0.000858683779 -0.000662480772 -0.00162767502 0.00621710438 -0.00654170522 -0.00453148177 -0.00188665639 0.00713468483 0.00178467948 -0.0020378409 0.0103082154 -0.00382980495 -0.00152657018 -0.0047144196 -0.0030142488 0.00741126202 -0.00674895663 0.00964967627 -0.000648709713 -0.00119855034 0.00511057442 -0.0114674009 0.00046345382 -0.015530494 -0.000273523096 0.00318000303 0.0025669001 0.00703306589 0.00402461132 0.00516261579 0.00652070343 -0.0101717673 0.00566466199 0.000513482664 -0.0140382163 -0.000410698762 -0.00197161594 0.00952325854 0.000344870146 0.0117923263 0.0120320749 0.00697554648 4.90896018e-05 -0.00312783453 0.0089474069 0.00364050642 0.000261986919 0.0121442806
2
2
kH
1
3
2015-05-16 17:48:05 +12:00
2
6
finput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
51
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
6
weight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
52
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
2
64 288
288 1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
53
3
V 1
17
torch.CudaStorage
18432
-0.0374797657 0.0285335649 -0.0121265491 -0.0195699967 0.0898627043 0.0702863559 -0.0171222556 -0.078619495 -0.0673705488 -0.0727457106 -0.0298613999 -0.103202671 -0.0242039748 -0.0284241252 0.0187585223 0.0350213721 0.0193189085 -0.0292647798 -0.0572850406 0.0466605984 -0.00571662514 -0.0171444118 -0.014603667 0.0346747562 0.0549096055 0.0562472232 -0.0408888869 0.0780212507 0.0679700449 -0.0844756067 -0.00692465389 0.0459475033 0.00453677634 0.024833357 -0.0764120594 0.0656346455 0.00900755636 0.0608043708 -0.0154360803 -0.0680931583 -0.0546973236 -0.13383469 0.0353975631 -0.0676268414 0.0173762403 0.00350708165 -0.0689714998 -0.0275609978 0.0520608425 0.000284443027 0.0757684633 0.00184949231 0.061014466 -0.0265759323 0.0803649649 -0.00650521368 0.111080252 -0.0597699024 -0.00612786924 -0.0346929468 -0.0457132198 0.0374663211 -0.00786183216 -0.00101911474 0.0678296611 -0.0617099404 0.0491468385 -0.0575685315 0.0999131799 0.0439813323 0.00256105093 -0.0393202342 0.0204390492 -0.00482547889 0.130455911 -0.0161748547 -0.0728622898 0.00278920704 -0.0150180366 -0.0239271577 -0.000106585059 -0.0491613559 -0.150917605 0.00908350851 -0.103924632 0.0268465895 0.0514787063 -0.000278742431 -0.0112379454 -0.0478742681 0.0259143338 0.0783927068 -0.0020876599 -0.0423641689 0.0784597918 0.117581367 -0.136944219 -0.0945973992 0.0410055257 0.0388003699 0.0438415743 0.0270342752 -0.00713793933 -0.0813974068 -0.0146842645 0.0784108341 -0.0137344738 0.031987343 0.00804737117 -0.0573907532 -0.00479490403 -0.0415784195 -0.0585252382 -0.185032681 0.0873155519 0.026740279 -0.0918753594 -0.0134761613 -0.1854406 0.0447211042 0.165088758 -0.0512924269 0.0184479691 0.0606492348 -0.022572685 0.00449140463 0.0303222183 0.113127492 -0.0270388871 0.0356121957 -0.0043805954 -0.0586611703 -0.0150533412 0.00807497278 -0.00684072077 0.0326538421 -0.0163332243 0.0391568094 -0.0461683683 0.0445941947 -0.175318584 -0.067648977 -0.00364142423 -0.128974766 -0.00046283516 -0.0633632094 -0.0583035424 -0.00423823949 -0.0770407543 -0.00260086241 -0.0201324616 -0.0248577595 0.00638154289 -0.021735847 -0.0715559274 -0.0462736823 0.0102233039 0.0708098486 0.0150872162 -0.0575095825 0.0110233687 -0.068792738 -0.0171608906 0.0268916544 -0.00781810749 0.13685438 0.0823470876 -0.0337371677 -0.101857103 0.0167654529 -0.0897708014 -0.0138491215 -0.0906844735 -0.133468136 0.00828199182 -0.0181507431 -0.0960036591 0.0658954829 -0.00706293341 -0.0999302194 -0.11223112 -0.12076389 0.0725299045 0.104528412 -0.0662507415 0.00454826374 0.0301449504 -0.00191159279 0.0516722277 -0.00339191337 0.031667158 -0.00735133654 -0.0378236696 -0.0139026185 -0.0965641513 -0.0557603091 -0.012086466 -0.101692863 0.0180710088 -0.0274740495 -0.0603300817 -0.0512239411 0.0470826887 -0.00293959561 -0.0264763758 -0.0491028652 0.000777970068 -0.0495859459 0.0428512916 0.0959715024 0.0214379877 -0.109444864 -0.110096104 -0.0161000416 -0.0483781584 0.0354741961 0.0162971783 0.015472875 -0.0478830375 0.0671018511 0.119554736 -0.0509749688 -0.0437650643 -0.00421422534 0.0746310651 0.0351864584 -0.0922495574 -0.104308002 0.0643043295 -0.021218637 -0.138872072 0.0053083417 0.0452574976 0.0301979911 -0.0451214127 -0.0463286936 0.0632651076 0.028578911 -0.0176705979 -0.147355393 0.0523338579 0.0794474632 -0.0153927384 -0.0195979476 -0.0151370522 -0.0192003343 -0.0461026467 -0.0736794099 0.029561121 0.0593396761 0.0268149357 0.0261905808 0.00174991088 0.10133031 0.00573500246 -0.0590256862 -0.0362817906 -0.00729169929 -0.0132683385 0.000543025206 0.0168904774 0.0843430385 0.0134473182 -0.0953573734 0.060153015 0.0499967597 -0.0281706825 0.0412881114 0.116945222 -0.00843531732 0.0355140194 0.0384279527 -0.00881349575 -0.0669684559 0.00173490483 0.0201525968 0.0656545162 -0.0551903993 0.0394647196 -0.0750482678 -0.0465503298 0.0350863747 0.0664806142 0.00838837586 -0.0369247831 -0.0393625833 -0.0042210212 0.0899487436 -0.087480396 0.0220218003 0.0537381507 -0.117035531 -0.0835344195 0.0311013777 0.0268419143 -0.070538044 -0.032280121 0.0285371728 -0.132691801 0.0380527042 -0.033722043 -0.0596729852 0.0410189
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
10
gradWeight
4
2016-01-23 13:36:54 +13:00
54
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
2
64 288
288 1
2016-01-23 13:36:54 +13:00
1
4
2016-01-23 13:36:54 +13:00
55
3
V 1
17
torch.CudaStorage
18432
7.27991556e-10 -1.41658023e-15 -7.47760516e-15 9.92933446e-17 -3.05844698e-16 1.11228182e-14 -3.1350973e-14 4.98476382e-14 1.949053e-18 7.51528423e-11 1.16731699e-22 -2.01904595e-12 8.6683753e-16 1.48093424e-13 2.00894633e-14 5.62829422e-15 -3.24837423e-12 -4.18980475e-11 -2.18484764e-16 7.20800225e-11 2.6256639e-13 6.51467019e-18 -3.6062833e-18 -2.85976799e-16 5.13116771e-14 -1.29089363e-14 3.2518647e-17 7.82840158e-17 4.87746308e-24 3.96238359e-29 5.95341051e-12 -2.37465846e-16 7.16403742e-11 2.30914147e-33 -6.89239443e-10 8.10007417e-17 1.8899038e-18 -1.16422128e-12 -1.39786047e-11 2.73443941e-11 -1.00725672e-09 6.10969001e-12 -2.43413091e-13 -4.37055184e-11 -3.68925945e-10 -5.61372456e-20 -4.05125491e-14 7.27750615e-22 5.21388014e-14 3.49047681e-21 3.41045661e-15 -1.91397543e-13 -3.39224785e-12 -3.82546476e-19 1.43664369e-11 -7.84816517e-11 2.71774485e-14 -4.51447282e-11 -1.08776422e-18 3.23341572e-15 9.55682075e-16 1.29317542e-12 7.14073693e-19 2.59352999e-12 -8.95138672e-13 3.92245351e-12 -5.246132e-11 1.53542656e-23 8.25950974e-10 -7.34806421e-13 -6.57329744e-18 6.07599285e-16 2.41646069e-12 -7.19831773e-14 -1.03440626e-13 -5.64694385e-12 -1.24052109e-13 -8.93593034e-13 -9.30883669e-12 -4.04156328e-21 -5.44077532e-16 4.11231379e-12 -2.07791666e-11 -5.1605531e-10 2.04900089e-13 1.03841815e-20 3.88045232e-18 -6.34657737e-19 1.58944228e-13 1.09622458e-24 -4.58159243e-23 -5.9341787e-10 9.37933645e-18 -5.65951105e-11 1.61830429e-13 -2.02918151e-16 -9.44943688e-15 -4.57274288e-11 -9.52498804e-23 1.81871141e-19 2.01291636e-16 -3.22402532e-13 5.51823737e-14 5.60746642e-20 -1.17905712e-13 -1.71335535e-25 9.85989888e-15 9.19536843e-12 6.7134243e-13 5.98039671e-16 6.9124273e-16 1.37329401e-11 -1.76361756e-17 2.35195257e-15 2.78430301e-09 -5.25585974e-25 2.19483421e-13 1.03589602e-16 1.54340513e-10 8.05820559e-15 7.23465002e-17 7.85728843e-11 8.53125188e-13 2.84160614e-14 8.77618778e-16 4.20511426e-17 -2.95391763e-14 3.67704255e-16 6.13534848e-15 1.37653209e-15 -3.02108436e-13 -1.20082557e-14 -4.11947765e-13 1.80447373e-13 4.66764041e-12 -1.95114734e-17 -2.84357702e-14 -1.78621434e-16 3.25028313e-17 -4.23900021e-16 3.70496605e-16 1.60121986e-15 -4.41021204e-16 1.29212851e-12 3.75867252e-13 -5.37492102e-17 3.37629821e-16 2.81062854e-15 2.78957933e-11 1.90477706e-13 9.92662023e-19 1.18890986e-12 1.15334429e-14 -5.93340402e-20 4.01100964e-16 1.06432146e-14 -6.0454172e-20 -1.92074445e-14 -2.07044641e-13 -2.14471513e-18 5.67424094e-11 1.10967815e-11 -1.00207385e-33 -8.52941134e-17 4.3214271e-16 1.07525234e-16 -2.37584509e-14 9.92396627e-20 3.81676637e-14 6.50810709e-27 7.70867837e-20 2.04545383e-20 -8.12136354e-16 -9.22574249e-24 -4.87781707e-18 -1.06585187e-16 2.38104512e-13 -1.090735e-14 9.66362552e-17 -1.45514437e-15 -8.23064795e-15 5.11072797e-18 5.52992263e-19 5.96736935e-16 1.9717481e-19 -2.66434842e-17 -7.89711391e-38 -6.56809785e-13 -9.33591841e-20 -4.59745002e-12 -1.95326871e-12 9.96311154e-15 -5.05661728e-21 6.48707298e-14 -3.88553137e-17 -7.84005638e-10 1.27507569e-13 2.31820348e-14 6.6858506e-14 -1.05245385e-14 -3.00616863e-30 -4.49748356e-17 -2.56561753e-13 1.68030927e-13 7.14140148e-16 -5.48886633e-14 -7.44340895e-13 -2.79516557e-12 -2.19296051e-14 -1.17639765e-21 1.06905197e-11 2.51940864e-11 5.33653593e-14 -2.2297233e-13 -3.63973107e-14 -1.78512323e-19 1.72765873e-18 8.01771708e-14 -7.12381456e-12 2.21780772e-09 -5.16423626e-23 -4.56432575e-19 -5.8951632e-18 2.13798536e-14 4.31995528e-15 4.45911277e-12 -2.20566803e-14 -6.36006664e-11 -6.69563623e-12 -5.90139382e-10 -4.31618939e-17 -6.35181342e-19 1.18025988e-11 -3.49004341e-14 -1.8447475e-13 1.92687093e-13 -2.08554772e-13 4.1062544e-12 1.09880481e-15 3.00555538e-18 2.31015895e-17 4.79896194e-18 1.02318609e-12 6.36450985e-15 1.62146503e-19 -1.05178521e-13 1.14517354e-10 7.4142471e-15 -1.12479928e-14 1.80006597e-14 8.70259099e-15 3.47046405e-17 7.42391953e-17 5.55062742e-17 5.11827268e-17 -7.32440503e-14 5.68478883e-17 -5.32334853e-12 -4.32233083e-10 -3.48567052e-12 1.36162552e-18 1.53408688e-12 -5.20102277e-12 3.65346046e-13 8.88262039e-14 -1.082883e-14 4.82185581e
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
10
fgradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
56
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
4
padH
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2
dH
2015-05-16 17:48:05 +12:00
1
1
2016-01-23 13:36:54 +13:00
2
2
kW
1
3
2015-05-16 17:48:05 +12:00
1
6
4
2016-01-23 13:36:54 +13:00
57
2015-05-16 17:48:05 +12:00
3
V 1
2016-01-23 13:36:54 +13:00
14
w2nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
58
5
2
5
train
2015-05-16 17:48:05 +12:00
5
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
8
negative
4
2016-01-23 13:36:54 +13:00
59
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
60
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
61
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
14
negative_scale
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
1
7
4
2016-01-23 13:36:54 +13:00
62
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
63
17
2
2016-01-23 13:36:54 +13:00
4
padW
2015-05-16 17:48:05 +12:00
1
0
2015-05-16 17:48:05 +12:00
2
11
nInputPlane
1
64
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
64
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
65
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2
2016-01-23 13:36:54 +13:00
8
gradBias
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
66
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-01-23 13:36:54 +13:00
64
1
1
4
67
3
V 1
17
torch.CudaStorage
64
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2
2
dW
1
1
2015-05-16 17:48:05 +12:00
2
12
nOutputPlane
1
64
2
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
bias
4
68
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
1
64
1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
69
2015-05-16 17:48:05 +12:00
3
V 1
17
torch.CudaStorage
2016-01-23 13:36:54 +13:00
64
-0.00285809278 -0.00575977052 -0.00371189136 0.00378575688 0.000717258488 -0.0089932736 0.00370385824 -0.00202250783 -0.000424195867 -0.00614302186 -0.000796524691 -0.00396962417 -0.00402833708 -7.82608113e-05 -0.00430071913 -0.0136341676 -0.00278358767 -0.0496610738 -0.00292490586 0.00266140024 -0.00224003964 0.000215846725 -0.0035353438 -0.0101558585 -0.00255184644 0.00207824376 -0.00679548644 -0.00311430288 -0.0257270522 -0.00143698021 -0.00302135269 -0.00629209587 -0.00471361028 -0.00370171806 -0.00414326554 -0.00291928812 0.00176364987 -0.000140756747 -0.000794686377 0.000853210979 0.00221084384 -0.00212393794 -0.00105676195 0.00261621224 0.00330722495 -0.00349173462 0.00306986738 -0.00521846302 -0.00329850777 -0.00244905427 -0.00438312069 -0.00298126391 -0.0294039436 -0.00952285808 -0.00595995132 -0.00252243481 -0.00191353948 0.0174076818 -0.000768227212 -0.0081197191 0.00163688068 -0.00823359098 0.00201241625 -0.00456011342
2
2
kH
1
3
2015-05-16 17:48:05 +12:00
2
6
finput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
70
3
V 1
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
0
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
6
weight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
71
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
2
64 576
576 1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
72
3
V 1
17
torch.CudaStorage
36864
-0.0643625632 -0.0258680433 -0.0137389973 -0.0289045926 -0.115011401 -0.033801537 -0.00897837244 -0.135576352 -0.0721494183 0.0231738482 0.0398886502 0.205320999 -0.109656468 -0.0348567031 -0.01054083 0.033759743 0.00843236502 0.113370903 -0.112106383 0.00644466002 -0.0539257228 -0.104569957 -0.0352494754 0.141271278 0.0493872203 0.00579199987 0.0157340989 -0.0567425527 0.0647241101 -0.0150999492 -0.0532865711 -0.0714307502 0.121729448 0.0190370865 0.0628368109 -0.0371861905 -0.0319733694 0.0736239105 0.00914199278 -0.0207212809 -0.0677258819 -0.0486498177 0.0938648432 -0.0505171046 0.0220305491 -0.0400733016 -0.0843999833 0.0599294789 0.0559733734 -0.103617705 -0.118756115 -0.0253547374 0.0196685661 -0.0676484406 -0.036225576 0.101781227 0.00699654454 0.0279425103 -0.0637357235 -0.0631956831 0.0434255637 0.145731509 -0.0420645848 -0.0364409909 -0.0902278647 0.0124285473 -0.169708461 -0.16542992 0.00687023532 -0.0551007651 -0.0156507213 -0.0330281369 -0.0492034927 0.0675017461 -0.0107180728 -0.0810508952 0.00369134988 -0.0583401211 -0.0843170732 -0.0614334233 0.119226038 -0.0318365879 0.00697253039 -0.106675982 0.0917767063 -0.0975836739 0.0636631325 0.110047132 0.0386228561 -0.070850268 0.0335197821 0.028883636 -0.0600230061 0.0560913906 -0.0183364823 -0.0344541296 0.0522440672 0.00839310978 0.00580644747 -0.0987221599 -0.0635417178 -0.00299644121 -0.0167808421 0.170918599 0.00571484398 -0.0359560624 -0.023996314 -0.0240146257 0.0455100052 0.129836291 -0.00298087485 -0.0171711389 -0.0701447874 0.109377235 -0.172774464 -0.104443774 0.1053361 -0.0135813924 0.0255887341 -0.0481536612 -0.0147241438 0.0093918005 -0.0324127525 0.0227328893 0.0769326389 0.0385698117 0.0295021832 -0.0217051357 0.0386576168 0.0435161367 -0.0394965708 0.0392276049 0.108378708 0.0500491075 -0.0474629067 -0.0056258305 0.0848952979 -0.00397760188 0.164257228 0.078120321 0.0602046922 -0.0302206371 -0.0408231281 -0.0539282933 -0.0465777628 -0.0361705199 -0.0363225527 -0.0376169719 -0.0257946681 -0.0552997142 -0.0508841239 -0.0946082175 -0.0368939638 -0.0783524215 -0.00441262871 0.063734524 -0.0482355952 -0.064345412 0.0177491307 -0.0536506921 0.0351287201 0.0156683084 0.0108144982 0.000272969017 -0.0588269196 -0.016562067 0.0109872473 -0.0288676005 0.068790637 -0.113603301 -0.0470348895 0.0235733241 0.0743588433 4.26924526e-05 -0.101797812 0.079492487 0.0295000784 -0.0127572995 0.0187183805 0.0387189388 0.00876602717 0.161168978 -0.065583095 0.0273727402 -0.129896954 -0.0020305689 -0.0393483639 -0.176678374 -0.0561362952 -0.00272426894 0.0517498106 0.0877034813 -0.0413436592 0.0869017839 0.0185325369 -0.0214992967 0.115372896 0.0607961528 0.0775552988 0.0553595684 0.143544257 0.0770191401 0.0242990926 0.0504197069 0.0312644094 0.0270204302 0.103906929 0.00710674888 -0.00280293985 0.0363297909 0.0638327003 -0.120683379 0.0183498617 0.0257041641 0.0473183952 0.0485429056 -0.0070648035 0.0727644041 0.00269128429 0.0947292373 -0.0138093065 -0.0523425713 0.0160676874 0.0075743692 -0.0445784144 -0.0854800791 -0.0597172864 -0.00583503302 -0.0145857977 -0.109893963 -0.108785138 -0.0229081195 -0.161921993 -0.0612902865 0.0219255015 0.0177233275 -0.102614887 -0.107326359 0.0822573677 0.00881591532 -0.168504164 0.0259517971 -0.00221079215 -0.0175358392 -0.0473130122 0.0944109932 -0.110842407 -0.0785870701 0.091935575 -0.0540940799 -0.0340195633 0.0996309668 -0.0123480177 0.00684144767 0.0218659472 0.0105717303 -0.0274712965 -0.0285677053 -0.0052140695 -0.0429572649 -0.0779050589 0.0495864525 -0.0101934839 0.0524876863 -0.0387461521 0.00581003539 0.0113923643 -0.0995740891 -0.000866655784 0.0339153409 0.0138326287 -0.075639002 0.00464294665 0.0706868768 -0.123245969 0.0809039474 0.00459687784 -0.157994777 -0.0438588262 0.00388948037 -0.00830683019 -0.0491460152 0.0345131867 0.00968331471 0.000546517142 0.00709762191 -0.0962270573 -0.0645047128 0.0135188149 0.0540387444 -0.0375299752 -0.081379883 -0.088137351 -0.0187348705 0.0173027851 -0.0410810523 -0.019209858 -0.0640669018 0.0553471111 0.0454785638 0.0114772553 0.0364018753 0.0747614279 -0.0767677799 -0.030985
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
10
gradWeight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
73
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
64 576
576 1
2016-01-23 13:36:54 +13:00
1
4
2016-01-23 13:36:54 +13:00
74
3
V 1
17
torch.CudaStorage
36864
-7.8188931e-12 -1.28391369e-15 9.1196324e-14 4.01404916e-14 2.69055315e-15 -5.74624864e-15 2.29321939e-14 3.83239244e-14 -1.7269331e-15 2.15107827e-11 4.69735622e-12 4.56811552e-17 8.36563631e-12 -7.08856324e-12 -6.09407408e-14 1.05984167e-14 -1.99153844e-12 -1.71428198e-18 5.22500405e-14 -9.48800398e-15 -5.04806255e-22 2.94947171e-18 -3.76224573e-17 1.77432206e-13 1.38782957e-10 7.00303204e-19 5.12196694e-16 -4.39266882e-14 4.11129525e-21 2.70619022e-15 1.60019109e-16 1.27575592e-15 8.4877585e-17 -9.86585099e-18 -5.47546966e-14 3.3981329e-12 2.1291115e-13 -1.01120395e-16 -2.3919132e-19 2.21615938e-13 -9.85434607e-13 -1.98894048e-18 -6.02586882e-13 7.32175063e-13 5.45606942e-17 -2.46561861e-17 -1.79810415e-40 1.32292503e-14 -5.61492496e-24 -4.19597422e-14 3.28237654e-12 -7.4477657e-19 2.25996894e-17 -7.47027088e-14 4.07565214e-12 -2.82719711e-21 2.96045808e-14 -1.43675389e-14 3.0414501e-15 -1.03658046e-14 -4.56906694e-18 -1.07423778e-10 1.07409329e-12 5.41554237e-18 2.96092183e-15 8.18448444e-16 1.02157594e-19 1.45367609e-19 -7.85350552e-17 1.21736106e-12 -6.48208836e-14 -2.22174666e-12 1.73796308e-19 -1.81025578e-13 -2.34539659e-18 6.70381198e-12 -2.4062379e-19 5.3370876e-18 -3.19264743e-13 -4.47803619e-14 2.61182131e-10 -3.51060674e-19 -1.31936366e-16 1.89597443e-13 4.33812287e-19 1.03146047e-16 9.45827834e-13 1.5865888e-16 6.89775854e-17 -5.26921375e-14 1.35484902e-14 -3.99851379e-16 -1.88439216e-21 1.46690754e-12 -6.67501005e-17 -1.59022974e-22 1.42964288e-14 -3.68587996e-15 1.47869585e-12 1.26488409e-16 -2.29612709e-14 3.46878513e-20 2.40839163e-12 6.31603797e-11 1.15141573e-17 -2.37572116e-15 -2.7262346e-12 9.3681169e-17 -2.70487901e-24 5.90319464e-21 8.60481395e-16 1.54946844e-17 -5.16564534e-14 8.6130054e-15 -2.01270317e-15 -6.5044634e-19 -1.44073168e-15 -1.05694667e-20 2.1845925e-14 2.47615132e-15 4.26949397e-14 -3.3064242e-25 3.06340728e-13 4.3440164e-24 -1.04321814e-14 6.5231707e-25 3.57883398e-17 2.16929637e-13 -4.03161147e-17 3.52045693e-12 -9.05718699e-17 -3.05891854e-17 -7.90518584e-13 -1.16448496e-12 -1.09432103e-12 -1.19591074e-14 -4.19055236e-15 -2.05648297e-13 -3.74419607e-28 3.7709385e-14 -2.03047045e-14 -1.40704132e-11 -3.52233939e-16 3.90682224e-13 2.67697366e-12 2.5635636e-11 -1.87399031e-27 -7.93697506e-16 6.5390088e-15 8.16236693e-15 -7.87475458e-21 1.14366871e-12 9.29754693e-15 4.83172824e-23 3.35294455e-13 -1.58011977e-13 -9.61794223e-15 -2.72786249e-14 -1.64088501e-18 -1.1481779e-20 -2.73363112e-14 -1.83987725e-24 -5.01166383e-12 3.03550072e-17 -1.53484776e-14 -2.20925892e-23 -1.53870216e-13 3.08968508e-13 3.62020241e-16 -8.38608188e-15 -9.2751804e-23 -3.57981693e-16 -4.91272558e-17 -9.26858591e-20 5.51346782e-17 -9.63842043e-14 3.52763269e-14 3.71881808e-17 9.24825402e-14 1.8831803e-13 -8.21341085e-12 4.98746213e-14 -2.08908844e-16 -4.22666756e-21 -3.15543794e-23 -1.92386949e-18 1.03452742e-13 1.34262869e-25 1.09086409e-14 -3.41899469e-12 8.60316506e-12 2.24181611e-11 -7.65065656e-19 -5.10502373e-23 5.34062879e-14 -2.98721754e-14 3.17963841e-12 -2.07121461e-21 7.7126026e-15 2.07369488e-14 2.14817978e-17 -6.88001347e-16 8.71842339e-14 1.18039415e-11 -2.16904436e-14 7.13822838e-12 1.42734084e-17 4.11053519e-23 1.97354499e-18 2.52936692e-24 -1.59019309e-13 -1.12932812e-16 3.16089752e-19 -1.18904304e-15 2.2718247e-14 -7.91849704e-16 -6.20912089e-12 1.32968953e-16 7.05989194e-25 1.84693537e-12 3.42849143e-12 -1.01031285e-13 1.1761764e-13 5.60397974e-13 1.24117665e-12 8.65183699e-16 -1.17472818e-15 2.66927822e-15 -3.32292448e-14 -5.20771986e-18 -1.67167891e-21 -3.63316233e-15 -2.78539565e-20 -4.90868518e-19 -8.87762194e-13 -2.41659503e-13 -1.43484465e-18 -7.88069019e-14 -3.43605691e-14 1.12642151e-14 -1.58757612e-16 -5.08824956e-14 -1.08419818e-11 4.30066613e-14 -8.76121685e-14 -1.25848052e-12 -1.12151083e-17 2.66532978e-14 1.71236858e-13 2.79637458e-14 5.48489409e-14 -8.81771219e-12 1.58929779e-11 -1.95133329e-17 3.53405373e-17 -8.02698077e-13 6.74798221e-16 -5.70302361e-10 -2.60151988e-22 1.66046884e-14 -3.85573434e-17 -1.84423764e-18 2.26972528e-18 -1.13619229e-17 -1.74992047e-13 -7.14282827e-1
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
10
fgradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
75
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
4
padH
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2
dH
2015-05-16 17:48:05 +12:00
1
1
2016-01-23 13:36:54 +13:00
2
2
kW
1
3
2015-05-16 17:48:05 +12:00
1
8
4
2016-01-23 13:36:54 +13:00
76
2015-05-16 17:48:05 +12:00
3
V 1
2016-01-23 13:36:54 +13:00
14
w2nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
77
2015-05-16 17:48:05 +12:00
5
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
2015-05-16 17:48:05 +12:00
8
negative
4
2016-01-23 13:36:54 +13:00
78
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
79
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
80
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
14
negative_scale
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
1
9
4
2016-01-23 13:36:54 +13:00
81
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
82
17
2
2016-01-23 13:36:54 +13:00
4
padW
2015-05-16 17:48:05 +12:00
1
0
2015-05-16 17:48:05 +12:00
2
11
nInputPlane
1
64
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
83
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
84
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2
2016-01-23 13:36:54 +13:00
8
gradBias
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
85
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-01-23 13:36:54 +13:00
128
1
1
4
86
3
V 1
17
torch.CudaStorage
128
0 0 0 0 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186
2
2
dW
1
1
2
12
nOutputPlane
2015-05-16 17:48:05 +12:00
1
128
2015-05-16 17:48:05 +12:00
2
4
2016-01-23 13:36:54 +13:00
bias
4
87
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
1
128
1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
88
2015-05-16 17:48:05 +12:00
3
V 1
17
torch.CudaStorage
2016-01-23 13:36:54 +13:00
128
-0.00223906315 -0.00564192981 -0.000587438699 -0.00330007123 -0.0019847299 -0.00346800219 -0.000640053826 -0.00456859823 -0.00170282915 -0.0022419272 -0.00154888479 -0.00320651871 -0.00496730441 -0.00187510159 -0.00410069898 -0.00349972374 -0.00277556339 -0.00202337676 -0.00297997333 -0.00231172447 -0.00495815417 -0.000779633468 -0.0040405062 0.00326331914 -0.00151972042 -0.00302759325 -0.00405702228 -0.00239663781 -0.00152731896 -0.00490829721 -0.00504754297 -0.00187284348 0.00127841381 -0.00374626438 -0.00384993386 -0.0032561908 -0.00253101951 -0.00916513801 -0.00196512928 -0.001295669 -0.00136179372 -6.00566673e-05 -0.00156397955 -0.00268514385 0.000507580698 -0.00219269074 -0.00578507222 -0.000533854822 -0.00192734867 -0.00246739457 -0.00308515807 -0.00102079508 -0.00138736004 -0.00274345977 -0.00248077163 -0.003405581 -0.00275326753 -0.00367268268 -0.00276740384 -0.00293721841 -0.00324599189 4.50888219e-05 0.0117115788 -0.00288088177 -0.00264649489 -0.001961431 -0.00329766725 -0.00201006676 -0.00361273391 -0.00303857937 -0.00255476101 -0.00587752555 0.000791140657 -0.00183710654 -0.00182586268 -0.00118170062 -0.00352250179 -0.00347002875 -0.00113971275 -0.00306446874 -0.00430353777 -0.00181000098 -0.00262577226 0.00275858236 -0.000576139893 -0.00140603923 -0.00373536488 -0.00277283113 -0.00729715824 -0.00510852085 -0.00175524573 -0.00152605504 -0.00287000439 -0.00358065567 -0.00246332865 -0.000865404436 -0.000979622826 -0.0044229459 -0.000767677149 -0.00141597423 -0.00251506781 -0.0025236709 -0.000229029843 -0.000216791464 -0.0021874893 -0.00263210572 0.00100636878 -0.00269071059 -0.00111422734 -0.00197166344 -0.00143779465 -0.00108963833 0.000167679158 -0.00317580556 -0.00244263303 -0.000948414381 -0.0023896452 -0.00227860361 -0.00410877261 -0.00218364224 -0.00094621873 -0.00249538012 -0.00322123989 -0.00387443881 -0.00137600454 -0.00328964717 -0.00905571226 -0.00286922161
2
2
kH
1
3
2015-05-16 17:48:05 +12:00
2
6
finput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
89
3
V 1
16
torch.CudaTensor
0
2015-05-16 17:48:05 +12:00
1
0
2
2016-01-23 13:36:54 +13:00
6
weight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
90
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
2
128 576
576 1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
91
3
V 1
17
torch.CudaStorage
73728
-0.0639486089 0.0507110059 0.0581143312 -0.0413753875 0.0085651949 0.12872915 -0.0244468637 -0.0448594801 -0.02909307 -0.0441230536 -0.036086008 0.0207542591 -0.0768679902 -0.0360676274 0.0892052427 -0.00357042183 0.0102316868 -0.00797057152 0.110115305 0.00663193082 0.0180179365 -0.0225018859 -0.00695590116 -0.00352965808 -0.0173423309 -0.0689607859 -0.0456825122 -0.00688724499 0.0490705557 -0.0268124584 -0.0438884608 -0.0654067025 0.0526400134 -0.0573983081 0.0286624394 0.0701014698 0.0926500186 0.105845131 -0.0201315098 0.0476317592 -0.0425289758 0.0570359454 0.0351527445 -0.00217129779 0.0220436957 -0.0242458824 0.00908900704 0.0494639724 -0.0212628283 -0.0435516052 0.0340492316 0.0422502831 -0.060492795 0.0113678323 0.109308518 0.0398647226 -0.0388981402 0.0415344574 0.0414027385 0.00569209643 -0.0125053087 -0.00776875438 0.0137257129 -0.0381870195 0.0257881824 -0.000285399787 -0.0875598118 -0.0205913782 0.032968007 0.0965893939 0.0232098028 0.00392417097 -0.0748917535 -0.0424878672 0.0453223102 -0.139484972 -0.0288423616 0.0626257136 -0.124088533 -0.100739963 0.00555635057 0.0771999285 -0.116458185 0.0363215506 -0.0388342105 0.0454020128 -0.111697711 -0.0601412691 -0.0411670618 0.00127731333 -0.00629131962 -0.0252490621 -0.00320236944 -0.0501189157 0.0813165382 -0.0490916669 -0.0587035045 -0.000896489597 0.0527687222 -0.0342738442 0.0345124342 -0.0035688146 -0.0602004752 0.062783584 0.0412950441 -0.0743808076 -0.203249618 0.0831377283 0.0869094431 -0.110949315 -0.0156863444 -4.16268376e-05 0.0859856308 -0.074030295 -0.0100482609 -0.00466052955 0.00935512595 -0.0181866735 0.135321572 -0.0476934798 0.0271401107 0.0612289533 -0.0473434292 -0.0076686861 -0.068121843 0.00709619047 -0.0788855553 -0.0422128476 -0.0410357863 0.00386965089 0.00022240293 -0.0061468794 7.6152799e-05 0.0676732957 -0.0267101191 -0.0215029456 -0.116271846 -0.039861083 0.0227447022 -0.105358705 -0.0189924166 0.0441764258 0.00106705062 0.0233832337 0.021353554 0.0538178235 0.0880188197 -0.0980543494 0.0125292316 0.0679634288 -0.0653312653 -0.00144493999 -0.0560980923 0.0323222317 -0.0598481074 -0.0615543239 0.0111308303 -0.112719171 0.00944978092 -0.0919810534 -0.0695775822 -0.000483827782 -0.112097062 -0.00522450265 -0.084120661 -0.0175515544 0.0109363506 -0.0166239906 -0.019394543 -0.028752245 0.0677276179 -0.0186019912 -0.0405116007 0.0419634394 0.0946924016 -0.0041591553 -0.0236843117 0.0440292619 -0.118492335 0.0568900444 0.0164462961 -0.00579112908 -0.102494255 0.00941327307 -0.0157481916 0.00346715469 -0.0966883376 -0.00314031751 0.0850265622 0.0513992645 0.0266812313 0.0240473263 0.0287467465 0.0286135543 -0.0167474579 0.0516456738 0.12017931 -0.085002251 -0.0425869152 0.0902682766 -0.0439751372 -0.0179351065 -0.0572434701 0.0382519253 -0.0457536839 0.0151565447 0.0266079027 0.0274193622 -0.0109072085 0.0439492911 -0.0693976283 0.00106373231 -0.0232171491 0.0286410376 0.0109481383 0.00797311217 0.0333026648 0.0444752462 0.0573555492 0.0856771842 0.0895676464 0.0403928645 0.0152198225 -0.06748344 -0.0268928893 -0.00681458088 -0.0665264279 -0.0973685458 -0.0980488658 0.0450405739 -0.0089623807 -0.124396034 -0.0962544158 0.0381435901 -0.0447320342 0.0378432646 0.0566829145 -0.00988580287 0.106138252 0.0129292449 -0.0547625422 -0.00658249948 0.0863131285 0.1244298 0.107844666 -0.0407100953 -0.0333855413 0.00070688891 0.0121234544 0.00665215543 -0.012680606 -0.0783395544 -0.0222706143 -0.050198406 -0.0590017028 -0.0426145382 -0.0658667758 -0.0308767837 0.0385813117 0.122501366 0.0289310589 -0.0861140192 0.0149862571 -0.119373493 0.0778462514 0.00208398071 0.058219187 0.0771981254 -0.0886703804 -0.125639319 0.0430764109 0.0565429293 0.123243667 -0.0658384115 -0.0183514804 -0.0198131632 0.0134401983 -0.00318840658 0.0452178717 0.00378029211 0.042977456 -0.046553608 0.00499832444 -0.156650499 0.0175341535 0.0104571907 0.000234752064 0.0157815907 -0.0672010407 0.0269012786 -0.0907446966 0.0133643737 -0.032993082 -0.0497179478 -0.0288374294 -0.0394011773 0.0427763127 0.0430263169 0.150980547 0.0258503091 0.00199761242 -0.107541062 0.0341623463 -0.0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
10
gradWeight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
92
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
128 576
576 1
2016-01-23 13:36:54 +13:00
1
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
93
3
V 1
17
torch.CudaStorage
73728
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
10
fgradInput
4
2016-01-23 13:36:54 +13:00
94
3
V 1
16
torch.CudaTensor
0
2015-05-16 17:48:05 +12:00
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
4
padH
1
2016-01-23 13:36:54 +13:00
0
2
2
dH
1
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
2
2
kW
1
3
2015-05-16 17:48:05 +12:00
1
10
4
2016-01-23 13:36:54 +13:00
95
2015-05-16 17:48:05 +12:00
3
V 1
2016-01-23 13:36:54 +13:00
14
w2nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
96
5
2
2015-05-16 17:48:05 +12:00
5
2016-01-23 13:36:54 +13:00
train
5
0
2015-05-16 17:48:05 +12:00
2
8
negative
4
2016-01-23 13:36:54 +13:00
97
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
98
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
99
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
14
negative_scale
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
1
11
4
2016-01-23 13:36:54 +13:00
100
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
101
17
2
2016-01-23 13:36:54 +13:00
4
padW
2015-05-16 17:48:05 +12:00
1
0
2015-05-16 17:48:05 +12:00
2
11
nInputPlane
1
128
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
102
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
103
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2
2016-01-23 13:36:54 +13:00
8
gradBias
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
104
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
1
2016-01-23 13:36:54 +13:00
128
1
1
4
105
3
V 1
17
torch.CudaStorage
128
0 0 0 0 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186
2
2
dW
1
1
2015-05-16 17:48:05 +12:00
2
12
nOutputPlane
1
128
2
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
bias
4
106
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
1
128
1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
107
2015-05-16 17:48:05 +12:00
3
V 1
17
torch.CudaStorage
2016-01-23 13:36:54 +13:00
128
-0.0174424239 -0.00441494398 -0.00187091203 -0.0165972766 0.00520471577 -0.0113483509 -0.00601622881 -0.0155359032 -0.0145057924 -0.00650269166 -0.00170008885 0.00861964095 -0.0180432945 -0.0101737436 -0.0126734665 0.00169225538 0.0134350071 -0.01622965 -0.00284516532 -0.0101969559 -0.0204207134 -0.017208958 0.013335526 -0.00619678479 -0.00112969801 -0.00205706782 0.00169970118 -0.010120634 -0.0292057805 0.0036877871 -0.00973385852 -0.0193506107 -0.00288015697 -0.00541405194 -0.0181581769 -0.0124297356 -0.0204882678 -0.0202166028 -0.0172857568 -0.00462766457 0.00163251476 0.00128503167 0.0135944039 -0.0261138026 0.00154586066 -0.0157007147 -0.0199298505 -0.0197797194 -0.0128447739 -0.0131284324 -0.00141243555 -0.00224168831 -0.0383181721 -0.00396215776 -0.020214403 -0.0139810415 -0.00234710309 -0.0153242182 -0.0112998066 -0.0165441465 -0.0159104597 -0.015582785 -0.00528914481 -0.0244423039 0.0155893872 -0.0221247356 0.000490907405 -0.0145985559 -0.0116234263 -0.000833198021 0.00371065619 -0.00710693514 -0.0193245448 -0.00396315567 0.00883803423 0.0078935558 0.0432194173 -0.0194785874 -0.0148788011 0.00382716837 -0.00877236202 -0.0174138453 0.0105593596 0.00627399376 -0.0016207532 -0.00386415073 -0.00256360369 -0.0117404778 -0.00696396362 -0.0091393413 -0.0031364141 -0.0214340258 0.00298002292 -0.00435995497 -0.00125136808 -0.0187083986 -0.0191166587 -0.0181855224 -5.26358563e-05 -0.00116594823 -0.00649671489 -0.0133659756 -0.0075565055 0.0117118238 0.0113961753 -0.0189972464 -0.0177347306 -0.000115355899 -0.00205061096 -0.00974817015 -0.00553580886 0.000548948359 -0.0218877327 -0.00875826459 0.013626406 -0.0139888665 -0.0213929247 0.00512529491 -0.0141332233 -0.00318096671 0.0151170716 -0.00874019321 0.00335070072 -0.00161071855 -0.020651577 0.008146693 -0.00450002356 -0.00983705092
2
2
kH
1
3
2015-05-16 17:48:05 +12:00
2
6
finput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
108
3
V 1
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
0
2016-01-23 13:36:54 +13:00
1
0
2015-05-16 17:48:05 +12:00
2
6
weight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
109
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
128 1152
1152 1
2016-01-23 13:36:54 +13:00
1
4
2016-01-23 13:36:54 +13:00
110
3
V 1
17
torch.CudaStorage
147456
-0.0400025584 -0.0042039738 0.00169666356 0.0458069444 0.0677332655 0.0131308166 -0.031839177 0.0236217808 -0.0351602063 0.019839393 -0.0544954054 -0.00102115516 0.056029994 -0.0586317293 -0.0085212104 -0.00815554988 -0.0436800569 -0.013675157 -0.037730854 -0.0517962873 -0.0258446131 -0.0260447189 0.00241112942 0.00925655849 0.0229062904 -0.00683342572 0.0307585187 -0.0334891342 0.0284872632 0.013185163 -0.0948907807 -0.0430105478 -0.0626673251 -0.0151871424 0.01351678 0.0325069278 -0.00808036979 -0.0281409603 0.0389364846 -0.0428257212 0.0405550189 -0.0125908582 0.0558850057 -0.0243568718 0.0426079109 0.00622297591 -0.0627290681 -0.10408245 -0.0103451349 -0.0513008684 0.0111329053 0.0390800498 -0.0162936039 0.0681585073 -0.00641185977 -0.0912840441 0.00292424182 -0.0586039126 -0.0387858637 0.00478885928 -0.0969754308 -0.00944207888 -0.0337207913 -0.00977702532 0.00490565458 -0.0496883914 -0.00427272171 0.0278977696 -0.0743431225 -0.073481366 0.0248727817 -0.00130628864 -0.0220025424 0.00364075648 0.0656089336 0.00386532699 -0.0779560581 -0.00618968019 0.0763303414 -0.0696482658 0.0230925456 -0.00247984123 -0.0330726467 -0.0339083113 0.040401198 0.0551991388 0.0812772661 0.00379758794 -0.000533841725 0.00270939642 -0.107911795 -0.0389300846 0.129131183 -0.0289271139 -0.0108544296 0.0371486135 -0.0362867489 0.00352974725 0.00393852498 0.0225885361 0.0687984228 -0.0029700317 -0.0215473007 -0.0217019878 -0.0207775459 -0.0447932445 -0.0581997558 -0.0531094484 -0.0688454136 -0.0306101944 -0.0259571262 -0.0144862877 0.041382499 0.0284589063 -0.0178175885 0.0381619409 -0.00189444993 0.000963551924 0.0655809268 0.0246106982 -0.018788144 -0.023623744 -0.060844101 -0.0172380619 -0.0262618847 0.0439704843 0.000825537252 0.0626597106 0.0039220741 -0.0108984746 0.00393659528 -0.0130860191 -0.0542089939 -0.055677101 -0.0407339372 0.0390117988 0.0108563239 -0.0777411759 -0.00537123531 -0.033498887 0.0295539163 -0.00171005412 0.00107509666 0.0458830893 -0.117478199 0.0574893206 0.045923762 0.00100308063 0.0497508571 -0.0167082101 0.0373627953 0.0170439053 0.0183969084 0.0456029251 0.0219675899 -0.0214629974 0.0544017032 0.00636878517 -0.00400591781 -0.0581276156 -0.0287636463 0.0164341703 0.0210215393 -0.0552279092 -0.0431153253 -0.00821072794 0.00266696396 -0.0415195525 0.0770807713 -0.00733295968 0.0583500899 -0.0308851711 -0.00999466609 -0.0181834754 -0.0174055714 -0.022853693 -0.0171624515 -0.00134173303 -0.0257898401 -0.0527768172 0.00561083481 -0.115395099 -0.00924043171 -0.0263971128 -0.0321370326 0.056985382 0.0265887193 0.062056113 -0.0374869183 -0.0351216421 0.0231997333 0.0405739248 0.00376361585 -0.00591063034 0.0650092363 -0.000734779635 -0.028649319 0.00905133225 -0.0170206968 0.00356774684 -0.0201012064 0.0840449631 -0.02928886 -0.0294103678 -0.0331360213 0.0355949588 0.0538641848 -0.0531014428 -0.0717715919 -0.000815480831 0.0282980762 -0.0307574999 -0.0193714388 0.0143439705 0.0126337437 0.0414530784 -0.0203958545 0.0243392475 0.0336823501 -0.0439205319 0.00352710485 0.00341798319 0.0524835736 0.021519497 -0.0402918458 -0.0536110513 -0.0245709457 0.00758181559 0.0911753029 0.0788967609 0.0456496887 -0.00408682739 0.0153079415 0.0317227766 0.0524746589 0.0485638268 -0.0562876947 0.00534088677 0.0469402112 0.0406302959 -0.0240072235 -0.0467492156 0.0713637248 0.000732375367 0.0457775705 0.00731991371 -0.0118685784 0.0264630672 0.0264361314 0.0205160677 0.0125145586 -0.0102423821 -0.0326954126 -0.0198787153 -0.0383117273 -0.0335361771 -0.0861471444 -0.0288037658 -0.0334756896 0.0121201202 0.044227168 -0.0432401672 0.0455558747 -0.115723431 -0.016386928 0.0102228913 0.0401716791 -0.0447459891 -0.00483643776 -0.0159691572 -0.0247466806 -0.00981854182 0.023699956 0.0277846362 0.0413833261 -0.0305659771 0.058231242 0.0464493744 0.0459218062 -0.00233360543 0.0740427151 -0.00711035542 -0.0541210473 0.0393931717 0.0426493138 0.0204852931 -0.0117088947 -0.00452799629 -0.00534024742 -0.0228706226 0.00128425611 -0.0385257713 -0.0748715252 0.00234347046 0.00243146392 0.0293783806 0.0563902929 0.0270989835 -0.0328624323 0.0481743105
2015-05-16 17:48:05 +12:00
2
5
train
5
2015-05-16 17:48:05 +12:00
0
2
2016-01-23 13:36:54 +13:00
10
gradWeight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
111
3
V 1
16
torch.CudaTensor
2
128 1152
1152 1
1
4
112
3
V 1
17
torch.CudaStorage
147456
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
2
10
fgradInput
4
113
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
4
padH
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2
dH
2015-05-16 17:48:05 +12:00
1
1
2016-01-23 13:36:54 +13:00
2
2
kW
1
3
2015-05-16 17:48:05 +12:00
1
12
4
2016-01-23 13:36:54 +13:00
114
2015-05-16 17:48:05 +12:00
3
V 1
2016-01-23 13:36:54 +13:00
14
w2nn.LeakyReLU
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
115
5
2
5
train
2015-05-16 17:48:05 +12:00
5
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
8
negative
4
2016-01-23 13:36:54 +13:00
116
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
117
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
118
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2016-01-23 13:36:54 +13:00
2
14
negative_scale
1
0.10000000000000001
2015-05-16 17:48:05 +12:00
1
13
4
2016-01-23 13:36:54 +13:00
119
2015-05-16 17:48:05 +12:00
3
V 1
23
nn.SpatialConvolutionMM
2015-05-16 17:48:05 +12:00
3
2016-01-23 13:36:54 +13:00
120
17
2
2016-01-23 13:36:54 +13:00
4
padW
2015-05-16 17:48:05 +12:00
1
0
2015-05-16 17:48:05 +12:00
2
11
nInputPlane
1
128
2
2016-01-23 13:36:54 +13:00
6
output
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
121
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
9
gradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
122
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
0
1
0
2
2016-01-23 13:36:54 +13:00
8
gradBias
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
123
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
1
1
1
1
4
124
3
V 1
17
torch.CudaStorage
2015-05-16 17:48:05 +12:00
1
0
2
2016-01-23 13:36:54 +13:00
2
dW
1
1
2
12
nOutputPlane
2015-05-16 17:48:05 +12:00
1
1
2
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
bias
4
125
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
1
1
1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
126
3
V 1
17
torch.CudaStorage
2016-01-23 13:36:54 +13:00
1
0.013519641
2
2
kH
1
3
2015-05-16 17:48:05 +12:00
2
6
finput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
127
3
V 1
16
torch.CudaTensor
2015-05-16 17:48:05 +12:00
0
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
6
weight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
128
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2016-01-23 13:36:54 +13:00
2
1 1152
1152 1
2015-05-16 17:48:05 +12:00
1
4
2016-01-23 13:36:54 +13:00
129
3
V 1
17
torch.CudaStorage
1152
-0.968702197 -0.246173263 -0.395858705 0.0110244025 -0.0879306644 0.797686398 0.273706704 0.579370201 -0.658405423 -0.593995214 0.978542507 -0.0317310132 0.54101783 -0.152037203 -0.971608043 -0.262601584 -0.153601333 0.645044029 -0.0964138731 0.049069684 0.133027658 -0.0599497259 -0.484507799 0.533405602 -0.522983074 0.822757304 -0.371751308 -0.131353572 0.123249166 0.136895314 -0.554207385 -0.165685833 -1.15162587 -0.174625605 0.71651268 0.488345057 -0.440272868 -0.580357432 0.301537246 0.900098443 -0.207345232 -0.615330338 -0.291121334 1.01440406 -0.0833065882 -0.415342003 0.456117034 -0.263134778 -0.0470403172 0.236230195 -0.00790908001 0.211072177 -0.316390634 0.104367696 0.398835391 0.464314401 -0.712735891 -0.0719743744 -0.713236153 -0.311996669 0.239423156 0.532402039 0.176131666 0.578389108 -0.433353752 -0.204792917 0.17000331 0.437562406 0.413770109 0.572913229 0.394384295 0.797740936 -0.0999602675 0.487336159 -0.288182348 0.132213697 -0.181360096 -0.946593761 -0.240955427 -0.551261961 -0.089645043 -0.0603546463 0.22348918 0.0318763815 0.0746425539 -0.21006313 -0.10775663 -0.0320423655 0.308666259 -0.219774425 -0.50052011 0.0252206046 -0.12409015 0.830070496 0.114050299 -0.819277704 0.0173523538 0.238986656 0.206469372 -0.85600847 -0.0172789171 -0.107213371 -0.263297737 0.24157232 0.68258965 0.359622926 0.311884344 0.122122355 -0.441600561 0.679896891 0.0467527583 -0.156419367 -0.382695824 0.735375762 -0.368394822 -0.473194987 -0.466644496 -0.015559095 1.17887628 -0.460936844 -0.216077223 0.370923549 -0.268609852 -0.0132896313 0.286484331 0.442310393 -0.841442764 0.0478871614 -1.06191254 0.0937394723 0.0845699757 0.125022337 0.0761328116 0.325423896 0.286850959 0.382697314 -0.484265774 -0.139350325 0.352079481 0.601100743 0.183256954 -0.8327806 0.566294134 -0.639612734 0.862764537 0.31221357 -0.51698035 -1.43437541 0.842086256 -0.129325926 -0.269518763 0.419676065 0.388411939 -0.0276997294 0.998208225 -0.259724975 0.628245533 0.170235768 -0.683086157 -0.359109789 -0.417966753 -0.696670175 -0.169076785 0.0075237914 0.149537861 0.248381644 -0.167030707 0.490167856 -0.225653186 -0.0162650645 -0.329533756 1.07293355 -0.678185821 0.0714018121 0.323557079 -0.532817185 0.471875846 -0.635474086 -0.528421164 0.272524774 -0.593776524 0.320925325 -0.279917002 -0.247242004 -0.138337314 -0.63525784 0.0867606252 0.39261049 0.209438786 0.353735864 -0.566178024 0.173131973 0.155094489 -0.0387236364 -0.565529764 -0.646077335 0.0392730422 -0.794628501 -0.770301104 -0.078206636 0.220839873 0.399475515 -0.0321876407 0.320349097 -0.216572434 0.482001305 0.396315962 0.198529616 -0.187269658 -0.478836 -0.65039748 -0.186620668 0.0884803683 0.664992332 0.511987805 0.0407209471 -0.0853051469 -0.174727336 0.253174782 0.251588166 -0.247844681 0.150150284 -0.159300223 0.315300792 -0.300532192 -0.317200124 -0.294903457 0.248781651 0.0266141817 0.165646091 0.85802871 -0.374940008 0.240412101 -0.551757574 -0.0913677812 0.238110036 -0.473404586 0.502745986 -0.275217056 0.25128755 0.0522941127 -0.463459581 0.258424133 0.705428421 0.18117775 -0.549119771 0.441701859 -0.567652822 -0.17298986 0.156832725 -0.079443641 -0.114284441 0.059267506 -0.00974048115 0.168199211 0.376075506 -0.421881109 -0.215551749 0.168736607 0.0959603563 -0.179641098 -0.0868079439 0.115968391 -0.00479767565 -0.384005964 -0.244407281 0.452850342 -0.180145845 0.101040356 0.634118259 -0.19870849 -0.677038014 -0.293557853 0.51081425 -0.223032132 0.468648434 -0.668595552 0.499418885 0.583018601 0.435648382 -0.0798027366 -0.0170586854 -0.155411482 -0.492242157 -0.0291862674 -0.934619308 -0.157041609 -0.0758312568 0.0945741013 0.315035939 1.14061475 -0.733324587 -0.216986477 0.0992171019 0.022143662 0.232617706 -0.833373725 0.0974382684 0.0753710642 0.663383842 -0.187995821 -0.477160275 -0.190994039 0.0255216397 -0.107552201 0.106859446 0.15032205 -1.52112138 -0.251844883 -0.0580662042 0.460049272 -0.21003972 0.594682217 -0.118438385 -0.56948024 -0.479211241 0.236374199 0.390817583 0.482779711 0.0809394047 -0.619490921 -0.353634834 -0.0167122595 -0.458844811 -0.172951683 -0.6666
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
5
train
5
0
2
10
gradWeight
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
130
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
2
1 1152
1152 1
2016-01-23 13:36:54 +13:00
1
4
2016-01-23 13:36:54 +13:00
131
3
V 1
17
torch.CudaStorage
1152
0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781250186 0.00781
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
10
fgradInput
2015-05-16 17:48:05 +12:00
4
2016-01-23 13:36:54 +13:00
132
2015-05-16 17:48:05 +12:00
3
V 1
16
torch.CudaTensor
0
1
0
2015-05-16 17:48:05 +12:00
2
2016-01-23 13:36:54 +13:00
4
padH
2015-05-16 17:48:05 +12:00
1
2016-01-23 13:36:54 +13:00
0
2015-05-16 17:48:05 +12:00
2
2
dH
2015-05-16 17:48:05 +12:00
1
1
2016-01-23 13:36:54 +13:00
2
2
kW
1
3
2015-05-16 17:48:05 +12:00
1
14
4
2016-01-23 13:36:54 +13:00
133
2015-05-16 17:48:05 +12:00
3
V 1
7
nn.View
3
2016-01-23 13:36:54 +13:00
134
4
2
12
numInputDims
1
3
2015-05-16 17:48:05 +12:00
2
4
size
4
2016-01-23 13:36:54 +13:00
135
3
V 1
17
torch.LongStorage
1
-1
2
2016-01-23 13:36:54 +13:00
5
train
5
2015-05-16 17:48:05 +12:00
0
2
11
numElements
1
1
2015-05-16 17:48:05 +12:00
2
5
train
5
0