osbm commited on
Commit
fa601fc
1 Parent(s): d7b0121

Delete Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres

Browse files
Files changed (16) hide show
  1. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/dataset.json +0 -18
  2. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/dataset_fingerprint.json +0 -1426
  3. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/.ipynb_checkpoints/progress-checkpoint.png +0 -3
  4. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/checkpoint_best.pth +0 -3
  5. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/checkpoint_latest.pth +0 -3
  6. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/debug.json +0 -52
  7. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/progress.png +0 -3
  8. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/training_log_2023_7_24_00_01_52.txt +0 -1194
  9. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/.ipynb_checkpoints/Untitled-checkpoint.ipynb +0 -6
  10. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/.ipynb_checkpoints/progress-checkpoint.png +0 -3
  11. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/Untitled.ipynb +0 -112
  12. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/checkpoint_best.pth +0 -3
  13. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/debug.json +0 -52
  14. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/progress.png +0 -3
  15. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/training_log_2023_7_24_09_31_46.txt +0 -170
  16. Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/plans.json +0 -342
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/dataset.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "name": "Prostate158",
3
- "description": "Prostate cancer segmentation dataset",
4
- "channel_names": {
5
- "0": "T2",
6
- "1": "ADC",
7
- "2": "DFI"
8
- },
9
- "labels": {
10
- "background": 0,
11
- "prostate_inner": 1,
12
- "prostate_outer": 2,
13
- "tumor": 3
14
- },
15
- "numTraining": 139,
16
- "numTest": 19,
17
- "file_ending": ".nii.gz"
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/dataset_fingerprint.json DELETED
@@ -1,1426 +0,0 @@
1
- {
2
- "foreground_intensity_properties_per_channel": {
3
- "0": {
4
- "max": 1694.0,
5
- "mean": 267.35308837890625,
6
- "median": 242.0,
7
- "min": 0.0,
8
- "percentile_00_5": 36.0,
9
- "percentile_99_5": 768.0,
10
- "std": 136.11251831054688
11
- },
12
- "1": {
13
- "max": 3557.286865234375,
14
- "mean": 1215.81591796875,
15
- "median": 1203.8331298828125,
16
- "min": 0.0,
17
- "percentile_00_5": 0.0,
18
- "percentile_99_5": 2259.82861328125,
19
- "std": 338.6748352050781
20
- },
21
- "2": {
22
- "max": 198.95455932617188,
23
- "mean": 72.26309204101562,
24
- "median": 70.3214340209961,
25
- "min": 0.0,
26
- "percentile_00_5": 34.534385681152344,
27
- "percentile_99_5": 132.71939086914062,
28
- "std": 18.909290313720703
29
- }
30
- },
31
- "median_relative_size_after_cropping": 1.0,
32
- "shapes_after_crop": [
33
- [
34
- 26,
35
- 270,
36
- 270
37
- ],
38
- [
39
- 26,
40
- 270,
41
- 270
42
- ],
43
- [
44
- 24,
45
- 270,
46
- 270
47
- ],
48
- [
49
- 25,
50
- 442,
51
- 442
52
- ],
53
- [
54
- 24,
55
- 270,
56
- 270
57
- ],
58
- [
59
- 24,
60
- 270,
61
- 270
62
- ],
63
- [
64
- 24,
65
- 270,
66
- 270
67
- ],
68
- [
69
- 24,
70
- 270,
71
- 270
72
- ],
73
- [
74
- 25,
75
- 442,
76
- 442
77
- ],
78
- [
79
- 24,
80
- 270,
81
- 270
82
- ],
83
- [
84
- 24,
85
- 270,
86
- 270
87
- ],
88
- [
89
- 26,
90
- 270,
91
- 270
92
- ],
93
- [
94
- 26,
95
- 270,
96
- 270
97
- ],
98
- [
99
- 24,
100
- 270,
101
- 270
102
- ],
103
- [
104
- 24,
105
- 270,
106
- 270
107
- ],
108
- [
109
- 24,
110
- 270,
111
- 270
112
- ],
113
- [
114
- 26,
115
- 270,
116
- 270
117
- ],
118
- [
119
- 24,
120
- 270,
121
- 270
122
- ],
123
- [
124
- 24,
125
- 270,
126
- 270
127
- ],
128
- [
129
- 28,
130
- 270,
131
- 270
132
- ],
133
- [
134
- 24,
135
- 270,
136
- 270
137
- ],
138
- [
139
- 26,
140
- 270,
141
- 270
142
- ],
143
- [
144
- 25,
145
- 442,
146
- 442
147
- ],
148
- [
149
- 25,
150
- 442,
151
- 442
152
- ],
153
- [
154
- 25,
155
- 442,
156
- 442
157
- ],
158
- [
159
- 24,
160
- 270,
161
- 270
162
- ],
163
- [
164
- 24,
165
- 270,
166
- 270
167
- ],
168
- [
169
- 27,
170
- 232,
171
- 232
172
- ],
173
- [
174
- 24,
175
- 270,
176
- 270
177
- ],
178
- [
179
- 30,
180
- 232,
181
- 232
182
- ],
183
- [
184
- 27,
185
- 232,
186
- 232
187
- ],
188
- [
189
- 24,
190
- 270,
191
- 270
192
- ],
193
- [
194
- 25,
195
- 442,
196
- 442
197
- ],
198
- [
199
- 24,
200
- 270,
201
- 270
202
- ],
203
- [
204
- 24,
205
- 270,
206
- 270
207
- ],
208
- [
209
- 26,
210
- 270,
211
- 270
212
- ],
213
- [
214
- 24,
215
- 270,
216
- 270
217
- ],
218
- [
219
- 24,
220
- 270,
221
- 270
222
- ],
223
- [
224
- 27,
225
- 442,
226
- 442
227
- ],
228
- [
229
- 25,
230
- 442,
231
- 442
232
- ],
233
- [
234
- 25,
235
- 442,
236
- 442
237
- ],
238
- [
239
- 35,
240
- 270,
241
- 270
242
- ],
243
- [
244
- 25,
245
- 442,
246
- 442
247
- ],
248
- [
249
- 24,
250
- 270,
251
- 270
252
- ],
253
- [
254
- 24,
255
- 270,
256
- 270
257
- ],
258
- [
259
- 24,
260
- 270,
261
- 270
262
- ],
263
- [
264
- 26,
265
- 270,
266
- 270
267
- ],
268
- [
269
- 24,
270
- 270,
271
- 270
272
- ],
273
- [
274
- 24,
275
- 270,
276
- 270
277
- ],
278
- [
279
- 35,
280
- 442,
281
- 442
282
- ],
283
- [
284
- 26,
285
- 270,
286
- 270
287
- ],
288
- [
289
- 24,
290
- 270,
291
- 270
292
- ],
293
- [
294
- 24,
295
- 270,
296
- 270
297
- ],
298
- [
299
- 31,
300
- 442,
301
- 442
302
- ],
303
- [
304
- 26,
305
- 270,
306
- 270
307
- ],
308
- [
309
- 30,
310
- 442,
311
- 442
312
- ],
313
- [
314
- 27,
315
- 232,
316
- 232
317
- ],
318
- [
319
- 27,
320
- 232,
321
- 232
322
- ],
323
- [
324
- 25,
325
- 442,
326
- 442
327
- ],
328
- [
329
- 24,
330
- 270,
331
- 270
332
- ],
333
- [
334
- 24,
335
- 270,
336
- 270
337
- ],
338
- [
339
- 24,
340
- 270,
341
- 270
342
- ],
343
- [
344
- 24,
345
- 270,
346
- 270
347
- ],
348
- [
349
- 28,
350
- 270,
351
- 270
352
- ],
353
- [
354
- 24,
355
- 270,
356
- 270
357
- ],
358
- [
359
- 31,
360
- 442,
361
- 442
362
- ],
363
- [
364
- 24,
365
- 270,
366
- 270
367
- ],
368
- [
369
- 26,
370
- 270,
371
- 270
372
- ],
373
- [
374
- 26,
375
- 270,
376
- 270
377
- ],
378
- [
379
- 24,
380
- 270,
381
- 270
382
- ],
383
- [
384
- 24,
385
- 270,
386
- 270
387
- ],
388
- [
389
- 24,
390
- 270,
391
- 270
392
- ],
393
- [
394
- 25,
395
- 442,
396
- 442
397
- ],
398
- [
399
- 26,
400
- 270,
401
- 270
402
- ],
403
- [
404
- 27,
405
- 232,
406
- 232
407
- ],
408
- [
409
- 24,
410
- 270,
411
- 270
412
- ],
413
- [
414
- 29,
415
- 270,
416
- 270
417
- ],
418
- [
419
- 24,
420
- 270,
421
- 270
422
- ],
423
- [
424
- 26,
425
- 270,
426
- 270
427
- ],
428
- [
429
- 26,
430
- 232,
431
- 232
432
- ],
433
- [
434
- 24,
435
- 270,
436
- 270
437
- ],
438
- [
439
- 27,
440
- 232,
441
- 232
442
- ],
443
- [
444
- 26,
445
- 270,
446
- 270
447
- ],
448
- [
449
- 25,
450
- 442,
451
- 442
452
- ],
453
- [
454
- 26,
455
- 270,
456
- 270
457
- ],
458
- [
459
- 24,
460
- 270,
461
- 270
462
- ],
463
- [
464
- 24,
465
- 270,
466
- 270
467
- ],
468
- [
469
- 30,
470
- 442,
471
- 442
472
- ],
473
- [
474
- 28,
475
- 270,
476
- 270
477
- ],
478
- [
479
- 24,
480
- 270,
481
- 270
482
- ],
483
- [
484
- 24,
485
- 270,
486
- 270
487
- ],
488
- [
489
- 24,
490
- 270,
491
- 270
492
- ],
493
- [
494
- 25,
495
- 442,
496
- 442
497
- ],
498
- [
499
- 31,
500
- 442,
501
- 442
502
- ],
503
- [
504
- 25,
505
- 442,
506
- 442
507
- ],
508
- [
509
- 25,
510
- 442,
511
- 442
512
- ],
513
- [
514
- 30,
515
- 232,
516
- 232
517
- ],
518
- [
519
- 24,
520
- 270,
521
- 270
522
- ],
523
- [
524
- 26,
525
- 270,
526
- 270
527
- ],
528
- [
529
- 31,
530
- 442,
531
- 442
532
- ],
533
- [
534
- 24,
535
- 270,
536
- 270
537
- ],
538
- [
539
- 26,
540
- 270,
541
- 270
542
- ],
543
- [
544
- 26,
545
- 270,
546
- 270
547
- ],
548
- [
549
- 24,
550
- 270,
551
- 270
552
- ],
553
- [
554
- 25,
555
- 442,
556
- 442
557
- ],
558
- [
559
- 24,
560
- 270,
561
- 270
562
- ],
563
- [
564
- 26,
565
- 270,
566
- 270
567
- ],
568
- [
569
- 27,
570
- 442,
571
- 442
572
- ],
573
- [
574
- 24,
575
- 270,
576
- 270
577
- ],
578
- [
579
- 24,
580
- 270,
581
- 270
582
- ],
583
- [
584
- 24,
585
- 270,
586
- 270
587
- ],
588
- [
589
- 28,
590
- 270,
591
- 270
592
- ],
593
- [
594
- 24,
595
- 270,
596
- 270
597
- ],
598
- [
599
- 30,
600
- 270,
601
- 270
602
- ],
603
- [
604
- 24,
605
- 270,
606
- 270
607
- ],
608
- [
609
- 24,
610
- 270,
611
- 270
612
- ],
613
- [
614
- 25,
615
- 442,
616
- 442
617
- ],
618
- [
619
- 24,
620
- 270,
621
- 270
622
- ],
623
- [
624
- 24,
625
- 270,
626
- 270
627
- ],
628
- [
629
- 24,
630
- 270,
631
- 270
632
- ],
633
- [
634
- 28,
635
- 270,
636
- 270
637
- ],
638
- [
639
- 24,
640
- 270,
641
- 270
642
- ],
643
- [
644
- 40,
645
- 270,
646
- 270
647
- ],
648
- [
649
- 24,
650
- 270,
651
- 270
652
- ],
653
- [
654
- 24,
655
- 270,
656
- 270
657
- ],
658
- [
659
- 27,
660
- 232,
661
- 232
662
- ],
663
- [
664
- 24,
665
- 270,
666
- 270
667
- ],
668
- [
669
- 26,
670
- 270,
671
- 270
672
- ],
673
- [
674
- 25,
675
- 442,
676
- 442
677
- ],
678
- [
679
- 27,
680
- 232,
681
- 232
682
- ],
683
- [
684
- 26,
685
- 270,
686
- 270
687
- ],
688
- [
689
- 24,
690
- 270,
691
- 270
692
- ],
693
- [
694
- 25,
695
- 442,
696
- 442
697
- ],
698
- [
699
- 24,
700
- 270,
701
- 270
702
- ],
703
- [
704
- 24,
705
- 270,
706
- 270
707
- ],
708
- [
709
- 24,
710
- 270,
711
- 270
712
- ],
713
- [
714
- 24,
715
- 270,
716
- 270
717
- ],
718
- [
719
- 25,
720
- 442,
721
- 442
722
- ],
723
- [
724
- 24,
725
- 270,
726
- 270
727
- ]
728
- ],
729
- "spacings": [
730
- [
731
- 3.0,
732
- 0.4017857015132904,
733
- 0.4017857015132904
734
- ],
735
- [
736
- 3.0,
737
- 0.4017857015132904,
738
- 0.4017857015132904
739
- ],
740
- [
741
- 3.0,
742
- 0.4017857015132904,
743
- 0.4017857015132904
744
- ],
745
- [
746
- 3.000002384185791,
747
- 0.27173900604248047,
748
- 0.27173900604248047
749
- ],
750
- [
751
- 3.000000238418579,
752
- 0.4017857015132904,
753
- 0.4017857015132904
754
- ],
755
- [
756
- 3.000000238418579,
757
- 0.4017857015132904,
758
- 0.4017857015132904
759
- ],
760
- [
761
- 2.999999761581421,
762
- 0.4017857015132904,
763
- 0.4017857015132904
764
- ],
765
- [
766
- 3.000000238418579,
767
- 0.4017857015132904,
768
- 0.4017857015132904
769
- ],
770
- [
771
- 3.0000030994415283,
772
- 0.27173900604248047,
773
- 0.27173900604248047
774
- ],
775
- [
776
- 3.0,
777
- 0.4017857015132904,
778
- 0.4017857015132904
779
- ],
780
- [
781
- 2.999999761581421,
782
- 0.4017857015132904,
783
- 0.4017857015132904
784
- ],
785
- [
786
- 3.0,
787
- 0.4464285671710968,
788
- 0.4464285671710968
789
- ],
790
- [
791
- 3.0,
792
- 0.4017857015132904,
793
- 0.4017857015132904
794
- ],
795
- [
796
- 3.0,
797
- 0.4017857015132904,
798
- 0.4017857015132904
799
- ],
800
- [
801
- 3.000000238418579,
802
- 0.4017857015132904,
803
- 0.4017857015132904
804
- ],
805
- [
806
- 3.0,
807
- 0.4017857015132904,
808
- 0.4017857015132904
809
- ],
810
- [
811
- 3.0,
812
- 0.4017857015132904,
813
- 0.4017857015132904
814
- ],
815
- [
816
- 3.0,
817
- 0.4017857015132904,
818
- 0.4017857015132904
819
- ],
820
- [
821
- 3.0,
822
- 0.4017857015132904,
823
- 0.4017857015132904
824
- ],
825
- [
826
- 3.0,
827
- 0.4017857015132904,
828
- 0.4017857015132904
829
- ],
830
- [
831
- 3.0,
832
- 0.4017857015132904,
833
- 0.4017857015132904
834
- ],
835
- [
836
- 3.000000238418579,
837
- 0.4017857015132904,
838
- 0.4017857015132904
839
- ],
840
- [
841
- 2.999998092651367,
842
- 0.27173900604248047,
843
- 0.27173900604248047
844
- ],
845
- [
846
- 2.9999985694885254,
847
- 0.27173900604248047,
848
- 0.27173900604248047
849
- ],
850
- [
851
- 3.000016689300537,
852
- 0.27173900604248047,
853
- 0.27173900604248047
854
- ],
855
- [
856
- 3.0,
857
- 0.4017857015132904,
858
- 0.4017857015132904
859
- ],
860
- [
861
- 3.0,
862
- 0.4017857015132904,
863
- 0.4017857015132904
864
- ],
865
- [
866
- 3.0000035762786865,
867
- 0.46875,
868
- 0.46875
869
- ],
870
- [
871
- 3.0,
872
- 0.4017857015132904,
873
- 0.4017857015132904
874
- ],
875
- [
876
- 2.9999992847442627,
877
- 0.46875,
878
- 0.46875
879
- ],
880
- [
881
- 2.9999945163726807,
882
- 0.46875,
883
- 0.46875
884
- ],
885
- [
886
- 3.000000238418579,
887
- 0.4017857015132904,
888
- 0.4017857015132904
889
- ],
890
- [
891
- 3.0000016689300537,
892
- 0.27173900604248047,
893
- 0.27173900604248047
894
- ],
895
- [
896
- 3.0,
897
- 0.4017857015132904,
898
- 0.4017857015132904
899
- ],
900
- [
901
- 3.0,
902
- 0.4017857015132904,
903
- 0.4017857015132904
904
- ],
905
- [
906
- 3.0,
907
- 0.4017857015132904,
908
- 0.4017857015132904
909
- ],
910
- [
911
- 3.0,
912
- 0.4017857015132904,
913
- 0.4017857015132904
914
- ],
915
- [
916
- 3.0,
917
- 0.4017857015132904,
918
- 0.4017857015132904
919
- ],
920
- [
921
- 2.9999990463256836,
922
- 0.27173900604248047,
923
- 0.27173900604248047
924
- ],
925
- [
926
- 2.9999988079071045,
927
- 0.27173900604248047,
928
- 0.27173900604248047
929
- ],
930
- [
931
- 2.999993085861206,
932
- 0.27173900604248047,
933
- 0.27173900604248047
934
- ],
935
- [
936
- 3.0,
937
- 0.4017857015132904,
938
- 0.4017857015132904
939
- ],
940
- [
941
- 2.999997854232788,
942
- 0.27173900604248047,
943
- 0.27173900604248047
944
- ],
945
- [
946
- 3.0,
947
- 0.4017857015132904,
948
- 0.4017857015132904
949
- ],
950
- [
951
- 3.0,
952
- 0.4017857015132904,
953
- 0.4017857015132904
954
- ],
955
- [
956
- 3.0,
957
- 0.4017857015132904,
958
- 0.4017857015132904
959
- ],
960
- [
961
- 3.000000238418579,
962
- 0.4017857015132904,
963
- 0.4017857015132904
964
- ],
965
- [
966
- 3.0,
967
- 0.4017857015132904,
968
- 0.4017857015132904
969
- ],
970
- [
971
- 3.0,
972
- 0.4017857015132904,
973
- 0.4017857015132904
974
- ],
975
- [
976
- 3.0000011920928955,
977
- 0.27173900604248047,
978
- 0.27173900604248047
979
- ],
980
- [
981
- 3.0,
982
- 0.4017857015132904,
983
- 0.4017857015132904
984
- ],
985
- [
986
- 3.0,
987
- 0.4017857015132904,
988
- 0.4017857015132904
989
- ],
990
- [
991
- 3.0,
992
- 0.4017857015132904,
993
- 0.4017857015132904
994
- ],
995
- [
996
- 3.000004768371582,
997
- 0.27173900604248047,
998
- 0.27173900604248047
999
- ],
1000
- [
1001
- 3.0000159740448,
1002
- 0.4017859995365143,
1003
- 0.4017859995365143
1004
- ],
1005
- [
1006
- 2.9999992847442627,
1007
- 0.27173900604248047,
1008
- 0.27173900604248047
1009
- ],
1010
- [
1011
- 2.9999992847442627,
1012
- 0.46875,
1013
- 0.46875
1014
- ],
1015
- [
1016
- 2.9999849796295166,
1017
- 0.46875,
1018
- 0.46875
1019
- ],
1020
- [
1021
- 3.0000016689300537,
1022
- 0.27173900604248047,
1023
- 0.27173900604248047
1024
- ],
1025
- [
1026
- 3.0,
1027
- 0.4017857015132904,
1028
- 0.4017857015132904
1029
- ],
1030
- [
1031
- 3.000000238418579,
1032
- 0.4017857015132904,
1033
- 0.4017857015132904
1034
- ],
1035
- [
1036
- 3.0,
1037
- 0.4017857015132904,
1038
- 0.4017857015132904
1039
- ],
1040
- [
1041
- 3.0,
1042
- 0.4017857015132904,
1043
- 0.4017857015132904
1044
- ],
1045
- [
1046
- 3.0,
1047
- 0.4017857015132904,
1048
- 0.4017857015132904
1049
- ],
1050
- [
1051
- 3.0,
1052
- 0.4017857015132904,
1053
- 0.4017857015132904
1054
- ],
1055
- [
1056
- 2.999997854232788,
1057
- 0.27173900604248047,
1058
- 0.27173900604248047
1059
- ],
1060
- [
1061
- 3.0,
1062
- 0.4017857015132904,
1063
- 0.4017857015132904
1064
- ],
1065
- [
1066
- 3.0,
1067
- 0.4017857015132904,
1068
- 0.4017857015132904
1069
- ],
1070
- [
1071
- 3.0,
1072
- 0.4017857015132904,
1073
- 0.4017857015132904
1074
- ],
1075
- [
1076
- 3.0,
1077
- 0.4017857015132904,
1078
- 0.4017857015132904
1079
- ],
1080
- [
1081
- 3.0,
1082
- 0.4017857015132904,
1083
- 0.4017857015132904
1084
- ],
1085
- [
1086
- 3.0,
1087
- 0.4017857015132904,
1088
- 0.4017857015132904
1089
- ],
1090
- [
1091
- 3.0,
1092
- 0.27173900604248047,
1093
- 0.27173900604248047
1094
- ],
1095
- [
1096
- 3.0,
1097
- 0.4017857015132904,
1098
- 0.4017857015132904
1099
- ],
1100
- [
1101
- 3.0000014305114746,
1102
- 0.46875,
1103
- 0.46875
1104
- ],
1105
- [
1106
- 2.999999761581421,
1107
- 0.4017857015132904,
1108
- 0.4017857015132904
1109
- ],
1110
- [
1111
- 3.0,
1112
- 0.4017857015132904,
1113
- 0.4017857015132904
1114
- ],
1115
- [
1116
- 3.0,
1117
- 0.4017857015132904,
1118
- 0.4017857015132904
1119
- ],
1120
- [
1121
- 3.0,
1122
- 0.4017857015132904,
1123
- 0.4017857015132904
1124
- ],
1125
- [
1126
- 3.0,
1127
- 0.46875,
1128
- 0.46875
1129
- ],
1130
- [
1131
- 3.0,
1132
- 0.4017857015132904,
1133
- 0.4017857015132904
1134
- ],
1135
- [
1136
- 3.00002121925354,
1137
- 0.46875,
1138
- 0.46875
1139
- ],
1140
- [
1141
- 3.0,
1142
- 0.4017857015132904,
1143
- 0.4017857015132904
1144
- ],
1145
- [
1146
- 3.0,
1147
- 0.27173900604248047,
1148
- 0.27173900604248047
1149
- ],
1150
- [
1151
- 3.0,
1152
- 0.4017857015132904,
1153
- 0.4017857015132904
1154
- ],
1155
- [
1156
- 3.0,
1157
- 0.4017857015132904,
1158
- 0.4017857015132904
1159
- ],
1160
- [
1161
- 3.0,
1162
- 0.4017857015132904,
1163
- 0.4017857015132904
1164
- ],
1165
- [
1166
- 3.000000476837158,
1167
- 0.27173900604248047,
1168
- 0.27173900604248047
1169
- ],
1170
- [
1171
- 3.0,
1172
- 0.4017857015132904,
1173
- 0.4017857015132904
1174
- ],
1175
- [
1176
- 3.0,
1177
- 0.4017857015132904,
1178
- 0.4017857015132904
1179
- ],
1180
- [
1181
- 3.000000238418579,
1182
- 0.4017857015132904,
1183
- 0.4017857015132904
1184
- ],
1185
- [
1186
- 3.0,
1187
- 0.4017857015132904,
1188
- 0.4017857015132904
1189
- ],
1190
- [
1191
- 3.0,
1192
- 0.27173900604248047,
1193
- 0.27173900604248047
1194
- ],
1195
- [
1196
- 3.0000033378601074,
1197
- 0.27173900604248047,
1198
- 0.27173900604248047
1199
- ],
1200
- [
1201
- 2.9999983310699463,
1202
- 0.27173900604248047,
1203
- 0.27173900604248047
1204
- ],
1205
- [
1206
- 3.0,
1207
- 0.27173900604248047,
1208
- 0.27173900604248047
1209
- ],
1210
- [
1211
- 2.9999992847442627,
1212
- 0.46875,
1213
- 0.46875
1214
- ],
1215
- [
1216
- 3.0,
1217
- 0.4017857015132904,
1218
- 0.4017857015132904
1219
- ],
1220
- [
1221
- 3.0,
1222
- 0.4017857015132904,
1223
- 0.4017857015132904
1224
- ],
1225
- [
1226
- 2.9999990463256836,
1227
- 0.27173900604248047,
1228
- 0.27173900604248047
1229
- ],
1230
- [
1231
- 3.0,
1232
- 0.4017857015132904,
1233
- 0.4017857015132904
1234
- ],
1235
- [
1236
- 3.0,
1237
- 0.4017857015132904,
1238
- 0.4017857015132904
1239
- ],
1240
- [
1241
- 3.0,
1242
- 0.4017857015132904,
1243
- 0.4017857015132904
1244
- ],
1245
- [
1246
- 3.0,
1247
- 0.4017857015132904,
1248
- 0.4017857015132904
1249
- ],
1250
- [
1251
- 3.000020742416382,
1252
- 0.27173900604248047,
1253
- 0.27173900604248047
1254
- ],
1255
- [
1256
- 2.999999761581421,
1257
- 0.4017857015132904,
1258
- 0.4017857015132904
1259
- ],
1260
- [
1261
- 3.0,
1262
- 0.4017857015132904,
1263
- 0.4017857015132904
1264
- ],
1265
- [
1266
- 2.9999923706054688,
1267
- 0.27173900604248047,
1268
- 0.27173900604248047
1269
- ],
1270
- [
1271
- 3.0,
1272
- 0.4017857015132904,
1273
- 0.4017857015132904
1274
- ],
1275
- [
1276
- 3.0,
1277
- 0.4017857015132904,
1278
- 0.4017857015132904
1279
- ],
1280
- [
1281
- 3.0,
1282
- 0.4017857015132904,
1283
- 0.4017857015132904
1284
- ],
1285
- [
1286
- 3.0,
1287
- 0.4017857015132904,
1288
- 0.4017857015132904
1289
- ],
1290
- [
1291
- 3.0,
1292
- 0.4017857015132904,
1293
- 0.4017857015132904
1294
- ],
1295
- [
1296
- 3.0,
1297
- 0.4017857015132904,
1298
- 0.4017857015132904
1299
- ],
1300
- [
1301
- 3.0,
1302
- 0.4017857015132904,
1303
- 0.4017857015132904
1304
- ],
1305
- [
1306
- 3.0,
1307
- 0.4017857015132904,
1308
- 0.4017857015132904
1309
- ],
1310
- [
1311
- 2.999995470046997,
1312
- 0.27173900604248047,
1313
- 0.27173900604248047
1314
- ],
1315
- [
1316
- 3.0,
1317
- 0.4017857015132904,
1318
- 0.4017857015132904
1319
- ],
1320
- [
1321
- 3.0,
1322
- 0.4017857015132904,
1323
- 0.4017857015132904
1324
- ],
1325
- [
1326
- 3.0,
1327
- 0.4017857015132904,
1328
- 0.4017857015132904
1329
- ],
1330
- [
1331
- 3.0,
1332
- 0.4017857015132904,
1333
- 0.4017857015132904
1334
- ],
1335
- [
1336
- 2.999999761581421,
1337
- 0.4017857015132904,
1338
- 0.4017857015132904
1339
- ],
1340
- [
1341
- 3.0,
1342
- 0.4017857015132904,
1343
- 0.4017857015132904
1344
- ],
1345
- [
1346
- 3.0,
1347
- 0.4017857015132904,
1348
- 0.4017857015132904
1349
- ],
1350
- [
1351
- 3.0,
1352
- 0.4017857015132904,
1353
- 0.4017857015132904
1354
- ],
1355
- [
1356
- 2.999994993209839,
1357
- 0.46875,
1358
- 0.46875
1359
- ],
1360
- [
1361
- 3.0,
1362
- 0.4017857015132904,
1363
- 0.4017857015132904
1364
- ],
1365
- [
1366
- 3.0,
1367
- 0.4017857015132904,
1368
- 0.4017857015132904
1369
- ],
1370
- [
1371
- 3.0,
1372
- 0.27173900604248047,
1373
- 0.27173900604248047
1374
- ],
1375
- [
1376
- 2.9999983310699463,
1377
- 0.46875,
1378
- 0.46875
1379
- ],
1380
- [
1381
- 3.0,
1382
- 0.4017857015132904,
1383
- 0.4017857015132904
1384
- ],
1385
- [
1386
- 3.0,
1387
- 0.4017857015132904,
1388
- 0.4017857015132904
1389
- ],
1390
- [
1391
- 2.999985694885254,
1392
- 0.27173900604248047,
1393
- 0.27173900604248047
1394
- ],
1395
- [
1396
- 3.000000238418579,
1397
- 0.4017857015132904,
1398
- 0.4017857015132904
1399
- ],
1400
- [
1401
- 3.0,
1402
- 0.4017857015132904,
1403
- 0.4017857015132904
1404
- ],
1405
- [
1406
- 3.0,
1407
- 0.4017857015132904,
1408
- 0.4017857015132904
1409
- ],
1410
- [
1411
- 3.0,
1412
- 0.4464285671710968,
1413
- 0.4464285671710968
1414
- ],
1415
- [
1416
- 3.0,
1417
- 0.27173900604248047,
1418
- 0.27173900604248047
1419
- ],
1420
- [
1421
- 3.000000238418579,
1422
- 0.4017857015132904,
1423
- 0.4017857015132904
1424
- ]
1425
- ]
1426
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/.ipynb_checkpoints/progress-checkpoint.png DELETED

Git LFS Details

  • SHA256: acad0880086620effdb8b1b34d02327b5ed2419fd875e25ddf0521c7d45f6fd7
  • Pointer size: 131 Bytes
  • Size of remote file: 699 kB
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/checkpoint_best.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:00ea40d5a8879983fd37b31da67b30e4af0746331e43f9557f00db4d0d9c1720
3
- size 356831121
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/checkpoint_latest.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:32c4e68fcceeded035c49c3e0530a70cd6c3d96f942d0d1e5a997ea23e6fffdc
3
- size 356854997
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/debug.json DELETED
@@ -1,52 +0,0 @@
1
- {
2
- "_best_ema": "None",
3
- "batch_size": "2",
4
- "configuration_manager": "{'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [28, 256, 256], 'median_image_size_in_voxels': [25.0, 270.0, 270.0], 'spacing': [2.999998998641968, 0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [2, 6, 6], 'pool_op_kernel_sizes': [[1, 1, 1], [1, 2, 2], [1, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2], [1, 2, 2]], 'conv_kernel_sizes': [[1, 3, 3], [1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'unet_max_num_features': 320, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': False}",
5
- "configuration_name": "3d_fullres",
6
- "cudnn_version": 8500,
7
- "current_epoch": "0",
8
- "dataloader_train": "<nnunetv2.training.data_augmentation.custom_transforms.limited_length_multithreaded_augmenter.LimitedLenWrapper object at 0x7fe17c7ca020>",
9
- "dataloader_train.generator": "<nnunetv2.training.dataloading.data_loader_3d.nnUNetDataLoader3D object at 0x7fe17c7c8c40>",
10
- "dataloader_train.num_processes": "4",
11
- "dataloader_train.transform": "Compose ( [Convert3DTo2DTransform( apply_to_keys = ('data', 'seg') ), SpatialTransform( independent_scale_for_each_axis = False, p_rot_per_sample = 0.2, p_scale_per_sample = 0.2, p_el_per_sample = 0, data_key = 'data', label_key = 'seg', patch_size = [256, 256], patch_center_dist_from_border = None, do_elastic_deform = False, alpha = (0, 0), sigma = (0, 0), do_rotation = True, angle_x = (-3.141592653589793, 3.141592653589793), angle_y = (0, 0), angle_z = (0, 0), do_scale = True, scale = (0.7, 1.4), border_mode_data = 'constant', border_cval_data = 0, order_data = 3, border_mode_seg = 'constant', border_cval_seg = -1, order_seg = 1, random_crop = False, p_rot_per_axis = 1, p_independent_scale_per_axis = 1 ), Convert2DTo3DTransform( apply_to_keys = ('data', 'seg') ), GaussianNoiseTransform( p_per_sample = 0.1, data_key = 'data', noise_variance = (0, 0.1), p_per_channel = 1, per_channel = False ), GaussianBlurTransform( p_per_sample = 0.2, different_sigma_per_channel = True, p_per_channel = 0.5, data_key = 'data', blur_sigma = (0.5, 1.0), different_sigma_per_axis = False, p_isotropic = 0 ), BrightnessMultiplicativeTransform( p_per_sample = 0.15, data_key = 'data', multiplier_range = (0.75, 1.25), per_channel = True ), ContrastAugmentationTransform( p_per_sample = 0.15, data_key = 'data', contrast_range = (0.75, 1.25), preserve_range = True, per_channel = True, p_per_channel = 1 ), SimulateLowResolutionTransform( order_upsample = 3, order_downsample = 0, channels = None, per_channel = True, p_per_channel = 0.5, p_per_sample = 0.25, data_key = 'data', zoom_range = (0.5, 1), ignore_axes = (0,) ), GammaTransform( p_per_sample = 0.1, retain_stats = True, per_channel = True, data_key = 'data', gamma_range = (0.7, 1.5), invert_image = True ), GammaTransform( p_per_sample = 0.3, retain_stats = True, per_channel = True, data_key = 'data', gamma_range = (0.7, 1.5), invert_image = False ), MirrorTransform( p_per_sample = 1, data_key = 'data', label_key = 'seg', axes = (0, 1, 2) ), RemoveLabelTransform( output_key = 'seg', input_key = 'seg', replace_with = 0, remove_label = -1 ), RenameTransform( delete_old = True, out_key = 'target', in_key = 'seg' ), DownsampleSegForDSTransform2( axes = None, output_key = 'target', input_key = 'target', order = 0, ds_scales = [[1.0, 1.0, 1.0], [1.0, 0.5, 0.5], [1.0, 0.25, 0.25], [0.5, 0.125, 0.125], [0.25, 0.0625, 0.0625], [0.25, 0.03125, 0.03125]] ), NumpyToTensor( keys = ['data', 'target'], cast_to = 'float' )] )",
12
- "dataloader_val": "<nnunetv2.training.data_augmentation.custom_transforms.limited_length_multithreaded_augmenter.LimitedLenWrapper object at 0x7fe17c7c90f0>",
13
- "dataloader_val.generator": "<nnunetv2.training.dataloading.data_loader_3d.nnUNetDataLoader3D object at 0x7fe17c7c90c0>",
14
- "dataloader_val.num_processes": "2",
15
- "dataloader_val.transform": "Compose ( [RemoveLabelTransform( output_key = 'seg', input_key = 'seg', replace_with = 0, remove_label = -1 ), RenameTransform( delete_old = True, out_key = 'target', in_key = 'seg' ), DownsampleSegForDSTransform2( axes = None, output_key = 'target', input_key = 'target', order = 0, ds_scales = [[1.0, 1.0, 1.0], [1.0, 0.5, 0.5], [1.0, 0.25, 0.25], [0.5, 0.125, 0.125], [0.25, 0.0625, 0.0625], [0.25, 0.03125, 0.03125]] ), NumpyToTensor( keys = ['data', 'target'], cast_to = 'float' )] )",
16
- "dataset_json": "{'name': 'Prostate158', 'description': 'Prostate cancer segmentation dataset', 'channel_names': {'0': 'T2', '1': 'ADC', '2': 'DFI'}, 'labels': {'background': 0, 'prostate_inner': 1, 'prostate_outer': 2, 'tumor': 3}, 'numTraining': 139, 'numTest': 19, 'file_ending': '.nii.gz'}",
17
- "device": "cuda:0",
18
- "disable_checkpointing": "False",
19
- "fold": "0",
20
- "folder_with_segs_from_previous_stage": "None",
21
- "gpu_name": "NVIDIA A10G",
22
- "grad_scaler": "<torch.cuda.amp.grad_scaler.GradScaler object at 0x7fe17d646560>",
23
- "hostname": "s-osbm-jupyter-f0b83-8689bbb555-5t6kn",
24
- "inference_allowed_mirroring_axes": "(0, 1, 2)",
25
- "initial_lr": "0.01",
26
- "is_cascaded": "False",
27
- "is_ddp": "False",
28
- "label_manager": "<nnunetv2.utilities.label_handling.label_handling.LabelManager object at 0x7fe17d6464a0>",
29
- "local_rank": "0",
30
- "log_file": "nnUNet_results/Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/training_log_2023_7_24_00_01_52.txt",
31
- "logger": "<nnunetv2.training.logging.nnunet_logger.nnUNetLogger object at 0x7fe17d646410>",
32
- "loss": "DeepSupervisionWrapper(\n (loss): DC_and_CE_loss(\n (ce): RobustCrossEntropyLoss()\n (dc): MemoryEfficientSoftDiceLoss()\n )\n)",
33
- "lr_scheduler": "<nnunetv2.training.lr_scheduler.polylr.PolyLRScheduler object at 0x7fe17d646470>",
34
- "my_init_kwargs": "{'plans': {'dataset_name': 'Dataset001_Prostate158', 'plans_name': 'nnUNetPlans', 'original_median_spacing_after_transp': [3.0, 0.4017857015132904, 0.4017857015132904], 'original_median_shape_after_transp': [25, 270, 270], 'image_reader_writer': 'SimpleITKIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'configurations': {'2d': {'data_identifier': 'nnUNetPlans_2d', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 31, 'patch_size': [320, 320], 'median_image_size_in_voxels': [270.0, 270.0], 'spacing': [0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [6, 6], 'pool_op_kernel_sizes': [[1, 1], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2]], 'conv_kernel_sizes': [[3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3]], 'unet_max_num_features': 512, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': True}, '3d_fullres': {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [28, 256, 256], 'median_image_size_in_voxels': [25.0, 270.0, 270.0], 'spacing': [2.999998998641968, 0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [2, 6, 6], 'pool_op_kernel_sizes': [[1, 1, 1], [1, 2, 2], [1, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2], [1, 2, 2]], 'conv_kernel_sizes': [[1, 3, 3], [1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'unet_max_num_features': 320, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': False}}, 'experiment_planner_used': 'ExperimentPlanner', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 1694.0, 'mean': 267.35308837890625, 'median': 242.0, 'min': 0.0, 'percentile_00_5': 36.0, 'percentile_99_5': 768.0, 'std': 136.11251831054688}, '1': {'max': 3557.286865234375, 'mean': 1215.81591796875, 'median': 1203.8331298828125, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 2259.82861328125, 'std': 338.6748352050781}, '2': {'max': 198.95455932617188, 'mean': 72.26309204101562, 'median': 70.3214340209961, 'min': 0.0, 'percentile_00_5': 34.534385681152344, 'percentile_99_5': 132.71939086914062, 'std': 18.909290313720703}}}, 'configuration': '3d_fullres', 'fold': 0, 'dataset_json': {'name': 'Prostate158', 'description': 'Prostate cancer segmentation dataset', 'channel_names': {'0': 'T2', '1': 'ADC', '2': 'DFI'}, 'labels': {'background': 0, 'prostate_inner': 1, 'prostate_outer': 2, 'tumor': 3}, 'numTraining': 139, 'numTest': 19, 'file_ending': '.nii.gz'}, 'unpack_dataset': True, 'device': device(type='cuda')}",
35
- "network": "PlainConvUNet",
36
- "num_epochs": "1000",
37
- "num_input_channels": "3",
38
- "num_iterations_per_epoch": "250",
39
- "num_val_iterations_per_epoch": "50",
40
- "optimizer": "SGD (\nParameter Group 0\n dampening: 0\n differentiable: False\n foreach: None\n initial_lr: 0.01\n lr: 0.01\n maximize: False\n momentum: 0.99\n nesterov: True\n weight_decay: 3e-05\n)",
41
- "output_folder": "nnUNet_results/Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0",
42
- "output_folder_base": "nnUNet_results/Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres",
43
- "oversample_foreground_percent": "0.33",
44
- "plans_manager": "{'dataset_name': 'Dataset001_Prostate158', 'plans_name': 'nnUNetPlans', 'original_median_spacing_after_transp': [3.0, 0.4017857015132904, 0.4017857015132904], 'original_median_shape_after_transp': [25, 270, 270], 'image_reader_writer': 'SimpleITKIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'configurations': {'2d': {'data_identifier': 'nnUNetPlans_2d', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 31, 'patch_size': [320, 320], 'median_image_size_in_voxels': [270.0, 270.0], 'spacing': [0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [6, 6], 'pool_op_kernel_sizes': [[1, 1], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2]], 'conv_kernel_sizes': [[3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3]], 'unet_max_num_features': 512, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': True}, '3d_fullres': {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [28, 256, 256], 'median_image_size_in_voxels': [25.0, 270.0, 270.0], 'spacing': [2.999998998641968, 0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [2, 6, 6], 'pool_op_kernel_sizes': [[1, 1, 1], [1, 2, 2], [1, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2], [1, 2, 2]], 'conv_kernel_sizes': [[1, 3, 3], [1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'unet_max_num_features': 320, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': False}}, 'experiment_planner_used': 'ExperimentPlanner', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 1694.0, 'mean': 267.35308837890625, 'median': 242.0, 'min': 0.0, 'percentile_00_5': 36.0, 'percentile_99_5': 768.0, 'std': 136.11251831054688}, '1': {'max': 3557.286865234375, 'mean': 1215.81591796875, 'median': 1203.8331298828125, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 2259.82861328125, 'std': 338.6748352050781}, '2': {'max': 198.95455932617188, 'mean': 72.26309204101562, 'median': 70.3214340209961, 'min': 0.0, 'percentile_00_5': 34.534385681152344, 'percentile_99_5': 132.71939086914062, 'std': 18.909290313720703}}}",
45
- "preprocessed_dataset_folder": "nnUNet_preprocessed/Dataset001_Prostate158/nnUNetPlans_3d_fullres",
46
- "preprocessed_dataset_folder_base": "nnUNet_preprocessed/Dataset001_Prostate158",
47
- "save_every": "50",
48
- "torch_version": "2.0.1+cu117",
49
- "unpack_dataset": "True",
50
- "was_initialized": "True",
51
- "weight_decay": "3e-05"
52
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/progress.png DELETED

Git LFS Details

  • SHA256: 1853459deab216f0dccaee8913ac5be19856556711ddfb2d5734cd720aaf559c
  • Pointer size: 131 Bytes
  • Size of remote file: 701 kB
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_0/training_log_2023_7_24_00_01_52.txt DELETED
@@ -1,1194 +0,0 @@
1
-
2
- #######################################################################
3
- Please cite the following paper when using nnU-Net:
4
- Isensee, F., Jaeger, P. F., Kohl, S. A., Petersen, J., & Maier-Hein, K. H. (2021). nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nature methods, 18(2), 203-211.
5
- #######################################################################
6
-
7
-
8
- This is the configuration used by this training:
9
- Configuration name: 3d_fullres
10
- {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [28, 256, 256], 'median_image_size_in_voxels': [25.0, 270.0, 270.0], 'spacing': [2.999998998641968, 0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [2, 6, 6], 'pool_op_kernel_sizes': [[1, 1, 1], [1, 2, 2], [1, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2], [1, 2, 2]], 'conv_kernel_sizes': [[1, 3, 3], [1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'unet_max_num_features': 320, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': False}
11
-
12
- These are the global plan.json settings:
13
- {'dataset_name': 'Dataset001_Prostate158', 'plans_name': 'nnUNetPlans', 'original_median_spacing_after_transp': [3.0, 0.4017857015132904, 0.4017857015132904], 'original_median_shape_after_transp': [25, 270, 270], 'image_reader_writer': 'SimpleITKIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'experiment_planner_used': 'ExperimentPlanner', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 1694.0, 'mean': 267.35308837890625, 'median': 242.0, 'min': 0.0, 'percentile_00_5': 36.0, 'percentile_99_5': 768.0, 'std': 136.11251831054688}, '1': {'max': 3557.286865234375, 'mean': 1215.81591796875, 'median': 1203.8331298828125, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 2259.82861328125, 'std': 338.6748352050781}, '2': {'max': 198.95455932617188, 'mean': 72.26309204101562, 'median': 70.3214340209961, 'min': 0.0, 'percentile_00_5': 34.534385681152344, 'percentile_99_5': 132.71939086914062, 'std': 18.909290313720703}}}
14
-
15
- 2023-07-24 00:01:53.902022: unpacking dataset...
16
- 2023-07-24 00:02:07.139120: unpacking done...
17
- 2023-07-24 00:02:07.196625: do_dummy_2d_data_aug: True
18
- 2023-07-24 00:02:07.197410: Creating new 5-fold cross-validation split...
19
- 2023-07-24 00:02:07.198499: Desired fold for training: 0
20
- 2023-07-24 00:02:07.198552: This split has 111 training and 28 validation cases.
21
- 2023-07-24 00:02:14.074938: Unable to plot network architecture:
22
- 2023-07-24 00:02:14.075182: module 'torch.onnx' has no attribute '_optimize_trace'
23
- 2023-07-24 00:02:14.123964:
24
- 2023-07-24 00:02:14.124048: Epoch 0
25
- 2023-07-24 00:02:14.124163: Current learning rate: 0.01
26
- 2023-07-24 00:06:16.568555: train_loss -0.0413
27
- 2023-07-24 00:06:16.568763: val_loss -0.1561
28
- 2023-07-24 00:06:16.568854: Pseudo dice [0.6633, 0.3715, 0.0]
29
- 2023-07-24 00:06:16.568946: Epoch time: 242.45 s
30
- 2023-07-24 00:06:16.569015: Yayy! New best EMA pseudo Dice: 0.3449
31
- 2023-07-24 00:06:19.487634:
32
- 2023-07-24 00:06:19.487758: Epoch 1
33
- 2023-07-24 00:06:19.487858: Current learning rate: 0.00999
34
- 2023-07-24 00:09:51.638274: train_loss -0.2546
35
- 2023-07-24 00:09:51.638539: val_loss -0.3228
36
- 2023-07-24 00:09:51.638626: Pseudo dice [0.7605, 0.5865, 0.0]
37
- 2023-07-24 00:09:51.638798: Epoch time: 212.15 s
38
- 2023-07-24 00:09:51.638938: Yayy! New best EMA pseudo Dice: 0.3553
39
- 2023-07-24 00:09:54.319529:
40
- 2023-07-24 00:09:54.319663: Epoch 2
41
- 2023-07-24 00:09:54.319769: Current learning rate: 0.00998
42
- 2023-07-24 00:13:25.194643: train_loss -0.324
43
- 2023-07-24 00:13:25.194925: val_loss -0.3233
44
- 2023-07-24 00:13:25.195090: Pseudo dice [0.7874, 0.5763, 0.0]
45
- 2023-07-24 00:13:25.195179: Epoch time: 210.88 s
46
- 2023-07-24 00:13:25.195318: Yayy! New best EMA pseudo Dice: 0.3652
47
- 2023-07-24 00:13:27.399833:
48
- 2023-07-24 00:13:27.399949: Epoch 3
49
- 2023-07-24 00:13:27.400062: Current learning rate: 0.00997
50
- 2023-07-24 00:17:09.770445: train_loss -0.3629
51
- 2023-07-24 00:17:09.770754: val_loss -0.3386
52
- 2023-07-24 00:17:09.771257: Pseudo dice [0.7865, 0.6213, 0.0]
53
- 2023-07-24 00:17:09.771582: Epoch time: 222.37 s
54
- 2023-07-24 00:17:09.771953: Yayy! New best EMA pseudo Dice: 0.3756
55
- 2023-07-24 00:17:13.443594:
56
- 2023-07-24 00:17:13.443887: Epoch 4
57
- 2023-07-24 00:17:13.444007: Current learning rate: 0.00996
58
- 2023-07-24 00:20:25.372916: train_loss -0.4038
59
- 2023-07-24 00:20:25.373092: val_loss -0.4029
60
- 2023-07-24 00:20:25.373185: Pseudo dice [0.821, 0.6488, 0.2694]
61
- 2023-07-24 00:20:25.373266: Epoch time: 191.93 s
62
- 2023-07-24 00:20:25.373332: Yayy! New best EMA pseudo Dice: 0.3961
63
- 2023-07-24 00:20:27.527769:
64
- 2023-07-24 00:20:27.527886: Epoch 5
65
- 2023-07-24 00:20:27.527997: Current learning rate: 0.00995
66
- 2023-07-24 00:23:59.957678: train_loss -0.4395
67
- 2023-07-24 00:23:59.957844: val_loss -0.4287
68
- 2023-07-24 00:23:59.957929: Pseudo dice [0.8103, 0.634, 0.4383]
69
- 2023-07-24 00:23:59.958011: Epoch time: 212.43 s
70
- 2023-07-24 00:23:59.958077: Yayy! New best EMA pseudo Dice: 0.4192
71
- 2023-07-24 00:24:01.987802:
72
- 2023-07-24 00:24:01.987912: Epoch 6
73
- 2023-07-24 00:24:01.988024: Current learning rate: 0.00995
74
- 2023-07-24 00:27:44.002411: train_loss -0.4486
75
- 2023-07-24 00:27:44.002615: val_loss -0.4226
76
- 2023-07-24 00:27:44.002702: Pseudo dice [0.8368, 0.6463, 0.3067]
77
- 2023-07-24 00:27:44.002805: Epoch time: 222.02 s
78
- 2023-07-24 00:27:44.002879: Yayy! New best EMA pseudo Dice: 0.4369
79
- 2023-07-24 00:27:46.502084:
80
- 2023-07-24 00:27:46.502206: Epoch 7
81
- 2023-07-24 00:27:46.502326: Current learning rate: 0.00994
82
- 2023-07-24 00:31:13.044165: train_loss -0.4879
83
- 2023-07-24 00:31:13.044357: val_loss -0.4522
84
- 2023-07-24 00:31:13.044453: Pseudo dice [0.8328, 0.6665, 0.389]
85
- 2023-07-24 00:31:13.044543: Epoch time: 206.54 s
86
- 2023-07-24 00:31:13.050315: Yayy! New best EMA pseudo Dice: 0.4562
87
- 2023-07-24 00:31:15.979445:
88
- 2023-07-24 00:31:15.979643: Epoch 8
89
- 2023-07-24 00:31:15.979780: Current learning rate: 0.00993
90
- 2023-07-24 00:34:55.354722: train_loss -0.5049
91
- 2023-07-24 00:34:55.355004: val_loss -0.4647
92
- 2023-07-24 00:34:55.355092: Pseudo dice [0.8501, 0.6851, 0.4456]
93
- 2023-07-24 00:34:55.355233: Epoch time: 219.38 s
94
- 2023-07-24 00:34:55.355299: Yayy! New best EMA pseudo Dice: 0.4766
95
- 2023-07-24 00:34:57.734633:
96
- 2023-07-24 00:34:57.734877: Epoch 9
97
- 2023-07-24 00:34:57.734988: Current learning rate: 0.00992
98
- 2023-07-24 00:38:36.038373: train_loss -0.5101
99
- 2023-07-24 00:38:36.038563: val_loss -0.4518
100
- 2023-07-24 00:38:36.038657: Pseudo dice [0.8332, 0.6664, 0.4652]
101
- 2023-07-24 00:38:36.038745: Epoch time: 218.3 s
102
- 2023-07-24 00:38:36.038840: Yayy! New best EMA pseudo Dice: 0.4944
103
- 2023-07-24 00:38:39.940725:
104
- 2023-07-24 00:38:39.941038: Epoch 10
105
- 2023-07-24 00:38:39.941158: Current learning rate: 0.00991
106
- 2023-07-24 00:42:04.394748: train_loss -0.5153
107
- 2023-07-24 00:42:04.398857: val_loss -0.4822
108
- 2023-07-24 00:42:04.399046: Pseudo dice [0.8432, 0.6794, 0.4922]
109
- 2023-07-24 00:42:04.399148: Epoch time: 204.46 s
110
- 2023-07-24 00:42:04.399216: Yayy! New best EMA pseudo Dice: 0.5121
111
- 2023-07-24 00:42:08.116886:
112
- 2023-07-24 00:42:08.117016: Epoch 11
113
- 2023-07-24 00:42:08.117131: Current learning rate: 0.0099
114
- 2023-07-24 00:45:37.933030: train_loss -0.5128
115
- 2023-07-24 00:45:37.933209: val_loss -0.4819
116
- 2023-07-24 00:45:37.933295: Pseudo dice [0.8511, 0.6729, 0.4875]
117
- 2023-07-24 00:45:37.933378: Epoch time: 209.82 s
118
- 2023-07-24 00:45:37.933442: Yayy! New best EMA pseudo Dice: 0.528
119
- 2023-07-24 00:45:41.223381:
120
- 2023-07-24 00:45:41.223522: Epoch 12
121
- 2023-07-24 00:45:41.223654: Current learning rate: 0.00989
122
- 2023-07-24 00:49:18.457412: train_loss -0.5139
123
- 2023-07-24 00:49:18.457591: val_loss -0.4663
124
- 2023-07-24 00:49:18.457679: Pseudo dice [0.8304, 0.6664, 0.4751]
125
- 2023-07-24 00:49:18.457762: Epoch time: 217.24 s
126
- 2023-07-24 00:49:18.457826: Yayy! New best EMA pseudo Dice: 0.5409
127
- 2023-07-24 00:49:21.452374:
128
- 2023-07-24 00:49:21.452548: Epoch 13
129
- 2023-07-24 00:49:21.452664: Current learning rate: 0.00988
130
- 2023-07-24 00:52:53.617019: train_loss -0.5306
131
- 2023-07-24 00:52:53.617260: val_loss -0.4947
132
- 2023-07-24 00:52:53.617345: Pseudo dice [0.8434, 0.6944, 0.4883]
133
- 2023-07-24 00:52:53.617487: Epoch time: 212.17 s
134
- 2023-07-24 00:52:53.617552: Yayy! New best EMA pseudo Dice: 0.5544
135
- 2023-07-24 00:52:55.726403:
136
- 2023-07-24 00:52:55.726518: Epoch 14
137
- 2023-07-24 00:52:55.726615: Current learning rate: 0.00987
138
- 2023-07-24 00:56:22.506680: train_loss -0.5446
139
- 2023-07-24 00:56:22.506907: val_loss -0.4714
140
- 2023-07-24 00:56:22.507013: Pseudo dice [0.8411, 0.671, 0.4914]
141
- 2023-07-24 00:56:22.507121: Epoch time: 206.78 s
142
- 2023-07-24 00:56:22.507219: Yayy! New best EMA pseudo Dice: 0.5657
143
- 2023-07-24 00:56:24.667025:
144
- 2023-07-24 00:56:24.667140: Epoch 15
145
- 2023-07-24 00:56:24.667252: Current learning rate: 0.00986
146
- 2023-07-24 00:59:50.747144: train_loss -0.5394
147
- 2023-07-24 00:59:50.747421: val_loss -0.4625
148
- 2023-07-24 00:59:50.747515: Pseudo dice [0.8447, 0.654, 0.5712]
149
- 2023-07-24 00:59:50.747690: Epoch time: 206.08 s
150
- 2023-07-24 00:59:50.747818: Yayy! New best EMA pseudo Dice: 0.5781
151
- 2023-07-24 00:59:53.732963:
152
- 2023-07-24 00:59:53.733104: Epoch 16
153
- 2023-07-24 00:59:53.733224: Current learning rate: 0.00986
154
- 2023-07-24 01:03:24.895722: train_loss -0.5551
155
- 2023-07-24 01:03:24.895922: val_loss -0.496
156
- 2023-07-24 01:03:24.896017: Pseudo dice [0.8705, 0.6946, 0.4975]
157
- 2023-07-24 01:03:24.896102: Epoch time: 211.16 s
158
- 2023-07-24 01:03:24.896200: Yayy! New best EMA pseudo Dice: 0.5891
159
- 2023-07-24 01:03:27.655639:
160
- 2023-07-24 01:03:27.655761: Epoch 17
161
- 2023-07-24 01:03:27.655877: Current learning rate: 0.00985
162
- 2023-07-24 01:07:09.366258: train_loss -0.5496
163
- 2023-07-24 01:07:09.371549: val_loss -0.4745
164
- 2023-07-24 01:07:09.371795: Pseudo dice [0.8679, 0.6897, 0.514]
165
- 2023-07-24 01:07:09.371953: Epoch time: 221.71 s
166
- 2023-07-24 01:07:09.372037: Yayy! New best EMA pseudo Dice: 0.5992
167
- 2023-07-24 01:07:12.715889:
168
- 2023-07-24 01:07:12.716015: Epoch 18
169
- 2023-07-24 01:07:12.716132: Current learning rate: 0.00984
170
- 2023-07-24 01:10:59.095254: train_loss -0.5406
171
- 2023-07-24 01:10:59.095483: val_loss -0.483
172
- 2023-07-24 01:10:59.095583: Pseudo dice [0.8634, 0.6795, 0.5385]
173
- 2023-07-24 01:10:59.095723: Epoch time: 226.38 s
174
- 2023-07-24 01:10:59.095788: Yayy! New best EMA pseudo Dice: 0.6087
175
- 2023-07-24 01:11:01.356395:
176
- 2023-07-24 01:11:01.356525: Epoch 19
177
- 2023-07-24 01:11:01.356640: Current learning rate: 0.00983
178
- 2023-07-24 01:14:19.901330: train_loss -0.5567
179
- 2023-07-24 01:14:19.901525: val_loss -0.4783
180
- 2023-07-24 01:14:19.901613: Pseudo dice [0.8492, 0.6553, 0.535]
181
- 2023-07-24 01:14:19.901693: Epoch time: 198.55 s
182
- 2023-07-24 01:14:19.901747: Yayy! New best EMA pseudo Dice: 0.6158
183
- 2023-07-24 01:14:22.511663:
184
- 2023-07-24 01:14:22.511957: Epoch 20
185
- 2023-07-24 01:14:22.512064: Current learning rate: 0.00982
186
- 2023-07-24 01:17:55.925512: train_loss -0.5745
187
- 2023-07-24 01:17:55.925694: val_loss -0.4715
188
- 2023-07-24 01:17:55.925791: Pseudo dice [0.8548, 0.653, 0.5559]
189
- 2023-07-24 01:17:55.925879: Epoch time: 213.41 s
190
- 2023-07-24 01:17:55.925951: Yayy! New best EMA pseudo Dice: 0.623
191
- 2023-07-24 01:17:58.448946:
192
- 2023-07-24 01:17:58.449071: Epoch 21
193
- 2023-07-24 01:17:58.449180: Current learning rate: 0.00981
194
- 2023-07-24 01:21:29.617508: train_loss -0.5761
195
- 2023-07-24 01:21:29.617684: val_loss -0.4919
196
- 2023-07-24 01:21:29.617774: Pseudo dice [0.8628, 0.7072, 0.4682]
197
- 2023-07-24 01:21:29.617859: Epoch time: 211.17 s
198
- 2023-07-24 01:21:29.617927: Yayy! New best EMA pseudo Dice: 0.6286
199
- 2023-07-24 01:21:32.232873:
200
- 2023-07-24 01:21:32.232998: Epoch 22
201
- 2023-07-24 01:21:32.233115: Current learning rate: 0.0098
202
- 2023-07-24 01:24:54.282193: train_loss -0.5831
203
- 2023-07-24 01:24:54.282391: val_loss -0.5023
204
- 2023-07-24 01:24:54.282501: Pseudo dice [0.8587, 0.6942, 0.4825]
205
- 2023-07-24 01:24:54.282606: Epoch time: 202.05 s
206
- 2023-07-24 01:24:54.282692: Yayy! New best EMA pseudo Dice: 0.6336
207
- 2023-07-24 01:24:57.618637:
208
- 2023-07-24 01:24:57.618829: Epoch 23
209
- 2023-07-24 01:24:57.618946: Current learning rate: 0.00979
210
- 2023-07-24 01:28:25.124920: train_loss -0.5864
211
- 2023-07-24 01:28:25.125104: val_loss -0.5029
212
- 2023-07-24 01:28:25.125192: Pseudo dice [0.8623, 0.7014, 0.5049]
213
- 2023-07-24 01:28:25.125279: Epoch time: 207.51 s
214
- 2023-07-24 01:28:25.125346: Yayy! New best EMA pseudo Dice: 0.6392
215
- 2023-07-24 01:28:29.015366:
216
- 2023-07-24 01:28:29.015487: Epoch 24
217
- 2023-07-24 01:28:29.015594: Current learning rate: 0.00978
218
- 2023-07-24 01:32:03.197192: train_loss -0.5843
219
- 2023-07-24 01:32:03.197452: val_loss -0.478
220
- 2023-07-24 01:32:03.197542: Pseudo dice [0.8586, 0.6702, 0.5603]
221
- 2023-07-24 01:32:03.197691: Epoch time: 214.18 s
222
- 2023-07-24 01:32:03.197757: Yayy! New best EMA pseudo Dice: 0.6449
223
- 2023-07-24 01:32:05.726707:
224
- 2023-07-24 01:32:05.726868: Epoch 25
225
- 2023-07-24 01:32:05.726998: Current learning rate: 0.00977
226
- 2023-07-24 01:35:41.851906: train_loss -0.5882
227
- 2023-07-24 01:35:41.852082: val_loss -0.4719
228
- 2023-07-24 01:35:41.852175: Pseudo dice [0.862, 0.6743, 0.3966]
229
- 2023-07-24 01:35:41.852259: Epoch time: 216.13 s
230
- 2023-07-24 01:35:43.534244:
231
- 2023-07-24 01:35:43.534373: Epoch 26
232
- 2023-07-24 01:35:43.534485: Current learning rate: 0.00977
233
- 2023-07-24 01:39:02.738217: train_loss -0.6018
234
- 2023-07-24 01:39:02.743619: val_loss -0.4994
235
- 2023-07-24 01:39:02.743780: Pseudo dice [0.8766, 0.733, 0.4566]
236
- 2023-07-24 01:39:02.743932: Epoch time: 199.21 s
237
- 2023-07-24 01:39:02.744023: Yayy! New best EMA pseudo Dice: 0.6493
238
- 2023-07-24 01:39:05.142646:
239
- 2023-07-24 01:39:05.142778: Epoch 27
240
- 2023-07-24 01:39:05.142897: Current learning rate: 0.00976
241
- 2023-07-24 01:42:29.737434: train_loss -0.5995
242
- 2023-07-24 01:42:29.737619: val_loss -0.4867
243
- 2023-07-24 01:42:29.737704: Pseudo dice [0.8689, 0.689, 0.4957]
244
- 2023-07-24 01:42:29.737784: Epoch time: 204.6 s
245
- 2023-07-24 01:42:29.737848: Yayy! New best EMA pseudo Dice: 0.6528
246
- 2023-07-24 01:42:32.621387:
247
- 2023-07-24 01:42:32.621503: Epoch 28
248
- 2023-07-24 01:42:32.621617: Current learning rate: 0.00975
249
- 2023-07-24 01:46:07.936492: train_loss -0.5998
250
- 2023-07-24 01:46:07.936731: val_loss -0.4998
251
- 2023-07-24 01:46:07.943375: Pseudo dice [0.8649, 0.6936, 0.5489]
252
- 2023-07-24 01:46:07.943589: Epoch time: 215.32 s
253
- 2023-07-24 01:46:07.943666: Yayy! New best EMA pseudo Dice: 0.6578
254
- 2023-07-24 01:46:12.240499:
255
- 2023-07-24 01:46:12.240789: Epoch 29
256
- 2023-07-24 01:46:12.240907: Current learning rate: 0.00974
257
- 2023-07-24 01:49:49.765490: train_loss -0.6192
258
- 2023-07-24 01:49:49.765681: val_loss -0.4884
259
- 2023-07-24 01:49:49.765781: Pseudo dice [0.8677, 0.6745, 0.5335]
260
- 2023-07-24 01:49:49.765875: Epoch time: 217.53 s
261
- 2023-07-24 01:49:49.765948: Yayy! New best EMA pseudo Dice: 0.6612
262
- 2023-07-24 01:49:54.074142:
263
- 2023-07-24 01:49:54.074317: Epoch 30
264
- 2023-07-24 01:49:54.074434: Current learning rate: 0.00973
265
- 2023-07-24 01:53:33.757728: train_loss -0.6115
266
- 2023-07-24 01:53:33.758031: val_loss -0.4838
267
- 2023-07-24 01:53:33.758127: Pseudo dice [0.8786, 0.6666, 0.5728]
268
- 2023-07-24 01:53:33.758209: Epoch time: 219.69 s
269
- 2023-07-24 01:53:33.758281: Yayy! New best EMA pseudo Dice: 0.6656
270
- 2023-07-24 01:53:37.008034:
271
- 2023-07-24 01:53:37.008158: Epoch 31
272
- 2023-07-24 01:53:37.008272: Current learning rate: 0.00972
273
- 2023-07-24 01:57:03.417002: train_loss -0.625
274
- 2023-07-24 01:57:03.417286: val_loss -0.5044
275
- 2023-07-24 01:57:03.417374: Pseudo dice [0.8686, 0.7071, 0.5431]
276
- 2023-07-24 01:57:03.417525: Epoch time: 206.41 s
277
- 2023-07-24 01:57:03.417588: Yayy! New best EMA pseudo Dice: 0.6697
278
- 2023-07-24 01:57:05.575807:
279
- 2023-07-24 01:57:05.575927: Epoch 32
280
- 2023-07-24 01:57:05.576024: Current learning rate: 0.00971
281
- 2023-07-24 02:00:42.226546: train_loss -0.6184
282
- 2023-07-24 02:00:42.232471: val_loss -0.4894
283
- 2023-07-24 02:00:42.232700: Pseudo dice [0.8702, 0.7033, 0.4937]
284
- 2023-07-24 02:00:42.232790: Epoch time: 216.65 s
285
- 2023-07-24 02:00:42.232924: Yayy! New best EMA pseudo Dice: 0.6716
286
- 2023-07-24 02:00:45.014342:
287
- 2023-07-24 02:00:45.014468: Epoch 33
288
- 2023-07-24 02:00:45.014586: Current learning rate: 0.0097
289
- 2023-07-24 02:04:19.757532: train_loss -0.6217
290
- 2023-07-24 02:04:19.757771: val_loss -0.4804
291
- 2023-07-24 02:04:19.757856: Pseudo dice [0.8748, 0.693, 0.5362]
292
- 2023-07-24 02:04:19.758005: Epoch time: 214.74 s
293
- 2023-07-24 02:04:19.758069: Yayy! New best EMA pseudo Dice: 0.6746
294
- 2023-07-24 02:04:24.430918:
295
- 2023-07-24 02:04:24.431120: Epoch 34
296
- 2023-07-24 02:04:24.431232: Current learning rate: 0.00969
297
- 2023-07-24 02:08:00.604949: train_loss -0.6087
298
- 2023-07-24 02:08:00.605173: val_loss -0.4999
299
- 2023-07-24 02:08:00.605262: Pseudo dice [0.8614, 0.7024, 0.5684]
300
- 2023-07-24 02:08:00.605341: Epoch time: 216.18 s
301
- 2023-07-24 02:08:00.605400: Yayy! New best EMA pseudo Dice: 0.6782
302
- 2023-07-24 02:08:05.228923:
303
- 2023-07-24 02:08:05.229156: Epoch 35
304
- 2023-07-24 02:08:05.229259: Current learning rate: 0.00968
305
- 2023-07-24 02:11:46.128471: train_loss -0.6278
306
- 2023-07-24 02:11:46.128723: val_loss -0.4975
307
- 2023-07-24 02:11:46.128807: Pseudo dice [0.8697, 0.6883, 0.5635]
308
- 2023-07-24 02:11:46.128886: Epoch time: 220.9 s
309
- 2023-07-24 02:11:46.128947: Yayy! New best EMA pseudo Dice: 0.6811
310
- 2023-07-24 02:11:48.352022:
311
- 2023-07-24 02:11:48.352140: Epoch 36
312
- 2023-07-24 02:11:48.352254: Current learning rate: 0.00968
313
- 2023-07-24 02:15:18.130978: train_loss -0.6341
314
- 2023-07-24 02:15:18.131156: val_loss -0.514
315
- 2023-07-24 02:15:18.131239: Pseudo dice [0.8717, 0.7153, 0.5774]
316
- 2023-07-24 02:15:18.131315: Epoch time: 209.78 s
317
- 2023-07-24 02:15:18.131555: Yayy! New best EMA pseudo Dice: 0.6852
318
- 2023-07-24 02:15:20.282215:
319
- 2023-07-24 02:15:20.282343: Epoch 37
320
- 2023-07-24 02:15:20.282458: Current learning rate: 0.00967
321
- 2023-07-24 02:18:53.374968: train_loss -0.6275
322
- 2023-07-24 02:18:53.375212: val_loss -0.5119
323
- 2023-07-24 02:18:53.375304: Pseudo dice [0.867, 0.6981, 0.5894]
324
- 2023-07-24 02:18:53.375387: Epoch time: 213.09 s
325
- 2023-07-24 02:18:53.375454: Yayy! New best EMA pseudo Dice: 0.6885
326
- 2023-07-24 02:18:56.726088:
327
- 2023-07-24 02:18:56.726211: Epoch 38
328
- 2023-07-24 02:18:56.726327: Current learning rate: 0.00966
329
- 2023-07-24 02:22:22.025652: train_loss -0.6328
330
- 2023-07-24 02:22:22.031740: val_loss -0.503
331
- 2023-07-24 02:22:22.032003: Pseudo dice [0.8649, 0.7161, 0.5203]
332
- 2023-07-24 02:22:22.032098: Epoch time: 205.3 s
333
- 2023-07-24 02:22:22.032236: Yayy! New best EMA pseudo Dice: 0.6897
334
- 2023-07-24 02:22:25.923321:
335
- 2023-07-24 02:22:25.923668: Epoch 39
336
- 2023-07-24 02:22:25.923775: Current learning rate: 0.00965
337
- 2023-07-24 02:25:55.050353: train_loss -0.6391
338
- 2023-07-24 02:25:55.050546: val_loss -0.5015
339
- 2023-07-24 02:25:55.050636: Pseudo dice [0.8618, 0.6972, 0.566]
340
- 2023-07-24 02:25:55.050725: Epoch time: 209.13 s
341
- 2023-07-24 02:25:55.050815: Yayy! New best EMA pseudo Dice: 0.6915
342
- 2023-07-24 02:25:57.131437:
343
- 2023-07-24 02:25:57.131554: Epoch 40
344
- 2023-07-24 02:25:57.131677: Current learning rate: 0.00964
345
- 2023-07-24 02:29:24.101611: train_loss -0.6453
346
- 2023-07-24 02:29:24.101807: val_loss -0.4857
347
- 2023-07-24 02:29:24.101905: Pseudo dice [0.8591, 0.6952, 0.5601]
348
- 2023-07-24 02:29:24.102006: Epoch time: 206.97 s
349
- 2023-07-24 02:29:24.102089: Yayy! New best EMA pseudo Dice: 0.6929
350
- 2023-07-24 02:29:26.722635:
351
- 2023-07-24 02:29:26.722777: Epoch 41
352
- 2023-07-24 02:29:26.722900: Current learning rate: 0.00963
353
- 2023-07-24 02:33:01.503387: train_loss -0.6384
354
- 2023-07-24 02:33:01.503580: val_loss -0.495
355
- 2023-07-24 02:33:01.503671: Pseudo dice [0.8646, 0.7111, 0.4851]
356
- 2023-07-24 02:33:01.503757: Epoch time: 214.78 s
357
- 2023-07-24 02:33:03.298025:
358
- 2023-07-24 02:33:03.298147: Epoch 42
359
- 2023-07-24 02:33:03.298252: Current learning rate: 0.00962
360
- 2023-07-24 02:36:28.531214: train_loss -0.6454
361
- 2023-07-24 02:36:28.531463: val_loss -0.4999
362
- 2023-07-24 02:36:28.531549: Pseudo dice [0.8693, 0.691, 0.6198]
363
- 2023-07-24 02:36:28.531719: Epoch time: 205.23 s
364
- 2023-07-24 02:36:28.531906: Yayy! New best EMA pseudo Dice: 0.6957
365
- 2023-07-24 02:36:30.675420:
366
- 2023-07-24 02:36:30.675563: Epoch 43
367
- 2023-07-24 02:36:30.675693: Current learning rate: 0.00961
368
- 2023-07-24 02:39:57.254115: train_loss -0.6449
369
- 2023-07-24 02:39:57.254383: val_loss -0.5029
370
- 2023-07-24 02:39:57.254475: Pseudo dice [0.8713, 0.6894, 0.6393]
371
- 2023-07-24 02:39:57.254557: Epoch time: 206.58 s
372
- 2023-07-24 02:39:57.254622: Yayy! New best EMA pseudo Dice: 0.6995
373
- 2023-07-24 02:39:59.978240:
374
- 2023-07-24 02:39:59.978516: Epoch 44
375
- 2023-07-24 02:39:59.978635: Current learning rate: 0.0096
376
- 2023-07-24 02:43:33.330351: train_loss -0.6418
377
- 2023-07-24 02:43:33.330585: val_loss -0.4937
378
- 2023-07-24 02:43:33.330672: Pseudo dice [0.8727, 0.6956, 0.5291]
379
- 2023-07-24 02:43:33.330842: Epoch time: 213.35 s
380
- 2023-07-24 02:43:35.411677:
381
- 2023-07-24 02:43:35.411899: Epoch 45
382
- 2023-07-24 02:43:35.412007: Current learning rate: 0.00959
383
- 2023-07-24 02:47:16.638136: train_loss -0.6484
384
- 2023-07-24 02:47:16.638322: val_loss -0.5065
385
- 2023-07-24 02:47:16.638409: Pseudo dice [0.8781, 0.7181, 0.5566]
386
- 2023-07-24 02:47:16.638503: Epoch time: 221.23 s
387
- 2023-07-24 02:47:16.638575: Yayy! New best EMA pseudo Dice: 0.7013
388
- 2023-07-24 02:47:18.875849:
389
- 2023-07-24 02:47:18.876126: Epoch 46
390
- 2023-07-24 02:47:18.876246: Current learning rate: 0.00959
391
- 2023-07-24 02:50:43.664102: train_loss -0.6594
392
- 2023-07-24 02:50:43.664368: val_loss -0.5021
393
- 2023-07-24 02:50:43.664471: Pseudo dice [0.8832, 0.7081, 0.5237]
394
- 2023-07-24 02:50:43.664641: Epoch time: 204.79 s
395
- 2023-07-24 02:50:43.664771: Yayy! New best EMA pseudo Dice: 0.7016
396
- 2023-07-24 02:50:45.913680:
397
- 2023-07-24 02:50:45.913988: Epoch 47
398
- 2023-07-24 02:50:45.914099: Current learning rate: 0.00958
399
- 2023-07-24 02:54:09.407397: train_loss -0.6559
400
- 2023-07-24 02:54:09.407646: val_loss -0.4876
401
- 2023-07-24 02:54:09.407736: Pseudo dice [0.8737, 0.6605, 0.6181]
402
- 2023-07-24 02:54:09.407889: Epoch time: 203.49 s
403
- 2023-07-24 02:54:09.407957: Yayy! New best EMA pseudo Dice: 0.7032
404
- 2023-07-24 02:54:12.936934:
405
- 2023-07-24 02:54:12.937068: Epoch 48
406
- 2023-07-24 02:54:12.937182: Current learning rate: 0.00957
407
- 2023-07-24 02:57:47.539666: train_loss -0.6583
408
- 2023-07-24 02:57:47.539865: val_loss -0.5048
409
- 2023-07-24 02:57:47.540035: Pseudo dice [0.8792, 0.7214, 0.4783]
410
- 2023-07-24 02:57:47.540195: Epoch time: 214.6 s
411
- 2023-07-24 02:57:49.232162:
412
- 2023-07-24 02:57:49.232288: Epoch 49
413
- 2023-07-24 02:57:49.232400: Current learning rate: 0.00956
414
- 2023-07-24 03:01:31.963185: train_loss -0.6562
415
- 2023-07-24 03:01:31.963373: val_loss -0.5071
416
- 2023-07-24 03:01:31.963463: Pseudo dice [0.8811, 0.7189, 0.5632]
417
- 2023-07-24 03:01:31.963546: Epoch time: 222.73 s
418
- 2023-07-24 03:01:32.441789: Yayy! New best EMA pseudo Dice: 0.7041
419
- 2023-07-24 03:01:34.494628:
420
- 2023-07-24 03:01:34.494856: Epoch 50
421
- 2023-07-24 03:01:34.494962: Current learning rate: 0.00955
422
- 2023-07-24 03:05:00.456790: train_loss -0.66
423
- 2023-07-24 03:05:00.457008: val_loss -0.4885
424
- 2023-07-24 03:05:00.457094: Pseudo dice [0.8722, 0.7199, 0.3882]
425
- 2023-07-24 03:05:00.457232: Epoch time: 205.96 s
426
- 2023-07-24 03:05:01.920852:
427
- 2023-07-24 03:05:01.921130: Epoch 51
428
- 2023-07-24 03:05:01.921252: Current learning rate: 0.00954
429
- 2023-07-24 03:08:21.863105: train_loss -0.6615
430
- 2023-07-24 03:08:21.863394: val_loss -0.5063
431
- 2023-07-24 03:08:21.863482: Pseudo dice [0.8652, 0.6923, 0.5877]
432
- 2023-07-24 03:08:21.863767: Epoch time: 199.94 s
433
- 2023-07-24 03:08:23.565506:
434
- 2023-07-24 03:08:23.565624: Epoch 52
435
- 2023-07-24 03:08:23.565736: Current learning rate: 0.00953
436
- 2023-07-24 03:11:58.618505: train_loss -0.6541
437
- 2023-07-24 03:11:58.618745: val_loss -0.5208
438
- 2023-07-24 03:11:58.618862: Pseudo dice [0.8769, 0.7111, 0.5391]
439
- 2023-07-24 03:11:58.619015: Epoch time: 215.05 s
440
- 2023-07-24 03:12:00.422971:
441
- 2023-07-24 03:12:00.423093: Epoch 53
442
- 2023-07-24 03:12:00.423211: Current learning rate: 0.00952
443
- 2023-07-24 03:15:34.016732: train_loss -0.6717
444
- 2023-07-24 03:15:34.016971: val_loss -0.4904
445
- 2023-07-24 03:15:34.017070: Pseudo dice [0.8808, 0.7097, 0.4256]
446
- 2023-07-24 03:15:34.017220: Epoch time: 213.59 s
447
- 2023-07-24 03:15:36.404579:
448
- 2023-07-24 03:15:36.404704: Epoch 54
449
- 2023-07-24 03:15:36.404819: Current learning rate: 0.00951
450
- 2023-07-24 03:19:13.253185: train_loss -0.677
451
- 2023-07-24 03:19:13.253387: val_loss -0.4902
452
- 2023-07-24 03:19:13.253475: Pseudo dice [0.8606, 0.7035, 0.5965]
453
- 2023-07-24 03:19:13.253560: Epoch time: 216.85 s
454
- 2023-07-24 03:19:14.880905:
455
- 2023-07-24 03:19:14.881205: Epoch 55
456
- 2023-07-24 03:19:14.881327: Current learning rate: 0.0095
457
- 2023-07-24 03:22:41.606614: train_loss -0.6675
458
- 2023-07-24 03:22:41.606848: val_loss -0.4941
459
- 2023-07-24 03:22:41.606937: Pseudo dice [0.873, 0.7194, 0.3798]
460
- 2023-07-24 03:22:41.607017: Epoch time: 206.73 s
461
- 2023-07-24 03:22:45.052177:
462
- 2023-07-24 03:22:45.052385: Epoch 56
463
- 2023-07-24 03:22:45.052505: Current learning rate: 0.00949
464
- 2023-07-24 03:26:17.637499: train_loss -0.6827
465
- 2023-07-24 03:26:17.637738: val_loss -0.5062
466
- 2023-07-24 03:26:17.637826: Pseudo dice [0.8737, 0.71, 0.6042]
467
- 2023-07-24 03:26:17.637978: Epoch time: 212.59 s
468
- 2023-07-24 03:26:21.340287:
469
- 2023-07-24 03:26:21.340517: Epoch 57
470
- 2023-07-24 03:26:21.340631: Current learning rate: 0.00949
471
- 2023-07-24 03:29:47.100365: train_loss -0.6838
472
- 2023-07-24 03:29:47.100567: val_loss -0.4964
473
- 2023-07-24 03:29:47.100666: Pseudo dice [0.8891, 0.7083, 0.4048]
474
- 2023-07-24 03:29:47.100763: Epoch time: 205.76 s
475
- 2023-07-24 03:29:48.555212:
476
- 2023-07-24 03:29:48.555346: Epoch 58
477
- 2023-07-24 03:29:48.555466: Current learning rate: 0.00948
478
- 2023-07-24 03:33:20.722145: train_loss -0.6697
479
- 2023-07-24 03:33:20.722324: val_loss -0.5062
480
- 2023-07-24 03:33:20.722407: Pseudo dice [0.8666, 0.6973, 0.5627]
481
- 2023-07-24 03:33:20.722488: Epoch time: 212.17 s
482
- 2023-07-24 03:33:23.647374:
483
- 2023-07-24 03:33:23.647504: Epoch 59
484
- 2023-07-24 03:33:23.647635: Current learning rate: 0.00947
485
- 2023-07-24 03:36:54.070232: train_loss -0.6887
486
- 2023-07-24 03:36:54.070504: val_loss -0.5134
487
- 2023-07-24 03:36:54.070593: Pseudo dice [0.8732, 0.7361, 0.4845]
488
- 2023-07-24 03:36:54.070743: Epoch time: 210.42 s
489
- 2023-07-24 03:36:56.444733:
490
- 2023-07-24 03:36:56.444867: Epoch 60
491
- 2023-07-24 03:36:56.444985: Current learning rate: 0.00946
492
- 2023-07-24 03:40:22.488861: train_loss -0.6985
493
- 2023-07-24 03:40:22.498857: val_loss -0.4992
494
- 2023-07-24 03:40:22.499064: Pseudo dice [0.8846, 0.7076, 0.5453]
495
- 2023-07-24 03:40:22.499147: Epoch time: 206.05 s
496
- 2023-07-24 03:40:25.415841:
497
- 2023-07-24 03:40:25.415988: Epoch 61
498
- 2023-07-24 03:40:25.416104: Current learning rate: 0.00945
499
- 2023-07-24 03:43:58.801075: train_loss -0.6876
500
- 2023-07-24 03:43:58.801260: val_loss -0.5122
501
- 2023-07-24 03:43:58.801349: Pseudo dice [0.8586, 0.716, 0.5498]
502
- 2023-07-24 03:43:58.801434: Epoch time: 213.39 s
503
- 2023-07-24 03:44:01.958030:
504
- 2023-07-24 03:44:01.970955: Epoch 62
505
- 2023-07-24 03:44:01.971086: Current learning rate: 0.00944
506
- 2023-07-24 03:47:44.738477: train_loss -0.6943
507
- 2023-07-24 03:47:44.738670: val_loss -0.5106
508
- 2023-07-24 03:47:44.738753: Pseudo dice [0.882, 0.7326, 0.544]
509
- 2023-07-24 03:47:44.738875: Epoch time: 222.78 s
510
- 2023-07-24 03:47:48.039309:
511
- 2023-07-24 03:47:48.039480: Epoch 63
512
- 2023-07-24 03:47:48.039597: Current learning rate: 0.00943
513
- 2023-07-24 03:51:16.951300: train_loss -0.6731
514
- 2023-07-24 03:51:16.951505: val_loss -0.4852
515
- 2023-07-24 03:51:16.951604: Pseudo dice [0.8708, 0.7122, 0.4841]
516
- 2023-07-24 03:51:16.951797: Epoch time: 208.91 s
517
- 2023-07-24 03:51:18.621170:
518
- 2023-07-24 03:51:18.621321: Epoch 64
519
- 2023-07-24 03:51:18.621441: Current learning rate: 0.00942
520
- 2023-07-24 03:54:44.957338: train_loss -0.6761
521
- 2023-07-24 03:54:44.957540: val_loss -0.4724
522
- 2023-07-24 03:54:44.957629: Pseudo dice [0.8598, 0.6921, 0.5167]
523
- 2023-07-24 03:54:44.957714: Epoch time: 206.34 s
524
- 2023-07-24 03:54:46.812274:
525
- 2023-07-24 03:54:46.812412: Epoch 65
526
- 2023-07-24 03:54:46.812517: Current learning rate: 0.00941
527
- 2023-07-24 03:58:23.950075: train_loss -0.6927
528
- 2023-07-24 03:58:23.950717: val_loss -0.5027
529
- 2023-07-24 03:58:23.951065: Pseudo dice [0.8701, 0.7152, 0.4785]
530
- 2023-07-24 03:58:23.951216: Epoch time: 217.14 s
531
- 2023-07-24 03:58:25.802012:
532
- 2023-07-24 03:58:25.802195: Epoch 66
533
- 2023-07-24 03:58:25.802299: Current learning rate: 0.0094
534
- 2023-07-24 04:01:56.722209: train_loss -0.6963
535
- 2023-07-24 04:01:56.722415: val_loss -0.5062
536
- 2023-07-24 04:01:56.722507: Pseudo dice [0.8726, 0.7042, 0.5969]
537
- 2023-07-24 04:01:56.722586: Epoch time: 210.92 s
538
- 2023-07-24 04:01:59.771096:
539
- 2023-07-24 04:01:59.771243: Epoch 67
540
- 2023-07-24 04:01:59.771358: Current learning rate: 0.00939
541
- 2023-07-24 04:05:40.925208: train_loss -0.6952
542
- 2023-07-24 04:05:40.925490: val_loss -0.5088
543
- 2023-07-24 04:05:40.925585: Pseudo dice [0.8802, 0.729, 0.5521]
544
- 2023-07-24 04:05:40.925745: Epoch time: 221.16 s
545
- 2023-07-24 04:05:43.897002:
546
- 2023-07-24 04:05:43.898616: Epoch 68
547
- 2023-07-24 04:05:43.898738: Current learning rate: 0.00939
548
- 2023-07-24 04:09:15.657769: train_loss -0.6889
549
- 2023-07-24 04:09:15.658007: val_loss -0.516
550
- 2023-07-24 04:09:15.658100: Pseudo dice [0.8782, 0.7278, 0.5643]
551
- 2023-07-24 04:09:15.658307: Epoch time: 211.76 s
552
- 2023-07-24 04:09:15.658453: Yayy! New best EMA pseudo Dice: 0.7051
553
- 2023-07-24 04:09:18.116199:
554
- 2023-07-24 04:09:18.116321: Epoch 69
555
- 2023-07-24 04:09:18.116441: Current learning rate: 0.00938
556
- 2023-07-24 04:12:42.753379: train_loss -0.6904
557
- 2023-07-24 04:12:42.753681: val_loss -0.4671
558
- 2023-07-24 04:12:42.753862: Pseudo dice [0.8749, 0.7138, 0.3681]
559
- 2023-07-24 04:12:42.753966: Epoch time: 204.64 s
560
- 2023-07-24 04:12:44.327600:
561
- 2023-07-24 04:12:44.327721: Epoch 70
562
- 2023-07-24 04:12:44.327840: Current learning rate: 0.00937
563
- 2023-07-24 04:16:22.761208: train_loss -0.6958
564
- 2023-07-24 04:16:22.761458: val_loss -0.5012
565
- 2023-07-24 04:16:22.761544: Pseudo dice [0.8879, 0.7126, 0.5304]
566
- 2023-07-24 04:16:22.761694: Epoch time: 218.43 s
567
- 2023-07-24 04:16:24.628846:
568
- 2023-07-24 04:16:24.629004: Epoch 71
569
- 2023-07-24 04:16:24.629131: Current learning rate: 0.00936
570
- 2023-07-24 04:19:50.557256: train_loss -0.6919
571
- 2023-07-24 04:19:50.557556: val_loss -0.4675
572
- 2023-07-24 04:19:50.557651: Pseudo dice [0.8698, 0.6855, 0.5056]
573
- 2023-07-24 04:19:50.557837: Epoch time: 205.93 s
574
- 2023-07-24 04:19:53.142438:
575
- 2023-07-24 04:19:53.142620: Epoch 72
576
- 2023-07-24 04:19:53.142735: Current learning rate: 0.00935
577
- 2023-07-24 04:23:22.464141: train_loss -0.6977
578
- 2023-07-24 04:23:22.464365: val_loss -0.5149
579
- 2023-07-24 04:23:22.464458: Pseudo dice [0.8859, 0.7116, 0.5175]
580
- 2023-07-24 04:23:22.464544: Epoch time: 209.32 s
581
- 2023-07-24 04:23:24.605744:
582
- 2023-07-24 04:23:24.605880: Epoch 73
583
- 2023-07-24 04:23:24.605996: Current learning rate: 0.00934
584
- 2023-07-24 04:26:50.364227: train_loss -0.6983
585
- 2023-07-24 04:26:50.364492: val_loss -0.4938
586
- 2023-07-24 04:26:50.364594: Pseudo dice [0.8637, 0.7213, 0.5451]
587
- 2023-07-24 04:26:50.364779: Epoch time: 205.76 s
588
- 2023-07-24 04:26:52.800122:
589
- 2023-07-24 04:26:52.800305: Epoch 74
590
- 2023-07-24 04:26:52.800421: Current learning rate: 0.00933
591
- 2023-07-24 04:30:26.828684: train_loss -0.6956
592
- 2023-07-24 04:30:26.828918: val_loss -0.4781
593
- 2023-07-24 04:30:26.829004: Pseudo dice [0.8634, 0.6982, 0.4993]
594
- 2023-07-24 04:30:26.829283: Epoch time: 214.03 s
595
- 2023-07-24 04:30:28.715835:
596
- 2023-07-24 04:30:28.716055: Epoch 75
597
- 2023-07-24 04:30:28.716175: Current learning rate: 0.00932
598
- 2023-07-24 04:33:56.996001: train_loss -0.6978
599
- 2023-07-24 04:33:56.996192: val_loss -0.5014
600
- 2023-07-24 04:33:56.996281: Pseudo dice [0.8721, 0.6953, 0.5428]
601
- 2023-07-24 04:33:56.996363: Epoch time: 208.28 s
602
- 2023-07-24 04:34:01.112788:
603
- 2023-07-24 04:34:01.112929: Epoch 76
604
- 2023-07-24 04:34:01.113038: Current learning rate: 0.00931
605
- 2023-07-24 04:37:36.118190: train_loss -0.6992
606
- 2023-07-24 04:37:36.118396: val_loss -0.5014
607
- 2023-07-24 04:37:36.118491: Pseudo dice [0.8768, 0.7055, 0.5652]
608
- 2023-07-24 04:37:36.118583: Epoch time: 215.01 s
609
- 2023-07-24 04:37:39.211814:
610
- 2023-07-24 04:37:39.211969: Epoch 77
611
- 2023-07-24 04:37:39.212115: Current learning rate: 0.0093
612
- 2023-07-24 04:41:10.694273: train_loss -0.7058
613
- 2023-07-24 04:41:10.694464: val_loss -0.4895
614
- 2023-07-24 04:41:10.694552: Pseudo dice [0.8602, 0.7049, 0.5613]
615
- 2023-07-24 04:41:10.694637: Epoch time: 211.48 s
616
- 2023-07-24 04:41:13.013998:
617
- 2023-07-24 04:41:13.014146: Epoch 78
618
- 2023-07-24 04:41:13.014262: Current learning rate: 0.0093
619
- 2023-07-24 04:44:39.537027: train_loss -0.7016
620
- 2023-07-24 04:44:39.537224: val_loss -0.4884
621
- 2023-07-24 04:44:39.537313: Pseudo dice [0.8797, 0.7206, 0.4807]
622
- 2023-07-24 04:44:39.537394: Epoch time: 206.52 s
623
- 2023-07-24 04:44:41.752397:
624
- 2023-07-24 04:44:41.752547: Epoch 79
625
- 2023-07-24 04:44:41.752668: Current learning rate: 0.00929
626
- 2023-07-24 04:48:09.494605: train_loss -0.6964
627
- 2023-07-24 04:48:09.494830: val_loss -0.5074
628
- 2023-07-24 04:48:09.494932: Pseudo dice [0.8874, 0.7203, 0.6069]
629
- 2023-07-24 04:48:09.495023: Epoch time: 207.74 s
630
- 2023-07-24 04:48:11.746876:
631
- 2023-07-24 04:48:11.747036: Epoch 80
632
- 2023-07-24 04:48:11.747147: Current learning rate: 0.00928
633
- 2023-07-24 04:51:42.348668: train_loss -0.7065
634
- 2023-07-24 04:51:42.348896: val_loss -0.4837
635
- 2023-07-24 04:51:42.348983: Pseudo dice [0.8784, 0.7032, 0.5621]
636
- 2023-07-24 04:51:42.349133: Epoch time: 210.6 s
637
- 2023-07-24 04:51:42.349199: Yayy! New best EMA pseudo Dice: 0.7061
638
- 2023-07-24 04:51:44.484021:
639
- 2023-07-24 04:51:44.484140: Epoch 81
640
- 2023-07-24 04:51:44.484257: Current learning rate: 0.00927
641
- 2023-07-24 04:55:09.696577: train_loss -0.714
642
- 2023-07-24 04:55:09.696886: val_loss -0.491
643
- 2023-07-24 04:55:09.696984: Pseudo dice [0.8814, 0.7251, 0.4554]
644
- 2023-07-24 04:55:09.697072: Epoch time: 205.21 s
645
- 2023-07-24 04:55:11.259889:
646
- 2023-07-24 04:55:11.260105: Epoch 82
647
- 2023-07-24 04:55:11.260218: Current learning rate: 0.00926
648
- 2023-07-24 04:58:42.167684: train_loss -0.7151
649
- 2023-07-24 04:58:42.167853: val_loss -0.5149
650
- 2023-07-24 04:58:42.167938: Pseudo dice [0.8747, 0.6999, 0.632]
651
- 2023-07-24 04:58:42.168024: Epoch time: 210.91 s
652
- 2023-07-24 04:58:42.168089: Yayy! New best EMA pseudo Dice: 0.7073
653
- 2023-07-24 04:58:46.756355:
654
- 2023-07-24 04:58:46.756490: Epoch 83
655
- 2023-07-24 04:58:46.756597: Current learning rate: 0.00925
656
- 2023-07-24 05:02:22.008022: train_loss -0.7232
657
- 2023-07-24 05:02:22.008212: val_loss -0.5229
658
- 2023-07-24 05:02:22.008301: Pseudo dice [0.8914, 0.7236, 0.6229]
659
- 2023-07-24 05:02:22.008386: Epoch time: 215.25 s
660
- 2023-07-24 05:02:22.008452: Yayy! New best EMA pseudo Dice: 0.7112
661
- 2023-07-24 05:02:24.066390:
662
- 2023-07-24 05:02:24.066524: Epoch 84
663
- 2023-07-24 05:02:24.066619: Current learning rate: 0.00924
664
- 2023-07-24 05:05:50.327326: train_loss -0.725
665
- 2023-07-24 05:05:50.327611: val_loss -0.5159
666
- 2023-07-24 05:05:50.327698: Pseudo dice [0.8756, 0.7106, 0.5817]
667
- 2023-07-24 05:05:50.327851: Epoch time: 206.26 s
668
- 2023-07-24 05:05:50.327918: Yayy! New best EMA pseudo Dice: 0.7123
669
- 2023-07-24 05:05:54.485906:
670
- 2023-07-24 05:05:54.486033: Epoch 85
671
- 2023-07-24 05:05:54.486146: Current learning rate: 0.00923
672
- 2023-07-24 05:09:37.322994: train_loss -0.723
673
- 2023-07-24 05:09:37.323187: val_loss -0.5029
674
- 2023-07-24 05:09:37.323277: Pseudo dice [0.8783, 0.7171, 0.4587]
675
- 2023-07-24 05:09:37.323362: Epoch time: 222.84 s
676
- 2023-07-24 05:09:38.723594:
677
- 2023-07-24 05:09:38.723815: Epoch 86
678
- 2023-07-24 05:09:38.723936: Current learning rate: 0.00922
679
- 2023-07-24 05:13:09.316705: train_loss -0.7182
680
- 2023-07-24 05:13:09.316886: val_loss -0.4834
681
- 2023-07-24 05:13:09.316982: Pseudo dice [0.8828, 0.7206, 0.4715]
682
- 2023-07-24 05:13:09.317072: Epoch time: 210.59 s
683
- 2023-07-24 05:13:12.422567:
684
- 2023-07-24 05:13:12.422688: Epoch 87
685
- 2023-07-24 05:13:12.422828: Current learning rate: 0.00921
686
- 2023-07-24 05:16:46.310607: train_loss -0.7222
687
- 2023-07-24 05:16:46.310821: val_loss -0.502
688
- 2023-07-24 05:16:46.310914: Pseudo dice [0.8901, 0.6998, 0.6513]
689
- 2023-07-24 05:16:46.310997: Epoch time: 213.89 s
690
- 2023-07-24 05:16:48.065801:
691
- 2023-07-24 05:16:48.065928: Epoch 88
692
- 2023-07-24 05:16:48.066038: Current learning rate: 0.0092
693
- 2023-07-24 05:20:26.412961: train_loss -0.7227
694
- 2023-07-24 05:20:26.413243: val_loss -0.5113
695
- 2023-07-24 05:20:26.413336: Pseudo dice [0.885, 0.7263, 0.6205]
696
- 2023-07-24 05:20:26.413419: Epoch time: 218.35 s
697
- 2023-07-24 05:20:26.413479: Yayy! New best EMA pseudo Dice: 0.7149
698
- 2023-07-24 05:20:28.466174:
699
- 2023-07-24 05:20:28.466462: Epoch 89
700
- 2023-07-24 05:20:28.466586: Current learning rate: 0.0092
701
- 2023-07-24 05:23:53.870799: train_loss -0.7281
702
- 2023-07-24 05:23:53.871042: val_loss -0.5177
703
- 2023-07-24 05:23:53.871127: Pseudo dice [0.8774, 0.7281, 0.5178]
704
- 2023-07-24 05:23:53.871280: Epoch time: 205.41 s
705
- 2023-07-24 05:23:56.666574:
706
- 2023-07-24 05:23:56.666709: Epoch 90
707
- 2023-07-24 05:23:56.666846: Current learning rate: 0.00919
708
- 2023-07-24 05:27:24.511108: train_loss -0.7268
709
- 2023-07-24 05:27:24.511282: val_loss -0.4883
710
- 2023-07-24 05:27:24.511378: Pseudo dice [0.8808, 0.7023, 0.5429]
711
- 2023-07-24 05:27:24.511463: Epoch time: 207.85 s
712
- 2023-07-24 05:27:26.253584:
713
- 2023-07-24 05:27:26.253713: Epoch 91
714
- 2023-07-24 05:27:26.253826: Current learning rate: 0.00918
715
- 2023-07-24 05:31:00.680078: train_loss -0.7187
716
- 2023-07-24 05:31:00.680262: val_loss -0.5126
717
- 2023-07-24 05:31:00.680357: Pseudo dice [0.8854, 0.734, 0.5313]
718
- 2023-07-24 05:31:00.680444: Epoch time: 214.43 s
719
- 2023-07-24 05:31:03.875111:
720
- 2023-07-24 05:31:03.875342: Epoch 92
721
- 2023-07-24 05:31:03.875453: Current learning rate: 0.00917
722
- 2023-07-24 05:34:36.700777: train_loss -0.7261
723
- 2023-07-24 05:34:36.701067: val_loss -0.4824
724
- 2023-07-24 05:34:36.701155: Pseudo dice [0.8677, 0.7068, 0.5625]
725
- 2023-07-24 05:34:36.701310: Epoch time: 212.83 s
726
- 2023-07-24 05:34:38.725615:
727
- 2023-07-24 05:34:38.725760: Epoch 93
728
- 2023-07-24 05:34:38.725878: Current learning rate: 0.00916
729
- 2023-07-24 05:38:14.723124: train_loss -0.7303
730
- 2023-07-24 05:38:14.723329: val_loss -0.4918
731
- 2023-07-24 05:38:14.723435: Pseudo dice [0.8725, 0.7201, 0.5379]
732
- 2023-07-24 05:38:14.723536: Epoch time: 216.0 s
733
- 2023-07-24 05:38:18.938028:
734
- 2023-07-24 05:38:18.938344: Epoch 94
735
- 2023-07-24 05:38:18.938459: Current learning rate: 0.00915
736
- 2023-07-24 05:41:49.722474: train_loss -0.7321
737
- 2023-07-24 05:41:49.722725: val_loss -0.5192
738
- 2023-07-24 05:41:49.722838: Pseudo dice [0.8869, 0.73, 0.5719]
739
- 2023-07-24 05:41:49.722987: Epoch time: 210.79 s
740
- 2023-07-24 05:41:49.723047: Yayy! New best EMA pseudo Dice: 0.7151
741
- 2023-07-24 05:41:52.238441:
742
- 2023-07-24 05:41:52.238770: Epoch 95
743
- 2023-07-24 05:41:52.238896: Current learning rate: 0.00914
744
- 2023-07-24 05:45:28.216917: train_loss -0.7294
745
- 2023-07-24 05:45:28.217144: val_loss -0.4859
746
- 2023-07-24 05:45:28.217231: Pseudo dice [0.8868, 0.7071, 0.502]
747
- 2023-07-24 05:45:28.217372: Epoch time: 215.98 s
748
- 2023-07-24 05:45:30.203134:
749
- 2023-07-24 05:45:30.203313: Epoch 96
750
- 2023-07-24 05:45:30.203428: Current learning rate: 0.00913
751
- 2023-07-24 05:49:07.118325: train_loss -0.7275
752
- 2023-07-24 05:49:07.130988: val_loss -0.4979
753
- 2023-07-24 05:49:07.131268: Pseudo dice [0.8742, 0.7169, 0.5796]
754
- 2023-07-24 05:49:07.131369: Epoch time: 216.92 s
755
- 2023-07-24 05:49:10.592634:
756
- 2023-07-24 05:49:10.592791: Epoch 97
757
- 2023-07-24 05:49:10.592889: Current learning rate: 0.00912
758
- 2023-07-24 05:52:37.449055: train_loss -0.7377
759
- 2023-07-24 05:52:37.449250: val_loss -0.4932
760
- 2023-07-24 05:52:37.449338: Pseudo dice [0.878, 0.7154, 0.5691]
761
- 2023-07-24 05:52:37.449421: Epoch time: 206.86 s
762
- 2023-07-24 05:52:37.449489: Yayy! New best EMA pseudo Dice: 0.7151
763
- 2023-07-24 05:52:39.613982:
764
- 2023-07-24 05:52:39.614139: Epoch 98
765
- 2023-07-24 05:52:39.614302: Current learning rate: 0.00911
766
- 2023-07-24 05:56:12.741833: train_loss -0.725
767
- 2023-07-24 05:56:12.742075: val_loss -0.5039
768
- 2023-07-24 05:56:12.742163: Pseudo dice [0.8862, 0.711, 0.6208]
769
- 2023-07-24 05:56:12.742243: Epoch time: 213.13 s
770
- 2023-07-24 05:56:12.742309: Yayy! New best EMA pseudo Dice: 0.7175
771
- 2023-07-24 05:56:16.737276:
772
- 2023-07-24 05:56:16.737576: Epoch 99
773
- 2023-07-24 05:56:16.737698: Current learning rate: 0.0091
774
- 2023-07-24 05:59:45.462096: train_loss -0.7306
775
- 2023-07-24 05:59:45.462285: val_loss -0.5266
776
- 2023-07-24 05:59:45.462388: Pseudo dice [0.8861, 0.7386, 0.513]
777
- 2023-07-24 05:59:45.462479: Epoch time: 208.73 s
778
- 2023-07-24 05:59:50.668950:
779
- 2023-07-24 05:59:50.669163: Epoch 100
780
- 2023-07-24 05:59:50.669280: Current learning rate: 0.0091
781
- 2023-07-24 06:03:33.285699: train_loss -0.7382
782
- 2023-07-24 06:03:33.285881: val_loss -0.5096
783
- 2023-07-24 06:03:33.285975: Pseudo dice [0.8703, 0.712, 0.5639]
784
- 2023-07-24 06:03:33.286058: Epoch time: 222.62 s
785
- 2023-07-24 06:03:35.851426:
786
- 2023-07-24 06:03:35.851692: Epoch 101
787
- 2023-07-24 06:03:35.851813: Current learning rate: 0.00909
788
- 2023-07-24 06:07:07.771703: train_loss -0.7398
789
- 2023-07-24 06:07:07.771932: val_loss -0.5006
790
- 2023-07-24 06:07:07.772034: Pseudo dice [0.8863, 0.723, 0.5644]
791
- 2023-07-24 06:07:07.772126: Epoch time: 211.92 s
792
- 2023-07-24 06:07:07.772207: Yayy! New best EMA pseudo Dice: 0.7176
793
- 2023-07-24 06:07:09.785565:
794
- 2023-07-24 06:07:09.785925: Epoch 102
795
- 2023-07-24 06:07:09.786036: Current learning rate: 0.00908
796
- 2023-07-24 06:10:43.597014: train_loss -0.7306
797
- 2023-07-24 06:10:43.597273: val_loss -0.5015
798
- 2023-07-24 06:10:43.599555: Pseudo dice [0.8817, 0.7251, 0.5061]
799
- 2023-07-24 06:10:43.599782: Epoch time: 213.81 s
800
- 2023-07-24 06:10:45.487350:
801
- 2023-07-24 06:10:45.487549: Epoch 103
802
- 2023-07-24 06:10:45.487668: Current learning rate: 0.00907
803
- 2023-07-24 06:14:14.858136: train_loss -0.7477
804
- 2023-07-24 06:14:14.858318: val_loss -0.4896
805
- 2023-07-24 06:14:14.858405: Pseudo dice [0.8898, 0.7226, 0.5152]
806
- 2023-07-24 06:14:14.858487: Epoch time: 209.37 s
807
- 2023-07-24 06:14:18.636439:
808
- 2023-07-24 06:14:18.636607: Epoch 104
809
- 2023-07-24 06:14:18.636723: Current learning rate: 0.00906
810
- 2023-07-24 06:17:56.407210: train_loss -0.7288
811
- 2023-07-24 06:17:56.407393: val_loss -0.4908
812
- 2023-07-24 06:17:56.407491: Pseudo dice [0.8777, 0.7242, 0.4401]
813
- 2023-07-24 06:17:56.407592: Epoch time: 217.77 s
814
- 2023-07-24 06:17:59.395841:
815
- 2023-07-24 06:17:59.406929: Epoch 105
816
- 2023-07-24 06:17:59.407068: Current learning rate: 0.00905
817
- 2023-07-24 06:21:34.379379: train_loss -0.7297
818
- 2023-07-24 06:21:34.379545: val_loss -0.5001
819
- 2023-07-24 06:21:34.379629: Pseudo dice [0.8812, 0.7174, 0.5466]
820
- 2023-07-24 06:21:34.379707: Epoch time: 215.0 s
821
- 2023-07-24 06:21:36.480978:
822
- 2023-07-24 06:21:36.481113: Epoch 106
823
- 2023-07-24 06:21:36.481229: Current learning rate: 0.00904
824
- 2023-07-24 06:24:59.495173: train_loss -0.7319
825
- 2023-07-24 06:24:59.495373: val_loss -0.4684
826
- 2023-07-24 06:24:59.495463: Pseudo dice [0.8719, 0.699, 0.4954]
827
- 2023-07-24 06:24:59.495545: Epoch time: 203.02 s
828
- 2023-07-24 06:25:01.416497:
829
- 2023-07-24 06:25:01.416672: Epoch 107
830
- 2023-07-24 06:25:01.416785: Current learning rate: 0.00903
831
- 2023-07-24 06:28:38.100931: train_loss -0.7425
832
- 2023-07-24 06:28:38.101100: val_loss -0.5066
833
- 2023-07-24 06:28:38.101199: Pseudo dice [0.8828, 0.7349, 0.551]
834
- 2023-07-24 06:28:38.101286: Epoch time: 216.69 s
835
- 2023-07-24 06:28:39.838402:
836
- 2023-07-24 06:28:39.838579: Epoch 108
837
- 2023-07-24 06:28:39.838694: Current learning rate: 0.00902
838
- 2023-07-24 06:32:19.977618: train_loss -0.7335
839
- 2023-07-24 06:32:19.990841: val_loss -0.4957
840
- 2023-07-24 06:32:19.990945: Pseudo dice [0.8713, 0.7344, 0.4371]
841
- 2023-07-24 06:32:19.991140: Epoch time: 220.14 s
842
- 2023-07-24 06:32:22.127137:
843
- 2023-07-24 06:32:22.127293: Epoch 109
844
- 2023-07-24 06:32:22.127414: Current learning rate: 0.00901
845
- 2023-07-24 06:35:56.862900: train_loss -0.7247
846
- 2023-07-24 06:35:56.863097: val_loss -0.4827
847
- 2023-07-24 06:35:56.863201: Pseudo dice [0.8801, 0.6793, 0.5831]
848
- 2023-07-24 06:35:56.863297: Epoch time: 214.74 s
849
- 2023-07-24 06:36:00.448839:
850
- 2023-07-24 06:36:00.449008: Epoch 110
851
- 2023-07-24 06:36:00.449116: Current learning rate: 0.009
852
- 2023-07-24 06:39:25.954977: train_loss -0.7307
853
- 2023-07-24 06:39:25.955158: val_loss -0.4855
854
- 2023-07-24 06:39:25.955249: Pseudo dice [0.8795, 0.6975, 0.5762]
855
- 2023-07-24 06:39:25.955334: Epoch time: 205.51 s
856
- 2023-07-24 06:39:28.934479:
857
- 2023-07-24 06:39:28.934601: Epoch 111
858
- 2023-07-24 06:39:28.934713: Current learning rate: 0.009
859
- 2023-07-24 06:42:53.226294: train_loss -0.7364
860
- 2023-07-24 06:42:53.226513: val_loss -0.4781
861
- 2023-07-24 06:42:53.226613: Pseudo dice [0.8844, 0.7035, 0.4998]
862
- 2023-07-24 06:42:53.226707: Epoch time: 204.29 s
863
- 2023-07-24 06:42:57.119614:
864
- 2023-07-24 06:42:57.119749: Epoch 112
865
- 2023-07-24 06:42:57.119859: Current learning rate: 0.00899
866
- 2023-07-24 06:46:27.868778: train_loss -0.7428
867
- 2023-07-24 06:46:27.868973: val_loss -0.5043
868
- 2023-07-24 06:46:27.869063: Pseudo dice [0.8869, 0.6979, 0.6164]
869
- 2023-07-24 06:46:27.869147: Epoch time: 210.75 s
870
- 2023-07-24 06:46:29.738551:
871
- 2023-07-24 06:46:29.738678: Epoch 113
872
- 2023-07-24 06:46:29.738820: Current learning rate: 0.00898
873
- 2023-07-24 06:49:47.016800: train_loss -0.7404
874
- 2023-07-24 06:49:47.017059: val_loss -0.484
875
- 2023-07-24 06:49:47.017146: Pseudo dice [0.8878, 0.7123, 0.5259]
876
- 2023-07-24 06:49:47.017297: Epoch time: 197.28 s
877
- 2023-07-24 06:49:49.028085:
878
- 2023-07-24 06:49:49.028212: Epoch 114
879
- 2023-07-24 06:49:49.028322: Current learning rate: 0.00897
880
- 2023-07-24 06:53:22.595026: train_loss -0.7403
881
- 2023-07-24 06:53:22.595240: val_loss -0.4923
882
- 2023-07-24 06:53:22.595333: Pseudo dice [0.8744, 0.7065, 0.5055]
883
- 2023-07-24 06:53:22.595417: Epoch time: 213.57 s
884
- 2023-07-24 06:53:25.630225:
885
- 2023-07-24 06:53:25.630350: Epoch 115
886
- 2023-07-24 06:53:25.630466: Current learning rate: 0.00896
887
- 2023-07-24 06:56:54.846274: train_loss -0.7407
888
- 2023-07-24 06:56:54.846479: val_loss -0.5206
889
- 2023-07-24 06:56:54.846571: Pseudo dice [0.8874, 0.7354, 0.4706]
890
- 2023-07-24 06:56:54.846656: Epoch time: 209.22 s
891
- 2023-07-24 06:56:58.530087:
892
- 2023-07-24 06:56:58.530329: Epoch 116
893
- 2023-07-24 06:56:58.530463: Current learning rate: 0.00895
894
- 2023-07-24 07:00:33.757893: train_loss -0.7397
895
- 2023-07-24 07:00:33.758083: val_loss -0.4963
896
- 2023-07-24 07:00:33.758173: Pseudo dice [0.8784, 0.7196, 0.5139]
897
- 2023-07-24 07:00:33.758259: Epoch time: 215.23 s
898
- 2023-07-24 07:00:36.559549:
899
- 2023-07-24 07:00:36.559805: Epoch 117
900
- 2023-07-24 07:00:36.559918: Current learning rate: 0.00894
901
- 2023-07-24 07:04:09.798198: train_loss -0.737
902
- 2023-07-24 07:04:09.798404: val_loss -0.4986
903
- 2023-07-24 07:04:09.809834: Pseudo dice [0.879, 0.706, 0.5266]
904
- 2023-07-24 07:04:09.810041: Epoch time: 213.24 s
905
- 2023-07-24 07:04:13.458022:
906
- 2023-07-24 07:04:13.458441: Epoch 118
907
- 2023-07-24 07:04:13.458563: Current learning rate: 0.00893
908
- 2023-07-24 07:07:55.203243: train_loss -0.7376
909
- 2023-07-24 07:07:55.203503: val_loss -0.5011
910
- 2023-07-24 07:07:55.203593: Pseudo dice [0.8852, 0.7273, 0.5817]
911
- 2023-07-24 07:07:55.203745: Epoch time: 221.75 s
912
- 2023-07-24 07:07:57.539746:
913
- 2023-07-24 07:07:57.539875: Epoch 119
914
- 2023-07-24 07:07:57.539999: Current learning rate: 0.00892
915
- 2023-07-24 07:11:23.694811: train_loss -0.7511
916
- 2023-07-24 07:11:23.706887: val_loss -0.5173
917
- 2023-07-24 07:11:23.707143: Pseudo dice [0.8818, 0.7369, 0.559]
918
- 2023-07-24 07:11:23.707327: Epoch time: 206.16 s
919
- 2023-07-24 07:11:26.298614:
920
- 2023-07-24 07:11:26.298933: Epoch 120
921
- 2023-07-24 07:11:26.299044: Current learning rate: 0.00891
922
- 2023-07-24 07:14:58.226911: train_loss -0.7376
923
- 2023-07-24 07:14:58.227126: val_loss -0.4984
924
- 2023-07-24 07:14:58.227228: Pseudo dice [0.8841, 0.7261, 0.4873]
925
- 2023-07-24 07:14:58.227322: Epoch time: 211.93 s
926
- 2023-07-24 07:15:00.409765:
927
- 2023-07-24 07:15:00.409939: Epoch 121
928
- 2023-07-24 07:15:00.410051: Current learning rate: 0.0089
929
- 2023-07-24 07:18:26.223230: train_loss -0.7331
930
- 2023-07-24 07:18:26.223434: val_loss -0.4727
931
- 2023-07-24 07:18:26.223520: Pseudo dice [0.8632, 0.6888, 0.5576]
932
- 2023-07-24 07:18:26.223612: Epoch time: 205.81 s
933
- 2023-07-24 07:18:28.232954:
934
- 2023-07-24 07:18:28.233098: Epoch 122
935
- 2023-07-24 07:18:28.233214: Current learning rate: 0.00889
936
- 2023-07-24 07:22:00.937126: train_loss -0.739
937
- 2023-07-24 07:22:00.937339: val_loss -0.4894
938
- 2023-07-24 07:22:00.937483: Pseudo dice [0.8769, 0.7199, 0.4559]
939
- 2023-07-24 07:22:00.937608: Epoch time: 212.71 s
940
- 2023-07-24 07:22:04.448639:
941
- 2023-07-24 07:22:04.448863: Epoch 123
942
- 2023-07-24 07:22:04.448967: Current learning rate: 0.00889
943
- 2023-07-24 07:25:39.226343: train_loss -0.7513
944
- 2023-07-24 07:25:39.226538: val_loss -0.4662
945
- 2023-07-24 07:25:39.226624: Pseudo dice [0.8728, 0.7032, 0.4345]
946
- 2023-07-24 07:25:39.226705: Epoch time: 214.78 s
947
- 2023-07-24 07:25:41.940106:
948
- 2023-07-24 07:25:41.940430: Epoch 124
949
- 2023-07-24 07:25:41.940552: Current learning rate: 0.00888
950
- 2023-07-24 07:29:31.867979: train_loss -0.7385
951
- 2023-07-24 07:29:31.868199: val_loss -0.4822
952
- 2023-07-24 07:29:31.868285: Pseudo dice [0.8879, 0.7072, 0.4662]
953
- 2023-07-24 07:29:31.868367: Epoch time: 229.93 s
954
- 2023-07-24 07:29:33.666378:
955
- 2023-07-24 07:29:33.666722: Epoch 125
956
- 2023-07-24 07:29:33.666862: Current learning rate: 0.00887
957
- 2023-07-24 07:33:13.897824: train_loss -0.735
958
- 2023-07-24 07:33:13.898085: val_loss -0.5056
959
- 2023-07-24 07:33:13.898172: Pseudo dice [0.8899, 0.7388, 0.4273]
960
- 2023-07-24 07:33:13.898322: Epoch time: 220.23 s
961
- 2023-07-24 07:33:15.701629:
962
- 2023-07-24 07:33:15.701919: Epoch 126
963
- 2023-07-24 07:33:15.702039: Current learning rate: 0.00886
964
- 2023-07-24 07:36:46.580624: train_loss -0.7412
965
- 2023-07-24 07:36:46.590938: val_loss -0.4682
966
- 2023-07-24 07:36:46.591583: Pseudo dice [0.8875, 0.707, 0.464]
967
- 2023-07-24 07:36:46.591727: Epoch time: 210.88 s
968
- 2023-07-24 07:36:49.059450:
969
- 2023-07-24 07:36:49.059957: Epoch 127
970
- 2023-07-24 07:36:49.060068: Current learning rate: 0.00885
971
- 2023-07-24 07:40:16.085344: train_loss -0.7467
972
- 2023-07-24 07:40:16.085670: val_loss -0.5007
973
- 2023-07-24 07:40:16.085762: Pseudo dice [0.8923, 0.7309, 0.4699]
974
- 2023-07-24 07:40:16.085928: Epoch time: 207.03 s
975
- 2023-07-24 07:40:18.069315:
976
- 2023-07-24 07:40:18.069484: Epoch 128
977
- 2023-07-24 07:40:18.069586: Current learning rate: 0.00884
978
- 2023-07-24 07:43:48.285688: train_loss -0.7613
979
- 2023-07-24 07:43:48.285896: val_loss -0.5026
980
- 2023-07-24 07:43:48.285983: Pseudo dice [0.8685, 0.7138, 0.5262]
981
- 2023-07-24 07:43:48.286072: Epoch time: 210.22 s
982
- 2023-07-24 07:43:50.040725:
983
- 2023-07-24 07:43:50.040847: Epoch 129
984
- 2023-07-24 07:43:50.040961: Current learning rate: 0.00883
985
- 2023-07-24 07:47:20.600014: train_loss -0.749
986
- 2023-07-24 07:47:20.600197: val_loss -0.4715
987
- 2023-07-24 07:47:20.600303: Pseudo dice [0.8669, 0.7066, 0.5142]
988
- 2023-07-24 07:47:20.600403: Epoch time: 210.56 s
989
- 2023-07-24 07:47:22.825845:
990
- 2023-07-24 07:47:22.825982: Epoch 130
991
- 2023-07-24 07:47:22.826100: Current learning rate: 0.00882
992
- 2023-07-24 07:51:01.366344: train_loss -0.7542
993
- 2023-07-24 07:51:01.372711: val_loss -0.4867
994
- 2023-07-24 07:51:01.373079: Pseudo dice [0.8709, 0.7212, 0.5275]
995
- 2023-07-24 07:51:01.373174: Epoch time: 218.54 s
996
- 2023-07-24 07:51:04.639824:
997
- 2023-07-24 07:51:04.640065: Epoch 131
998
- 2023-07-24 07:51:04.640186: Current learning rate: 0.00881
999
- 2023-07-24 07:54:25.996458: train_loss -0.7539
1000
- 2023-07-24 07:54:25.996660: val_loss -0.493
1001
- 2023-07-24 07:54:25.996764: Pseudo dice [0.879, 0.7073, 0.5086]
1002
- 2023-07-24 07:54:25.996862: Epoch time: 201.36 s
1003
- 2023-07-24 07:54:27.833304:
1004
- 2023-07-24 07:54:27.833434: Epoch 132
1005
- 2023-07-24 07:54:27.833548: Current learning rate: 0.0088
1006
- 2023-07-24 07:58:06.308678: train_loss -0.7424
1007
- 2023-07-24 07:58:06.308868: val_loss -0.4781
1008
- 2023-07-24 07:58:06.308955: Pseudo dice [0.8749, 0.7132, 0.4636]
1009
- 2023-07-24 07:58:06.309040: Epoch time: 218.48 s
1010
- 2023-07-24 07:58:07.995495:
1011
- 2023-07-24 07:58:07.995820: Epoch 133
1012
- 2023-07-24 07:58:07.995934: Current learning rate: 0.00879
1013
- 2023-07-24 08:01:39.319222: train_loss -0.7583
1014
- 2023-07-24 08:01:39.319408: val_loss -0.4887
1015
- 2023-07-24 08:01:39.319496: Pseudo dice [0.8776, 0.6941, 0.6211]
1016
- 2023-07-24 08:01:39.319587: Epoch time: 211.32 s
1017
- 2023-07-24 08:01:42.469862:
1018
- 2023-07-24 08:01:42.470145: Epoch 134
1019
- 2023-07-24 08:01:42.470260: Current learning rate: 0.00879
1020
- 2023-07-24 08:05:12.546540: train_loss -0.7622
1021
- 2023-07-24 08:05:12.546793: val_loss -0.487
1022
- 2023-07-24 08:05:12.546890: Pseudo dice [0.877, 0.6954, 0.6036]
1023
- 2023-07-24 08:05:12.547057: Epoch time: 210.08 s
1024
- 2023-07-24 08:05:15.318394:
1025
- 2023-07-24 08:05:15.318724: Epoch 135
1026
- 2023-07-24 08:05:15.318935: Current learning rate: 0.00878
1027
- 2023-07-24 08:08:54.008573: train_loss -0.757
1028
- 2023-07-24 08:08:54.008773: val_loss -0.4737
1029
- 2023-07-24 08:08:54.008857: Pseudo dice [0.8687, 0.6997, 0.5345]
1030
- 2023-07-24 08:08:54.008935: Epoch time: 218.69 s
1031
- 2023-07-24 08:08:57.009470:
1032
- 2023-07-24 08:08:57.009609: Epoch 136
1033
- 2023-07-24 08:08:57.009723: Current learning rate: 0.00877
1034
- 2023-07-24 08:12:31.627689: train_loss -0.7468
1035
- 2023-07-24 08:12:31.628009: val_loss -0.4989
1036
- 2023-07-24 08:12:31.628209: Pseudo dice [0.8809, 0.7079, 0.5713]
1037
- 2023-07-24 08:12:31.628382: Epoch time: 214.62 s
1038
- 2023-07-24 08:12:34.747013:
1039
- 2023-07-24 08:12:34.747181: Epoch 137
1040
- 2023-07-24 08:12:34.747300: Current learning rate: 0.00876
1041
- 2023-07-24 08:16:01.029350: train_loss -0.7588
1042
- 2023-07-24 08:16:01.029534: val_loss -0.4855
1043
- 2023-07-24 08:16:01.029621: Pseudo dice [0.8791, 0.6898, 0.6106]
1044
- 2023-07-24 08:16:01.029705: Epoch time: 206.28 s
1045
- 2023-07-24 08:16:02.771766:
1046
- 2023-07-24 08:16:02.771899: Epoch 138
1047
- 2023-07-24 08:16:02.772012: Current learning rate: 0.00875
1048
- 2023-07-24 08:19:44.406471: train_loss -0.747
1049
- 2023-07-24 08:19:44.406666: val_loss -0.5084
1050
- 2023-07-24 08:19:44.406777: Pseudo dice [0.8869, 0.7204, 0.5542]
1051
- 2023-07-24 08:19:44.406874: Epoch time: 221.64 s
1052
- 2023-07-24 08:19:47.105821:
1053
- 2023-07-24 08:19:47.105972: Epoch 139
1054
- 2023-07-24 08:19:47.106068: Current learning rate: 0.00874
1055
- 2023-07-24 08:23:18.237113: train_loss -0.7528
1056
- 2023-07-24 08:23:18.237421: val_loss -0.4871
1057
- 2023-07-24 08:23:18.237525: Pseudo dice [0.8776, 0.706, 0.5431]
1058
- 2023-07-24 08:23:18.237715: Epoch time: 211.13 s
1059
- 2023-07-24 08:23:20.150906:
1060
- 2023-07-24 08:23:20.151026: Epoch 140
1061
- 2023-07-24 08:23:20.151122: Current learning rate: 0.00873
1062
- 2023-07-24 08:26:46.623374: train_loss -0.7637
1063
- 2023-07-24 08:26:46.623651: val_loss -0.4917
1064
- 2023-07-24 08:26:46.623740: Pseudo dice [0.8785, 0.6989, 0.6665]
1065
- 2023-07-24 08:26:46.623897: Epoch time: 206.47 s
1066
- 2023-07-24 08:26:48.784806:
1067
- 2023-07-24 08:26:48.785008: Epoch 141
1068
- 2023-07-24 08:26:48.785114: Current learning rate: 0.00872
1069
- 2023-07-24 08:30:30.098001: train_loss -0.7605
1070
- 2023-07-24 08:30:30.098207: val_loss -0.499
1071
- 2023-07-24 08:30:30.098298: Pseudo dice [0.8795, 0.7093, 0.5429]
1072
- 2023-07-24 08:30:30.098378: Epoch time: 221.31 s
1073
- 2023-07-24 08:30:32.113620:
1074
- 2023-07-24 08:30:32.113780: Epoch 142
1075
- 2023-07-24 08:30:32.113915: Current learning rate: 0.00871
1076
- 2023-07-24 08:33:58.876559: train_loss -0.7434
1077
- 2023-07-24 08:33:58.876756: val_loss -0.4936
1078
- 2023-07-24 08:33:58.876848: Pseudo dice [0.8844, 0.7121, 0.5416]
1079
- 2023-07-24 08:33:58.876932: Epoch time: 206.76 s
1080
- 2023-07-24 08:34:01.494818:
1081
- 2023-07-24 08:34:01.495366: Epoch 143
1082
- 2023-07-24 08:34:01.495486: Current learning rate: 0.0087
1083
- 2023-07-24 08:37:26.398371: train_loss -0.7416
1084
- 2023-07-24 08:37:26.398579: val_loss -0.5108
1085
- 2023-07-24 08:37:26.398680: Pseudo dice [0.8855, 0.7182, 0.5946]
1086
- 2023-07-24 08:37:26.398790: Epoch time: 204.91 s
1087
- 2023-07-24 08:37:30.655529:
1088
- 2023-07-24 08:37:30.655930: Epoch 144
1089
- 2023-07-24 08:37:30.656043: Current learning rate: 0.00869
1090
- 2023-07-24 08:40:55.134679: train_loss -0.7574
1091
- 2023-07-24 08:40:55.134889: val_loss -0.4785
1092
- 2023-07-24 08:40:55.134991: Pseudo dice [0.8803, 0.7162, 0.3907]
1093
- 2023-07-24 08:40:55.135085: Epoch time: 204.48 s
1094
- 2023-07-24 08:40:57.261309:
1095
- 2023-07-24 08:40:57.261564: Epoch 145
1096
- 2023-07-24 08:40:57.261678: Current learning rate: 0.00868
1097
- 2023-07-24 08:44:23.101116: train_loss -0.7516
1098
- 2023-07-24 08:44:23.106944: val_loss -0.4902
1099
- 2023-07-24 08:44:23.107199: Pseudo dice [0.8784, 0.7066, 0.5606]
1100
- 2023-07-24 08:44:23.107291: Epoch time: 205.84 s
1101
- 2023-07-24 08:44:25.099223:
1102
- 2023-07-24 08:44:25.099430: Epoch 146
1103
- 2023-07-24 08:44:25.099553: Current learning rate: 0.00868
1104
- 2023-07-24 08:47:56.039377: train_loss -0.759
1105
- 2023-07-24 08:47:56.039689: val_loss -0.5251
1106
- 2023-07-24 08:47:56.039789: Pseudo dice [0.8786, 0.7236, 0.5975]
1107
- 2023-07-24 08:47:56.039928: Epoch time: 210.94 s
1108
- 2023-07-24 08:47:58.670197:
1109
- 2023-07-24 08:47:58.670312: Epoch 147
1110
- 2023-07-24 08:47:58.670412: Current learning rate: 0.00867
1111
- 2023-07-24 08:51:23.394095: train_loss -0.7657
1112
- 2023-07-24 08:51:23.394388: val_loss -0.4863
1113
- 2023-07-24 08:51:23.394477: Pseudo dice [0.8613, 0.7275, 0.4779]
1114
- 2023-07-24 08:51:23.394634: Epoch time: 204.72 s
1115
- 2023-07-24 08:51:25.151827:
1116
- 2023-07-24 08:51:25.152007: Epoch 148
1117
- 2023-07-24 08:51:25.152129: Current learning rate: 0.00866
1118
- 2023-07-24 08:55:00.305970: train_loss -0.759
1119
- 2023-07-24 08:55:00.306156: val_loss -0.4877
1120
- 2023-07-24 08:55:00.306260: Pseudo dice [0.8708, 0.7051, 0.5229]
1121
- 2023-07-24 08:55:00.306348: Epoch time: 215.16 s
1122
- 2023-07-24 08:55:03.166358:
1123
- 2023-07-24 08:55:03.166571: Epoch 149
1124
- 2023-07-24 08:55:03.166703: Current learning rate: 0.00865
1125
- 2023-07-24 08:58:36.158555: train_loss -0.7584
1126
- 2023-07-24 08:58:36.163698: val_loss -0.5075
1127
- 2023-07-24 08:58:36.163839: Pseudo dice [0.8796, 0.7186, 0.6129]
1128
- 2023-07-24 08:58:36.163924: Epoch time: 212.99 s
1129
- 2023-07-24 08:58:39.431387:
1130
- 2023-07-24 08:58:39.431509: Epoch 150
1131
- 2023-07-24 08:58:39.431633: Current learning rate: 0.00864
1132
- 2023-07-24 09:02:08.121246: train_loss -0.7586
1133
- 2023-07-24 09:02:08.121416: val_loss -0.5081
1134
- 2023-07-24 09:02:08.121502: Pseudo dice [0.8847, 0.727, 0.6269]
1135
- 2023-07-24 09:02:08.121582: Epoch time: 208.69 s
1136
- 2023-07-24 09:02:10.686369:
1137
- 2023-07-24 09:02:10.686602: Epoch 151
1138
- 2023-07-24 09:02:10.686719: Current learning rate: 0.00863
1139
- 2023-07-24 09:05:44.706747: train_loss -0.7629
1140
- 2023-07-24 09:05:44.706952: val_loss -0.4625
1141
- 2023-07-24 09:05:44.707042: Pseudo dice [0.8698, 0.6773, 0.509]
1142
- 2023-07-24 09:05:44.707127: Epoch time: 214.02 s
1143
- 2023-07-24 09:05:48.504527:
1144
- 2023-07-24 09:05:48.504706: Epoch 152
1145
- 2023-07-24 09:05:48.504832: Current learning rate: 0.00862
1146
- 2023-07-24 09:09:19.069351: train_loss -0.7578
1147
- 2023-07-24 09:09:19.069588: val_loss -0.5
1148
- 2023-07-24 09:09:19.069682: Pseudo dice [0.8778, 0.7217, 0.5765]
1149
- 2023-07-24 09:09:19.069833: Epoch time: 210.57 s
1150
- 2023-07-24 09:09:21.075281:
1151
- 2023-07-24 09:09:21.075402: Epoch 153
1152
- 2023-07-24 09:09:21.075518: Current learning rate: 0.00861
1153
- 2023-07-24 09:12:58.820027: train_loss -0.7531
1154
- 2023-07-24 09:12:58.826884: val_loss -0.4773
1155
- 2023-07-24 09:12:58.827152: Pseudo dice [0.8738, 0.7111, 0.5466]
1156
- 2023-07-24 09:12:58.827244: Epoch time: 217.75 s
1157
- 2023-07-24 09:13:02.862645:
1158
- 2023-07-24 09:13:02.862877: Epoch 154
1159
- 2023-07-24 09:13:02.862992: Current learning rate: 0.0086
1160
- 2023-07-24 09:16:28.130270: train_loss -0.7474
1161
- 2023-07-24 09:16:28.130463: val_loss -0.4756
1162
- 2023-07-24 09:16:28.130564: Pseudo dice [0.8795, 0.7165, 0.489]
1163
- 2023-07-24 09:16:28.130655: Epoch time: 205.27 s
1164
- 2023-07-24 09:16:30.896615:
1165
- 2023-07-24 09:16:30.896794: Epoch 155
1166
- 2023-07-24 09:16:30.896921: Current learning rate: 0.00859
1167
- 2023-07-24 09:20:10.606297: train_loss -0.7662
1168
- 2023-07-24 09:20:10.609706: val_loss -0.4819
1169
- 2023-07-24 09:20:10.609968: Pseudo dice [0.8773, 0.711, 0.5174]
1170
- 2023-07-24 09:20:10.610058: Epoch time: 219.71 s
1171
- 2023-07-24 09:20:13.662477:
1172
- 2023-07-24 09:20:13.662828: Epoch 156
1173
- 2023-07-24 09:20:13.662953: Current learning rate: 0.00858
1174
- 2023-07-24 09:23:55.274102: train_loss -0.7626
1175
- 2023-07-24 09:23:55.274285: val_loss -0.4917
1176
- 2023-07-24 09:23:55.274372: Pseudo dice [0.8786, 0.7163, 0.5242]
1177
- 2023-07-24 09:23:55.274455: Epoch time: 221.61 s
1178
- 2023-07-24 09:23:56.839917:
1179
- 2023-07-24 09:23:56.840061: Epoch 157
1180
- 2023-07-24 09:23:56.840175: Current learning rate: 0.00858
1181
- 2023-07-24 09:27:20.148015: train_loss -0.7542
1182
- 2023-07-24 09:27:20.148191: val_loss -0.5016
1183
- 2023-07-24 09:27:20.148277: Pseudo dice [0.8869, 0.7229, 0.5561]
1184
- 2023-07-24 09:27:20.148360: Epoch time: 203.31 s
1185
- 2023-07-24 09:27:21.869529:
1186
- 2023-07-24 09:27:21.869649: Epoch 158
1187
- 2023-07-24 09:27:21.869765: Current learning rate: 0.00857
1188
- 2023-07-24 09:30:58.994606: train_loss -0.7622
1189
- 2023-07-24 09:30:58.994871: val_loss -0.4855
1190
- 2023-07-24 09:30:58.994982: Pseudo dice [0.8842, 0.7165, 0.4867]
1191
- 2023-07-24 09:30:58.995074: Epoch time: 217.13 s
1192
- 2023-07-24 09:31:03.135907:
1193
- 2023-07-24 09:31:03.136067: Epoch 159
1194
- 2023-07-24 09:31:03.136214: Current learning rate: 0.00856
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/.ipynb_checkpoints/Untitled-checkpoint.ipynb DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "cells": [],
3
- "metadata": {},
4
- "nbformat": 4,
5
- "nbformat_minor": 5
6
- }
 
 
 
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/.ipynb_checkpoints/progress-checkpoint.png DELETED

Git LFS Details

  • SHA256: 8fc57b4caaa26cf16f39f128306a43910994edbe76b8c3ae7ccf055fd818ebf7
  • Pointer size: 131 Bytes
  • Size of remote file: 463 kB
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/Untitled.ipynb DELETED
@@ -1,112 +0,0 @@
1
- {
2
- "cells": [
3
- {
4
- "cell_type": "code",
5
- "execution_count": 7,
6
- "id": "f69d6289-cca0-4546-baf8-c3e0bd052370",
7
- "metadata": {},
8
- "outputs": [
9
- {
10
- "name": "stdout",
11
- "output_type": "stream",
12
- "text": [
13
- " \u001b[0m\u001b[01;34mProstate158\u001b[0m/ create_nnunet_dataset.py \u001b[01;34mnnUNet_results\u001b[0m/\n",
14
- "'Untitled (2) (1) (2).ipynb' \u001b[01;34mnnUNet_preprocessed\u001b[0m/\n",
15
- " Untitled.ipynb \u001b[01;34mnnUNet_raw\u001b[0m/\n"
16
- ]
17
- }
18
- ],
19
- "source": [
20
- "ls"
21
- ]
22
- },
23
- {
24
- "cell_type": "code",
25
- "execution_count": 5,
26
- "id": "d358c520-7bfb-4846-a9a7-e01a151b6912",
27
- "metadata": {},
28
- "outputs": [
29
- {
30
- "name": "stderr",
31
- "output_type": "stream",
32
- "text": [
33
- "/usr/local/lib/python3.10/dist-packages/tqdm/auto.py:21: TqdmWarning: IProgress not found. Please update jupyter and ipywidgets. See https://ipywidgets.readthedocs.io/en/stable/user_install.html\n",
34
- " from .autonotebook import tqdm as notebook_tqdm\n"
35
- ]
36
- },
37
- {
38
- "ename": "RepositoryNotFoundError",
39
- "evalue": "404 Client Error. (Request ID: Root=1-64be540a-13608ebb4f5e72723d9907a8;bec748ef-4d79-4099-914c-5e806f83f5e0)\n\nRepository Not Found for url: https://huggingface.co/api/spaces/username/my-cool-space/preupload/main.\nPlease make sure you specified the correct `repo_id` and `repo_type`.\nIf you are trying to access a private or gated repo, make sure you are authenticated.\nNote: Creating a commit assumes that the repo already exists on the Huggingface Hub. Please use `create_repo` if it's not the case.",
40
- "output_type": "error",
41
- "traceback": [
42
- "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
43
- "\u001b[0;31mHTTPError\u001b[0m Traceback (most recent call last)",
44
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py:261\u001b[0m, in \u001b[0;36mhf_raise_for_status\u001b[0;34m(response, endpoint_name)\u001b[0m\n\u001b[1;32m 260\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[0;32m--> 261\u001b[0m \u001b[43mresponse\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mraise_for_status\u001b[49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 262\u001b[0m \u001b[38;5;28;01mexcept\u001b[39;00m HTTPError \u001b[38;5;28;01mas\u001b[39;00m e:\n",
45
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/requests/models.py:1021\u001b[0m, in \u001b[0;36mResponse.raise_for_status\u001b[0;34m(self)\u001b[0m\n\u001b[1;32m 1020\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m http_error_msg:\n\u001b[0;32m-> 1021\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m HTTPError(http_error_msg, response\u001b[38;5;241m=\u001b[39m\u001b[38;5;28mself\u001b[39m)\n",
46
- "\u001b[0;31mHTTPError\u001b[0m: 404 Client Error: Not Found for url: https://huggingface.co/api/spaces/username/my-cool-space/preupload/main",
47
- "\nThe above exception was the direct cause of the following exception:\n",
48
- "\u001b[0;31mRepositoryNotFoundError\u001b[0m Traceback (most recent call last)",
49
- "Cell \u001b[0;32mIn[5], line 6\u001b[0m\n\u001b[1;32m 2\u001b[0m api \u001b[38;5;241m=\u001b[39m HfApi()\n\u001b[1;32m 4\u001b[0m \u001b[38;5;66;03m# Upload all the content from the local folder to your remote Space.\u001b[39;00m\n\u001b[1;32m 5\u001b[0m \u001b[38;5;66;03m# By default, files are uploaded at the root of the repo\u001b[39;00m\n\u001b[0;32m----> 6\u001b[0m \u001b[43mapi\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mupload_folder\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 7\u001b[0m \u001b[43m \u001b[49m\u001b[43mfolder_path\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m,\u001b[49m\n\u001b[1;32m 8\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_id\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43musername/my-cool-space\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m,\u001b[49m\n\u001b[1;32m 9\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_type\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43mspace\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m,\u001b[49m\n\u001b[1;32m 10\u001b[0m \u001b[43m)\u001b[49m\n",
50
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py:118\u001b[0m, in \u001b[0;36mvalidate_hf_hub_args.<locals>._inner_fn\u001b[0;34m(*args, **kwargs)\u001b[0m\n\u001b[1;32m 115\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m check_use_auth_token:\n\u001b[1;32m 116\u001b[0m kwargs \u001b[38;5;241m=\u001b[39m smoothly_deprecate_use_auth_token(fn_name\u001b[38;5;241m=\u001b[39mfn\u001b[38;5;241m.\u001b[39m\u001b[38;5;18m__name__\u001b[39m, has_token\u001b[38;5;241m=\u001b[39mhas_token, kwargs\u001b[38;5;241m=\u001b[39mkwargs)\n\u001b[0;32m--> 118\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfn\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n",
51
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/hf_api.py:828\u001b[0m, in \u001b[0;36mfuture_compatible.<locals>._inner\u001b[0;34m(self, *args, **kwargs)\u001b[0m\n\u001b[1;32m 825\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mrun_as_future(fn, \u001b[38;5;28mself\u001b[39m, \u001b[38;5;241m*\u001b[39margs, \u001b[38;5;241m*\u001b[39m\u001b[38;5;241m*\u001b[39mkwargs)\n\u001b[1;32m 827\u001b[0m \u001b[38;5;66;03m# Otherwise, call the function normally\u001b[39;00m\n\u001b[0;32m--> 828\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfn\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n",
52
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/hf_api.py:3509\u001b[0m, in \u001b[0;36mHfApi.upload_folder\u001b[0;34m(self, repo_id, folder_path, path_in_repo, commit_message, commit_description, token, repo_type, revision, create_pr, parent_commit, allow_patterns, ignore_patterns, delete_patterns, multi_commits, multi_commits_verbose, run_as_future)\u001b[0m\n\u001b[1;32m 3497\u001b[0m pr_url \u001b[38;5;241m=\u001b[39m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mcreate_commits_on_pr(\n\u001b[1;32m 3498\u001b[0m repo_id\u001b[38;5;241m=\u001b[39mrepo_id,\n\u001b[1;32m 3499\u001b[0m repo_type\u001b[38;5;241m=\u001b[39mrepo_type,\n\u001b[0;32m (...)\u001b[0m\n\u001b[1;32m 3506\u001b[0m verbose\u001b[38;5;241m=\u001b[39mmulti_commits_verbose,\n\u001b[1;32m 3507\u001b[0m )\n\u001b[1;32m 3508\u001b[0m \u001b[38;5;28;01melse\u001b[39;00m:\n\u001b[0;32m-> 3509\u001b[0m commit_info \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mcreate_commit\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 3510\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_type\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrepo_type\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3511\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_id\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrepo_id\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3512\u001b[0m \u001b[43m \u001b[49m\u001b[43moperations\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mcommit_operations\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3513\u001b[0m \u001b[43m \u001b[49m\u001b[43mcommit_message\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mcommit_message\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3514\u001b[0m \u001b[43m \u001b[49m\u001b[43mcommit_description\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mcommit_description\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3515\u001b[0m \u001b[43m \u001b[49m\u001b[43mtoken\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mtoken\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3516\u001b[0m \u001b[43m \u001b[49m\u001b[43mrevision\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrevision\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3517\u001b[0m \u001b[43m \u001b[49m\u001b[43mcreate_pr\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mcreate_pr\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3518\u001b[0m \u001b[43m \u001b[49m\u001b[43mparent_commit\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mparent_commit\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3519\u001b[0m \u001b[43m \u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 3520\u001b[0m pr_url \u001b[38;5;241m=\u001b[39m commit_info\u001b[38;5;241m.\u001b[39mpr_url\n\u001b[1;32m 3522\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m create_pr \u001b[38;5;129;01mand\u001b[39;00m pr_url \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m:\n",
53
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py:118\u001b[0m, in \u001b[0;36mvalidate_hf_hub_args.<locals>._inner_fn\u001b[0;34m(*args, **kwargs)\u001b[0m\n\u001b[1;32m 115\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m check_use_auth_token:\n\u001b[1;32m 116\u001b[0m kwargs \u001b[38;5;241m=\u001b[39m smoothly_deprecate_use_auth_token(fn_name\u001b[38;5;241m=\u001b[39mfn\u001b[38;5;241m.\u001b[39m\u001b[38;5;18m__name__\u001b[39m, has_token\u001b[38;5;241m=\u001b[39mhas_token, kwargs\u001b[38;5;241m=\u001b[39mkwargs)\n\u001b[0;32m--> 118\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfn\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n",
54
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/hf_api.py:828\u001b[0m, in \u001b[0;36mfuture_compatible.<locals>._inner\u001b[0;34m(self, *args, **kwargs)\u001b[0m\n\u001b[1;32m 825\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mrun_as_future(fn, \u001b[38;5;28mself\u001b[39m, \u001b[38;5;241m*\u001b[39margs, \u001b[38;5;241m*\u001b[39m\u001b[38;5;241m*\u001b[39mkwargs)\n\u001b[1;32m 827\u001b[0m \u001b[38;5;66;03m# Otherwise, call the function normally\u001b[39;00m\n\u001b[0;32m--> 828\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfn\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n",
55
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/hf_api.py:2675\u001b[0m, in \u001b[0;36mHfApi.create_commit\u001b[0;34m(self, repo_id, operations, commit_message, commit_description, token, repo_type, revision, create_pr, num_threads, parent_commit, run_as_future)\u001b[0m\n\u001b[1;32m 2672\u001b[0m warn_on_overwriting_operations(operations)\n\u001b[1;32m 2674\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[0;32m-> 2675\u001b[0m upload_modes \u001b[38;5;241m=\u001b[39m \u001b[43mfetch_upload_modes\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 2676\u001b[0m \u001b[43m \u001b[49m\u001b[43madditions\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43madditions\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 2677\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_type\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrepo_type\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 2678\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_id\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrepo_id\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 2679\u001b[0m \u001b[43m \u001b[49m\u001b[43mtoken\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mtoken\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;129;43;01mor\u001b[39;49;00m\u001b[43m \u001b[49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mtoken\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 2680\u001b[0m \u001b[43m \u001b[49m\u001b[43mrevision\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrevision\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 2681\u001b[0m \u001b[43m \u001b[49m\u001b[43mendpoint\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mendpoint\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 2682\u001b[0m \u001b[43m \u001b[49m\u001b[43mcreate_pr\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mcreate_pr\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 2683\u001b[0m \u001b[43m \u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 2684\u001b[0m \u001b[38;5;28;01mexcept\u001b[39;00m RepositoryNotFoundError \u001b[38;5;28;01mas\u001b[39;00m e:\n\u001b[1;32m 2685\u001b[0m e\u001b[38;5;241m.\u001b[39mappend_to_message(_CREATE_COMMIT_NO_REPO_ERROR_MESSAGE)\n",
56
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py:118\u001b[0m, in \u001b[0;36mvalidate_hf_hub_args.<locals>._inner_fn\u001b[0;34m(*args, **kwargs)\u001b[0m\n\u001b[1;32m 115\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m check_use_auth_token:\n\u001b[1;32m 116\u001b[0m kwargs \u001b[38;5;241m=\u001b[39m smoothly_deprecate_use_auth_token(fn_name\u001b[38;5;241m=\u001b[39mfn\u001b[38;5;241m.\u001b[39m\u001b[38;5;18m__name__\u001b[39m, has_token\u001b[38;5;241m=\u001b[39mhas_token, kwargs\u001b[38;5;241m=\u001b[39mkwargs)\n\u001b[0;32m--> 118\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfn\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n",
57
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/_commit_api.py:483\u001b[0m, in \u001b[0;36mfetch_upload_modes\u001b[0;34m(additions, repo_type, repo_id, token, revision, endpoint, create_pr)\u001b[0m\n\u001b[1;32m 465\u001b[0m payload \u001b[38;5;241m=\u001b[39m {\n\u001b[1;32m 466\u001b[0m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mfiles\u001b[39m\u001b[38;5;124m\"\u001b[39m: [\n\u001b[1;32m 467\u001b[0m {\n\u001b[0;32m (...)\u001b[0m\n\u001b[1;32m 474\u001b[0m ]\n\u001b[1;32m 475\u001b[0m }\n\u001b[1;32m 477\u001b[0m resp \u001b[38;5;241m=\u001b[39m get_session()\u001b[38;5;241m.\u001b[39mpost(\n\u001b[1;32m 478\u001b[0m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mendpoint\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m/api/\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mrepo_type\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124ms/\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mrepo_id\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m/preupload/\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mrevision\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m\"\u001b[39m,\n\u001b[1;32m 479\u001b[0m json\u001b[38;5;241m=\u001b[39mpayload,\n\u001b[1;32m 480\u001b[0m headers\u001b[38;5;241m=\u001b[39mheaders,\n\u001b[1;32m 481\u001b[0m params\u001b[38;5;241m=\u001b[39m{\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mcreate_pr\u001b[39m\u001b[38;5;124m\"\u001b[39m: \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124m1\u001b[39m\u001b[38;5;124m\"\u001b[39m} \u001b[38;5;28;01mif\u001b[39;00m create_pr \u001b[38;5;28;01melse\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m,\n\u001b[1;32m 482\u001b[0m )\n\u001b[0;32m--> 483\u001b[0m \u001b[43mhf_raise_for_status\u001b[49m\u001b[43m(\u001b[49m\u001b[43mresp\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 484\u001b[0m preupload_info \u001b[38;5;241m=\u001b[39m _validate_preupload_info(resp\u001b[38;5;241m.\u001b[39mjson())\n\u001b[1;32m 485\u001b[0m upload_modes\u001b[38;5;241m.\u001b[39mupdate(\u001b[38;5;241m*\u001b[39m\u001b[38;5;241m*\u001b[39m{file[\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mpath\u001b[39m\u001b[38;5;124m\"\u001b[39m]: file[\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124muploadMode\u001b[39m\u001b[38;5;124m\"\u001b[39m] \u001b[38;5;28;01mfor\u001b[39;00m file \u001b[38;5;129;01min\u001b[39;00m preupload_info[\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mfiles\u001b[39m\u001b[38;5;124m\"\u001b[39m]})\n",
58
- "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py:293\u001b[0m, in \u001b[0;36mhf_raise_for_status\u001b[0;34m(response, endpoint_name)\u001b[0m\n\u001b[1;32m 279\u001b[0m \u001b[38;5;28;01melif\u001b[39;00m error_code \u001b[38;5;241m==\u001b[39m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mRepoNotFound\u001b[39m\u001b[38;5;124m\"\u001b[39m \u001b[38;5;129;01mor\u001b[39;00m response\u001b[38;5;241m.\u001b[39mstatus_code \u001b[38;5;241m==\u001b[39m \u001b[38;5;241m401\u001b[39m:\n\u001b[1;32m 280\u001b[0m \u001b[38;5;66;03m# 401 is misleading as it is returned for:\u001b[39;00m\n\u001b[1;32m 281\u001b[0m \u001b[38;5;66;03m# - private and gated repos if user is not authenticated\u001b[39;00m\n\u001b[1;32m 282\u001b[0m \u001b[38;5;66;03m# - missing repos\u001b[39;00m\n\u001b[1;32m 283\u001b[0m \u001b[38;5;66;03m# => for now, we process them as `RepoNotFound` anyway.\u001b[39;00m\n\u001b[1;32m 284\u001b[0m \u001b[38;5;66;03m# See https://gist.github.com/Wauplin/46c27ad266b15998ce56a6603796f0b9\u001b[39;00m\n\u001b[1;32m 285\u001b[0m message \u001b[38;5;241m=\u001b[39m (\n\u001b[1;32m 286\u001b[0m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mresponse\u001b[38;5;241m.\u001b[39mstatus_code\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m Client Error.\u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 287\u001b[0m \u001b[38;5;241m+\u001b[39m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;130;01m\\n\u001b[39;00m\u001b[38;5;130;01m\\n\u001b[39;00m\u001b[38;5;124m\"\u001b[39m\n\u001b[0;32m (...)\u001b[0m\n\u001b[1;32m 291\u001b[0m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124m make sure you are authenticated.\u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 292\u001b[0m )\n\u001b[0;32m--> 293\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m RepositoryNotFoundError(message, response) \u001b[38;5;28;01mfrom\u001b[39;00m \u001b[38;5;21;01me\u001b[39;00m\n\u001b[1;32m 295\u001b[0m \u001b[38;5;28;01melif\u001b[39;00m response\u001b[38;5;241m.\u001b[39mstatus_code \u001b[38;5;241m==\u001b[39m \u001b[38;5;241m400\u001b[39m:\n\u001b[1;32m 296\u001b[0m message \u001b[38;5;241m=\u001b[39m (\n\u001b[1;32m 297\u001b[0m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;130;01m\\n\u001b[39;00m\u001b[38;5;130;01m\\n\u001b[39;00m\u001b[38;5;124mBad request for \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mendpoint_name\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m endpoint:\u001b[39m\u001b[38;5;124m\"\u001b[39m \u001b[38;5;28;01mif\u001b[39;00m endpoint_name \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m \u001b[38;5;28;01melse\u001b[39;00m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;130;01m\\n\u001b[39;00m\u001b[38;5;130;01m\\n\u001b[39;00m\u001b[38;5;124mBad request:\u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 298\u001b[0m )\n",
59
- "\u001b[0;31mRepositoryNotFoundError\u001b[0m: 404 Client Error. (Request ID: Root=1-64be540a-13608ebb4f5e72723d9907a8;bec748ef-4d79-4099-914c-5e806f83f5e0)\n\nRepository Not Found for url: https://huggingface.co/api/spaces/username/my-cool-space/preupload/main.\nPlease make sure you specified the correct `repo_id` and `repo_type`.\nIf you are trying to access a private or gated repo, make sure you are authenticated.\nNote: Creating a commit assumes that the repo already exists on the Huggingface Hub. Please use `create_repo` if it's not the case."
60
- ]
61
- }
62
- ],
63
- "source": [
64
- "from huggingface_hub import HfApi\n",
65
- "api = HfApi()\n",
66
- "\n",
67
- "api.create_repo(\n",
68
- " repo_id=\"osbm/prostate158_nnUNet_results_3d_fullres2\",\n",
69
- " repo_type=\"dataset\",\n",
70
- " exist_ok=True,\n",
71
- " private=False,\n",
72
- ")\n",
73
- "# Upload all the content from the local folder to your remote Space.\n",
74
- "# By default, files are uploaded at the root of the repo\n",
75
- "api.upload_folder(\n",
76
- " folder_path=\"\",\n",
77
- " repo_id=\"username/my-cool-space\",\n",
78
- " repo_type=\"space\",\n",
79
- ")"
80
- ]
81
- },
82
- {
83
- "cell_type": "code",
84
- "execution_count": null,
85
- "id": "9dc0670f-31b2-4b0d-b756-043b798d865e",
86
- "metadata": {},
87
- "outputs": [],
88
- "source": []
89
- }
90
- ],
91
- "metadata": {
92
- "kernelspec": {
93
- "display_name": "Python 3 (ipykernel)",
94
- "language": "python",
95
- "name": "python3"
96
- },
97
- "language_info": {
98
- "codemirror_mode": {
99
- "name": "ipython",
100
- "version": 3
101
- },
102
- "file_extension": ".py",
103
- "mimetype": "text/x-python",
104
- "name": "python",
105
- "nbconvert_exporter": "python",
106
- "pygments_lexer": "ipython3",
107
- "version": "3.10.6"
108
- }
109
- },
110
- "nbformat": 4,
111
- "nbformat_minor": 5
112
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/checkpoint_best.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b1c91cbd76cbc97c1a5c93420cf3496051fab1e0972623ab6a21416aab02550
3
- size 356790417
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/debug.json DELETED
@@ -1,52 +0,0 @@
1
- {
2
- "_best_ema": "None",
3
- "batch_size": "2",
4
- "configuration_manager": "{'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [28, 256, 256], 'median_image_size_in_voxels': [25.0, 270.0, 270.0], 'spacing': [2.999998998641968, 0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [2, 6, 6], 'pool_op_kernel_sizes': [[1, 1, 1], [1, 2, 2], [1, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2], [1, 2, 2]], 'conv_kernel_sizes': [[1, 3, 3], [1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'unet_max_num_features': 320, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': False}",
5
- "configuration_name": "3d_fullres",
6
- "cudnn_version": 8500,
7
- "current_epoch": "0",
8
- "dataloader_train": "<nnunetv2.training.data_augmentation.custom_transforms.limited_length_multithreaded_augmenter.LimitedLenWrapper object at 0x7f974c2f5780>",
9
- "dataloader_train.generator": "<nnunetv2.training.dataloading.data_loader_3d.nnUNetDataLoader3D object at 0x7f974c2f5090>",
10
- "dataloader_train.num_processes": "4",
11
- "dataloader_train.transform": "Compose ( [Convert3DTo2DTransform( apply_to_keys = ('data', 'seg') ), SpatialTransform( independent_scale_for_each_axis = False, p_rot_per_sample = 0.2, p_scale_per_sample = 0.2, p_el_per_sample = 0, data_key = 'data', label_key = 'seg', patch_size = [256, 256], patch_center_dist_from_border = None, do_elastic_deform = False, alpha = (0, 0), sigma = (0, 0), do_rotation = True, angle_x = (-3.141592653589793, 3.141592653589793), angle_y = (0, 0), angle_z = (0, 0), do_scale = True, scale = (0.7, 1.4), border_mode_data = 'constant', border_cval_data = 0, order_data = 3, border_mode_seg = 'constant', border_cval_seg = -1, order_seg = 1, random_crop = False, p_rot_per_axis = 1, p_independent_scale_per_axis = 1 ), Convert2DTo3DTransform( apply_to_keys = ('data', 'seg') ), GaussianNoiseTransform( p_per_sample = 0.1, data_key = 'data', noise_variance = (0, 0.1), p_per_channel = 1, per_channel = False ), GaussianBlurTransform( p_per_sample = 0.2, different_sigma_per_channel = True, p_per_channel = 0.5, data_key = 'data', blur_sigma = (0.5, 1.0), different_sigma_per_axis = False, p_isotropic = 0 ), BrightnessMultiplicativeTransform( p_per_sample = 0.15, data_key = 'data', multiplier_range = (0.75, 1.25), per_channel = True ), ContrastAugmentationTransform( p_per_sample = 0.15, data_key = 'data', contrast_range = (0.75, 1.25), preserve_range = True, per_channel = True, p_per_channel = 1 ), SimulateLowResolutionTransform( order_upsample = 3, order_downsample = 0, channels = None, per_channel = True, p_per_channel = 0.5, p_per_sample = 0.25, data_key = 'data', zoom_range = (0.5, 1), ignore_axes = (0,) ), GammaTransform( p_per_sample = 0.1, retain_stats = True, per_channel = True, data_key = 'data', gamma_range = (0.7, 1.5), invert_image = True ), GammaTransform( p_per_sample = 0.3, retain_stats = True, per_channel = True, data_key = 'data', gamma_range = (0.7, 1.5), invert_image = False ), MirrorTransform( p_per_sample = 1, data_key = 'data', label_key = 'seg', axes = (0, 1, 2) ), RemoveLabelTransform( output_key = 'seg', input_key = 'seg', replace_with = 0, remove_label = -1 ), RenameTransform( delete_old = True, out_key = 'target', in_key = 'seg' ), DownsampleSegForDSTransform2( axes = None, output_key = 'target', input_key = 'target', order = 0, ds_scales = [[1.0, 1.0, 1.0], [1.0, 0.5, 0.5], [1.0, 0.25, 0.25], [0.5, 0.125, 0.125], [0.25, 0.0625, 0.0625], [0.25, 0.03125, 0.03125]] ), NumpyToTensor( keys = ['data', 'target'], cast_to = 'float' )] )",
12
- "dataloader_val": "<nnunetv2.training.data_augmentation.custom_transforms.limited_length_multithreaded_augmenter.LimitedLenWrapper object at 0x7f974c2f4880>",
13
- "dataloader_val.generator": "<nnunetv2.training.dataloading.data_loader_3d.nnUNetDataLoader3D object at 0x7f974c2f4be0>",
14
- "dataloader_val.num_processes": "2",
15
- "dataloader_val.transform": "Compose ( [RemoveLabelTransform( output_key = 'seg', input_key = 'seg', replace_with = 0, remove_label = -1 ), RenameTransform( delete_old = True, out_key = 'target', in_key = 'seg' ), DownsampleSegForDSTransform2( axes = None, output_key = 'target', input_key = 'target', order = 0, ds_scales = [[1.0, 1.0, 1.0], [1.0, 0.5, 0.5], [1.0, 0.25, 0.25], [0.5, 0.125, 0.125], [0.25, 0.0625, 0.0625], [0.25, 0.03125, 0.03125]] ), NumpyToTensor( keys = ['data', 'target'], cast_to = 'float' )] )",
16
- "dataset_json": "{'name': 'Prostate158', 'description': 'Prostate cancer segmentation dataset', 'channel_names': {'0': 'T2', '1': 'ADC', '2': 'DFI'}, 'labels': {'background': 0, 'prostate_inner': 1, 'prostate_outer': 2, 'tumor': 3}, 'numTraining': 139, 'numTest': 19, 'file_ending': '.nii.gz'}",
17
- "device": "cuda:0",
18
- "disable_checkpointing": "False",
19
- "fold": "1",
20
- "folder_with_segs_from_previous_stage": "None",
21
- "gpu_name": "NVIDIA A10G",
22
- "grad_scaler": "<torch.cuda.amp.grad_scaler.GradScaler object at 0x7f974d1d1a50>",
23
- "hostname": "s-osbm-jupyter-f0b83-8689bbb555-5t6kn",
24
- "inference_allowed_mirroring_axes": "(0, 1, 2)",
25
- "initial_lr": "0.01",
26
- "is_cascaded": "False",
27
- "is_ddp": "False",
28
- "label_manager": "<nnunetv2.utilities.label_handling.label_handling.LabelManager object at 0x7f974d1d1c00>",
29
- "local_rank": "0",
30
- "log_file": "nnUNet_results/Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/training_log_2023_7_24_09_31_46.txt",
31
- "logger": "<nnunetv2.training.logging.nnunet_logger.nnUNetLogger object at 0x7f974d1d1ae0>",
32
- "loss": "DeepSupervisionWrapper(\n (loss): DC_and_CE_loss(\n (ce): RobustCrossEntropyLoss()\n (dc): MemoryEfficientSoftDiceLoss()\n )\n)",
33
- "lr_scheduler": "<nnunetv2.training.lr_scheduler.polylr.PolyLRScheduler object at 0x7f974d1d1b40>",
34
- "my_init_kwargs": "{'plans': {'dataset_name': 'Dataset001_Prostate158', 'plans_name': 'nnUNetPlans', 'original_median_spacing_after_transp': [3.0, 0.4017857015132904, 0.4017857015132904], 'original_median_shape_after_transp': [25, 270, 270], 'image_reader_writer': 'SimpleITKIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'configurations': {'2d': {'data_identifier': 'nnUNetPlans_2d', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 31, 'patch_size': [320, 320], 'median_image_size_in_voxels': [270.0, 270.0], 'spacing': [0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [6, 6], 'pool_op_kernel_sizes': [[1, 1], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2]], 'conv_kernel_sizes': [[3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3]], 'unet_max_num_features': 512, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': True}, '3d_fullres': {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [28, 256, 256], 'median_image_size_in_voxels': [25.0, 270.0, 270.0], 'spacing': [2.999998998641968, 0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [2, 6, 6], 'pool_op_kernel_sizes': [[1, 1, 1], [1, 2, 2], [1, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2], [1, 2, 2]], 'conv_kernel_sizes': [[1, 3, 3], [1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'unet_max_num_features': 320, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': False}}, 'experiment_planner_used': 'ExperimentPlanner', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 1694.0, 'mean': 267.35308837890625, 'median': 242.0, 'min': 0.0, 'percentile_00_5': 36.0, 'percentile_99_5': 768.0, 'std': 136.11251831054688}, '1': {'max': 3557.286865234375, 'mean': 1215.81591796875, 'median': 1203.8331298828125, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 2259.82861328125, 'std': 338.6748352050781}, '2': {'max': 198.95455932617188, 'mean': 72.26309204101562, 'median': 70.3214340209961, 'min': 0.0, 'percentile_00_5': 34.534385681152344, 'percentile_99_5': 132.71939086914062, 'std': 18.909290313720703}}}, 'configuration': '3d_fullres', 'fold': 1, 'dataset_json': {'name': 'Prostate158', 'description': 'Prostate cancer segmentation dataset', 'channel_names': {'0': 'T2', '1': 'ADC', '2': 'DFI'}, 'labels': {'background': 0, 'prostate_inner': 1, 'prostate_outer': 2, 'tumor': 3}, 'numTraining': 139, 'numTest': 19, 'file_ending': '.nii.gz'}, 'unpack_dataset': True, 'device': device(type='cuda')}",
35
- "network": "PlainConvUNet",
36
- "num_epochs": "1000",
37
- "num_input_channels": "3",
38
- "num_iterations_per_epoch": "250",
39
- "num_val_iterations_per_epoch": "50",
40
- "optimizer": "SGD (\nParameter Group 0\n dampening: 0\n differentiable: False\n foreach: None\n initial_lr: 0.01\n lr: 0.01\n maximize: False\n momentum: 0.99\n nesterov: True\n weight_decay: 3e-05\n)",
41
- "output_folder": "nnUNet_results/Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1",
42
- "output_folder_base": "nnUNet_results/Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres",
43
- "oversample_foreground_percent": "0.33",
44
- "plans_manager": "{'dataset_name': 'Dataset001_Prostate158', 'plans_name': 'nnUNetPlans', 'original_median_spacing_after_transp': [3.0, 0.4017857015132904, 0.4017857015132904], 'original_median_shape_after_transp': [25, 270, 270], 'image_reader_writer': 'SimpleITKIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'configurations': {'2d': {'data_identifier': 'nnUNetPlans_2d', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 31, 'patch_size': [320, 320], 'median_image_size_in_voxels': [270.0, 270.0], 'spacing': [0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [6, 6], 'pool_op_kernel_sizes': [[1, 1], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2], [2, 2]], 'conv_kernel_sizes': [[3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3], [3, 3]], 'unet_max_num_features': 512, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': True}, '3d_fullres': {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [28, 256, 256], 'median_image_size_in_voxels': [25.0, 270.0, 270.0], 'spacing': [2.999998998641968, 0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [2, 6, 6], 'pool_op_kernel_sizes': [[1, 1, 1], [1, 2, 2], [1, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2], [1, 2, 2]], 'conv_kernel_sizes': [[1, 3, 3], [1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'unet_max_num_features': 320, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': False}}, 'experiment_planner_used': 'ExperimentPlanner', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 1694.0, 'mean': 267.35308837890625, 'median': 242.0, 'min': 0.0, 'percentile_00_5': 36.0, 'percentile_99_5': 768.0, 'std': 136.11251831054688}, '1': {'max': 3557.286865234375, 'mean': 1215.81591796875, 'median': 1203.8331298828125, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 2259.82861328125, 'std': 338.6748352050781}, '2': {'max': 198.95455932617188, 'mean': 72.26309204101562, 'median': 70.3214340209961, 'min': 0.0, 'percentile_00_5': 34.534385681152344, 'percentile_99_5': 132.71939086914062, 'std': 18.909290313720703}}}",
45
- "preprocessed_dataset_folder": "nnUNet_preprocessed/Dataset001_Prostate158/nnUNetPlans_3d_fullres",
46
- "preprocessed_dataset_folder_base": "nnUNet_preprocessed/Dataset001_Prostate158",
47
- "save_every": "50",
48
- "torch_version": "2.0.1+cu117",
49
- "unpack_dataset": "True",
50
- "was_initialized": "True",
51
- "weight_decay": "3e-05"
52
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/progress.png DELETED

Git LFS Details

  • SHA256: 05023ef21b368b31d2fe9a7dc0cbaf5e216e7d6e9e31c2c5237409409428e3e4
  • Pointer size: 131 Bytes
  • Size of remote file: 530 kB
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/fold_1/training_log_2023_7_24_09_31_46.txt DELETED
@@ -1,170 +0,0 @@
1
-
2
- #######################################################################
3
- Please cite the following paper when using nnU-Net:
4
- Isensee, F., Jaeger, P. F., Kohl, S. A., Petersen, J., & Maier-Hein, K. H. (2021). nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nature methods, 18(2), 203-211.
5
- #######################################################################
6
-
7
-
8
- This is the configuration used by this training:
9
- Configuration name: 3d_fullres
10
- {'data_identifier': 'nnUNetPlans_3d_fullres', 'preprocessor_name': 'DefaultPreprocessor', 'batch_size': 2, 'patch_size': [28, 256, 256], 'median_image_size_in_voxels': [25.0, 270.0, 270.0], 'spacing': [2.999998998641968, 0.4017857015132904, 0.4017857015132904], 'normalization_schemes': ['ZScoreNormalization', 'ZScoreNormalization', 'ZScoreNormalization'], 'use_mask_for_norm': [False, False, False], 'UNet_class_name': 'PlainConvUNet', 'UNet_base_num_features': 32, 'n_conv_per_stage_encoder': [2, 2, 2, 2, 2, 2, 2], 'n_conv_per_stage_decoder': [2, 2, 2, 2, 2, 2], 'num_pool_per_axis': [2, 6, 6], 'pool_op_kernel_sizes': [[1, 1, 1], [1, 2, 2], [1, 2, 2], [2, 2, 2], [2, 2, 2], [1, 2, 2], [1, 2, 2]], 'conv_kernel_sizes': [[1, 3, 3], [1, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3], [3, 3, 3]], 'unet_max_num_features': 320, 'resampling_fn_data': 'resample_data_or_seg_to_shape', 'resampling_fn_seg': 'resample_data_or_seg_to_shape', 'resampling_fn_data_kwargs': {'is_seg': False, 'order': 3, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_seg_kwargs': {'is_seg': True, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'resampling_fn_probabilities': 'resample_data_or_seg_to_shape', 'resampling_fn_probabilities_kwargs': {'is_seg': False, 'order': 1, 'order_z': 0, 'force_separate_z': None}, 'batch_dice': False}
11
-
12
- These are the global plan.json settings:
13
- {'dataset_name': 'Dataset001_Prostate158', 'plans_name': 'nnUNetPlans', 'original_median_spacing_after_transp': [3.0, 0.4017857015132904, 0.4017857015132904], 'original_median_shape_after_transp': [25, 270, 270], 'image_reader_writer': 'SimpleITKIO', 'transpose_forward': [0, 1, 2], 'transpose_backward': [0, 1, 2], 'experiment_planner_used': 'ExperimentPlanner', 'label_manager': 'LabelManager', 'foreground_intensity_properties_per_channel': {'0': {'max': 1694.0, 'mean': 267.35308837890625, 'median': 242.0, 'min': 0.0, 'percentile_00_5': 36.0, 'percentile_99_5': 768.0, 'std': 136.11251831054688}, '1': {'max': 3557.286865234375, 'mean': 1215.81591796875, 'median': 1203.8331298828125, 'min': 0.0, 'percentile_00_5': 0.0, 'percentile_99_5': 2259.82861328125, 'std': 338.6748352050781}, '2': {'max': 198.95455932617188, 'mean': 72.26309204101562, 'median': 70.3214340209961, 'min': 0.0, 'percentile_00_5': 34.534385681152344, 'percentile_99_5': 132.71939086914062, 'std': 18.909290313720703}}}
14
-
15
- 2023-07-24 09:31:48.662707: unpacking dataset...
16
- 2023-07-24 09:31:51.254056: unpacking done...
17
- 2023-07-24 09:31:51.254646: do_dummy_2d_data_aug: True
18
- 2023-07-24 09:31:51.255391: Using splits from existing split file: nnUNet_preprocessed/Dataset001_Prostate158/splits_final.json
19
- 2023-07-24 09:31:51.255736: The split file contains 5 splits.
20
- 2023-07-24 09:31:51.255795: Desired fold for training: 1
21
- 2023-07-24 09:31:51.255847: This split has 111 training and 28 validation cases.
22
- 2023-07-24 09:31:55.429484: Unable to plot network architecture:
23
- 2023-07-24 09:31:55.429709: module 'torch.onnx' has no attribute '_optimize_trace'
24
- 2023-07-24 09:31:55.479973:
25
- 2023-07-24 09:31:55.480082: Epoch 0
26
- 2023-07-24 09:31:55.480217: Current learning rate: 0.01
27
- 2023-07-24 09:35:51.149318: train_loss -0.0686
28
- 2023-07-24 09:35:51.149595: val_loss -0.1977
29
- 2023-07-24 09:35:51.149694: Pseudo dice [0.6772, 0.3986, 0.0]
30
- 2023-07-24 09:35:51.149858: Epoch time: 235.67 s
31
- 2023-07-24 09:35:51.149975: Yayy! New best EMA pseudo Dice: 0.3586
32
- 2023-07-24 09:35:52.944725:
33
- 2023-07-24 09:35:52.944855: Epoch 1
34
- 2023-07-24 09:35:52.945019: Current learning rate: 0.00999
35
- 2023-07-24 09:39:26.457915: train_loss -0.2509
36
- 2023-07-24 09:39:26.458209: val_loss -0.2695
37
- 2023-07-24 09:39:26.458309: Pseudo dice [0.7352, 0.5579, 0.0]
38
- 2023-07-24 09:39:26.458386: Epoch time: 213.51 s
39
- 2023-07-24 09:39:26.458445: Yayy! New best EMA pseudo Dice: 0.3658
40
- 2023-07-24 09:39:30.325983:
41
- 2023-07-24 09:39:30.326120: Epoch 2
42
- 2023-07-24 09:39:30.326235: Current learning rate: 0.00998
43
- 2023-07-24 09:42:59.369360: train_loss -0.3283
44
- 2023-07-24 09:42:59.369581: val_loss -0.2917
45
- 2023-07-24 09:42:59.369678: Pseudo dice [0.7702, 0.5453, 0.0]
46
- 2023-07-24 09:42:59.369768: Epoch time: 209.04 s
47
- 2023-07-24 09:42:59.369841: Yayy! New best EMA pseudo Dice: 0.3731
48
- 2023-07-24 09:43:02.898393:
49
- 2023-07-24 09:43:02.898701: Epoch 3
50
- 2023-07-24 09:43:02.898844: Current learning rate: 0.00997
51
- 2023-07-24 09:46:30.423032: train_loss -0.3623
52
- 2023-07-24 09:46:30.423313: val_loss -0.362
53
- 2023-07-24 09:46:30.423406: Pseudo dice [0.8084, 0.6354, 0.0]
54
- 2023-07-24 09:46:30.423570: Epoch time: 207.53 s
55
- 2023-07-24 09:46:30.423706: Yayy! New best EMA pseudo Dice: 0.3839
56
- 2023-07-24 09:46:32.679859:
57
- 2023-07-24 09:46:32.679986: Epoch 4
58
- 2023-07-24 09:46:32.680083: Current learning rate: 0.00996
59
- 2023-07-24 09:50:16.112560: train_loss -0.4055
60
- 2023-07-24 09:50:16.112745: val_loss -0.3814
61
- 2023-07-24 09:50:16.112836: Pseudo dice [0.8423, 0.6152, 0.3433]
62
- 2023-07-24 09:50:16.112920: Epoch time: 223.43 s
63
- 2023-07-24 09:50:16.112987: Yayy! New best EMA pseudo Dice: 0.4056
64
- 2023-07-24 09:50:18.759251:
65
- 2023-07-24 09:50:18.759383: Epoch 5
66
- 2023-07-24 09:50:18.759503: Current learning rate: 0.00995
67
- 2023-07-24 09:53:41.007815: train_loss -0.4426
68
- 2023-07-24 09:53:41.007999: val_loss -0.4302
69
- 2023-07-24 09:53:41.008113: Pseudo dice [0.8359, 0.6439, 0.4462]
70
- 2023-07-24 09:53:41.008215: Epoch time: 202.25 s
71
- 2023-07-24 09:53:41.008303: Yayy! New best EMA pseudo Dice: 0.4292
72
- 2023-07-24 09:53:43.767199:
73
- 2023-07-24 09:53:43.767324: Epoch 6
74
- 2023-07-24 09:53:43.767419: Current learning rate: 0.00995
75
- 2023-07-24 09:57:16.562120: train_loss -0.4301
76
- 2023-07-24 09:57:16.562339: val_loss -0.4084
77
- 2023-07-24 09:57:16.562424: Pseudo dice [0.8011, 0.6376, 0.48]
78
- 2023-07-24 09:57:16.562501: Epoch time: 212.8 s
79
- 2023-07-24 09:57:16.562591: Yayy! New best EMA pseudo Dice: 0.4502
80
- 2023-07-24 09:57:19.469062:
81
- 2023-07-24 09:57:19.469288: Epoch 7
82
- 2023-07-24 09:57:19.469407: Current learning rate: 0.00994
83
- 2023-07-24 10:00:48.194506: train_loss -0.4578
84
- 2023-07-24 10:00:48.194689: val_loss -0.4189
85
- 2023-07-24 10:00:48.194808: Pseudo dice [0.8137, 0.6457, 0.4479]
86
- 2023-07-24 10:00:48.194900: Epoch time: 208.73 s
87
- 2023-07-24 10:00:48.194967: Yayy! New best EMA pseudo Dice: 0.4688
88
- 2023-07-24 10:00:52.373474:
89
- 2023-07-24 10:00:52.373721: Epoch 8
90
- 2023-07-24 10:00:52.373865: Current learning rate: 0.00993
91
- 2023-07-24 10:04:19.517230: train_loss -0.498
92
- 2023-07-24 10:04:19.517421: val_loss -0.4187
93
- 2023-07-24 10:04:19.517504: Pseudo dice [0.8555, 0.6358, 0.3124]
94
- 2023-07-24 10:04:19.517583: Epoch time: 207.15 s
95
- 2023-07-24 10:04:19.517648: Yayy! New best EMA pseudo Dice: 0.482
96
- 2023-07-24 10:04:21.806987:
97
- 2023-07-24 10:04:21.807117: Epoch 9
98
- 2023-07-24 10:04:21.807218: Current learning rate: 0.00992
99
- 2023-07-24 10:07:51.359195: train_loss -0.5027
100
- 2023-07-24 10:07:51.370858: val_loss -0.4306
101
- 2023-07-24 10:07:51.371073: Pseudo dice [0.8521, 0.6531, 0.3801]
102
- 2023-07-24 10:07:51.371159: Epoch time: 209.55 s
103
- 2023-07-24 10:07:51.371223: Yayy! New best EMA pseudo Dice: 0.4967
104
- 2023-07-24 10:07:54.287527:
105
- 2023-07-24 10:07:54.287659: Epoch 10
106
- 2023-07-24 10:07:54.287776: Current learning rate: 0.00991
107
- 2023-07-24 10:11:21.020047: train_loss -0.5007
108
- 2023-07-24 10:11:21.020299: val_loss -0.4555
109
- 2023-07-24 10:11:21.020390: Pseudo dice [0.8525, 0.6574, 0.5038]
110
- 2023-07-24 10:11:21.020547: Epoch time: 206.73 s
111
- 2023-07-24 10:11:21.020615: Yayy! New best EMA pseudo Dice: 0.5141
112
- 2023-07-24 10:11:23.595996:
113
- 2023-07-24 10:11:23.596125: Epoch 11
114
- 2023-07-24 10:11:23.596240: Current learning rate: 0.0099
115
- 2023-07-24 10:14:50.060935: train_loss -0.5114
116
- 2023-07-24 10:14:50.061202: val_loss -0.4435
117
- 2023-07-24 10:14:50.061292: Pseudo dice [0.8367, 0.6786, 0.3972]
118
- 2023-07-24 10:14:50.061454: Epoch time: 206.47 s
119
- 2023-07-24 10:14:50.061521: Yayy! New best EMA pseudo Dice: 0.5265
120
- 2023-07-24 10:14:54.044929:
121
- 2023-07-24 10:14:54.045076: Epoch 12
122
- 2023-07-24 10:14:54.045194: Current learning rate: 0.00989
123
- 2023-07-24 10:18:35.033546: train_loss -0.5262
124
- 2023-07-24 10:18:35.033730: val_loss -0.4482
125
- 2023-07-24 10:18:35.033818: Pseudo dice [0.8485, 0.6596, 0.5738]
126
- 2023-07-24 10:18:35.033900: Epoch time: 220.99 s
127
- 2023-07-24 10:18:35.033965: Yayy! New best EMA pseudo Dice: 0.5432
128
- 2023-07-24 10:18:37.229037:
129
- 2023-07-24 10:18:37.229155: Epoch 13
130
- 2023-07-24 10:18:37.229270: Current learning rate: 0.00988
131
- 2023-07-24 10:22:07.718349: train_loss -0.5297
132
- 2023-07-24 10:22:07.718728: val_loss -0.4473
133
- 2023-07-24 10:22:07.718979: Pseudo dice [0.8557, 0.6552, 0.5333]
134
- 2023-07-24 10:22:07.719130: Epoch time: 210.49 s
135
- 2023-07-24 10:22:07.719215: Yayy! New best EMA pseudo Dice: 0.557
136
- 2023-07-24 10:22:12.377048:
137
- 2023-07-24 10:22:12.377682: Epoch 14
138
- 2023-07-24 10:22:12.377826: Current learning rate: 0.00987
139
- 2023-07-24 10:25:35.042624: train_loss -0.544
140
- 2023-07-24 10:25:35.042880: val_loss -0.4672
141
- 2023-07-24 10:25:35.042980: Pseudo dice [0.8415, 0.6909, 0.5861]
142
- 2023-07-24 10:25:35.043072: Epoch time: 202.67 s
143
- 2023-07-24 10:25:35.043142: Yayy! New best EMA pseudo Dice: 0.5719
144
- 2023-07-24 10:25:39.141425:
145
- 2023-07-24 10:25:39.141566: Epoch 15
146
- 2023-07-24 10:25:39.141699: Current learning rate: 0.00986
147
- 2023-07-24 10:29:17.006086: train_loss -0.536
148
- 2023-07-24 10:29:17.006276: val_loss -0.458
149
- 2023-07-24 10:29:17.006367: Pseudo dice [0.8684, 0.6802, 0.4725]
150
- 2023-07-24 10:29:17.006453: Epoch time: 217.87 s
151
- 2023-07-24 10:29:17.006521: Yayy! New best EMA pseudo Dice: 0.5821
152
- 2023-07-24 10:29:19.086107:
153
- 2023-07-24 10:29:19.086326: Epoch 16
154
- 2023-07-24 10:29:19.086447: Current learning rate: 0.00986
155
- 2023-07-24 10:32:52.937337: train_loss -0.5573
156
- 2023-07-24 10:32:52.937526: val_loss -0.4805
157
- 2023-07-24 10:32:52.937625: Pseudo dice [0.8545, 0.6796, 0.6108]
158
- 2023-07-24 10:32:52.937717: Epoch time: 213.85 s
159
- 2023-07-24 10:32:52.937808: Yayy! New best EMA pseudo Dice: 0.5954
160
- 2023-07-24 10:32:55.486784:
161
- 2023-07-24 10:32:55.486909: Epoch 17
162
- 2023-07-24 10:32:55.487025: Current learning rate: 0.00985
163
- 2023-07-24 10:36:29.203138: train_loss -0.5585
164
- 2023-07-24 10:36:29.203351: val_loss -0.4583
165
- 2023-07-24 10:36:29.203443: Pseudo dice [0.8662, 0.6791, 0.4972]
166
- 2023-07-24 10:36:29.203535: Epoch time: 213.72 s
167
- 2023-07-24 10:36:29.203619: Yayy! New best EMA pseudo Dice: 0.604
168
- 2023-07-24 10:36:34.055434:
169
- 2023-07-24 10:36:34.055742: Epoch 18
170
- 2023-07-24 10:36:34.055855: Current learning rate: 0.00984
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Dataset001_Prostate158/nnUNetTrainer__nnUNetPlans__3d_fullres/plans.json DELETED
@@ -1,342 +0,0 @@
1
- {
2
- "dataset_name": "Dataset001_Prostate158",
3
- "plans_name": "nnUNetPlans",
4
- "original_median_spacing_after_transp": [
5
- 3.0,
6
- 0.4017857015132904,
7
- 0.4017857015132904
8
- ],
9
- "original_median_shape_after_transp": [
10
- 25,
11
- 270,
12
- 270
13
- ],
14
- "image_reader_writer": "SimpleITKIO",
15
- "transpose_forward": [
16
- 0,
17
- 1,
18
- 2
19
- ],
20
- "transpose_backward": [
21
- 0,
22
- 1,
23
- 2
24
- ],
25
- "configurations": {
26
- "2d": {
27
- "data_identifier": "nnUNetPlans_2d",
28
- "preprocessor_name": "DefaultPreprocessor",
29
- "batch_size": 31,
30
- "patch_size": [
31
- 320,
32
- 320
33
- ],
34
- "median_image_size_in_voxels": [
35
- 270.0,
36
- 270.0
37
- ],
38
- "spacing": [
39
- 0.4017857015132904,
40
- 0.4017857015132904
41
- ],
42
- "normalization_schemes": [
43
- "ZScoreNormalization",
44
- "ZScoreNormalization",
45
- "ZScoreNormalization"
46
- ],
47
- "use_mask_for_norm": [
48
- false,
49
- false,
50
- false
51
- ],
52
- "UNet_class_name": "PlainConvUNet",
53
- "UNet_base_num_features": 32,
54
- "n_conv_per_stage_encoder": [
55
- 2,
56
- 2,
57
- 2,
58
- 2,
59
- 2,
60
- 2,
61
- 2
62
- ],
63
- "n_conv_per_stage_decoder": [
64
- 2,
65
- 2,
66
- 2,
67
- 2,
68
- 2,
69
- 2
70
- ],
71
- "num_pool_per_axis": [
72
- 6,
73
- 6
74
- ],
75
- "pool_op_kernel_sizes": [
76
- [
77
- 1,
78
- 1
79
- ],
80
- [
81
- 2,
82
- 2
83
- ],
84
- [
85
- 2,
86
- 2
87
- ],
88
- [
89
- 2,
90
- 2
91
- ],
92
- [
93
- 2,
94
- 2
95
- ],
96
- [
97
- 2,
98
- 2
99
- ],
100
- [
101
- 2,
102
- 2
103
- ]
104
- ],
105
- "conv_kernel_sizes": [
106
- [
107
- 3,
108
- 3
109
- ],
110
- [
111
- 3,
112
- 3
113
- ],
114
- [
115
- 3,
116
- 3
117
- ],
118
- [
119
- 3,
120
- 3
121
- ],
122
- [
123
- 3,
124
- 3
125
- ],
126
- [
127
- 3,
128
- 3
129
- ],
130
- [
131
- 3,
132
- 3
133
- ]
134
- ],
135
- "unet_max_num_features": 512,
136
- "resampling_fn_data": "resample_data_or_seg_to_shape",
137
- "resampling_fn_seg": "resample_data_or_seg_to_shape",
138
- "resampling_fn_data_kwargs": {
139
- "is_seg": false,
140
- "order": 3,
141
- "order_z": 0,
142
- "force_separate_z": null
143
- },
144
- "resampling_fn_seg_kwargs": {
145
- "is_seg": true,
146
- "order": 1,
147
- "order_z": 0,
148
- "force_separate_z": null
149
- },
150
- "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
151
- "resampling_fn_probabilities_kwargs": {
152
- "is_seg": false,
153
- "order": 1,
154
- "order_z": 0,
155
- "force_separate_z": null
156
- },
157
- "batch_dice": true
158
- },
159
- "3d_fullres": {
160
- "data_identifier": "nnUNetPlans_3d_fullres",
161
- "preprocessor_name": "DefaultPreprocessor",
162
- "batch_size": 2,
163
- "patch_size": [
164
- 28,
165
- 256,
166
- 256
167
- ],
168
- "median_image_size_in_voxels": [
169
- 25.0,
170
- 270.0,
171
- 270.0
172
- ],
173
- "spacing": [
174
- 2.999998998641968,
175
- 0.4017857015132904,
176
- 0.4017857015132904
177
- ],
178
- "normalization_schemes": [
179
- "ZScoreNormalization",
180
- "ZScoreNormalization",
181
- "ZScoreNormalization"
182
- ],
183
- "use_mask_for_norm": [
184
- false,
185
- false,
186
- false
187
- ],
188
- "UNet_class_name": "PlainConvUNet",
189
- "UNet_base_num_features": 32,
190
- "n_conv_per_stage_encoder": [
191
- 2,
192
- 2,
193
- 2,
194
- 2,
195
- 2,
196
- 2,
197
- 2
198
- ],
199
- "n_conv_per_stage_decoder": [
200
- 2,
201
- 2,
202
- 2,
203
- 2,
204
- 2,
205
- 2
206
- ],
207
- "num_pool_per_axis": [
208
- 2,
209
- 6,
210
- 6
211
- ],
212
- "pool_op_kernel_sizes": [
213
- [
214
- 1,
215
- 1,
216
- 1
217
- ],
218
- [
219
- 1,
220
- 2,
221
- 2
222
- ],
223
- [
224
- 1,
225
- 2,
226
- 2
227
- ],
228
- [
229
- 2,
230
- 2,
231
- 2
232
- ],
233
- [
234
- 2,
235
- 2,
236
- 2
237
- ],
238
- [
239
- 1,
240
- 2,
241
- 2
242
- ],
243
- [
244
- 1,
245
- 2,
246
- 2
247
- ]
248
- ],
249
- "conv_kernel_sizes": [
250
- [
251
- 1,
252
- 3,
253
- 3
254
- ],
255
- [
256
- 1,
257
- 3,
258
- 3
259
- ],
260
- [
261
- 3,
262
- 3,
263
- 3
264
- ],
265
- [
266
- 3,
267
- 3,
268
- 3
269
- ],
270
- [
271
- 3,
272
- 3,
273
- 3
274
- ],
275
- [
276
- 3,
277
- 3,
278
- 3
279
- ],
280
- [
281
- 3,
282
- 3,
283
- 3
284
- ]
285
- ],
286
- "unet_max_num_features": 320,
287
- "resampling_fn_data": "resample_data_or_seg_to_shape",
288
- "resampling_fn_seg": "resample_data_or_seg_to_shape",
289
- "resampling_fn_data_kwargs": {
290
- "is_seg": false,
291
- "order": 3,
292
- "order_z": 0,
293
- "force_separate_z": null
294
- },
295
- "resampling_fn_seg_kwargs": {
296
- "is_seg": true,
297
- "order": 1,
298
- "order_z": 0,
299
- "force_separate_z": null
300
- },
301
- "resampling_fn_probabilities": "resample_data_or_seg_to_shape",
302
- "resampling_fn_probabilities_kwargs": {
303
- "is_seg": false,
304
- "order": 1,
305
- "order_z": 0,
306
- "force_separate_z": null
307
- },
308
- "batch_dice": false
309
- }
310
- },
311
- "experiment_planner_used": "ExperimentPlanner",
312
- "label_manager": "LabelManager",
313
- "foreground_intensity_properties_per_channel": {
314
- "0": {
315
- "max": 1694.0,
316
- "mean": 267.35308837890625,
317
- "median": 242.0,
318
- "min": 0.0,
319
- "percentile_00_5": 36.0,
320
- "percentile_99_5": 768.0,
321
- "std": 136.11251831054688
322
- },
323
- "1": {
324
- "max": 3557.286865234375,
325
- "mean": 1215.81591796875,
326
- "median": 1203.8331298828125,
327
- "min": 0.0,
328
- "percentile_00_5": 0.0,
329
- "percentile_99_5": 2259.82861328125,
330
- "std": 338.6748352050781
331
- },
332
- "2": {
333
- "max": 198.95455932617188,
334
- "mean": 72.26309204101562,
335
- "median": 70.3214340209961,
336
- "min": 0.0,
337
- "percentile_00_5": 34.534385681152344,
338
- "percentile_99_5": 132.71939086914062,
339
- "std": 18.909290313720703
340
- }
341
- }
342
- }