train.lua 4.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152
  1. require './lib/portable'
  2. require 'optim'
  3. require 'xlua'
  4. require 'pl'
  5. local settings = require './lib/settings'
  6. local minibatch_adam = require './lib/minibatch_adam'
  7. local iproc = require './lib/iproc'
  8. local reconstruct = require './lib/reconstruct'
  9. local pairwise_transform = require './lib/pairwise_transform'
  10. local image_loader = require './lib/image_loader'
  11. local function save_test_scale(model, rgb, file)
  12. local up = reconstruct.scale(model, settings.scale, rgb, settings.block_offset)
  13. image.save(file, up)
  14. end
  15. local function save_test_jpeg(model, rgb, file)
  16. local im, count = reconstruct.image(model, rgb, settings.block_offset)
  17. image.save(file, im)
  18. end
  19. local function split_data(x, test_size)
  20. local index = torch.randperm(#x)
  21. local train_size = #x - test_size
  22. local train_x = {}
  23. local valid_x = {}
  24. for i = 1, train_size do
  25. train_x[i] = x[index[i]]
  26. end
  27. for i = 1, test_size do
  28. valid_x[i] = x[index[train_size + i]]
  29. end
  30. return train_x, valid_x
  31. end
  32. local function make_validation_set(x, transformer, n)
  33. n = n or 4
  34. local data = {}
  35. for i = 1, #x do
  36. for k = 1, n do
  37. local x, y = transformer(x[i], true)
  38. table.insert(data, {x = x:reshape(1, x:size(1), x:size(2), x:size(3)),
  39. y = y:reshape(1, y:size(1), y:size(2), y:size(3))})
  40. end
  41. xlua.progress(i, #x)
  42. collectgarbage()
  43. end
  44. return data
  45. end
  46. local function validate(model, criterion, data)
  47. local loss = 0
  48. for i = 1, #data do
  49. local z = model:forward(data[i].x:cuda())
  50. loss = loss + criterion:forward(z, data[i].y:cuda())
  51. xlua.progress(i, #data)
  52. if i % 10 == 0 then
  53. collectgarbage()
  54. end
  55. end
  56. return loss / #data
  57. end
  58. local function train()
  59. local model, offset = settings.create_model()
  60. assert(offset == settings.block_offset)
  61. local criterion = nn.MSECriterion():cuda()
  62. local x = torch.load(settings.images)
  63. local lrd_count = 0
  64. local train_x, valid_x = split_data(x,
  65. math.floor(settings.validation_ratio * #x),
  66. settings.validation_crops)
  67. local test = image_loader.load_float(settings.test)
  68. local adam_config = {
  69. learningRate = settings.learning_rate,
  70. xBatchSize = settings.batch_size,
  71. }
  72. local transformer = function(x, is_validation)
  73. if is_validation == nil then is_validation = false end
  74. if settings.method == "scale" then
  75. return pairwise_transform.scale(x,
  76. settings.scale,
  77. settings.crop_size, offset,
  78. { color_augment = not is_validation,
  79. random_half = settings.random_half})
  80. elseif settings.method == "noise" then
  81. return pairwise_transform.jpeg(x,
  82. settings.noise_level,
  83. settings.crop_size, offset,
  84. { color_augment = not is_validation,
  85. random_half = settings.random_half})
  86. elseif settings.method == "noise_scale" then
  87. return pairwise_transform.jpeg_scale(x,
  88. settings.scale,
  89. settings.noise_level,
  90. settings.crop_size, offset,
  91. { color_augment = not is_validation,
  92. random_half = settings.random_half
  93. })
  94. end
  95. end
  96. local best_score = 100000.0
  97. print("# make validation-set")
  98. local valid_xy = make_validation_set(valid_x, transformer, 20)
  99. valid_x = nil
  100. collectgarbage()
  101. model:cuda()
  102. print("load .. " .. #train_x)
  103. for epoch = 1, settings.epoch do
  104. model:training()
  105. print("# " .. epoch)
  106. print(minibatch_adam(model, criterion, train_x, adam_config,
  107. transformer,
  108. {1, settings.crop_size, settings.crop_size},
  109. {1, settings.crop_size - offset * 2, settings.crop_size - offset * 2}
  110. ))
  111. model:evaluate()
  112. print("# validation")
  113. local score = validate(model, criterion, valid_xy)
  114. if score < best_score then
  115. lrd_count = 0
  116. best_score = score
  117. print("* update best model")
  118. torch.save(settings.model_file, model)
  119. if settings.method == "noise" then
  120. local log = path.join(settings.model_dir,
  121. ("noise%d_best.png"):format(settings.noise_level))
  122. save_test_jpeg(model, test, log)
  123. elseif settings.method == "scale" then
  124. local log = path.join(settings.model_dir,
  125. ("scale%.1f_best.png"):format(settings.scale))
  126. save_test_scale(model, test, log)
  127. elseif settings.method == "noise_scale" then
  128. local log = path.join(settings.model_dir,
  129. ("noise%d_scale%.1f_best.png"):format(settings.noise_level,
  130. settings.scale))
  131. save_test_scale(model, test, log)
  132. end
  133. else
  134. lrd_count = lrd_count + 1
  135. if lrd_count > 5 then
  136. lrd_count = 0
  137. adam_config.learningRate = adam_config.learningRate * 0.8
  138. print("* learning rate decay: " .. adam_config.learningRate)
  139. end
  140. end
  141. print("current: " .. score .. ", best: " .. best_score)
  142. collectgarbage()
  143. end
  144. end
  145. torch.manualSeed(settings.seed)
  146. cutorch.manualSeed(settings.seed)
  147. print(settings)
  148. train()