No Description

xword.py 8.9KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262
  1. import math
  2. import cv2
  3. import numpy as np
  4. import copy
  5. import argparse
  6. def non_greys_to_white(img, threshold=48):
  7. b, g, r = cv2.split(img)
  8. rgb_diff = cv2.subtract(cv2.max(cv2.max(b, g), r), cv2.min(cv2.min(b, g), r))
  9. filtered = img.copy()
  10. filtered[np.where(rgb_diff > threshold)] = (255, 255, 255)
  11. return filtered
  12. def load_image_as_greyscale(file_name, filter_colours, colour_filter_threshold):
  13. img = cv2.imread(file_name)
  14. if img is None:
  15. raise RuntimeError("Failed to load image")
  16. if filter_colours:
  17. img = non_greys_to_white(img, colour_filter_threshold)
  18. return cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
  19. def preprocess_image(original, gaussian_blur_size, adaptive_threshold_block_size, adaptive_threshold_mean_adjustment, num_dilations):
  20. img = cv2.GaussianBlur(original, (gaussian_blur_size, gaussian_blur_size), 0)
  21. img = cv2.adaptiveThreshold(img, 255, cv2.ADAPTIVE_THRESH_GAUSSIAN_C, cv2.THRESH_BINARY_INV, adaptive_threshold_block_size, adaptive_threshold_mean_adjustment)
  22. kernel = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]], np.uint8)
  23. for i in range(num_dilations):
  24. img = cv2.dilate(img, kernel)
  25. return img
  26. def morph_open_image(img, kernel_size, iterations=1):
  27. kernel = cv2.getStructuringElement(cv2.MORPH_RECT, kernel_size)
  28. return cv2.morphologyEx(img, cv2.MORPH_OPEN, kernel, iterations=iterations)
  29. def get_fundamental_frequency(fft):
  30. mag = abs(fft[0:len(fft) // 2])
  31. mag[0] = 0
  32. return int(np.argmax(mag))
  33. def get_line_fft(img, line_detector_element_size, axis):
  34. lines = morph_open_image(img, (line_detector_element_size, 1) if axis == 1 else (1, line_detector_element_size))
  35. return np.fft.fft(np.sum(lines, axis=axis))
  36. def get_line_frequency(img, line_detector_element_size, axis):
  37. return get_fundamental_frequency(get_line_fft(img, line_detector_element_size, axis))
  38. def find_biggest_contour(img):
  39. contours, hierarchy = cv2.findContours(img, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE)
  40. biggest = None
  41. max_area = 0
  42. for contour in contours:
  43. area = cv2.contourArea(contour)
  44. if area > max_area:
  45. biggest = contour
  46. max_area = area
  47. return biggest
  48. def erode_contour(img_shape, contour, erosion_kernel_size, iterations):
  49. contour_img = np.zeros(img_shape, dtype=np.uint8)
  50. cv2.drawContours(contour_img, [contour], 0, 255, -1)
  51. contour_img = morph_open_image(contour_img, (erosion_kernel_size, erosion_kernel_size), iterations)
  52. return find_biggest_contour(contour_img)
  53. def get_contour_corners(img, contour):
  54. height, width = img.shape
  55. top_left = [width, height]
  56. top_right = [-1, height]
  57. bottom_left = [width, -1]
  58. bottom_right = [-1, -1]
  59. for vertex in contour:
  60. point = vertex[0]
  61. sum = point[0] + point[1]
  62. diff = point[0] - point[1]
  63. if sum < top_left[0] + top_left[1]:
  64. top_left = point
  65. if sum > bottom_right[0] + bottom_right[1]:
  66. bottom_right = point
  67. if diff < bottom_left[0] - bottom_left[1]:
  68. bottom_left = point
  69. if diff > top_right[0] - top_right[1]:
  70. top_right = point
  71. return top_left, top_right, bottom_right, bottom_left
  72. def segment_length(p1, p2):
  73. dx = p1[0] - p2[0]
  74. dy = p1[1] - p2[1]
  75. return math.sqrt(dx ** 2 + dy ** 2)
  76. def get_longest_side(poly):
  77. previous = poly[-1]
  78. max = 0
  79. for current in poly:
  80. len = segment_length(previous, current)
  81. if len > max:
  82. max = len
  83. previous = current
  84. return max
  85. def extract_square(img, top_left, top_right, bottom_right, bottom_left):
  86. src = [top_left, top_right, bottom_right, bottom_left]
  87. longest = get_longest_side(src)
  88. dst = [[0, 0], [longest - 1, 0], [longest - 1, longest - 1], [0, longest - 1]]
  89. m = cv2.getPerspectiveTransform(np.array(src, dtype=np.float32), np.array(dst, dtype=np.float32))
  90. return cv2.warpPerspective(img, m, (int(longest), int(longest)))
  91. def get_threshold_from_quantile(img, quantile):
  92. height, width = img.shape
  93. num_pixels = height * width
  94. pixels = np.sort(np.reshape(img, num_pixels))
  95. return pixels[int(num_pixels * quantile)]
  96. def extract_grid_colours(img, num_rows, num_cols, sampling_block_size_ratio):
  97. height, width = img.shape
  98. row_delta = int(height * sampling_block_size_ratio / num_rows / 2)
  99. col_delta = int(width * sampling_block_size_ratio / num_cols / 2)
  100. sampling_block_area = (2 * row_delta + 1) * (2 * col_delta + 1)
  101. grid = []
  102. for row in range(num_rows):
  103. line = []
  104. y = int(((row + 0.5) / num_rows) * height)
  105. for col in range(num_cols):
  106. sum = 0
  107. x = int(((col + 0.5) / num_cols) * width)
  108. for dy in range(-row_delta, row_delta + 1):
  109. for dx in range(-col_delta, col_delta + 1):
  110. sum += img[y + dy, x + dx]
  111. line.append(sum / sampling_block_area)
  112. grid.append(line)
  113. return grid
  114. def grid_colours_to_blocks(grid_colours, num_rows, num_cols, sampling_threshold):
  115. grid = copy.deepcopy(grid_colours)
  116. warning = False
  117. for row in range(round(num_rows / 2)):
  118. for col in range(num_cols):
  119. row2 = num_rows - row - 1
  120. col2 = num_cols - col - 1
  121. delta1 = grid_colours[row][col] - sampling_threshold
  122. delta2 = grid_colours[row2][col2] - sampling_threshold
  123. if (delta1 > 0) and (delta2 > 0):
  124. block = 0
  125. elif (delta1 < 0) and (delta2 < 0):
  126. block = 1
  127. else:
  128. warning = True
  129. if abs(delta1) > abs(delta2):
  130. block = 1 if delta1 < 0 else 0
  131. else:
  132. block = 1 if delta2 < 0 else 0
  133. grid[row][col] = grid[row2][col2] = block
  134. return warning, grid
  135. def draw_point(image, point, colour):
  136. height, width, _ = image.shape
  137. for dx in range(-10, 11):
  138. for dy in range(-10, 11):
  139. x = point[0] + dx
  140. y = point[1] + dy
  141. if (x >= 0) and (y >= 0) and (x < width) and (y < height):
  142. image[y, x] = colour
  143. def show_image(image):
  144. cv2.namedWindow('xword', cv2.WINDOW_NORMAL)
  145. cv2.imshow('xword', image)
  146. while cv2.waitKey() & 0xFF != ord('q'):
  147. pass
  148. cv2.destroyAllWindows()
  149. def extract_crossword(
  150. file_name,
  151. filter_colours=False,
  152. colour_filter_threshold=48,
  153. gaussian_blur_size=11,
  154. adaptive_threshold_block_size=11,
  155. adaptive_threshold_mean_adjustment=2,
  156. square=True,
  157. num_dilations=1,
  158. contour_erosion_kernel_size=5,
  159. contour_erosion_iterations=6,
  160. line_detector_element_size=51,
  161. sampling_block_size_ratio=0.25,
  162. sampling_threshold_quantile=0.3,
  163. sampling_threshold=None,
  164. grid_line_thickness=4,
  165. grid_square_size=64,
  166. grid_border_size=20,
  167. ):
  168. warnings = []
  169. original = load_image_as_greyscale(file_name, filter_colours, colour_filter_threshold)
  170. img = preprocess_image(original, gaussian_blur_size, adaptive_threshold_block_size, adaptive_threshold_mean_adjustment, num_dilations)
  171. biggest = find_biggest_contour(img)
  172. biggest = erode_contour(img.shape, biggest, contour_erosion_kernel_size, contour_erosion_iterations)
  173. top_left, top_right, bottom_right, bottom_left = get_contour_corners(img, biggest)
  174. img = extract_square(img, top_left, top_right, bottom_right, bottom_left)
  175. num_rows = get_line_frequency(img, line_detector_element_size, 1)
  176. num_cols = get_line_frequency(img, line_detector_element_size, 0)
  177. if square and (num_rows != num_cols):
  178. warnings.append("Crossword is not square")
  179. block_img = extract_square(original, top_left, top_right, bottom_right, bottom_left)
  180. if sampling_threshold is None:
  181. sampling_threshold = get_threshold_from_quantile(block_img, sampling_threshold_quantile)
  182. else:
  183. sampling_threshold = sampling_threshold
  184. grid_colours = extract_grid_colours(block_img, num_rows, num_cols, sampling_block_size_ratio)
  185. warning, grid = grid_colours_to_blocks(grid_colours, num_rows, num_cols, sampling_threshold)
  186. if warning:
  187. warnings.append("Some blocks may be the wrong colour")
  188. step = grid_square_size + grid_line_thickness
  189. grid_height = num_rows * step + grid_line_thickness
  190. grid_width = num_cols * step + grid_line_thickness
  191. output = np.full([2 * grid_border_size + grid_height, 2 * grid_border_size + grid_width], 255, dtype=np.uint8)
  192. cv2.rectangle(output, (grid_border_size, grid_border_size), (grid_border_size + grid_width - 1, grid_border_size + grid_height - 1), 0, -1)
  193. for row in range(num_rows):
  194. y = row * step + grid_line_thickness + grid_border_size
  195. for col in range(num_cols):
  196. if grid[row][col] == 0:
  197. x = col * step + grid_line_thickness + grid_border_size
  198. cv2.rectangle(output, (x, y), (x + grid_square_size - 1, y + grid_square_size - 1), 255, -1)
  199. _, png = cv2.imencode('.png', output)
  200. return png.tobytes(), warnings