import os
import cv2
import numpy as np
import matplotlib.pyplot as plt
import glob
from zipfile import ZipFile
from urllib.request import urlretrieve
%matplotlib inline
Image Alignment
In this page we will take the original form, and a filled out copy of the same form, but is laid out on a table and a picture is taken of it at an angle. Our job is to take that image of the form and find enough keypoints to align it to match the original form
Homography
A homography transforms a square to arbitrary quad.
- Images of two planes are already related by a Homography
- We need 4 corresponding points to estimate Homography
Setup
.
Download Assets
def download_and_unzip(url, save_path):
print(f"Downloading and extracting assests....", end="")
# Downloading zip file using urllib package.
urlretrieve(url, save_path)
try:
# Extracting zip file using the zipfile package.
with ZipFile(save_path) as z:
# Extract ZIP file contents in the same directory.
0])
z.extractall(os.path.split(save_path)[
print("Done")
except Exception as e:
print("\nInvalid file.", e)
= r"https://www.dropbox.com/s/zuwnn6rqe0f4zgh/opencv_bootcamp_assets_NB8.zip?dl=1"
URL
= os.path.join(os.getcwd(), "opencv_bootcamp_assets_NB8.zip")
asset_zip_path
# Download if assest ZIP does not exists.
if not os.path.exists(asset_zip_path):
download_and_unzip(URL, asset_zip_path)
Read Image
# Read reference image
#Remember that opencv uses BGR instead of RGB so if we plot the image before converting over to RGB you'll see what it looks like below
= "form.jpg"
refFilename print("Reading reference image:", refFilename)
= cv2.imread(refFilename, cv2.IMREAD_COLOR)
im_1 = cv2.cvtColor(im_1, cv2.COLOR_BGR2RGB)
im1
# Read image to be aligned
= "scanned-form.jpg"
imFilename print("Reading image to align:", imFilename)
= cv2.imread(imFilename, cv2.IMREAD_COLOR)
im2 = cv2.cvtColor(im2, cv2.COLOR_BGR2RGB) im2
# Display both images, note the subplot() has arguments of: rows, columns, index number
=[20, 10]);
plt.figure(figsize131); plt.axis('off'); plt.imshow(im_1); plt.title("Before Conversion")
plt.subplot(132); plt.axis('off'); plt.imshow(im1); plt.title("Original Form")
plt.subplot(133); plt.axis('off'); plt.imshow(im2); plt.title("Scanned Form") plt.subplot(
Find Keypoints
Note that the more keypoints we can find the more accurate our transformation will be.
- First we will convert both images to grayscale because the code requires grayscale images
- Create an orb class, which is extracting features from images, features that relate to that specific image itself
- Use the orb object to detect keypoints and descripters for each image
- Keypoints are associated with sharp edges and shapes and there is a list of descriptors associated with each keypoint which is a vector which describes the region around that keypoint
- So we can use the descriptors of two edges to compare them
- Keypoints are the red circle, along with their centers and the descriptors describe the patch where that keypoint is located
- List of keypoints for fig 1 & 2 are overlapping, and so we will try to find the one in common that match so we can align the new image
# Convert images to grayscale
= cv2.cvtColor(im1, cv2.COLOR_BGR2GRAY)
im1_gray = cv2.cvtColor(im2, cv2.COLOR_BGR2GRAY)
im2_gray
# Detect ORB features and compute descriptors.
= 500
MAX_NUM_FEATURES = cv2.ORB_create(MAX_NUM_FEATURES)
orb = orb.detectAndCompute(im1_gray, None)
keypoints1, descriptors1 = orb.detectAndCompute(im2_gray, None)
keypoints2, descriptors2
# Display
= cv2.drawKeypoints(im1, keypoints1, outImage=np.array([]),
im1_display =(255, 0, 0), flags=cv2.DRAW_MATCHES_FLAGS_DRAW_RICH_KEYPOINTS)
color
= cv2.drawKeypoints(im2, keypoints2, outImage=np.array([]),
im2_display =(255, 0, 0), flags=cv2.DRAW_MATCHES_FLAGS_DRAW_RICH_KEYPOINTS) color
=[20,10])
plt.figure(figsize121); plt.axis('off'); plt.imshow(im1_display); plt.title("Original Form");
plt.subplot(122); plt.axis('off'); plt.imshow(im2_display); plt.title("Scanned Form"); plt.subplot(
Match Keypoints
# Match features.
= cv2.DescriptorMatcher_create(cv2.DESCRIPTOR_MATCHER_BRUTEFORCE_HAMMING)
matcher
# Converting to list for sorting as tuples are immutable objects.
= list(matcher.match(descriptors1, descriptors2, None))
matches
# Sort matches by score
=lambda x: x.distance, reverse=False)
matches.sort(key
# Remove not so good matches to only the top 10%
= int(len(matches) * 0.1)
numGoodMatches = matches[:numGoodMatches] matches
# Draw top matches between the two images. If the descriptors match close enough a keypoint match is drawn. There are some false positive but we still have a large number of true matches to accomplish the conversion
= cv2.drawMatches(im1, keypoints1, im2, keypoints2, matches, None)
im_matches
=[40, 10])
plt.figure(figsize;plt.axis("off");plt.title("Original Form") plt.imshow(im_matches)
Find Homography
# Extract location of good matches
= np.zeros((len(matches), 2), dtype=np.float32)
points1 = np.zeros((len(matches), 2), dtype=np.float32)
points2
for i, match in enumerate(matches):
= keypoints1[match.queryIdx].pt
points1[i, :] = keypoints2[match.trainIdx].pt
points2[i, :]
# Find homography using the RANSAC algorithm
= cv2.findHomography(points2, points1, cv2.RANSAC) h, mask
Warp Image
# Pass the homography to warp image
= im1.shape
height, width, channels = cv2.warpPerspective(im2, h, (width, height))
im2_reg
# Display results
=[20, 10])
plt.figure(figsize121);plt.imshow(im1); plt.axis("off");plt.title("Original Form")
plt.subplot(122);plt.imshow(im2_reg);plt.axis("off");plt.title("Scanned Form") plt.subplot(
So as you can see we’ve changed the outlook of the uploaded image, and now we can process the form automatically
Panorama
It is advisable to use the same light setting, and a fixed camera position to take the multitude of images to stitch to form a panorama view. This is similar to image alignment above except that we will use the stitching function once we find the keypoints.’
Import Assets
= r"https://www.dropbox.com/s/0o5yqql1ynx31bi/opencv_bootcamp_assets_NB9.zip?dl=1"
URL
= os.path.join(os.getcwd(), "opencv_bootcamp_assets_NB9.zip")
asset_zip_path
# Download if assest ZIP does not exists.
if not os.path.exists(asset_zip_path):
download_and_unzip(URL, asset_zip_path)
Here are the steps needed to create a panorama view:
- Find keypoints in all images
- Find pairwise correspondences
- Estimate pairwise Homographies
- Refine Homographies
- Stitch with Blending
# Read Images
= glob.glob(f"boat{os.sep}*")
imagefiles
imagefiles.sort()
= []
images for filename in imagefiles:
= cv2.imread(filename)
img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)
img
images.append(img)
= len(images) num_images
import math
# Display Images
=[30, 10])
plt.figure(figsize= 3
num_cols = math.ceil(num_images / num_cols)
num_rows for i in range(0, num_images):
+ 1)
plt.subplot(num_rows, num_cols, i "off")
plt.axis( plt.imshow(images[i])
# Stitch Images using sticher class
= cv2.Stitcher_create()
stitcher = stitcher.stitch(images)
status, result
if status == 0:
=[30, 10])
plt.figure(figsize plt.imshow(result)