Jump into digital image structures and learn to process them! Extract data, transform and analyze images using NumPy and Scikit-image. With just a few lines of code, you will convert RGB images to grayscale, get data from them, obtain histograms containing very useful information, and separate objects from the background! This is the Summary of lecture "Image Processing in Python", via datacamp.
import numpy as np
import matplotlib.pyplot as plt
Whats the main difference between the images shown below?
These images have been preloaded as coffee_image and coins_image from the scikit-image data module using:
coffee_image = data.coffee()
coins_image = data.coins()
def show_image(image, title='Image', cmap_type='gray'):
plt.imshow(image, cmap=cmap_type)
plt.title(title)
plt.axis('off')
from skimage import data
coffee_image = data.coffee()
coins_image = data.coins()
coffee_image.shape
(400, 600, 3)
coins_image.shape
(303, 384)
In this exercise you will load an image from scikit-image module data and make it grayscale, then compare both of them in the output.
from skimage import data, color
# Load the rocket image
rocket = data.rocket()
# Convert the image to grayscale
gray_scaled_rocket = color.rgb2gray(rocket)
# Show the original image
show_image(rocket, 'Original RGB image');
# Show the grayscale image
show_image(gray_scaled_rocket, 'Grayscale image')
As a prank, someone has turned an image from a photo album of a trip to Seville upside-down and back-to-front! Now, we need to straighten the image, by flipping it.
Using the NumPy methods learned in the course, flip the image horizontally and vertically. Then display the corrected image using the show_image()
function.
flipped_seville = plt.imread('./dataset/sevilleup.jpg')
# Flip the image vertically
seville_vertical_flip = np.flipud(flipped_seville)
# Flip the previous image horizontally
seville_horizontal_flip = np.fliplr(seville_vertical_flip)
# Show the resulting image
show_image(seville_horizontal_flip, 'Seville')
In this exercise, you will analyze the amount of red in the image. To do this, the histogram of the red channel will be computed for the image shown below: Extracting information from images is a fundamental part of image enhancement. This way you can balance the red and blue to make the image look colder or warmer.
You will use hist()
to display the 256 different intensities of the red color. And ravel()
to make these color values an array of one flat dimension.
image = plt.imread('./dataset/portrait.png')
# Obtain the red channel
red_channel = image[:, :, 0]
# Plot the the red histogram with bins in a range of 256
plt.hist(red_channel.ravel(), bins=256, color='red');
# Set title
plt.title('Red Histogram');
With this histogram we see that the image is quite reddish, meaning it has a sensation of warmness. This is because it has a wide and large distribution of bright red pixels, from 0 to around 150.
Apply global thresholding In this exercise, you'll transform a photograph to binary so you can separate the foreground from the background.
To do so, you need to import the required modules, load the image, obtain the optimal thresh value using threshold_otsu()
and apply it to the image.
You'll see the resulting binarized image when using the show_image()
function, previously explained.
from skimage.filters import threshold_otsu
chess_pieces_image = plt.imread('./dataset/bw.jpg')
# Make the image grayscale using rgb2gray
chess_pieces_image_gray = color.rgb2gray(chess_pieces_image)
# Obtain the optimal threshold value with otsu
thresh = threshold_otsu(chess_pieces_image_gray)
# Apply thresholding to the image
binary = chess_pieces_image_gray > thresh
# Show the image
show_image(binary, 'Binary image')
Sometimes, it isn't that obvious to identify the background. If the image background is relatively uniform, then you can use a global threshold value as we practiced before, using threshold_otsu()
. However, if there's uneven background illumination, adaptive thresholding threshold_local()
(a.k.a. local thresholding) may produce better results.
In this exercise, you will compare both types of thresholding methods (global and local), to find the optimal way to obtain the binary image we need.
page_image = plt.imread('./dataset/text_page.png')
# Make the image grayscale using rgb2gray
page_image = color.rgb2gray(page_image)
# Obtain the optimal otsu global thresh value
global_thresh = threshold_otsu(page_image)
# Obtain the binary image by applying global thresholding
binary_global = page_image > global_thresh
# Show the binary image obtained
show_image(binary_global, 'Global thresholding')
from skimage.filters import threshold_local
# Set the block size to 35
block_size = 35
# Obtain the optimal local thresholding
local_thresh = threshold_local(page_image, block_size, offset=0.1)
# Obtain the binary image by applying local thresholding
binary_local = page_image > local_thresh
# Show the binary image
show_image(binary_local, 'Local thresholding')
As we saw in the video, not being sure about what thresholding method to use isn't a problem. In fact, scikit-image provides us with a function to check multiple methods and see for ourselves what the best option is. It returns a figure comparing the outputs of different global thresholding methods.
from skimage.filters import try_all_threshold
fruits_image = plt.imread('./dataset/fruits-2.jpg')
# Turn the fruits_image to grayscale
grayscale = color.rgb2gray(fruits_image)
# Use the try all method on the resulting grayscale image
fig, ax = try_all_threshold(grayscale, verbose=False);
In this exercise, you will decide what type of thresholding is best used to binarize an image of knitting and craft tools. In doing so, you will be able to see the shapes of the objects, from paper hearts to scissors more clearly.
What type of thresholding would you use judging by the characteristics of the image? Is the background illumination and intensity even or uneven?
tools_image = plt.imread('./dataset/shapes52.jpg')
# Turn the image grayscale
gray_tools_image = color.rgb2gray(tools_image)
# Obtain the optimal thresh
thresh = threshold_otsu(gray_tools_image)
# Obtain the binary image by applying thresholding
binary_image = gray_tools_image > thresh
# Show the resulting binary image
show_image(binary_image, 'Binarized image')
By using a global thresholding method, you obtained the precise binarized image. If you would have used local instead nothing would have been segmented.