Railway Track Crack Detection with Python Computer vision πŸš†πŸš§

parmarjatin4911@gmail.com - Jan 11 - - Dev Community

πŸš‚πŸ‘€ Detecting Railway Track Cracks with Python & Computer Vision! πŸ•΅οΈβ€β™‚οΈπŸ” Transforming images into actionable insights using OpenCV. πŸ“ΈπŸ”§ Enhance rail safety and reliability. πŸ›€οΈπŸ’»

Implementation code

Now, let's dive into the implementation. Railway Track Detection Crack Track Railway Detection with Python Computer vision the OpenCV library. This assumes that you have prior knowledge of setting up Python and installing the necessary libraries:

Run the following commands to install required libraries:

pip install opencv-python && pip install argparse && pip install numpy && pip install scipy

main.py

import cv2
import numpy as np
import argparse
from scipy.special import comb

args setting

parser = argparse.ArgumentParser(description='Process some integers.')
parser.add_argument('-i', "--input", help="input file video")
parser.add_argument('--leftPoint', type=int, help="Left rail offset", default=450)
parser.add_argument('--rightPoint', type=int, help="Right rail offset", default=840)
parser.add_argument('--topPoint', type=int, help="Top rail offset", default=330)
args = parser.parse_args()

def main():

load video class

cap = VideoCapture(args.input)

desired_width = 1920 # Your desired width
desired_height = 1080 # Your desired height

cap = cv2.VideoCapture(args.input)
cap.set(cv2.CAP_PROP_FRAME_WIDTH, desired_width)
cap.set(cv2.CAP_PROP_FRAME_HEIGHT, desired_height)

initialization for line detection

expt_startLeft = args.leftPoint
expt_startRight = args.rightPoint
expt_startTop = args.topPoint

value initialize

left_maxpoint = [0] * 50
right_maxpoint = [195] * 50

convolution filter

kernel = np.array([
[-1, 1, 0, 1, -1],
[-1, 1, 0, 1, -1],
[-1, 1, 0, 1, -1],
[-1, 1, 0, 1, -1],
[-1, 1, 0, 1, -1]
])

Next frame availability

r = True
first = True

while r is True:
r, frame = cap.read()
if frame is None:
break

cut away invalid frame area

valid_frame = frame[expt_startTop:, expt_startLeft:expt_startRight]
original_frame = valid_frame.copy()

gray scale transform

gray_frame = cv2.cvtColor(valid_frame, cv2.COLOR_BGR2GRAY)

histogram equalization image

histeqaul_frame = cv2.equalizeHist(gray_frame)

apply gaussian blur

blur_frame = cv2.GaussianBlur(histeqaul_frame, (5, 5), 5)

merge current frame and last frame

if first is True:
merge_frame = blur_frame
first = False
old_valid_frame = merge_frame.copy()
else:
merge_frame = cv2.addWeighted(blur_frame, 0.2, old_valid_frame, 0.8, 0)
old_valid_frame = merge_frame.copy()

convolution filter

conv_frame = cv2.filter2D(merge_frame, -1, kernel)

initialization for sliding window property

sliding_window = [20, 190, 200, 370]
slide_interval = 15
slide_height = 15
slide_width = 60

initialization for bezier curve variables

left_points = []
right_points = []

define count value

count = 0
for i in range(340, 40, -slide_interval):

get edges in sliding window

left_edge = conv_frame[i:i + slide_height, sliding_window[0]:sliding_window[1]].sum(axis=0)
right_edge = conv_frame[i:i + slide_height, sliding_window[2]:sliding_window[3]].sum(axis=0)

left railroad line processing

if left_edge.argmax() > 0:
left_maxindex = sliding_window[0] + left_edge.argmax()
left_maxpoint[count] = left_maxindex
cv2.line(valid_frame, (left_maxindex, i + int(slide_height / 2)),
(left_maxindex, i + int(slide_height / 2)), (255, 255, 255), 5, cv2.LINE_AA)
left_points.append([left_maxindex, i + int(slide_height / 2)])
sliding_window[0] = max(0, left_maxindex - int(slide_width / 4 + (slide_width + 10) / (count + 1)))
sliding_window[1] = min(390, left_maxindex + int(slide_width / 4 + (slide_width + 10) / (count + 1)))

cv2.rectangle(valid_frame, (sliding_window[0], i + slide_height), (sliding_window[1], i), (0, 255, 0),1)

Add text for left railroad line

Add more accurate and attractive text for the left railroad line

cv2.putText(valid_frame, f'Left Railroad Line: {left_maxindex:.2f}', (10, 10),
cv2.FONT_HERSHEY_SIMPLEX, 0.8, (255, 255, 255), 2, cv2.LINE_AA)

right railroad line processing

if right_edge.argmax() > 0:
right_maxindex = sliding_window[2] + right_edge.argmax()
right_maxpoint[count] = right_maxindex
cv2.line(valid_frame, (right_maxindex, i + int(slide_height / 2)),
(right_maxindex, i + int(slide_height / 2)), (255, 255, 255), 5, cv2.LINE_AA)
right_points.append([right_maxindex, i + int(slide_height / 2)])
sliding_window[2] = max(0, right_maxindex - int(slide_width / 4 + (slide_width + 10) / (count + 1)))
sliding_window[3] = min(390, right_maxindex + int(slide_width / 4 + (slide_width + 10) / (count + 1)))

cv2.rectangle(valid_frame, (sliding_window[2], i + slide_height), (sliding_window[3], i), (0, 0, 255),1)

Add more accurate and attractive text for the right railroad line

cv2.putText(valid_frame, f'Left Railroad Line: {left_maxindex:.2f}', (10, 10),
cv2.FONT_HERSHEY_SIMPLEX, 0.8, (255, 255, 255), 2, cv2.LINE_AA)

count += 1

bezier curve process

bezier_left_xval, bezier_left_yval = bezier_curve(left_points, 50)
bezier_right_xval, bezier_right_yval = bezier_curve(right_points, 50)

bezier_left_points = []
bezier_right_points = []
try:
old_point = (bezier_left_xval[0], bezier_left_yval[0])
for point in zip(bezier_left_xval, bezier_left_yval):
cv2.line(valid_frame, old_point, point, (0, 0, 255), 2, cv2.LINE_AA)
old_point = point
bezier_left_points.append(point)

old_point = (bezier_right_xval[0], bezier_right_yval[0])
for point in zip(bezier_right_xval, bezier_right_yval):
cv2.line(valid_frame, old_point, point, (255, 0, 0), 2, cv2.LINE_AA)
old_point = point
bezier_right_points.append(point)
except IndexError:
pass

Assuming you have defined font properties earlier in your code

Assuming you have defined font properties earlier in your code

Assuming you have defined font properties earlier in your code

font = cv2.FONT_HERSHEY_SIMPLEX
font_scale = 0.5
font_thickness = 1
font_color = (255, 255, 255) # White color

...

cv2.imshow('frame', np.vstack([
np.hstack([
cv2.putText(valid_frame, 'Railway is Going Clear', (40, 20), font, font_scale, font_color, font_thickness),
cv2.putText(original_frame, 'Original Frame', (10, 20), font, font_scale, font_color, font_thickness),
cv2.putText(cv2.cvtColor(histeqaul_frame, cv2.COLOR_GRAY2BGR), 'Histeqaul Frame', (10, 20), font, font_scale, font_color, font_thickness),
]),
np.hstack([
cv2.putText(cv2.cvtColor(blur_frame, cv2.COLOR_GRAY2BGR), 'Blur Frame', (10, 20), font, font_scale, font_color, font_thickness),
cv2.putText(cv2.cvtColor(merge_frame, cv2.COLOR_GRAY2BGR), 'Merge Frame', (10, 20), font, font_scale, font_color, font_thickness),
cv2.putText(cv2.cvtColor(conv_frame, cv2.COLOR_GRAY2BGR), 'Conv Frame', (5, 10), font, font_scale, font_color, font_thickness),
])
]))

Adding text indicating "Railway is Going Clear"

cv2.putText(frame, 'Railway is Going Clear', (5, 5), font, font_scale, font_color, font_thickness)

Display the frame

cv2.imshow('frame', frame)

cv2.imshow('Video', valid_frame)
cv2.waitKey(1)
print('finish')

class for reading video

class VideoCapture:
def init(self, path):

Using OpenCV to capture from device 0. If you have trouble capturing

from a webcam, comment the line below out and use a video file

instead.

self.video = cv2.VideoCapture(path)

If you decide to use video.mp4, you must have this file in the folder

as the main.py.

self.video = cv2.VideoCapture('video.mp4')

def del(self):
self.video.release()

def read(self):

Grab a single frame of video

ret, frame = self.video.read()
return frame is not None, frame

bezier curve function

def bezier_curve(points, ntimes=1000):
"""
Given a set of control points, return the
bezier curve defined by the control points.

points should be a list of lists, or list of tuples
such as [ [1,1],
[2,3],
[4,5], ..[Xn, Yn] ]
ntimes is the number of time steps, defaults to 1000

See http://processingjs.nihongoresources.com/bezierinfo/
"""

def bernstein_poly(i, n, t):
"""
The Bernstein polynomial of n, i as a function of t
"""
return comb(n, i) * (t ** (n - i)) * (1 - t) ** i

nPoints = len(points)
xPoints = np.array([p[0] for p in points])
yPoints = np.array([p[1] for p in points])

t = np.linspace(0.0, 1.0, ntimes)

polynomial_array = np.array([bernstein_poly(i, nPoints - 1, t) for i in range(0, nPoints)])

xvals = np.dot(xPoints, polynomial_array)
yvals = np.dot(yPoints, polynomial_array)
return xvals.astype('int32'), yvals.astype('int32')

def nothing(value):
pass

if name == 'main':
main()

test.py

import the necessary packages

from collections import deque
import numpy as np
import argparse
import imutils
import cv2
import time
import math
import scipy.ndimage
from matplotlib import pyplot as plt

construct the argument parse and parse the arguments

font = cv2.FONT_HERSHEY_SIMPLEX

def orientated_non_max_suppression(mag, ang):
ang_quant = np.round(ang / (np.pi/4)) % 4
winE = np.array([[0, 0, 0], [1, 1, 1], [0, 0, 0]])
winSE = np.array([[1, 0, 0], [0, 1, 0], [0, 0, 1]])
winS = np.array([[0, 1, 0], [0, 1, 0], [0, 1, 0]])
winSW = np.array([[0, 0, 1], [0, 1, 0], [1, 0, 0]])

magE = non_max_suppression(mag, winE)
magSE = non_max_suppression(mag, winSE)
magS = non_max_suppression(mag, winS)
magSW = non_max_suppression(mag, winSW)

mag[ang_quant == 0] = magE[ang_quant == 0]
mag[ang_quant == 1] = magSE[ang_quant == 1]
mag[ang_quant == 2] = magS[ang_quant == 2]
mag[ang_quant == 3] = magSW[ang_quant == 3]
return mag

def non_max_suppression(data, win):
data_max = scipy.ndimage.filters.maximum_filter(
data, footprint=win, mode='constant')
data_max[data != data_max] = 0
return data_max

ap = argparse.ArgumentParser()
ap.add_argument("-v", "--video",
help="path to the (optional) video file")
ap.add_argument("-b", "--buffer", type=int, default=64,
help="max buffer size")
args = vars(ap.parse_args())

define the lower and upper boundaries of the "green"

ball in the HSV color space, then initialize the

list of tracked points

[0, 0, 0, 87, 255, 222]

greenLower = (0, 0, 0)
greenUpper = (87, 255, 222)
pts = deque(maxlen=args["buffer"])

if a video path was not supplied, grab the reference

to the webcam

if not args.get("video", False):
camera = cv2.VideoCapture("crac.mp4")

otherwise, grab a reference to the video file

else:
camera = cv2.VideoCapture(args["video"])

keep looping

while True:

grab the current frame

(grabbed, frame) = camera.read()

if we are viewing a video and we did not grab a frame,

then we have reached the end of the video

if args.get("video") and not grabbed:
break
frame = imutils.resize(frame, width=800)
gray_image = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)
img = frame
with_nmsup = True # apply non-maximal suppression
fudgefactor = 1.5 # with this threshold you can play a little bit
sigma = 50 # for Gaussian Kernel
kernel = 2*math.ceil(2*sigma)+1 # Kernel size

gray_image = gray_image/255.0
blur = cv2.GaussianBlur(gray_image, (kernel, kernel), sigma)
gray_image = cv2.subtract(gray_image, blur)

csudo raspi-configompute sobel response

sobelx = cv2.Sobel(gray_image, cv2.CV_64F, 1, 0, ksize=3)
sobely = cv2.Sobel(gray_image, cv2.CV_64F, 0, 1, ksize=3)

mag = np.hypot(sobelx, sobely)
ang = np.arctan2(sobely, sobelx)

threshold

threshold = 4 * fudgefactor * np.mean(mag)
mag[mag < threshold] = 0

either get edges directly

if with_nmsup is False:
mag = cv2.normalize(mag, 0, 255, cv2.NORM_MINMAX)
kernel = np.ones((5, 5), np.uint8)
result = cv2.morphologyEx(mag, cv2.MORPH_CLOSE, kernel)
num_white = np.sum(result == 255)
num_black = np.sum(result == 0)
print(num_white)
print(num_black)
print((num_white/num_black)*100)

print(result)

cv2.imshow('im', img)
cv2.imshow('im2', result)

cv2.waitKey()

or apply a non-maximal suppression

else:

non-maximal suppression

mag = orientated_non_max_suppression(mag, ang)

create mask

mag[mag > 0] = 255
mag = mag.astype(np.uint8)

kernel = np.ones((5, 5), np.uint8)
result = cv2.morphologyEx(mag, cv2.MORPH_CLOSE, kernel)
num_white = np.sum(result == 255)
num_black = np.sum(result == 0)
print(num_white)
print(num_black)
ratio = (num_white/num_black)*100
print(ratio)
if ratio > 0.7:
print("Cracked")
cv2.putText(img, 'Cracked', (0, 30),
font, 0.8, (0, 0, 255), 2, cv2.LINE_AA)

else:
cv2.putText(img, 'Not Cracked', (0, 30),
font, 0.8, (0, 0, 255), 2, cv2.LINE_AA)
print("Not Cracked")

plt.hist(result.ravel(), 256, [0, 256])

print(result)

cv2.imshow('im', img)
cv2.imshow('im2', result)

plt.show()

key = cv2.waitKey(1) & 0xFF

if the 'q' key is pressed, stop the loop

if key == ord("q"):
break

cleanup the camera and close any open windows

camera.release()
cv2.destroyAllWindows()

Save the above script as a Python file (e.g: run python main.py ) and script is save test.py which is crack detection and run it on your os system,result will show from the machine, performs analyze track detection , and displays the feedback.

Conclusion

In conclusion, leveraging Python and Computer Vision for Railway Track Detection and Crack Identification is a promising approach to enhance rail safety. By employing OpenCV, we can efficiently process images, detect railway tracks, and pinpoint potential cracks. This technology holds immense potential for optimizing maintenance processes and ensuring the integrity of rail infrastructure. Implementing such solutions is crucial for proactive maintenance, reducing downtime, and ultimately enhancing the safety and reliability of railway systems. As we advance in the field of computer vision, we open up new possibilities for automation and real-time monitoring, contributing to the efficiency and resilience of our rail networks. Embracing innovation in railway track inspection is a key step toward building smarter, safer, and more sustainable transportation systems. πŸš„πŸŒ #RailTech #Innovation #SafetyFirst πŸ›€οΈπŸ”

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Terabox Video Player