2024-01-16 15:30:46 +01:00
|
|
|
#from ultralytics import YOLO
|
|
|
|
from flask import request, Flask, jsonify
|
|
|
|
from waitress import serve
|
|
|
|
from PIL import Image
|
|
|
|
import onnxruntime as ort
|
|
|
|
import numpy as np
|
|
|
|
import requests
|
|
|
|
|
|
|
|
#my changes
|
|
|
|
import os
|
|
|
|
|
|
|
|
script_dir = os.path.dirname(os.path.abspath(__file__))
|
|
|
|
# Change the working directory to the script's directory
|
|
|
|
os.chdir(script_dir)
|
|
|
|
|
|
|
|
yolo_classes = ["b_fully_ripened",
|
|
|
|
"b_half_ripened",
|
|
|
|
"b_green",
|
|
|
|
"l_fully_ripened",
|
|
|
|
"l_half_ripened",
|
|
|
|
"l_green"
|
|
|
|
]
|
|
|
|
|
|
|
|
#app start
|
|
|
|
|
|
|
|
app = Flask(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
@app.route("/")
|
|
|
|
def root():
|
|
|
|
"""
|
|
|
|
Site main page handler function.
|
|
|
|
:return: Content of index.html file
|
|
|
|
"""
|
|
|
|
with open("index.html") as file:
|
|
|
|
return file.read()
|
|
|
|
|
|
|
|
@app.route("/detect", methods=["POST"])
|
|
|
|
def detect():
|
|
|
|
buf = request.files["image_file"]
|
|
|
|
crop_type = request.form.get("cropType")
|
|
|
|
location = request.form.get("location")
|
|
|
|
variety = request.form.get("variety")
|
|
|
|
boxes, orientation = detect_objects_on_image(buf.stream)
|
|
|
|
# Do something with crop_type, location, and variety here (e.g., store in database)
|
|
|
|
|
|
|
|
return jsonify(boxes)
|
|
|
|
|
|
|
|
|
|
|
|
def detect_objects_on_image(buf):
|
|
|
|
input, img_width, img_height = prepare_input(buf)
|
|
|
|
output = run_model(input)
|
|
|
|
orientation = get_orientation(buf)
|
|
|
|
processed_output = process_output(output, img_width, img_height, orientation)
|
|
|
|
return processed_output, orientation
|
|
|
|
|
|
|
|
def prepare_input(buf):
|
|
|
|
img = Image.open(buf)
|
|
|
|
img_width, img_height = img.size
|
|
|
|
img = img.resize((640, 640))
|
|
|
|
img = img.convert("RGB")
|
|
|
|
input = np.array(img)
|
|
|
|
input = input.transpose(2, 0, 1)
|
|
|
|
input = input.reshape(1, 3, 640, 640) / 255.0
|
|
|
|
return input.astype(np.float32), img_width, img_height
|
|
|
|
|
|
|
|
def run_model(input):
|
|
|
|
model = ort.InferenceSession("best.onnx", providers=['CPUExecutionProvider'])
|
|
|
|
outputs = model.run(["output0"], {"images":input})
|
|
|
|
return outputs[0]
|
|
|
|
|
|
|
|
def process_output(output, img_width, img_height, orientation):
|
|
|
|
output = output[0].astype(float)
|
|
|
|
output = output.transpose()
|
|
|
|
|
|
|
|
boxes = []
|
|
|
|
for row in output:
|
|
|
|
prob = row[4:].max()
|
|
|
|
if prob < 0.5:
|
|
|
|
continue
|
|
|
|
|
|
|
|
class_id = row[4:].argmax()
|
|
|
|
label = yolo_classes[class_id]
|
|
|
|
xc, yc, w, h = row[:4]
|
|
|
|
x1 = (xc - w/2) / 640 * img_width
|
|
|
|
y1 = (yc - h/2) / 640 * img_height
|
|
|
|
x2 = (xc + w/2) / 640 * img_width
|
|
|
|
y2 = (yc + h/2) / 640 * img_height
|
|
|
|
|
|
|
|
boxes.append([x1, y1, x2, y2, label, prob])
|
|
|
|
|
|
|
|
# Adjust boxes based on orientation
|
|
|
|
adjusted_boxes = adjust_boxes_for_orientation(boxes, orientation, img_width, img_height)
|
|
|
|
|
|
|
|
# Sort and apply non-max suppression as before
|
|
|
|
adjusted_boxes.sort(key=lambda x: x[5], reverse=True)
|
|
|
|
result = []
|
|
|
|
while len(adjusted_boxes) > 0:
|
|
|
|
result.append(adjusted_boxes[0])
|
|
|
|
adjusted_boxes = [box for box in adjusted_boxes if iou(box, adjusted_boxes[0]) < 0.7]
|
|
|
|
|
|
|
|
return result
|
|
|
|
|
|
|
|
|
|
|
|
def iou(box1,box2):
|
|
|
|
return intersection(box1,box2)/union(box1,box2)
|
|
|
|
|
|
|
|
def union(box1,box2):
|
|
|
|
box1_x1,box1_y1,box1_x2,box1_y2 = box1[:4]
|
|
|
|
box2_x1,box2_y1,box2_x2,box2_y2 = box2[:4]
|
|
|
|
box1_area = (box1_x2-box1_x1)*(box1_y2-box1_y1)
|
|
|
|
box2_area = (box2_x2-box2_x1)*(box2_y2-box2_y1)
|
|
|
|
return box1_area + box2_area - intersection(box1,box2)
|
|
|
|
|
|
|
|
def intersection(box1,box2):
|
|
|
|
box1_x1,box1_y1,box1_x2,box1_y2 = box1[:4]
|
|
|
|
box2_x1,box2_y1,box2_x2,box2_y2 = box2[:4]
|
|
|
|
x1 = max(box1_x1,box2_x1)
|
|
|
|
y1 = max(box1_y1,box2_y1)
|
|
|
|
x2 = min(box1_x2,box2_x2)
|
|
|
|
y2 = min(box1_y2,box2_y2)
|
|
|
|
return (x2-x1)*(y2-y1)
|
|
|
|
|
|
|
|
def get_orientation(image_path):
|
|
|
|
with Image.open(image_path) as img:
|
|
|
|
if hasattr(img, '_getexif'):
|
|
|
|
exif_data = img._getexif()
|
|
|
|
if exif_data is not None:
|
|
|
|
return exif_data.get(274, 1) # Default to normal orientation
|
|
|
|
return 1 # Default orientation if no EXIF data
|
|
|
|
|
|
|
|
def adjust_boxes_for_orientation(boxes, orientation, img_width, img_height):
|
|
|
|
adjusted_boxes = []
|
|
|
|
for box in boxes:
|
|
|
|
x1, y1, x2, y2, label, prob = box
|
|
|
|
|
|
|
|
# Apply transformations based on orientation
|
|
|
|
if orientation == 3: # 180 degrees
|
|
|
|
x1, y1, x2, y2 = img_width - x2, img_height - y2, img_width - x1, img_height - y1
|
|
|
|
elif orientation == 6: # 270 degrees (or -90 degrees)
|
|
|
|
x1, y1, x2, y2 = img_height - y2, x1, img_height - y1, x2
|
|
|
|
elif orientation == 8: # 90 degrees
|
|
|
|
x1, y1, x2, y2 = y1, img_width - x2, y2, img_width - x1
|
|
|
|
|
|
|
|
adjusted_boxes.append([x1, y1, x2, y2, label, prob])
|
|
|
|
|
|
|
|
return adjusted_boxes
|
|
|
|
|
|
|
|
|
|
|
|
""" def detect_objects_on_image(buf):
|
|
|
|
"""
|
|
|
|
""""
|
|
|
|
Function receives an image,
|
|
|
|
passes it through YOLOv8 neural network
|
|
|
|
and returns an array of detected objects
|
|
|
|
and their bounding boxes
|
|
|
|
:param buf: Input image file stream
|
|
|
|
:return: Array of bounding boxes in format [[x1,y1,x2,y2,object_type,probability],..]
|
|
|
|
"""
|
|
|
|
"""""
|
|
|
|
model = YOLO("best.pt")
|
|
|
|
results = model.predict(Image.open(buf))
|
|
|
|
result = results[0]
|
|
|
|
output = []
|
|
|
|
for box in result.boxes:
|
|
|
|
x1, y1, x2, y2 = [
|
|
|
|
round(x) for x in box.xyxy[0].tolist()
|
|
|
|
]
|
|
|
|
class_id = box.cls[0].item()
|
|
|
|
prob = round(box.conf[0].item(), 2)
|
|
|
|
output.append([
|
|
|
|
x1, y1, x2, y2, result.names[class_id], prob
|
|
|
|
])
|
|
|
|
return output
|
|
|
|
|
|
|
|
"""
|
|
|
|
serve(app, host='0.0.0.0', port=8080)
|