引入yolo监狱识别代码
This commit is contained in:
153
npu_yolo_onnx_person_car_phone.py
Normal file
153
npu_yolo_onnx_person_car_phone.py
Normal file
@@ -0,0 +1,153 @@
|
|||||||
|
# 文件名: npu_yolo_onnx.py
|
||||||
|
import cv2
|
||||||
|
import numpy as np
|
||||||
|
import onnxruntime as ort
|
||||||
|
import os
|
||||||
|
import time
|
||||||
|
|
||||||
|
def letterbox(img, new_shape=(640, 640), color=(114, 114, 114)):
|
||||||
|
shape = img.shape[:2] # h, w
|
||||||
|
r = min(new_shape[0] / shape[0], new_shape[1] / shape[1])
|
||||||
|
new_unpad = int(round(shape[1] * r)), int(round(shape[0] * r))
|
||||||
|
dw, dh = new_shape[1] - new_unpad[0], new_shape[0] - new_unpad[1]
|
||||||
|
dw /= 2
|
||||||
|
dh /= 2
|
||||||
|
if shape[::-1] != new_unpad:
|
||||||
|
img = cv2.resize(img, new_unpad, interpolation=cv2.INTER_LINEAR)
|
||||||
|
top, bottom = int(round(dh - 0.1)), int(round(dh + 0.1))
|
||||||
|
left, right = int(round(dw - 0.1)), int(round(dw + 0.1))
|
||||||
|
img = cv2.copyMakeBorder(img, top, bottom, left, right, cv2.BORDER_CONSTANT, value=color)
|
||||||
|
return img, r, (dw, dh)
|
||||||
|
|
||||||
|
class YOLOv8_ONNX:
|
||||||
|
def __init__(self, onnx_path, conf_threshold=0.25, iou_threshold=0.45, input_size=640):
|
||||||
|
providers = [("CANNExecutionProvider", {
|
||||||
|
"device_id": 0,
|
||||||
|
"arena_extend_strategy": "kNextPowerOfTwo",
|
||||||
|
"npu_mem_limit": 16 * 1024 * 1024 * 1024,
|
||||||
|
"precision_mode": "allow_fp32_to_fp16",
|
||||||
|
"op_select_impl_mode": "high_precision",
|
||||||
|
"enable_cann_graph": True,
|
||||||
|
}),
|
||||||
|
"CUDAExecutionProvider",
|
||||||
|
"CPUExecutionProvider",
|
||||||
|
]
|
||||||
|
|
||||||
|
self.session = ort.InferenceSession(onnx_path, providers=providers)
|
||||||
|
actual_providers = self.session.get_providers()
|
||||||
|
print("YOLO Providers:", actual_providers)
|
||||||
|
|
||||||
|
if "CANNExecutionProvider" in actual_providers:
|
||||||
|
print("[INFO] YOLO 使用 CANNExecutionProvider(昇腾 NPU)")
|
||||||
|
elif 'CUDAExecutionProvider' in actual_providers:
|
||||||
|
print("[INFO] YOLO 使用 CUDAExecutionProvider(NVIDIA GPU)")
|
||||||
|
else:
|
||||||
|
print("[INFO] YOLO 使用 CPUExecutionProvider")
|
||||||
|
|
||||||
|
self.conf_threshold = conf_threshold
|
||||||
|
self.iou_threshold = iou_threshold
|
||||||
|
self.input_name = self.session.get_inputs()[0].name
|
||||||
|
self.input_size = (input_size, input_size) if isinstance(input_size, int) else input_size
|
||||||
|
|
||||||
|
print(f"模型输入名称: {self.input_name}")
|
||||||
|
print(f"模型输入形状: {self.session.get_inputs()[0].shape}")
|
||||||
|
print(f"模型输出形状: {self.session.get_outputs()[0].shape}")
|
||||||
|
|
||||||
|
def preprocess(self, img):
|
||||||
|
self.orig_shape = img.shape[:2]
|
||||||
|
img, self.ratio, (self.dw, self.dh) = letterbox(img, self.input_size)
|
||||||
|
img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)
|
||||||
|
img = img.transpose(2, 0, 1).astype(np.float32)
|
||||||
|
img /= 255.0
|
||||||
|
img = np.expand_dims(img, axis=0)
|
||||||
|
return img
|
||||||
|
|
||||||
|
def postprocess(self, pred, im0_shape):
|
||||||
|
# 1. 转置:从 [1, 4+cls, 8400] -> [8400, 4+cls]
|
||||||
|
pred = pred[0].T
|
||||||
|
|
||||||
|
# 2. 获取数据
|
||||||
|
boxes = pred[:, :4] # cx, cy, w, h
|
||||||
|
scores = pred[:, 4:]
|
||||||
|
|
||||||
|
# 3. 获取最大置信度和类别
|
||||||
|
conf = np.max(scores, axis=1)
|
||||||
|
class_pred = np.argmax(scores, axis=1)
|
||||||
|
|
||||||
|
# 4. 初步过滤
|
||||||
|
mask = conf > self.conf_threshold
|
||||||
|
if not mask.any():
|
||||||
|
return []
|
||||||
|
|
||||||
|
boxes = boxes[mask]
|
||||||
|
conf = conf[mask]
|
||||||
|
class_pred = class_pred[mask]
|
||||||
|
|
||||||
|
# =========================================================
|
||||||
|
# 还原坐标 (逆 Letterbox)
|
||||||
|
# =========================================================
|
||||||
|
boxes[:, 0] = (boxes[:, 0] - self.dw) / self.ratio # cx
|
||||||
|
boxes[:, 1] = (boxes[:, 1] - self.dh) / self.ratio # cy
|
||||||
|
boxes[:, 2] = boxes[:, 2] / self.ratio # w
|
||||||
|
boxes[:, 3] = boxes[:, 3] / self.ratio # h
|
||||||
|
|
||||||
|
# 转换格式:Center(cx,cy) -> TopLeft(x,y)
|
||||||
|
x = boxes[:, 0] - boxes[:, 2] / 2
|
||||||
|
y = boxes[:, 1] - boxes[:, 3] / 2
|
||||||
|
w = boxes[:, 2]
|
||||||
|
h = boxes[:, 3]
|
||||||
|
|
||||||
|
# 原始框(用于最终输出)
|
||||||
|
bboxes_original = np.stack([x, y, w, h], axis=1)
|
||||||
|
|
||||||
|
# =========================================================
|
||||||
|
# 【核心修复】:Class-Aware NMS (偏移量法)
|
||||||
|
# 给不同类别的框增加不同的偏移量,使得不同类别的框绝对不会重叠
|
||||||
|
# 从而避免 "车" 把 "人" 过滤掉的情况
|
||||||
|
# =========================================================
|
||||||
|
max_wh = 4096 # 只要大于图片最大分辨率即可
|
||||||
|
class_offset = class_pred * max_wh
|
||||||
|
|
||||||
|
# NMS 专用的框坐标 (加上了偏移量)
|
||||||
|
bboxes_for_nms = bboxes_original.copy()
|
||||||
|
bboxes_for_nms[:, 0] += class_offset
|
||||||
|
bboxes_for_nms[:, 1] += class_offset
|
||||||
|
|
||||||
|
# =========================================================
|
||||||
|
# 执行 NMS
|
||||||
|
# =========================================================
|
||||||
|
indices = cv2.dnn.NMSBoxes(
|
||||||
|
bboxes_for_nms.tolist(),
|
||||||
|
conf.tolist(),
|
||||||
|
self.conf_threshold,
|
||||||
|
self.iou_threshold
|
||||||
|
)
|
||||||
|
|
||||||
|
result = []
|
||||||
|
if len(indices) > 0:
|
||||||
|
indices = indices.flatten()
|
||||||
|
for i in indices:
|
||||||
|
# 注意:这里取数据要从 bboxes_original 取 (没有加偏移量的)
|
||||||
|
bx, by, bw, bh = bboxes_original[i]
|
||||||
|
|
||||||
|
# 转换回 x1, y1, x2, y2 供业务代码画图使用
|
||||||
|
x1 = np.clip(bx, 0, im0_shape[1])
|
||||||
|
y1 = np.clip(by, 0, im0_shape[0])
|
||||||
|
x2 = np.clip(bx + bw, 0, im0_shape[1])
|
||||||
|
y2 = np.clip(by + bh, 0, im0_shape[0])
|
||||||
|
|
||||||
|
result.append([
|
||||||
|
float(x1),
|
||||||
|
float(y1),
|
||||||
|
float(x2),
|
||||||
|
float(y2),
|
||||||
|
float(conf[i]),
|
||||||
|
int(class_pred[i])
|
||||||
|
])
|
||||||
|
return result
|
||||||
|
|
||||||
|
def __call__(self, frame):
|
||||||
|
input_data = self.preprocess(frame)
|
||||||
|
pred = self.session.run(None, {self.input_name: input_data})[0]
|
||||||
|
results = self.postprocess(pred, frame.shape[:2])
|
||||||
|
return results
|
||||||
@@ -4,7 +4,8 @@ RTSP 服务模块 - 简洁版本,直接使用原始服务
|
|||||||
|
|
||||||
import threading
|
import threading
|
||||||
# from rtsp_service_ws_1217 import RTSPService
|
# from rtsp_service_ws_1217 import RTSPService
|
||||||
from rtsp_service_ws_prison import RTSPService
|
# from rtsp_service_ws_prison import RTSPService
|
||||||
|
from rtsp_service_ws_0108 import RTSPService
|
||||||
|
|
||||||
|
|
||||||
class SimpleRTSPServer:
|
class SimpleRTSPServer:
|
||||||
|
|||||||
644
rtsp_service_ws_0108.py
Normal file
644
rtsp_service_ws_0108.py
Normal file
@@ -0,0 +1,644 @@
|
|||||||
|
# rtsp_service_kadian.py
|
||||||
|
# 融合 Kadian_Detect_1221.py + rtsp_service_ws.py
|
||||||
|
# 支持多路RTSP、抽帧、分段保存MP4、WebSocket推送图像与告警
|
||||||
|
|
||||||
|
import cv2
|
||||||
|
import numpy as np
|
||||||
|
import os
|
||||||
|
import time
|
||||||
|
import threading
|
||||||
|
import queue
|
||||||
|
import yaml
|
||||||
|
import json
|
||||||
|
import base64
|
||||||
|
import asyncio
|
||||||
|
import websockets
|
||||||
|
from dataclasses import dataclass
|
||||||
|
from typing import Dict, Any, Tuple
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
|
# 导入人脸识别算法
|
||||||
|
try:
|
||||||
|
from api.routes.algorithm_router import video_face_prison_biz
|
||||||
|
|
||||||
|
print("[INFO] 成功导入人脸识别算法")
|
||||||
|
except Exception as e:
|
||||||
|
print(f"[WARN] 无法导入人脸识别算法: {e}")
|
||||||
|
|
||||||
|
# -------------------------- Kadian 检测相关导入 --------------------------
|
||||||
|
from npu_yolo_onnx_person_car_phone import YOLOv8_ONNX # 主检测模型(人/车/后备箱/手机)
|
||||||
|
|
||||||
|
from yolox.tracker.byte_tracker import BYTETracker
|
||||||
|
|
||||||
|
|
||||||
|
# ========================= 配置区 =========================
|
||||||
|
# Kadian 模型路径与ROI(可根据实际情况修改)
|
||||||
|
police_prisoner_model_path = 'YOLO_Weight/prisoner_model.onnx'
|
||||||
|
|
||||||
|
FACE_RECOGNITION_ENABLED = True # 是否启用人脸识别
|
||||||
|
|
||||||
|
# 输入尺寸
|
||||||
|
police_prisoner_input_size = 1280
|
||||||
|
|
||||||
|
# RTSP 服务配置
|
||||||
|
RTSP_TARGET_FPS = 30.0
|
||||||
|
FRAMES_PER_SEGMENT = 1800
|
||||||
|
VIDEO_OUTPUT_DIR = "./videos"
|
||||||
|
WS_HOST = "0.0.0.0"
|
||||||
|
WS_PORT = 8765
|
||||||
|
|
||||||
|
# WebSocket 客户端集合
|
||||||
|
ws_clients = set()
|
||||||
|
|
||||||
|
|
||||||
|
# ========================= 数据结构 =========================
|
||||||
|
@dataclass
|
||||||
|
class CameraConfig:
|
||||||
|
id: int
|
||||||
|
name: str
|
||||||
|
rtsp_url: str
|
||||||
|
|
||||||
|
|
||||||
|
# ========================= Kadian TrafficMonitor(精简版,专为服务设计) =========================
|
||||||
|
class KadianDetector:
|
||||||
|
def __init__(self):
|
||||||
|
# 模型加载
|
||||||
|
|
||||||
|
self.police_prisoner_detector = YOLOv8_ONNX(police_prisoner_model_path, conf_threshold=0.5, iou_threshold=0.45,
|
||||||
|
input_size=police_prisoner_input_size)
|
||||||
|
|
||||||
|
# ByteTracker
|
||||||
|
class TrackerArgs:
|
||||||
|
track_thresh = 0.25
|
||||||
|
track_buffer = 30
|
||||||
|
match_thresh = 0.8
|
||||||
|
mot20 = False
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
self.police_prisoner_track_role = {}
|
||||||
|
|
||||||
|
self.fps = RTSP_TARGET_FPS
|
||||||
|
|
||||||
|
self.tracker = BYTETracker(TrackerArgs(), frame_rate=self.fps)
|
||||||
|
|
||||||
|
# ==========================================
|
||||||
|
# 超参数设置 (Hyperparameters)
|
||||||
|
# ==========================================
|
||||||
|
|
||||||
|
# 1. 业务判定时间阈值
|
||||||
|
# self.TIME_THRESHOLD_NOBODY = 2.0 # 无人在场判定时长
|
||||||
|
|
||||||
|
self.TIME_THRESHOLD_POLICE = 1.0 # 警察判定时长
|
||||||
|
self.TIME_TOLERANCE_POLICE = 0.5 # 警察失缓冲时间(防抖动)
|
||||||
|
|
||||||
|
self.TIME_THRESHOLD_PRISONER = 1.0 # 犯人判定时长
|
||||||
|
self.TIME_TOLERANCE_PRISONER = 0.5 # 犯人丢失缓冲时间(防抖动)
|
||||||
|
|
||||||
|
# 无人在场帧数阈值
|
||||||
|
# self.frame_thresh_nobody = int(self.TIME_THRESHOLD_NOBODY * self.fps)
|
||||||
|
|
||||||
|
# 警察检测帧数阈值
|
||||||
|
self.frame_thresh_police = int(self.TIME_THRESHOLD_POLICE * self.fps)
|
||||||
|
self.frame_buffer_police = int(self.TIME_TOLERANCE_POLICE * self.fps)
|
||||||
|
|
||||||
|
# 犯人检测帧数阈值
|
||||||
|
self.frame_thresh_prisoner = int(self.TIME_THRESHOLD_PRISONER * self.fps)
|
||||||
|
self.frame_buffer_prisoner = int(self.TIME_TOLERANCE_PRISONER * self.fps)
|
||||||
|
|
||||||
|
print(f"\n超参数设置:")
|
||||||
|
print(f" FPS: {self.fps:.2f}")
|
||||||
|
# print(f" 判定 'Nobody' 需连续: {self.frame_thresh_nobody} 帧")
|
||||||
|
print(f" 判定 'police Detected' 需累计检测: {self.frame_thresh_police} 帧")
|
||||||
|
print(f" 警察丢失缓冲帧数: {self.frame_buffer_police} 帧")
|
||||||
|
print(f" 判定 'prisoner Detected' 需累计检测: {self.frame_thresh_prisoner} 帧")
|
||||||
|
print(f" 犯人丢失缓冲帧数: {self.frame_buffer_prisoner} 帧")
|
||||||
|
|
||||||
|
# ==========================================
|
||||||
|
# 状态变量初始化
|
||||||
|
# ==========================================
|
||||||
|
|
||||||
|
self.current_frame_idx = 0
|
||||||
|
|
||||||
|
# 无人在场检测状态变量
|
||||||
|
self.cnt_frame_nobody = 0
|
||||||
|
|
||||||
|
# 警察检测状态变量
|
||||||
|
self.police_detection_frames = 0 # 连续检测到警察的帧数
|
||||||
|
self.police_missing_frames = 0 # 连续未检测到警察的帧数
|
||||||
|
self.police_alert_active = False # 警察报警是否激活
|
||||||
|
|
||||||
|
# 犯人检测状态变量
|
||||||
|
self.prisoner_detection_frames = 0 # 连续检测到犯人的帧数
|
||||||
|
self.prisoner_missing_frames = 0 # 连续未检测到犯人的帧数
|
||||||
|
self.prisoner_alert_active = False # 犯人报警是否激活
|
||||||
|
|
||||||
|
|
||||||
|
def compute_iou(self,boxA, boxB):
|
||||||
|
# box = [x1, y1, x2, y2]
|
||||||
|
xA = max(boxA[0], boxB[0])
|
||||||
|
yA = max(boxA[1], boxB[1])
|
||||||
|
xB = min(boxA[2], boxB[2])
|
||||||
|
yB = min(boxA[3], boxB[3])
|
||||||
|
|
||||||
|
interW = max(0, xB - xA)
|
||||||
|
interH = max(0, yB - yA)
|
||||||
|
interArea = interW * interH
|
||||||
|
|
||||||
|
boxAArea = (boxA[2] - boxA[0]) * (boxA[3] - boxA[1])
|
||||||
|
boxBArea = (boxB[2] - boxB[0]) * (boxB[3] - boxB[1])
|
||||||
|
|
||||||
|
unionArea = boxAArea + boxBArea - interArea
|
||||||
|
if unionArea == 0:
|
||||||
|
return 0.0
|
||||||
|
|
||||||
|
return interArea / unionArea
|
||||||
|
|
||||||
|
def draw_alert(self, frame, text, color=(0, 0, 255), sub_text=None, offset_y=0):
|
||||||
|
"""在右上角绘制警告文字 (支持垂直偏移,防止文字重叠)"""
|
||||||
|
font_scale = 1.5
|
||||||
|
thickness = 3
|
||||||
|
font = cv2.FONT_HERSHEY_SIMPLEX
|
||||||
|
|
||||||
|
(text_w, text_h), _ = cv2.getTextSize(text, font, font_scale, thickness)
|
||||||
|
x = self.width - text_w - 20
|
||||||
|
y = 50 + text_h + offset_y # 增加 Y 轴偏移
|
||||||
|
|
||||||
|
cv2.rectangle(frame, (x - 10, y - text_h - 10), (x + text_w + 10, y + 10), (0, 0, 0), -1)
|
||||||
|
cv2.putText(frame, text, (x, y), font, font_scale, color, thickness)
|
||||||
|
|
||||||
|
if sub_text:
|
||||||
|
cv2.putText(frame, sub_text, (x, y + 40), font, 0.7, (200, 200, 200), 2)
|
||||||
|
|
||||||
|
def process_frame(self, frame, camera_id: int, timestamp: float) -> Dict[str, Any]:
|
||||||
|
h, w = frame.shape[:2]
|
||||||
|
self.width, self.height = w, h
|
||||||
|
|
||||||
|
self.current_frame_idx += 1
|
||||||
|
|
||||||
|
current_time_sec = timestamp
|
||||||
|
|
||||||
|
# ========= 警察和犯人检测 =========
|
||||||
|
police_prisoner_results = self.police_prisoner_detector(frame)
|
||||||
|
|
||||||
|
police_prisoner_dets_xyxy = []
|
||||||
|
police_prisoner_dets_roles = []
|
||||||
|
police_prisoner_dets_for_tracker = []
|
||||||
|
|
||||||
|
# ========= 当前帧所有警告列表(关键改动)==========
|
||||||
|
current_frame_alerts = [] # 每帧清空,重新收集
|
||||||
|
|
||||||
|
if police_prisoner_results:
|
||||||
|
for det in police_prisoner_results:
|
||||||
|
x1, y1, x2, y2, conf, cls_id = det # x1, y1, x2, y2为角点坐标,x1 y1为左上角,x2 y2为右下角
|
||||||
|
police_prisoner_dets_xyxy.append([x1, y1, x2, y2])
|
||||||
|
police_prisoner_dets_for_tracker.append([x1, y1, x2, y2, conf])
|
||||||
|
if cls_id == 0:
|
||||||
|
police_prisoner_dets_roles.append("police")
|
||||||
|
elif cls_id == 1:
|
||||||
|
police_prisoner_dets_roles.append("prisoner")
|
||||||
|
|
||||||
|
ppolice_prisoner_dets = np.array(police_prisoner_dets_for_tracker, dtype=np.float32) if len(
|
||||||
|
police_prisoner_dets_for_tracker) else np.empty((0, 5))
|
||||||
|
|
||||||
|
police_prisoner_dets_tracks = self.tracker.update(
|
||||||
|
ppolice_prisoner_dets,
|
||||||
|
[self.height, self.width],
|
||||||
|
[self.height, self.width]
|
||||||
|
)
|
||||||
|
# ========= 单帧统计变量 =========
|
||||||
|
current_police_count = 0
|
||||||
|
current_prisoner_count = 0
|
||||||
|
|
||||||
|
# ========= 警察和犯人检测 =========
|
||||||
|
for t in police_prisoner_dets_tracks:
|
||||||
|
# print("t: {}".format(t))
|
||||||
|
tid = t.track_id
|
||||||
|
# cls_id = -1
|
||||||
|
|
||||||
|
# IoU 匹配角色
|
||||||
|
|
||||||
|
if tid not in self.police_prisoner_track_role:
|
||||||
|
best_iou = 0
|
||||||
|
best_role = "unknown"
|
||||||
|
|
||||||
|
t_box = list(map(float, t.tlbr)) # [x1,y1,x2,y2]
|
||||||
|
|
||||||
|
for i, box in enumerate(police_prisoner_dets_xyxy):
|
||||||
|
iou_val = self.compute_iou(t_box, box)
|
||||||
|
if iou_val > best_iou:
|
||||||
|
best_iou = iou_val
|
||||||
|
best_role = police_prisoner_dets_roles[i]
|
||||||
|
if best_iou > 0.1:
|
||||||
|
self.police_prisoner_track_role[tid] = best_role
|
||||||
|
else:
|
||||||
|
self.police_prisoner_track_role[tid] = "unknown"
|
||||||
|
|
||||||
|
role = self.police_prisoner_track_role.get(tid, "unknown")
|
||||||
|
cls_id = -1
|
||||||
|
if role == "police":
|
||||||
|
cls_id = 0
|
||||||
|
elif role == "prisoner":
|
||||||
|
cls_id = 1
|
||||||
|
# print("tid: {}, role: {}, cls: {}".format(tid, role,cls_id))
|
||||||
|
|
||||||
|
x1, y1, x2, y2 = map(int, t.tlbr)
|
||||||
|
|
||||||
|
cx, cy = (x1 + x2) // 2, (y1 + y2) // 2
|
||||||
|
|
||||||
|
color = None
|
||||||
|
label = None
|
||||||
|
|
||||||
|
if cls_id == 0: # Person
|
||||||
|
current_police_count += 1
|
||||||
|
color = (255, 0, 255)
|
||||||
|
label = "police"
|
||||||
|
|
||||||
|
elif cls_id == 1: # Phone(主模型已支持)
|
||||||
|
current_prisoner_count += 1
|
||||||
|
color = (0, 0, 139)
|
||||||
|
label = "prisoner"
|
||||||
|
else:
|
||||||
|
color = (255, 255, 255)
|
||||||
|
label = "Unknown"
|
||||||
|
|
||||||
|
# label = f"ID:{tid} IN"
|
||||||
|
|
||||||
|
cv2.rectangle(frame, (x1, y1), (x2, y2), color, 2)
|
||||||
|
cv2.putText(frame, label, (x1, y1 - 10), cv2.FONT_HERSHEY_SIMPLEX, 0.6, color, 2)
|
||||||
|
|
||||||
|
# ==========================================
|
||||||
|
# 犯人检测
|
||||||
|
# ==========================================
|
||||||
|
if current_prisoner_count > 0:
|
||||||
|
# 检测到犯人框
|
||||||
|
self.prisoner_detection_frames += 1
|
||||||
|
self.prisoner_missing_frames = 0 # 重置丢失计数器
|
||||||
|
|
||||||
|
# 当检测累计达到阈值时,激活报警
|
||||||
|
if self.prisoner_detection_frames >= self.frame_thresh_prisoner:
|
||||||
|
self.prisoner_alert_active = True
|
||||||
|
else:
|
||||||
|
# 未检测到犯人框
|
||||||
|
self.prisoner_missing_frames += 1
|
||||||
|
|
||||||
|
# 如果之前检测到手机,重置检测计数器
|
||||||
|
if self.prisoner_detection_frames > 0:
|
||||||
|
# 只有在连续丢失超过缓冲帧数时才重置
|
||||||
|
if self.prisoner_missing_frames >= self.frame_buffer_prisoner:
|
||||||
|
self.prisoner_detection_frames = 0
|
||||||
|
self.prisoner_alert_active = False
|
||||||
|
else:
|
||||||
|
# 从未检测到犯人,保持状态
|
||||||
|
pass
|
||||||
|
# ==========================================
|
||||||
|
# 警察检测
|
||||||
|
# ==========================================
|
||||||
|
if current_police_count > 0:
|
||||||
|
# 检测到犯人框
|
||||||
|
self.police_detection_frames += 1
|
||||||
|
self.police_missing_frames = 0 # 重置丢失计数器
|
||||||
|
|
||||||
|
# 当检测累计达到阈值时,激活报警
|
||||||
|
if self.police_detection_frames >= self.frame_thresh_police:
|
||||||
|
self.police_alert_active = True
|
||||||
|
else:
|
||||||
|
# 未检测到犯人框
|
||||||
|
self.police_missing_frames += 1
|
||||||
|
|
||||||
|
# 如果之前检测到手机,重置检测计数器
|
||||||
|
if self.police_detection_frames > 0:
|
||||||
|
# 只有在连续丢失超过缓冲帧数时才重置
|
||||||
|
if self.police_missing_frames >= self.frame_buffer_police:
|
||||||
|
self.police_detection_frames = 0
|
||||||
|
self.police_alert_active = False
|
||||||
|
else:
|
||||||
|
# 从未检测到犯人,保持状态
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
alert_offset = 0
|
||||||
|
|
||||||
|
# A. 有犯人
|
||||||
|
if self.prisoner_alert_active:
|
||||||
|
duration_seconds = self.prisoner_detection_frames / self.fps
|
||||||
|
current_frame_alerts.append(
|
||||||
|
{
|
||||||
|
'time': current_time_sec,
|
||||||
|
'action': 'prisoner',
|
||||||
|
'confidence': 1.0, # 固定为1.0(规则判定)
|
||||||
|
'details': f"Detected for {duration_seconds:.1f}s"
|
||||||
|
}
|
||||||
|
)
|
||||||
|
self.draw_alert(frame, "prisoner", (0, 0, 255), offset_y=alert_offset)
|
||||||
|
alert_offset += 100
|
||||||
|
|
||||||
|
# ==========================================
|
||||||
|
# 11. 统一显示当前帧所有警告(可替换原分层显示)
|
||||||
|
# ==========================================
|
||||||
|
debug_info = f" prisoner: {current_prisoner_count}"
|
||||||
|
cv2.putText(frame, debug_info, (20, 40), cv2.FONT_HERSHEY_SIMPLEX, 0.7, (0, 255, 0), 2)
|
||||||
|
|
||||||
|
# 统一警告显示区
|
||||||
|
alert_y_start = 150
|
||||||
|
for i, alert in enumerate(current_frame_alerts):
|
||||||
|
action = alert['action']
|
||||||
|
details = alert.get('details', '')
|
||||||
|
color = (0, 0, 255) # 默认红色警告
|
||||||
|
|
||||||
|
if action == 'prisoner':
|
||||||
|
color = (255, 255, 255)
|
||||||
|
|
||||||
|
main_text = action
|
||||||
|
if details:
|
||||||
|
main_text += f" ({details})"
|
||||||
|
|
||||||
|
y_pos = alert_y_start + i * 50
|
||||||
|
cv2.rectangle(frame, (20, y_pos - 40), (900, y_pos + 10), (0, 0, 0), -1)
|
||||||
|
cv2.putText(frame, main_text, (30, y_pos), cv2.FONT_HERSHEY_SIMPLEX, 1.0, color, 2)
|
||||||
|
|
||||||
|
return {
|
||||||
|
"image": frame,
|
||||||
|
|
||||||
|
"alerts":current_frame_alerts
|
||||||
|
}
|
||||||
|
|
||||||
|
# ========================= WebSocket 服务线程 =========================
|
||||||
|
class WebSocketSender(threading.Thread):
|
||||||
|
def __init__(self, send_queue: queue.Queue, stop_event: threading.Event):
|
||||||
|
super().__init__(daemon=True)
|
||||||
|
self.send_queue = send_queue
|
||||||
|
self.stop_event = stop_event
|
||||||
|
|
||||||
|
async def _ws_handler(self, websocket):
|
||||||
|
ws_clients.add(websocket)
|
||||||
|
try:
|
||||||
|
async for _ in websocket:
|
||||||
|
pass
|
||||||
|
finally:
|
||||||
|
ws_clients.discard(websocket)
|
||||||
|
|
||||||
|
async def _broadcaster(self):
|
||||||
|
while not self.stop_event.is_set():
|
||||||
|
try:
|
||||||
|
msg = await asyncio.to_thread(self.send_queue.get, timeout=0.5)
|
||||||
|
except queue.Empty:
|
||||||
|
continue
|
||||||
|
data = json.dumps(msg)
|
||||||
|
dead = []
|
||||||
|
for ws in list(ws_clients):
|
||||||
|
try:
|
||||||
|
await ws.send(data)
|
||||||
|
except:
|
||||||
|
dead.append(ws)
|
||||||
|
for ws in dead:
|
||||||
|
ws_clients.discard(ws)
|
||||||
|
self.send_queue.task_done()
|
||||||
|
|
||||||
|
async def _run_async(self):
|
||||||
|
async with websockets.serve(self._ws_handler, WS_HOST, WS_PORT):
|
||||||
|
print(f"[INFO] WebSocket server started at ws://{WS_HOST}:{WS_PORT}")
|
||||||
|
await self._broadcaster()
|
||||||
|
|
||||||
|
def run(self):
|
||||||
|
asyncio.run(self._run_async())
|
||||||
|
|
||||||
|
|
||||||
|
# ========================= RTSP 抓流线程 =========================
|
||||||
|
class RTSPCaptureWorker(threading.Thread):
|
||||||
|
def __init__(self, camera_cfg: CameraConfig, raw_queue: queue.Queue, stop_event: threading.Event):
|
||||||
|
super().__init__(daemon=True)
|
||||||
|
self.camera_cfg = camera_cfg
|
||||||
|
self.raw_queue = raw_queue
|
||||||
|
self.stop_event = stop_event
|
||||||
|
|
||||||
|
def run(self):
|
||||||
|
cap = cv2.VideoCapture(self.camera_cfg.rtsp_url, cv2.CAP_FFMPEG)
|
||||||
|
if not cap.isOpened():
|
||||||
|
print(f"[ERROR] Cannot open RTSP: {self.camera_cfg.rtsp_url}")
|
||||||
|
return
|
||||||
|
print(f"[INFO] Capturing {self.camera_cfg.name} (ID:{self.camera_cfg.id})")
|
||||||
|
while not self.stop_event.is_set():
|
||||||
|
ret, frame = cap.read()
|
||||||
|
if not ret:
|
||||||
|
time.sleep(0.2)
|
||||||
|
continue
|
||||||
|
item = {
|
||||||
|
"camera_id": self.camera_cfg.id,
|
||||||
|
"camera_name": self.camera_cfg.name,
|
||||||
|
"timestamp": time.time(),
|
||||||
|
"frame": frame,
|
||||||
|
}
|
||||||
|
try:
|
||||||
|
self.raw_queue.put(item, timeout=1.0)
|
||||||
|
except queue.Full:
|
||||||
|
pass
|
||||||
|
cap.release()
|
||||||
|
|
||||||
|
|
||||||
|
# ========================= 帧处理线程 =========================
|
||||||
|
class FrameProcessorWorker(threading.Thread):
|
||||||
|
def __init__(self,
|
||||||
|
raw_frame_queue: "queue.Queue[Dict[str, Any]]",
|
||||||
|
ws_send_queue: "queue.Queue[Dict[str, Any]]",
|
||||||
|
stop_event: threading.Event):
|
||||||
|
super().__init__(daemon=True)
|
||||||
|
self.raw_queue = raw_frame_queue
|
||||||
|
self.ws_queue = ws_send_queue
|
||||||
|
self.stop_event = stop_event
|
||||||
|
|
||||||
|
self.video_writers: Dict[int, cv2.VideoWriter] = {}
|
||||||
|
self.video_counts: Dict[int, int] = {}
|
||||||
|
self.last_ts: Dict[int, float] = {}
|
||||||
|
self.video_files: Dict[int, str] = {}
|
||||||
|
|
||||||
|
|
||||||
|
os.makedirs(VIDEO_OUTPUT_DIR, exist_ok=True)
|
||||||
|
|
||||||
|
# 每个摄像头一个独立的 Kadian 检测器实例
|
||||||
|
self.kadian_detectors: Dict[int, KadianDetector] = {}
|
||||||
|
|
||||||
|
def _get_writer(self, camera_id: int, frame) -> Tuple[cv2.VideoWriter, str]:
|
||||||
|
if camera_id in self.video_writers:
|
||||||
|
return self.video_writers[camera_id], self.video_files[camera_id]
|
||||||
|
|
||||||
|
h, w = frame.shape[:2]
|
||||||
|
ts_str = datetime.now().strftime("%Y%m%d_%H%M%S")
|
||||||
|
filepath = os.path.join(VIDEO_OUTPUT_DIR, f"{ts_str}_cam{camera_id}.mp4")
|
||||||
|
writer = cv2.VideoWriter(filepath, cv2.VideoWriter_fourcc(*"mp4v"), RTSP_TARGET_FPS, (w, h))
|
||||||
|
|
||||||
|
self.video_writers[camera_id] = writer
|
||||||
|
self.video_files[camera_id] = filepath
|
||||||
|
self.video_counts[camera_id] = 0
|
||||||
|
print(f"[INFO] New segment: {filepath}")
|
||||||
|
return writer, filepath
|
||||||
|
|
||||||
|
def _close_segment_if_needed(self, camera_id: int):
|
||||||
|
count = self.video_counts.get(camera_id, 0)
|
||||||
|
if count >= FRAMES_PER_SEGMENT:
|
||||||
|
writer = self.video_writers.get(camera_id)
|
||||||
|
if writer is not None:
|
||||||
|
writer.release()
|
||||||
|
print(f"[INFO] Close segment: camera={camera_id}, file={self.video_files[camera_id]}")
|
||||||
|
|
||||||
|
self.video_writers.pop(camera_id, None)
|
||||||
|
self.video_counts.pop(camera_id, None)
|
||||||
|
self.last_ts.pop(camera_id, None)
|
||||||
|
self.video_files.pop(camera_id, None)
|
||||||
|
|
||||||
|
def _encode_image_to_base64(self, image) -> str:
|
||||||
|
ok, buf = cv2.imencode(".jpg", image)
|
||||||
|
if not ok:
|
||||||
|
raise RuntimeError("Failed to encode image to JPEG")
|
||||||
|
return base64.b64encode(buf.tobytes()).decode("ascii")
|
||||||
|
|
||||||
|
def run(self):
|
||||||
|
target_interval = 1.0 / RTSP_TARGET_FPS
|
||||||
|
while not self.stop_event.is_set():
|
||||||
|
try:
|
||||||
|
item = self.raw_queue.get(timeout=0.5)
|
||||||
|
except queue.Empty:
|
||||||
|
continue
|
||||||
|
|
||||||
|
cam_id = item["camera_id"]
|
||||||
|
ts = item["timestamp"]
|
||||||
|
frame = item["frame"]
|
||||||
|
|
||||||
|
# 抽帧控制
|
||||||
|
if ts - self.last_ts.get(cam_id, 0) < target_interval:
|
||||||
|
self.raw_queue.task_done()
|
||||||
|
continue
|
||||||
|
self.last_ts[cam_id] = ts
|
||||||
|
|
||||||
|
# 获取检测器实例
|
||||||
|
if cam_id not in self.kadian_detectors:
|
||||||
|
self.kadian_detectors[cam_id] = KadianDetector()
|
||||||
|
detector = self.kadian_detectors[cam_id]
|
||||||
|
|
||||||
|
# 2) 进行人脸识别(如果启用)
|
||||||
|
current_face_alert = None
|
||||||
|
face_results = []
|
||||||
|
face_processing_time = 0
|
||||||
|
if video_face_prison_biz is not None and FACE_RECOGNITION_ENABLED:
|
||||||
|
try:
|
||||||
|
# 处理当前帧 - 获取人脸识别结果
|
||||||
|
processed_frame_for_face, face_results, face_processing_time = video_face_prison_biz.process_frame(
|
||||||
|
frame.copy())
|
||||||
|
|
||||||
|
for result in face_results:
|
||||||
|
if result['has_passed']:
|
||||||
|
print(f"[INFO] 犯人带出: {result['passed_person_id']}")
|
||||||
|
|
||||||
|
# 记录当前帧人脸告警信息
|
||||||
|
current_face_alert = {
|
||||||
|
"person_name": result['passed_person_id'],
|
||||||
|
"timestamp": ts
|
||||||
|
}
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
print(f"[WARN] 人脸识别处理失败: {e}")
|
||||||
|
|
||||||
|
# 执行检测
|
||||||
|
result = detector.process_frame(frame.copy(), cam_id, ts)
|
||||||
|
|
||||||
|
result_img = result["image"]
|
||||||
|
result_type = result["alerts"]
|
||||||
|
|
||||||
|
# 绘制人脸识别结果
|
||||||
|
if video_face_prison_biz is not None and face_results:
|
||||||
|
result_img = video_face_prison_biz.draw_detections(result_img, face_results)
|
||||||
|
|
||||||
|
# 添加人脸识别统计信息
|
||||||
|
match_count = sum(1 for r in face_results if r['is_match'])
|
||||||
|
face_info_text = f"Faces: {len(face_results)} | Matches: {match_count}"
|
||||||
|
cv2.putText(result_img, face_info_text, (10, result_img.shape[0] - 20),
|
||||||
|
cv2.FONT_HERSHEY_SIMPLEX, 0.6, (0, 255, 255), 2)
|
||||||
|
|
||||||
|
# 写视频
|
||||||
|
writer, video_path = self._get_writer(cam_id, frame)
|
||||||
|
writer.write(result_img)
|
||||||
|
self.video_counts[cam_id] = self.video_counts.get(cam_id, 0) + 1
|
||||||
|
|
||||||
|
# 5) 通过 WebSocket 发送帧结果
|
||||||
|
try:
|
||||||
|
img_b64 = self._encode_image_to_base64(result_img)
|
||||||
|
except Exception as e:
|
||||||
|
print(f"[ERROR] Encode image failed: {e}")
|
||||||
|
img_b64 = None
|
||||||
|
|
||||||
|
if img_b64 is not None:
|
||||||
|
# 将abnormal_actions对象数组转换为字符串数组
|
||||||
|
action_names = [action_info['action'] for action_info in result_type]
|
||||||
|
|
||||||
|
if current_face_alert is not None:
|
||||||
|
action_names.append("face")
|
||||||
|
|
||||||
|
msg = {
|
||||||
|
"msg_type": "frame",
|
||||||
|
"camera_id": cam_id,
|
||||||
|
"timestamp": ts,
|
||||||
|
"result_type": action_names,
|
||||||
|
"image_base64": img_b64,
|
||||||
|
}
|
||||||
|
try:
|
||||||
|
self.ws_queue.put(msg, timeout=1.0)
|
||||||
|
except queue.Full:
|
||||||
|
print("[WARN] ws_send_queue full, drop frame message")
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
self._close_segment_if_needed(cam_id)
|
||||||
|
self.raw_queue.task_done()
|
||||||
|
|
||||||
|
self.video_counts[cam_id] = self.video_counts.get(cam_id, 0) + 1
|
||||||
|
|
||||||
|
# 清理
|
||||||
|
for w in self.video_writers.values():
|
||||||
|
w.release()
|
||||||
|
|
||||||
|
|
||||||
|
# ========================= 服务主类 =========================
|
||||||
|
class RTSPService:
|
||||||
|
def __init__(self, config_path: str = "config.yaml"):
|
||||||
|
with open(config_path, "r", encoding="utf-8") as f:
|
||||||
|
cfg = yaml.safe_load(f)
|
||||||
|
self.cameras = [CameraConfig(id=c["id"], name=c.get("name", f"cam_{c['id']}"), rtsp_url=c["rtsp_url"])
|
||||||
|
for c in cfg.get("cameras", [])]
|
||||||
|
|
||||||
|
self.stop_event = threading.Event()
|
||||||
|
self.raw_queue = queue.Queue(maxsize=500)
|
||||||
|
self.ws_queue = queue.Queue(maxsize=1000)
|
||||||
|
|
||||||
|
self.capture_workers = []
|
||||||
|
self.processor = FrameProcessorWorker(self.raw_queue, self.ws_queue, self.stop_event)
|
||||||
|
self.ws_sender = WebSocketSender(self.ws_queue, self.stop_event)
|
||||||
|
|
||||||
|
def start(self):
|
||||||
|
self.ws_sender.start()
|
||||||
|
self.processor.start()
|
||||||
|
for cam in self.cameras:
|
||||||
|
w = RTSPCaptureWorker(cam, self.raw_queue, self.stop_event)
|
||||||
|
w.start()
|
||||||
|
self.capture_workers.append(w)
|
||||||
|
print("[INFO] Kadian RTSP Service started")
|
||||||
|
|
||||||
|
def stop(self):
|
||||||
|
self.stop_event.set()
|
||||||
|
self.raw_queue.join()
|
||||||
|
self.ws_queue.join()
|
||||||
|
for w in self.capture_workers:
|
||||||
|
w.join(timeout=2.0)
|
||||||
|
self.processor.join(timeout=2.0)
|
||||||
|
self.ws_sender.join(timeout=2.0)
|
||||||
|
print("[INFO] Service stopped")
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
service = RTSPService("config.yaml")
|
||||||
|
service.start()
|
||||||
|
try:
|
||||||
|
while True:
|
||||||
|
time.sleep(1)
|
||||||
|
except KeyboardInterrupt:
|
||||||
|
service.stop()
|
||||||
Reference in New Issue
Block a user