引入yolo监狱识别代码
This commit is contained in:
153
npu_yolo_onnx_person_car_phone.py
Normal file
153
npu_yolo_onnx_person_car_phone.py
Normal file
@@ -0,0 +1,153 @@
|
||||
# 文件名: npu_yolo_onnx.py
|
||||
import cv2
|
||||
import numpy as np
|
||||
import onnxruntime as ort
|
||||
import os
|
||||
import time
|
||||
|
||||
def letterbox(img, new_shape=(640, 640), color=(114, 114, 114)):
|
||||
shape = img.shape[:2] # h, w
|
||||
r = min(new_shape[0] / shape[0], new_shape[1] / shape[1])
|
||||
new_unpad = int(round(shape[1] * r)), int(round(shape[0] * r))
|
||||
dw, dh = new_shape[1] - new_unpad[0], new_shape[0] - new_unpad[1]
|
||||
dw /= 2
|
||||
dh /= 2
|
||||
if shape[::-1] != new_unpad:
|
||||
img = cv2.resize(img, new_unpad, interpolation=cv2.INTER_LINEAR)
|
||||
top, bottom = int(round(dh - 0.1)), int(round(dh + 0.1))
|
||||
left, right = int(round(dw - 0.1)), int(round(dw + 0.1))
|
||||
img = cv2.copyMakeBorder(img, top, bottom, left, right, cv2.BORDER_CONSTANT, value=color)
|
||||
return img, r, (dw, dh)
|
||||
|
||||
class YOLOv8_ONNX:
|
||||
def __init__(self, onnx_path, conf_threshold=0.25, iou_threshold=0.45, input_size=640):
|
||||
providers = [("CANNExecutionProvider", {
|
||||
"device_id": 0,
|
||||
"arena_extend_strategy": "kNextPowerOfTwo",
|
||||
"npu_mem_limit": 16 * 1024 * 1024 * 1024,
|
||||
"precision_mode": "allow_fp32_to_fp16",
|
||||
"op_select_impl_mode": "high_precision",
|
||||
"enable_cann_graph": True,
|
||||
}),
|
||||
"CUDAExecutionProvider",
|
||||
"CPUExecutionProvider",
|
||||
]
|
||||
|
||||
self.session = ort.InferenceSession(onnx_path, providers=providers)
|
||||
actual_providers = self.session.get_providers()
|
||||
print("YOLO Providers:", actual_providers)
|
||||
|
||||
if "CANNExecutionProvider" in actual_providers:
|
||||
print("[INFO] YOLO 使用 CANNExecutionProvider(昇腾 NPU)")
|
||||
elif 'CUDAExecutionProvider' in actual_providers:
|
||||
print("[INFO] YOLO 使用 CUDAExecutionProvider(NVIDIA GPU)")
|
||||
else:
|
||||
print("[INFO] YOLO 使用 CPUExecutionProvider")
|
||||
|
||||
self.conf_threshold = conf_threshold
|
||||
self.iou_threshold = iou_threshold
|
||||
self.input_name = self.session.get_inputs()[0].name
|
||||
self.input_size = (input_size, input_size) if isinstance(input_size, int) else input_size
|
||||
|
||||
print(f"模型输入名称: {self.input_name}")
|
||||
print(f"模型输入形状: {self.session.get_inputs()[0].shape}")
|
||||
print(f"模型输出形状: {self.session.get_outputs()[0].shape}")
|
||||
|
||||
def preprocess(self, img):
|
||||
self.orig_shape = img.shape[:2]
|
||||
img, self.ratio, (self.dw, self.dh) = letterbox(img, self.input_size)
|
||||
img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)
|
||||
img = img.transpose(2, 0, 1).astype(np.float32)
|
||||
img /= 255.0
|
||||
img = np.expand_dims(img, axis=0)
|
||||
return img
|
||||
|
||||
def postprocess(self, pred, im0_shape):
|
||||
# 1. 转置:从 [1, 4+cls, 8400] -> [8400, 4+cls]
|
||||
pred = pred[0].T
|
||||
|
||||
# 2. 获取数据
|
||||
boxes = pred[:, :4] # cx, cy, w, h
|
||||
scores = pred[:, 4:]
|
||||
|
||||
# 3. 获取最大置信度和类别
|
||||
conf = np.max(scores, axis=1)
|
||||
class_pred = np.argmax(scores, axis=1)
|
||||
|
||||
# 4. 初步过滤
|
||||
mask = conf > self.conf_threshold
|
||||
if not mask.any():
|
||||
return []
|
||||
|
||||
boxes = boxes[mask]
|
||||
conf = conf[mask]
|
||||
class_pred = class_pred[mask]
|
||||
|
||||
# =========================================================
|
||||
# 还原坐标 (逆 Letterbox)
|
||||
# =========================================================
|
||||
boxes[:, 0] = (boxes[:, 0] - self.dw) / self.ratio # cx
|
||||
boxes[:, 1] = (boxes[:, 1] - self.dh) / self.ratio # cy
|
||||
boxes[:, 2] = boxes[:, 2] / self.ratio # w
|
||||
boxes[:, 3] = boxes[:, 3] / self.ratio # h
|
||||
|
||||
# 转换格式:Center(cx,cy) -> TopLeft(x,y)
|
||||
x = boxes[:, 0] - boxes[:, 2] / 2
|
||||
y = boxes[:, 1] - boxes[:, 3] / 2
|
||||
w = boxes[:, 2]
|
||||
h = boxes[:, 3]
|
||||
|
||||
# 原始框(用于最终输出)
|
||||
bboxes_original = np.stack([x, y, w, h], axis=1)
|
||||
|
||||
# =========================================================
|
||||
# 【核心修复】:Class-Aware NMS (偏移量法)
|
||||
# 给不同类别的框增加不同的偏移量,使得不同类别的框绝对不会重叠
|
||||
# 从而避免 "车" 把 "人" 过滤掉的情况
|
||||
# =========================================================
|
||||
max_wh = 4096 # 只要大于图片最大分辨率即可
|
||||
class_offset = class_pred * max_wh
|
||||
|
||||
# NMS 专用的框坐标 (加上了偏移量)
|
||||
bboxes_for_nms = bboxes_original.copy()
|
||||
bboxes_for_nms[:, 0] += class_offset
|
||||
bboxes_for_nms[:, 1] += class_offset
|
||||
|
||||
# =========================================================
|
||||
# 执行 NMS
|
||||
# =========================================================
|
||||
indices = cv2.dnn.NMSBoxes(
|
||||
bboxes_for_nms.tolist(),
|
||||
conf.tolist(),
|
||||
self.conf_threshold,
|
||||
self.iou_threshold
|
||||
)
|
||||
|
||||
result = []
|
||||
if len(indices) > 0:
|
||||
indices = indices.flatten()
|
||||
for i in indices:
|
||||
# 注意:这里取数据要从 bboxes_original 取 (没有加偏移量的)
|
||||
bx, by, bw, bh = bboxes_original[i]
|
||||
|
||||
# 转换回 x1, y1, x2, y2 供业务代码画图使用
|
||||
x1 = np.clip(bx, 0, im0_shape[1])
|
||||
y1 = np.clip(by, 0, im0_shape[0])
|
||||
x2 = np.clip(bx + bw, 0, im0_shape[1])
|
||||
y2 = np.clip(by + bh, 0, im0_shape[0])
|
||||
|
||||
result.append([
|
||||
float(x1),
|
||||
float(y1),
|
||||
float(x2),
|
||||
float(y2),
|
||||
float(conf[i]),
|
||||
int(class_pred[i])
|
||||
])
|
||||
return result
|
||||
|
||||
def __call__(self, frame):
|
||||
input_data = self.preprocess(frame)
|
||||
pred = self.session.run(None, {self.input_name: input_data})[0]
|
||||
results = self.postprocess(pred, frame.shape[:2])
|
||||
return results
|
||||
@@ -4,7 +4,8 @@ RTSP 服务模块 - 简洁版本,直接使用原始服务
|
||||
|
||||
import threading
|
||||
# from rtsp_service_ws_1217 import RTSPService
|
||||
from rtsp_service_ws_prison import RTSPService
|
||||
# from rtsp_service_ws_prison import RTSPService
|
||||
from rtsp_service_ws_0108 import RTSPService
|
||||
|
||||
|
||||
class SimpleRTSPServer:
|
||||
|
||||
644
rtsp_service_ws_0108.py
Normal file
644
rtsp_service_ws_0108.py
Normal file
@@ -0,0 +1,644 @@
|
||||
# rtsp_service_kadian.py
|
||||
# 融合 Kadian_Detect_1221.py + rtsp_service_ws.py
|
||||
# 支持多路RTSP、抽帧、分段保存MP4、WebSocket推送图像与告警
|
||||
|
||||
import cv2
|
||||
import numpy as np
|
||||
import os
|
||||
import time
|
||||
import threading
|
||||
import queue
|
||||
import yaml
|
||||
import json
|
||||
import base64
|
||||
import asyncio
|
||||
import websockets
|
||||
from dataclasses import dataclass
|
||||
from typing import Dict, Any, Tuple
|
||||
from datetime import datetime
|
||||
|
||||
# 导入人脸识别算法
|
||||
try:
|
||||
from api.routes.algorithm_router import video_face_prison_biz
|
||||
|
||||
print("[INFO] 成功导入人脸识别算法")
|
||||
except Exception as e:
|
||||
print(f"[WARN] 无法导入人脸识别算法: {e}")
|
||||
|
||||
# -------------------------- Kadian 检测相关导入 --------------------------
|
||||
from npu_yolo_onnx_person_car_phone import YOLOv8_ONNX # 主检测模型(人/车/后备箱/手机)
|
||||
|
||||
from yolox.tracker.byte_tracker import BYTETracker
|
||||
|
||||
|
||||
# ========================= 配置区 =========================
|
||||
# Kadian 模型路径与ROI(可根据实际情况修改)
|
||||
police_prisoner_model_path = 'YOLO_Weight/prisoner_model.onnx'
|
||||
|
||||
FACE_RECOGNITION_ENABLED = True # 是否启用人脸识别
|
||||
|
||||
# 输入尺寸
|
||||
police_prisoner_input_size = 1280
|
||||
|
||||
# RTSP 服务配置
|
||||
RTSP_TARGET_FPS = 30.0
|
||||
FRAMES_PER_SEGMENT = 1800
|
||||
VIDEO_OUTPUT_DIR = "./videos"
|
||||
WS_HOST = "0.0.0.0"
|
||||
WS_PORT = 8765
|
||||
|
||||
# WebSocket 客户端集合
|
||||
ws_clients = set()
|
||||
|
||||
|
||||
# ========================= 数据结构 =========================
|
||||
@dataclass
|
||||
class CameraConfig:
|
||||
id: int
|
||||
name: str
|
||||
rtsp_url: str
|
||||
|
||||
|
||||
# ========================= Kadian TrafficMonitor(精简版,专为服务设计) =========================
|
||||
class KadianDetector:
|
||||
def __init__(self):
|
||||
# 模型加载
|
||||
|
||||
self.police_prisoner_detector = YOLOv8_ONNX(police_prisoner_model_path, conf_threshold=0.5, iou_threshold=0.45,
|
||||
input_size=police_prisoner_input_size)
|
||||
|
||||
# ByteTracker
|
||||
class TrackerArgs:
|
||||
track_thresh = 0.25
|
||||
track_buffer = 30
|
||||
match_thresh = 0.8
|
||||
mot20 = False
|
||||
|
||||
|
||||
|
||||
self.police_prisoner_track_role = {}
|
||||
|
||||
self.fps = RTSP_TARGET_FPS
|
||||
|
||||
self.tracker = BYTETracker(TrackerArgs(), frame_rate=self.fps)
|
||||
|
||||
# ==========================================
|
||||
# 超参数设置 (Hyperparameters)
|
||||
# ==========================================
|
||||
|
||||
# 1. 业务判定时间阈值
|
||||
# self.TIME_THRESHOLD_NOBODY = 2.0 # 无人在场判定时长
|
||||
|
||||
self.TIME_THRESHOLD_POLICE = 1.0 # 警察判定时长
|
||||
self.TIME_TOLERANCE_POLICE = 0.5 # 警察失缓冲时间(防抖动)
|
||||
|
||||
self.TIME_THRESHOLD_PRISONER = 1.0 # 犯人判定时长
|
||||
self.TIME_TOLERANCE_PRISONER = 0.5 # 犯人丢失缓冲时间(防抖动)
|
||||
|
||||
# 无人在场帧数阈值
|
||||
# self.frame_thresh_nobody = int(self.TIME_THRESHOLD_NOBODY * self.fps)
|
||||
|
||||
# 警察检测帧数阈值
|
||||
self.frame_thresh_police = int(self.TIME_THRESHOLD_POLICE * self.fps)
|
||||
self.frame_buffer_police = int(self.TIME_TOLERANCE_POLICE * self.fps)
|
||||
|
||||
# 犯人检测帧数阈值
|
||||
self.frame_thresh_prisoner = int(self.TIME_THRESHOLD_PRISONER * self.fps)
|
||||
self.frame_buffer_prisoner = int(self.TIME_TOLERANCE_PRISONER * self.fps)
|
||||
|
||||
print(f"\n超参数设置:")
|
||||
print(f" FPS: {self.fps:.2f}")
|
||||
# print(f" 判定 'Nobody' 需连续: {self.frame_thresh_nobody} 帧")
|
||||
print(f" 判定 'police Detected' 需累计检测: {self.frame_thresh_police} 帧")
|
||||
print(f" 警察丢失缓冲帧数: {self.frame_buffer_police} 帧")
|
||||
print(f" 判定 'prisoner Detected' 需累计检测: {self.frame_thresh_prisoner} 帧")
|
||||
print(f" 犯人丢失缓冲帧数: {self.frame_buffer_prisoner} 帧")
|
||||
|
||||
# ==========================================
|
||||
# 状态变量初始化
|
||||
# ==========================================
|
||||
|
||||
self.current_frame_idx = 0
|
||||
|
||||
# 无人在场检测状态变量
|
||||
self.cnt_frame_nobody = 0
|
||||
|
||||
# 警察检测状态变量
|
||||
self.police_detection_frames = 0 # 连续检测到警察的帧数
|
||||
self.police_missing_frames = 0 # 连续未检测到警察的帧数
|
||||
self.police_alert_active = False # 警察报警是否激活
|
||||
|
||||
# 犯人检测状态变量
|
||||
self.prisoner_detection_frames = 0 # 连续检测到犯人的帧数
|
||||
self.prisoner_missing_frames = 0 # 连续未检测到犯人的帧数
|
||||
self.prisoner_alert_active = False # 犯人报警是否激活
|
||||
|
||||
|
||||
def compute_iou(self,boxA, boxB):
|
||||
# box = [x1, y1, x2, y2]
|
||||
xA = max(boxA[0], boxB[0])
|
||||
yA = max(boxA[1], boxB[1])
|
||||
xB = min(boxA[2], boxB[2])
|
||||
yB = min(boxA[3], boxB[3])
|
||||
|
||||
interW = max(0, xB - xA)
|
||||
interH = max(0, yB - yA)
|
||||
interArea = interW * interH
|
||||
|
||||
boxAArea = (boxA[2] - boxA[0]) * (boxA[3] - boxA[1])
|
||||
boxBArea = (boxB[2] - boxB[0]) * (boxB[3] - boxB[1])
|
||||
|
||||
unionArea = boxAArea + boxBArea - interArea
|
||||
if unionArea == 0:
|
||||
return 0.0
|
||||
|
||||
return interArea / unionArea
|
||||
|
||||
def draw_alert(self, frame, text, color=(0, 0, 255), sub_text=None, offset_y=0):
|
||||
"""在右上角绘制警告文字 (支持垂直偏移,防止文字重叠)"""
|
||||
font_scale = 1.5
|
||||
thickness = 3
|
||||
font = cv2.FONT_HERSHEY_SIMPLEX
|
||||
|
||||
(text_w, text_h), _ = cv2.getTextSize(text, font, font_scale, thickness)
|
||||
x = self.width - text_w - 20
|
||||
y = 50 + text_h + offset_y # 增加 Y 轴偏移
|
||||
|
||||
cv2.rectangle(frame, (x - 10, y - text_h - 10), (x + text_w + 10, y + 10), (0, 0, 0), -1)
|
||||
cv2.putText(frame, text, (x, y), font, font_scale, color, thickness)
|
||||
|
||||
if sub_text:
|
||||
cv2.putText(frame, sub_text, (x, y + 40), font, 0.7, (200, 200, 200), 2)
|
||||
|
||||
def process_frame(self, frame, camera_id: int, timestamp: float) -> Dict[str, Any]:
|
||||
h, w = frame.shape[:2]
|
||||
self.width, self.height = w, h
|
||||
|
||||
self.current_frame_idx += 1
|
||||
|
||||
current_time_sec = timestamp
|
||||
|
||||
# ========= 警察和犯人检测 =========
|
||||
police_prisoner_results = self.police_prisoner_detector(frame)
|
||||
|
||||
police_prisoner_dets_xyxy = []
|
||||
police_prisoner_dets_roles = []
|
||||
police_prisoner_dets_for_tracker = []
|
||||
|
||||
# ========= 当前帧所有警告列表(关键改动)==========
|
||||
current_frame_alerts = [] # 每帧清空,重新收集
|
||||
|
||||
if police_prisoner_results:
|
||||
for det in police_prisoner_results:
|
||||
x1, y1, x2, y2, conf, cls_id = det # x1, y1, x2, y2为角点坐标,x1 y1为左上角,x2 y2为右下角
|
||||
police_prisoner_dets_xyxy.append([x1, y1, x2, y2])
|
||||
police_prisoner_dets_for_tracker.append([x1, y1, x2, y2, conf])
|
||||
if cls_id == 0:
|
||||
police_prisoner_dets_roles.append("police")
|
||||
elif cls_id == 1:
|
||||
police_prisoner_dets_roles.append("prisoner")
|
||||
|
||||
ppolice_prisoner_dets = np.array(police_prisoner_dets_for_tracker, dtype=np.float32) if len(
|
||||
police_prisoner_dets_for_tracker) else np.empty((0, 5))
|
||||
|
||||
police_prisoner_dets_tracks = self.tracker.update(
|
||||
ppolice_prisoner_dets,
|
||||
[self.height, self.width],
|
||||
[self.height, self.width]
|
||||
)
|
||||
# ========= 单帧统计变量 =========
|
||||
current_police_count = 0
|
||||
current_prisoner_count = 0
|
||||
|
||||
# ========= 警察和犯人检测 =========
|
||||
for t in police_prisoner_dets_tracks:
|
||||
# print("t: {}".format(t))
|
||||
tid = t.track_id
|
||||
# cls_id = -1
|
||||
|
||||
# IoU 匹配角色
|
||||
|
||||
if tid not in self.police_prisoner_track_role:
|
||||
best_iou = 0
|
||||
best_role = "unknown"
|
||||
|
||||
t_box = list(map(float, t.tlbr)) # [x1,y1,x2,y2]
|
||||
|
||||
for i, box in enumerate(police_prisoner_dets_xyxy):
|
||||
iou_val = self.compute_iou(t_box, box)
|
||||
if iou_val > best_iou:
|
||||
best_iou = iou_val
|
||||
best_role = police_prisoner_dets_roles[i]
|
||||
if best_iou > 0.1:
|
||||
self.police_prisoner_track_role[tid] = best_role
|
||||
else:
|
||||
self.police_prisoner_track_role[tid] = "unknown"
|
||||
|
||||
role = self.police_prisoner_track_role.get(tid, "unknown")
|
||||
cls_id = -1
|
||||
if role == "police":
|
||||
cls_id = 0
|
||||
elif role == "prisoner":
|
||||
cls_id = 1
|
||||
# print("tid: {}, role: {}, cls: {}".format(tid, role,cls_id))
|
||||
|
||||
x1, y1, x2, y2 = map(int, t.tlbr)
|
||||
|
||||
cx, cy = (x1 + x2) // 2, (y1 + y2) // 2
|
||||
|
||||
color = None
|
||||
label = None
|
||||
|
||||
if cls_id == 0: # Person
|
||||
current_police_count += 1
|
||||
color = (255, 0, 255)
|
||||
label = "police"
|
||||
|
||||
elif cls_id == 1: # Phone(主模型已支持)
|
||||
current_prisoner_count += 1
|
||||
color = (0, 0, 139)
|
||||
label = "prisoner"
|
||||
else:
|
||||
color = (255, 255, 255)
|
||||
label = "Unknown"
|
||||
|
||||
# label = f"ID:{tid} IN"
|
||||
|
||||
cv2.rectangle(frame, (x1, y1), (x2, y2), color, 2)
|
||||
cv2.putText(frame, label, (x1, y1 - 10), cv2.FONT_HERSHEY_SIMPLEX, 0.6, color, 2)
|
||||
|
||||
# ==========================================
|
||||
# 犯人检测
|
||||
# ==========================================
|
||||
if current_prisoner_count > 0:
|
||||
# 检测到犯人框
|
||||
self.prisoner_detection_frames += 1
|
||||
self.prisoner_missing_frames = 0 # 重置丢失计数器
|
||||
|
||||
# 当检测累计达到阈值时,激活报警
|
||||
if self.prisoner_detection_frames >= self.frame_thresh_prisoner:
|
||||
self.prisoner_alert_active = True
|
||||
else:
|
||||
# 未检测到犯人框
|
||||
self.prisoner_missing_frames += 1
|
||||
|
||||
# 如果之前检测到手机,重置检测计数器
|
||||
if self.prisoner_detection_frames > 0:
|
||||
# 只有在连续丢失超过缓冲帧数时才重置
|
||||
if self.prisoner_missing_frames >= self.frame_buffer_prisoner:
|
||||
self.prisoner_detection_frames = 0
|
||||
self.prisoner_alert_active = False
|
||||
else:
|
||||
# 从未检测到犯人,保持状态
|
||||
pass
|
||||
# ==========================================
|
||||
# 警察检测
|
||||
# ==========================================
|
||||
if current_police_count > 0:
|
||||
# 检测到犯人框
|
||||
self.police_detection_frames += 1
|
||||
self.police_missing_frames = 0 # 重置丢失计数器
|
||||
|
||||
# 当检测累计达到阈值时,激活报警
|
||||
if self.police_detection_frames >= self.frame_thresh_police:
|
||||
self.police_alert_active = True
|
||||
else:
|
||||
# 未检测到犯人框
|
||||
self.police_missing_frames += 1
|
||||
|
||||
# 如果之前检测到手机,重置检测计数器
|
||||
if self.police_detection_frames > 0:
|
||||
# 只有在连续丢失超过缓冲帧数时才重置
|
||||
if self.police_missing_frames >= self.frame_buffer_police:
|
||||
self.police_detection_frames = 0
|
||||
self.police_alert_active = False
|
||||
else:
|
||||
# 从未检测到犯人,保持状态
|
||||
pass
|
||||
|
||||
|
||||
|
||||
alert_offset = 0
|
||||
|
||||
# A. 有犯人
|
||||
if self.prisoner_alert_active:
|
||||
duration_seconds = self.prisoner_detection_frames / self.fps
|
||||
current_frame_alerts.append(
|
||||
{
|
||||
'time': current_time_sec,
|
||||
'action': 'prisoner',
|
||||
'confidence': 1.0, # 固定为1.0(规则判定)
|
||||
'details': f"Detected for {duration_seconds:.1f}s"
|
||||
}
|
||||
)
|
||||
self.draw_alert(frame, "prisoner", (0, 0, 255), offset_y=alert_offset)
|
||||
alert_offset += 100
|
||||
|
||||
# ==========================================
|
||||
# 11. 统一显示当前帧所有警告(可替换原分层显示)
|
||||
# ==========================================
|
||||
debug_info = f" prisoner: {current_prisoner_count}"
|
||||
cv2.putText(frame, debug_info, (20, 40), cv2.FONT_HERSHEY_SIMPLEX, 0.7, (0, 255, 0), 2)
|
||||
|
||||
# 统一警告显示区
|
||||
alert_y_start = 150
|
||||
for i, alert in enumerate(current_frame_alerts):
|
||||
action = alert['action']
|
||||
details = alert.get('details', '')
|
||||
color = (0, 0, 255) # 默认红色警告
|
||||
|
||||
if action == 'prisoner':
|
||||
color = (255, 255, 255)
|
||||
|
||||
main_text = action
|
||||
if details:
|
||||
main_text += f" ({details})"
|
||||
|
||||
y_pos = alert_y_start + i * 50
|
||||
cv2.rectangle(frame, (20, y_pos - 40), (900, y_pos + 10), (0, 0, 0), -1)
|
||||
cv2.putText(frame, main_text, (30, y_pos), cv2.FONT_HERSHEY_SIMPLEX, 1.0, color, 2)
|
||||
|
||||
return {
|
||||
"image": frame,
|
||||
|
||||
"alerts":current_frame_alerts
|
||||
}
|
||||
|
||||
# ========================= WebSocket 服务线程 =========================
|
||||
class WebSocketSender(threading.Thread):
|
||||
def __init__(self, send_queue: queue.Queue, stop_event: threading.Event):
|
||||
super().__init__(daemon=True)
|
||||
self.send_queue = send_queue
|
||||
self.stop_event = stop_event
|
||||
|
||||
async def _ws_handler(self, websocket):
|
||||
ws_clients.add(websocket)
|
||||
try:
|
||||
async for _ in websocket:
|
||||
pass
|
||||
finally:
|
||||
ws_clients.discard(websocket)
|
||||
|
||||
async def _broadcaster(self):
|
||||
while not self.stop_event.is_set():
|
||||
try:
|
||||
msg = await asyncio.to_thread(self.send_queue.get, timeout=0.5)
|
||||
except queue.Empty:
|
||||
continue
|
||||
data = json.dumps(msg)
|
||||
dead = []
|
||||
for ws in list(ws_clients):
|
||||
try:
|
||||
await ws.send(data)
|
||||
except:
|
||||
dead.append(ws)
|
||||
for ws in dead:
|
||||
ws_clients.discard(ws)
|
||||
self.send_queue.task_done()
|
||||
|
||||
async def _run_async(self):
|
||||
async with websockets.serve(self._ws_handler, WS_HOST, WS_PORT):
|
||||
print(f"[INFO] WebSocket server started at ws://{WS_HOST}:{WS_PORT}")
|
||||
await self._broadcaster()
|
||||
|
||||
def run(self):
|
||||
asyncio.run(self._run_async())
|
||||
|
||||
|
||||
# ========================= RTSP 抓流线程 =========================
|
||||
class RTSPCaptureWorker(threading.Thread):
|
||||
def __init__(self, camera_cfg: CameraConfig, raw_queue: queue.Queue, stop_event: threading.Event):
|
||||
super().__init__(daemon=True)
|
||||
self.camera_cfg = camera_cfg
|
||||
self.raw_queue = raw_queue
|
||||
self.stop_event = stop_event
|
||||
|
||||
def run(self):
|
||||
cap = cv2.VideoCapture(self.camera_cfg.rtsp_url, cv2.CAP_FFMPEG)
|
||||
if not cap.isOpened():
|
||||
print(f"[ERROR] Cannot open RTSP: {self.camera_cfg.rtsp_url}")
|
||||
return
|
||||
print(f"[INFO] Capturing {self.camera_cfg.name} (ID:{self.camera_cfg.id})")
|
||||
while not self.stop_event.is_set():
|
||||
ret, frame = cap.read()
|
||||
if not ret:
|
||||
time.sleep(0.2)
|
||||
continue
|
||||
item = {
|
||||
"camera_id": self.camera_cfg.id,
|
||||
"camera_name": self.camera_cfg.name,
|
||||
"timestamp": time.time(),
|
||||
"frame": frame,
|
||||
}
|
||||
try:
|
||||
self.raw_queue.put(item, timeout=1.0)
|
||||
except queue.Full:
|
||||
pass
|
||||
cap.release()
|
||||
|
||||
|
||||
# ========================= 帧处理线程 =========================
|
||||
class FrameProcessorWorker(threading.Thread):
|
||||
def __init__(self,
|
||||
raw_frame_queue: "queue.Queue[Dict[str, Any]]",
|
||||
ws_send_queue: "queue.Queue[Dict[str, Any]]",
|
||||
stop_event: threading.Event):
|
||||
super().__init__(daemon=True)
|
||||
self.raw_queue = raw_frame_queue
|
||||
self.ws_queue = ws_send_queue
|
||||
self.stop_event = stop_event
|
||||
|
||||
self.video_writers: Dict[int, cv2.VideoWriter] = {}
|
||||
self.video_counts: Dict[int, int] = {}
|
||||
self.last_ts: Dict[int, float] = {}
|
||||
self.video_files: Dict[int, str] = {}
|
||||
|
||||
|
||||
os.makedirs(VIDEO_OUTPUT_DIR, exist_ok=True)
|
||||
|
||||
# 每个摄像头一个独立的 Kadian 检测器实例
|
||||
self.kadian_detectors: Dict[int, KadianDetector] = {}
|
||||
|
||||
def _get_writer(self, camera_id: int, frame) -> Tuple[cv2.VideoWriter, str]:
|
||||
if camera_id in self.video_writers:
|
||||
return self.video_writers[camera_id], self.video_files[camera_id]
|
||||
|
||||
h, w = frame.shape[:2]
|
||||
ts_str = datetime.now().strftime("%Y%m%d_%H%M%S")
|
||||
filepath = os.path.join(VIDEO_OUTPUT_DIR, f"{ts_str}_cam{camera_id}.mp4")
|
||||
writer = cv2.VideoWriter(filepath, cv2.VideoWriter_fourcc(*"mp4v"), RTSP_TARGET_FPS, (w, h))
|
||||
|
||||
self.video_writers[camera_id] = writer
|
||||
self.video_files[camera_id] = filepath
|
||||
self.video_counts[camera_id] = 0
|
||||
print(f"[INFO] New segment: {filepath}")
|
||||
return writer, filepath
|
||||
|
||||
def _close_segment_if_needed(self, camera_id: int):
|
||||
count = self.video_counts.get(camera_id, 0)
|
||||
if count >= FRAMES_PER_SEGMENT:
|
||||
writer = self.video_writers.get(camera_id)
|
||||
if writer is not None:
|
||||
writer.release()
|
||||
print(f"[INFO] Close segment: camera={camera_id}, file={self.video_files[camera_id]}")
|
||||
|
||||
self.video_writers.pop(camera_id, None)
|
||||
self.video_counts.pop(camera_id, None)
|
||||
self.last_ts.pop(camera_id, None)
|
||||
self.video_files.pop(camera_id, None)
|
||||
|
||||
def _encode_image_to_base64(self, image) -> str:
|
||||
ok, buf = cv2.imencode(".jpg", image)
|
||||
if not ok:
|
||||
raise RuntimeError("Failed to encode image to JPEG")
|
||||
return base64.b64encode(buf.tobytes()).decode("ascii")
|
||||
|
||||
def run(self):
|
||||
target_interval = 1.0 / RTSP_TARGET_FPS
|
||||
while not self.stop_event.is_set():
|
||||
try:
|
||||
item = self.raw_queue.get(timeout=0.5)
|
||||
except queue.Empty:
|
||||
continue
|
||||
|
||||
cam_id = item["camera_id"]
|
||||
ts = item["timestamp"]
|
||||
frame = item["frame"]
|
||||
|
||||
# 抽帧控制
|
||||
if ts - self.last_ts.get(cam_id, 0) < target_interval:
|
||||
self.raw_queue.task_done()
|
||||
continue
|
||||
self.last_ts[cam_id] = ts
|
||||
|
||||
# 获取检测器实例
|
||||
if cam_id not in self.kadian_detectors:
|
||||
self.kadian_detectors[cam_id] = KadianDetector()
|
||||
detector = self.kadian_detectors[cam_id]
|
||||
|
||||
# 2) 进行人脸识别(如果启用)
|
||||
current_face_alert = None
|
||||
face_results = []
|
||||
face_processing_time = 0
|
||||
if video_face_prison_biz is not None and FACE_RECOGNITION_ENABLED:
|
||||
try:
|
||||
# 处理当前帧 - 获取人脸识别结果
|
||||
processed_frame_for_face, face_results, face_processing_time = video_face_prison_biz.process_frame(
|
||||
frame.copy())
|
||||
|
||||
for result in face_results:
|
||||
if result['has_passed']:
|
||||
print(f"[INFO] 犯人带出: {result['passed_person_id']}")
|
||||
|
||||
# 记录当前帧人脸告警信息
|
||||
current_face_alert = {
|
||||
"person_name": result['passed_person_id'],
|
||||
"timestamp": ts
|
||||
}
|
||||
|
||||
except Exception as e:
|
||||
print(f"[WARN] 人脸识别处理失败: {e}")
|
||||
|
||||
# 执行检测
|
||||
result = detector.process_frame(frame.copy(), cam_id, ts)
|
||||
|
||||
result_img = result["image"]
|
||||
result_type = result["alerts"]
|
||||
|
||||
# 绘制人脸识别结果
|
||||
if video_face_prison_biz is not None and face_results:
|
||||
result_img = video_face_prison_biz.draw_detections(result_img, face_results)
|
||||
|
||||
# 添加人脸识别统计信息
|
||||
match_count = sum(1 for r in face_results if r['is_match'])
|
||||
face_info_text = f"Faces: {len(face_results)} | Matches: {match_count}"
|
||||
cv2.putText(result_img, face_info_text, (10, result_img.shape[0] - 20),
|
||||
cv2.FONT_HERSHEY_SIMPLEX, 0.6, (0, 255, 255), 2)
|
||||
|
||||
# 写视频
|
||||
writer, video_path = self._get_writer(cam_id, frame)
|
||||
writer.write(result_img)
|
||||
self.video_counts[cam_id] = self.video_counts.get(cam_id, 0) + 1
|
||||
|
||||
# 5) 通过 WebSocket 发送帧结果
|
||||
try:
|
||||
img_b64 = self._encode_image_to_base64(result_img)
|
||||
except Exception as e:
|
||||
print(f"[ERROR] Encode image failed: {e}")
|
||||
img_b64 = None
|
||||
|
||||
if img_b64 is not None:
|
||||
# 将abnormal_actions对象数组转换为字符串数组
|
||||
action_names = [action_info['action'] for action_info in result_type]
|
||||
|
||||
if current_face_alert is not None:
|
||||
action_names.append("face")
|
||||
|
||||
msg = {
|
||||
"msg_type": "frame",
|
||||
"camera_id": cam_id,
|
||||
"timestamp": ts,
|
||||
"result_type": action_names,
|
||||
"image_base64": img_b64,
|
||||
}
|
||||
try:
|
||||
self.ws_queue.put(msg, timeout=1.0)
|
||||
except queue.Full:
|
||||
print("[WARN] ws_send_queue full, drop frame message")
|
||||
|
||||
|
||||
|
||||
self._close_segment_if_needed(cam_id)
|
||||
self.raw_queue.task_done()
|
||||
|
||||
self.video_counts[cam_id] = self.video_counts.get(cam_id, 0) + 1
|
||||
|
||||
# 清理
|
||||
for w in self.video_writers.values():
|
||||
w.release()
|
||||
|
||||
|
||||
# ========================= 服务主类 =========================
|
||||
class RTSPService:
|
||||
def __init__(self, config_path: str = "config.yaml"):
|
||||
with open(config_path, "r", encoding="utf-8") as f:
|
||||
cfg = yaml.safe_load(f)
|
||||
self.cameras = [CameraConfig(id=c["id"], name=c.get("name", f"cam_{c['id']}"), rtsp_url=c["rtsp_url"])
|
||||
for c in cfg.get("cameras", [])]
|
||||
|
||||
self.stop_event = threading.Event()
|
||||
self.raw_queue = queue.Queue(maxsize=500)
|
||||
self.ws_queue = queue.Queue(maxsize=1000)
|
||||
|
||||
self.capture_workers = []
|
||||
self.processor = FrameProcessorWorker(self.raw_queue, self.ws_queue, self.stop_event)
|
||||
self.ws_sender = WebSocketSender(self.ws_queue, self.stop_event)
|
||||
|
||||
def start(self):
|
||||
self.ws_sender.start()
|
||||
self.processor.start()
|
||||
for cam in self.cameras:
|
||||
w = RTSPCaptureWorker(cam, self.raw_queue, self.stop_event)
|
||||
w.start()
|
||||
self.capture_workers.append(w)
|
||||
print("[INFO] Kadian RTSP Service started")
|
||||
|
||||
def stop(self):
|
||||
self.stop_event.set()
|
||||
self.raw_queue.join()
|
||||
self.ws_queue.join()
|
||||
for w in self.capture_workers:
|
||||
w.join(timeout=2.0)
|
||||
self.processor.join(timeout=2.0)
|
||||
self.ws_sender.join(timeout=2.0)
|
||||
print("[INFO] Service stopped")
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
service = RTSPService("config.yaml")
|
||||
service.start()
|
||||
try:
|
||||
while True:
|
||||
time.sleep(1)
|
||||
except KeyboardInterrupt:
|
||||
service.stop()
|
||||
Reference in New Issue
Block a user