What can CV-based gun detection actually detect? Computer vision gun detection systems analyse video feeds to identify firearms in the scene. The detection models are trained on image datasets of handguns, rifles, and other weapons, and flag frames where a weapon is detected with a confidence score above a configured threshold. Current systems detect three categories: visible firearms (weapons held or carried openly), brandished weapons (weapons pointed or raised in a threatening posture), and abandoned weapons (weapons placed on surfaces). Detection accuracy varies significantly across categories: Detection Category Controlled Test Accuracy Real-World Accuracy Key Challenge Visible firearm (open carry) 90β95% 75β85% Occlusion by clothing, bags Brandished weapon 85β92% 70β80% Pose variation, distance Abandoned weapon 80β88% 65β75% Small object, varied backgrounds The gap between controlled test accuracy and real-world accuracy is the central challenge. Manufacturing environments add complexity: PPE (hard hats, safety vests, gloves), carried tools (drills, pneumatic tools) that visually resemble weapons, and variable lighting from welding, machinery, and loading dock doors. What causes false positives in manufacturing settings? False positives β non-weapon objects flagged as weapons β are the primary operational concern. In manufacturing environments, we have documented false positives triggered by: power drills, pneumatic nail guns, spray paint guns, hand-held barcode scanners, black-coloured L-shaped tools, and dark umbrella handles. The visual features that distinguish a handgun from a drill at 15 metres and 1080p resolution are subtle β both are dark, L-shaped, handheld objects. Reducing false positives requires environment-specific model tuning: retraining or fine-tuning the detection model with images from the specific deployment environment, including the common tool types that trigger false positives. We typically collect 2β4 weeks of video from the deployment site, annotate false positive events, and use these as negative examples during fine-tuning. This process reduces false positive rates by 40β60% compared to off-the-shelf models. For the broader comparison of machine vision and computer vision approaches to manufacturing safety, our analysis of machine vision vs computer vision covers the architectural differences. How should threat detection be deployed responsibly? Deploying AI-based threat detection in manufacturing facilities raises ethical and practical considerations. False positive alerts trigger security responses β armed response teams, lockdowns, evacuations β that have real safety and productivity costs. A system with a 2% false positive rate processing 50 cameras at 1 frame per second generates approximately 86,400 false detections per day, of which the alert system must filter the vast majority before human review. Our deployment architecture uses a two-stage detection pipeline: a fast, sensitive first-stage detector flags potential weapons with high recall (catching most real weapons at the cost of many false positives), followed by a slower, precise second-stage classifier that analyses flagged frames at higher resolution and with temporal context (is the object consistent across multiple frames?). The second stage reduces false positive alerts to fewer than 5 per day per 50-camera system β a manageable review workload for security personnel. Privacy and legal compliance require documented policies: what data is collected, how long it is retained, who has access, and what actions are triggered by detections. Manufacturing environments with unionised workforces may require negotiation with labour representatives before deploying surveillance AI. We advise clients to engage legal and HR stakeholders early in the deployment planning process. How do you evaluate gun detection system performance? Evaluating gun detection system performance requires metrics beyond simple accuracy. The relevant metrics for a deployment decision: True positive rate (recall): What percentage of actual weapons does the system detect? For threat detection, recall is the priority metric β a missed weapon is a safety failure. Target: β₯95% on visible firearms in the deployment environment (after environment-specific tuning). False positive rate: How many non-weapon objects trigger alerts? Measured per camera per hour. Target: <0.5 false alerts per camera per hour after tuning (approximately 1 false alert every 2 hours per camera). Detection latency: How quickly does the system alert after a weapon appears in frame? Measured from weapon appearance to alert generation. Target: <3 seconds for edge-processed, <10 seconds for cloud-processed. Environmental robustness: Does accuracy degrade under specific conditions? Test across: day/night, indoor/outdoor, crowded/empty, winter clothing/summer clothing, various tool-carrying scenarios specific to the manufacturing environment. We structure evaluation in three phases: (1) laboratory testing with staged scenarios (controlled, repeatable, measures baseline capability), (2) passive deployment (system runs and logs detections but does not generate alerts β measures real-world accuracy without operational disruption), (3) active deployment with human review (alerts are generated and reviewed by security staff β measures operational effectiveness including response time and workflow integration). The passive deployment phase typically lasts 4β6 weeks and is essential for understanding the false positive profile in the specific environment before going live. Integration with existing security infrastructure (access control systems, alarm panels, VMS platforms) determines whether the detection system produces actionable outcomes or creates an additional monitoring burden. We design integrations that inject weapon detection alerts into the existing security monitoring workflow β appearing on the same screens, triggering the same escalation procedures β rather than requiring security staff to monitor a separate system. This integration approach reduces alert response time by 40β60% compared to standalone monitoring.