Beyond the Myths: The Truth About Vision AI Security, Ethics, and Privacy

She didn’t see him—AI could have.

On a late hospital shift, an exhausted nurse looked up and locked eyes with someone she never expected to see—a former patient with a history of aggression. He wasn’t supposed to be there. A restraining order was in place. But security had no idea.

This isn’t about surveillance—it’s about prevention. AI for good means identifying real threats before harm happens. “When you think about security, AI, biometrics, and behavior detection, a lot of red flags may go up,” said Ben Thomas, host of Pro AV Today. A longtime technology analyst, Ben has seen how misinformation distorts public perception. “And a lot of times, those red flags go up because there’s a lot of miseducation out there.”

In our latest episode of Innovation Obsessed, Ben Thomas and Shawn Fontenot, VP of Global Marketing at Oosto, tackle the biggest misconceptions about AI security. They explore how misinformation stalls progress and why AI, when done right, is a force for good—protecting people before it’s too late.

Responsible AI Considerations for Security

Responsible AI narrows the focus to specific known threats.

“We’re not looking for everyone,” said Shawn. “In fact, when we start with a customer, we start with an empty database. We don’t scrape data from the internet. We don’t track people’s movements. We don’t even know who you are unless you’re on a security watchlist—like a known bad actor, a banned individual, or someone with a restraining order.”

With Oosto, customers populate and control watchlists and create rules to uncover safety and security threats. The system completely ignores individuals not on those watchlists.

Real-World Example: AI in Highly Regulated Industries

Casinos, some of the most heavily regulated security environments in the world, rely on facial recognition for targeted threat detection and for identifying self-excluders: people who want help in breaking their addiction to gambling.

“Casinos aren’t just sending out a squad of people when an alert pops up,” explained Shawn. “They go through multiple layers of verification before taking action. It’s about identifying repeat offenders, fraudsters, and those banned from the premises—not tracking guests.”

This model is used in airports, hospitals, and critical infrastructure where security matters most. Instead of relying on human guards to remember hundreds or thousands of faces—a process prone to error—AI helps ensure only pre-identified threats are flagged.

Addressing Bias and Accuracy

AI is only as good as the data that built it, and bias arises when algorithms are trained on limited or unrepresentative datasets.

“Some applications of facial recognition have given the industry a bad name,” said Shawn. “But the truth is, most people we’re looking for aren’t standing still, looking at a camera. They’re in motion, in crowds, in bad lighting. Our AI is trained for those conditions—not just perfect, static images.”

In this episode, Ben and Shawn discuss what makes ethical AI security different:

  • Diverse Training Data: Oosto’s AI is trained on millions of images across all demographics, ensuring real-world accuracy, not just lab-tested results.
  • Continuous Testing & Validation: Unlike AI systems that rely solely on controlled, static images, Oosto’s technology is tested in real-world environments—where conditions are far more unpredictable.
  • Independent Regulatory Scrutiny: Oosto’s AI is used in 95% of Las Vegas casinos, which must re-evaluate and validate facial recognition accuracy every six months under Nevada Gaming Commission regulations.

“Our models are constantly being refined and tested in the real world, which is why we continue to meet and exceed regulatory standards,” Shawn explained.

Privacy Considerations

“I always tell folks, you watch CSI and they keep saying ‘enhance, enhance’ until they zoom in all the way to a credit card number. That’s just not how it works,” Ben said with a laugh.

A major misconception about AI security is that it collects, stores, and tracks personal data. Ethical AI systems do not operate that way.
“We don’t even store personal data. When a face is scanned, it’s converted into a mathematical vector—there’s nothing to steal,” explained Shawn.

They discuss how responsible AI security follows strict privacy-first principles:

No open-source data scraping: Unlike some facial recognition providers, Oosto does not train its AI on social media images or public databases.

No mass data retention: AI security systems must follow data minimization principles, ensuring data is only stored as long as necessary.

Human Oversight on Every Alert: Every alert is verified by human reviewers before any action is taken.

Real-World Example: Protecting Healthcare Workers

One of the most practical applications of AI security today is in hospitals, where staff—including nurses and doctors—face growing threats of violence.

“If you’re a nurse, and there’s someone you have a restraining order against, you shouldn’t have to worry about whether security will recognize them at the door,” said Shawn.

Instead of relying on human memory and manual monitoring, AI can alert security the moment a flagged individual enters the facility—giving staff time to take preventative action before an incident occurs.

AI Security: A Force for Good

The conversation around AI security shouldn’t be framed as a battle between privacy and protection—it should be about how we use technology to create safer environments while upholding ethical standards.

As Shawn focuses on in the episode, Oosto’s mission is to harness Vision AI for Good—a commitment to using artificial intelligence responsibly to protect people, businesses, and communities without compromising privacy.

Ethical AI security is about proactive protection, not surveillance. It’s about ensuring healthcare workers, retail employees, university students, corporate staff, and the public are safeguarded from known threats—without indiscriminate monitoring or data misuse.

By prioritizing transparency, responsible deployment, and privacy-first design, AI security becomes an essential tool for preventing harm before it happens.

With the right approach, security and privacy don’t have to be at odds. We can, and must, have both.

Recent Episodes

In the rapidly advancing field of cancer immunotherapy, accurately modeling the tumor microenvironment (TME) has become essential to improving the predictive power of preclinical drug testing. As immune-modulating therapies surge forward, with over 4,000 immune modulators in development globally, scientists are refining assay technologies that maintain the complexity of patient-specific tumor biology. In vitro platforms…

As cancer immunotherapy continues to reshape treatment landscapes, fine-tuning T-cell responses has become a critical frontier. Recent advances in 3D organoid models and high-content imaging are enabling scientists to closely mimic patient-specific tumor environments—unlocking insights into how T cells behave, respond, and falter under immune checkpoint blockade. With over 4,000 immune modulators in clinical…

As immunotherapy revolutionizes cancer treatment, the need for physiologically relevant preclinical models becomes more urgent than ever. Despite the success of immune checkpoint inhibitors, a large majority of patients fail to achieve long-lasting responses, prompting researchers to explore more complex and predictive assays. The cancer immunity cycle, first described in 2013, remains a central framework…