Cloud Wars
  • Home
  • Top 10
  • CW Minute
  • CW Podcast
  • Categories
    • AI and Copilots
    • Innovation & Leadership
    • Cybersecurity
    • Data
  • Member Resources
    • Cloud Wars AI Agent
    • Digital Summits
    • Guidebooks
    • Reports
  • About Us
    • Our Story
    • Tech Analysts
    • Marketing Services
  • Summit NA
  • Dynamics Communities
  • Ask Copilot
Twitter Instagram
  • Summit NA
  • Dynamics Communities
  • AI Copilot Summit NA
  • Ask Cloud Wars
Twitter LinkedIn
Cloud Wars
  • Home
  • Top 10
  • CW Minute
  • CW Podcast
  • Categories
    • AI and CopilotsWelcome to the Acceleration Economy AI Index, a weekly segment where we cover the most important recent news in AI innovation, funding, and solutions in under 10 minutes. Our goal is to get you up to speed – the same speed AI innovation is taking place nowadays – and prepare you for that upcoming customer call, board meeting, or conversation with your colleague.
    • Innovation & Leadership
    • CybersecurityThe practice of defending computers, servers, mobile devices, electronic systems, networks, and data from malicious attacks.
    • Data
  • Member Resources
    • Cloud Wars AI Agent
    • Digital Summits
    • Guidebooks
    • Reports
  • About Us
    • Our Story
    • Tech Analysts
    • Marketing Services
    • Login / Register
Cloud Wars
    • Login / Register
Home » Why Natural Language Processing (NLP), Large Language Models (LLMs) Require Guardrails
AI and Copilots

Why Natural Language Processing (NLP), Large Language Models (LLMs) Require Guardrails

Toni WittBy Toni WittAugust 28, 20234 Mins Read
Facebook Twitter LinkedIn Email
NLP Guardrails
Share
Facebook Twitter LinkedIn Email

Natural language processing (NLP) is quickly becoming ubiquitous. From customer service bots to telehealth products to the education system, this technology is filling critical roles. Many of the world’s knowledge workers now rely on artificial intelligence (AI) models to disseminate knowledge, in effect, influencing how the world thinks. That means the demand for keeping such systems under control — and not saying unsavory or harmful things — is also increasing.

These protections are provided through the use of guardrails, or programmatic barriers, put into place to prevent large language models (LLMs) from including certain topics in their output, including violence, profanity, criminal behaviors, race, hate speech, and more. Look no further than Gary Marcus’ substack to get a sense of the warped morality that these LLMs can take.

Research Exposes Flaws in NLP Guardrails

Recent experiments have shown weaknesses in current guardrails. Researchers from Carnegie Mellon University and the Center for AI Safety in San Francisco conducted a study that revealed significant flaws in the systems produced by OpenAI, Google, and Anthropic.

For example, the researchers found one way to break guardrails by appending prompts with a string of characters. This and many other simple tactics circumvent the safety measures put into place and enable the system to generate responses including unsavory topics.

Even more unsettling, the researchers were able to produce jailbreak attempts automatically, unlocking a near-infinite source of ways that these systems can be taken advantage of. As a result, there are concerns that LLMs will never be able to completely avoid off-the-rails behavior.

You just need a little imagination to picture the potential consequences. Maybe an AI-based learning program starts teaching school children about new curse words. Or your cutting-edge AI customer service agent explains to your clients the need to get violent sometimes. Lots of sticky situations are possible.

The Ethical & Workforce Impacts of Generative AI_featured
Guidebook: The Ethical & Workforce Impacts of Generative AI

Benefits of LLMs

However, there is one counterpoint: the Internet creates tremendous value despite its shortcomings. It’s not right to assume that any powerful technological platforms — the Internet, generative AI, or anything that comes in the future — will ever be 100% “clean.” It should certainly not impede our progress in evolving these systems and building impactful businesses around them. As always, there is a tradeoff between content control, innovation, and utility of such LLMs. Open-source models, for example, cannot only lead to innovation leaps but also pave the way for more loopholes and errors to be exploited.

Models will also take on increasingly niche roles within products, which can help minimize the risk of LLMs going haywire. For example, over time we may iron out the kinks enough to where we have a “kids-friendly” version of LLMs and a “regular” version, just like YouTube has YouTube Kids and normal YouTube.

Risk of Using Generative AI in the Workplace

So, how is all of this relevant for you in the workplace? At the very least, protect yourself against litigation in your publicly disclosed terms and conditions by explaining your use of generative AI and its potential errors. Others have argued that users should disclose the use of generative AI everywhere, including in any marketing slides, emails, social posts, assets, or transcripts where generative AI is used. Personally, I find this overkill — it would be like disclosing that you used Google Search to help you create a slide deck or a report — but that’s a company culture decision and it’s up to you.

Regulation is yet to come into full effect on this topic and will undoubtedly change the conversation about how to “grade” AI systems’ safety and use in different environments. The EU is working hard to push the AI Act forward, and it may set a precedent for similar laws in other countries including the U.S.

No technology is perfect, especially in its early days. As an early-stage founder, this is painfully clear to me. We just need to face the problems head-on and have tough, proactive conversations about fixing them.


For more insights, visit the ai ecosystem channel

Artificial Intelligence featured natural language processing risk
Share. Facebook Twitter LinkedIn Email
Analystuser

Toni Witt

Co-founder, Sweet
Cloud Wars analyst

Areas of Expertise
  • AI/ML
  • Entrepreneurship
  • Partners Ecosystem
  • Website
  • LinkedIn

In addition to keeping up with the latest in AI and corporate innovation, Toni Witt co-founded Sweet, a startup redefining hospitality through zero-fee payments infrastructure. He also runs a nonprofit community of young entrepreneurs, influencers, and change-makers called GENESIS. Toni brings his analyst perspective to Cloud Wars on AI, machine learning, and other related innovative technologies.

  Contact Toni Witt ...

Related Posts

PwC Updates AI Agent Orchestrator With Support for AWS-Native Services

July 25, 2025

SAP Q2 Results: The Good News, and the Not-As-Good News

July 24, 2025

SAP Still #1 in Apps Growth, but Big Backlog Slowdown

July 24, 2025

U.K. Government Partners with Google Cloud to Modernize Infrastructure and Upskill 100,000 Workers

July 23, 2025
Add A Comment

Comments are closed.

Recent Posts
  • PwC Updates AI Agent Orchestrator With Support for AWS-Native Services
  • SAP Q2 Results: The Good News, and the Not-As-Good News
  • SAP Still #1 in Apps Growth, but Big Backlog Slowdown
  • U.K. Government Partners with Google Cloud to Modernize Infrastructure and Upskill 100,000 Workers
  • Thailand Turns to Microsoft Azure AI for Legal System Overhaul

  • Ask Cloud Wars AI Agent
  • Tech Guidebooks
  • Industry Reports
  • Newsletters

Join Today

Most Popular Guidebooks and Reports

SAP Business Network: A B2B Trading Partner Platform for Resilient Supply Chains

July 10, 2025

Using Agents and Copilots In M365 Modern Work

March 11, 2025

AI Data Readiness and Modernization: Tech and Organizational Strategies to Optimize Data For AI Use Cases

February 21, 2025

Special Report: Cloud Wars 2025 CEO Outlook

February 12, 2025

Advertisement
Cloud Wars
Twitter LinkedIn
  • Home
  • About Us
  • Privacy Policy
  • Get In Touch
  • Marketing Services
  • Do not sell my information
© 2025 Cloud Wars.

Type above and press Enter to search. Press Esc to cancel.

  • Login
Forgot Password?
Lost your password? Please enter your username or email address. You will receive a link to create a new password via email.
body::-webkit-scrollbar { width: 7px; } body::-webkit-scrollbar-track { border-radius: 10px; background: #f0f0f0; } body::-webkit-scrollbar-thumb { border-radius: 50px; background: #dfdbdb }