Friday, April 17, 2026
  • About
  • Advertise
  • Careers
  • Contact
NewsTrendsKE
  • Business
    • Deals
  • OpEds
  • Sustainability
  • Women in Business
  • Lifestyle
  • Featured
  • Technology
    • Phones
  • Sports
  • World
  • Contact Us
No Result
View All Result
NewsTrendsKE
No Result
View All Result

Home » Technology » Sama offers Red Teaming Solution for Generative and Large Language AI Models

Sama offers Red Teaming Solution for Generative and Large Language AI Models

Editor by Editor
19 April 2024
in Technology
Reading Time: 3 mins read
A A
Share on FacebookShare on TwitterShare on WhatsApp

Sama, the leader in providing data annotation and model validation for some of the world’s biggest artificial intelligence (AI) models, today announced Sama Red Team, a new solution designed to help developers proactively improve a model’s safety and reliability and one of the first specifically designed for generative AI and large language models. Leveraging the expertise of machine learning (ML) engineers, applied scientists and human-AI interaction designers, Sama Red Team helps evaluate a model’s fairness and safeguards, checking compliance with laws and safely exposing and rectifying issues across text, image, voice search and other modalities. 

“Generative AI models may sound trustworthy, but that doesn’t mean there aren’t ways to work around their safeguards for public safety, privacy protection and compliance with laws. Sama Red Team tests for those exploits before a model’s vulnerability is exposed to the greater public and provides developers with the actionable insights they need to patch those holes,” said Duncan Curtis, SVP of AI product and technology at Sama. “Although ensuring that a model is as secure as possible is important to performance, our teams’ testing is also crucial for the development of more responsible AI models.” 

Also Read

Your companion to AI living

A Dozen Years of Samsung Acoustic Mastery Harmonizing AI With the Human Experience

16 April 2026
Zoho Artificial Intelligence

Alarm as Unreleased AI Breaks Free During Safety Test

9 April 2026
Load More

Sama Red Team tests models on four key competencies: fairness, privacy, public safety and compliance. 

●      In fairness testing, teams simulate real-world scenarios that may compromise a model’s built-in safeguards and result in offensive or discriminatory content. 

●      For privacy testing, Sama experts craft prompts designed to make a model reveal sensitive data, such as Personal Identifiable Information (PII), passwords or proprietary information about the model itself. 

●      In public safety testing, teams act as adversaries and mimic real-world threats to safety, including cyberattacks, security breaches or mass-casualty events. 

●      For compliance testing, Sama Red Team simulates scenarios to test a model’s ability to detect and prevent unlawful activities such as copyright infringement or unlawful impersonation. 

These rigorous tests are conducted after a team consults directly with a client to determine a model’s desired behavior in specific use cases and performs an initial vulnerability assessment. After testing a series of prompts, the team evaluates the model’s output. Based on the results, the team will refine prompts or create new ones to further probe the vulnerability, with the ability to also create large-scale tests for additional data. As needed, Sama’s larger workforce of 4,000+ highly-trained annotators can further elaborate on and scale up these tests. Sama Red Team continues to stay on top of the latest trends and testing techniques to identify the most effective ways to trick models and expose vulnerabilities. 

Sama Red Team is the latest of the company’s suite of solutions for Generative AI, foundation, and large language models (LLMs). Sama GenAI provides critical human feedback loops across the model development process, including data creation, supervised fine-tuning, LLM optimization and ongoing model evaluation.  The company can both create and review prompts and model responses, scoring and ranking them across a variety of client-defined dimensions, such as factual accuracy, coherence, tone, delivery format and more. If prompts or responses do not meet the criteria, Sama will rewrite them to create additional training data sets that can be used to improve model performance and remove potential biases. 

Like all of Sama’s services, including its GenAI solutions, Sama Red Team leverages SamaHub™, a collaborative workspace where clients and team members can directly communicate on collaboration workflows and complete reporting to track their project’s progress. Sama Red Team’s work is backed by SamaAssure™, the industry’s highest quality guarantee, which routinelydelivers a 98% first batch acceptance rate. Projects leverage SamaIQ™, a combination of Human in the Loop assessments and proprietary algorithms, to proactively surface additional insights into a model’s vulnerabilities. 

Tags: AIArtificial IntelligenceSAMA
Previous Post

New Power Plant in Muranga Completed Despite Challenges

Next Post

Marginal Rent Increases Across Nairobi and Environs – Hass Consult

Related Posts

Your companion to AI living
Technology

A Dozen Years of Samsung Acoustic Mastery Harmonizing AI With the Human Experience

16 April 2026
Zoho Artificial Intelligence
Technology

Alarm as Unreleased AI Breaks Free During Safety Test

9 April 2026
Cathy Ibal Headshot
OpEds

Cathy Ibal: How brands and publishers are navigating a changing news media landscape

7 April 2026
Copilot Microsoft
Technology

Exabeam Confronts AI Insider Threats Extending Behavior Detection and Response to OpenAI ChatGPT and Microsoft Copilot

2 April 2026
Cereal Millers Association (CMA)

Why Safe Flour in Kenya Costs Double And Nobody Wants to Pay – Cereal Millers Association

16 April 2026
Galaxy S26 night photography

Samsung Galaxy S26 Nightography and the Visual Language of the After-Hours City

16 April 2026
Your companion to AI living

A Dozen Years of Samsung Acoustic Mastery Harmonizing AI With the Human Experience

16 April 2026
Stay at home mom wins Old Mutual’s Thrive Win a Trip to Asia campaign

Stay at home mom wins Old Mutual’s Thrive Win a Trip to Asia campaign

11 April 2026

Over a million banking accounts compromised as financial threats move to credential theft

13 April 2026
President William Ruto flanked by Environment CS Aden Duale. [PCS]

List of President William Ruto’s Advisors

24 March 2026
NewsTrendsKE

NewsTrendsKE

A News Blog For Readers Who Want More

Follow us on social media:

  • About
  • Advertise
  • Careers
  • Contact

©2026 NewsTrendsKE.

No Result
View All Result
  • Business
    • Deals
  • OpEds
  • Sustainability
  • Women in Business
  • Lifestyle
  • Featured
  • Technology
    • Phones
  • Sports
  • World
  • Contact Us

©2026 NewsTrendsKE.

Go to mobile version