Is Llama 3.1 Content Detectable? – Originality.AI (original) (raw)

Meta released Llama 3.1 alongside its prized 405B model, achieving state-of-the-art performance across key benchmarks and becoming the first-ever open-sourced frontier model, marking a major milestone in open-source AI development.

It’s the first time an open-source AI model matches or outperforms top closed AI models like OpenAI’s GPT-4o. By offering a private, customizable alternative to closed AI systems, Meta is enabling anyone to create their own tailored AI and with this, it’s more necessary than ever to understand the accuracy of AI detectors.

This brief study looks at 1000 Llama 3.1 generated text results to find out whether the Originality.ai AI Detector can detect Llama 3.1.

TL;DR — Is Llama 3.1 AI-Generated Content Detectable?

Try the Originality.ai AI Detector. Then, learn about AI content detection accuracy and Originality’s exceptional performance in a meta-analysis of third-party studies.

Note: Standard is now retired. Get the latest details about model updates in our guide on which AI detector model is best for you!

Dataset

To evaluate the detectability of Llama 3.1, we prepared a dataset of 1000 Llama 3.1-generated text samples.

The Method: gathering AI-generated text data

For AI-text generation, we used Llama 3.1 based on three approaches:

  1. Rewrite prompts: The first approach generated content by providing the model with a custom prompt and articles (probably generated by LLMs) as a reference to rewrite from (450 Samples).
  2. Rewrite human-written text: The second approach involved generating content with the aim of finding out whether AI rewrites of human-written text could bypass AI detection. The samples for this approach were fetched from an open-source dataset (325 Samples).
    1. One-Class Learning for AI-Generated Essay Detection
      1. Paper: https://www.mdpi.com/2076-3417/13/13/7901
      2. Dataset: https://github.com/rcorizzo/one-class-essay-detection
  3. Write articles from scratch: The third approach, generated articles from scratch based on topics ranging from fiction to nonfiction, such as history, medicine, mental health, content marketing, social media, literature, robots, the future, etc. (225 Samples).

Evaluation

To evaluate the efficacy, we used the Open Source AI Detection Efficacy tool that we released:

A brief overview of the Originality.ai AI detection models

Originality.ai has three models, 3.0.0 Turbo, 2.0.1 Standard, and 1.0.0 Lite,for AI text detection.

For additional information on each of these models, check out our AI detector and read our AI detection accuracy guide.

Evaluating the AI detection models

The open-source testing tool returns a variety of metrics for each detector you test, each of which reports on a different aspect of that detector’s performance, including:

For a detailed discussion of these metrics, what they mean, how they're calculated, and why we chose them, check out our blog post on AI detector evaluation. For a succinct snapshot, the confusion matrix is an excellent representation of a model's performance.

Below is an evaluation of all these models on the above dataset.

Confusion Matrix

Figure 1. Confusion Matrix on AI only dataset with Model 1.0.0 Lite

Figure 2. Confusion Matrix on AI only dataset with Model 2.0.1 Standard

Figure 3. Confusion Matrix on AI only dataset with Model 3.0.0 Turbo

Results of the Evaluation :

For this smaller test to identify the ability of Originality.ai’s AI detector to detect Llama 3.1 content, we reviewed the True Positive Rate or the percentage (%) of time that the model correctly identified AI text as AI out of 1000 samples of Llama 3.1 content.

1.0.0 Lite:

2.0.1 Standard:

3.0.0 Turbo:

Final Thoughts

Overall, Originality.ai continues to demonstrate an outstanding capability to identify AI-generated content, including the latest releases of AI models such as OpenAI’s GPT-4o, Claude 3.5, Gemini 1.5 Pro, and GPT-4o-mini.

Each of Originality.ai’s AI detection models detected Llama 3.1 with exceptional accuracy from 3.0.0 Turbo with 99.6% accuracy to 2.0.1 Standard with 98.8% accuracy, and 1.0.0 Lite with 99.1% accuracy.