Back in February, Google announced its AI model Gemini 1.5 Pro with its 1 million token context window resulting in a need to understand our AI detectors accuracy.
This quick study looks at 1000 Gemini 1.5 Pro generated text results to answer if it is able to be detected.
Try our AI Detector here.
In order to evaluate the detectability of Gemini 1.5 Pro, we prepared a dataset of 1000 Gemini 1.5 Pro generated text samples.
For AI-text generation, we used Gemini 1.5 Pro based on three approaches given below:
To evaluate the efficacy we used the Open Source AI Detection Efficacy tool that we have released:
Originality.AI has two models namely Model 3.0 Turbo and Model 2.0 Standard for the purpose of AI Text Detection.
The open-source testing tool returns a variety of metrics for each detector you test, each of which reports on a different aspect of that detectors performance, including:
If you'd like a detailed discussion of these metrics, what they mean, how they're calculated, and why we chose them, check out our blog post on AI detector evaluation. For a succinct upshot, though, we think the confusion matrix is an excellent representation of a model's performance.
Below is an evaluation of both the models on the above dataset.
For this smaller test to be able to identify the ability for Originality.ai’s AI detector to identify Gemini 1.5 Pro content we look at True Positive Rate or the % of the time that the model correctly identified AI text as AI out of a 1000 sample Gemini 1.5 Pro content.
Model 2.0 Standard:
Model 3.0 Turbo:
Content creators, editors and writers are entering uncharted territory with regard to AI. In the past, it was easy to identify AI-generated content at-a-glance. However, with new developments in machine learning and natural language processing, the waters become even more murky.