Me vs. AI: A Blogging Experiment for Content Creators

15–22 minutes

It’s a quiet Friday afternoon, I’ve had a busy working week, full of writing software, writing reports, reading reports – you know the drill. So, I thought it would be interesting to let the AI write something for me. As we are being encouraged to use these AI tools more and more, why not do a quick experiment? I just want to know if an AI could replace me! Would you consider using AI to help you create content?

I, like many people, sometimes have very busy weeks, or I want to write about complex topics and I just don’t know how to start. Using an AI to help me, could be a great super beneficial, but I don’t want using AI to make my content feel disjoint from my other work. In this post, I’m going to run a quick experiment. Let’s see what happens when I ask ChatGPT to write a blog post for me. I’ll analyse the differences between a post I have written and a post an LLM has written. We can look through the results together, and perhaps this will help you understand how you might leverage AI in your content creation. So let’s get started.

Can we use AI, such as LLMs to speed up our content creation, could they replace us entirely?

TLDR;

With an example of the style of content I wanted like to create, and a description of the topic that I wanted to write about, ChatGPT was able to create a convincing and pretty good post. There were some issues with American vs British spelling, and the examples it used were a little flat and used frequently in other content. But on the whole – it was convincing! Whether you believe in using these tools for improving productivity or whether you think they will be the killer of human creativity, they are getting pretty good and that can’t be ignored.

The Experiment

This is a very professional, very rigorous experiment, so let’s do what any good scientist should do and outline the experiment, to make this a somewhat fair, albeit short, experiment.

Hypothesis

ChatGPT can write content which passes for my writing.

Tools

We don’t need much, a blog topic and a method of writing the blog. I’m going to use ChatGPT, which is the GPT-4-turbo model (I’m using a free account). For the blog topic, I will describe what precision, recall and accuracy mean for ML model evaluation. Don’t worry if you have never heard of these words, you’re about to be told three different ways. I’ll keep it short!

Methodology
  1. I will write a very short blog, how I would normally (in a fraction of the time). No AI will be used – I promise, all my own words and some help from Wikipedia.
  2. Then I’ll go and ask ChatGPT (the free one!) to do the same thing. I’ll give it no other prompting other than “Please write a short blog post for an AI explainer website explaining precision, recall and accuracy”
  3. Then I’ll ask it again, in a new session, to do the same thing, but this time I will share one of my previous posts (like this one on AI agents) and ask it to write the post in the same style.
Metrics

Measuring whether two pieces of text are from the same author is both subjective and complicated to measure. This is a fundamental challenge with LLM Evaluation (which I will not dwell on here). So for this experiment, I will read and decide how similar they are to what I feel my writing style is – based on a previous post of mine. I will also leverage ChatGPT to decide how similar the three blog posts are to this same previous post. For each post, and each method of marking we will get a score out of 10.

Let’s see what it does! Can ChatGPT sound like me? And can we tell?

Post 1: I will write the post first

Everything below I have written, without AI aid.



This is probably not my finest work, but gloss over what I’m saying, you’re about to read it two more times! I’ve tried sticking to my usual format – a longer subheading, a few metaphors and a friendly tone.

Post 2: Lightly Prompted ChatGPT Created Post

Let’s try the LLM, with only a little bit of guidance. Here we go – my prompt to ChatGPT is:

I’m writing a blog for an AI explainer website, which explains what accuracy recall and precision are. I don’t need titles – I just need an introduction and an explanation of accuracy recall and precision.

The AI generated blog is below.



The content is correct and is clearly written. I do feel like it is missing something. There are no metaphors to help you understand, the text doesn’t feel like it is written to ‘you’ as a reader. Essentially feels like it is coming from a textbook.

Post 3: Let’s get the AI generate it in the style of me

In ChatGPT’s defence, my prompt was lightweight on details and I didn’t give it a helping hand for style and tone. So this time, I’m giving it a link to one of my blogs as a style inspiration. I will just add the same formatting I normally do – but I will not tweak the words.

I’m writing a blog for an AI explainer website, which explains what accuracy recall and precision are. I don’t need titles – I just need an introduction and an explanation of accuracy recall and precision. Could you copy the style of the post body from here: https://thealgorithmedit.com/2025/02/26/understanding-ai-agents-types-and-functions/

Everything below is written by AI:


In the realm of artificial intelligence and machine learning, evaluating the performance of models is crucial to ensure their effectiveness and reliability. Three fundamental metrics—accuracy, precision, and recall—serve as the cornerstone for this evaluation, each offering unique insights into a model’s performance.

Accuracy represents the proportion of correct predictions made by the model out of all predictions. It provides a general sense of the model’s overall performance. However, in datasets where classes are imbalanced, accuracy can be misleading. For instance, if 95% of emails are non-spam and 5% are spam, a model that always predicts “non-spam” would achieve 95% accuracy but fail to identify any spam emails.

Precision focuses on the quality of positive predictions. It is defined as the number of true positive predictions divided by the total number of positive predictions (true positives plus false positives). High precision indicates that when the model predicts a positive class, it is usually correct. This metric is particularly important in scenarios where false positives are costly. For example, in email spam detection, high precision ensures that legitimate emails are not mislabeled as spam.

Recall, also known as sensitivity, measures the model’s ability to identify all actual positive cases. It is calculated by dividing the number of true positive predictions by the total number of actual positives (true positives plus false negatives). High recall indicates that the model successfully captures most of the positive instances. This is crucial in situations where missing a positive case has serious consequences, such as in medical diagnoses where failing to detect a disease could be life-threatening.

Balancing precision and recall is often necessary, as improving one can lead to a decline in the other. Depending on the specific application and the associated costs of false positives and false negatives, one metric may be prioritized over the other to optimize the model’s performance effectively.


I can already see some mistakes! I use the British spelling – optimize, and prioritize are not how I would spell these words. I think perhaps it hasn’t read the article properly, I don’t feel like this is in my style at all.

Post 4: A final go, let’s copy in the inspiration directly.

I got a little worried that maybe ChatGPT has difficulty reading the text from the link. The website page has more than just the blog post text. So this final time I’m pasting some of the body of my previous article in, using the same prompt as above, but with the first three paragraphs of that link. Let’s see if it has better luck with the content embedding directly into the prompt.



That last one was way better right? The introduction was much more my style, with questions, directly talking to the reader, and outlining what is in this post. There are some examples, but it is light on metaphors still. Yes, I can still see the American spellings!

What am I looking for?

When reading this text, I’m going to bear the following things in mind. These are some tricks I’ve picked up along the way which help determine if the content is AI-written or not.

That’s just plain wrong, AI Can Hallucinate Facts

Are there errors – factual or grammatical? AI does tend to make up facts, which sound plausible (known as hallucination). Humans are much more likely to make grammatical errors or spelling errors.

What a pleasant AI, the AI will aim to please with politeness

Is the tone of the text too polite or formal? Many language models have been specifically created to match with human preferences (known as human alignment), so often respond in a very polite tone. Human creators are more likely to have an inconsistent tone, especially in longer, unedited texts.

You’re giving my mixed massages, what do you actually believe?

Does the text avoid overly strong opinions? Linking back to human alignment, AI tends to avoid strong opinions. This reflects what most of us prefer in a conversation partner. We, the humans, are often not afraid to voice our opinions (wanted or not)!

I’ve heard this all before, the AI has no new opinions!

Are there any unique viewpoints, or are they just summarised facts? A fundamental weakness of AI (at least the AI of today) is that it can’t innovate new viewpoints, it just has an incredible ability to file away all the opinions it has read during its creation and summarise them when prompted.

The AI might be very polite, say wrong facts, give mixed opinions or not have an original opinion. These are telltale signs you might look out for.

Findings

So let’s see what the two evaluations resulted in, the higher the score – the more the content sounded like my content.

Test Blog Post[Scores] Comments (Me/ AI)
Take 1 (My Writing)[10 / 10] My writing, so you know I can’t comment.

[8 / 10] GPT had a few things to say though – The post effectively explains key AI evaluation metrics—accuracy, precision, and recall—in a clear and accessible manner. The tone is informative and engaging, with relatable analogies. Minor improvements could be made in capitalization consistency (e.g., “A models” → “a model’s”) and terminology uniformity.
Take 2 (Lightly Prompted)[4 / 10] I thought this was too formal, not engaging with the reader at all, though the technical detail was clear and correct.

[9 / 10] AI thought this was well-structured and precise, with clear explanations and the tone aligns well with The Algorithm Edit. It is slightly more formal than the reference post, lacking some of the conversational elements.
Take 3 (Linked Blog in Prompt)[6 / 10] Again I thought the was too formal, but had clear explanations with examples. There were some American spellings. The structure didn’t quite match (subheadings aren’t there) and doesn’t engage directly with the reader.

[9 / 10] A technically accurate and well-organized post that matches the target audience. The explanations and examples are strong, though the style leans slightly more formal than the reference article.
Take 4 (Pasted Blog in Prompt)[9 / 10] I thought this was very close to my style, with the introduction, questions, talking to the reader directly, examples and the subheadings as sentences. Again minus a point for American spellings.

[9.5 / 10] Engaging, structured, and highly readable. The conversational yet informative tone, use of rhetorical questions, and bold headings improve clarity. Slightly more casual than the reference post, but this enhances readability rather than detracting from it.
The AI and I only agreed on one post (out of FOUR)

So ChatGPT didn’t like my post, which is understandable – I am a human after all! But what is interesting is we both agreed that the final (Take 4) post strongly aligned with my style, whereas the others felt more formal and textbook-like. Re-read Posts 2 and 3 and see that the text never interacts with you, the reader in any way.

Example Content Helps with Style Matching

You can see the improvements each time we used this LLM to generate content for me. Giving an example of the style to the model has helped it match the fit of the content that I usually write. It uses things like rhetorical questions:

AI models are often judged on how well they perform, but what does “perform well” actually mean? (Post 4)

It included subheadings of short sentences, which is something I frequently use:

Accuracy Measures Overall Correctness (Post 4)

However, it still used American spelling! It didn’t quite pick up on that.

… medical test might prioritize recall…

The AI insn’t Inventive when it comes to examples

It’s funny that three different AI-generated posts use the same two illustrative examples to help explain the differences (medical diagnosis and email spam filter). I imagine these are probably used frequently to describe the differences over the internet.

Conclusions

After this experiment, I’ll be using AI to enhance my work, and wherever I do, I’ll be incorporating some of my other content to help it sound more natural. I’ve seen how well it can match style and tone with specific instructions. The area that I will still need to contribute, is providing examples, opinions and editing.

Some people might argue that AI will replace content creators, and yes, we’ve seen that the quality of AI-generated work can be pretty impressive. However, I believe that AI isn’t here to replace us. You’ll still need a human creator to plan the work, weave in personal experiences, and use helpful metaphors to add that special “flair” and authenticity.

So, what do you think? Can you spot any signs that AI might have written this text? Will you be using AI in your content creation, and if so, how? Drop a comment below!

Leave a comment