BETA
This is a BETA experience. You may opt-out by clicking here

More From Forbes

Edit Story

AI Showdown, Part 2: ChatGPT, Claude, Bing, And Bard Tackle Blogging

Following

In my last article, Part 1 of the AI Showdown series, I shared my surprise when speaking with a prospect who hadn’t heard of Claude or Bard, two of the major generative AI tools making news. That surprise led me to organize this AI showdown.

The first article introduced the contenders: ChatGPT, Claude, Bing Chat, and Bard. For this article, I tested each model by having it create content for a popular marketing use case—blogging.

Since I ran the experiment, Google announced an update to Bard, which is supposed to have made it the “most capable model yet.” I double-checked the results using the new model, but nothing changed. Bard performed just as it had in the first run.

My original hypothesis stands: ChatGPT will outperform the other AI tools. Was I right? Let’s start with the content brief I prepared to put the chatbots on even ground.

Creating the content brief

What’s the first thing you do when you want a freelance content creator to write a blog post for you? You craft a content brief. That’s what I did for this experiment, too.

When creating a content brief for an AI tool, include all the information it needs to write the content you want, for the audience you want, in the voice and style you want.

If you’ve been following my Forbes.com articles on ChatGPT, you may notice that this article's content brief has more elements because I’ve expanded what goes into my standard brief. In the following brief, I included the new elements and omitted others because they were irrelevant to the task.

Elements of the content brief

Let’s dive into the brief, starting with the first element: the AI’s role in the work to come.

  • The model’s role

You’re a world-class expert in content marketing. For the task to come, think deeply about how you’ll go about the work given the context in this brief.

  • The task

Write a listicle blog post on 10 ways content teams can generate ideas for the content calendar. Also, to drive people to the blog post, write four social media posts—one each for LinkedIn, Twitter, Facebook, and Instagram—and one email.

  • The objective

To showcase my creativity, ideation power, and 20+ years of experience in content marketing. The objective of the blog post is to book an exploratory call. The objective of the email and social posts is to drive people to read the blog post.

  • Brand voice, tone, and style

[Note: The following content is included in ChatGPT’s custom instructions so the AI can access it for each conversation. The other AI tools do not have the same functionality, so I included the details in the content brief.]

My brand voice is expert, insightful, passionate, and empathetic. I position myself as an expert and share deep insights about the importance of the holistic content experience—how content looks, sounds, feels, and functions, all from the reader’s perspective. My love for great content shines through as I describe its effect on the reading experience. I empathize with readers because I understand the frustration of encountering poorly conceived, written, designed, and delivered content.

I use a conversational tone that makes my content approachable and easy to digest. I use reflective questions to engage readers in dialogue. I incorporate humor in my writing, using playful analogies to emphasize the transformative power of exceptional content experiences. I also encourage content marketers to build respectful and trusting relationships with readers.

My style is narrative, descriptive, and persuasive. I use storytelling elements to draw readers in, such as personal anecdotes and quotes that set the stage for the discussion. I vividly describe the outcomes of great content experiences, painting pictures of smooth reading and hearts filled with hope and trust. I persuasively argue for the value of great content experiences, showcasing their ability to smooth the reader's learning and buying journeys.

  • Audience

The audience comprises CEOs, founders, marketing leaders, and content teams in B2B SaaS, technology, and consulting startups and established companies.

  • Content type and format

The content will be a blog post, an email message, and social media posts with informative and engaging narratives.

  • Key messages and themes

The key message is the connection between consistently producing excellent, reader-friendly content and increased leads and sales.

Clarifying the brief

In my experiments with AI chatbots for content creation, I discovered that asking the AI tools if they had questions led to better content.

My second step was to feed the brief into each tool with the following instructions:

I’ll give you a content brief and then ask you to create some content. First, consume the brief and let me know if you have any questions.

Let’s see what they came up with.

ChatGPT’s clarifying questions

ChatGPT asked a few questions, but nothing that would tip the way I created content.

Claude’s clarifying questions

Claude asked more profound, business-focused questions than ChatGPT.

I was surprised by Claude’s question about tone of voice because the creative brief included three paragraphs about it.

Even so, the question made me pause: Do I want to write the piece in my brand voice as described in the brief? Or do I want to write it in a less conversational, more informational voice?

I was impressed that Claude understands the nuance of voice enough to cause me to stop and reflect, just as a good editor’s questions would.

Bing Chat’s clarifying questions

Because ChatGPT and Claude asked questions, I was surprised that Bing Chat had no questions for me.

Let’s see what Bard had to say.

Bard’s clarifying questions

In a race to be the biggest and best, AI chatbots continually add new functionality. Sometimes, the developers trumpet new goodness through press releases and platform notes; other times, the new functionality just appears. That’s what happened when I visited Bard.

In the following screenshot, notice how:

  1. The chatbot provided three “drafts” at the top of its response; this functionality is new since I worked with Bard yesterday.
  2. In Draft 1—also its initial response—Bard said it had no questions. But in Drafts 2 and 3, Bard did have questions. Which is it, Bard?
  3. Bard immediately presented an outline for the blog post, even though I didn’t ask. I clipped the screenshot to show the first of the “10 ways”—researching industry trends.
  4. Bard also presented ideas for the social media posts and the content of the email. The screenshot shows two of the social ideas; the email content isn’t pictured.

Except for the questions hiding in Bard’s second and third drafts, ChatGPT and Claude were the only two models to ask follow-up questions. Because both asked different questions and because I wanted to keep each model on the same page for this experiment, I fed my aggregate answers into each AI tool. For your reference, this is the text I submitted:

Here are the answers to your questions, plus a little more information that may be helpful.

  1. If you can find a humorous quote to open the post with—a quote that ties into generating content ideas—please do so. Otherwise, open the post with the challenges content teams and writers face when constantly coming up with new content and statistics or scholarly papers that show why it’s important to publish regularly.
  2. For the structure of the post, each of the 10 ideas should have a brief description, a hypothetical example, or, better, an example from a real-life company. Also, weave in details about the importance of creating a complete, enjoyable, holistic content experience. Each item should focus on a specific tactic or top for ideation. Lead into the list of 10 ways with content about the broader perspective on overall approaches to content ideation.
  3. For the social posts, please suggest images or memes that grab attention in the newsfeed but keep them professional.
  4. Open the email with a short first sentence, five words or fewer.
  5. The tone of voice for the blog post should be conversational and personal, using first and second person as if I’m speaking directly to the reader. But keep it professional for busy corporate readers.
  6. The CTA in the blog post is to schedule an exploratory call to see how I can help companies, content teams, and individuals with their content strategy and content marketing training.

After feeding in my answers to the clarifying questions, I told each AI to write the blog post. Here’s what they delivered.

ChatGPT’s blog post

ChatGPT’s blog post came in at 723 words, an appropriate length for a blog post designed to generate awareness.

Although Yoast, the SEO tool built into WordPress, says to write 300 words, minimum, for a regular blog post, I recommend at least 750 words to ensure enough content depth to interest readers. If I had asked the models to write cornerstone blog posts, I’d expect much longer content—at least 2,500 words.

I’ll present the blog post content first, and then I’ll evaluate it.

Analysis of ChatGPT’s blog post

I’ll evaluate each chatbot’s post on the same elements.

  • Relevance and accuracy

The content ChatGPT provided is relevant and, as far as I can tell, accurate. If I were publishing this blog post, I’d double-check and link to each example. I’d also include screenshots and, if possible, quotes from one or more companies mentioned.

I’d also want to make the content more tactical and how-to, perhaps by writing a fuller, deeper blog post on each way and linking to those posts from this original. As it is now, the post “looks” weak, with just two or three sentences per heading.

  • Creativity and innovation

I don’t see anything innovative in any of the 10 ways. They’re all old news, likely written to death about. What would make the content innovative in my eyes would be to include new, rich examples for each of the 10 ways. For instance, under “Collaborative content,” ChatGPT said, “Look at the collaborations between Adobe and various artists….” What collaborations? What artists? What should I “look at?”

  • Reader experience

I’m not a fan of the “we” voice for a personal brand. I generally reserve “we” writing for content coming from companies. The content also has a few strange elements, such as the overly conversational, out-of-flow “See that?” in the first paragraph.

  • Narrative and descriptive abilities

I also see nothing special about ChatGPT’s narrative and descriptive abilities. It did incorporate two analogies—one about a treadmill and one about a well—in the introduction, but other than that, the narration falls flat.

That said, I’d never expect to publish a blog post directly from the mouth of ChatGPT or any other tool. The tools simply give you a starting point. After that, it’s up to you and your team to be humans in the loop, asking questions, researching, expounding, and adding the human elements that make blog posts exciting and worth reading.

  • Humor and engagement

Nothing of note here.

  • Technical aspects of writing

ChatGPT gets most grammar correct, although it referred to “Moz” as a “who” when it’s actually the name of a company.

It also failed to use the same consistent form for each heading. For instance, to match the noun forms of the other headings, “Leveraging user-generated content” should be “User-generated content.”

  • Overall assessment

ChatGPT’s draft is, as expected, average. It’s usable as a starting point, but I’d need to do much more back and forth or research before publishing it.

Claude’s blog post

Claude’s blog post was lighter than ChatGPT’s, coming in at 588 words.

Here’s the post; my analysis follows.

Before my analysis, let’s talk about the elephant in the blog post—Claude’s use of a biblical framework, complete with a title using the 10 commandments and the words thou, thy, and thee in the content.

I have no idea why Claude would draw on such an analogy in our global world, where it makes more sense to rely on what unites than on what divides.

I didn’t want to proceed with the draft because I couldn’t bear to read the content in that state. Kudos to you if you were able to slog through it.

I asked Claude to try again using another framework or approach. The result follows.

With the biblical framework gone, I could read and analyze the post.

Analysis of Claude’s blog post

  • Relevance and accuracy

Claude’s post is relevant and possibly accurate but too short to be immediately useful. The post presents some neat ideas—like filtering topics through a journalistic lens—but just like ChatGPT’s post, Claude’s needs a lot more substance. The stat in the introduction is a nice touch, but I’d check its accuracy before publishing.

  • Creativity and innovation

Whereas ChatGPT took the route of naming each “way” with a noun phrase, for instance, “Analytical insights,” Claude used more creative language: “Let analytics uncover resonance.” I’m not saying Claude’s language is better, just more creative.

The proposed ways for generating more content ideas are also more creative. For instance, Claude said to draw from everyday experiences for ideas. ChatGPT talked about a “feedback loop,” which could be similar.

  • Reader experience

I like Claude’s writing in this blog post more than ChatGPT’s. It feels richer and flows better. That said, I’m disappointed by the lack of depth. I also cringed at the word “skyrocket” in the last sentence.

  • Narrative and descriptive abilities

Claude’s use of language is tighter and more elevated than ChatGPT’s. The first sentence describing each “way” is more creative than ChatGPT’s offerings. Claude also included a hypothetical example with each, helping readers understand how they might use the advice.

  • Humor and engagement

Nothing to note.

  • Technical aspects

No issues noted.

  • Overall assessment

Claude’s post is better than ChatGPT’s in terms of creativity and reader experience. However, the post would need much work to be useful to readers. I would not want to carry on building out the post with Claude’s draft.

Now, let’s see how Bing Chat performed.

Bing Chat’s blog post

I was blown away when Bing’s content began populating my screen. “It’s a tome!” I thought.

Then Bing stopped writing after about 1,800 words, in the middle of the fourth point. “Can you continue writing the post?” I asked.

It did, this time getting through the start of item 10 before stopping again—but only briefly. Then, without prompting, Bing continued responding—but with the exact words it had already written.

I knew then that Bing was confused; the following screenshot shows where it happened. Notice how Bing had already started writing the tenth point but suddenly stopped and returned to the fourth.

In my experience, it’s difficult to unconfuse a confused model, so I didn’t ask Bing to finish item 10 or to write a conclusion.

In the end, even in its unfinished state, Bing’s blog post weighed in at 3,155 words.

Warning: Because the post is so long, there are 10 screenshots.

Analysis of Bing Chat’s blog post

Even though I knew Bing Chat’s reputation for being chatty, I was surprised by the length and depth of the post. Based on my experience using the chatbot since its release, I expected Bing to be one of the weaker competitors in the AI showdown. That proved not to be the case.

  • Relevance and accuracy

Just because a post is long doesn’t mean it’s relevant and accurate.

Bing Chat included many links in its draft, but most led to 404 “not found” errors, a significant and common problem with all four chatbots in this experiment.

Broken links are a major problem because you have to ask the chatbots for correct links (a fruitless task, I’ve discovered), research the included facts yourself, or come up with new sources and examples for which you can find links. Either way, it’s a time suck.

Besides the accuracy issue, the post seems relevant because it touches on good ideas for generating ideas for the content calendar.

  • Creativity and innovation

Although Bing’s post is entirely formulaic instead of innovative, formulaic isn’t always bad. For each numbered “way,” Bing presents the tactic, how it helps with idea generation, an example, and tools to help with the idea.

I don’t see anything innovative or new in the post, though.

  • Reader experience

Bing presents the best reader experience. The content is long, suggesting depth. It’s well structured, with many bullet points to break up the text. Bing also suggests images and examples for the content, elements that positively affect the reader’s experience.

  • Narrative and descriptive abilities

Bing’s narrative and descriptive abilities exceeded ChatGPT’s and Claude’s for this experiment. Each example Bing presented relates to content marketers, showing how they might use the idea in practice.

  • Humor and engagement

I didn’t see any humor in Bing’s post, but the potential for engagement is high based on other factors I mentioned.

  • Technical aspects

Bing gets good grades for technical writing ability; although I didn’t scour the post word by word, I didn’t notice any glaring errors.

  • Overall assessment

Bing delivered the most thorough, detailed draft, making it the winner thus far. Despite the broken links—a major letdown—I’d still want to work with this draft as a starting point.

Next up—Bard’s draft.

Bard’s blog post

Based on my earlier experiences with Bard, I expected the AI to bomb in this experiment. It didn’t bomb, but it didn’t blow me away.

Take a look.

Analysis of Bard’s blog post

As a fellow content creator, I bet you’d agree with me when I say Bing Chat’s 3,000+ word draft makes any draft with fewer words seem poor by comparison. Bard’s post was much poorer, coming in at 549 words.

  • Relevance and accuracy

The content is relevant because it targets content creators and meets the brief. It’s accurate, too, but only because it lacks the content to be inaccurate.

  • Creativity and innovation

I saw neither creativity nor innovation in the content of this post. It’s just a typical “meh” blog post, like thousands—likely millions—of others online.

  • Reader experience

Bard gave a good experience, speaking directly to readers and asking many questions. The part about the holistic content experience is tacked on at the end, though. And the quote at the end felt out of place, which left me out of sorts.

  • Narrative and descriptive abilities

Bard excelled at narrative and connection in the introduction by opening with a quote. “I’m so tired of writing blog posts!” it wrote, followed by, “I hear content marketers say this all the time.” Of all the introductions, this one grabbed my attention the most.

  • Humor and engagement

The opening quote and the questions throughout cause readers to be engaged. I didn’t see humor, though. The supposedly funny quote at the end of the post seemed misplaced.

  • Technical aspects

Bard can write correctly but needs to learn finesse and flow. Almost every “way” has one sentence out of a few that begins with, “This is a great way to….”

  • Overall assessment

Bard’s post cries out for added substance, meaning more work querying the AI or more research on my part. I wouldn’t want to use this draft as a starting point.

Whew! It’s been a lot of work to get to this point of the experiment. So far, based solely on the blog posts, here are the winners and losers:

  1. Bing Chat, first place
  2. ChatGPT, second place
  3. Claude, third place
  4. Bard, last place

What’s next? Writing the social media posts and email

In the third and final article in this AI showdown series, see how ChatGPT, Claude, Bing Chat, and Bard performed when creating social media content and an email to drive readers to the blog post.

Follow me on Twitter or LinkedInCheck out my website