Search This Blog

Showing posts with label generative AI. Show all posts
Showing posts with label generative AI. Show all posts

Thursday, November 27, 2025

How Thankful Should We Be for Generative AI?

 I thought I'd like to generate a picture of Jane Austen in context of an American Thanksgiving celebration for a holiday greeting post on the social media accounts I have associated with my Totally Jane Austen brand. I opted to use Google Gemini. 

Here's how it went: 

First prompt: Can you generate an image of Jane Austen celebrating an American Thanksgiving

Here's what it delivered with this message: 

"Here is an image of Jane Austen celebrating American Thanksgiving."



Second prompt included the portrait of Jane Austen that I referenced: Let's have fewer people, no signs, and a woman who is dressed like Jane Austen is in her portrait 



Second iteration came back with this: "Here is a Thanksgiving celebration with fewer people, no signs, and a woman dressed like Jane Austen in her portrait."




I attached the Jane Austen portrait again and wrote: "Nope you didn't do it. That's almost the same picture. Try doing it with Jane Austen herself looking like this."

Gemini then delivered this with the note, "My apologies, I misunderstood the previous request. Here is an image of Jane Austen herself, looking like in her portrait, celebrating Thanksgiving with fewer people and no signs." 


My response was to again attach the portrait with this prompt:  "That's the same picture again. Delete that and just put this picture of Jane Austen in a context of a table with turkey, etc."


Gemini then offered this monstrosity with this phrase "You are absolutely right, I apologize for the repeated image. I will generate a new image for you, placing Jane Austen from the provided portrait into a Thanksgiving setting. "


I then wrote:  "She's out of proportion to the people. Just take them out and get the proportions right."


Gemini responded with "My apologies for the scale issue. Let me correct that for you. Here is Jane Austen, in proper proportion, at a Thanksgiving table by herself." and delivered this: 


I wasn't thrilled with the results for several reasons but was tired of the back-and-forth, so I said, "It'll have to do."

Gemini's response: "I understand. I apologize that the image generations were not meeting your expectations and appreciate your patience with the process.

Is there anything else I can help you generate or create today?"


Related:

Aim higher than SEO for your marketing content


Conversation with Gemini

can you generate an image of Jane Austen celebrating an American Thanksgiving

Let's have fewer people, no signs, and a woman who is dressed like Jane Austen is in her portrait

Nope you didn't do it. That's almost the same picture. Try doing it with Jane Austen herself looking like this

That's the same picture again. Delete that and just put this picture of Jane Austen in a context of a table with turkey, etc.

She's out of proportion to the people. Just take them out and get the proportions right.

It'll have to do.

Sunday, August 25, 2024

Ouroboros, an apt symbol for AI model collapse

Engraving of a wyvern-type ouroboros by Lucas Jennis, in the 1625 alchemical tract De Lapide Philosophico

by Ariella Brown


AI hits the ouroboros (sometimes written uroboros) stage. You've likely seen it in the form of a snake in a circle, eating its own tail. The ancient symbol also sometimes showed dragons or a wyvern, so I chose this engraving by Lucas Jennis intended to represent mercury in the 1625 alchemical tract "De Lapide Philosophico," for my illustration instead of just going with something as prosaic as "model collapse"


To get a bit meta and bring generative AI into the picture (pun intended, I'm afraid) here's an ouroboros image
made with generative AI. ked Google

Ouroboros image generated by Google Gemini



Model collapse is what the researchers who published their take on this in Nature called the phenomenon of large language models (LLMs) doing the equivalent of eating their own tails when ingesting LLM output for new generation. They insist that the models should be limited to"data collected about genuine human interactions."

From the abstract:
"Here we consider what may happen to GPT-{n} once LLMs contribute much of the text found online. We find that indiscriminate use of model-generated content in training causes irreversible defects in the resulting models, in which tails of the original content distribution disappear. We refer to this effect as ‘model collapse’ and show that it can occur in LLMs as well as in variational autoencoders (VAEs) and Gaussian mixture models (GMMs). We build theoretical intuition behind the phenomenon and portray its ubiquity among all learned generative models. We demonstrate that it must be taken seriously if we are to sustain the benefits of training from large-scale data scraped from the web. Indeed, the value of data collected about genuine human interactions with systems will be increasingly valuable in the presence of LLM-generated content in data crawled from the Internet."

Shumailov, I., Shumaylov, Z., Zhao, Y. et al. AI models collapse when trained on recursively generated data. Nature 631, 755–759 (2024).

Let me know in the comments which illustration you like more. 

Friday, June 28, 2024

An Apology to Generative AI

ChatGPT spelled out in Scrabble tiles

By Ariella Brown


I'm not a generative AI fangirl. If anything, I'd consider myself more of a skeptic because people tend to not just use it as a tool to improve their writing but as a tool to replace the work of research, composition, and revision that is essential to good writing.

It is generally embraced by people who consider online research to be too much work and who believe that anything that comes out of a machine that will charge them no more than $20 a month for writing to be too good a deal to pass up. 

For those of us who actually read, the output of ChatGPT and similar LLMs is not exactly something to write home about. Unless you know how to prompt it and train it to write in a truly readable style, it will default to the worst of wordy, opaque corporate style text. 

But this isn't the fault of the technology. It's the fault of the mediocre content that dominates the internet that trained it. Below is one example that I pulled off  the "About" section of a real LinkedIn profile (first name Kerri maintained in the screenshot that proves this is real and not something I made up):  LinkedIn screenshot

As a strategic thinker, problem-solver, and mediator, I thrive in managing multiple, sometimes differing inputs to achieve optimal messaging and positioning. My proactive nature drives me to partner with leaders across marketing teams and internal business units, aligning efforts, connecting dots, and adding context to enable flawless execution of communication strategies and tactics.


In fast-paced, fluid environments, I excel in effectively prioritizing tasks and ensuring they are completed efficiently. I have a proven track record of setting and meeting strict deadlines and budgets, leveraging my ability to navigate dynamic landscapes seamlessly.

Driven by natural curiosity, I am constantly seeking to understand and implement the latest trends, technologies, and tactics essential for driving B2B sales opportunities. My keen interest in exploring new channels for messaging and content distribution fuels my passion for innovation and continuous improvement to not just meet but exceed expectations.

Let’s connect to explore how we can drive success together.

You know what sounds exactly like this? Cover letters you ask ChatGPT to compose for you. 

I've tried those out a few times and never been happy with the results because they always sounds like the text above. Trying to tell it to sound less stiff doesn't make it sound any less canned, and forget about getting it to copy my own writing style.

It's possible that Kerri used ChatGPT to create her "About" section. Given that she's been in the marketing biz for some time, though, I'd think she had to have had something filled out for years before ChatGPT was available, and it likely sounded very much like this even if she did let some LLM or something like Grammarly tweak it for her.  

People like Kerri, who ignore all writing advice from the masters like Orwell, White (watch for a upcoming  blog about him), and others made this the public face of corporate communication who are to blame for the bombastic and soulless style that LLMs replicate at scale. 


That's the reason for this apology too ChatGPT for mocking its output. You're not the one at fault. You had no way of knowing better. Humans do, and they should have provided you with better models for writing. 

Note on the title: I thought of giving this post the title "Apology" intended in the classical sense of a defense or justification for something others take as wrong with the hint of an apology to AI. Knowing that that wouldn't be clear to some readers, I opted to make this just a straight apology instead. 

Related:

A new generative AI comparison




Sunday, June 23, 2024

A new generative AI comparison






Photo by Andrew Neel: https://www.pexels.com/photo/monitor-screen-showing-chatgpt-landing-page-15863066/



In the past, I've compared Chat GPT to Google Bard (now Gemini). For today's exercise, I put the same query to Claude, ChatGPT, Gemini, and Copilot. I put in the name of a real person and identified the location for that individual in Queens, NY.  None could give me the information I sough. 
What's interesting, though, is noting the differences in how they contextualized their responses. Out of the four, Claude shows the greatest self-awareness, and I'd say that Copilot shows the least.  

Claude's response

As this appears to be asking about a relatively obscure individual, I want to remind you that while I try to be accurate, I may hallucinate or generate incorrect information when asked about very specific people who are not widely known public figures. If you're looking for factual information about this person, I'd recommend checking other reliable sources.

I award Claude points for warning me about the possibility of hallucinations occurring when generative AI attempts to answer questions in the absence of data. See 

ChatGPT's response

I don't have specific information about an individual named ____from Queens, NY. It's possible that she is a private individual or someone not widely known beyond her personal circles. If she is known for something specific like public contributions, achievements, or notable activities, that information might be available through local news sources or community records.

ChatGPT's response failed to warn about hallucination, though it's interesting that it inferred the person in question was a she (based on the first name, I suppose), a gender identification that Claude managed to avoid.

Google Gemini's response

Gemini just bowed out, practically chiding me for giving it this query that it could not answer:
 
I do not have enough information about that person to help with your request. I am a large language model, and I am able to communicate and generate human-like text in response to a wide range of prompts and questions, but my knowledge about this person is limited. Is there anything else I can do to help you with this request?


Copilot's response

Copilot went in a completely different direction. I wouldn't call it a hallucination, but it is still inaccurate. It wanted to find a match, so in one case it named a business with the last name appearing twice as a partnership, though it admitted that was Brooklyn rather than Queens. Its second suggestion was a male doctor in Queens with a similar last name, though it is spelled differently. Honestly, I find it hard to believe that these are the only two results for the last name in all of NYC. I mean once you're expanding the parameters, there really should be a lot more than two possible results. 


RelatedAn A/B test of generative AI
Bard brings in the clowns


You can also follow Ariella Brown.  

Wednesday, February 28, 2024

Building foundations for castles in the air

By Ariella Brown

“If you have built castles in the air, your work need not be lost; that is where they should be. Now put the foundations under them.”

― Henry David Thoreau 


Two results of the Gemini-generated pics for my prompt based on the quote. Let me know which one you prefer in the comments


This is the essence of creative content marketing. You can start with the creative idea to construct something truly different and unique. But if it is not grounded in reality, which is the actual experiences of the target customer, then it will never be a sustainable structure.

HT to Alfredo A. Lopez for making the comment on Tom Goodwin's post that made me think of this.

I'm wondering if Tom's dichotomy was sparked by reading the "bothism" advocated by Mark Ritson. https://www.marketingweek.com/ritson-bothism-cure-marketers-fascination-conflict/in which he says:

We’ve had the decade-long custard-pie fight over ‘digital’ versus ‘traditional’ forms of communication. I remain completely bemused as to what these terms actually mean in 2020, given almost every form of communication from radio to outdoor is now demonstratively digital in delivery. And yet we have spent all this time pushing the benefits of one side and then defending the honour of the other.

Twenty minutes with a decent data set and an open mind would demonstrate to any marketer that when you adopt a Bothist view of communications and add a dash of traditional media to the digital cake mix, the whole confection improves dramatically as a result.


Related

Don't just aim for different
Visualizing the customer journey
What makes content marketing effective
Trying to stand out