The Setup

This week I'm running Prism, my iridescent AI mascot, through five days of structured NB2 testing. Thumbnails. Book covers. Merch. Ad campaigns. Custom model training. One character, five use cases, 132 images, all scored and documented.

Today's question: can Nano Banana 2 (@NanoBanana), a partner model available for free inside Adobe Firefly (@AdobeFirefly), generate usable YouTube thumbnails?

Twenty-four images later, I got my answer. But I also stumbled onto something bigger. NB2 might have the most capable text rendering of any image generation model I've tested.

Here's everything I found, with scores and prompts you can steal.

I tested six thumbnail styles using Prism as the subject, with the character sheet uploaded as a reference image for every generation. Each style got four images, scored across five dimensions: Visual Quality (30%), Prompt Alignment (25%), Consistency (15%), Uniqueness (15%), and X Engagement Potential (15%). This rubric stays constant across all five days.

But here's where it gets interesting. Instead of using the same text across all six, I escalated the text complexity from one word to four separate text elements.

It passed every level.

Variation A's single word versus Variation F's four text blocks. Same model. Same session. NB2 handled both without breaking a sweat.

The Text Rendering Results

Let me lay this out because it's the headline finding.

Level 1. Single word ("PRISM"): Perfect. All four images. Not surprising. Most models can handle one word.

Level 2. Title plus subtitle ("PRISM" + "Creative AI Tips"): Perfect. Clear size hierarchy. NB2 understood the subtitle should be smaller. Already impressive.

Level 2 with numbers. Dual position ("PRISM" top left + "EP. 47" top right): Perfect. Period, space, digits all clean. NB2 placed text in two distinct locations reliably.

Level 3. Three text blocks ("PRISM" + "IS AI ART DEAD?" + "WATCH NOW"): This is where I expected failure. Three separate pieces of text, three positions, a question mark, different sizes. NB2 nailed it. And then it did something I didn't ask for. It rendered "WATCH NOW" as an actual button with a rounded rectangle border.

D-4's unprompted button CTA. Nobody asked for that rounded rectangle. NB2 understood "WATCH NOW" is a call-to-action and treated it like one.

Level 3 with special characters. ("PRISM" + "How I Made $10K" + "with AI Art"): Dollar sign. Numbers. Letter-as-abbreviation. All clean across four images. NB2 even made "$10K" slightly bolder than the surrounding text without being prompted.

E-1's special character rendering. Dollar signs, digits, and letter-as-abbreviation all sharp. NB2 emphasized "$10K" with bolder weight on its own.

Level 4. Four separate text elements ("PRISM" + "The Ultimate Guide" + "to AI Image Generation" + "2026 Edition"): The stress test. Four blocks, four positions, a four-digit year. Every image came back clean. The "PRISM" text got a brushed metallic gold 3D treatment that looks like a movie title card.

F-1 scored 9.25, the session high. Four text blocks, four positions, metallic gold 3D treatment on the brand name. Four was the most I tested, and NB2 didn't flinch.

That's the result: NB2 handled everything I threw at it from one word to four text blocks, including special characters, numbers, question marks, and multi-word phrases. Four was the highest I tested, and it passed clean.

What NB2 Does With Text That Other Models Don't

The accuracy is one thing. What surprised me more was the intelligence behind the text rendering.

NB2 doesn't just put words on the image. It makes design decisions.

When I tested split lighting (warm amber on one side, cool blue on the other), NB2 rendered "PRISM" in dark text on the warm side and "EP. 47" in white on the cool side. It chose colors for maximum readability against the local background. I didn't tell it to do this.

When I asked for a brand name, a hook question, and a call-to-action, NB2 gave each one a different visual treatment. Neon outline glow for the brand. Bold 3D gradient for the hook. A literal button shape for the CTA. It understood these are different types of text with different functions.

When the brand text "PRISM" appeared alongside a character with iridescent rainbow skin, NB2 color-matched the text to the character, giving it a rainbow gradient that tied the brand to the mascot visually. Zero prompting for this.

D-2's rainbow gradient text. NB2 matched the PRISM text to the character's iridescent skin without any prompting. That's not rendering. That's design.

This isn't text rendering. This is text design.

The Thumbnail Ranking

Now the visual findings. Six styles, ranked by average composite score:

#1: Dark Luxe (9.03 avg). Deep black background, spotlight, gold accent lighting, smoke wisps. Best for educational and premium content. Dark backgrounds give text the most breathing room and scored highest across every metric. If you want maximum text, go dark.

#2: Split Lighting (8.83 avg). Warm amber on one side, cool blue on the other. Best for entertainment and episodic content. The dramatic contrast creates instant visual tension. If your character has interesting skin textures or colors, split lighting will show them off.

#3: Chaotic Neon (8.71 avg). Neon particles, light streaks, electric purple and hot pink. Best for gaming and reaction content. Maximum energy, maximum scroll-stop power. Text still renders cleanly even in busy scenes. NB2 handles visual noise better than expected.

#4: Contextual Office (8.35 avg). Character at desk with monitor, natural window light. Best for tutorial and lifestyle content. The most "authentic" looking thumbnails. NB2 put Photoshop on the screen without being asked. Lower engagement scores but higher relatability.

#5: Warm Dramatic (8.04 avg). Red and orange lighting, bokeh, dark vignette. Best for general use. Solid cinematic feel. The A-3 image with the forward lean proved that body language matters more than background.

#6: Cool Professional (8.03 avg). Blue and teal gradient, clean studio. Best for business and education. Highest consistency (9/10) but lowest energy. If you need reliable, repeatable, professional results every time, this is your formula.

The full lineup ranked by score. Dark Luxe at 9.03 on top, Cool Professional at 8.03 on the bottom. Same character, same reference image, six completely different moods.

The Formula That Emerged

After scoring all 24 images, the winning pattern was clear:

Upper body framing + dramatic lighting + text positioned in negative space + character leaning forward or displaying strong expression = scores above 8.5 consistently.

Full body shots underperformed every time. Static poses underperformed dynamic ones. Clean backgrounds scored higher on consistency but lower on engagement. Dark backgrounds made text more readable.

The single most impactful mid-session change? Adding "upper body framing" to every prompt after the first two variations. That alone raised the floor from 7.70 to 8.60.

Known Limitations

Honesty matters. Here's what didn't work perfectly. I'll document these once here and track them through the rest of the week.

Eye and pupil rendering. Prism's glowing amber eyes sometimes got standard pupils instead of the solid glow. Happened in roughly half the images. Easy fix in Adobe Firefly Boards or Photoshop, but it's consistent enough to flag.

Shirt logo dropout. The "Adobe Firefly Ambassador" logo on Prism's shirt disappeared in about 12% of images, particularly in complex scenes. The reference image carries it most of the time, but not always.

Both of these are things I'll be targeting when I train a custom Firefly model on Friday.

Text Z-ordering. In one image out of 24, large background text was partially hidden behind the character instead of in front. Not a rendering failure, a layout and layering issue. Happened when brand text was very large and character was centered.

The Prompts (Copy-Paste Ready)

Here's the winning formula as a template. Replace the bracketed sections with your content:

[image reference] YouTube thumbnail, upper body framing, [LIGHTING STYLE], [CHARACTER POSE/EXPRESSION], [TEXT ELEMENT 1 with position], [TEXT ELEMENT 2 with position], [ATMOSPHERE DESCRIPTION]

For maximum text + premium feel:

[image reference] YouTube thumbnail, upper body framing, deep black background with subtle gold accent lighting, elegant smoke wisps, single dramatic spotlight from above on character, large metallic gold text '[BRAND]' top left, white text '[MAIN TITLE]' center right, smaller text '[SUBTITLE]' below it, small text '[DETAIL]' in bottom right corner, luxurious moody atmosphere, character emerging from darkness

For maximum engagement:

[image reference] YouTube thumbnail, upper body framing, split lighting one side warm amber glow other side cool blue shadow, dramatic diagonal light division across character face, character leaning forward with intense expression, bold text '[BRAND]' at top left and text '[EPISODE/DETAIL]' at top right, strong contrast, cinematic color grading, volumetric light rays cutting through scene

What This Means for Creators

If you're making YouTube thumbnails, social media headers, or any content that needs text baked into the image, NB2 through Adobe Firefly can do it. Not approximately. Not "close enough to fix in post." Actually, usably, production-ready text rendering with intelligent design decisions.

Is it perfect? The eyes need work and the shirt logo drops occasionally. But those are character-specific reference image issues, not text issues. The text is rock solid.

Next up tomorrow: I'm putting Prism on book covers across five genres. If NB2 can handle "THE PRISM EFFECT" in distressed thriller font and "Finding Prism" in elegant romance script, we'll know this model isn't just a thumbnail tool. It's a design tool.

Testing methodology: Nano Banana 2 (@NanoBanana), a partner model inside Adobe Firefly (@AdobeFirefly). All images scored using a weighted 5-dimension rubric. Minimum 4 generations per variation before drawing conclusions.

Keep Reading