Architectural Prompting: Nano Banana 2, the Grounding revolution with Google Search and the leap towards 4K

|

Emma Potter

Nano Banana 2: what changes

Here’s what changes from today with the new update:

  • Resolution up to 4K: finally the output we’ve been waiting for. We can generate images with sharpness ready for printing and professional use, overcoming the limitations of old formats.
  • Grounding with Google Search: The model can use Google Search to base images on real-world data. This guarantees greater historical and geographical accuracy in the representation of monuments, specific objects or technical details.
  • Subject Coherence (Character Consistency): a fundamental function. It is now possible to maintain the identity of objects (up to 14) or characters across multiple generations. Indispensable for creating storyboards or series of coherent renders.
  • Conversational editing: just regenerate everything from scratch. Now we can modify an existing image by talking to it: changing the color of a finish or removing an element from the background is a matter of a simple message.
  • Precision in the text: great progress in rendering legible and correct writing, even in different languages, directly within the graphics.
  • New formats (Aspect Ratio): maximum creative flexibility with extreme aspect ratios, such as 1:8 or 4:1, for panoramic or vertical details.

Grounding with Google Search

The new feature that struck me most, however, is Grounding with Google Search. Often the limit of AI in image generation is invention: the model “hallucinates” because it does not know the physical reality of our project site. With Grounding, the system stops guessing and searches Google Images and the web in real time to see exactly what it needs to generate.

For our photoinsertion work, this is a fundamental support for several reasons:

  • Accuracy of the urban context: if you ask to view a building next to the Velasca Tower in Milan, the system retrieves the real data of the monument. Proportions, materials and shades of concrete will be consistent with reality, avoiding distortions.
  • Fidelity to technical materials: the difference between a reflective glass and a selective one is enormous. We can request specific coverings (e.g. “zinc-titanium of a certain brand”) and the model will look for what that material actually looks like under different lights.
  • Consistency with real light: the system can analyze the solar orientation and weather of a specific street via Street View, helping us generate shadows and reflections perfectly integrated into the surrounding environment.
  • Local vegetation and street furniture: goodbye plants out of context. Thanks to the research, the model identifies the tree species typical of the project’s climate zone, making the rendering credible and less “artificial”.
  • Historical referencing: in the concept phase, we can recall specific styles – such as Como rationalism – and the model will actively look for key elements of that current one to apply them correctly.

Nano Banana 2, let’s put it to the test with a little test

To see if the theory corresponded to reality, I wanted to do a little field test. I asked Nano Banana 2 to insert a modern pop-up café in glass and steel right in the heart of Turin.

The result? It’s really interesting. The AI ​​did not limit itself to placing a generic box in any square, but perfectly captured the architectural characteristics of the historic center of Turin:

  • Architectural DNA: the nineteenth-century porticoes and typical cream-colored facades of Turin have been rendered with incredible precision.
  • Details on the ground: even the stone paving – which is usually a weak point of the AI ​​- reflects the real textures and colors of the square.
  • The “touch of class”: the thing that struck me the most was the tram. It’s not a generic wagon; it is an excellent reproduction of the historic green trams of Turin.

It is clear that the system is no longer just “drawing”, but is consulting reality. For those of us who deal with photo insertions, this means that the days of manually correcting lights and local details are numbered.

In summary

Grounding transforms Nano Banana into an assistant who “studies” the project site or an object before starting to draw. It drastically reduces manual post-production times and gives us a base that already knows the physical reality of the site. We are faced with an increasingly mature tool suitable for professional work, capable of amplifying our story of the project rather than replacing it. Until next time viewing!

The weekly column “Architectural Prompting” is edited by experts Luciana Mastrolia, Giovanna Panucci and Andrea Tinazzo
>> If you are interested in these topics, also sign up to the free Linkedin Newsletter AI & Design for Technicians, we’ll talk about it here!