top of page

Design with AI, Not by AI: Lessons from a Full App Concept Experiment

  • Writer: Davide Arcoraci
    Davide Arcoraci
  • 2 hours ago
  • 5 min read


Can an Entire Application Concept Be Generated by AI?


Short answer: Yes.

Longer answer: Yes.. if you treat AI as a collaborator, not a shortcut.


Over the past year, many design teams have experimented with AI tools to generate UI screens, layouts, and even entire flows. In most cases, the first outcome is a visually pleasant but shallow, something nice to look at, yet detached from real user needs.


This article documents a different kind of experiment:
Can an entire application concept from research to an interactive prototype be generated using only AI tools, while still respecting good UX practices?


The results were, honestly, impressive. But not for the reasons you might expect.



The Use Case: Why This Experiment Made Sense


The experiment took place during a short vision-focused design engagement with a large American post-secondary institution, aimed at defining a future-state concept for their internal Grant Management system.


This was not a simple product.



 The existing system is:

  • Highly complex

  • Strongly dependent on manual human operations

  • Rich in domain-specific terminology and workflows


From the beginning, we knew the goal was not to deliver a production-ready solution. The outcome was meant to be a conceptual prototype, something that could communicate a future vision, explore possibilities, and align stakeholders.

That made it a good and meaningful space to experiment.


Internally, we had already been exploring AI generative tools for more than a year.

This project felt like the right opportunity to:

  • Push those tools to their limits

  • Test a fully AI-assisted design workflow

  • Start shaping a repeatable internal framework that other designers could use


It’s important to clarify upfront:


We set one rule: AI tools first.



Every application artifact, from research synthesis to mockups, had to be generated using an AI tool.

This constraint forced us to slow down, be more intentional, and rethink how AI fits into the design process, not just where it appears.


The established framework is not set in stone.


It is the result of hands-on experimentation, continuous iteration, and insights drawn from both personal experience and industry references, and it is meant to evolve as tools and practices continue to evolve.



Designing Screens with AI: A Pilot Framework



The framework that emerged is divided into two main parts:


1. The Core: Knowledge, Research, and Architecture


This is where the foundation is built:

  • Research synthesis

  • Context building

  • Information architecture

  • Early pattern definition


This phase is driven mostly in ChatGPT, which acts as a central brain for organizing, summarizing, and structuring knowledge.


2. The Interface: UI Creation and Refinement


This is where ideas become tangible:

  • Screens

  • Flows

  • Visual hierarchy

  • Interaction details


This phase happens mainly in Figma (still one of the most powerful design environments) and is the shortest bridge to development.

The separation is intentional.


A strong interface cannot exist without a strong core.



Step 1 – Set the Project in ChatGPT



We started by creating a dedicated ChatGPT project.


This might sound trivial, but it was one of the most important decisions in the process.


What happened


  • All research outputs were collected in one place (documents, notes, data, screenshots)

  • Findings were summarized and structured

  • A shared project context was built to support every next step


Why start here?


ChatGPT turned out to be extremely effective at:

  • Organizing complexity

  • Holding long-term context

  • Creating continuity across decisions


Instead of scattered notes and disconnected artifacts, we had a single evolving knowledge base.


Output


  • Synthesized insights and requirements

  • A solid foundation for prompt design and screen generation



Step 2 – Generating the Design Concept



With a strong research backbone in place, we moved into concept generation.


What happened


  • Defined the information architecture of the application

  • Mapped primary user flows and key scenarios

  • Generated early, low-fidelity screen structures

  • Explored multiple layouts and interaction patterns


This phase was still largely handled in ChatGPT, focusing on structure rather than visuals.


Why this matters


By keeping things intentionally low-fidelity, we avoided the classic AI trap:
jumping straight to a polished UI without understanding the problem.


Output


  • Information architecture and core user flows

  • Low-fidelity screen structures



Step 3 – Refining the Design in Figma Make



Once the concept was clear, we moved into visual refinement.


What happened


  • Imported code from ChatGPT into Figma Make

  • Recreated the same screens and flows

  • Connected the project to an existing design system

  • Refined layout, spacing, typography, and visual hierarchy


At this point, AI was still heavily involved, but the designer’s judgment became even more critical.


Output


  • A more polished, design-ready prototype

  • A strong foundation for interaction design and validation



Step 4 – Exporting Mockups



The final step is almost optional; it focuses mostly on communicating the design outputs using the traditional boards in Figma Design.


What happened


  • Defined screens representing key user states and flows

  • Exported screens from Figma Make

  • Assembled screen flows in Figma Design

  • Used the mockups for presentation and discussion


Output


  • Fully editable Figma mockups with Auto Layout

  • High-fidelity, scenario-specific screens



The Final Output: More Than a “Nice Prototype”



The result was not the typical AI-generated prototype we see in most experiments made by using just a few generic prompts.


It was:


  • Scenario-driven, based on real user processes and constraints

  • Built around end-to-end workflows, not disconnected screens

  • Created to actually work, not just look good


Users could follow realistic paths that reflected actual grant-management scenarios, not just click through disconnected screens.


This required multiple iterations, guiding the AI step by step through each page, state, and transition.



A critical reflection: AI can easily generate interfaces, but UX Design still starts with strong research


One of the strongest learnings from this experiment is also one of the simplest:


AI does not replace UX research. It amplifies it or completely breaks without it.


If research inputs are vague, incomplete, or biased:


  • AI will make false assumptions not based on the real context

  • The outputs will often be visually convincing but meaningless for the real users


Important discovery steps such as user interviews, domain analysis, and current-state mapping cannot be skipped. AI can speed them up, summarize them, and help extract patterns, but it cannot invent understanding.



Final Considerations from the Experiment


A few practical lessons stood out:


  • Learn and use technical terminology.
Knowing the correct terminology, especially when generating UI content, is fundamental during the AI prompting process. Precise language enables faster and more accurate UI changes. Incorrect or ambiguous naming of UI elements can easily lead to hours of unnecessary iteration.


  • Always provide context.
Background information and clear goals help AI align its output with the overall vision. For example, I initially underestimated the value of adding visual UI references to my prompts, but in practice, they proved extremely effective for guiding the AI toward the desired result in specific cases.


  • Set clear boundaries.
Explicitly state what should be included and what should be avoided in each prompt. Many AI tools tend to introduce additional elements that were never requested. While this can occasionally lead to interesting insights, more often it results in unnecessary or distracting output.


  • Understand AI’s capabilities and limits.
Knowing what AI can and cannot do helps set realistic expectations. During this experiment, I often realized that I was underestimating AI’s potential because I was still operating within a traditional design mental model. In many cases, AI proved to be significantly more efficient than humans at purely mechanical tasks. Continued hands-on experience with AI will help designers better understand its logic and gradually reduce this gap.


  • Experiment constantly.
Trying different prompts and approaches often leads to unexpected and valuable solutions. We are still in a phase where no AI-driven workflows can be considered industry standards. Even though many AI tools have introduced credit-based limitations that make exploration more expensive, the most effective way to achieve strong results remains testing, iterating, and exploring multiple approaches.


But more important than all the above:




Collaborate with AI.




After this experience, one idea became very clear to me:


Treat AI like a junior coworker.


Someone who can:


  • Work incredibly fast

  • Support you across many tasks

  • Offer ideas and alternatives


But who:


  • Needs supervision

  • Sometimes lacks real context understanding

  • Cannot fully replace your judgment


Use it to challenge assumptions, accelerate exploration, and handle complexity but keep ownership of the decisions.


That’s where meaningful design still lives.

 
 
 
direction.png

Ready to get started?

Let us help transform your business on the industry leading modern application platform

img

Accelerated Focus is an OutSystems Premier Partner based in North America and Europe.

© Accelerated Focus 2026

CONTACT US

We don't just build apps, we build businesses

img
  • LinkedIn
  • Twitter
bottom of page