Armando Gil – Portfolio

Armando Gil – Portfolio

TeamShifts: Designing an AI experience from magical to truly usable

AI Product Design + UX Research

TimeShifts schedule managing

SOME CONTEXT

The initial idea felt almost magical: type a simple prompt and instantly generate perfect employee schedules. In theory, AI could remove hours of manual planning and transform a highly operational process into a natural conversation. In reality, the challenge was far more complex. This project became an exploration of how to design AI-powered products that balance automation, structure, and trust.

CHALLENGE

Shift planning is not just about assigning employees to dates. Every generated schedule needs to balance availability, legal constraints, staffing coverage, fairness, and individual preferences, while still remaining operationally viable. What initially looked like a technical problem quickly revealed itself as a design problem. The difficulty was not generating schedules, but generating schedules users could trust.

TeamShifts prompting

EARLY TESTING

The first usability tests exposed a critical issue. While the AI could generate schedules, results were often inconsistent or incomplete. Users repeatedly rewrote prompts trying to “guess” the correct phrasing. Many were unsure why the system failed or how to improve the output. The issue was not the AI itself, but the amount of ambiguity users had to resolve on their own.

Check AI interpretation

FIRST ITERATION

The advanced flow transformed workforce planning into a sequence of focused steps. Users could define the scheduling period, import employees, configure staffing rules, and specify preferences before generating the final schedule. This changed the role of AI completely. Instead of interpreting vague instructions, the system received structured inputs that dramatically improved output quality and predictability.

Schedule output

At this point he final product became a hybrid system where AI handled complexity while the interface provided clarity and control. The prompt remained part of the experience, but as an entry point rather than the primary workflow.

RESEARCH

We tested both approaches with 24 participants including HR managers, operations teams, and scheduling coordinators. Initial behavior strongly favored the AI-first experience:

%

started with the prompt interface

Eighty-five percent of users who initially chose the prompt eventually switched to the guided flow. Most described the prompt as attractive and fast, but unreliable for completing real scheduling tasks. Meanwhile, users who started with the advanced flow rarely abandoned it, largely because it gave them a stronger sense of control and confidence.

%

of users who started with the prompt tried again using the guided flow

Prompt-only generation achieved a 38% success rate on the first attempt, while the structured advanced flow reached 87%. Confidence scores followed the same pattern. Users preferred spending a bit more time configuring the system if the outcome felt predictable and reliable.

DISCOVERY VS. DELIVERY

By combining AI prototyping tools with a Dual-Track approach instead of a traditional Scrum-only process, we rapidly built an MVP to validate our core hypothesis with real users. This allowed us to test, learn, and iterate continuously within days rather than weeks, turning AI-driven speed into faster product learning and a more effective final solution.

Dual track diagram

CONCLUSIONS

This project fundamentally changed the way I think about AI product design. The best AI experiences are not the ones that remove interfaces completely. They are the ones that carefully balance flexibility, guidance, and trust in situations where the outcome truly matters.