AI and Inclusivity: Making Virtual Try-Ons Work for Every Skin Tone and Eye Shape
A deep dive into inclusive AI, virtual try-on bias, and the shopper-brand checklist for fair beauty tech.
AI and Inclusivity: Making Virtual Try-Ons Work for Every Skin Tone and Eye Shape
Virtual try-on tools are no longer a novelty. They are quickly becoming a core part of how shoppers discover foundation shades, test lipstick finishes, preview eyeliner styles, and compare beauty looks before buying. But as AI becomes more embedded in the beauty journey, a hard truth has surfaced: many tools still work better for some faces than others. Skin tone representation, eye shape mapping, age diversity, and lighting robustness all affect whether a try-on feels accurate or misleading. For shoppers, that can mean wasted money and frustration; for brands, it can mean lost trust and a biased product experience. This guide breaks down the real risks, the technical reasons they happen, and the practical checklist both shoppers and brands can use to demand inclusive AI that performs across faces, ages, and undertones.
The beauty industry is already being rewritten by AI, with retailers using facial analysis and first-party data to personalize recommendations and streamline shopping. Ulta’s push into AI-assisted consultation is a good signal of where the market is headed, especially as retailers race to meet shoppers where they already are: on their phones, in apps, and in social commerce. Yet personalization only works if the underlying model sees everyone clearly. That is why beauty tech ethics is not a side conversation; it is part of product quality. If you want to understand where this fits inside broader beauty tech shifts, it helps to read our guide on AI’s Beauty Makeover: Personalization Without the Creepy Factor and compare it with our deeper retail lens on what brands should demand when agencies use agentic tools in pitches.
Why inclusivity in virtual try-on is now a product requirement, not a bonus
AI beauty tools are only as good as the faces they were trained on
Most virtual try-on systems rely on a combination of computer vision, facial landmark detection, segmentation, and color rendering. Those steps determine where the lips end, where the eyelid begins, how reflective a gloss looks, and whether a blush placement seems believable. If the training data overrepresents a narrow range of skin tones or face structures, the model may misread or oversimplify other faces. That creates a “works on me, fails on you” problem that shoppers notice instantly, even if the underlying engine is technically advanced.
This is why model training data is not just an engineering detail; it is the foundation of trust. In adjacent AI fields, the difference between a strong dataset and a weak one can be the difference between reliable automation and a costly failure, as shown in our piece on MLOps for hospitals and the practical lessons in benchmarking OCR accuracy across scanned documents. The same principle applies here: if the model does not encounter enough diversity during training, it cannot generalize gracefully in the wild.
Beauty shoppers judge accuracy by feel, not by model metrics
A brand might report improved detection rates, but shoppers evaluate a try-on by whether it matches their face in a mirror or in a selfie. If a foundation preview makes deep skin tones look gray, if a shimmer shadow disappears on rich eyelids, or if eye-lift placement floats above hooded lids, the experience feels broken. Accuracy is therefore both technical and emotional. The tool can be “objectively” functional while still feeling unusable to the person trying it.
That gap is why businesses should not measure only click-through or conversion. They need qualitative feedback from real users, especially from underrepresented groups. To see how feedback loops can shape launch decisions, our article on auditing comment quality as a launch signal offers a useful model for gathering real-world evidence rather than relying on vanity metrics.
Retailers now have a commercial incentive to get inclusivity right
As AI becomes a shopping assistant, a beauty retailer’s reputation will increasingly be tied to whether the tool is fair, useful, and representative. A shopper who cannot trust a try-on will abandon it and often abandon the purchase altogether. That is especially true in prestige beauty, where customers expect a high-touch experience and strong shade fidelity. Inclusive AI is therefore not just ethics; it is conversion infrastructure.
We see the same pattern in other sectors where trust determines adoption. Consider how companies think about security posture disclosure or how creators evaluate authentic narratives that build long-term trust. In beauty tech, the trust layer is the interface between the shopper’s face and the model’s interpretation of it.
Where virtual try-on bias shows up: skin tone, eye shape, age, and lighting
Skin tone representation and undertone drift
One of the most common failures in virtual try-on is undertone drift. On some skin tones, a foundation shade may look too ashy, too orange, or too flat because the model is rendering color without accounting for melanin depth, surface reflectance, and camera white balance. The result is that the product preview may not resemble the real finish once applied. That can lead shoppers to choose the wrong shade, especially when they are already uncertain.
Brands should test for representation across the full range of skin tones, from very fair to very deep, including neutral, cool, and warm undertones. Shoppers should also treat any tool as a preview, not a verdict, and compare it against customer swatches, user photos, and retailer shade-finding support. For a practical consumer lens on claims and labeling, see our guide to verifying claims before you buy and our checklist on tools that help you verify coupons before you buy, which uses a similar proof-before-purchase mindset.
Eye shape mapping and eyelid geometry
Eye shape mapping is where many tools fall down. Hooded eyes, monolids, deep-set eyes, round eyes, upturned eyes, and downturned eyes each require different placement logic for eyeliner, shadow cut-crease effects, and lash overlays. If a model assumes one universal eyelid geometry, the virtual look will float in the wrong position or warp when the user turns their head. This is more than a cosmetic defect; it is a sign that the system has been overfit to a narrow facial template.
Brands building eye-focused tools should think in terms of landmark flexibility, not one-size-fits-all masks. That means checking whether the segmentation works in profile, with glasses, with lashes, and with varying lid space. For product teams looking to build stronger category boundaries in AI experiences, our article on building fuzzy search for AI products with clear product boundaries is a useful parallel: the model must know what it is actually mapping, and what it should not assume.
Age diversity and facial texture realism
Age is often ignored in beauty AI, but it changes almost everything about how makeup sits on the face. Fine lines, texture, under-eye shadowing, eyelid laxity, and changes in lip contour can all affect whether a try-on appears realistic. If a system smooths everything aggressively, it may look flattering in a demo and misleading in use. If it cannot render mature skin accurately, it risks excluding a huge and valuable audience.
Accessible tech design means accounting for life-stage diversity, not just young, highly lit creators. This is an area where broader product strategy matters, similar to how teams serving older adults should design with real user constraints in mind, as discussed in product ideas and partnerships for tech-savvy older adults. Inclusivity is not limited to race or gender; it includes age, facial movement, and the way skin changes over time.
Lighting, camera quality, and device bias
Sometimes the bias is not in the person; it is in the camera pipeline. Low-light selfies, front-facing cameras with aggressive smoothing, and color-correcting phone software can distort the source image before the AI even begins analysis. That means a shopper may see different results depending on the device they use, which is a major accessibility and consistency problem. Tools should be tested across devices, bandwidth conditions, and lighting scenarios so that beauty recommendations remain stable.
In other digital systems, engineers already understand the value of reliability under imperfect conditions, as shown in edge computing for smart homes and routing resilience in logistics. Beauty AI needs the same mindset: robustness first, then polish.
What inclusive AI looks like in beauty tech
Diverse model training data with documented coverage
Inclusive AI starts with training data that reflects the real world. That means capturing a wide range of skin tones, eye shapes, ages, genders, and camera conditions. It also means documenting the dataset so internal teams can audit coverage and spot blind spots before release. If a vendor cannot explain what populations were represented, how images were labeled, or how bias was measured, that is a serious warning sign.
For brands, the strongest standard is to treat dataset documentation like any other quality control artifact. Ask what percent of the data comes from each skin tone range, what age bands are included, whether multiple lighting environments were used, and whether the model was re-validated after retraining. In retail and procurement environments, clear documentation has long been the difference between confidence and chaos, which is why guides like designing trading-grade cloud systems for volatile markets and how data quality claims impact bot trading are surprisingly relevant to beauty AI procurement.
User testing with real shoppers, not only internal staff
Internal demos are not enough. The people building the system are often the least likely to notice bias because they share assumptions, devices, and familiarity with the product. Inclusive user testing should include participants across multiple skin tones, eye shapes, facial structures, ages, and accessibility needs. It should also include people who do not identify with conventional beauty categories, because AI tools often fail at the edges first.
Ask testers to complete tasks such as shade matching, eyeliner placement, lash previewing, and before-and-after comparison under different lighting. Then record not only accuracy scores but emotional responses: Did the result look like them? Did it make them more confident or less? For a framework on collecting trusted feedback loops, see how to turn executive interviews into a high-trust live series and adapt the principle of high-trust dialogue to beauty usability sessions.
Accessible interfaces and clear controls
A truly inclusive try-on should also be accessible. That means readable contrast, keyboard-friendly interactions, clear labels, adjustable camera controls, and alternatives for users who do not want to upload a face photo. Some users may prefer manual shade matching or a guided quiz, while others may need reduced motion or lower visual complexity. A tool that is technically accurate but hard to use is still excluding people.
Accessible tech is strongest when it reduces cognitive load, not just visual clutter. That principle shows up in product categories well beyond beauty, including mobile device accessory planning and building a creator resource hub that gets found in traditional and AI search. In all cases, the interface should help users make confident decisions with minimal friction.
Practical checklist for shoppers: how to tell if a virtual try-on is trustworthy
Before you trust the result, test the system like a skeptic
Shoppers do not need a machine-learning degree to spot red flags. Start by trying the tool in different lighting conditions and, if possible, on more than one device. Compare the results for accuracy across skin tone, lip color, and eye shape. If the tool seems to “clean up” your face too aggressively, or if makeup placement looks offset, assume the product is prioritizing visual polish over realistic rendering.
Also check whether the brand offers multiple paths to purchase. A trustworthy beauty tech experience may combine try-on, customer reviews, ingredient information, and human advice. That multi-signal approach is similar to the way smart shoppers evaluate deals using several sources, as seen in best tech deals under the radar and coupon verification tools.
Look for proof of diversity, not just a diversity statement
Many brands say they value inclusion. Fewer show evidence. Look for model cards, fairness statements, user testing summaries, or before-and-after examples across multiple skin tones and face shapes. If the brand only showcases light skin and a single eye shape in its marketing, that is usually a clue that the model may be trained or tuned in a similarly narrow way. Transparency does not guarantee accuracy, but the absence of transparency is a meaningful risk signal.
Shoppers can also read reviews strategically. Search for commentary from users with similar skin tone, eye shape, or age range to your own. If several users report that the same shade looks different in person than in the app, treat that as evidence rather than anecdote. For a structured review lens, our guide on programmatic strategies to replace fading local news audiences shows how to identify patterns in noisy feedback.
Use try-ons as a starting point, then validate in the real world
Think of virtual try-on as a high-speed filter, not the final decision. If a foundation appears promising in the app, check swatches on the neck in daylight or confirm shade matches with a sample or return-friendly policy. For eye makeup, compare the virtual result with tutorials or creator looks from people with similar eye shape. This step is especially important for fine-grain products like concealer, contour, and nude lipstick, where minor inaccuracies become highly visible on the face.
The smartest shoppers use AI as one input among many, not a substitute for judgment. That mindset is similar to how consumers compare categories before spending, whether they are evaluating daily commuter credit cards or hunting for last-minute conference deals. The more expensive or face-specific the product, the more validation you need.
Practical checklist for brands: how to reduce bias and improve representation
Audit data, labels, and performance by subgroup
Brands should not ship a virtual try-on without subgroup evaluation. Break results down by skin tone, age bracket, eye shape, device type, and lighting condition. Track where performance degrades, and do not average away the problem with a single overall score. A model that performs well on the majority but fails a minority group is not inclusive; it is selectively accurate.
Set explicit thresholds for acceptable variance. If error rates are higher on deeper skin tones or monolid eye shapes, retrain, rebalance, or redesign before launch. Teams that work in high-stakes environments already know this discipline, which is why articles such as vendor security for competitor tools and automating HR with agentic assistants matter: governance has to be built into the workflow, not added after an incident.
Bring in domain experts and community testers early
Inclusive beauty AI benefits from makeup artists, dermatology-informed advisors, accessibility experts, and testers from the communities most likely to be misrepresented. Early feedback is cheaper than post-launch rework, and it is more honest because users are not yet trying to defend a finished product. Brands should compensate testers fairly and treat their input as product data, not marketing decoration.
If your team is used to working with agencies or outside vendors, apply the same standards you would use for other strategic projects. Demand documentation, testing protocols, and escalation plans, much like the due diligence recommended in what brands should demand when agencies use agentic tools in pitches and the process rigor found in AI for managing freelancers and editorial queues.
Design for transparency, consent, and user control
Users should know what happens to their face data. Are images stored? Are they used to improve the model? Can a shopper opt out of training? Can they use the tool without creating an account? These questions matter because facial analysis tools can feel invasive if the privacy model is vague. Clear consent language and simple controls are part of beauty tech ethics, not legal fine print.
For brands, transparency can become a differentiator. The companies that explain how their system works, where it is weak, and how they are improving will earn more trust over time than those that hide behind “AI magic.” That honesty-driven approach is similar to the trust-building strategies in founder storytelling without the hype and the operational clarity discussed in navigating change in martech.
Comparison table: what to check before trusting a virtual try-on
| Checkpoint | What good looks like | Red flags | Why it matters |
|---|---|---|---|
| Skin tone representation | Results look natural across light, medium, deep, and undertone variations | Gray cast, orange cast, or washed-out rendering on deeper skin tones | Prevents shade mismatch and trust loss |
| Eye shape mapping | Liner, shadow, and lash overlays follow hooded, monolid, round, and deep-set eyes correctly | Floating makeup, warped overlays, or identical placement on every eye | Ensures realistic eye looks and usable guidance |
| Age diversity | Model handles mature skin texture, fine lines, and different lid structure | Over-smoothing or breaking on older faces | Expands relevance and prevents exclusion |
| Lighting robustness | Stable output across indoor, outdoor, low-light, and phone-camera conditions | Major changes from one device or light source to another | Improves consistency in real shopping conditions |
| Transparency | Brand explains data sources, limitations, and consent terms clearly | No documentation, no fairness info, vague privacy language | Supports beauty tech ethics and informed choice |
| User testing | Feedback from diverse real shoppers is visible in product iteration | Only internal demos or influencer-only validation | Reduces blind spots and improves performance |
How brands can operationalize inclusive AI without slowing innovation
Make bias testing part of the release gate
Inclusive AI works best when it is not a side project. Put bias testing into the release process the same way payment, privacy, and security checks are handled. If subgroup performance fails, the product does not ship. This is the same logic used in reliable engineering teams that treat system readiness as non-negotiable, as discussed in secure data pipelines from wearables to EHR and automating IT admin tasks.
Invest in dataset refreshes, not one-time launches
Faces change, camera technology changes, and beauty trends change. A dataset that was strong two years ago may already be stale. Brands need scheduled refreshes, new user panels, and continuous monitoring for drift. That matters especially for seasonal colors, new product textures, and demographic shifts in the user base.
Think of the model like a living retail asset, not a static asset. The same long-term optimization mindset appears in when to buy an industry report versus DIY and marginal ROI for tech teams: the best results come from ongoing measurement, not one-time investment.
Use the commercial upside to justify the ethical work
Some teams still treat inclusivity as a cost center. In reality, better representation often improves conversion, returns, repeat purchase, and brand affinity. If a try-on feels accurate across more faces, more shoppers can use it confidently. That means fewer shade returns, fewer abandoned carts, and stronger loyalty in a category where trust is everything.
The business case is visible across beauty retail as AI becomes a shopping companion rather than a novelty. Retailers that combine first-party data, smart personalization, and reliable visualization are likely to win more basket share. But the winners will also be the ones that avoid the “creepy factor” by building systems that respect people while helping them shop, much like the balanced approach outlined in our personalization and privacy guide.
Key takeaways for shoppers, brands, and product teams
For shoppers: trust, but verify
Use virtual try-ons as a discovery tool, not the final authority. Compare results across lighting, device types, and user reviews from people with similar features. If the tool consistently misses your skin tone or eye shape, do not blame yourself; treat it as a product limitation. Demand better, because the market will improve only if users keep pointing out where it falls short.
For brands: inclusive AI is a measurable standard
Do not settle for broad promises about diversity. Test subgroup performance, document dataset coverage, involve diverse users, and make privacy and consent clear. The brands that do this well will stand out because they are solving the most important problem in AI beauty: making the technology work for real humans, not idealized templates.
For product teams: build for representation from day one
Do not wait until launch to discover that the model fails on hooded eyes or deeper skin tones. Bake inclusive AI into data collection, annotation, QA, and user testing from the beginning. This is how accessible tech earns both trust and adoption, and how beauty tech can grow without excluding the very people it claims to serve.
Pro Tip: If a virtual try-on looks impressive in a polished demo but fails when you switch skin tone, eye shape, device, or lighting, the model is optimized for presentation — not representation.
FAQ: inclusive AI and virtual try-ons
How can I tell if a virtual try-on is biased?
Test it against your real features in different lighting and compare the result to photos, swatches, or in-store samples. If makeup placement shifts dramatically, skin looks gray or overly warm, or eye makeup floats away from your actual lid shape, the tool is likely biased or poorly calibrated. Consistent errors across similar users are another major warning sign.
What should brands ask AI vendors before buying a virtual try-on system?
Ask what skin tones, eye shapes, ages, and camera conditions were represented in the training data. Request subgroup performance reports, documentation of labeling methods, privacy terms, and details on how often the model is retrained. Also ask whether the vendor has tested the system with real users from underrepresented groups, not just internal staff.
Is a shade-matching tool enough to call a beauty app inclusive?
No. Shade matching is only one piece of the experience. Inclusive AI should also handle eye shape mapping, age diversity, facial texture, device variation, and accessibility needs. A tool can be strong in one area and still be exclusionary in others.
Why does eye shape mapping matter so much in makeup try-ons?
Because eye makeup is highly geometry-dependent. Hooded, monolid, deep-set, and round eyes all need different placement rules for shadows, eyeliner, and lashes. If the model assumes one eye structure, the preview will look unrealistic or unusable for many shoppers.
What is the biggest ethical issue with facial analysis in beauty tech?
Consent and transparency are the biggest issues, followed closely by bias. Users should know how their face data is processed, whether it is stored, and if it may be used to train future models. Without clear disclosure, even a helpful feature can feel invasive.
How often should brands retest inclusive AI tools?
At minimum, every major model update, every significant camera or app change, and on a recurring schedule because user behavior and hardware change over time. Continuous monitoring is better than annual reviews, especially in fast-moving beauty tech categories where new products and formats launch constantly.
Related Reading
- AI’s Beauty Makeover: Personalization Without the Creepy Factor - Learn how brands can personalize beauty without crossing privacy lines.
- What Brands Should Demand When Agencies Use Agentic Tools in Pitches - A practical checklist for vendor scrutiny and responsible AI procurement.
- MLOps for Hospitals: Productionizing Predictive Models that Clinicians Trust - A governance-first look at deploying reliable AI systems.
- Building Fuzzy Search for AI Products with Clear Product Boundaries: Chatbot, Agent, or Copilot? - A useful framework for defining AI product behavior and limits.
- Benchmarking OCR Accuracy Across Scanned Contracts, Forms, and Procurement Documents - Why dataset quality and testing coverage determine real-world accuracy.
Related Topics
Maya Thompson
Senior Beauty Tech Editor
Senior editor and content strategist. Writing about technology, design, and the future of digital media. Follow along for deep dives into the industry's moving parts.
Up Next
More stories handpicked for you
Ingredient Decoder: How to Read Cosmetic Labels and Choose Safer Skincare
Build the Perfect Everyday Makeup Kit: Essentials for Every Age and Budget
Stream your Style: The Intersection of Live Commerce and Beauty Education
Are Eyeshadow Palettes Dying? What the TikTok Decline and Market Data Really Mean for Your Makeup Bag
Late Bloomers, Big Wins: What a 78-Year-Old Author's Pivot Teaches Beauty Entrepreneurs
From Our Network
Trending stories across our publication group