Get GenAI guide

Access HaxiTAG GenAI research content, trends and predictions.

Friday, July 18, 2025

OpenAI’s Seven Key Lessons and Case Studies in Enterprise AI Adoption

AI is Transforming How Enterprises Work

OpenAI recently released a comprehensive guide on enterprise AI deployment, openai-ai-in-the-enterprise.pdf, based on firsthand experiences from its research, application, and deployment teams. It identified three core areas where AI is already delivering substantial and measurable improvements for organizations:

  • Enhancing Employee Performance: Empowering employees to deliver higher-quality output in less time

  • Automating Routine Operations: Freeing employees from repetitive tasks so they can focus on higher-value work

  • Enabling Product Innovation: Delivering more relevant and responsive customer experiences

However, AI implementation differs fundamentally from traditional software development or cloud deployment. The most successful organizations treat AI as a new paradigm, adopting an experimental and iterative approach that accelerates value creation and drives faster user and stakeholder adoption.

OpenAI’s integrated approach — combining foundational research, applied model development, and real-world deployment — follows a rapid iteration cycle. This means frequent updates, real-time feedback collection, and continuous improvements to performance and safety.

Seven Key Lessons for Enterprise AI Deployment

Lesson 1: Start with Rigorous Evaluation
Case: How Morgan Stanley Ensures Quality and Safety through Iteration

As a global leader in financial services, Morgan Stanley places relationships at the core of its business. Faced with the challenge of introducing AI into highly personalized and sensitive workflows, the company began with rigorous evaluations (evals) for every proposed use case.

Evaluation is a structured process that assesses model performance against benchmarks within specific applications. It also supports continuous process improvement, reinforced with expert feedback at each step.

In its early stages, Morgan Stanley focused on improving the efficiency and effectiveness of its financial advisors. The hypothesis was simple: if advisors could retrieve information faster and reduce time spent on repetitive tasks, they could provide more and better insights to clients.

Three initial evaluation tracks were launched:

  • Translation Accuracy: Measuring the quality of AI-generated translations

  • Summarization: Evaluating AI’s ability to condense information using metrics for accuracy, relevance, and coherence

  • Human Comparison: Comparing AI outputs to expert responses, scored on accuracy and relevance

Results: Today, 98% of Morgan Stanley advisors use OpenAI tools daily. Document access has increased from 20% to 80%, and search times have dropped dramatically. Advisors now spend more time on client relationships, supported by task automation and faster insights. Feedback has been overwhelmingly positive — tasks that once took days now take hours.

Lesson 2: Embed AI into Products
Case: How Indeed Humanized Job Matching

AI’s strength lies in handling vast datasets from multiple sources, enabling companies to automate repetitive work while making user experiences more relevant and personalized.

Indeed, the world’s largest job site, now uses GPT-4o mini to redefine job matching.

The “Why” Factor: Recommending good-fit jobs is just the beginning — it’s equally important to explain why a particular role is suggested.

By leveraging GPT-4o mini’s analytical and language capabilities, Indeed crafts natural-language explanations in its messages and emails to job seekers. Its popular "invite to apply" feature also explains how a candidate’s background makes them a great fit.

When tested against the prior matching engine, the GPT-powered version showed:

  • A 20% increase in job application starts

  • A 13% improvement in downstream hiring success

Given that Indeed sends over 20 million messages monthly and serves 350 million visits, these improvements translate to major business impact.

Scaling posed a challenge due to token usage. To improve efficiency, OpenAI and Indeed fine-tuned a smaller model that achieved similar results with 60% fewer tokens.

Helping candidates understand why they’re a fit for a role is a deeply human experience. With AI, Indeed is enabling more people to find the right job faster — a win for everyone.

Lesson 3: Start Early, Invest Ahead of Time
Case: Klarna’s Compounding Returns from AI Adoption

AI solutions rarely work out-of-the-box. Use cases grow in complexity and impact through iteration. Early adoption helps organizations realize compounding gains.

Klarna, a global payments and shopping platform, launched a new AI assistant to streamline customer service. Within months, the assistant handled two-thirds of all service chats — doing the work of hundreds of agents and reducing average resolution time from 11 to 2 minutes. It’s expected to drive $40 million in profit improvement, with customer satisfaction scores on par with human agents.

This wasn’t an overnight success. Klarna achieved these results through constant testing and iteration.

Today, 90% of Klarna’s employees use AI in their daily work, enabling faster internal launches and continuous customer experience improvements. By investing early and fostering broad adoption, Klarna is reaping ongoing returns across the organization.

Lesson 4: Customize and Fine-Tune Models
Case: How Lowe’s Improved Product Search

The most successful enterprises using AI are those that invest in customizing and fine-tuning models to fit their data and goals. OpenAI has invested heavily in making model customization easier — through both self-service tools and enterprise-grade support.

OpenAI partnered with Lowe’s, a Fortune 50 home improvement retailer, to improve e-commerce search accuracy and relevance. With thousands of suppliers, Lowe’s deals with inconsistent or incomplete product data.

Effective product search requires both accurate descriptions and an understanding of how shoppers search — which can vary by category. This is where fine-tuning makes a difference.

By fine-tuning OpenAI models, Lowe’s achieved:

  • A 20% improvement in labeling accuracy

  • A 60% increase in error detection

Fine-tuning allows organizations to train models on proprietary data such as product catalogs or internal FAQs, leading to:

  • Higher accuracy and relevance

  • Better understanding of domain-specific terms and user behavior

  • Consistent tone and voice, essential for brand experience or legal formatting

  • Faster output with less manual review

Lesson 5: Empower Domain Experts
Case: BBVA’s Expert-Led AI Adoption

Employees often know their problems best — making them ideal candidates to lead AI-driven solutions. Empowering domain experts can be more impactful than building generic tools.

BBVA, a global banking leader with over 125,000 employees, launched ChatGPT Enterprise across its operations. Employees were encouraged to explore their own use cases, supported by legal, compliance, and IT security teams to ensure responsible use.

“Traditionally, prototyping in companies like ours required engineering resources,” said Elena Alfaro, Global Head of AI Adoption at BBVA. “With custom GPTs, anyone can build tools to solve unique problems — getting started is easy.”

In just five months, BBVA staff created over 2,900 custom GPTs, leading to significant time savings and cross-departmental impact:

  • Credit risk teams: Faster, more accurate creditworthiness assessments

  • Legal teams: Handling 40,000+ annual policy and compliance queries

  • Customer service teams: Automating sentiment analysis of NPS surveys

The initiative is now expanding into marketing, risk, operations, and more — because AI was placed in the hands of people who know how to use it.

Lesson 6: Remove Developer Bottlenecks
Case: Mercado Libre Accelerates AI Development

In many organizations, developer resources are the primary bottleneck. When engineering teams are overwhelmed, innovation slows, and ideas remain stuck in backlogs.

Mercado Libre, Latin America's largest e-commerce and fintech company, partnered with OpenAI to build Verdi, a developer platform powered by GPT-4o and GPT-4o mini.

Verdi integrates language models, Python, and APIs into a scalable, unified platform where developers use natural language as the primary interface. This empowers 17,000 developers to build consistently high-quality AI applications quickly — without deep code dives. Guardrails and routing logic are built-in.

Key results include:

  • 100x increase in cataloged products via automated listings using GPT-4o mini Vision

  • 99% accuracy in fraud detection through daily evaluation of millions of product listings

  • Multilingual product descriptions adapted to regional dialects

  • Automated review summarization to help customers understand feedback at a glance

  • Personalized notifications that drive engagement and boost recommendations

Next up: using Verdi to enhance logistics, reduce delivery delays, and tackle more high-impact problems across the enterprise.

Lesson 7: Set Bold Automation Goals
Case: How OpenAI Automates Its Own Work

At OpenAI, we work alongside AI every day — constantly discovering new ways to automate our own tasks.

One challenge was our support team’s workflow: navigating systems, understanding context, crafting responses, and executing actions — all manually.

We built an internal automation platform that layers on top of existing tools, streamlining repetitive tasks and accelerating insight-to-action workflows.

First use case: Working on top of Gmail to compose responses and trigger actions. The platform pulls in relevant customer data and support knowledge, then embeds results into emails or takes actions like opening support tickets.

By integrating AI into daily workflows, the support team became more efficient, responsive, and customer-centric. The platform now handles hundreds of thousands of tasks per month — freeing teams to focus on higher-impact work.

It all began because we chose to set bold automation goals, not settle for inefficient processes.

Key Takeaways

As these OpenAI case studies show, every organization has untapped potential to use AI for better outcomes. Use cases may vary by industry, but the principles remain universal.

The Common Thread: AI deployment thrives on open, experimental thinking — grounded in rigorous evaluation and strong safety measures. The best-performing companies don’t rush to inject AI everywhere. Instead, they align on high-ROI, low-friction use cases, learn through iteration, and expand based on that learning.

The Result: Faster and more accurate workflows, more personalized customer experiences, and more meaningful work — as people focus on what humans do best.

We’re now seeing companies automate increasingly complex workflows — often with AI agents, tools, and resources working in concert to deliver impact at scale.

Related topic:

Exploring HaxiTAG Studio: The Future of Enterprise Intelligent Transformation
Leveraging HaxiTAG AI for ESG Reporting and Sustainable Development
Revolutionizing Market Research with HaxiTAG AI
How HaxiTAG AI Enhances Enterprise Intelligent Knowledge Management
The Application of HaxiTAG AI in Intelligent Data Analysis
The Application and Prospects of HaxiTAG AI Solutions in Digital Asset Compliance Management
Report on Public Relations Framework and Content Marketing Strategies