AI team structure: the optimal setup
Most organizations build AI teams backward, hiring specialists before defining what they actually need to build. The most effective university AI lab setup starts with three core functions, cloud infrastructure, and a hybrid collaboration model that scales with real problems.

The pattern at university AI labs is almost scripted. Someone gets funding. Job postings go up. Suddenly there’s a ten-specialist hiring spree before anyone defines what the team is actually supposed to build.
This fails. Consistently.
Institutions spend months assembling dream teams that never ship because nobody defined the underlying functions first. AI teams need data scientists, ML engineers, and AI architects working alongside business domain experts. Most organizations confuse roles with functions and end up with expensive overlap and zero accountability. Meanwhile, 87% of tech leaders already can’t find the skilled AI workers they need.
Why most AI teams fail before they start
The problem isn’t talent. It’s structure.
Stanford HAI’s 2025 AI Index found 78% of organizations now deploy AI in at least one function. Only 6% report meaningful ROI. That gap should alarm anyone planning an AI lab from scratch.
Recent research on agentic organizations paints a different picture: small outcome-focused teams of 2-5 people can now supervise 50-100 specialized AI agents running end-to-end processes. Universities keep building teams like it’s 2018. Massive. Centralized. Disconnected from actual use cases.
When Princeton built their AI Lab, they didn’t start with dozens of researchers. They created shared infrastructure first: 300 H100 GPUs, administrative support, research software engineers. Then specific projects attracted the right specialists.
The University of Tokyo went further. Their Matsuo-Iwasawa Laboratory equipped actual hardware environments including robot arms, mobile manipulators, simulators, and VR devices. They grew from core faculty to 50 members through a research community model that attracted talent to problems, not positions.
Start with infrastructure and clear functions.
Talent follows.
The three roles that actually matter
Forget the ten-specialist fantasy. A working university AI lab needs three core roles that map to actual work.
Research engineers who experiment and prototype. These are the people testing hypotheses, exploring new approaches, figuring out what’s actually possible with current technology. Not pure theorists. Not production engineers. Researchers who code.
ML engineers who move prototypes into production. These engineers focus on transitioning models from research to systems that operate in real environments. The numbers from Talent500’s job trends analysis make this clear: the majority of enterprise AI initiatives struggle without dedicated operational support, which is why MLOps skills are now minimum requirements, not differentiators.
Infrastructure specialists who keep systems running. Data engineers construct and maintain the data pipelines that make AI development possible. AI certifications like Google ML Engineer and AWS ML Specialty are linked to 20-25% salary premiums for data engineers. Without solid infrastructure, both research and production grind to a halt.
Everything else - data scientists, ethicists, NLP specialists, security officers - maps to these three functions or gets added when specific projects demand it. The IT skills shortage is projected to cause trillions in cumulative losses according to IDC. You won’t hire your way into ten distinct roles. You’ll burn budget trying.
Build the three core functions first. Specialists emerge from project needs.
Cloud versus on-premise for university labs
On-premise infrastructure requires massive upfront investment. Hardware, cooling, power, maintenance staff, physical security. The cost math is unfavorable: on-premise AI workloads need substantial initial capital plus ongoing costs for power consumption, cooling systems, space, and maintenance. On-premise can be more cost-effective over time for organizations running AI workloads continuously, but most university labs don’t fit that profile.
Universities don’t run AI workloads continuously. That’s the part most lab planners miss.
Classes happen in bursts. Research projects ramp up and wind down. Student projects spike during semesters then disappear. Does any of that justify paying for continuous hardware capacity? No. And yet universities keep making this mistake.
CloudLabs and similar platforms solve this by providing cloud-based, customizable learning environments. Students get dedicated access to Big Data Analytics, Deep Learning, and NLP labs hosted on AWS, Azure, and GCP. When class ends, you’re not paying for idle GPUs gathering dust.
The Minnesota Supercomputing Institute took a different approach. They built shared on-premise HPC clusters that individual departments can access without each one buying its own hardware. Researchers run large-scale experiments concurrently on shared infrastructure, avoiding per-department capital spending even though the model itself is on-premise.
For teaching and research that varies by semester, cloud wins on economics and student experience. Students learn the same platforms they’ll use professionally. Universities avoid hardware refresh cycles and maintenance overhead. Reserve on-premise for the rare cases where sustained, predictable workloads actually justify the capital investment.
Hybrid models beat pure centralization
The debate shouldn’t be centralized versus decentralized. It should be about which elements belong in each category.
AWS published a useful piece on generative AI operating models that recommends centralizing foundations - infrastructure, data governance, security standards - while distributing innovation across business domains. This hybrid approach keeps AI governance solid while letting teams move fast on delivery.
Pure centralization creates bottlenecks. Every department waits for the central AI team to get around to their project. TDWI’s research on AI team structures backs this up: mid-size organizations tend to fully centralize, but this sacrifices speed and domain alignment as they grow.
Pure decentralization fragments everything. Each department builds its own solutions that don’t talk to each other, all reinventing the same infrastructure and governance patterns. The numbers are sobering: only about 5% of companies qualify as “future-built” for AI, and they’re 1.5x more likely to adopt shared ownership between business and IT departments.
The hybrid or federated model, sometimes called hub-and-spoke, centralizes infrastructure, security, and standards while embedding AI specialists in department teams. University AI lab setup maintains consistent data quality and security while letting departments move fast on domain-specific problems.
Airbnb learned this through experience. They transitioned from fully centralized data science to a hybrid model as they grew, keeping the data science team together for career development and standards while splitting it into sub-teams aligned with specific product areas.
Build your hub first. Grow spokes as departments prove they’re ready.
Building skills instead of buying talent
The math doesn’t work on hiring.
The WEF’s Future of Jobs Report 2025 puts a number on it: 63% of employers cite the skills gap as the key barrier to business transformation. Nearly 40% of global jobs are exposed to AI-driven change, and skill demands are evolving at a much faster clip in AI-exposed roles.
You can’t compete with tech companies offering equity and unlimited budgets. I think most university leaders already know this, but badly underestimate how much it limits their options. The alternative is developing internal talent.
85% of employers now plan to offer upskilling, and 77% provide AI training according to the WEF. This works because AI expertise builds on existing domain knowledge. Your biology faculty who understand the research problems just need the technical tools, not another PhD.
The key skills aren’t mysterious. Hiring managers surveyed recently ranked on-the-job training, industry certifications, and university coursework as the top pathways into AI roles. Real-world projects and applied skills matter most. A CS degree isn’t the only entry point anymore.
Universities have genuine advantages here. Cloud-based ML certifications like AWS Machine Learning Specialty are linked to roughly 20% salary boosts in existing data and engineering roles. AWS leads cloud market share for ML workloads, and 73% of organizations actively prioritize AI-certified talent.
The infrastructure to train your own people exists. Use it before burning budget on hiring battles you’ll lose.
Stop planning the perfect team. Give three people who want to learn some cloud credits and real problems, then grow from there. The organizations that do well with AI don’t have the biggest teams or the most PhDs. They have clear functions, appropriate infrastructure, and people who learn by shipping real things.
In two years, the labs that started small and shipped fast will have lapped the ones still writing hiring plans. That gap only widens.
About the Author
Amit Kothari is an experienced consultant, advisor, coach, and educator specializing in AI and operations for executives and their companies. With 25+ years of experience and as the founder of Tallyfy (raised $3.6m), he helps mid-size companies identify, plan, and implement practical AI solutions that actually work. Originally British and now based in St. Louis, MO, Amit combines deep technical expertise with real-world business understanding.
Disclaimer: The content in this article represents personal opinions based on extensive research and practical experience. While every effort has been made to ensure accuracy through data analysis and source verification, this should not be considered professional advice. Always consult with qualified professionals for decisions specific to your situation.