Humane developed a foundational model from scratch trained on proprietary Arabic data. The primary goals were not to compete with global leaders, but to understand cultural nuances, address language biases, and, most importantly, train the internal team on building the entire AI stack from the ground up.

Related Insights

Popular benchmarks like MMLU are inadequate for evaluating sovereign AI models. They primarily test multiple-choice knowledge extraction but miss a model's ability to generate culturally nuanced, fluent, and appropriate long-form text. This necessitates creating new, culturally specific evaluation tools.

Reports that OpenAI hasn't completed a new full-scale pre-training run since May 2024 suggest a strategic shift. The race for raw model scale may be less critical than enhancing existing models with better reasoning and product features that customers demand. The business goal is profit, not necessarily achieving the next level of model intelligence.

The key for enterprises isn't integrating general AI like ChatGPT but creating "proprietary intelligence." This involves fine-tuning smaller, custom models on their unique internal data and workflows, creating a competitive moat that off-the-shelf solutions cannot replicate.

ElevenLabs' defense against giants isn't just a better text-to-speech model. Their strategy focuses on building deep, workflow-specific platforms for agents and creatives. This includes features like CRM integrations and collaboration tools, creating a sticky application layer that a foundational model alone cannot replicate.

Humane was founded after its CEO discovered it took oil giant Aramco nine months just to procure and deploy AI infrastructure. This massive delay, even for a well-resourced company, highlighted the foundational opportunity to build a national AI champion and regional digital hub for the Middle East.

Rather than committing to a single LLM provider like OpenAI or Gemini, Hux uses multiple commercial models. They've found that different models excel at different tasks within their app. This multi-model strategy allows them to optimize for quality and latency on a per-workflow basis, avoiding a one-size-fits-all compromise.

Initially, even OpenAI believed a single, ultimate 'model to rule them all' would emerge. This thinking has completely changed to favor a proliferation of specialized models, creating a healthier, less winner-take-all ecosystem where different models serve different needs.

Powerful AI products are built with LLMs as a core architectural primitive, not as a retrofitted feature. This "native AI" approach creates a deep technical moat that is difficult for incumbents with legacy architectures to replicate, similar to the on-prem to cloud-native shift.

To escape platform risk and high API costs, startups are building their own AI models. The strategy involves taking powerful, state-subsidized open-source models from China and fine-tuning them for specific use cases, creating a competitive alternative to relying on APIs from OpenAI or Anthropic.

For specific, high-leverage tasks like conversation summarization and re-ranking search results, Intercom trains its own custom models. These smaller, fine-tuned models have proven to be cheaper, faster, and higher quality than using general-purpose frontier models from vendors.