OpenAI trained GPT-4 on the internet. A disproportionate share of that internet is in English, by Americans, about American concerns, shaped by American values and legal frameworks. This is not a conspiracy — it is an engineering inevitability. You build what you have data for.
India has 1.4 billion people. 22 scheduled languages. A legal system descended from British common law but modified beyond recognition over 75 years of sovereign jurisprudence. A healthcare system split between allopathy, Ayurveda, and 10,000 quacks. A financial system where half the population has never filed a tax return.
None of this is well-represented in any major AI model today.
What "Sovereign AI" Actually Means
The phrase has been captured by politicians who use it to mean "AI made in India by Indians." That's nationalism, not a technical specification.
True sovereign AI means four things:
- Data sovereignty: The training data reflects Indian context — not translated English. Indian legal judgements, not US case law. Indian dialects, not machine-translated Hindi.
- Infrastructure sovereignty: Compute hosted on Indian soil, under Indian law, not subject to US export controls or CLOUD Act jurisdiction.
- Model sovereignty: Weights owned and controlled by Indian entities, not licensed from foreign providers who can terminate access.
- Value alignment: The model's embedded values, biases, and assumptions reflect Indian pluralism — not a particular coastal American worldview.
Why Nobody Is Actually Building It
The honest reason: it's hard and it doesn't pay immediately.
Training a genuinely Indian-aligned LLM requires:
- Curating high-quality Indian language corpora — a 3–5 year academic project, not a weekend hackathon.
- Building RLHF pipelines with Indian annotators who understand regional, religious, and linguistic nuance.
- Navigating India's fragmented data landscape — there is no Indian equivalent of The Pile or Common Crawl with meaningful Indic coverage.
- Securing GPU compute at a cost that doesn't require a Series B before you have product-market fit.
The companies that have money (Reliance, TCS, Infosys) are not AI-native. The companies that are AI-native don't have the money. The government-funded initiatives (AIRAWAT, NLP4Bharat) are real but slow.
What We're Actually Doing
Lexcore's approach is not to build a foundation model. We are not competing with Meta or Google on pretraining. We are building the application and fine-tuning layer on top of open-weight models — specifically LLaMA 3.1 — using Indian data, for Indian use cases, on Indian infrastructure.
This is pragmatic, not idealistic. The pragmatic path to sovereign AI in India is:
- Use the best open-weight base models available.
- Fine-tune aggressively on Indian domain data.
- Serve on infrastructure you control.
- Build the product experience around Indian user behaviour.
Naira AI is the public face of this effort. Cortina DUC is the enterprise deployment vehicle. The Cortina Infinity platform is the cloud-based service layer. Together, they constitute a stack that is more "sovereign" than anything a foreign provider can offer — even if they localise their interface.
The Stakes
This is not just a business opportunity. AI is going to reshape how India's 700 million internet users access information, legal recourse, medical advice, and financial guidance. If that AI is built by foreign companies, trained on foreign data, serving foreign interests — India's digital future will be a colonial project with better UX.
That is not a future we're willing to accept.