Go back
Authors:
Felix Kim & Redrob Research Labs
Date:
Oct 15, 2025
Abstract
General-purpose large language models (LLMs) demonstrate suboptimal performance in region-specific HR and sales workflows due to lack of exposure to local nomenclature, cultural context, and domain conventions. This paper examines how incorporating localized resume corpora and workflow data through fine-tuning, retrieval-augmented generation (RAG), and continued pretraining improves task-specific performance while reducing computational costs by 15-40× compared to frontier models. We demonstrate that models trained on 2.3M Indian resumes and 450K regional job descriptions achieve 34% higher accuracy on candidate matching tasks and 28% improvement in sales lead qualification, while operating at 20× lower inference costs.
Introduction
Modern LLMs trained on predominantly Western datasets exhibit systematic biases when deployed in emerging market contexts. In HR applications, these models struggle with non-standardized educational credentials (e.g., Indian polytechnic degrees vs. traditional B.Tech programs), regional naming conventions, and localized career trajectories. Similarly, in sales workflows, general-purpose models fail to capture region-specific business hierarchies, decision-making structures, and cultural communication patterns.
The cost structure of frontier models further compounds accessibility challenges. At $200/month per seat for enterprise GPT-4 access, deployment becomes economically infeasible for organizations operating in markets where median professional salaries range from $300-800/month. This creates an imperative for both performance optimization and cost reduction through domain-specific model development.
Methodology
Data Collection and Preprocessing
We assembled a proprietary corpus comprising:
2.3M anonymized resumes from Indian job platforms (2018-2024)
450K job descriptions across 15 industry verticals
1.8M recruiter-candidate interaction logs
340K sales conversation transcripts from B2B engagements in Southeast Asian markets
All personally identifiable information was removed through automated redaction pipelines. The dataset was structured to preserve domain-specific terminology, credential hierarchies, and workflow patterns unique to target regions.
Technical Approaches
Continued Pretraining (CPT)
We performed continued pretraining on a Llama 3.1-8B base model using our domain corpus for 50,000 additional steps. This approach allows the model to internalize region-specific vocabulary and semantic relationships without catastrophic forgetting of general capabilities. Training utilized 8×H100 GPUs over 72 hours with a learning rate of 1e-5 and batch size of 128.
Supervised Fine-Tuning (SFT)
Following CPT, we fine-tuned on 85,000 labeled examples covering:
Resume parsing and entity extraction
Candidate-job matching scores
Sales lead qualification
Email generation for recruiting and outreach
Fine-tuning employed LoRA (Low-Rank Adaptation) with rank=16 to reduce trainable parameters by 99.7% while maintaining performance gains, enabling training completion in 18 hours on 4×A100 GPUs.
Retrieval-Augmented Generation (RAG)
For deployment, we implemented a hybrid RAG system combining:
Vector database of 2.3M resume embeddings (1024-dimensional)
Semantic search using cosine similarity with 0.85 threshold
Dynamic context injection limited to top-k=5 most relevant documents
This architecture allows real-time incorporation of new data without model retraining, maintaining performance as the corpus grows.
Model Distillation
To further reduce costs, we distilled our fine-tuned 8B model into a 1.5B student model using knowledge distillation with temperature τ=2.5. The distilled model retains 94% of performance while achieving 5.3× faster inference and 6.2× lower memory footprint.
Results
Performance Gains
Resume Parsing Accuracy
Our domain-adapted model achieved 91.2% F1-score on entity extraction from Indian resumes compared to 67.4% for GPT-4 and 58.3% for base Llama 3.1. The model correctly identified non-standard credentials (e.g., "12th Standard" vs. "High School Diploma") in 89% of cases versus 41% for general-purpose models.
Candidate Matching
On a held-out test set of 12,000 job-candidate pairs rated by human recruiters, our model achieved 0.83 Spearman correlation with human judgments versus 0.62 for GPT-4. The model demonstrated particular strength in evaluating non-linear career progressions common in emerging markets, where professionals frequently shift between sectors.
Sales Lead Qualification
For B2B sales workflows, our model correctly classified high-intent leads with 82% precision and 78% recall, compared to 64% precision and 59% recall for general-purpose alternatives. The model learned to identify culturally-specific buying signals, such as hierarchical approval language patterns in Southeast Asian business communications.
Multilingual Performance
While primarily trained on English-language data with Hindi/Tamil code-switching, the model demonstrated 73% accuracy on Hindi-only resumes through transfer learning, suggesting effective cross-lingual generalization within the target demographic.
Cost Efficiency
Inference Cost Reduction
Our 8B fine-tuned model operates at $0.12 per 1M tokens versus $2.40 for GPT-4, representing a 20× cost reduction. The distilled 1.5B model further reduces costs to $0.02 per 1M tokens—a 120× improvement—while maintaining 86% of GPT-4's accuracy on domain tasks.
Training Cost Amortization
Total training costs (CPT + SFT + distillation) amounted to $14,000 for the full pipeline. At current deployment scale processing 500M tokens monthly, training costs are recovered within 2 months, after which operational savings compound indefinitely.
Latency Improvements
Our distilled model generates responses with p95 latency of 680ms versus 2,100ms for GPT-4 API calls, enabling real-time interactive applications at scale. This 3× latency reduction improves user experience while reducing infrastructure costs through higher throughput per GPU.
Discussion
Why Local Data Matters
The performance gains observed stem from three key factors:
Vocabulary Alignment: Emerging market resumes contain region-specific terminology absent from general training corpora. Terms like "articleship" (CA internship in India), "Matric" (10th grade), or "ITI diploma" rarely appear in Western datasets. Without exposure, general models either hallucinate interpretations or fail to extract these entities entirely.
Structural Patterns: Resume formatting conventions vary significantly by region. Indian resumes typically lead with educational credentials and family details, while Western resumes emphasize work experience. Models trained on localized data learn these structural priors, improving parsing accuracy by 23-31%.
Cultural Context: Sales and recruiting communications embed cultural norms around hierarchy, formality, and decision-making. Our models learned that phrases like "I will check with management" signal genuine interest in Indian B2B contexts, whereas general models interpret this as a soft rejection. These subtle contextual cues drive 15-20% improvement in lead scoring accuracy.
Scalability and Generalization
The methodology extends beyond HR and sales verticals. Any domain with:
Localized terminology and conventions
High-volume proprietary data availability
Clear evaluation metrics
represents a candidate for similar optimization. Early experiments in legal contract analysis for Indian corporate law and medical record processing for Southeast Asian healthcare systems show comparable 25-35% accuracy gains over general-purpose models.
Limitations
Current work focuses on English and code-mixed data. Pure vernacular language performance (Hindi-only, Tamil-only) remains 18-25% below English baselines, though this gap narrows with additional language-specific pretraining. Future work will expand to fully multilingual architectures.
The approach requires substantial upfront data collection and curation costs. Organizations without existing data repositories face 6-12 month collection periods before training becomes viable. However, network effects emerge as deployed models generate new training data through user interactions, creating a virtuous cycle.
Conclusion
Domain-specific data integration through continued pretraining, fine-tuning, and RAG enables simultaneous performance improvements and cost reductions of 15-40× for LLM deployment in regional HR and sales applications. Our results demonstrate that models trained on localized corpora achieve 28-34% higher accuracy than frontier general-purpose alternatives while operating at dramatically lower computational costs.
For organizations serving emerging markets, this approach transforms LLM economics from prohibitive to accessible. A $200/month ChatGPT Pro subscription becomes a $10/month domain-optimized deployment, enabling profitability at regional price points while delivering superior user experiences.
The methodology provides a template for democratizing AI access: rather than waiting for frontier models to incorporate niche regional data, organizations can achieve world-class performance through targeted optimization of smaller, efficient models. As emerging markets represent the next 3 billion AI users, this paradigm shift from universal to specialized models becomes essential infrastructure for global AI accessibility.
Privacy policy