Qwen2.5-3B-Instruct Fine-tuned for RBI Regulations Q&A

A specialized model for answering questions about Reserve Bank of India (RBI) regulations and banking policies.

Performance: 57.6% accuracy (8.2x improvement over base model's 7%)

Quick Facts

  • 🎯 Accuracy: 57.6% on 1000-sample evaluation set
  • πŸ“š Coverage: 100+ regulation areas (Basel III, FEMA, AML, PSL, etc.)
  • πŸš€ Training: 47K QA pairs with rephrased variants
  • ⚑ Efficient: 3B parameters, optimized for deployment

Performance Highlights

Category Base Model Fine-tuned Improvement
Overall 7.0% 57.6% +50.6%
Fact-based 6.8% 57.6% +50.8%
Reasoning 37.5% 62.5% +25.0%

Top Categories (70%+ accuracy): Anti-Money Laundering, Digital Payments, Government Banking, MSME Finance, Currency Management

Training Details

  • Method: LoRA fine-tuning (r=16, alpha=32)
  • Dataset: RBI-Circular-QA-Dataset (47K samples)
  • Training: 1 epoch, 2 hours on NVIDIA L40S
  • Loss: 0.79 β†’ 0.57 (train), 0.58 (eval)

Code & Resources

πŸ“ Technical Deep Dive

Want to understand the theory and methodology behind this model?

Read the full article: Fine-tuning Qwen 2.5 3B for RBI Regulations: Achieving 8x Performance with Smart Data Augmentation

The article covers:

  • 🧠 LoRA Theory: Why and how Low-Rank Adaptation works
  • ⚑ Unsloth Deep Dive: Technical advantages and performance optimizations
  • πŸ“Š Data Augmentation Strategy: The rephrasing technique that delivered 40% of improvement
  • πŸŽ“ Hyperparameter Analysis: Detailed explanation of every training choice
  • πŸ“ˆ Evaluation Methodology: Stratified sampling and LLM-as-judge approach
  • πŸ”¬ Ablation Studies: What really mattered for the 8x improvement

Perfect for ML engineers who want to replicate or adapt this approach for their own domain-specific fine-tuning projects.

Use Cases

βœ… Banking compliance chatbots
βœ… Regulatory Q&A systems
βœ… Training tools for banking professionals
βœ… RBI circular analysis

⚠️ Not for: Legal compliance decisions (requires expert review), real-time updates

Limitations

  • Knowledge cutoff: 2024 regulations
  • 57.6% accuracy means ~42% of complex queries need verification
  • Optimized for English only

Citation

  @misc{
    qwen25-3b-rbi-qa,
    author = {Vishva007},
    title = {Qwen2.5-3B-Instruct Fine-tuned for RBI Regulations Q&A},
    year = {2025},
    publisher = {HuggingFace},
    howpublished = {\url{https://huggingface.co/Vishva007/Qwen2.5-3B-Instruct-RBI-QA}},
  }

License

Apache 2.0 (inherits from base model)


Author: @Vishva007 | Updated: November 2025

Downloads last month
31
Safetensors
Model size
3B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Vishva007/Qwen2.5-3B-Instruct-RBI-QA

Base model

Qwen/Qwen2.5-3B
Adapter
(607)
this model

Dataset used to train Vishva007/Qwen2.5-3B-Instruct-RBI-QA