|
--- |
|
license: mit |
|
datasets: |
|
- Meforgers/Aixr-Thinkable-V1 |
|
language: |
|
- tr |
|
- en |
|
base_model: |
|
- meta-llama/Llama-3.1-8B |
|
tags: |
|
- code |
|
- medical |
|
- math |
|
- turkish |
|
- türkçe |
|
- coding |
|
- yazılım |
|
- programlama |
|
- thinkable |
|
- düşünebilen |
|
- düşünen |
|
new_version: Aixr/Aixr |
|
--- |
|
# LLama-3.1-Thinkable: Bilingual AI Expert in Mathematics and Programming |
|
|
|
LLama-3.1-Thinkable is a fine-tuned version of LLama 3.1, specifically designed to excel in **bilingual (Turkish and English)** communication, advanced **mathematics**, and **programming** tasks. This model combines enhanced reasoning capabilities with strong multilingual proficiency, offering a cutting-edge solution for users in diverse fields. |
|
|
|
--- |
|
|
|
## 🚀 Features |
|
|
|
1. **Bilingual Expertise** |
|
- Fluent in both **Turkish** and **English**. |
|
- Designed to seamlessly understand and respond in either language. |
|
- Ideal for users who switch between these languages or require multilingual solutions. |
|
|
|
2. **Mathematics Mastery** |
|
- Excels in solving advanced mathematical problems, including algebra, calculus, and statistics. |
|
- Provides step-by-step explanations for better understanding. |
|
|
|
3. **Programming Proficiency** |
|
- Supports a wide range of programming languages, including **Python**, **JavaScript**, **C++**, and more. |
|
- Assists with debugging, algorithm design, and code optimization. |
|
- Generates clear and efficient code snippets for complex tasks. |
|
|
|
4. **Thinkable AI: Enhanced Reasoning** |
|
- Fine-tuned for improved logical and critical thinking. |
|
- Capable of breaking down complex concepts into understandable insights. |
|
|
|
--- |
|
|
|
## 🔧 Technical Details |
|
|
|
- **Base Model:** LLama 3.1 |
|
- **Fine-tuning Dataset:** |
|
- High-quality bilingual datasets (Turkish-English). |
|
- Specialized datasets for mathematics and programming tasks. |
|
- **Parameter Count:** 5.25B & 8B |
|
--- |
|
|
|
## 📚 Use Cases |
|
|
|
- **Education:** |
|
- Learn programming and advanced mathematics with detailed explanations. |
|
- Solve bilingual academic tasks in Turkish and English. |
|
|
|
- **Development:** |
|
- Generate production-ready code. |
|
- Debug complex applications and find optimized solutions. |
|
|
|
- **AI Research:** |
|
- Experiment with a high-performance bilingual model in NLP tasks. |
|
|
|
--- |
|
|
|
## 🛠️ How to Use |
|
|
|
Here’s how you can get started with LLama-3.1-Thinkable: |
|
|
|
### Installation |
|
```bash |
|
pip install transformers |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
model_name = "aixr/llama-3.1-thinkable" |
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
model = AutoModelForCausalLM.from_pretrained(model_name) |
|
|
|
# Generate a response |
|
inputs = tokenizer("Explain recursion in programming:", return_tensors="pt") |
|
outputs = model.generate(inputs["input_ids"], max_length=200) |
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
|
``` |