OGAI-R1: Oil & Gas AI Model for Engineering & Technical Knowledge
OGAI-R1 is a fine-tuned version of TinyR1-32B, designed specifically for oil and gas engineering applications. It is optimized for engineering calculations, wellbore stability analysis, reservoir management, and document-based retrieval-augmented generation (RAG).
The model has been trained using GainEnergy's GPT-4o Oil & Gas Training Set, incorporating expert knowledge, technical formulas, and structured query-response interactions.
π Why Use OGAI-R1?
- π Fine-tuned for oil & gas engineering tasks (drilling, production, reservoir, and refining).
- π‘ Optimized for RAG β Enhanced document understanding and retrieval.
- π Long-Context Retention β Handles up to 32K tokens for complex engineering workflows.
- β‘ LoRA Fine-Tuning on TinyR1-32B β Enables efficient inference and quick knowledge retrieval.
π How to Use OGAI-R1
1οΈβ£ Install Required Dependencies
pip install torch transformers accelerate bitsandbytes
2οΈβ£ Load the Model
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "GainEnergy/OGAI-R1"
# Load tokenizer
tokenizer = AutoTokenizer.from_pretrained(model_id)
# Load model
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
# Run inference
prompt = "Explain the principles of reservoir simulation in petroleum engineering."
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
π¦ Model Variants
Model Name | Base Model | Precision | Context Window | Use Case |
---|---|---|---|---|
OGAI-R1 | TinyR1-32B | FP16 | 32K tokens | Engineering Calculations & RAG |
OGAI-8x7B | Mixtral-8x7B | 4-bit | 32K tokens | Oil & Gas AI Assistant |
OGAI-Reasoner | DeepSeek-R1 | FP16 | 128K tokens | Logical Reasoning & AI Simulation |
π Key Capabilities
β
Engineering Calculations β Computes reservoir volumes, wellbore stability, mud weight, casing depth, and more.
β
Technical Document Understanding β Trained on oil and gas technical literature, drilling reports, and engineering manuals.
β
Retrieval-Augmented Generation (RAG) β Enhances AI-driven document retrieval for faster decision-making.
β
High-Context Retention (32K tokens) β Supports long technical reports, operational workflows, and AI-driven engineering analysis.
π Use Cases
- Wellbore Stability & Drilling Optimization
- Hydraulics & Fluid Flow Simulations
- Reservoir Engineering & Petrophysics Analysis
- AI-Powered Document Retrieval & RAG Workflows
- Technical Compliance & Regulatory Document Processing
π‘ Deployment Options
Platform | Compatible? | Recommended Setup |
---|---|---|
Hugging Face Inference API | β Yes | Deploy via hf.co/GainEnergy/OGAI-R1 |
RunPod.io (Serverless GPU) | β Yes | A100-40GB or RTX 4090 |
AWS EC2 (G5 Instances) | β Yes | ml.g5.2xlarge (8 vCPUs, 32GB RAM) |
Local GPU (Consumer Hardware) | β Yes | Requires β₯16GB VRAM (RTX 3090, 4090) |
β οΈ Limitations
π§ Optimized for Oil & Gas Engineering β Not designed for general-purpose AI tasks.
π§ Requires domain-specific expertise β Outputs should be validated by industry experts.
π§ Computational requirements β Running the full TinyR1-32B model requires high-end GPUs.
π Resources
- GainEnergy AI Platform β Explore AI-powered drilling automation.
- Hugging Face Model Hub β Download & deploy the model.
π Citing OGAI-R1
@article{ogai-r1-2025,
title={OGAI-R1: An AI Model for Oil & Gas Engineering Optimization},
author={GainEnergy AI Team},
year={2025},
publisher={Hugging Face Models}
}
Model tree for GainEnergy/OGAI-r1
Base model
deepseek-ai/DeepSeek-R1-Distill-Qwen-32BEvaluation results
- Engineering Calculations Accuracy on GainEnergy GPT-4o Oil & Gas Training Setself-reported94.300
- Technical Document Retrieval Precision on GainEnergy GPT-4o Oil & Gas Training Setself-reported90.500
- Context Retention on GainEnergy GPT-4o Oil & Gas Training Setself-reportedHigh