MonicaDasari commited on
Commit
3b1dd8a
·
verified ·
1 Parent(s): 0fb973d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +99 -11
README.md CHANGED
@@ -1,14 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
- title: FinalProject
3
- emoji: 🔥
4
- colorFrom: gray
5
- colorTo: yellow
6
- sdk: gradio
7
- sdk_version: 5.8.0
8
- app_file: app.py
9
- pinned: false
10
- license: mit
11
- short_description: Translation
 
 
12
  ---
13
 
14
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
1
+ Here’s a **README** template for your project, designed to highlight the models used, evaluation methodology, and key results. You can adapt this for Hugging Face or any similar platform.
2
+
3
+ ---
4
+
5
+ # **English-to-Japanese Translation Project**
6
+
7
+ ## **Overview**
8
+ This project focuses on building a robust system for English-to-Japanese translation using state-of-the-art multilingual models. Two models were used: **mT5** as the primary model and **mBART** as the secondary model. Together, they ensure high-quality translations and versatility in multilingual tasks.
9
+
10
+ ---
11
+
12
+ ## **Models Used**
13
+
14
+ ### **1. mT5 (Primary Model)**
15
+ - **Reason for Selection**:
16
+ - mT5 is highly versatile and trained on a broad multilingual dataset, making it suitable for translation and other tasks like summarization or answering questions.
17
+ - It performs well without extensive fine-tuning, saving computational resources.
18
+
19
+ - **Strengths**:
20
+ - Handles translation naturally with minimal training.
21
+ - Can perform additional tasks beyond translation.
22
+
23
+ - **Limitations**:
24
+ - Sometimes lacks precision in detailed translations.
25
+
26
+ ---
27
+
28
+ ### **2. mBART (Secondary Model)**
29
+ - **Reason for Selection**:
30
+ - mBART specializes in multilingual translation tasks and provides highly accurate translations when fine-tuned.
31
+
32
+ - **Strengths**:
33
+ - Optimized for translation accuracy, especially for long sentences and contextual consistency.
34
+ - Handles grammatical and contextual errors well.
35
+
36
+ - **Limitations**:
37
+ - Less flexible for tasks like summarization or question answering compared to mT5.
38
+
39
+ ---
40
+
41
+ ## **Evaluation Strategy**
42
+
43
+ To evaluate model performance, the following metrics were used:
44
+
45
+ 1. **BLEU Score**:
46
+ - Measures how close the model's output is to the correct translation.
47
+ - Chosen because it is a standard for evaluating translation accuracy.
48
+
49
+ 2. **Training Loss**:
50
+ - Tracks how well the model is learning during training.
51
+ - A lower loss shows better learning and fewer errors.
52
+
53
+ 3. **Perplexity**:
54
+ - Checks the confidence of the model’s predictions.
55
+ - Lower perplexity means fewer mistakes and more fluent translations.
56
+
57
+ ---
58
+
59
+ ## **Steps Taken**
60
+ 1. Fine-tuned both models using a dataset of English-Japanese text pairs to improve translation accuracy.
61
+ 2. Tested the models on unseen data to measure their real-world performance.
62
+ 3. Applied optimizations like **4-bit quantization** to reduce memory usage and make the models faster during evaluation.
63
+
64
+ ---
65
+
66
+ ## **Results**
67
+ - **mT5**:
68
+ - Performed well in handling translations and additional tasks like summarization and answering questions.
69
+ - Showed versatility but sometimes lacked detailed accuracy for translations.
70
+
71
+ - **mBART**:
72
+ - Delivered precise and contextually accurate translations, especially for longer sentences.
73
+ - Required fine-tuning but outperformed mT5 in translation-focused tasks.
74
+
75
+ - **Overall Conclusion**:
76
+ mT5 is a flexible model for multilingual tasks, while mBART ensures high-quality translations. Together, they balance versatility and accuracy, making them ideal for English-to-Japanese translations.
77
+
78
+ ---
79
+
80
+ ## **How to Use**
81
+ 1. Load the models from Hugging Face:
82
+ - [mT5 Model on Hugging Face](https://huggingface.co/google/mt5-small)
83
+ - [mBART Model on Hugging Face](https://huggingface.co/facebook/mbart-large-50)
84
+
85
+ 2. Fine-tune the models for your dataset using English-Japanese text pairs.
86
+ 3. Evaluate performance using BLEU Score, training loss, and perplexity.
87
+
88
  ---
89
+
90
+ ## **Future Work**
91
+ - Expand the dataset for better fine-tuning.
92
+ - Explore task-specific fine-tuning for mT5 to improve its translation accuracy.
93
+ - Optimize the models further for deployment in resource-constrained environments.
94
+
95
+ ---
96
+
97
+ ## **References**
98
+ - [mT5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer](https://arxiv.org/abs/2010.11934)
99
+ - [mBART: Multilingual Denoising Pretraining for Neural Machine Translation](https://arxiv.org/abs/2001.08210)
100
+
101
  ---
102