Update README.md
Browse files
README.md
CHANGED
|
@@ -32,6 +32,13 @@ For example, if you prompt it with the "Two Sum" problem and the input `nums = [
|
|
| 32 |
|
| 33 |
**Finetuned from model:** `unsloth/Llama-3.2-1B-unsloth-bnb-4bit`
|
| 34 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
## Uses
|
| 36 |
|
| 37 |
### Direct Use
|
|
@@ -91,11 +98,4 @@ It is a specialized model designed only to replicate the input-output pairs from
|
|
| 91 |
|
| 92 |
The model was fine-tuned on a dataset by newfacade from here: https://huggingface.co/datasets/newfacade/LeetCodeDataset
|
| 93 |
|
| 94 |
-
I just used 5k samples from it and trained only for 1 epoch
|
| 95 |
-
|
| 96 |
-
# NOTE
|
| 97 |
-
|
| 98 |
-
I used only two target models ("q_proj", "v_proj") which focuses only on the attention blocks and kept rank value (r=8).
|
| 99 |
-
Why? I have neither money nor time to run the model.
|
| 100 |
-
|
| 101 |
-
If you like to waste your time on this, the notebook is available here: https://www.kaggle.com/code/yashasnadig/leetcode2output
|
|
|
|
| 32 |
|
| 33 |
**Finetuned from model:** `unsloth/Llama-3.2-1B-unsloth-bnb-4bit`
|
| 34 |
|
| 35 |
+
## NOTE
|
| 36 |
+
|
| 37 |
+
I used only two target models ("q_proj", "v_proj") which focuses only on the attention blocks and kept rank value (r=8).
|
| 38 |
+
Why? I have neither money nor time to run the model.
|
| 39 |
+
|
| 40 |
+
If you like to waste your time on this, the notebook is available here: https://www.kaggle.com/code/yashasnadig/leetcode2output
|
| 41 |
+
|
| 42 |
## Uses
|
| 43 |
|
| 44 |
### Direct Use
|
|
|
|
| 98 |
|
| 99 |
The model was fine-tuned on a dataset by newfacade from here: https://huggingface.co/datasets/newfacade/LeetCodeDataset
|
| 100 |
|
| 101 |
+
I just used 5k samples from it and trained only for 1 epoch.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|