Update inference-cache-config/llama.json
Browse files
inference-cache-config/llama.json
CHANGED
|
@@ -42,5 +42,21 @@
|
|
| 42 |
"num_cores": 8,
|
| 43 |
"auto_cast_type": "bf16"
|
| 44 |
}
|
| 45 |
-
]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
}
|
|
|
|
| 42 |
"num_cores": 8,
|
| 43 |
"auto_cast_type": "bf16"
|
| 44 |
}
|
| 45 |
+
],
|
| 46 |
+
"meta-llama/Llama-2-7b-hf": [
|
| 47 |
+
{
|
| 48 |
+
"batch_size": 1,
|
| 49 |
+
"sequence_length": 2048,
|
| 50 |
+
"num_cores": 2,
|
| 51 |
+
"auto_cast_type": "bf16"
|
| 52 |
+
}
|
| 53 |
+
],
|
| 54 |
+
"meta-llama/Llama-2-13b-hf": [
|
| 55 |
+
{
|
| 56 |
+
"batch_size": 1,
|
| 57 |
+
"sequence_length": 2048,
|
| 58 |
+
"num_cores": 8,
|
| 59 |
+
"auto_cast_type": "bf16"
|
| 60 |
+
}
|
| 61 |
+
],
|
| 62 |
}
|