File size: 286 Bytes
03fa3a5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
---
base_model: []
library_name: transformers
tags:
- mergekit
- merge

---
# lobotollama6.5b 26/32 layer prune




This model was merged using the passthrough merge method.

### Models Merged

The following models were included in the merge:
* /home/meow/hf/Meta-Llama-3-8B-Instruct