nisten's picture
Update README.md
16a02eb verified
metadata
base_model:
  - meta-llama/Meta-Llama-3.1-405B
library_name: transformers
tags:
  - mergekit
  - merge

lobotollama-368b prune Meta-Llama-3.1-405B-Base.

This is a negative-merge of pre-trained language models created using mergekit.

Just so you meow, this did not turn out all that great in the perplexity benchmarks. Needs healing, you'll probably need 32xh100 to do a full finetune.

Model was designed to fin in a M2 mac-studio 192gb in 4bit.

perplexity: 167.37 seconds per pass - ETA 33.47 minutes - meta-405b-base - q8_0 - newest base was identical in bf16 and q8_0
[1]1.3927,[2]1.6952,[3]1.5905,[4]1.4674,[5]1.3652,[6]1.3054,[7]1.2885,[8]1.2673,[9]1.2397,[10]1.2179,[11]1.2149,[12]1.2162,
Final estimate: PPL = 1.2162 +/- 0.02128

perplexity: 2197.87 seconds per pass - ETA 1 hours 49.88 minutes -- llama 405b - instruct - old BF16 -8head
[1]2.1037,[2]2.4201,[3]2.0992,[4]1.8446,[5]1.6823,[6]1.5948,[7]1.5575,[8]1.5121,[9]1.4750,[10]1.4570,[11]1.4567,[12]1.4666,
Final estimate: PPL = 1.4666 +/- 0.03184

./llama-perplexity -m /scratch-10/lobotollama-q8_0.gguf -f wiki.test.raw -t 96  --chunks 12 -b 1024
perplexity: 331.47 seconds per pass - ETA 33.13 minutes
[1]2.6744,[2]3.4041,[3]2.9683,[4]2.8669,[5]2.7924,[6]2.7590,[7]2.8274,[8]2.8306,[9]2.7943,[10]2.7910,[11]2.8164,[12]2.9396,
Final estimate: PPL = 2.9396 +/- 0.09497

Merge Details

Merge Method

This model was merged using the passthrough merge method.

Models Merged

The following models were included in the merge:

  • /Meta-Llama-3.1-405B

Configuration

The following YAML configuration was used to produce this model:

dtype: bfloat16
merge_method: passthrough
slices:
- sources:
  - layer_range: [0, 29]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [30, 35]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [36, 40]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [41, 45]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [46, 49]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [50, 54]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [55, 59]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [60, 64]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [65, 69]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [70, 74]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [75, 79]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [80, 84]
    model: /Meta-Llama-3.1-405B
- sources:
  - layer_range: [85, 126]
    model: /Meta-Llama-3.1-405B