@@ -46,7 +46,7 @@ Enabled with [bitsandbytes](https://github.com/TimDettmers/bitsandbytes). Check
46
46
Uses the normalized float 4 (nf4) data type. This is recommended over "fp4" based on the paper's experimental results and theoretical analysis.
47
47
48
48
``` bash
49
- pip install scipy bitsandbytes # scipy is required until https://github.com/TimDettmers/bitsandbytes/pull/525 is released
49
+ pip install bitsandbytes
50
50
51
51
litgpt generate base --quantize bnb.nf4 --checkpoint_dir checkpoints/tiiuae/falcon-7b --precision bf16-true --max_new_tokens 256
52
52
...
@@ -62,7 +62,7 @@ Enabled with [bitsandbytes](https://github.com/TimDettmers/bitsandbytes). Check
62
62
In average, this amounts to about 0.37 bits per parameter (approximately 3 GB for a 65B model).
63
63
64
64
``` bash
65
- pip install scipy bitsandbytes # scipy is required until https://github.com/TimDettmers/bitsandbytes/pull/525 is released
65
+ pip install bitsandbytes
66
66
67
67
litgpt generate base --quantize bnb.nf4-dq --checkpoint_dir checkpoints/tiiuae/falcon-7b --precision bf16-true --max_new_tokens 256
68
68
...
@@ -77,7 +77,7 @@ Enabled with [bitsandbytes](https://github.com/TimDettmers/bitsandbytes). Check
77
77
Uses pure FP4 quantization.
78
78
79
79
``` bash
80
- pip install scipy bitsandbytes # scipy is required until https://github.com/TimDettmers/bitsandbytes/pull/525 is released
80
+ pip install bitsandbytes
81
81
82
82
litgpt generate base --quantize bnb.fp4 --checkpoint_dir checkpoints/tiiuae/falcon-7b --precision bf16-true --max_new_tokens 256
83
83
...
@@ -93,7 +93,7 @@ Enabled with [bitsandbytes](https://github.com/TimDettmers/bitsandbytes). Check
93
93
In average, this amounts to about 0.37 bits per parameter (approximately 3 GB for a 65B model).
94
94
95
95
``` bash
96
- pip install scipy bitsandbytes # scipy is required until https://github.com/TimDettmers/bitsandbytes/pull/525 is released
96
+ pip install bitsandbytes
97
97
98
98
litgpt generate base --quantize bnb.fp4-dq --checkpoint_dir checkpoints/tiiuae/falcon-7b --precision bf16-true --max_new_tokens 256
99
99
...
@@ -106,7 +106,7 @@ Memory used: 5.38 GB
106
106
Enabled with [ bitsandbytes] ( https://github.com/TimDettmers/bitsandbytes ) . Check out the [ paper] ( https://arxiv.org/abs/2110.02861 ) to learn more about how it works.
107
107
108
108
``` bash
109
- pip install scipy bitsandbytes # scipy is required until https://github.com/TimDettmers/bitsandbytes/pull/525 is released
109
+ pip install bitsandbytes
110
110
111
111
litgpt generate base --quantize bnb.int8 --checkpoint_dir checkpoints/tiiuae/falcon-7b --precision 16-true --max_new_tokens 256
112
112
...
0 commit comments