You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
/bamba-674f1388b9bbc98b413c7bab"> Bamba on Hugging Face</a>  | <ahref="https://huggingface.co/blog/bamba"> Bamba Blog</a> 
10
11
<be>
11
12
12
13
13
-
<!--Bamba is a repository for training and using [Bamba](https://huggingface.co/ibm-fms/Avengers-Mamba2-9B) models, which are derived from [Mamba](https://github.com/state-spaces/mamba) models.-->
14
+
<!--Bamba is a repository for training and using [Bamba](https://huggingface.co/ibm-ai-platform
15
+
/Avengers-Mamba2-9B) models, which are derived from [Mamba](https://github.com/state-spaces/mamba) models.-->
14
16
15
17
Bamba-9B is a decoder-only language model based on the [Mamba-2](https://github.com/state-spaces/mamba) architecture and is designed to handle a wide range of text generation tasks. It is trained from scratch using a two-stage training approach. In the first stage, the model is trained on 2 trillion tokens from the Dolma v1.7 dataset. In the second stage, it undergoes additional training on 200 billion tokens, leveraging a carefully curated blend of high-quality data to further refine its performance and enhance output quality.
You can find links to our model checkpoints here: [Bamba Models](https://huggingface.co/collections/ibm-fms/bamba-674f1388b9bbc98b413c7bab)
49
+
You can find links to our model checkpoints here: [Bamba Models](https://huggingface.co/collections/ibm-ai-platform
50
+
/bamba-674f1388b9bbc98b413c7bab)
48
51
49
52
## Inference
50
53
51
54
You can use the following command to perform text generation using one of our checkpoints provided above:
52
55
53
56
```python
54
-
python text_generation.py --model_path ibm-fms/Bamba-9B--tokenizer_path ibm-fms/Bamba-9B--prompt "The largest living mammal on Earth is "--max_new_tokens 128
/Bamba-9B--prompt "The largest living mammal on Earth is "--max_new_tokens 128
55
60
```
56
61
57
62
## Training
@@ -247,7 +252,8 @@ make -j
247
252
248
253
### Conversion to GGUF
249
254
250
-
You can use a pre-converted GGUF file from Huggingface (e.g. [bamba-9b.gguf](https://huggingface.co/ibm-fms/Bamba-9B/blob/main/bamba-9b.gguf)). If one doesn't exist, you can use the [convert_hf_to_gguf.py](https://github.com/gabe-l-hart/llama.cpp/blob/BambaArchitecture/convert_hf_to_gguf.py) script from Gabe's fork to perform the conversion manually.
255
+
You can use a pre-converted GGUF file from Huggingface (e.g. [bamba-9b.gguf](https://huggingface.co/ibm-ai-platform
256
+
/Bamba-9B/blob/main/bamba-9b.gguf)). If one doesn't exist, you can use the [convert_hf_to_gguf.py](https://github.com/gabe-l-hart/llama.cpp/blob/BambaArchitecture/convert_hf_to_gguf.py) script from Gabe's fork to perform the conversion manually.
Copy file name to clipboardExpand all lines: blog/bamba31T.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,7 +3,7 @@
3
3
During Christmas of 2024, IBM, Princeton, CMU, and UIUC [released](https://huggingface.co/blog/bamba), Bamba v1, a performant Mamba2 based pretrained model with full data lineage trained to 2T tokens. Since then, we have been busy cooking an update with new datasets. Today, we are excited to release Bamba v2, trained for an additional 1T tokens that significantly improves on Bamba v1. The L1 and L2 leaderboard scores outperform Llama 3.1 8B, which was trained with nearly 5x the amount of data. All of this with the inference speedup that we get from Mamba2 based architecture, which with the latest vLLM is 2-2.5x faster than similar sized transformer models.
4
4
5
5
## Artifacts 📦
6
-
1.[Hugging Face Bamba collection](https://huggingface.co/collections/ibm-fms/bamba-674f1388b9bbc98b413c7bab)
6
+
1.[Hugging Face Bamba collection](https://huggingface.co/collections/ibm-ai-platform/bamba-674f1388b9bbc98b413c7bab)
7
7
2.[GitHub repo with inference, training, and tuning scripts](https://github.com/foundation-model-stack/bamba)
0 commit comments