Falcon Mamba Playground
Generate chat responses using FalconMamba-7b model
This collection features the FalconMamba 7B base model, the instruction-tuned version, their 4-bit and GGUF variants, and the demo.
Generate chat responses using FalconMamba-7b model
Note FalconMamba technical report
Note First strong attention free model for general purpose usage, based on Mamba1 architecture
Note FalconMamba-7B fine-tuned on instruction data, for chat-like interaction with the model
Note Pre-decay stage checkpoint useful for continuous pretraining
Note FalconMamba-7B quantized in 4bit precision using `bitsandbytes` library for lighter memory requirements and smaller GPU hardwares
Note FalconMamba-7B-instruct quantized in 4bit precision using `bitsandbytes` library for lighter memory requirements and smaller GPU hardwares
Note Falcon Mamba 7b-instruct in GGUF format (compatible with llama.cpp) in BF16 format
Note Falcon Mamba 7b-instruct in GGUF format (compatible with llama.cpp) in F16 format
Note Falcon Mamba 7b-instruct in GGUF format (compatible with llama.cpp) in quantized Q8_0 format
Note Falcon Mamba 7b-instruct in GGUF format (compatible with llama.cpp) in quantized Q4_K_M format
Note Falcon Mamba 7b in GGUF format (compatible with llama.cpp) in BF16 format
Note Falcon Mamba 7b in GGUF format (compatible with llama.cpp) in quantized Q4_K_M format
Note Falcon Mamba 7b in GGUF format (compatible with llama.cpp) in F16 format
Note Falcon Mamba 7b in GGUF format (compatible with llama.cpp) in quantized Q8_0 format