LLM - MLX
					Collection
				
Text Generations Models in MLX format, hand picked by Nexa Team.
					• 
				6 items
				• 
				Updated
					
				•
					
					2
Run them directly with nexa-sdk installed In nexa-sdk CLI:
NexaAI/gpt-oss-20b-MLX-4bit
This is a 4-bit quantized version of the OpenAI GPT OSS 20B model, optimized for Apple Silicon using the MLX framework. The model was successfully converted from the original gpt_oss architecture to MLX format using the development version of mlx-lm.
Original model card: InferenceIllusionist/gpt-oss-20b-MLX-4bit
Base model
openai/gpt-oss-20b