PaperToRead Replacing softmax with ReLU in Vision Transformers Paper • 2309.08586 • Published Sep 15, 2023 • 18 AstroLLaMA: Towards Specialized Foundation Models in Astronomy Paper • 2309.06126 • Published Sep 12, 2023 • 17
Replacing softmax with ReLU in Vision Transformers Paper • 2309.08586 • Published Sep 15, 2023 • 18
AstroLLaMA: Towards Specialized Foundation Models in Astronomy Paper • 2309.06126 • Published Sep 12, 2023 • 17
1bitLLM The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 627
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 627
PaperToRead Replacing softmax with ReLU in Vision Transformers Paper • 2309.08586 • Published Sep 15, 2023 • 18 AstroLLaMA: Towards Specialized Foundation Models in Astronomy Paper • 2309.06126 • Published Sep 12, 2023 • 17
Replacing softmax with ReLU in Vision Transformers Paper • 2309.08586 • Published Sep 15, 2023 • 18
AstroLLaMA: Towards Specialized Foundation Models in Astronomy Paper • 2309.06126 • Published Sep 12, 2023 • 17
1bitLLM The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 627
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 627