raxtemur commited on
Commit
1d7bf0e
·
verified ·
1 Parent(s): 0e4299d

Initial upload (weights + code + README)

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -12,7 +12,7 @@ library_name: transformers
12
  pipeline_tag: text-generation
13
  ---
14
 
15
- # SONAR-LLM (1.3M) -- Text summarization checkopoint
16
 
17
  We present SONAR-LLM, a decoder-only transformer that "thinks" in the same continuous SONAR embedding space, yet is supervised through token-level cross-entropy propagated via the frozen SONAR decoder. This hybrid objective retains the semantic abstraction of LCM while eliminating its diffusion sampler and restoring a likelihood-based training signal. Across model sizes from 39M to 1.3B parameters, SONAR-LLM attains competitive generation quality.
18
 
 
12
  pipeline_tag: text-generation
13
  ---
14
 
15
+ # SONAR-LLM (1.3B) -- Text summarization checkopoint
16
 
17
  We present SONAR-LLM, a decoder-only transformer that "thinks" in the same continuous SONAR embedding space, yet is supervised through token-level cross-entropy propagated via the frozen SONAR decoder. This hybrid objective retains the semantic abstraction of LCM while eliminating its diffusion sampler and restoring a likelihood-based training signal. Across model sizes from 39M to 1.3B parameters, SONAR-LLM attains competitive generation quality.
18