Commit 
							
							·
						
						8fae5ad
	
1
								Parent(s):
							
							f912ca4
								
support transformers 4.34.0 (#1)
Browse files- support transformers 4.34.0 (4962895a2c3847e0f784be3f3cb775e9688d53ac)
    	
        triton_models/tokenizer/tokenization_internlm.py
    CHANGED
    
    | @@ -65,6 +65,13 @@ class InternLMTokenizer(PreTrainedTokenizer): | |
| 65 | 
             
                    **kwargs,
         | 
| 66 | 
             
                ):
         | 
| 67 | 
             
                    self.sp_model_kwargs = {} if sp_model_kwargs is None else sp_model_kwargs
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 68 | 
             
                    super().__init__(
         | 
| 69 | 
             
                        bos_token=bos_token,
         | 
| 70 | 
             
                        eos_token=eos_token,
         | 
| @@ -73,13 +80,6 @@ class InternLMTokenizer(PreTrainedTokenizer): | |
| 73 | 
             
                        clean_up_tokenization_spaces=clean_up_tokenization_spaces,
         | 
| 74 | 
             
                        **kwargs,
         | 
| 75 | 
             
                    )
         | 
| 76 | 
            -
                    self.vocab_file = vocab_file
         | 
| 77 | 
            -
                    self.add_bos_token = add_bos_token
         | 
| 78 | 
            -
                    self.add_eos_token = add_eos_token
         | 
| 79 | 
            -
                    self.decode_with_prefix_space = decode_with_prefix_space
         | 
| 80 | 
            -
                    self.sp_model = spm.SentencePieceProcessor(**self.sp_model_kwargs)
         | 
| 81 | 
            -
                    self.sp_model.Load(vocab_file)
         | 
| 82 | 
            -
                    self._no_prefix_space_tokens = None
         | 
| 83 |  | 
| 84 | 
             
                    """ Initialisation"""
         | 
| 85 |  | 
|  | |
| 65 | 
             
                    **kwargs,
         | 
| 66 | 
             
                ):
         | 
| 67 | 
             
                    self.sp_model_kwargs = {} if sp_model_kwargs is None else sp_model_kwargs
         | 
| 68 | 
            +
                    self.vocab_file = vocab_file
         | 
| 69 | 
            +
                    self.add_bos_token = add_bos_token
         | 
| 70 | 
            +
                    self.add_eos_token = add_eos_token
         | 
| 71 | 
            +
                    self.decode_with_prefix_space = decode_with_prefix_space
         | 
| 72 | 
            +
                    self.sp_model = spm.SentencePieceProcessor(**self.sp_model_kwargs)
         | 
| 73 | 
            +
                    self.sp_model.Load(vocab_file)
         | 
| 74 | 
            +
                    self._no_prefix_space_tokens = None
         | 
| 75 | 
             
                    super().__init__(
         | 
| 76 | 
             
                        bos_token=bos_token,
         | 
| 77 | 
             
                        eos_token=eos_token,
         | 
|  | |
| 80 | 
             
                        clean_up_tokenization_spaces=clean_up_tokenization_spaces,
         | 
| 81 | 
             
                        **kwargs,
         | 
| 82 | 
             
                    )
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 83 |  | 
| 84 | 
             
                    """ Initialisation"""
         | 
| 85 |  | 
