How can I disable streaming during inference? Also, can Deepseek OCR handle multiple images at the same time?

#91
by Tizzzzy - opened

How can I disable streaming during inference? Also, can Deepseek OCR handle multiple images at the same time?

Here is my current code:

tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
model = AutoModel.from_pretrained(
        model_name, 
        _attn_implementation='flash_attention_2', 
        trust_remote_code=True, 
        use_safetensors=True,
    )

res = model.infer(
            tokenizer, 
            prompt=prompt, 
            image_file=image_file, 
            output_path=output_path, 
            base_size=1024, 
            image_size=640, 
            crop_mode=True,   # Gundam mode (Dynamic Tiling) - Good for docs
            save_results=True, 
            test_compress=True,
            eval_mode=True
        )

Sign up or log in to comment