| from typing import Dict, Optional | |
| from torch import Tensor, nn | |
| from transformers import PreTrainedModel | |
| from .configuration_dptdepth import DPTDepthConfig | |
| from .models import DPTDepthModel as DPTDepth | |
| class DPTDepthModel(PreTrainedModel): | |
| """ | |
| The line that sets the config_class is not mandatory, | |
| unless you want to register your model with the auto classes | |
| """ | |
| config_class = DPTDepthConfig | |
| def __init__(self, config: DPTDepthConfig): | |
| super().__init__(config) | |
| self.model = DPTDepth() | |
| self.loss = nn.L1Loss() | |
| """ | |
| You can have your model return anything you want, | |
| but returning a dictionary with the loss included when labels are passed, | |
| will make your model directly usable inside the Trainer class. | |
| Using another output format is fine as long as you are planning on | |
| using your own training loop or another library for training. | |
| """ | |
| def forward(self, rgbs: Tensor, gts: Optional[Tensor] = None) -> Dict[str, Tensor]: | |
| logits = self.model(rgbs) | |
| if gts is not None: | |
| loss = self.loss(logits, gts) | |
| return {"loss": loss, "logits": logits} | |
| return {"logits": logits} | |