Source code for lightwood.encoder.image.img_2_vec

from typing import List, Tuple, Iterable
import torch
from lightwood.encoder.image.helpers.img_to_vec import Img2Vec
from lightwood.encoder.base import BaseEncoder
from lightwood.helpers.log import log

    import torchvision.transforms as transforms
    from PIL import Image
except ModuleNotFoundError:"No torchvision/pillow detected, image encoder not supported")

[docs]class Img2VecEncoder(BaseEncoder): """ Generates encoded representations for images using a pre-trained deep neural network. Inputs must be str-based location of the data. Without user-specified details, all input images are rescaled to a standard size of 224x224, and normalized using the mean and standard deviation of the ImageNet dataset (as it was used to train the underlying NN). This encoder currently does not support a `decode()` call; models with an image output will not work. For more information about the neural network this encoder uses, refer to the `lightwood.encoder.image.helpers.img_to_vec.Img2Vec`. """ # noqa is_trainable_encoder: bool = True def __init__( self, stop_after: float = 3600, is_target: bool = False, scale: Tuple[int, int] = (224, 224), mean: List[float] = [0.485, 0.456, 0.406], std: List[float] = [0.229, 0.224, 0.225], device: str = '', ): """ :param stop_after: time budget, in seconds. :param is_target: Whether encoder represents target or not :param scale: Resize scale of image (x, y) :param mean: Mean of pixel values :param std: Standard deviation of pixel values :param device: Name of the device that get_device_from_name will attempt to use """ # noqa assert not is_target super().__init__(is_target) self.is_prepared = False self.scale = scale self.mean = mean self.std = std self._scaler = transforms.Resize(scale) self._normalize = transforms.Normalize(mean=self.mean, std=self.std) self._to_tensor = transforms.ToTensor() self._img_to_tensor = transforms.Compose([ self._scaler, self._to_tensor, self._normalize ]) self.stop_after = stop_after self.device_name = device # pil_logger = logging.getLogger('PIL') # noqa # pil_logger.setLevel(logging.ERROR) # noqa
[docs] def prepare(self, train_priming_data: Iterable[str], dev_priming_data: Iterable[str]): # @TODO: finetune here? depending on time aim """ Sets an `Img2Vec` object (model) and sets the expected size for encoded representations. """ if self.is_prepared: raise Exception('You can only call "prepare" once for a given encoder.') self.model = Img2Vec(device=self.device_name) self.device = self.model.device self.device_name = f'{self.device}' self.output_size = self.model.output_size self.is_prepared = True
[docs] def to(self, device, available_devices=1): """ Changes device of model to support CPU/GPU :param device: will move the model to this device. :param available_devices: all available devices as reported by lightwood. :return: same object but moved to the target device. """, available_devices=1) return self
[docs] def encode(self, images: List[str]) -> torch.Tensor: """ Creates encodings for a list of images; each image is referenced by a filepath or url. :param images: list of images, each image is a path to a file or a url. :return: a torch.floatTensor """ if not self.is_prepared: raise Exception('You need to call "prepare" before calling "encode" or "decode".') img_tensors = [self._img_to_tensor( ) for img_path in images] vec_arr = [] self.model.eval() with torch.no_grad(): for img_tensor in img_tensors: vec = self.model(img_tensor.unsqueeze(0), batch=False) vec_arr.append(vec) return torch.stack(vec_arr).to('cpu')
[docs] def decode(self, encoded_values_tensor: torch.Tensor): """ Currently not supported """ raise Exception('This encoder is not bi-directional')