Spaces:
Running
Running
| # ------------------------------------------------------------------------ | |
| # Modified from OFA (https://github.com/OFA-Sys/OFA) | |
| # Copyright 2022 The OFA-Sys Team. | |
| # All rights reserved. | |
| # This source code is licensed under the Apache 2.0 license | |
| # found in the LICENSE file in the root directory. | |
| # ------------------------------------------------------------------------ | |
| # Modifications Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved. | |
| # SPDX-License-Identifier: Apache-2.0 | |
| try: | |
| from collections.abc import Iterable | |
| except ImportError: | |
| from collections import Iterable | |
| import contextlib | |
| import itertools | |
| import logging | |
| import re | |
| import warnings | |
| from typing import Optional, Tuple | |
| import numpy as np | |
| import torch | |
| from fairseq.file_io import PathManager | |
| from fairseq import utils | |
| import os | |
| logger = logging.getLogger(__name__) | |
| def infer_language_pair(path): | |
| """Infer language pair from filename: <split>.<lang1>-<lang2>.(...).idx""" | |
| src, dst = None, None | |
| for filename in PathManager.ls(path): | |
| parts = filename.split(".") | |
| if len(parts) >= 3 and len(parts[1].split("-")) == 2: | |
| return parts[1].split("-") | |
| return src, dst | |
| def collate_tokens( | |
| values, | |
| pad_idx, | |
| eos_idx=None, | |
| left_pad=False, | |
| move_eos_to_beginning=False, | |
| pad_to_length=None, | |
| pad_to_multiple=1, | |
| pad_to_bsz=None, | |
| ): | |
| """Convert a list of 1d tensors into a padded 2d tensor.""" | |
| size = max(v.size(0) for v in values) | |
| size = size if pad_to_length is None else max(size, pad_to_length) | |
| if pad_to_multiple != 1 and size % pad_to_multiple != 0: | |
| size = int(((size - 0.1) // pad_to_multiple + 1) * pad_to_multiple) | |
| def copy_tensor(src, dst): | |
| assert dst.numel() == src.numel() | |
| if move_eos_to_beginning: | |
| if eos_idx is None: | |
| # if no eos_idx is specified, then use the last token in src | |
| dst[0] = src[-1] | |
| else: | |
| dst[0] = eos_idx | |
| dst[1:] = src[:-1] | |
| else: | |
| dst.copy_(src) | |
| if values[0].dim() == 1: | |
| res = values[0].new(len(values), size).fill_(pad_idx) | |
| elif values[0].dim() == 2: | |
| assert move_eos_to_beginning is False | |
| res = values[0].new(len(values), size, values[0].size(1)).fill_(pad_idx) | |
| else: | |
| raise NotImplementedError | |
| for i, v in enumerate(values): | |
| copy_tensor(v, res[i][size - len(v) :] if left_pad else res[i][: len(v)]) | |
| return res | |
| def load_indexed_dataset( | |
| path, dictionary=None, dataset_impl=None, combine=False, default="cached" | |
| ): | |
| """A helper function for loading indexed datasets. | |
| Args: | |
| path (str): path to indexed dataset (e.g., 'data-bin/train') | |
| dictionary (~fairseq.data.Dictionary): data dictionary | |
| dataset_impl (str, optional): which dataset implementation to use. If | |
| not provided, it will be inferred automatically. For legacy indexed | |
| data we use the 'cached' implementation by default. | |
| combine (bool, optional): automatically load and combine multiple | |
| datasets. For example, if *path* is 'data-bin/train', then we will | |
| combine 'data-bin/train', 'data-bin/train1', ... and return a | |
| single ConcatDataset instance. | |
| """ | |
| import fairseq.data.indexed_dataset as indexed_dataset | |
| from fairseq.data.concat_dataset import ConcatDataset | |
| datasets = [] | |
| for k in itertools.count(): | |
| path_k = path + (str(k) if k > 0 else "") | |
| try: | |
| path_k = indexed_dataset.get_indexed_dataset_to_local(path_k) | |
| except Exception as e: | |
| if "StorageException: [404] Path not found" in str(e): | |
| logger.warning(f"path_k: {e} not found") | |
| else: | |
| raise e | |
| dataset_impl_k = dataset_impl | |
| if dataset_impl_k is None: | |
| dataset_impl_k = indexed_dataset.infer_dataset_impl(path_k) | |
| dataset = indexed_dataset.make_dataset( | |
| path_k, | |
| impl=dataset_impl_k or default, | |
| fix_lua_indexing=True, | |
| dictionary=dictionary, | |
| ) | |
| if dataset is None: | |
| break | |
| logger.info("loaded {:,} examples from: {}".format(len(dataset), path_k)) | |
| datasets.append(dataset) | |
| if not combine: | |
| break | |
| if len(datasets) == 0: | |
| return None | |
| elif len(datasets) == 1: | |
| return datasets[0] | |
| else: | |
| return ConcatDataset(datasets) | |
| def numpy_seed(seed, *addl_seeds): | |
| """Context manager which seeds the NumPy PRNG with the specified seed and | |
| restores the state afterward""" | |
| if seed is None: | |
| yield | |
| return | |
| if len(addl_seeds) > 0: | |
| seed = int(hash((seed, *addl_seeds)) % 1e6) | |
| state = np.random.get_state() | |
| np.random.seed(seed) | |
| try: | |
| yield | |
| finally: | |
| np.random.set_state(state) | |
| def collect_filtered(function, iterable, filtered): | |
| """ | |
| Similar to :func:`filter` but collects filtered elements in ``filtered``. | |
| Args: | |
| function (callable): function that returns ``False`` for elements that | |
| should be filtered | |
| iterable (iterable): iterable to filter | |
| filtered (list): list to store filtered elements | |
| """ | |
| for el in iterable: | |
| if function(el): | |
| yield el | |
| else: | |
| filtered.append(el) | |
| def _filter_by_size_dynamic(indices, size_fn, max_positions, raise_exception=False): | |
| def compare_leq(a, b): | |
| return a <= b if not isinstance(a, tuple) else max(a) <= b | |
| def check_size(idx): | |
| if isinstance(max_positions, float) or isinstance(max_positions, int): | |
| return size_fn(idx) <= max_positions | |
| elif isinstance(max_positions, dict): | |
| idx_size = size_fn(idx) | |
| assert isinstance(idx_size, dict) | |
| intersect_keys = set(max_positions.keys()) & set(idx_size.keys()) | |
| return all( | |
| all( | |
| a is None or b is None or a <= b | |
| for a, b in zip(idx_size[key], max_positions[key]) | |
| ) | |
| for key in intersect_keys | |
| ) | |
| else: | |
| # For MultiCorpusSampledDataset, will generalize it later | |
| if not isinstance(size_fn(idx), Iterable): | |
| return all(size_fn(idx) <= b for b in max_positions) | |
| return all( | |
| a is None or b is None or a <= b | |
| for a, b in zip(size_fn(idx), max_positions) | |
| ) | |
| ignored = [] | |
| itr = collect_filtered(check_size, indices, ignored) | |
| indices = np.fromiter(itr, dtype=np.int64, count=-1) | |
| return indices, ignored | |
| def filter_by_size(indices, dataset, max_positions, raise_exception=False): | |
| """ | |
| [deprecated] Filter indices based on their size. | |
| Use `FairseqDataset::filter_indices_by_size` instead. | |
| Args: | |
| indices (List[int]): ordered list of dataset indices | |
| dataset (FairseqDataset): fairseq dataset instance | |
| max_positions (tuple): filter elements larger than this size. | |
| Comparisons are done component-wise. | |
| raise_exception (bool, optional): if ``True``, raise an exception if | |
| any elements are filtered (default: False). | |
| """ | |
| warnings.warn( | |
| "data_utils.filter_by_size is deprecated. " | |
| "Use `FairseqDataset::filter_indices_by_size` instead.", | |
| stacklevel=2, | |
| ) | |
| if isinstance(max_positions, float) or isinstance(max_positions, int): | |
| if hasattr(dataset, "sizes") and isinstance(dataset.sizes, np.ndarray): | |
| ignored = indices[dataset.sizes[indices] > max_positions].tolist() | |
| indices = indices[dataset.sizes[indices] <= max_positions] | |
| elif ( | |
| hasattr(dataset, "sizes") | |
| and isinstance(dataset.sizes, list) | |
| and len(dataset.sizes) == 1 | |
| ): | |
| ignored = indices[dataset.sizes[0][indices] > max_positions].tolist() | |
| indices = indices[dataset.sizes[0][indices] <= max_positions] | |
| else: | |
| indices, ignored = _filter_by_size_dynamic( | |
| indices, dataset.size, max_positions | |
| ) | |
| else: | |
| indices, ignored = _filter_by_size_dynamic(indices, dataset.size, max_positions) | |
| if len(ignored) > 0 and raise_exception: | |
| raise Exception( | |
| ( | |
| "Size of sample #{} is invalid (={}) since max_positions={}, " | |
| "skip this example with --skip-invalid-size-inputs-valid-test" | |
| ).format(ignored[0], dataset.size(ignored[0]), max_positions) | |
| ) | |
| if len(ignored) > 0: | |
| logger.warning( | |
| ( | |
| "{} samples have invalid sizes and will be skipped, " | |
| "max_positions={}, first few sample ids={}" | |
| ).format(len(ignored), max_positions, ignored[:10]) | |
| ) | |
| return indices | |
| def filter_paired_dataset_indices_by_size(src_sizes, tgt_sizes, indices, max_sizes): | |
| """Filter a list of sample indices. Remove those that are longer | |
| than specified in max_sizes. | |
| Args: | |
| indices (np.array): original array of sample indices | |
| max_sizes (int or list[int] or tuple[int]): max sample size, | |
| can be defined separately for src and tgt (then list or tuple) | |
| Returns: | |
| np.array: filtered sample array | |
| list: list of removed indices | |
| """ | |
| if max_sizes is None: | |
| return indices, [] | |
| if type(max_sizes) in (int, float): | |
| max_src_size, max_tgt_size = max_sizes, max_sizes | |
| else: | |
| max_src_size, max_tgt_size = max_sizes | |
| if tgt_sizes is None: | |
| ignored = indices[src_sizes[indices] > max_src_size] | |
| else: | |
| ignored = indices[ | |
| (src_sizes[indices] > max_src_size) | (tgt_sizes[indices] > max_tgt_size) | |
| ] | |
| if len(ignored) > 0: | |
| if tgt_sizes is None: | |
| indices = indices[src_sizes[indices] <= max_src_size] | |
| else: | |
| indices = indices[ | |
| (src_sizes[indices] <= max_src_size) | |
| & (tgt_sizes[indices] <= max_tgt_size) | |
| ] | |
| return indices, ignored.tolist() | |
| def batch_by_size( | |
| indices, | |
| num_tokens_fn, | |
| num_tokens_vec=None, | |
| max_tokens=None, | |
| max_sentences=None, | |
| required_batch_size_multiple=1, | |
| fixed_shapes=None, | |
| ): | |
| """ | |
| Yield mini-batches of indices bucketed by size. Batches may contain | |
| sequences of different lengths. | |
| Args: | |
| indices (List[int]): ordered list of dataset indices | |
| num_tokens_fn (callable): function that returns the number of tokens at | |
| a given index | |
| num_tokens_vec (List[int], optional): precomputed vector of the number | |
| of tokens for each index in indices (to enable faster batch generation) | |
| max_tokens (int, optional): max number of tokens in each batch | |
| (default: None). | |
| max_sentences (int, optional): max number of sentences in each | |
| batch (default: None). | |
| required_batch_size_multiple (int, optional): require batch size to | |
| be less than N or a multiple of N (default: 1). | |
| fixed_shapes (List[Tuple[int, int]], optional): if given, batches will | |
| only be created with the given shapes. *max_sentences* and | |
| *required_batch_size_multiple* will be ignored (default: None). | |
| """ | |
| try: | |
| from fairseq.data.data_utils_fast import ( | |
| batch_by_size_fn, | |
| batch_by_size_vec, | |
| batch_fixed_shapes_fast, | |
| ) | |
| except ImportError: | |
| raise ImportError( | |
| "Please build Cython components with: " | |
| "`python setup.py build_ext --inplace`" | |
| ) | |
| except ValueError: | |
| raise ValueError( | |
| "Please build (or rebuild) Cython components with `python setup.py build_ext --inplace`." | |
| ) | |
| # added int() to avoid TypeError: an integer is required | |
| max_tokens = ( | |
| int(max_tokens) if max_tokens is not None else -1 | |
| ) | |
| max_sentences = max_sentences if max_sentences is not None else -1 | |
| bsz_mult = required_batch_size_multiple | |
| if not isinstance(indices, np.ndarray): | |
| indices = np.fromiter(indices, dtype=np.int64, count=-1) | |
| if num_tokens_vec is not None and not isinstance(num_tokens_vec, np.ndarray): | |
| num_tokens_vec = np.fromiter(num_tokens_vec, dtype=np.int64, count=-1) | |
| if fixed_shapes is None: | |
| if num_tokens_vec is None: | |
| return batch_by_size_fn( | |
| indices, | |
| num_tokens_fn, | |
| max_tokens, | |
| max_sentences, | |
| bsz_mult, | |
| ) | |
| else: | |
| return batch_by_size_vec( | |
| indices, | |
| num_tokens_vec, | |
| max_tokens, | |
| max_sentences, | |
| bsz_mult, | |
| ) | |
| else: | |
| fixed_shapes = np.array(fixed_shapes, dtype=np.int64) | |
| sort_order = np.lexsort( | |
| [ | |
| fixed_shapes[:, 1].argsort(), # length | |
| fixed_shapes[:, 0].argsort(), # bsz | |
| ] | |
| ) | |
| fixed_shapes_sorted = fixed_shapes[sort_order] | |
| return batch_fixed_shapes_fast(indices, num_tokens_fn, fixed_shapes_sorted) | |
| def post_process(sentence: str, symbol: str): | |
| if symbol == "sentencepiece": | |
| sentence = sentence.replace(" ", "").replace("\u2581", " ").strip() | |
| elif symbol == "wordpiece": | |
| sentence = sentence.replace(" ", "").replace("_", " ").strip() | |
| elif symbol == "letter": | |
| sentence = sentence.replace(" ", "").replace("|", " ").strip() | |
| elif symbol == "silence": | |
| import re | |
| sentence = sentence.replace("<SIL>", "") | |
| sentence = re.sub(' +', ' ', sentence).strip() | |
| elif symbol == "_EOW": | |
| sentence = sentence.replace(" ", "").replace("_EOW", " ").strip() | |
| elif symbol in {"subword_nmt", "@@ ", "@@"}: | |
| if symbol == "subword_nmt": | |
| symbol = "@@ " | |
| sentence = (sentence + " ").replace(symbol, "").rstrip() | |
| elif symbol == "none": | |
| pass | |
| elif symbol is not None: | |
| raise NotImplementedError(f"Unknown post_process option: {symbol}") | |
| return sentence | |
| def compute_mask_indices( | |
| shape: Tuple[int, int], | |
| padding_mask: Optional[torch.Tensor], | |
| mask_prob: float, | |
| mask_length: int, | |
| mask_type: str = "static", | |
| mask_other: float = 0.0, | |
| min_masks: int = 0, | |
| no_overlap: bool = False, | |
| min_space: int = 0, | |
| ) -> np.ndarray: | |
| """ | |
| Computes random mask spans for a given shape | |
| Args: | |
| shape: the the shape for which to compute masks. | |
| should be of size 2 where first element is batch size and 2nd is timesteps | |
| padding_mask: optional padding mask of the same size as shape, which will prevent masking padded elements | |
| mask_prob: probability for each token to be chosen as start of the span to be masked. this will be multiplied by | |
| number of timesteps divided by length of mask span to mask approximately this percentage of all elements. | |
| however due to overlaps, the actual number will be smaller (unless no_overlap is True) | |
| mask_type: how to compute mask lengths | |
| static = fixed size | |
| uniform = sample from uniform distribution [mask_other, mask_length*2] | |
| normal = sample from normal distribution with mean mask_length and stdev mask_other. mask is min 1 element | |
| poisson = sample from possion distribution with lambda = mask length | |
| min_masks: minimum number of masked spans | |
| no_overlap: if false, will switch to an alternative recursive algorithm that prevents spans from overlapping | |
| min_space: only used if no_overlap is True, this is how many elements to keep unmasked between spans | |
| """ | |
| bsz, all_sz = shape | |
| mask = np.full((bsz, all_sz), False) | |
| all_num_mask = int( | |
| # add a random number for probabilistic rounding | |
| mask_prob * all_sz / float(mask_length) | |
| + np.random.rand() | |
| ) | |
| all_num_mask = max(min_masks, all_num_mask) | |
| mask_idcs = [] | |
| for i in range(bsz): | |
| if padding_mask is not None: | |
| sz = all_sz - padding_mask[i].long().sum().item() | |
| num_mask = int( | |
| # add a random number for probabilistic rounding | |
| mask_prob * sz / float(mask_length) | |
| + np.random.rand() | |
| ) | |
| num_mask = max(min_masks, num_mask) | |
| else: | |
| sz = all_sz | |
| num_mask = all_num_mask | |
| if mask_type == "static": | |
| lengths = np.full(num_mask, mask_length) | |
| elif mask_type == "uniform": | |
| lengths = np.random.randint(mask_other, mask_length * 2 + 1, size=num_mask) | |
| elif mask_type == "normal": | |
| lengths = np.random.normal(mask_length, mask_other, size=num_mask) | |
| lengths = [max(1, int(round(x))) for x in lengths] | |
| elif mask_type == "poisson": | |
| lengths = np.random.poisson(mask_length, size=num_mask) | |
| lengths = [int(round(x)) for x in lengths] | |
| else: | |
| raise Exception("unknown mask selection " + mask_type) | |
| if sum(lengths) == 0: | |
| lengths[0] = min(mask_length, sz - 1) | |
| if no_overlap: | |
| mask_idc = [] | |
| def arrange(s, e, length, keep_length): | |
| span_start = np.random.randint(s, e - length) | |
| mask_idc.extend(span_start + i for i in range(length)) | |
| new_parts = [] | |
| if span_start - s - min_space >= keep_length: | |
| new_parts.append((s, span_start - min_space + 1)) | |
| if e - span_start - keep_length - min_space > keep_length: | |
| new_parts.append((span_start + length + min_space, e)) | |
| return new_parts | |
| parts = [(0, sz)] | |
| min_length = min(lengths) | |
| for length in sorted(lengths, reverse=True): | |
| lens = np.fromiter( | |
| (e - s if e - s >= length + min_space else 0 for s, e in parts), | |
| np.int, | |
| ) | |
| l_sum = np.sum(lens) | |
| if l_sum == 0: | |
| break | |
| probs = lens / np.sum(lens) | |
| c = np.random.choice(len(parts), p=probs) | |
| s, e = parts.pop(c) | |
| parts.extend(arrange(s, e, length, min_length)) | |
| mask_idc = np.asarray(mask_idc) | |
| else: | |
| min_len = min(lengths) | |
| if sz - min_len <= num_mask: | |
| min_len = sz - num_mask - 1 | |
| mask_idc = np.random.choice(sz - min_len, num_mask, replace=False) | |
| mask_idc = np.asarray( | |
| [ | |
| mask_idc[j] + offset | |
| for j in range(len(mask_idc)) | |
| for offset in range(lengths[j]) | |
| ] | |
| ) | |
| mask_idcs.append(np.unique(mask_idc[mask_idc < sz])) | |
| min_len = min([len(m) for m in mask_idcs]) | |
| for i, mask_idc in enumerate(mask_idcs): | |
| if len(mask_idc) > min_len: | |
| mask_idc = np.random.choice(mask_idc, min_len, replace=False) | |
| mask[i, mask_idc] = True | |
| return mask | |
| def get_mem_usage(): | |
| try: | |
| import psutil | |
| mb = 1024 * 1024 | |
| return f"used={psutil.virtual_memory().used / mb}Mb; avail={psutil.virtual_memory().available / mb}Mb" | |
| except ImportError: | |
| return "N/A" | |
| # lens: torch.LongTensor | |
| # returns: torch.BoolTensor | |
| def lengths_to_padding_mask(lens): | |
| bsz, max_lens = lens.size(0), torch.max(lens).item() | |
| mask = torch.arange(max_lens).to(lens.device).view(1, max_lens) | |
| mask = mask.expand(bsz, -1) >= lens.view(bsz, 1).expand(-1, max_lens) | |
| return mask | |
| # lens: torch.LongTensor | |
| # returns: torch.BoolTensor | |
| def lengths_to_mask(lens): | |
| return ~lengths_to_padding_mask(lens) | |
| def get_buckets(sizes, num_buckets): | |
| buckets = np.unique( | |
| np.percentile( | |
| sizes, | |
| np.linspace(0, 100, num_buckets + 1), | |
| interpolation='lower', | |
| )[1:] | |
| ) | |
| return buckets | |
| def get_bucketed_sizes(orig_sizes, buckets): | |
| sizes = np.copy(orig_sizes) | |
| assert np.min(sizes) >= 0 | |
| start_val = -1 | |
| for end_val in buckets: | |
| mask = (sizes > start_val) & (sizes <= end_val) | |
| sizes[mask] = end_val | |
| start_val = end_val | |
| return sizes | |
| def _find_extra_valid_paths(dataset_path: str) -> set: | |
| paths = utils.split_paths(dataset_path) | |
| all_valid_paths = set() | |
| for sub_dir in paths: | |
| contents = PathManager.ls(sub_dir) | |
| valid_paths = [c for c in contents if re.match("valid*[0-9].*", c) is not None] | |
| all_valid_paths |= {os.path.basename(p) for p in valid_paths} | |
| # Remove .bin, .idx etc | |
| roots = {os.path.splitext(p)[0] for p in all_valid_paths} | |
| return roots | |
| def raise_if_valid_subsets_unintentionally_ignored(train_cfg) -> None: | |
| """Raises if there are paths matching 'valid*[0-9].*' which are not combined or ignored.""" | |
| if ( | |
| train_cfg.dataset.ignore_unused_valid_subsets | |
| or train_cfg.dataset.combine_valid_subsets | |
| or train_cfg.dataset.disable_validation | |
| or not hasattr(train_cfg.task, "data") | |
| ): | |
| return | |
| other_paths = _find_extra_valid_paths(train_cfg.task.data) | |
| specified_subsets = train_cfg.dataset.valid_subset.split(",") | |
| ignored_paths = [p for p in other_paths if p not in specified_subsets] | |
| if ignored_paths: | |
| advice = "Set --combine-val to combine them or --ignore-unused-valid-subsets to ignore them." | |
| msg = f"Valid paths {ignored_paths} will be ignored. {advice}" | |
| raise ValueError(msg) | |