code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
# Lint as: python3 import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union SCREAMING_SNAKE_CASE = re.compile(R"^(?P<major>\d+)" R"\.(?P<minor>\d+)" R"\.(?P<patch>\d+)$") @total_ordering @dataclass class lowerCamelCase : '''simple docstring''' lowerCAmelCase_ : str lowerCAmelCase_ : Optional[str] = None lowerCAmelCase_ : Optional[Union[str, int]] = None lowerCAmelCase_ : Optional[Union[str, int]] = None lowerCAmelCase_ : Optional[Union[str, int]] = None def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = _str_to_version_tuple(self.version_str ) def __repr__( self ): return f'''{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}''' @property def A__ ( self ): return self.major, self.minor, self.patch def A__ ( self , lowerCAmelCase ): if isinstance(lowerCAmelCase , lowerCAmelCase ): return Version(lowerCAmelCase ) elif isinstance(lowerCAmelCase , lowerCAmelCase ): return other raise TypeError(f'''{other} (type {type(lowerCAmelCase )}) cannot be compared to version.''' ) def __eq__( self , lowerCAmelCase ): try: UpperCAmelCase_ = self._validate_operand(lowerCAmelCase ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__( self , lowerCAmelCase ): UpperCAmelCase_ = self._validate_operand(lowerCAmelCase ) return self.tuple < other.tuple def __hash__( self ): return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def A__ ( cls , lowerCAmelCase ): UpperCAmelCase_ = {f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def A__ ( self ): return self.version_str def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> List[Any]: UpperCAmelCase_ = _VERSION_REG.match(__SCREAMING_SNAKE_CASE ) if not res: raise ValueError(f'''Invalid version \'{version_str}\'. Format should be x.y.z with {{x,y,z}} being digits.''' ) return tuple(int(__SCREAMING_SNAKE_CASE ) for v in [res.group("major" ), res.group("minor" ), res.group("patch" )] ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: return ".".join(str(__SCREAMING_SNAKE_CASE ) for v in version_tuple )
23
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "xlm-roberta-base": "https://huggingface.co/xlm-roberta-base/resolve/main/config.json", "xlm-roberta-large": "https://huggingface.co/xlm-roberta-large/resolve/main/config.json", "xlm-roberta-large-finetuned-conll02-dutch": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll02-spanish": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-english": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-german": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json" ), } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = 'xlm-roberta' def __init__( self , lowerCAmelCase=3_0522 , lowerCAmelCase=768 , lowerCAmelCase=12 , lowerCAmelCase=12 , lowerCAmelCase=3072 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=1e-1_2 , lowerCAmelCase=1 , lowerCAmelCase=0 , lowerCAmelCase=2 , lowerCAmelCase="absolute" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__(pad_token_id=lowerCAmelCase , bos_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = position_embedding_type UpperCAmelCase_ = use_cache UpperCAmelCase_ = classifier_dropout class lowerCamelCase ( lowercase__ ): '''simple docstring''' @property def A__ ( self ): if self.task == "multiple-choice": UpperCAmelCase_ = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
23
1
from ..utils import DummyObject, requires_backends class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Tuple = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Dict = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Any = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : int = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Optional[Any] = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : str = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[Any] = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : int = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Dict = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : int = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Any = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) class lowerCamelCase ( metaclass=lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = ['flax'] def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(self , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] ) @classmethod def A__ ( cls , *lowerCAmelCase , **lowerCAmelCase ): requires_backends(cls , ["flax"] )
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = int(__SCREAMING_SNAKE_CASE ) if decimal in (0, 1): # Exit cases for the recursion return str(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ , UpperCAmelCase_ = divmod(__SCREAMING_SNAKE_CASE , 2 ) return binary_recursive(__SCREAMING_SNAKE_CASE ) + str(__SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = str(__SCREAMING_SNAKE_CASE ).strip() if not number: raise ValueError("No input value was provided" ) UpperCAmelCase_ = "-" if number.startswith("-" ) else "" UpperCAmelCase_ = number.lstrip("-" ) if not number.isnumeric(): raise ValueError("Input value is not an integer" ) return f'''{negative}0b{binary_recursive(int(__SCREAMING_SNAKE_CASE ) )}''' if __name__ == "__main__": from doctest import testmod testmod()
23
1
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 1 for i in range(1 , num + 1 ): fact *= i return fact def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 0 while number > 0: UpperCAmelCase_ = number % 10 sum_of_digits += last_digit UpperCAmelCase_ = number // 10 # Removing the last_digit from the given number return sum_of_digits def snake_case__ ( __SCREAMING_SNAKE_CASE = 100 ) -> int: UpperCAmelCase_ = factorial(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = split_and_add(__SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
23
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self ): UpperCAmelCase_ = AutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" , return_dict=lowerCAmelCase ).to(lowerCAmelCase ) UpperCAmelCase_ = AutoTokenizer.from_pretrained("google/mt5-small" ) UpperCAmelCase_ = tokenizer("Hello there" , return_tensors="pt" ).input_ids UpperCAmelCase_ = tokenizer("Hi I am" , return_tensors="pt" ).input_ids UpperCAmelCase_ = model(input_ids.to(lowerCAmelCase ) , labels=labels.to(lowerCAmelCase ) ).loss UpperCAmelCase_ = -(labels.shape[-1] * loss.item()) UpperCAmelCase_ = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
23
1
# Function to print upper half of diamond (pyramid) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: for i in range(0 , __SCREAMING_SNAKE_CASE ): for _ in range(0 , n - i - 1 ): # printing spaces print(" " , end="" ) for _ in range(0 , i + 1 ): # printing stars print("* " , end="" ) print() def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> List[str]: for i in range(__SCREAMING_SNAKE_CASE , 0 , -1 ): for _ in range(__SCREAMING_SNAKE_CASE , 0 , -1 ): # printing stars print("* " , end="" ) print() for _ in range(n - i + 1 , 0 , -1 ): # printing spaces print(" " , end="" ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Optional[int]: if n <= 0: print(" ... .... nothing printing :(" ) return floyd(__SCREAMING_SNAKE_CASE ) # upper half reverse_floyd(__SCREAMING_SNAKE_CASE ) # lower half if __name__ == "__main__": print(R"| /\ | |- | |- |--| |\ /| |-") print(R"|/ \| |- |_ |_ |__| | \/ | |_") SCREAMING_SNAKE_CASE = 1 while K: SCREAMING_SNAKE_CASE = int(input("enter the number and , and see the magic : ")) print() pretty_print(user_number) SCREAMING_SNAKE_CASE = int(input("press 0 to exit... and 1 to continue...")) print("Good Bye...")
23
def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ = 0 while b > 0: if b & 1: UpperCAmelCase_ = ((res % c) + (a % c)) % c a += a b >>= 1 return res
23
1
import datasets from .evaluate import evaluate SCREAMING_SNAKE_CASE = "\\n@article{hendrycks2021cuad,\n title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review},\n author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball},\n journal={arXiv preprint arXiv:2103.06268},\n year={2021}\n}\n" SCREAMING_SNAKE_CASE = "\nThis metric wrap the official scoring script for version 1 of the Contract\nUnderstanding Atticus Dataset (CUAD).\nContract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510\ncommercial legal contracts that have been manually labeled to identify 41 categories of important\nclauses that lawyers look for when reviewing contracts in connection with corporate transactions.\n" SCREAMING_SNAKE_CASE = "\nComputes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair as given in the references (see below)\n - 'prediction_text': list of possible texts for the answer, as a list of strings\n depending on a threshold on the confidence probability of each prediction.\n references: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair (see above),\n - 'answers': a Dict in the CUAD dataset format\n {\n 'text': list of possible texts for the answer, as a list of strings\n 'answer_start': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n 'exact_match': Exact match (the normalized answer exactly match the gold answer)\n 'f1': The F-score of predicted tokens versus the gold answer\n 'aupr': Area Under the Precision-Recall curve\n 'prec_at_80_recall': Precision at 80% recall\n 'prec_at_90_recall': Precision at 90% recall\nExamples:\n >>> predictions = [{'prediction_text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.'], 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}]\n >>> references = [{'answers': {'answer_start': [143, 49], 'text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.']}, 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}]\n >>> cuad_metric = datasets.load_metric(\"cuad\")\n >>> results = cuad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 100.0, 'f1': 100.0, 'aupr': 0.0, 'prec_at_80_recall': 1.0, 'prec_at_90_recall': 1.0}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase ( datasets.Metric ): '''simple docstring''' def A__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": { "id": datasets.Value("string" ), "prediction_text": datasets.features.Sequence(datasets.Value("string" ) ), }, "references": { "id": datasets.Value("string" ), "answers": datasets.features.Sequence( { "text": datasets.Value("string" ), "answer_start": datasets.Value("int32" ), } ), }, } ) , codebase_urls=["https://www.atticusprojectai.org/cuad"] , reference_urls=["https://www.atticusprojectai.org/cuad"] , ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = {prediction["id"]: prediction["prediction_text"] for prediction in predictions} UpperCAmelCase_ = [ { "paragraphs": [ { "qas": [ { "answers": [{"text": answer_text} for answer_text in ref["answers"]["text"]], "id": ref["id"], } for ref in references ] } ] } ] UpperCAmelCase_ = evaluate(dataset=lowerCAmelCase , predictions=lowerCAmelCase ) return score
23
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> List[List[ImageInput]]: if isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(__SCREAMING_SNAKE_CASE ): return [[videos]] raise ValueError(f'''Could not make batched video from {videos}''' ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[Any] = ['pixel_values'] def __init__( self , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = True , lowerCAmelCase = 1 / 255 , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = None , **lowerCAmelCase , ): super().__init__(**lowerCAmelCase ) UpperCAmelCase_ = size if size is not None else {"shortest_edge": 224} UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) UpperCAmelCase_ = crop_size if crop_size is not None else {"height": 224, "width": 224} UpperCAmelCase_ = get_size_dict(lowerCAmelCase , param_name="crop_size" ) UpperCAmelCase_ = do_resize UpperCAmelCase_ = size UpperCAmelCase_ = do_center_crop UpperCAmelCase_ = crop_size UpperCAmelCase_ = resample UpperCAmelCase_ = do_rescale UpperCAmelCase_ = rescale_factor UpperCAmelCase_ = do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) if "shortest_edge" in size: UpperCAmelCase_ = get_resize_output_image_size(lowerCAmelCase , size["shortest_edge"] , default_to_square=lowerCAmelCase ) elif "height" in size and "width" in size: UpperCAmelCase_ = (size["height"], size["width"]) else: raise ValueError(f'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = get_size_dict(lowerCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(lowerCAmelCase , size=(size["height"], size["width"]) , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return rescale(lowerCAmelCase , scale=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return normalize(lowerCAmelCase , mean=lowerCAmelCase , std=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , ): if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. UpperCAmelCase_ = to_numpy_array(lowerCAmelCase ) if do_resize: UpperCAmelCase_ = self.resize(image=lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase ) if do_center_crop: UpperCAmelCase_ = self.center_crop(lowerCAmelCase , size=lowerCAmelCase ) if do_rescale: UpperCAmelCase_ = self.rescale(image=lowerCAmelCase , scale=lowerCAmelCase ) if do_normalize: UpperCAmelCase_ = self.normalize(image=lowerCAmelCase , mean=lowerCAmelCase , std=lowerCAmelCase ) UpperCAmelCase_ = to_channel_dimension_format(lowerCAmelCase , lowerCAmelCase ) return image def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , **lowerCAmelCase , ): UpperCAmelCase_ = do_resize if do_resize is not None else self.do_resize UpperCAmelCase_ = resample if resample is not None else self.resample UpperCAmelCase_ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase_ = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase_ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase_ = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else self.image_mean UpperCAmelCase_ = image_std if image_std is not None else self.image_std UpperCAmelCase_ = size if size is not None else self.size UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) UpperCAmelCase_ = crop_size if crop_size is not None else self.crop_size UpperCAmelCase_ = get_size_dict(lowerCAmelCase , param_name="crop_size" ) if not valid_images(lowerCAmelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) UpperCAmelCase_ = make_batched(lowerCAmelCase ) UpperCAmelCase_ = [ [ self._preprocess_image( image=lowerCAmelCase , do_resize=lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase , do_center_crop=lowerCAmelCase , crop_size=lowerCAmelCase , do_rescale=lowerCAmelCase , rescale_factor=lowerCAmelCase , do_normalize=lowerCAmelCase , image_mean=lowerCAmelCase , image_std=lowerCAmelCase , data_format=lowerCAmelCase , ) for img in video ] for video in videos ] UpperCAmelCase_ = {"pixel_values": videos} return BatchFeature(data=lowerCAmelCase , tensor_type=lowerCAmelCase )
23
1
import heapq as hq import math from collections.abc import Iterator class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase ): UpperCAmelCase_ = str(id_ ) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = [] UpperCAmelCase_ = {} # {vertex:distance} def __lt__( self , lowerCAmelCase ): return self.key < other.key def __repr__( self ): return self.id def A__ ( self , lowerCAmelCase ): self.neighbors.append(lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = weight def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __SCREAMING_SNAKE_CASE ) graph[b - 1].add_edge(graph[a - 1] , __SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> list: UpperCAmelCase_ = [] for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = graph[:] while q: UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE ) q.remove(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Iterator[tuple]: for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = list(__SCREAMING_SNAKE_CASE ) hq.heapify(__SCREAMING_SNAKE_CASE ) while h: UpperCAmelCase_ = hq.heappop(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] hq.heapify(__SCREAMING_SNAKE_CASE ) for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def snake_case__ ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 1 for i in range(1 , num + 1 ): fact *= i return fact def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 0 while number > 0: UpperCAmelCase_ = number % 10 sum_of_digits += last_digit UpperCAmelCase_ = number // 10 # Removing the last_digit from the given number return sum_of_digits def snake_case__ ( __SCREAMING_SNAKE_CASE = 100 ) -> int: UpperCAmelCase_ = factorial(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = split_and_add(__SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
23
1
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Union[str, Any]: # Initialise PyTorch model UpperCAmelCase_ = AlbertConfig.from_json_file(__SCREAMING_SNAKE_CASE ) print(f'''Building PyTorch model from configuration: {config}''' ) UpperCAmelCase_ = AlbertForPreTraining(__SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint load_tf_weights_in_albert(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , __SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--albert_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained ALBERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_xlnet": ["XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["XLNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["XLNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "XLNetForMultipleChoice", "XLNetForQuestionAnswering", "XLNetForQuestionAnsweringSimple", "XLNetForSequenceClassification", "XLNetForTokenClassification", "XLNetLMHeadModel", "XLNetModel", "XLNetPreTrainedModel", "load_tf_weights_in_xlnet", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLNetForMultipleChoice", "TFXLNetForQuestionAnsweringSimple", "TFXLNetForSequenceClassification", "TFXLNetForTokenClassification", "TFXLNetLMHeadModel", "TFXLNetMainLayer", "TFXLNetModel", "TFXLNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER', 'False' ) ) is not True, reason='Skipping test because should only be run when releasing minor transformers version', ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.p3.16xlarge', 'results': {'train_runtime': 650, 'eval_accuracy': 0.7, 'eval_loss': 0.6}, }, { 'framework': 'pytorch', 'script': 'run_ddp.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.p3.16xlarge', 'results': {'train_runtime': 600, 'eval_accuracy': 0.7, 'eval_loss': 0.6}, }, { 'framework': 'tensorflow', 'script': 'run_tf_dist.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.p3.16xlarge', 'results': {'train_runtime': 600, 'eval_accuracy': 0.6, 'eval_loss': 0.7}, }, ] ) class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def A__ ( self ): if self.framework == "pytorch": subprocess.run( f'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding="utf-8" , check=lowerCAmelCase , ) assert hasattr(self , "env" ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = f'''{self.env.base_job_name}-{instance_count}-{'ddp' if 'ddp' in self.script else 'smd'}''' # distributed data settings UpperCAmelCase_ = {"smdistributed": {"dataparallel": {"enabled": True}}} if self.script != "run_ddp.py" else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=lowerCAmelCase , instance_count=lowerCAmelCase , instance_type=self.instance_type , debugger_hook_config=lowerCAmelCase , hyperparameters={**self.env.distributed_hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=lowerCAmelCase , py_version="py36" , ) def A__ ( self , lowerCAmelCase ): TrainingJobAnalytics(lowerCAmelCase ).export_csv(f'''{self.env.test_path}/{job_name}_metrics.csv''' ) @parameterized.expand([(2,)] ) def A__ ( self , lowerCAmelCase ): # create estimator UpperCAmelCase_ = self.create_estimator(lowerCAmelCase ) # run training estimator.fit() # result dataframe UpperCAmelCase_ = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCAmelCase_ = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) UpperCAmelCase_ = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCAmelCase_ = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 99_9999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(f'''{estimator.latest_training_job.name}.json''' , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , lowerCAmelCase )
23
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[Any]: # Initialise PyTorch model UpperCAmelCase_ = MobileBertConfig.from_json_file(__SCREAMING_SNAKE_CASE ) print(f'''Building PyTorch model from configuration: {config}''' ) UpperCAmelCase_ = MobileBertForPreTraining(__SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint UpperCAmelCase_ = load_tf_weights_in_mobilebert(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , __SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--mobilebert_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained MobileBERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
23
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE = { "configuration_table_transformer": [ "TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "TableTransformerConfig", "TableTransformerOnnxConfig", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TableTransformerForObjectDetection", "TableTransformerModel", "TableTransformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
import heapq as hq import math from collections.abc import Iterator class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase ): UpperCAmelCase_ = str(id_ ) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = [] UpperCAmelCase_ = {} # {vertex:distance} def __lt__( self , lowerCAmelCase ): return self.key < other.key def __repr__( self ): return self.id def A__ ( self , lowerCAmelCase ): self.neighbors.append(lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = weight def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __SCREAMING_SNAKE_CASE ) graph[b - 1].add_edge(graph[a - 1] , __SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> list: UpperCAmelCase_ = [] for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = graph[:] while q: UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE ) q.remove(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Iterator[tuple]: for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = list(__SCREAMING_SNAKE_CASE ) hq.heapify(__SCREAMING_SNAKE_CASE ) while h: UpperCAmelCase_ = hq.heappop(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] hq.heapify(__SCREAMING_SNAKE_CASE ) for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def snake_case__ ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
23
1
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "alibaba-damo/mgp-str-base": "https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json", } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Tuple = 'mgp-str' def __init__( self , lowerCAmelCase=[32, 128] , lowerCAmelCase=4 , lowerCAmelCase=3 , lowerCAmelCase=27 , lowerCAmelCase=38 , lowerCAmelCase=5_0257 , lowerCAmelCase=3_0522 , lowerCAmelCase=768 , lowerCAmelCase=12 , lowerCAmelCase=12 , lowerCAmelCase=4.0 , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase=1e-5 , lowerCAmelCase=0.0 , lowerCAmelCase=0.0 , lowerCAmelCase=0.0 , lowerCAmelCase=False , lowerCAmelCase=0.02 , **lowerCAmelCase , ): super().__init__(**lowerCAmelCase ) UpperCAmelCase_ = image_size UpperCAmelCase_ = patch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = max_token_length UpperCAmelCase_ = num_character_labels UpperCAmelCase_ = num_bpe_labels UpperCAmelCase_ = num_wordpiece_labels UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = mlp_ratio UpperCAmelCase_ = distilled UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = drop_rate UpperCAmelCase_ = qkv_bias UpperCAmelCase_ = attn_drop_rate UpperCAmelCase_ = drop_path_rate UpperCAmelCase_ = output_aa_attentions UpperCAmelCase_ = initializer_range
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_fnet": ["FNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "FNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["FNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["FNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "FNET_PRETRAINED_MODEL_ARCHIVE_LIST", "FNetForMaskedLM", "FNetForMultipleChoice", "FNetForNextSentencePrediction", "FNetForPreTraining", "FNetForQuestionAnswering", "FNetForSequenceClassification", "FNetForTokenClassification", "FNetLayer", "FNetModel", "FNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = torch.device("cpu") def snake_case__ ( ) -> Optional[int]: UpperCAmelCase_ = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ = Image.open(requests.get(__SCREAMING_SNAKE_CASE , stream=__SCREAMING_SNAKE_CASE ).raw ) return im def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Tuple: if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.17_03E00, 2.11_07E00, -2.08_11E00, 8.86_85E-01, 2.43_60E-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.96_36E-01, 2.34_78E-01, -1.69_63E00, -1.73_81E00, -8.63_37E-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.27_68E-01, -4.74_29E-01, -1.08_97E00, -1.02_48E00, 3.55_23E-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.53_30E-01, 2.42_11E-01, -6.01_85E-01, -8.27_89E-01, -6.04_46E-02] ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Dict: UpperCAmelCase_ = dct.pop(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = val def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ = [] for k in state_dict.keys(): UpperCAmelCase_ = k if ".pwconv" in k: UpperCAmelCase_ = k_new.replace(".pwconv" , ".point_wise_conv" ) if ".dwconv" in k: UpperCAmelCase_ = k_new.replace(".dwconv" , ".depth_wise_conv" ) if ".Proj." in k: UpperCAmelCase_ = k_new.replace(".Proj." , ".proj." ) if "patch_embed" in k_new: UpperCAmelCase_ = k_new.replace("patch_embed" , "swiftformer.patch_embed.patch_embedding" ) if "network" in k_new: UpperCAmelCase_ = k_new.split("." ) if ls[2].isdigit(): UpperCAmelCase_ = "swiftformer.encoder.network." + ls[1] + ".blocks." + ls[2] + "." + ".".join(ls[3:] ) else: UpperCAmelCase_ = k_new.replace("network" , "swiftformer.encoder.network" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[str]: UpperCAmelCase_ = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase_ = 1000 UpperCAmelCase_ = "huggingface/label-files" UpperCAmelCase_ = "imagenet-1k-id2label.json" UpperCAmelCase_ = json.load(open(hf_hub_download(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , repo_type="dataset" ) , "r" ) ) UpperCAmelCase_ = {int(__SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} UpperCAmelCase_ = idalabel UpperCAmelCase_ = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": UpperCAmelCase_ = [3, 3, 6, 4] UpperCAmelCase_ = [48, 56, 112, 220] elif swiftformer_name == "swiftformer_s": UpperCAmelCase_ = [3, 3, 9, 6] UpperCAmelCase_ = [48, 64, 168, 224] elif swiftformer_name == "swiftformer_l1": UpperCAmelCase_ = [4, 3, 10, 5] UpperCAmelCase_ = [48, 96, 192, 384] elif swiftformer_name == "swiftformer_l3": UpperCAmelCase_ = [4, 4, 12, 6] UpperCAmelCase_ = [64, 128, 320, 512] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("https" ): UpperCAmelCase_ = torch.hub.load_state_dict_from_url(__SCREAMING_SNAKE_CASE , map_location="cpu" , check_hash=__SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ = torch.load(__SCREAMING_SNAKE_CASE , map_location="cpu" ) UpperCAmelCase_ = checkpoint UpperCAmelCase_ = create_rename_keys(__SCREAMING_SNAKE_CASE ) for rename_key_src, rename_key_dest in rename_keys: rename_key(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # load HuggingFace model UpperCAmelCase_ = SwiftFormerForImageClassification(__SCREAMING_SNAKE_CASE ).eval() hf_model.load_state_dict(__SCREAMING_SNAKE_CASE ) # prepare test inputs UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = ViTImageProcessor.from_pretrained("preprocessor_config" ) UpperCAmelCase_ = processor(images=__SCREAMING_SNAKE_CASE , return_tensors="pt" ) # compare outputs from both models UpperCAmelCase_ = get_expected_output(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = hf_model(inputs["pixel_values"] ).logits assert hf_logits.shape == torch.Size([1, 1000] ) assert torch.allclose(hf_logits[0, 0:5] , __SCREAMING_SNAKE_CASE , atol=1E-3 ) Path(__SCREAMING_SNAKE_CASE ).mkdir(exist_ok=__SCREAMING_SNAKE_CASE ) print(f'''Saving model {swiftformer_name} to {pytorch_dump_folder_path}''' ) hf_model.save_pretrained(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--swiftformer_name", default="swiftformer_xs", choices=["swiftformer_xs", "swiftformer_s", "swiftformer_l1", "swiftformer_l3"], type=str, help="Name of the SwiftFormer model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default="./converted_outputs/", type=str, help="Path to the output PyTorch model directory.", ) parser.add_argument("--original_ckpt", default=None, type=str, help="Path to the original model checkpoint.") SCREAMING_SNAKE_CASE = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
23
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": 512, } SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": {"do_lower_case": True}, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = VOCAB_FILES_NAMES lowerCAmelCase_ : int = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Dict = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : List[str] = RetriBertTokenizer lowerCAmelCase_ : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase="[UNK]" , lowerCAmelCase="[SEP]" , lowerCAmelCase="[PAD]" , lowerCAmelCase="[CLS]" , lowerCAmelCase="[MASK]" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , do_lower_case=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , pad_token=lowerCAmelCase , cls_token=lowerCAmelCase , mask_token=lowerCAmelCase , tokenize_chinese_chars=lowerCAmelCase , strip_accents=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(lowerCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**lowerCAmelCase ) UpperCAmelCase_ = do_lower_case def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
1
import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCamelCase ( lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Tuple = LEDTokenizer lowerCAmelCase_ : int = LEDTokenizerFast lowerCAmelCase_ : Dict = True def A__ ( self ): super().setUp() UpperCAmelCase_ = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] UpperCAmelCase_ = dict(zip(lowerCAmelCase , range(len(lowerCAmelCase ) ) ) ) UpperCAmelCase_ = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] UpperCAmelCase_ = {"unk_token": "<unk>"} UpperCAmelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowerCAmelCase ) ) def A__ ( self , **lowerCAmelCase ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase ) def A__ ( self , **lowerCAmelCase ): kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): return "lower newer", "lower newer" @cached_property def A__ ( self ): return LEDTokenizer.from_pretrained("allenai/led-base-16384" ) @cached_property def A__ ( self ): return LEDTokenizerFast.from_pretrained("allenai/led-base-16384" ) @require_torch def A__ ( self ): UpperCAmelCase_ = ["A long paragraph for summarization.", "Another paragraph for summarization."] UpperCAmelCase_ = [0, 250, 251, 1_7818, 13, 3_9186, 1938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCAmelCase_ = tokenizer(lowerCAmelCase , max_length=len(lowerCAmelCase ) , padding=lowerCAmelCase , return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase , lowerCAmelCase ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) UpperCAmelCase_ = batch.input_ids.tolist()[0] self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) @require_torch def A__ ( self ): UpperCAmelCase_ = ["A long paragraph for summarization.", "Another paragraph for summarization."] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCAmelCase_ = tokenizer(lowerCAmelCase , padding=lowerCAmelCase , return_tensors="pt" ) self.assertIn("input_ids" , lowerCAmelCase ) self.assertIn("attention_mask" , lowerCAmelCase ) self.assertNotIn("labels" , lowerCAmelCase ) self.assertNotIn("decoder_attention_mask" , lowerCAmelCase ) @require_torch def A__ ( self ): UpperCAmelCase_ = [ "Summary of the text.", "Another summary.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCAmelCase_ = tokenizer(text_target=lowerCAmelCase , max_length=32 , padding="max_length" , return_tensors="pt" ) self.assertEqual(32 , targets["input_ids"].shape[1] ) @require_torch def A__ ( self ): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCAmelCase_ = tokenizer( ["I am a small frog" * 1024, "I am a small frog"] , padding=lowerCAmelCase , truncation=lowerCAmelCase , return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase , lowerCAmelCase ) self.assertEqual(batch.input_ids.shape , (2, 5122) ) @require_torch def A__ ( self ): UpperCAmelCase_ = ["A long paragraph for summarization."] UpperCAmelCase_ = [ "Summary of the text.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCAmelCase_ = tokenizer(lowerCAmelCase , return_tensors="pt" ) UpperCAmelCase_ = tokenizer(text_target=lowerCAmelCase , return_tensors="pt" ) UpperCAmelCase_ = inputs["input_ids"] UpperCAmelCase_ = targets["input_ids"] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def A__ ( self ): for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: UpperCAmelCase_ = ["Summary of the text.", "Another summary."] UpperCAmelCase_ = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] UpperCAmelCase_ = tokenizer(lowerCAmelCase , padding=lowerCAmelCase ) UpperCAmelCase_ = [[0] * len(lowerCAmelCase ) for x in encoded_output["input_ids"]] UpperCAmelCase_ = tokenizer.pad(lowerCAmelCase ) self.assertSequenceEqual(outputs["global_attention_mask"] , lowerCAmelCase ) def A__ ( self ): pass def A__ ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = self.tokenizer_class.from_pretrained(lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = "A, <mask> AllenNLP sentence." UpperCAmelCase_ = tokenizer_r.encode_plus(lowerCAmelCase , add_special_tokens=lowerCAmelCase , return_token_type_ids=lowerCAmelCase ) UpperCAmelCase_ = tokenizer_p.encode_plus(lowerCAmelCase , add_special_tokens=lowerCAmelCase , return_token_type_ids=lowerCAmelCase ) self.assertEqual(sum(tokens_r["token_type_ids"] ) , sum(tokens_p["token_type_ids"] ) ) self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) , sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) , ) UpperCAmelCase_ = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) UpperCAmelCase_ = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) self.assertSequenceEqual(tokens_p["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual( lowerCAmelCase , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( lowerCAmelCase , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] )
23
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-german-cased": ( "https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json" ), "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE = { "distilbert-base-uncased": 512, "distilbert-base-uncased-distilled-squad": 512, "distilbert-base-cased": 512, "distilbert-base-cased-distilled-squad": 512, "distilbert-base-german-cased": 512, "distilbert-base-multilingual-cased": 512, } SCREAMING_SNAKE_CASE = { "distilbert-base-uncased": {"do_lower_case": True}, "distilbert-base-uncased-distilled-squad": {"do_lower_case": True}, "distilbert-base-cased": {"do_lower_case": False}, "distilbert-base-cased-distilled-squad": {"do_lower_case": False}, "distilbert-base-german-cased": {"do_lower_case": False}, "distilbert-base-multilingual-cased": {"do_lower_case": False}, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Any = VOCAB_FILES_NAMES lowerCAmelCase_ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Union[str, Any] = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : int = ['input_ids', 'attention_mask'] lowerCAmelCase_ : str = DistilBertTokenizer def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase="[UNK]" , lowerCAmelCase="[SEP]" , lowerCAmelCase="[PAD]" , lowerCAmelCase="[CLS]" , lowerCAmelCase="[MASK]" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , do_lower_case=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , pad_token=lowerCAmelCase , cls_token=lowerCAmelCase , mask_token=lowerCAmelCase , tokenize_chinese_chars=lowerCAmelCase , strip_accents=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(lowerCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**lowerCAmelCase ) UpperCAmelCase_ = do_lower_case def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
1
import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version(">=", FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp from torch.distributed.checkpoint.default_planner import DefaultLoadPlanner, DefaultSavePlanner from torch.distributed.checkpoint.optimizer import load_sharded_optimizer_state_dict from torch.distributed.fsdp.fully_sharded_data_parallel import FullyShardedDataParallel as FSDP from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType SCREAMING_SNAKE_CASE = get_logger(__name__) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=0 ) -> int: os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) with FSDP.state_dict_type( __SCREAMING_SNAKE_CASE , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): UpperCAmelCase_ = model.state_dict() if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: UpperCAmelCase_ = f'''{MODEL_NAME}.bin''' if model_index == 0 else f'''{MODEL_NAME}_{model_index}.bin''' UpperCAmelCase_ = os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if accelerator.process_index == 0: logger.info(f'''Saving model to {output_model_file}''' ) torch.save(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) logger.info(f'''Model saved to {output_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: UpperCAmelCase_ = ( f'''{MODEL_NAME}_rank{accelerator.process_index}.bin''' if model_index == 0 else f'''{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin''' ) UpperCAmelCase_ = os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) logger.info(f'''Saving model to {output_model_file}''' ) torch.save(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) logger.info(f'''Model saved to {output_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: UpperCAmelCase_ = os.path.join(__SCREAMING_SNAKE_CASE , f'''{MODEL_NAME}_{model_index}''' ) os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) logger.info(f'''Saving model to {ckpt_dir}''' ) UpperCAmelCase_ = {"model": state_dict} dist_cp.save_state_dict( state_dict=__SCREAMING_SNAKE_CASE , storage_writer=dist_cp.FileSystemWriter(__SCREAMING_SNAKE_CASE ) , planner=DefaultSavePlanner() , ) logger.info(f'''Model saved to {ckpt_dir}''' ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=0 ) -> List[Any]: accelerator.wait_for_everyone() with FSDP.state_dict_type( __SCREAMING_SNAKE_CASE , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if type(__SCREAMING_SNAKE_CASE ) != FSDP and accelerator.process_index != 0: if not fsdp_plugin.sync_module_states: raise ValueError( "Set the `sync_module_states` flag to `True` so that model states are synced across processes when " "initializing FSDP object" ) return UpperCAmelCase_ = f'''{MODEL_NAME}.bin''' if model_index == 0 else f'''{MODEL_NAME}_{model_index}.bin''' UpperCAmelCase_ = os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) logger.info(f'''Loading model from {input_model_file}''' ) UpperCAmelCase_ = torch.load(__SCREAMING_SNAKE_CASE ) logger.info(f'''Model loaded from {input_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: UpperCAmelCase_ = ( f'''{MODEL_NAME}_rank{accelerator.process_index}.bin''' if model_index == 0 else f'''{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin''' ) UpperCAmelCase_ = os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) logger.info(f'''Loading model from {input_model_file}''' ) UpperCAmelCase_ = torch.load(__SCREAMING_SNAKE_CASE ) logger.info(f'''Model loaded from {input_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: UpperCAmelCase_ = ( os.path.join(__SCREAMING_SNAKE_CASE , f'''{MODEL_NAME}_{model_index}''' ) if f'''{MODEL_NAME}''' not in input_dir else input_dir ) logger.info(f'''Loading model from {ckpt_dir}''' ) UpperCAmelCase_ = {"model": model.state_dict()} dist_cp.load_state_dict( state_dict=__SCREAMING_SNAKE_CASE , storage_reader=dist_cp.FileSystemReader(__SCREAMING_SNAKE_CASE ) , planner=DefaultLoadPlanner() , ) UpperCAmelCase_ = state_dict["model"] logger.info(f'''Model loaded from {ckpt_dir}''' ) model.load_state_dict(__SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=0 ) -> Optional[Any]: os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) with FSDP.state_dict_type( __SCREAMING_SNAKE_CASE , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): UpperCAmelCase_ = FSDP.optim_state_dict(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if accelerator.process_index == 0: UpperCAmelCase_ = ( f'''{OPTIMIZER_NAME}.bin''' if optimizer_index == 0 else f'''{OPTIMIZER_NAME}_{optimizer_index}.bin''' ) UpperCAmelCase_ = os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) logger.info(f'''Saving Optimizer state to {output_optimizer_file}''' ) torch.save(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) logger.info(f'''Optimizer state saved in {output_optimizer_file}''' ) else: UpperCAmelCase_ = os.path.join(__SCREAMING_SNAKE_CASE , f'''{OPTIMIZER_NAME}_{optimizer_index}''' ) os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) logger.info(f'''Saving Optimizer state to {ckpt_dir}''' ) dist_cp.save_state_dict( state_dict={"optimizer": optim_state} , storage_writer=dist_cp.FileSystemWriter(__SCREAMING_SNAKE_CASE ) , planner=DefaultSavePlanner() , ) logger.info(f'''Optimizer state saved in {ckpt_dir}''' ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=0 ) -> Optional[Any]: accelerator.wait_for_everyone() with FSDP.state_dict_type( __SCREAMING_SNAKE_CASE , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: UpperCAmelCase_ = None # below check should work but currently it isn't working (mostly opytorch issue), # in the meantime disabling it at the cost of excess memory usage # if accelerator.process_index == 0 or not fsdp_plugin.optim_state_dict_config.rank0_only: UpperCAmelCase_ = ( f'''{OPTIMIZER_NAME}.bin''' if optimizer_index == 0 else f'''{OPTIMIZER_NAME}_{optimizer_index}.bin''' ) UpperCAmelCase_ = os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) logger.info(f'''Loading Optimizer state from {input_optimizer_file}''' ) UpperCAmelCase_ = torch.load(__SCREAMING_SNAKE_CASE ) logger.info(f'''Optimizer state loaded from {input_optimizer_file}''' ) else: UpperCAmelCase_ = ( os.path.join(__SCREAMING_SNAKE_CASE , f'''{OPTIMIZER_NAME}_{optimizer_index}''' ) if f'''{OPTIMIZER_NAME}''' not in input_dir else input_dir ) logger.info(f'''Loading Optimizer from {ckpt_dir}''' ) UpperCAmelCase_ = load_sharded_optimizer_state_dict( model_state_dict=model.state_dict() , optimizer_key="optimizer" , storage_reader=dist_cp.FileSystemReader(__SCREAMING_SNAKE_CASE ) , ) UpperCAmelCase_ = optim_state["optimizer"] logger.info(f'''Optimizer loaded from {ckpt_dir}''' ) UpperCAmelCase_ = FSDP.optim_state_dict_to_load(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) optimizer.load_state_dict(__SCREAMING_SNAKE_CASE )
23
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> np.array: UpperCAmelCase_ = f'''{sampling_rate}''' UpperCAmelCase_ = "1" UpperCAmelCase_ = "f32le" UpperCAmelCase_ = [ "ffmpeg", "-i", "pipe:0", "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-hide_banner", "-loglevel", "quiet", "pipe:1", ] try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: UpperCAmelCase_ = ffmpeg_process.communicate(__SCREAMING_SNAKE_CASE ) except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to load audio files from filename" ) from error UpperCAmelCase_ = output_stream[0] UpperCAmelCase_ = np.frombuffer(__SCREAMING_SNAKE_CASE , np.floataa ) if audio.shape[0] == 0: raise ValueError("Malformed soundfile" ) return audio def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = "f32le" , ) -> Dict: UpperCAmelCase_ = f'''{sampling_rate}''' UpperCAmelCase_ = "1" if format_for_conversion == "s16le": UpperCAmelCase_ = 2 elif format_for_conversion == "f32le": UpperCAmelCase_ = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) UpperCAmelCase_ = platform.system() if system == "Linux": UpperCAmelCase_ = "alsa" UpperCAmelCase_ = "default" elif system == "Darwin": UpperCAmelCase_ = "avfoundation" UpperCAmelCase_ = ":0" elif system == "Windows": UpperCAmelCase_ = "dshow" UpperCAmelCase_ = "default" UpperCAmelCase_ = [ "ffmpeg", "-f", format_, "-i", input_, "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-fflags", "nobuffer", "-hide_banner", "-loglevel", "quiet", "pipe:1", ] UpperCAmelCase_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample UpperCAmelCase_ = _ffmpeg_stream(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for item in iterator: yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = "f32le" , ) -> int: if stream_chunk_s is not None: UpperCAmelCase_ = stream_chunk_s else: UpperCAmelCase_ = chunk_length_s UpperCAmelCase_ = ffmpeg_microphone(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , format_for_conversion=__SCREAMING_SNAKE_CASE ) if format_for_conversion == "s16le": UpperCAmelCase_ = np.intaa UpperCAmelCase_ = 2 elif format_for_conversion == "f32le": UpperCAmelCase_ = np.floataa UpperCAmelCase_ = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: UpperCAmelCase_ = chunk_length_s / 6 UpperCAmelCase_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(__SCREAMING_SNAKE_CASE , (int, float) ): UpperCAmelCase_ = [stride_length_s, stride_length_s] UpperCAmelCase_ = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample UpperCAmelCase_ = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample UpperCAmelCase_ = datetime.datetime.now() UpperCAmelCase_ = datetime.timedelta(seconds=__SCREAMING_SNAKE_CASE ) for item in chunk_bytes_iter(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=(stride_left, stride_right) , stream=__SCREAMING_SNAKE_CASE ): # Put everything back in numpy scale UpperCAmelCase_ = np.frombuffer(item["raw"] , dtype=__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = ( item["stride"][0] // size_of_sample, item["stride"][1] // size_of_sample, ) UpperCAmelCase_ = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = False ) -> Dict: UpperCAmelCase_ = B"" UpperCAmelCase_ , UpperCAmelCase_ = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) UpperCAmelCase_ = 0 for raw in iterator: acc += raw if stream and len(__SCREAMING_SNAKE_CASE ) < chunk_len: UpperCAmelCase_ = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(__SCREAMING_SNAKE_CASE ) >= chunk_len: # We are flushing the accumulator UpperCAmelCase_ = (_stride_left, stride_right) UpperCAmelCase_ = {"raw": acc[:chunk_len], "stride": stride} if stream: UpperCAmelCase_ = False yield item UpperCAmelCase_ = stride_left UpperCAmelCase_ = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(__SCREAMING_SNAKE_CASE ) > stride_left: UpperCAmelCase_ = {"raw": acc, "stride": (_stride_left, 0)} if stream: UpperCAmelCase_ = False yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ = 2**24 # 16Mo try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdout=subprocess.PIPE , bufsize=__SCREAMING_SNAKE_CASE ) as ffmpeg_process: while True: UpperCAmelCase_ = ffmpeg_process.stdout.read(__SCREAMING_SNAKE_CASE ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to stream audio files from filename" ) from error
23
1
from __future__ import annotations def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> list[int]: UpperCAmelCase_ = 2 UpperCAmelCase_ = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(__SCREAMING_SNAKE_CASE ) if n > 1: factors.append(__SCREAMING_SNAKE_CASE ) return factors if __name__ == "__main__": import doctest doctest.testmod()
23
from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class lowerCamelCase ( lowercase__, lowercase__ ): '''simple docstring''' @register_to_config def __init__( self , lowerCAmelCase = 768 , ): super().__init__() UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(torch.ones(1 , lowerCAmelCase ) ) def A__ ( self , lowerCAmelCase = None , lowerCAmelCase = None , ): UpperCAmelCase_ = nn.Parameter(self.mean.to(lowerCAmelCase ).to(lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(self.std.to(lowerCAmelCase ).to(lowerCAmelCase ) ) return self def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (embeds - self.mean) * 1.0 / self.std return embeds def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (embeds * self.std) + self.mean return embeds
23
1
import pickle import numpy as np from matplotlib import pyplot as plt class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=0.2 , lowerCAmelCase=0.2 ): UpperCAmelCase_ = bp_numa UpperCAmelCase_ = bp_numa UpperCAmelCase_ = bp_numa UpperCAmelCase_ = conva_get[:2] UpperCAmelCase_ = conva_get[2] UpperCAmelCase_ = size_pa UpperCAmelCase_ = rate_w UpperCAmelCase_ = rate_t UpperCAmelCase_ = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] UpperCAmelCase_ = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) UpperCAmelCase_ = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) UpperCAmelCase_ = -2 * np.random.rand(self.conva[1] ) + 1 UpperCAmelCase_ = -2 * np.random.rand(self.num_bpa ) + 1 UpperCAmelCase_ = -2 * np.random.rand(self.num_bpa ) + 1 def A__ ( self , lowerCAmelCase ): # save model dict with pickle UpperCAmelCase_ = { "num_bp1": self.num_bpa, "num_bp2": self.num_bpa, "num_bp3": self.num_bpa, "conv1": self.conva, "step_conv1": self.step_conva, "size_pooling1": self.size_poolinga, "rate_weight": self.rate_weight, "rate_thre": self.rate_thre, "w_conv1": self.w_conva, "wkj": self.wkj, "vji": self.vji, "thre_conv1": self.thre_conva, "thre_bp2": self.thre_bpa, "thre_bp3": self.thre_bpa, } with open(lowerCAmelCase , "wb" ) as f: pickle.dump(lowerCAmelCase , lowerCAmelCase ) print(f'''Model saved: {save_path}''' ) @classmethod def A__ ( cls , lowerCAmelCase ): # read saved model with open(lowerCAmelCase , "rb" ) as f: UpperCAmelCase_ = pickle.load(lowerCAmelCase ) # noqa: S301 UpperCAmelCase_ = model_dic.get("conv1" ) conv_get.append(model_dic.get("step_conv1" ) ) UpperCAmelCase_ = model_dic.get("size_pooling1" ) UpperCAmelCase_ = model_dic.get("num_bp1" ) UpperCAmelCase_ = model_dic.get("num_bp2" ) UpperCAmelCase_ = model_dic.get("num_bp3" ) UpperCAmelCase_ = model_dic.get("rate_weight" ) UpperCAmelCase_ = model_dic.get("rate_thre" ) # create model instance UpperCAmelCase_ = CNN(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # modify model parameter UpperCAmelCase_ = model_dic.get("w_conv1" ) UpperCAmelCase_ = model_dic.get("wkj" ) UpperCAmelCase_ = model_dic.get("vji" ) UpperCAmelCase_ = model_dic.get("thre_conv1" ) UpperCAmelCase_ = model_dic.get("thre_bp2" ) UpperCAmelCase_ = model_dic.get("thre_bp3" ) return conv_ins def A__ ( self , lowerCAmelCase ): return 1 / (1 + np.exp(-1 * x )) def A__ ( self , lowerCAmelCase ): return round(lowerCAmelCase , 3 ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): # convolution process UpperCAmelCase_ = convs[0] UpperCAmelCase_ = convs[1] UpperCAmelCase_ = np.shape(lowerCAmelCase )[0] # get the data slice of original image data, data_focus UpperCAmelCase_ = [] for i_focus in range(0 , size_data - size_conv + 1 , lowerCAmelCase ): for j_focus in range(0 , size_data - size_conv + 1 , lowerCAmelCase ): UpperCAmelCase_ = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(lowerCAmelCase ) # calculate the feature map of every single kernel, and saved as list of matrix UpperCAmelCase_ = [] UpperCAmelCase_ = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(lowerCAmelCase ): UpperCAmelCase_ = [] for i_focus in range(len(lowerCAmelCase ) ): UpperCAmelCase_ = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(lowerCAmelCase ) ) UpperCAmelCase_ = np.asmatrix(lowerCAmelCase ).reshape( lowerCAmelCase , lowerCAmelCase ) data_featuremap.append(lowerCAmelCase ) # expanding the data slice to One dimenssion UpperCAmelCase_ = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(lowerCAmelCase ) ) UpperCAmelCase_ = np.asarray(lowerCAmelCase ) return focus_list, data_featuremap def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase="average_pool" ): # pooling process UpperCAmelCase_ = len(featuremaps[0] ) UpperCAmelCase_ = int(size_map / size_pooling ) UpperCAmelCase_ = [] for i_map in range(len(lowerCAmelCase ) ): UpperCAmelCase_ = featuremaps[i_map] UpperCAmelCase_ = [] for i_focus in range(0 , lowerCAmelCase , lowerCAmelCase ): for j_focus in range(0 , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(lowerCAmelCase ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(lowerCAmelCase ) ) UpperCAmelCase_ = np.asmatrix(lowerCAmelCase ).reshape(lowerCAmelCase , lowerCAmelCase ) featuremap_pooled.append(lowerCAmelCase ) return featuremap_pooled def A__ ( self , lowerCAmelCase ): # expanding three dimension data to one dimension list UpperCAmelCase_ = [] for i in range(len(lowerCAmelCase ) ): UpperCAmelCase_ = np.shape(data[i] ) UpperCAmelCase_ = data[i].reshape(1 , shapes[0] * shapes[1] ) UpperCAmelCase_ = data_listed.getA().tolist()[0] data_expanded.extend(lowerCAmelCase ) UpperCAmelCase_ = np.asarray(lowerCAmelCase ) return data_expanded def A__ ( self , lowerCAmelCase ): # expanding matrix to one dimension list UpperCAmelCase_ = np.asarray(lowerCAmelCase ) UpperCAmelCase_ = np.shape(lowerCAmelCase ) UpperCAmelCase_ = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = [] UpperCAmelCase_ = 0 for i_map in range(lowerCAmelCase ): UpperCAmelCase_ = np.ones((size_map, size_map) ) for i in range(0 , lowerCAmelCase , lowerCAmelCase ): for j in range(0 , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = pd_pool[ i_pool ] UpperCAmelCase_ = i_pool + 1 UpperCAmelCase_ = np.multiply( lowerCAmelCase , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(lowerCAmelCase ) return pd_all def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=bool ): # model traning print("----------------------Start Training-------------------------" ) print((" - - Shape: Train_Data ", np.shape(lowerCAmelCase )) ) print((" - - Shape: Teach_Data ", np.shape(lowerCAmelCase )) ) UpperCAmelCase_ = 0 UpperCAmelCase_ = [] UpperCAmelCase_ = 1_0000 while rp < n_repeat and mse >= error_accuracy: UpperCAmelCase_ = 0 print(f'''-------------Learning Time {rp}--------------''' ) for p in range(len(lowerCAmelCase ) ): # print('------------Learning Image: %d--------------'%p) UpperCAmelCase_ = np.asmatrix(datas_train[p] ) UpperCAmelCase_ = np.asarray(datas_teach[p] ) UpperCAmelCase_ , UpperCAmelCase_ = self.convolute( lowerCAmelCase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCAmelCase_ = self.pooling(lowerCAmelCase , self.size_poolinga ) UpperCAmelCase_ = np.shape(lowerCAmelCase ) UpperCAmelCase_ = self._expand(lowerCAmelCase ) UpperCAmelCase_ = data_bp_input UpperCAmelCase_ = np.dot(lowerCAmelCase , self.vji.T ) - self.thre_bpa UpperCAmelCase_ = self.sig(lowerCAmelCase ) UpperCAmelCase_ = np.dot(lowerCAmelCase , self.wkj.T ) - self.thre_bpa UpperCAmelCase_ = self.sig(lowerCAmelCase ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- UpperCAmelCase_ = np.multiply( (data_teach - bp_outa) , np.multiply(lowerCAmelCase , (1 - bp_outa) ) ) UpperCAmelCase_ = np.multiply( np.dot(lowerCAmelCase , self.wkj ) , np.multiply(lowerCAmelCase , (1 - bp_outa) ) ) UpperCAmelCase_ = np.dot(lowerCAmelCase , self.vji ) UpperCAmelCase_ = pd_i_all / (self.size_poolinga * self.size_poolinga) UpperCAmelCase_ = pd_conva_pooled.T.getA().tolist() UpperCAmelCase_ = self._calculate_gradient_from_pool( lowerCAmelCase , lowerCAmelCase , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): UpperCAmelCase_ = self._expand_mat(pd_conva_all[k_conv] ) UpperCAmelCase_ = self.rate_weight * np.dot(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) UpperCAmelCase_ = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer UpperCAmelCase_ = self.wkj + pd_k_all.T * bp_outa * self.rate_weight UpperCAmelCase_ = self.vji + pd_j_all.T * bp_outa * self.rate_weight UpperCAmelCase_ = self.thre_bpa - pd_k_all * self.rate_thre UpperCAmelCase_ = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image UpperCAmelCase_ = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) UpperCAmelCase_ = rp + 1 UpperCAmelCase_ = error_count / patterns all_mse.append(lowerCAmelCase ) def draw_error(): UpperCAmelCase_ = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(lowerCAmelCase , "+-" ) plt.plot(lowerCAmelCase , "r--" ) plt.xlabel("Learning Times" ) plt.ylabel("All_mse" ) plt.grid(lowerCAmelCase , alpha=0.5 ) plt.show() print("------------------Training Complished---------------------" ) print((" - - Training epoch: ", rp, f''' - - Mse: {mse:.6f}''') ) if draw_e: draw_error() return mse def A__ ( self , lowerCAmelCase ): # model predict UpperCAmelCase_ = [] print("-------------------Start Testing-------------------------" ) print((" - - Shape: Test_Data ", np.shape(lowerCAmelCase )) ) for p in range(len(lowerCAmelCase ) ): UpperCAmelCase_ = np.asmatrix(datas_test[p] ) UpperCAmelCase_ , UpperCAmelCase_ = self.convolute( lowerCAmelCase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCAmelCase_ = self.pooling(lowerCAmelCase , self.size_poolinga ) UpperCAmelCase_ = self._expand(lowerCAmelCase ) UpperCAmelCase_ = data_bp_input UpperCAmelCase_ = bp_outa * self.vji.T - self.thre_bpa UpperCAmelCase_ = self.sig(lowerCAmelCase ) UpperCAmelCase_ = bp_outa * self.wkj.T - self.thre_bpa UpperCAmelCase_ = self.sig(lowerCAmelCase ) produce_out.extend(bp_outa.getA().tolist() ) UpperCAmelCase_ = [list(map(self.do_round , lowerCAmelCase ) ) for each in produce_out] return np.asarray(lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): # return the data of image after convoluting process so we can check it out UpperCAmelCase_ = np.asmatrix(lowerCAmelCase ) UpperCAmelCase_ , UpperCAmelCase_ = self.convolute( lowerCAmelCase , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) UpperCAmelCase_ = self.pooling(lowerCAmelCase , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
23
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_VISION_2_SEQ_MAPPING if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_VISION_2_SEQ_MAPPING SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @add_end_docstrings(lowercase__ ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): super().__init__(*lowerCAmelCase , **lowerCAmelCase ) requires_backends(self , "vision" ) self.check_model_type( TF_MODEL_FOR_VISION_2_SEQ_MAPPING if self.framework == "tf" else MODEL_FOR_VISION_2_SEQ_MAPPING ) def A__ ( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None ): UpperCAmelCase_ = {} UpperCAmelCase_ = {} if prompt is not None: UpperCAmelCase_ = prompt if generate_kwargs is not None: UpperCAmelCase_ = generate_kwargs if max_new_tokens is not None: if "generate_kwargs" not in forward_kwargs: UpperCAmelCase_ = {} if "max_new_tokens" in forward_kwargs["generate_kwargs"]: raise ValueError( "'max_new_tokens' is defined twice, once in 'generate_kwargs' and once as a direct parameter," " please use only one" ) UpperCAmelCase_ = max_new_tokens return preprocess_params, forward_kwargs, {} def __call__( self , lowerCAmelCase , **lowerCAmelCase ): return super().__call__(lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = load_image(lowerCAmelCase ) if prompt is not None: if not isinstance(lowerCAmelCase , lowerCAmelCase ): raise ValueError( f'''Received an invalid text input, got - {type(lowerCAmelCase )} - but expected a single string. ''' "Note also that one single text can be provided for conditional image to text generation." ) UpperCAmelCase_ = self.model.config.model_type if model_type == "git": UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = self.tokenizer(text=lowerCAmelCase , add_special_tokens=lowerCAmelCase ).input_ids UpperCAmelCase_ = [self.tokenizer.cls_token_id] + input_ids UpperCAmelCase_ = torch.tensor(lowerCAmelCase ).unsqueeze(0 ) model_inputs.update({"input_ids": input_ids} ) elif model_type == "pix2struct": UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , header_text=lowerCAmelCase , return_tensors=self.framework ) elif model_type != "vision-encoder-decoder": # vision-encoder-decoder does not support conditional generation UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = self.tokenizer(lowerCAmelCase , return_tensors=self.framework ) model_inputs.update(lowerCAmelCase ) else: raise ValueError(f'''Model type {model_type} does not support conditional text generation''' ) else: UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) if self.model.config.model_type == "git" and prompt is None: UpperCAmelCase_ = None return model_inputs def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): # Git model sets `model_inputs["input_ids"] = None` in `preprocess` (when `prompt=None`). In batch model, the # pipeline will group them into a list of `None`, which fail `_forward`. Avoid this by checking it first. if ( "input_ids" in model_inputs and isinstance(model_inputs["input_ids"] , lowerCAmelCase ) and all(x is None for x in model_inputs["input_ids"] ) ): UpperCAmelCase_ = None if generate_kwargs is None: UpperCAmelCase_ = {} # FIXME: We need to pop here due to a difference in how `generation.py` and `generation.tf_utils.py` # parse inputs. In the Tensorflow version, `generate` raises an error if we don't use `input_ids` whereas # the PyTorch version matches it with `self.model.main_input_name` or `self.model.encoder.main_input_name` # in the `_prepare_model_inputs` method. UpperCAmelCase_ = model_inputs.pop(self.model.main_input_name ) UpperCAmelCase_ = self.model.generate(lowerCAmelCase , **lowerCAmelCase , **lowerCAmelCase ) return model_outputs def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = [] for output_ids in model_outputs: UpperCAmelCase_ = { "generated_text": self.tokenizer.decode( lowerCAmelCase , skip_special_tokens=lowerCAmelCase , ) } records.append(lowerCAmelCase ) return records
23
1
def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ = 0 while b > 0: if b & 1: UpperCAmelCase_ = ((res % c) + (a % c)) % c a += a b >>= 1 return res
23
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowerCamelCase ( lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : int = TextToVideoSDPipeline lowerCAmelCase_ : Dict = TEXT_TO_IMAGE_PARAMS lowerCAmelCase_ : Optional[Any] = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. lowerCAmelCase_ : Optional[Any] = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def A__ ( self ): torch.manual_seed(0 ) UpperCAmelCase_ = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) UpperCAmelCase_ = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowerCAmelCase , set_alpha_to_one=lowerCAmelCase , ) torch.manual_seed(0 ) UpperCAmelCase_ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) UpperCAmelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="gelu" , projection_dim=512 , ) UpperCAmelCase_ = CLIPTextModel(lowerCAmelCase ) UpperCAmelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCAmelCase_ = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def A__ ( self , lowerCAmelCase , lowerCAmelCase=0 ): if str(lowerCAmelCase ).startswith("mps" ): UpperCAmelCase_ = torch.manual_seed(lowerCAmelCase ) else: UpperCAmelCase_ = torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) UpperCAmelCase_ = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def A__ ( self ): UpperCAmelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ = self.get_dummy_components() UpperCAmelCase_ = TextToVideoSDPipeline(**lowerCAmelCase ) UpperCAmelCase_ = sd_pipe.to(lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase ) UpperCAmelCase_ = self.get_dummy_inputs(lowerCAmelCase ) UpperCAmelCase_ = "np" UpperCAmelCase_ = sd_pipe(**lowerCAmelCase ).frames UpperCAmelCase_ = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) UpperCAmelCase_ = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def A__ ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=3e-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def A__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=1e-2 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def A__ ( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def A__ ( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def A__ ( self ): pass def A__ ( self ): return super().test_progress_bar() @slow @skip_mps class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def A__ ( self ): UpperCAmelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy" ) UpperCAmelCase_ = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) UpperCAmelCase_ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) UpperCAmelCase_ = pipe.to("cuda" ) UpperCAmelCase_ = "Spiderman is surfing" UpperCAmelCase_ = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase_ = pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=25 , output_type="pt" ).frames UpperCAmelCase_ = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2 def A__ ( self ): UpperCAmelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy" ) UpperCAmelCase_ = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) UpperCAmelCase_ = pipe.to("cuda" ) UpperCAmelCase_ = "Spiderman is surfing" UpperCAmelCase_ = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase_ = pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=2 , output_type="pt" ).frames UpperCAmelCase_ = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2
23
1
SCREAMING_SNAKE_CASE = "\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n" SCREAMING_SNAKE_CASE = [{"type": "code", "content": INSTALL_CONTENT}] SCREAMING_SNAKE_CASE = { "{processor_class}": "FakeProcessorClass", "{model_class}": "FakeModelClass", "{object_class}": "FakeObjectClass", }
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = [1] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = 0, 0, 0 UpperCAmelCase_ = ugly_nums[ia] * 2 UpperCAmelCase_ = ugly_nums[ia] * 3 UpperCAmelCase_ = ugly_nums[ia] * 5 for _ in range(1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ugly_nums.append(__SCREAMING_SNAKE_CASE ) if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(f'''{ugly_numbers(200) = }''')
23
1
import jax.numpy as jnp from ...utils import logging from ..ta.modeling_flax_ta import FlaxTaEncoderModel, FlaxTaForConditionalGeneration, FlaxTaModel from .configuration_mta import MTaConfig SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = "T5Config" def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> jnp.ndarray: UpperCAmelCase_ = jnp.zeros_like(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = shifted_input_ids.at[:, 1:].set(input_ids[:, :-1] ) UpperCAmelCase_ = shifted_input_ids.at[:, 0].set(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = jnp.where(shifted_input_ids == -100 , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return shifted_input_ids class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[Any] = 'mt5' lowerCAmelCase_ : List[str] = MTaConfig class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = 'mt5' lowerCAmelCase_ : Optional[int] = MTaConfig class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : int = 'mt5' lowerCAmelCase_ : Optional[int] = MTaConfig
23
import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=3 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=224 , lowerCAmelCase=1000 , lowerCAmelCase=[3, 3, 6, 4] , lowerCAmelCase=[48, 56, 112, 220] , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = is_training UpperCAmelCase_ = use_labels UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = num_labels UpperCAmelCase_ = image_size UpperCAmelCase_ = layer_depths UpperCAmelCase_ = embed_dims def A__ ( self ): UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase_ = self.get_config() return config, pixel_values, labels def A__ ( self ): return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="gelu" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=lowerCAmelCase , layer_scale_init_value=1e-5 , ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = SwiftFormerModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = SwiftFormerForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) UpperCAmelCase_ = SwiftFormerForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A__ ( self ): ((UpperCAmelCase_) , (UpperCAmelCase_) , (UpperCAmelCase_)) = self.prepare_config_and_inputs() UpperCAmelCase_ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Optional[Any] = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () lowerCAmelCase_ : int = ( {'feature-extraction': SwiftFormerModel, 'image-classification': SwiftFormerForImageClassification} if is_torch_available() else {} ) lowerCAmelCase_ : List[Any] = False lowerCAmelCase_ : Dict = False lowerCAmelCase_ : int = False lowerCAmelCase_ : str = False lowerCAmelCase_ : Optional[Any] = False def A__ ( self ): UpperCAmelCase_ = SwiftFormerModelTester(self ) UpperCAmelCase_ = ConfigTester( self , config_class=lowerCAmelCase , has_text_modality=lowerCAmelCase , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def A__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="SwiftFormer does not use inputs_embeds" ) def A__ ( self ): pass def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase , nn.Linear ) ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase ) @slow def A__ ( self ): for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = SwiftFormerModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) @unittest.skip(reason="SwiftFormer does not output attentions" ) def A__ ( self ): pass def A__ ( self ): def check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = model_class(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCAmelCase_ = model(**self._prepare_for_class(lowerCAmelCase , lowerCAmelCase ) ) UpperCAmelCase_ = outputs.hidden_states UpperCAmelCase_ = 8 self.assertEqual(len(lowerCAmelCase ) , lowerCAmelCase ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(lowerCAmelCase ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = True check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ = True check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): def _config_zero_init(lowerCAmelCase ): UpperCAmelCase_ = copy.deepcopy(lowerCAmelCase ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(lowerCAmelCase , lowerCAmelCase , 1e-1_0 ) if isinstance(getattr(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) , lowerCAmelCase ): UpperCAmelCase_ = _config_zero_init(getattr(lowerCAmelCase , lowerCAmelCase ) ) setattr(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return configs_no_init UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = _config_zero_init(lowerCAmelCase ) for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(config=lowerCAmelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9) / 1e9).round().item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def A__ ( self ): pass def snake_case__ ( ) -> str: UpperCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def A__ ( self ): return ViTImageProcessor.from_pretrained("MBZUAI/swiftformer-xs" ) if is_vision_available() else None @slow def A__ ( self ): UpperCAmelCase_ = SwiftFormerForImageClassification.from_pretrained("MBZUAI/swiftformer-xs" ).to(lowerCAmelCase ) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(images=lowerCAmelCase , return_tensors="pt" ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase_ = model(**lowerCAmelCase ) # verify the logits UpperCAmelCase_ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase ) UpperCAmelCase_ = torch.tensor([[-2.1_7_0_3e0_0, 2.1_1_0_7e0_0, -2.0_8_1_1e0_0]] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase , atol=1e-4 ) )
23
1
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin SCREAMING_SNAKE_CASE = "▁" SCREAMING_SNAKE_CASE = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class lowerCamelCase ( lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : str = BigBirdTokenizer lowerCAmelCase_ : Union[str, Any] = BigBirdTokenizerFast lowerCAmelCase_ : Optional[int] = True lowerCAmelCase_ : Union[str, Any] = True def A__ ( self ): super().setUp() UpperCAmelCase_ = self.tokenizer_class(lowerCAmelCase , keep_accents=lowerCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def A__ ( self ): UpperCAmelCase_ = "<s>" UpperCAmelCase_ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase ) , lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase ) , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "[MASK]" ) self.assertEqual(len(lowerCAmelCase ) , 1004 ) def A__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def A__ ( self ): if not self.test_rust_tokenizer: return UpperCAmelCase_ = self.get_tokenizer() UpperCAmelCase_ = self.get_rust_tokenizer() UpperCAmelCase_ = "I was born in 92000, and this is falsé." UpperCAmelCase_ = tokenizer.tokenize(lowerCAmelCase ) UpperCAmelCase_ = rust_tokenizer.tokenize(lowerCAmelCase ) self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) UpperCAmelCase_ = rust_tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = self.get_rust_tokenizer() UpperCAmelCase_ = tokenizer.encode(lowerCAmelCase ) UpperCAmelCase_ = rust_tokenizer.encode(lowerCAmelCase ) self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = BigBirdTokenizer(lowerCAmelCase , keep_accents=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCAmelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase ) , [285, 46, 10, 170, 382] , ) UpperCAmelCase_ = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowerCAmelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) UpperCAmelCase_ = tokenizer.convert_tokens_to_ids(lowerCAmelCase ) self.assertListEqual( lowerCAmelCase , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(lowerCAmelCase ) self.assertListEqual( lowerCAmelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def A__ ( self ): return BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) @slow def A__ ( self ): UpperCAmelCase_ = "Hello World!" UpperCAmelCase_ = [65, 1_8536, 2260, 101, 66] self.assertListEqual(lowerCAmelCase , self.big_tokenizer.encode(lowerCAmelCase ) ) @slow def A__ ( self ): UpperCAmelCase_ = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) # fmt: off UpperCAmelCase_ = [65, 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 3_4324, 497, 391, 408, 1_1342, 1244, 385, 100, 938, 985, 456, 574, 362, 1_2597, 3200, 3129, 1172, 66] # noqa: E231 # fmt: on self.assertListEqual(lowerCAmelCase , self.big_tokenizer.encode(lowerCAmelCase ) ) @require_torch @slow def A__ ( self ): import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence UpperCAmelCase_ = list(self.big_tokenizer.get_vocab().keys() )[:10] UpperCAmelCase_ = " ".join(lowerCAmelCase ) UpperCAmelCase_ = self.big_tokenizer.encode_plus(lowerCAmelCase , return_tensors="pt" , return_token_type_ids=lowerCAmelCase ) UpperCAmelCase_ = self.big_tokenizer.batch_encode_plus( [sequence + " " + sequence] , return_tensors="pt" , return_token_type_ids=lowerCAmelCase ) UpperCAmelCase_ = BigBirdConfig(attention_type="original_full" ) UpperCAmelCase_ = BigBirdModel(lowerCAmelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowerCAmelCase ) model(**lowerCAmelCase ) @slow def A__ ( self ): UpperCAmelCase_ = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) UpperCAmelCase_ = tokenizer.decode(tokenizer("Paris is the [MASK]." ).input_ids ) self.assertTrue(decoded_text == "[CLS] Paris is the[MASK].[SEP]" ) @slow def A__ ( self ): # fmt: off UpperCAmelCase_ = {"input_ids": [[65, 3_9286, 458, 3_6335, 2001, 456, 1_3073, 1_3266, 455, 113, 7746, 1741, 1_1157, 391, 1_3073, 1_3266, 455, 113, 3967, 3_5412, 113, 4936, 109, 3870, 2377, 113, 3_0084, 4_5720, 458, 134, 1_7496, 112, 503, 1_1672, 113, 118, 112, 5665, 1_3347, 3_8687, 112, 1496, 3_1389, 112, 3268, 4_7264, 134, 962, 112, 1_6377, 8035, 2_3130, 430, 1_2169, 1_5518, 2_8592, 458, 146, 4_1697, 109, 391, 1_2169, 1_5518, 1_6689, 458, 146, 4_1358, 109, 452, 726, 4034, 111, 763, 3_5412, 5082, 388, 1903, 111, 9051, 391, 2870, 4_8918, 1900, 1123, 550, 998, 112, 9586, 1_5985, 455, 391, 410, 2_2955, 3_7636, 114, 66], [65, 448, 1_7496, 419, 3663, 385, 763, 113, 2_7533, 2870, 3283, 1_3043, 1639, 2_4713, 523, 656, 2_4013, 1_8550, 2521, 517, 2_7014, 2_1244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 1_1786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2169, 7687, 2_1932, 1_8146, 726, 363, 1_7032, 3391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase , model_name="google/bigbird-roberta-base" , revision="215c99f1600e06f83acce68422f2035b2b5c3510" , )
23
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available SCREAMING_SNAKE_CASE = {"tokenization_herbert": ["HerbertTokenizer"]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["HerbertTokenizerFast"] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "feature_projection.layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> int: for attribute in key.split("." ): UpperCAmelCase_ = getattr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if weight_type is not None: UpperCAmelCase_ = getattr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ).shape else: UpperCAmelCase_ = hf_pointer.shape assert hf_shape == value.shape, ( f'''Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be''' f''' {value.shape} for {full_name}''' ) if weight_type == "weight": UpperCAmelCase_ = value elif weight_type == "weight_g": UpperCAmelCase_ = value elif weight_type == "weight_v": UpperCAmelCase_ = value elif weight_type == "bias": UpperCAmelCase_ = value else: UpperCAmelCase_ = value logger.info(f'''{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.''' ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: UpperCAmelCase_ = [] UpperCAmelCase_ = fairseq_model.state_dict() UpperCAmelCase_ = hf_model.hubert.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase_ = False if "conv_layers" in name: load_conv_layer( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == "group" , ) UpperCAmelCase_ = True else: for key, mapped_key in MAPPING.items(): UpperCAmelCase_ = "hubert." + mapped_key if (is_finetuned and mapped_key != "lm_head") else mapped_key if key in name or (key.split("w2v_model." )[-1] == name.split("." )[0] and not is_finetuned): UpperCAmelCase_ = True if "*" in mapped_key: UpperCAmelCase_ = name.split(__SCREAMING_SNAKE_CASE )[0].split("." )[-2] UpperCAmelCase_ = mapped_key.replace("*" , __SCREAMING_SNAKE_CASE ) if "weight_g" in name: UpperCAmelCase_ = "weight_g" elif "weight_v" in name: UpperCAmelCase_ = "weight_v" elif "weight" in name: UpperCAmelCase_ = "weight" elif "bias" in name: UpperCAmelCase_ = "bias" else: UpperCAmelCase_ = None set_recursively(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(__SCREAMING_SNAKE_CASE ) logger.warning(f'''Unused weights: {unused_weights}''' ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ = full_name.split("conv_layers." )[-1] UpperCAmelCase_ = name.split("." ) UpperCAmelCase_ = int(items[0] ) UpperCAmelCase_ = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) UpperCAmelCase_ = value logger.info(f'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) UpperCAmelCase_ = value logger.info(f'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) UpperCAmelCase_ = value logger.info(f'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) UpperCAmelCase_ = value logger.info(f'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(__SCREAMING_SNAKE_CASE ) @torch.no_grad() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=True ) -> Dict: if config_path is not None: UpperCAmelCase_ = HubertConfig.from_pretrained(__SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ = HubertConfig() if is_finetuned: if dict_path: UpperCAmelCase_ = Dictionary.load(__SCREAMING_SNAKE_CASE ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq UpperCAmelCase_ = target_dict.pad_index UpperCAmelCase_ = target_dict.bos_index UpperCAmelCase_ = target_dict.eos_index UpperCAmelCase_ = len(target_dict.symbols ) UpperCAmelCase_ = os.path.join(__SCREAMING_SNAKE_CASE , "vocab.json" ) if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error("--pytorch_dump_folder_path ({}) should be a directory".format(__SCREAMING_SNAKE_CASE ) ) return os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) with open(__SCREAMING_SNAKE_CASE , "w" , encoding="utf-8" ) as vocab_handle: json.dump(target_dict.indices , __SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = WavaVecaCTCTokenizer( __SCREAMING_SNAKE_CASE , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="|" , do_lower_case=__SCREAMING_SNAKE_CASE , ) UpperCAmelCase_ = True if config.feat_extract_norm == "layer" else False UpperCAmelCase_ = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=__SCREAMING_SNAKE_CASE , return_attention_mask=__SCREAMING_SNAKE_CASE , ) UpperCAmelCase_ = WavaVecaProcessor(feature_extractor=__SCREAMING_SNAKE_CASE , tokenizer=__SCREAMING_SNAKE_CASE ) processor.save_pretrained(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = HubertForCTC(__SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ = HubertModel(__SCREAMING_SNAKE_CASE ) if is_finetuned: UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] )} ) else: UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) UpperCAmelCase_ = model[0].eval() recursively_load_weights(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) hf_wavavec.save_pretrained(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_hubert_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
23
import math def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> list[int]: UpperCAmelCase_ = [] UpperCAmelCase_ = 2 UpperCAmelCase_ = int(math.sqrt(__SCREAMING_SNAKE_CASE ) ) # Size of every segment UpperCAmelCase_ = [True] * (end + 1) UpperCAmelCase_ = [] while start <= end: if temp[start] is True: in_prime.append(__SCREAMING_SNAKE_CASE ) for i in range(start * start , end + 1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = False start += 1 prime += in_prime UpperCAmelCase_ = end + 1 UpperCAmelCase_ = min(2 * end , __SCREAMING_SNAKE_CASE ) while low <= n: UpperCAmelCase_ = [True] * (high - low + 1) for each in in_prime: UpperCAmelCase_ = math.floor(low / each ) * each if t < low: t += each for j in range(__SCREAMING_SNAKE_CASE , high + 1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = False for j in range(len(__SCREAMING_SNAKE_CASE ) ): if temp[j] is True: prime.append(j + low ) UpperCAmelCase_ = high + 1 UpperCAmelCase_ = min(high + end , __SCREAMING_SNAKE_CASE ) return prime print(sieve(10**6))
23
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE = { "configuration_gpt_bigcode": ["GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTBigCodeConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTBigCodeForSequenceClassification", "GPTBigCodeForTokenClassification", "GPTBigCodeForCausalLM", "GPTBigCodeModel", "GPTBigCodePreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : torch.FloatTensor class lowerCamelCase ( lowercase__, lowercase__ ): '''simple docstring''' @register_to_config def __init__( self , lowerCAmelCase = 32 , lowerCAmelCase = 64 , lowerCAmelCase = 20 , lowerCAmelCase = 768 , lowerCAmelCase=77 , lowerCAmelCase=4 , lowerCAmelCase = 0.0 , lowerCAmelCase = "silu" , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = "linear" , lowerCAmelCase = "prd" , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , ): super().__init__() UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = attention_head_dim UpperCAmelCase_ = num_attention_heads * attention_head_dim UpperCAmelCase_ = additional_embeddings UpperCAmelCase_ = time_embed_dim or inner_dim UpperCAmelCase_ = embedding_proj_dim or embedding_dim UpperCAmelCase_ = clip_embed_dim or embedding_dim UpperCAmelCase_ = Timesteps(lowerCAmelCase , lowerCAmelCase , 0 ) UpperCAmelCase_ = TimestepEmbedding(lowerCAmelCase , lowerCAmelCase , out_dim=lowerCAmelCase , act_fn=lowerCAmelCase ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) if embedding_proj_norm_type is None: UpperCAmelCase_ = None elif embedding_proj_norm_type == "layer": UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) else: raise ValueError(f'''unsupported embedding_proj_norm_type: {embedding_proj_norm_type}''' ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) if encoder_hid_proj_type is None: UpperCAmelCase_ = None elif encoder_hid_proj_type == "linear": UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) else: raise ValueError(f'''unsupported encoder_hid_proj_type: {encoder_hid_proj_type}''' ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , num_embeddings + additional_embeddings , lowerCAmelCase ) ) if added_emb_type == "prd": UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , 1 , lowerCAmelCase ) ) elif added_emb_type is None: UpperCAmelCase_ = None else: raise ValueError( f'''`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `\'prd\'` or `None`.''' ) UpperCAmelCase_ = nn.ModuleList( [ BasicTransformerBlock( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , dropout=lowerCAmelCase , activation_fn="gelu" , attention_bias=lowerCAmelCase , ) for d in range(lowerCAmelCase ) ] ) if norm_in_type == "layer": UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) elif norm_in_type is None: UpperCAmelCase_ = None else: raise ValueError(f'''Unsupported norm_in_type: {norm_in_type}.''' ) UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] , -10000.0 ) causal_attention_mask.triu_(1 ) UpperCAmelCase_ = causal_attention_mask[None, ...] self.register_buffer("causal_attention_mask" , lowerCAmelCase , persistent=lowerCAmelCase ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def A__ ( self ): UpperCAmelCase_ = {} def fn_recursive_add_processors(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if hasattr(lowerCAmelCase , "set_processor" ): UpperCAmelCase_ = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f'''{name}.{sub_name}''' , lowerCAmelCase , lowerCAmelCase ) return processors for name, module in self.named_children(): fn_recursive_add_processors(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return processors def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = len(self.attn_processors.keys() ) if isinstance(lowerCAmelCase , lowerCAmelCase ) and len(lowerCAmelCase ) != count: raise ValueError( f'''A dict of processors was passed, but the number of processors {len(lowerCAmelCase )} does not match the''' f''' number of attention layers: {count}. Please make sure to pass {count} processor classes.''' ) def fn_recursive_attn_processor(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if hasattr(lowerCAmelCase , "set_processor" ): if not isinstance(lowerCAmelCase , lowerCAmelCase ): module.set_processor(lowerCAmelCase ) else: module.set_processor(processor.pop(f'''{name}.processor''' ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f'''{name}.{sub_name}''' , lowerCAmelCase , lowerCAmelCase ) for name, module in self.named_children(): fn_recursive_attn_processor(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): self.set_attn_processor(AttnProcessor() ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = True , ): UpperCAmelCase_ = hidden_states.shape[0] UpperCAmelCase_ = timestep if not torch.is_tensor(lowerCAmelCase ): UpperCAmelCase_ = torch.tensor([timesteps] , dtype=torch.long , device=hidden_states.device ) elif torch.is_tensor(lowerCAmelCase ) and len(timesteps.shape ) == 0: UpperCAmelCase_ = timesteps[None].to(hidden_states.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML UpperCAmelCase_ = timesteps * torch.ones(lowerCAmelCase , dtype=timesteps.dtype , device=timesteps.device ) UpperCAmelCase_ = self.time_proj(lowerCAmelCase ) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. UpperCAmelCase_ = timesteps_projected.to(dtype=self.dtype ) UpperCAmelCase_ = self.time_embedding(lowerCAmelCase ) if self.embedding_proj_norm is not None: UpperCAmelCase_ = self.embedding_proj_norm(lowerCAmelCase ) UpperCAmelCase_ = self.embedding_proj(lowerCAmelCase ) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: UpperCAmelCase_ = self.encoder_hidden_states_proj(lowerCAmelCase ) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError("`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set" ) UpperCAmelCase_ = self.proj_in(lowerCAmelCase ) UpperCAmelCase_ = self.positional_embedding.to(hidden_states.dtype ) UpperCAmelCase_ = [] UpperCAmelCase_ = 0 if encoder_hidden_states is not None: additional_embeds.append(lowerCAmelCase ) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape ) == 2: UpperCAmelCase_ = proj_embeddings[:, None, :] if len(hidden_states.shape ) == 2: UpperCAmelCase_ = hidden_states[:, None, :] UpperCAmelCase_ = additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: UpperCAmelCase_ = self.prd_embedding.to(hidden_states.dtype ).expand(lowerCAmelCase , -1 , -1 ) additional_embeds.append(lowerCAmelCase ) UpperCAmelCase_ = torch.cat( lowerCAmelCase , dim=1 , ) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens UpperCAmelCase_ = additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: UpperCAmelCase_ = F.pad( lowerCAmelCase , ( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) , value=0.0 , ) UpperCAmelCase_ = hidden_states + positional_embeddings if attention_mask is not None: UpperCAmelCase_ = (1 - attention_mask.to(hidden_states.dtype )) * -10000.0 UpperCAmelCase_ = F.pad(lowerCAmelCase , (0, self.additional_embeddings) , value=0.0 ) UpperCAmelCase_ = (attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype ) UpperCAmelCase_ = attention_mask.repeat_interleave(self.config.num_attention_heads , dim=0 ) if self.norm_in is not None: UpperCAmelCase_ = self.norm_in(lowerCAmelCase ) for block in self.transformer_blocks: UpperCAmelCase_ = block(lowerCAmelCase , attention_mask=lowerCAmelCase ) UpperCAmelCase_ = self.norm_out(lowerCAmelCase ) if self.prd_embedding is not None: UpperCAmelCase_ = hidden_states[:, -1] else: UpperCAmelCase_ = hidden_states[:, additional_embeddings_len:] UpperCAmelCase_ = self.proj_to_clip_embeddings(lowerCAmelCase ) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (prior_latents * self.clip_std) + self.clip_mean return prior_latents
23
1
import numpy as np class lowerCamelCase : '''simple docstring''' def __init__( self ): UpperCAmelCase_ = (0, 0) UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = 0 UpperCAmelCase_ = 0 def __eq__( self , lowerCAmelCase ): return self.position == cell.position def A__ ( self ): print(self.position ) class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase=(5, 5) ): UpperCAmelCase_ = np.zeros(lowerCAmelCase ) UpperCAmelCase_ = world_size[0] UpperCAmelCase_ = world_size[1] def A__ ( self ): print(self.w ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] UpperCAmelCase_ = cell.position[0] UpperCAmelCase_ = cell.position[1] UpperCAmelCase_ = [] for n in neughbour_cord: UpperCAmelCase_ = current_x + n[0] UpperCAmelCase_ = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: UpperCAmelCase_ = Cell() UpperCAmelCase_ = (x, y) UpperCAmelCase_ = cell neighbours.append(lowerCAmelCase ) return neighbours def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: UpperCAmelCase_ = [] UpperCAmelCase_ = [] _open.append(__SCREAMING_SNAKE_CASE ) while _open: UpperCAmelCase_ = np.argmin([n.f for n in _open] ) UpperCAmelCase_ = _open[min_f] _closed.append(_open.pop(__SCREAMING_SNAKE_CASE ) ) if current == goal: break for n in world.get_neigbours(__SCREAMING_SNAKE_CASE ): for c in _closed: if c == n: continue UpperCAmelCase_ = current.g + 1 UpperCAmelCase_ , UpperCAmelCase_ = n.position UpperCAmelCase_ , UpperCAmelCase_ = goal.position UpperCAmelCase_ = (ya - ya) ** 2 + (xa - xa) ** 2 UpperCAmelCase_ = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = [] while current.parent is not None: path.append(current.position ) UpperCAmelCase_ = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": SCREAMING_SNAKE_CASE = Gridworld() # Start position and goal SCREAMING_SNAKE_CASE = Cell() SCREAMING_SNAKE_CASE = (0, 0) SCREAMING_SNAKE_CASE = Cell() SCREAMING_SNAKE_CASE = (4, 4) print(f'''path from {start.position} to {goal.position}''') SCREAMING_SNAKE_CASE = astar(world, start, goal) # Just for visual reasons. for i in s: SCREAMING_SNAKE_CASE = 1 print(world.w)
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE = { "configuration_gpt_bigcode": ["GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTBigCodeConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTBigCodeForSequenceClassification", "GPTBigCodeForTokenClassification", "GPTBigCodeForCausalLM", "GPTBigCodeModel", "GPTBigCodePreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @add_end_docstrings( lowercase__, r'\n top_k (`int`, defaults to 5):\n The number of predictions to return.\n targets (`str` or `List[str]`, *optional*):\n When passed, the model will limit the scores to the passed targets instead of looking up in the whole\n vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting\n token will be used (with a warning, and that might be slower).\n\n ', ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' def A__ ( self , lowerCAmelCase ): if self.framework == "tf": UpperCAmelCase_ = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": UpperCAmelCase_ = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=lowerCAmelCase ) else: raise ValueError("Unsupported framework" ) return masked_index def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = self.get_masked_index(lowerCAmelCase ) UpperCAmelCase_ = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( "fill-mask" , self.model.base_model_prefix , f'''No mask_token ({self.tokenizer.mask_token}) found on the input''' , ) def A__ ( self , lowerCAmelCase ): if isinstance(lowerCAmelCase , lowerCAmelCase ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input["input_ids"][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase=None , **lowerCAmelCase ): if return_tensors is None: UpperCAmelCase_ = self.framework UpperCAmelCase_ = self.tokenizer(lowerCAmelCase , return_tensors=lowerCAmelCase ) self.ensure_exactly_one_mask_token(lowerCAmelCase ) return model_inputs def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = self.model(**lowerCAmelCase ) UpperCAmelCase_ = model_inputs["input_ids"] return model_outputs def A__ ( self , lowerCAmelCase , lowerCAmelCase=5 , lowerCAmelCase=None ): # Cap top_k if there are targets if target_ids is not None and target_ids.shape[0] < top_k: UpperCAmelCase_ = target_ids.shape[0] UpperCAmelCase_ = model_outputs["input_ids"][0] UpperCAmelCase_ = model_outputs["logits"] if self.framework == "tf": UpperCAmelCase_ = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] UpperCAmelCase_ = outputs.numpy() UpperCAmelCase_ = outputs[0, masked_index, :] UpperCAmelCase_ = stable_softmax(lowerCAmelCase , axis=-1 ) if target_ids is not None: UpperCAmelCase_ = tf.gather_nd(tf.squeeze(lowerCAmelCase , 0 ) , target_ids.reshape(-1 , 1 ) ) UpperCAmelCase_ = tf.expand_dims(lowerCAmelCase , 0 ) UpperCAmelCase_ = tf.math.top_k(lowerCAmelCase , k=lowerCAmelCase ) UpperCAmelCase_ , UpperCAmelCase_ = topk.values.numpy(), topk.indices.numpy() else: UpperCAmelCase_ = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=lowerCAmelCase ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample UpperCAmelCase_ = outputs[0, masked_index, :] UpperCAmelCase_ = logits.softmax(dim=-1 ) if target_ids is not None: UpperCAmelCase_ = probs[..., target_ids] UpperCAmelCase_ , UpperCAmelCase_ = probs.topk(lowerCAmelCase ) UpperCAmelCase_ = [] UpperCAmelCase_ = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): UpperCAmelCase_ = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place UpperCAmelCase_ = input_ids.numpy().copy() if target_ids is not None: UpperCAmelCase_ = target_ids[p].tolist() UpperCAmelCase_ = p # Filter padding out: UpperCAmelCase_ = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back UpperCAmelCase_ = self.tokenizer.decode(lowerCAmelCase , skip_special_tokens=lowerCAmelCase ) UpperCAmelCase_ = {"score": v, "token": p, "token_str": self.tokenizer.decode([p] ), "sequence": sequence} row.append(lowerCAmelCase ) result.append(lowerCAmelCase ) if single_mask: return result[0] return result def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): if isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = [targets] try: UpperCAmelCase_ = self.tokenizer.get_vocab() except Exception: UpperCAmelCase_ = {} UpperCAmelCase_ = [] for target in targets: UpperCAmelCase_ = vocab.get(lowerCAmelCase , lowerCAmelCase ) if id_ is None: UpperCAmelCase_ = self.tokenizer( lowerCAmelCase , add_special_tokens=lowerCAmelCase , return_attention_mask=lowerCAmelCase , return_token_type_ids=lowerCAmelCase , max_length=1 , truncation=lowerCAmelCase , )["input_ids"] if len(lowerCAmelCase ) == 0: logger.warning( f'''The specified target token `{target}` does not exist in the model vocabulary. ''' "We cannot replace it with anything meaningful, ignoring it" ) continue UpperCAmelCase_ = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( f'''The specified target token `{target}` does not exist in the model vocabulary. ''' f'''Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.''' ) target_ids.append(id_ ) UpperCAmelCase_ = list(set(lowerCAmelCase ) ) if len(lowerCAmelCase ) == 0: raise ValueError("At least one target must be provided when passed." ) UpperCAmelCase_ = np.array(lowerCAmelCase ) return target_ids def A__ ( self , lowerCAmelCase=None , lowerCAmelCase=None ): UpperCAmelCase_ = {} if targets is not None: UpperCAmelCase_ = self.get_target_ids(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = target_ids if top_k is not None: UpperCAmelCase_ = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( "fill-mask" , self.model.base_model_prefix , "The tokenizer does not define a `mask_token`." ) return {}, {}, postprocess_params def __call__( self , lowerCAmelCase , *lowerCAmelCase , **lowerCAmelCase ): UpperCAmelCase_ = super().__call__(lowerCAmelCase , **lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) and len(lowerCAmelCase ) == 1: return outputs[0] return outputs
23
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "xlm-roberta-base": "https://huggingface.co/xlm-roberta-base/resolve/main/config.json", "xlm-roberta-large": "https://huggingface.co/xlm-roberta-large/resolve/main/config.json", "xlm-roberta-large-finetuned-conll02-dutch": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll02-spanish": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-english": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-german": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json" ), } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = 'xlm-roberta' def __init__( self , lowerCAmelCase=3_0522 , lowerCAmelCase=768 , lowerCAmelCase=12 , lowerCAmelCase=12 , lowerCAmelCase=3072 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=1e-1_2 , lowerCAmelCase=1 , lowerCAmelCase=0 , lowerCAmelCase=2 , lowerCAmelCase="absolute" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__(pad_token_id=lowerCAmelCase , bos_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = position_embedding_type UpperCAmelCase_ = use_cache UpperCAmelCase_ = classifier_dropout class lowerCamelCase ( lowercase__ ): '''simple docstring''' @property def A__ ( self ): if self.task == "multiple-choice": UpperCAmelCase_ = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
23
1
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @add_end_docstrings(lowercase__ ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , **lowerCAmelCase ): super().__init__(**lowerCAmelCase ) if self.framework == "tf": raise ValueError(f'''The {self.__class__} is only available in PyTorch.''' ) requires_backends(self , "vision" ) self.check_model_type(lowerCAmelCase ) def __call__( self , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): if "text_queries" in kwargs: UpperCAmelCase_ = kwargs.pop("text_queries" ) if isinstance(lowerCAmelCase , (str, Image.Image) ): UpperCAmelCase_ = {"image": image, "candidate_labels": candidate_labels} else: UpperCAmelCase_ = image UpperCAmelCase_ = super().__call__(lowerCAmelCase , **lowerCAmelCase ) return results def A__ ( self , **lowerCAmelCase ): UpperCAmelCase_ = {} if "threshold" in kwargs: UpperCAmelCase_ = kwargs["threshold"] if "top_k" in kwargs: UpperCAmelCase_ = kwargs["top_k"] return {}, {}, postprocess_params def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = load_image(inputs["image"] ) UpperCAmelCase_ = inputs["candidate_labels"] if isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = candidate_labels.split("," ) UpperCAmelCase_ = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(lowerCAmelCase ): UpperCAmelCase_ = self.tokenizer(lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = self.image_processor(lowerCAmelCase , return_tensors=self.framework ) yield { "is_last": i == len(lowerCAmelCase ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = model_inputs.pop("target_size" ) UpperCAmelCase_ = model_inputs.pop("candidate_label" ) UpperCAmelCase_ = model_inputs.pop("is_last" ) UpperCAmelCase_ = self.model(**lowerCAmelCase ) UpperCAmelCase_ = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def A__ ( self , lowerCAmelCase , lowerCAmelCase=0.1 , lowerCAmelCase=None ): UpperCAmelCase_ = [] for model_output in model_outputs: UpperCAmelCase_ = model_output["candidate_label"] UpperCAmelCase_ = BaseModelOutput(lowerCAmelCase ) UpperCAmelCase_ = self.image_processor.post_process_object_detection( outputs=lowerCAmelCase , threshold=lowerCAmelCase , target_sizes=model_output["target_size"] )[0] for index in outputs["scores"].nonzero(): UpperCAmelCase_ = outputs["scores"][index].item() UpperCAmelCase_ = self._get_bounding_box(outputs["boxes"][index][0] ) UpperCAmelCase_ = {"score": score, "label": label, "box": box} results.append(lowerCAmelCase ) UpperCAmelCase_ = sorted(lowerCAmelCase , key=lambda lowerCAmelCase : x["score"] , reverse=lowerCAmelCase ) if top_k: UpperCAmelCase_ = results[:top_k] return results def A__ ( self , lowerCAmelCase ): if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch." ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = box.int().tolist() UpperCAmelCase_ = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = int(__SCREAMING_SNAKE_CASE ) if decimal in (0, 1): # Exit cases for the recursion return str(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ , UpperCAmelCase_ = divmod(__SCREAMING_SNAKE_CASE , 2 ) return binary_recursive(__SCREAMING_SNAKE_CASE ) + str(__SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = str(__SCREAMING_SNAKE_CASE ).strip() if not number: raise ValueError("No input value was provided" ) UpperCAmelCase_ = "-" if number.startswith("-" ) else "" UpperCAmelCase_ = number.lstrip("-" ) if not number.isnumeric(): raise ValueError("Input value is not an integer" ) return f'''{negative}0b{binary_recursive(int(__SCREAMING_SNAKE_CASE ) )}''' if __name__ == "__main__": from doctest import testmod testmod()
23
1
from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "uw-madison/mra-base-512-4": "https://huggingface.co/uw-madison/mra-base-512-4/resolve/main/config.json", } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = 'mra' def __init__( self , lowerCAmelCase=5_0265 , lowerCAmelCase=768 , lowerCAmelCase=12 , lowerCAmelCase=12 , lowerCAmelCase=3072 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=1 , lowerCAmelCase=0.02 , lowerCAmelCase=1e-5 , lowerCAmelCase="absolute" , lowerCAmelCase=4 , lowerCAmelCase="full" , lowerCAmelCase=0 , lowerCAmelCase=0 , lowerCAmelCase=1 , lowerCAmelCase=0 , lowerCAmelCase=2 , **lowerCAmelCase , ): super().__init__(pad_token_id=lowerCAmelCase , bos_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = initializer_range UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = position_embedding_type UpperCAmelCase_ = block_per_row UpperCAmelCase_ = approx_mode UpperCAmelCase_ = initial_prior_first_n_blocks UpperCAmelCase_ = initial_prior_diagonal_n_blocks
23
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self ): UpperCAmelCase_ = AutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" , return_dict=lowerCAmelCase ).to(lowerCAmelCase ) UpperCAmelCase_ = AutoTokenizer.from_pretrained("google/mt5-small" ) UpperCAmelCase_ = tokenizer("Hello there" , return_tensors="pt" ).input_ids UpperCAmelCase_ = tokenizer("Hi I am" , return_tensors="pt" ).input_ids UpperCAmelCase_ = model(input_ids.to(lowerCAmelCase ) , labels=labels.to(lowerCAmelCase ) ).loss UpperCAmelCase_ = -(labels.shape[-1] * loss.item()) UpperCAmelCase_ = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
23
1
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase ( lowercase__, lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = AltDiffusionPipeline lowerCAmelCase_ : Optional[int] = TEXT_TO_IMAGE_PARAMS lowerCAmelCase_ : Tuple = TEXT_TO_IMAGE_BATCH_PARAMS lowerCAmelCase_ : Dict = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCAmelCase_ : Dict = TEXT_TO_IMAGE_IMAGE_PARAMS def A__ ( self ): torch.manual_seed(0 ) UpperCAmelCase_ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) UpperCAmelCase_ = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowerCAmelCase , set_alpha_to_one=lowerCAmelCase , ) torch.manual_seed(0 ) UpperCAmelCase_ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) UpperCAmelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5002 , ) UpperCAmelCase_ = CLIPTextModel(lowerCAmelCase ) UpperCAmelCase_ = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) UpperCAmelCase_ = 77 UpperCAmelCase_ = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def A__ ( self , lowerCAmelCase , lowerCAmelCase=0 ): if str(lowerCAmelCase ).startswith("mps" ): UpperCAmelCase_ = torch.manual_seed(lowerCAmelCase ) else: UpperCAmelCase_ = torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) UpperCAmelCase_ = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def A__ ( self ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def A__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def A__ ( self ): UpperCAmelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ = self.get_dummy_components() torch.manual_seed(0 ) UpperCAmelCase_ = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase_ = RobertaSeriesModelWithTransformation(lowerCAmelCase ) UpperCAmelCase_ = text_encoder UpperCAmelCase_ = AltDiffusionPipeline(**lowerCAmelCase ) UpperCAmelCase_ = alt_pipe.to(lowerCAmelCase ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase ) UpperCAmelCase_ = self.get_dummy_inputs(lowerCAmelCase ) UpperCAmelCase_ = "A photo of an astronaut" UpperCAmelCase_ = alt_pipe(**lowerCAmelCase ) UpperCAmelCase_ = output.images UpperCAmelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ = np.array( [0.5748162, 0.60447145, 0.48821217, 0.50100636, 0.5431185, 0.45763683, 0.49657696, 0.48132733, 0.47573093] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def A__ ( self ): UpperCAmelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ = self.get_dummy_components() UpperCAmelCase_ = PNDMScheduler(skip_prk_steps=lowerCAmelCase ) torch.manual_seed(0 ) UpperCAmelCase_ = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase_ = RobertaSeriesModelWithTransformation(lowerCAmelCase ) UpperCAmelCase_ = text_encoder UpperCAmelCase_ = AltDiffusionPipeline(**lowerCAmelCase ) UpperCAmelCase_ = alt_pipe.to(lowerCAmelCase ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase ) UpperCAmelCase_ = self.get_dummy_inputs(lowerCAmelCase ) UpperCAmelCase_ = alt_pipe(**lowerCAmelCase ) UpperCAmelCase_ = output.images UpperCAmelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ = np.array( [0.51605093, 0.5707241, 0.47365507, 0.50578886, 0.5633877, 0.4642503, 0.5182081, 0.48763484, 0.49084237] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def A__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A__ ( self ): # make sure here that pndm scheduler skips prk UpperCAmelCase_ = AltDiffusionPipeline.from_pretrained("BAAI/AltDiffusion" , safety_checker=lowerCAmelCase ) UpperCAmelCase_ = alt_pipe.to(lowerCAmelCase ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase ) UpperCAmelCase_ = "A painting of a squirrel eating a burger" UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = alt_pipe([prompt] , generator=lowerCAmelCase , guidance_scale=6.0 , num_inference_steps=20 , output_type="np" ) UpperCAmelCase_ = output.images UpperCAmelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase_ = np.array([0.1010, 0.0800, 0.0794, 0.0885, 0.0843, 0.0762, 0.0769, 0.0729, 0.0586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def A__ ( self ): UpperCAmelCase_ = DDIMScheduler.from_pretrained("BAAI/AltDiffusion" , subfolder="scheduler" ) UpperCAmelCase_ = AltDiffusionPipeline.from_pretrained("BAAI/AltDiffusion" , scheduler=lowerCAmelCase , safety_checker=lowerCAmelCase ) UpperCAmelCase_ = alt_pipe.to(lowerCAmelCase ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase ) UpperCAmelCase_ = "A painting of a squirrel eating a burger" UpperCAmelCase_ = torch.manual_seed(0 ) UpperCAmelCase_ = alt_pipe([prompt] , generator=lowerCAmelCase , num_inference_steps=2 , output_type="numpy" ) UpperCAmelCase_ = output.images UpperCAmelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase_ = np.array([0.4019, 0.4052, 0.3810, 0.4119, 0.3916, 0.3982, 0.4651, 0.4195, 0.5323] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
23
def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ = 0 while b > 0: if b & 1: UpperCAmelCase_ = ((res % c) + (a % c)) % c a += a b >>= 1 return res
23
1
import torch from torch import nn class lowerCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=1 , lowerCAmelCase=False ): super().__init__() UpperCAmelCase_ = n_token UpperCAmelCase_ = d_embed UpperCAmelCase_ = d_proj UpperCAmelCase_ = cutoffs + [n_token] UpperCAmelCase_ = [0] + self.cutoffs UpperCAmelCase_ = div_val UpperCAmelCase_ = self.cutoffs[0] UpperCAmelCase_ = len(self.cutoffs ) - 1 UpperCAmelCase_ = self.shortlist_size + self.n_clusters if self.n_clusters > 0: UpperCAmelCase_ = nn.Parameter(torch.zeros(self.n_clusters , self.d_embed ) ) UpperCAmelCase_ = nn.Parameter(torch.zeros(self.n_clusters ) ) UpperCAmelCase_ = nn.ModuleList() UpperCAmelCase_ = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs ) ): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(lowerCAmelCase , lowerCAmelCase ) ) ) else: self.out_projs.append(lowerCAmelCase ) self.out_layers.append(nn.Linear(lowerCAmelCase , lowerCAmelCase ) ) else: for i in range(len(self.cutoffs ) ): UpperCAmelCase_ , UpperCAmelCase_ = self.cutoff_ends[i], self.cutoff_ends[i + 1] UpperCAmelCase_ = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(lowerCAmelCase , lowerCAmelCase ) ) ) self.out_layers.append(nn.Linear(lowerCAmelCase , r_idx - l_idx ) ) UpperCAmelCase_ = keep_order def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if proj is None: UpperCAmelCase_ = nn.functional.linear(lowerCAmelCase , lowerCAmelCase , bias=lowerCAmelCase ) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: UpperCAmelCase_ = nn.functional.linear(lowerCAmelCase , proj.t().contiguous() ) UpperCAmelCase_ = nn.functional.linear(lowerCAmelCase , lowerCAmelCase , bias=lowerCAmelCase ) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def A__ ( self , lowerCAmelCase , lowerCAmelCase=None , lowerCAmelCase=False ): if labels is not None: # Shift so that tokens < n predict n UpperCAmelCase_ = hidden[..., :-1, :].contiguous() UpperCAmelCase_ = labels[..., 1:].contiguous() UpperCAmelCase_ = hidden.view(-1 , hidden.size(-1 ) ) UpperCAmelCase_ = labels.view(-1 ) if hidden.size(0 ) != labels.size(0 ): raise RuntimeError("Input and labels should have the same size in the batch dimension." ) else: UpperCAmelCase_ = hidden.view(-1 , hidden.size(-1 ) ) if self.n_clusters == 0: UpperCAmelCase_ = self._compute_logit(lowerCAmelCase , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) if labels is not None: UpperCAmelCase_ = labels != -100 UpperCAmelCase_ = torch.zeros_like(lowerCAmelCase , dtype=hidden.dtype , device=hidden.device ) UpperCAmelCase_ = ( -nn.functional.log_softmax(lowerCAmelCase , dim=-1 )[mask].gather(1 , labels[mask].unsqueeze(1 ) ).squeeze(1 ) ) else: UpperCAmelCase_ = nn.functional.log_softmax(lowerCAmelCase , dim=-1 ) else: # construct weights and biases UpperCAmelCase_ , UpperCAmelCase_ = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: UpperCAmelCase_ , UpperCAmelCase_ = self.cutoff_ends[i], self.cutoff_ends[i + 1] UpperCAmelCase_ = self.out_layers[0].weight[l_idx:r_idx] UpperCAmelCase_ = self.out_layers[0].bias[l_idx:r_idx] else: UpperCAmelCase_ = self.out_layers[i].weight UpperCAmelCase_ = self.out_layers[i].bias if i == 0: UpperCAmelCase_ = torch.cat([weight_i, self.cluster_weight] , dim=0 ) UpperCAmelCase_ = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(lowerCAmelCase ) biases.append(lowerCAmelCase ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = weights[0], biases[0], self.out_projs[0] UpperCAmelCase_ = self._compute_logit(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = nn.functional.log_softmax(lowerCAmelCase , dim=1 ) if labels is None: UpperCAmelCase_ = hidden.new_empty((head_logit.size(0 ), self.n_token) ) else: UpperCAmelCase_ = torch.zeros_like(lowerCAmelCase , dtype=hidden.dtype , device=hidden.device ) UpperCAmelCase_ = 0 UpperCAmelCase_ = [0] + self.cutoffs for i in range(len(lowerCAmelCase ) - 1 ): UpperCAmelCase_ , UpperCAmelCase_ = cutoff_values[i], cutoff_values[i + 1] if labels is not None: UpperCAmelCase_ = (labels >= l_idx) & (labels < r_idx) UpperCAmelCase_ = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue UpperCAmelCase_ = labels.index_select(0 , lowerCAmelCase ) - l_idx UpperCAmelCase_ = head_logprob.index_select(0 , lowerCAmelCase ) UpperCAmelCase_ = hidden.index_select(0 , lowerCAmelCase ) else: UpperCAmelCase_ = hidden if i == 0: if labels is not None: UpperCAmelCase_ = head_logprob_i.gather(1 , target_i[:, None] ).squeeze(1 ) else: UpperCAmelCase_ = head_logprob[:, : self.cutoffs[0]] else: UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = weights[i], biases[i], self.out_projs[i] UpperCAmelCase_ = self._compute_logit(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = nn.functional.log_softmax(lowerCAmelCase , dim=1 ) UpperCAmelCase_ = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: UpperCAmelCase_ = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None] ).squeeze(1 ) else: UpperCAmelCase_ = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i UpperCAmelCase_ = logprob_i if labels is not None: if (hasattr(self , "keep_order" ) and self.keep_order) or keep_order: out.index_copy_(0 , lowerCAmelCase , -logprob_i ) else: out[offset : offset + logprob_i.size(0 )].copy_(-logprob_i ) offset += logprob_i.size(0 ) return out def A__ ( self , lowerCAmelCase ): if self.n_clusters == 0: UpperCAmelCase_ = self._compute_logit(lowerCAmelCase , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) return nn.functional.log_softmax(lowerCAmelCase , dim=-1 ) else: # construct weights and biases UpperCAmelCase_ , UpperCAmelCase_ = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: UpperCAmelCase_ , UpperCAmelCase_ = self.cutoff_ends[i], self.cutoff_ends[i + 1] UpperCAmelCase_ = self.out_layers[0].weight[l_idx:r_idx] UpperCAmelCase_ = self.out_layers[0].bias[l_idx:r_idx] else: UpperCAmelCase_ = self.out_layers[i].weight UpperCAmelCase_ = self.out_layers[i].bias if i == 0: UpperCAmelCase_ = torch.cat([weight_i, self.cluster_weight] , dim=0 ) UpperCAmelCase_ = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(lowerCAmelCase ) biases.append(lowerCAmelCase ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = weights[0], biases[0], self.out_projs[0] UpperCAmelCase_ = self._compute_logit(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = hidden.new_empty((head_logit.size(0 ), self.n_token) ) UpperCAmelCase_ = nn.functional.log_softmax(lowerCAmelCase , dim=1 ) UpperCAmelCase_ = [0] + self.cutoffs for i in range(len(lowerCAmelCase ) - 1 ): UpperCAmelCase_ , UpperCAmelCase_ = cutoff_values[i], cutoff_values[i + 1] if i == 0: UpperCAmelCase_ = head_logprob[:, : self.cutoffs[0]] else: UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = weights[i], biases[i], self.out_projs[i] UpperCAmelCase_ = self._compute_logit(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = nn.functional.log_softmax(lowerCAmelCase , dim=1 ) UpperCAmelCase_ = head_logprob[:, -i] + tail_logprob_i UpperCAmelCase_ = logprob_i return out
23
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> List[List[ImageInput]]: if isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(__SCREAMING_SNAKE_CASE ): return [[videos]] raise ValueError(f'''Could not make batched video from {videos}''' ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[Any] = ['pixel_values'] def __init__( self , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = True , lowerCAmelCase = 1 / 255 , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = None , **lowerCAmelCase , ): super().__init__(**lowerCAmelCase ) UpperCAmelCase_ = size if size is not None else {"shortest_edge": 224} UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) UpperCAmelCase_ = crop_size if crop_size is not None else {"height": 224, "width": 224} UpperCAmelCase_ = get_size_dict(lowerCAmelCase , param_name="crop_size" ) UpperCAmelCase_ = do_resize UpperCAmelCase_ = size UpperCAmelCase_ = do_center_crop UpperCAmelCase_ = crop_size UpperCAmelCase_ = resample UpperCAmelCase_ = do_rescale UpperCAmelCase_ = rescale_factor UpperCAmelCase_ = do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) if "shortest_edge" in size: UpperCAmelCase_ = get_resize_output_image_size(lowerCAmelCase , size["shortest_edge"] , default_to_square=lowerCAmelCase ) elif "height" in size and "width" in size: UpperCAmelCase_ = (size["height"], size["width"]) else: raise ValueError(f'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = get_size_dict(lowerCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(lowerCAmelCase , size=(size["height"], size["width"]) , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return rescale(lowerCAmelCase , scale=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return normalize(lowerCAmelCase , mean=lowerCAmelCase , std=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , ): if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. UpperCAmelCase_ = to_numpy_array(lowerCAmelCase ) if do_resize: UpperCAmelCase_ = self.resize(image=lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase ) if do_center_crop: UpperCAmelCase_ = self.center_crop(lowerCAmelCase , size=lowerCAmelCase ) if do_rescale: UpperCAmelCase_ = self.rescale(image=lowerCAmelCase , scale=lowerCAmelCase ) if do_normalize: UpperCAmelCase_ = self.normalize(image=lowerCAmelCase , mean=lowerCAmelCase , std=lowerCAmelCase ) UpperCAmelCase_ = to_channel_dimension_format(lowerCAmelCase , lowerCAmelCase ) return image def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , **lowerCAmelCase , ): UpperCAmelCase_ = do_resize if do_resize is not None else self.do_resize UpperCAmelCase_ = resample if resample is not None else self.resample UpperCAmelCase_ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase_ = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase_ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase_ = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else self.image_mean UpperCAmelCase_ = image_std if image_std is not None else self.image_std UpperCAmelCase_ = size if size is not None else self.size UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) UpperCAmelCase_ = crop_size if crop_size is not None else self.crop_size UpperCAmelCase_ = get_size_dict(lowerCAmelCase , param_name="crop_size" ) if not valid_images(lowerCAmelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) UpperCAmelCase_ = make_batched(lowerCAmelCase ) UpperCAmelCase_ = [ [ self._preprocess_image( image=lowerCAmelCase , do_resize=lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase , do_center_crop=lowerCAmelCase , crop_size=lowerCAmelCase , do_rescale=lowerCAmelCase , rescale_factor=lowerCAmelCase , do_normalize=lowerCAmelCase , image_mean=lowerCAmelCase , image_std=lowerCAmelCase , data_format=lowerCAmelCase , ) for img in video ] for video in videos ] UpperCAmelCase_ = {"pixel_values": videos} return BatchFeature(data=lowerCAmelCase , tensor_type=lowerCAmelCase )
23
1
from __future__ import annotations import math import random from typing import Any class lowerCamelCase : '''simple docstring''' def __init__( self ): UpperCAmelCase_ = [] UpperCAmelCase_ = 0 UpperCAmelCase_ = 0 def A__ ( self ): return self.head == self.tail def A__ ( self , lowerCAmelCase ): self.data.append(lowerCAmelCase ) UpperCAmelCase_ = self.tail + 1 def A__ ( self ): UpperCAmelCase_ = self.data[self.head] UpperCAmelCase_ = self.head + 1 return ret def A__ ( self ): return self.tail - self.head def A__ ( self ): print(self.data ) print("**************" ) print(self.data[self.head : self.tail] ) class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase ): UpperCAmelCase_ = data UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = 1 def A__ ( self ): return self.data def A__ ( self ): return self.left def A__ ( self ): return self.right def A__ ( self ): return self.height def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = data def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = node def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = node def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = height def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: if node is None: return 0 return node.get_height() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> int: if a > b: return a return b def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> MyNode: print("left rotation node:" , node.get_data() ) UpperCAmelCase_ = node.get_left() assert ret is not None node.set_left(ret.get_right() ) ret.set_right(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(__SCREAMING_SNAKE_CASE ) return ret def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> MyNode: print("right rotation node:" , node.get_data() ) UpperCAmelCase_ = node.get_right() assert ret is not None node.set_right(ret.get_left() ) ret.set_left(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(__SCREAMING_SNAKE_CASE ) return ret def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> MyNode: UpperCAmelCase_ = node.get_left() assert left_child is not None node.set_left(left_rotation(__SCREAMING_SNAKE_CASE ) ) return right_rotation(__SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> MyNode: UpperCAmelCase_ = node.get_right() assert right_child is not None node.set_right(right_rotation(__SCREAMING_SNAKE_CASE ) ) return left_rotation(__SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> MyNode | None: if node is None: return MyNode(__SCREAMING_SNAKE_CASE ) if data < node.get_data(): node.set_left(insert_node(node.get_left() , __SCREAMING_SNAKE_CASE ) ) if ( get_height(node.get_left() ) - get_height(node.get_right() ) == 2 ): # an unbalance detected UpperCAmelCase_ = node.get_left() assert left_child is not None if ( data < left_child.get_data() ): # new node is the left child of the left child UpperCAmelCase_ = right_rotation(__SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ = lr_rotation(__SCREAMING_SNAKE_CASE ) else: node.set_right(insert_node(node.get_right() , __SCREAMING_SNAKE_CASE ) ) if get_height(node.get_right() ) - get_height(node.get_left() ) == 2: UpperCAmelCase_ = node.get_right() assert right_child is not None if data < right_child.get_data(): UpperCAmelCase_ = rl_rotation(__SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ = left_rotation(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(__SCREAMING_SNAKE_CASE ) return node def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Any: while True: UpperCAmelCase_ = root.get_right() if right_child is None: break UpperCAmelCase_ = right_child return root.get_data() def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Any: while True: UpperCAmelCase_ = root.get_left() if left_child is None: break UpperCAmelCase_ = left_child return root.get_data() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> MyNode | None: UpperCAmelCase_ = root.get_left() UpperCAmelCase_ = root.get_right() if root.get_data() == data: if left_child is not None and right_child is not None: UpperCAmelCase_ = get_left_most(__SCREAMING_SNAKE_CASE ) root.set_data(__SCREAMING_SNAKE_CASE ) root.set_right(del_node(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) elif left_child is not None: UpperCAmelCase_ = left_child elif right_child is not None: UpperCAmelCase_ = right_child else: return None elif root.get_data() > data: if left_child is None: print("No such data" ) return root else: root.set_left(del_node(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) else: # root.get_data() < data if right_child is None: return root else: root.set_right(del_node(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) if get_height(__SCREAMING_SNAKE_CASE ) - get_height(__SCREAMING_SNAKE_CASE ) == 2: assert right_child is not None if get_height(right_child.get_right() ) > get_height(right_child.get_left() ): UpperCAmelCase_ = left_rotation(__SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ = rl_rotation(__SCREAMING_SNAKE_CASE ) elif get_height(__SCREAMING_SNAKE_CASE ) - get_height(__SCREAMING_SNAKE_CASE ) == -2: assert left_child is not None if get_height(left_child.get_left() ) > get_height(left_child.get_right() ): UpperCAmelCase_ = right_rotation(__SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ = lr_rotation(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = my_max(get_height(root.get_right() ) , get_height(root.get_left() ) ) + 1 root.set_height(__SCREAMING_SNAKE_CASE ) return root class lowerCamelCase : '''simple docstring''' def __init__( self ): UpperCAmelCase_ = None def A__ ( self ): return get_height(self.root ) def A__ ( self , lowerCAmelCase ): print("insert:" + str(lowerCAmelCase ) ) UpperCAmelCase_ = insert_node(self.root , lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): print("delete:" + str(lowerCAmelCase ) ) if self.root is None: print("Tree is empty!" ) return UpperCAmelCase_ = del_node(self.root , lowerCAmelCase ) def __str__( self , ): # a level traversale, gives a more intuitive look on the tree UpperCAmelCase_ = "" UpperCAmelCase_ = MyQueue() q.push(self.root ) UpperCAmelCase_ = self.get_height() if layer == 0: return output UpperCAmelCase_ = 0 while not q.is_empty(): UpperCAmelCase_ = q.pop() UpperCAmelCase_ = " " * int(math.pow(2 , layer - 1 ) ) output += space if node is None: output += "*" q.push(lowerCAmelCase ) q.push(lowerCAmelCase ) else: output += str(node.get_data() ) q.push(node.get_left() ) q.push(node.get_right() ) output += space UpperCAmelCase_ = cnt + 1 for i in range(100 ): if cnt == math.pow(2 , lowerCAmelCase ) - 1: UpperCAmelCase_ = layer - 1 if layer == 0: output += "\n*************************************" return output output += "\n" break output += "\n*************************************" return output def snake_case__ ( ) -> None: import doctest doctest.testmod() if __name__ == "__main__": _test() SCREAMING_SNAKE_CASE = AVLtree() SCREAMING_SNAKE_CASE = list(range(10)) random.shuffle(lst) for i in lst: t.insert(i) print(str(t)) random.shuffle(lst) for i in lst: t.del_node(i) print(str(t))
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 1 for i in range(1 , num + 1 ): fact *= i return fact def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 0 while number > 0: UpperCAmelCase_ = number % 10 sum_of_digits += last_digit UpperCAmelCase_ = number // 10 # Removing the last_digit from the given number return sum_of_digits def snake_case__ ( __SCREAMING_SNAKE_CASE = 100 ) -> int: UpperCAmelCase_ = factorial(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = split_and_add(__SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
23
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "shi-labs/nat-mini-in1k-224": "https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json", # See all Nat models at https://huggingface.co/models?filter=nat } class lowerCamelCase ( lowercase__, lowercase__ ): '''simple docstring''' lowerCAmelCase_ : str = 'nat' lowerCAmelCase_ : Optional[Any] = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , lowerCAmelCase=4 , lowerCAmelCase=3 , lowerCAmelCase=64 , lowerCAmelCase=[3, 4, 6, 5] , lowerCAmelCase=[2, 4, 8, 16] , lowerCAmelCase=7 , lowerCAmelCase=3.0 , lowerCAmelCase=True , lowerCAmelCase=0.0 , lowerCAmelCase=0.0 , lowerCAmelCase=0.1 , lowerCAmelCase="gelu" , lowerCAmelCase=0.02 , lowerCAmelCase=1e-5 , lowerCAmelCase=0.0 , lowerCAmelCase=None , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__(**lowerCAmelCase ) UpperCAmelCase_ = patch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = embed_dim UpperCAmelCase_ = depths UpperCAmelCase_ = len(lowerCAmelCase ) UpperCAmelCase_ = num_heads UpperCAmelCase_ = kernel_size UpperCAmelCase_ = mlp_ratio UpperCAmelCase_ = qkv_bias UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = drop_path_rate UpperCAmelCase_ = hidden_act UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model UpperCAmelCase_ = int(embed_dim * 2 ** (len(lowerCAmelCase ) - 1) ) UpperCAmelCase_ = layer_scale_init_value UpperCAmelCase_ = ["stem"] + [f'''stage{idx}''' for idx in range(1 , len(lowerCAmelCase ) + 1 )] UpperCAmelCase_ , UpperCAmelCase_ = get_aligned_output_features_output_indices( out_features=lowerCAmelCase , out_indices=lowerCAmelCase , stage_names=self.stage_names )
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_xlnet": ["XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["XLNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["XLNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "XLNetForMultipleChoice", "XLNetForQuestionAnswering", "XLNetForQuestionAnsweringSimple", "XLNetForSequenceClassification", "XLNetForTokenClassification", "XLNetLMHeadModel", "XLNetModel", "XLNetPreTrainedModel", "load_tf_weights_in_xlnet", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLNetForMultipleChoice", "TFXLNetForQuestionAnsweringSimple", "TFXLNetForSequenceClassification", "TFXLNetForTokenClassification", "TFXLNetLMHeadModel", "TFXLNetMainLayer", "TFXLNetModel", "TFXLNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available SCREAMING_SNAKE_CASE = { "configuration_audio_spectrogram_transformer": [ "AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "ASTConfig", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "ASTForAudioClassification", "ASTModel", "ASTPreTrainedModel", ] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["ASTFeatureExtractor"] if TYPE_CHECKING: from .configuration_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ASTConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ASTForAudioClassification, ASTModel, ASTPreTrainedModel, ) try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_audio_spectrogram_transformer import ASTFeatureExtractor else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[Any]: # Initialise PyTorch model UpperCAmelCase_ = MobileBertConfig.from_json_file(__SCREAMING_SNAKE_CASE ) print(f'''Building PyTorch model from configuration: {config}''' ) UpperCAmelCase_ = MobileBertForPreTraining(__SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint UpperCAmelCase_ = load_tf_weights_in_mobilebert(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , __SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--mobilebert_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained MobileBERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
23
1
import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse("1.0.0a"): raise Exception("requires fairseq >= 1.0.0a") logging.set_verbosity_info() SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = "Hello world! cécé herlolip" def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Dict: UpperCAmelCase_ = FairseqRobertaModel.from_pretrained(__SCREAMING_SNAKE_CASE ) roberta.eval() # disable dropout UpperCAmelCase_ = roberta.model.encoder.sentence_encoder UpperCAmelCase_ = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings , hidden_size=roberta.cfg.model.encoder_embed_dim , num_hidden_layers=roberta.cfg.model.encoder_layers , num_attention_heads=roberta.cfg.model.encoder_attention_heads , intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=514 , type_vocab_size=1 , layer_norm_eps=1E-5 , ) if classification_head: UpperCAmelCase_ = roberta.model.classification_heads["mnli"].out_proj.weight.shape[0] print("Our RoBERTa config:" , __SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = XLMRobertaXLForSequenceClassification(__SCREAMING_SNAKE_CASE ) if classification_head else XLMRobertaXLForMaskedLM(__SCREAMING_SNAKE_CASE ) model.eval() # Now let's copy all the weights. # Embeddings UpperCAmelCase_ = roberta_sent_encoder.embed_tokens.weight UpperCAmelCase_ = roberta_sent_encoder.embed_positions.weight UpperCAmelCase_ = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. UpperCAmelCase_ = roberta_sent_encoder.layer_norm.weight UpperCAmelCase_ = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer UpperCAmelCase_ = model.roberta.encoder.layer[i] UpperCAmelCase_ = roberta_sent_encoder.layers[i] UpperCAmelCase_ = layer.attention UpperCAmelCase_ = roberta_layer.self_attn_layer_norm.weight UpperCAmelCase_ = roberta_layer.self_attn_layer_norm.bias # self attention UpperCAmelCase_ = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) UpperCAmelCase_ = roberta_layer.self_attn.q_proj.weight UpperCAmelCase_ = roberta_layer.self_attn.q_proj.bias UpperCAmelCase_ = roberta_layer.self_attn.k_proj.weight UpperCAmelCase_ = roberta_layer.self_attn.k_proj.bias UpperCAmelCase_ = roberta_layer.self_attn.v_proj.weight UpperCAmelCase_ = roberta_layer.self_attn.v_proj.bias # self-attention output UpperCAmelCase_ = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape UpperCAmelCase_ = roberta_layer.self_attn.out_proj.weight UpperCAmelCase_ = roberta_layer.self_attn.out_proj.bias # this one is final layer norm UpperCAmelCase_ = roberta_layer.final_layer_norm.weight UpperCAmelCase_ = roberta_layer.final_layer_norm.bias # intermediate UpperCAmelCase_ = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape UpperCAmelCase_ = roberta_layer.fca.weight UpperCAmelCase_ = roberta_layer.fca.bias # output UpperCAmelCase_ = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape UpperCAmelCase_ = roberta_layer.fca.weight UpperCAmelCase_ = roberta_layer.fca.bias # end of layer if classification_head: UpperCAmelCase_ = roberta.model.classification_heads["mnli"].dense.weight UpperCAmelCase_ = roberta.model.classification_heads["mnli"].dense.bias UpperCAmelCase_ = roberta.model.classification_heads["mnli"].out_proj.weight UpperCAmelCase_ = roberta.model.classification_heads["mnli"].out_proj.bias else: # LM Head UpperCAmelCase_ = roberta.model.encoder.lm_head.dense.weight UpperCAmelCase_ = roberta.model.encoder.lm_head.dense.bias UpperCAmelCase_ = roberta.model.encoder.lm_head.layer_norm.weight UpperCAmelCase_ = roberta.model.encoder.lm_head.layer_norm.bias UpperCAmelCase_ = roberta.model.encoder.lm_head.weight UpperCAmelCase_ = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. UpperCAmelCase_ = roberta.encode(__SCREAMING_SNAKE_CASE ).unsqueeze(0 ) # batch of size 1 UpperCAmelCase_ = model(__SCREAMING_SNAKE_CASE )[0] if classification_head: UpperCAmelCase_ = roberta.model.classification_heads["mnli"](roberta.extract_features(__SCREAMING_SNAKE_CASE ) ) else: UpperCAmelCase_ = roberta.model(__SCREAMING_SNAKE_CASE )[0] print(our_output.shape , their_output.shape ) UpperCAmelCase_ = torch.max(torch.abs(our_output - their_output ) ).item() print(f'''max_absolute_diff = {max_absolute_diff}''' ) # ~ 1e-7 UpperCAmelCase_ = torch.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1E-3 ) print("Do both models output the same tensors?" , "🔥" if success else "💩" ) if not success: raise Exception("Something went wRoNg" ) pathlib.Path(__SCREAMING_SNAKE_CASE ).mkdir(parents=__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--roberta_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--classification_head", action="store_true", help="Whether to convert a final classification head." ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
23
import heapq as hq import math from collections.abc import Iterator class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase ): UpperCAmelCase_ = str(id_ ) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = [] UpperCAmelCase_ = {} # {vertex:distance} def __lt__( self , lowerCAmelCase ): return self.key < other.key def __repr__( self ): return self.id def A__ ( self , lowerCAmelCase ): self.neighbors.append(lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = weight def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __SCREAMING_SNAKE_CASE ) graph[b - 1].add_edge(graph[a - 1] , __SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> list: UpperCAmelCase_ = [] for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = graph[:] while q: UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE ) q.remove(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Iterator[tuple]: for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = list(__SCREAMING_SNAKE_CASE ) hq.heapify(__SCREAMING_SNAKE_CASE ) while h: UpperCAmelCase_ = hq.heappop(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] hq.heapify(__SCREAMING_SNAKE_CASE ) for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def snake_case__ ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
23
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self ): UpperCAmelCase_ = AutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" , return_dict=lowerCAmelCase ).to(lowerCAmelCase ) UpperCAmelCase_ = AutoTokenizer.from_pretrained("google/mt5-small" ) UpperCAmelCase_ = tokenizer("Hello there" , return_tensors="pt" ).input_ids UpperCAmelCase_ = tokenizer("Hi I am" , return_tensors="pt" ).input_ids UpperCAmelCase_ = model(input_ids.to(lowerCAmelCase ) , labels=labels.to(lowerCAmelCase ) ).loss UpperCAmelCase_ = -(labels.shape[-1] * loss.item()) UpperCAmelCase_ = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_fnet": ["FNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "FNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["FNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["FNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "FNET_PRETRAINED_MODEL_ARCHIVE_LIST", "FNetForMaskedLM", "FNetForMultipleChoice", "FNetForNextSentencePrediction", "FNetForPreTraining", "FNetForQuestionAnswering", "FNetForSequenceClassification", "FNetForTokenClassification", "FNetLayer", "FNetModel", "FNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = dataset UpperCAmelCase_ = process UpperCAmelCase_ = params def __len__( self ): return len(self.dataset ) def __getitem__( self , lowerCAmelCase ): UpperCAmelCase_ = self.dataset[i] UpperCAmelCase_ = self.process(lowerCAmelCase , **self.params ) return processed class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = loader UpperCAmelCase_ = infer UpperCAmelCase_ = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether UpperCAmelCase_ = None UpperCAmelCase_ = loader_batch_size # Internal bookkeeping UpperCAmelCase_ = None UpperCAmelCase_ = None def __len__( self ): return len(self.loader ) def __iter__( self ): UpperCAmelCase_ = iter(self.loader ) return self def A__ ( self ): if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice UpperCAmelCase_ = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) UpperCAmelCase_ = {} for k, element in self._loader_batch_data.items(): if isinstance(lowerCAmelCase , lowerCAmelCase ): # Convert ModelOutput to tuple first UpperCAmelCase_ = element.to_tuple() if isinstance(element[0] , torch.Tensor ): UpperCAmelCase_ = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): UpperCAmelCase_ = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(lowerCAmelCase , lowerCAmelCase ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): UpperCAmelCase_ = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): UpperCAmelCase_ = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around UpperCAmelCase_ = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers UpperCAmelCase_ = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers UpperCAmelCase_ = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. UpperCAmelCase_ = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 UpperCAmelCase_ = self._loader_batch_data.__class__(lowerCAmelCase ) self._loader_batch_index += 1 return result def A__ ( self ): if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch UpperCAmelCase_ = next(self.iterator ) UpperCAmelCase_ = self.infer(lowerCAmelCase , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(lowerCAmelCase , torch.Tensor ): UpperCAmelCase_ = processed else: UpperCAmelCase_ = list(processed.keys() )[0] UpperCAmelCase_ = processed[key] if isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = len(lowerCAmelCase ) else: UpperCAmelCase_ = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. UpperCAmelCase_ = observed_batch_size # Setting internal index to unwrap the batch UpperCAmelCase_ = processed UpperCAmelCase_ = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=None ): super().__init__(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def __iter__( self ): UpperCAmelCase_ = iter(self.loader ) UpperCAmelCase_ = None return self def A__ ( self ): if self.subiterator is None: UpperCAmelCase_ = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item UpperCAmelCase_ = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators UpperCAmelCase_ = self.infer(next(self.iterator ) , **self.params ) UpperCAmelCase_ = next(self.subiterator ) return processed class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __iter__( self ): UpperCAmelCase_ = iter(self.loader ) return self def A__ ( self ): # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. UpperCAmelCase_ = False UpperCAmelCase_ = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: UpperCAmelCase_ = self.loader_batch_item() UpperCAmelCase_ = item.pop("is_last" ) accumulator.append(lowerCAmelCase ) if is_last: return accumulator while not is_last: UpperCAmelCase_ = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(lowerCAmelCase , torch.Tensor ): UpperCAmelCase_ = processed else: UpperCAmelCase_ = list(processed.keys() )[0] UpperCAmelCase_ = processed[key] if isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = len(lowerCAmelCase ) else: UpperCAmelCase_ = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. UpperCAmelCase_ = observed_batch_size UpperCAmelCase_ = processed UpperCAmelCase_ = 0 while self._loader_batch_index < self.loader_batch_size: UpperCAmelCase_ = self.loader_batch_item() UpperCAmelCase_ = item.pop("is_last" ) accumulator.append(lowerCAmelCase ) if is_last: return accumulator else: UpperCAmelCase_ = processed UpperCAmelCase_ = item.pop("is_last" ) accumulator.append(lowerCAmelCase ) return accumulator class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = dataset UpperCAmelCase_ = key def __len__( self ): return len(self.dataset ) def __getitem__( self , lowerCAmelCase ): return self.dataset[i][self.key] class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = dataset UpperCAmelCase_ = keya UpperCAmelCase_ = keya def __len__( self ): return len(self.dataset ) def __getitem__( self , lowerCAmelCase ): return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
23
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": 512, } SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": {"do_lower_case": True}, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = VOCAB_FILES_NAMES lowerCAmelCase_ : int = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Dict = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : List[str] = RetriBertTokenizer lowerCAmelCase_ : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase="[UNK]" , lowerCAmelCase="[SEP]" , lowerCAmelCase="[PAD]" , lowerCAmelCase="[CLS]" , lowerCAmelCase="[MASK]" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , do_lower_case=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , pad_token=lowerCAmelCase , cls_token=lowerCAmelCase , mask_token=lowerCAmelCase , tokenize_chinese_chars=lowerCAmelCase , strip_accents=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(lowerCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**lowerCAmelCase ) UpperCAmelCase_ = do_lower_case def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "spiece.model"} SCREAMING_SNAKE_CASE = { "vocab_file": { "bert_for_seq_generation": ( "https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model" ), } } SCREAMING_SNAKE_CASE = {"bert_for_seq_generation": 512} class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Any = VOCAB_FILES_NAMES lowerCAmelCase_ : Dict = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : List[int] = [] lowerCAmelCase_ : List[str] = ['input_ids', 'attention_mask'] def __init__( self , lowerCAmelCase , lowerCAmelCase="<s>" , lowerCAmelCase="</s>" , lowerCAmelCase="<unk>" , lowerCAmelCase="<pad>" , lowerCAmelCase="<::::>" , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=lowerCAmelCase , eos_token=lowerCAmelCase , unk_token=lowerCAmelCase , pad_token=lowerCAmelCase , sep_token=lowerCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase , ) UpperCAmelCase_ = vocab_file UpperCAmelCase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCAmelCase ) @property def A__ ( self ): return self.sp_model.get_piece_size() def A__ ( self ): UpperCAmelCase_ = {self.convert_ids_to_tokens(lowerCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): UpperCAmelCase_ = self.__dict__.copy() UpperCAmelCase_ = None return state def __setstate__( self , lowerCAmelCase ): UpperCAmelCase_ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): UpperCAmelCase_ = {} UpperCAmelCase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A__ ( self , lowerCAmelCase ): return self.sp_model.encode(lowerCAmelCase , out_type=lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): return self.sp_model.piece_to_id(lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = self.sp_model.IdToPiece(lowerCAmelCase ) return token def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = [] UpperCAmelCase_ = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowerCAmelCase ) + token UpperCAmelCase_ = [] else: current_sub_tokens.append(lowerCAmelCase ) out_string += self.sp_model.decode(lowerCAmelCase ) return out_string.strip() def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): if not os.path.isdir(lowerCAmelCase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase_ = os.path.join( lowerCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(lowerCAmelCase , "wb" ) as fi: UpperCAmelCase_ = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase ) return (out_vocab_file,)
23
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-german-cased": ( "https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json" ), "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE = { "distilbert-base-uncased": 512, "distilbert-base-uncased-distilled-squad": 512, "distilbert-base-cased": 512, "distilbert-base-cased-distilled-squad": 512, "distilbert-base-german-cased": 512, "distilbert-base-multilingual-cased": 512, } SCREAMING_SNAKE_CASE = { "distilbert-base-uncased": {"do_lower_case": True}, "distilbert-base-uncased-distilled-squad": {"do_lower_case": True}, "distilbert-base-cased": {"do_lower_case": False}, "distilbert-base-cased-distilled-squad": {"do_lower_case": False}, "distilbert-base-german-cased": {"do_lower_case": False}, "distilbert-base-multilingual-cased": {"do_lower_case": False}, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Any = VOCAB_FILES_NAMES lowerCAmelCase_ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Union[str, Any] = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : int = ['input_ids', 'attention_mask'] lowerCAmelCase_ : str = DistilBertTokenizer def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase="[UNK]" , lowerCAmelCase="[SEP]" , lowerCAmelCase="[PAD]" , lowerCAmelCase="[CLS]" , lowerCAmelCase="[MASK]" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , do_lower_case=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , pad_token=lowerCAmelCase , cls_token=lowerCAmelCase , mask_token=lowerCAmelCase , tokenize_chinese_chars=lowerCAmelCase , strip_accents=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(lowerCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**lowerCAmelCase ) UpperCAmelCase_ = do_lower_case def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
1
# # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def snake_case__ ( *__SCREAMING_SNAKE_CASE ) -> Dict: with open(__SCREAMING_SNAKE_CASE , "r" ) as fh: fcntl.flock(__SCREAMING_SNAKE_CASE , fcntl.LOCK_EX ) try: print(*__SCREAMING_SNAKE_CASE ) finally: fcntl.flock(__SCREAMING_SNAKE_CASE , fcntl.LOCK_UN ) SCREAMING_SNAKE_CASE = int(os.environ["LOCAL_RANK"]) torch.cuda.set_device(local_rank) SCREAMING_SNAKE_CASE = torch.device("cuda", local_rank) SCREAMING_SNAKE_CASE = socket.gethostname() SCREAMING_SNAKE_CASE = f'''[{hostname}-{local_rank}]''' try: # test distributed dist.init_process_group("nccl") dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank SCREAMING_SNAKE_CASE = dist.get_rank() SCREAMING_SNAKE_CASE = dist.get_world_size() printflock(f'''{gpu} is OK (global rank: {rank}/{world_size})''') dist.barrier() if rank == 0: printflock(f'''pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}''') except Exception: printflock(f'''{gpu} is broken''') raise
23
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> np.array: UpperCAmelCase_ = f'''{sampling_rate}''' UpperCAmelCase_ = "1" UpperCAmelCase_ = "f32le" UpperCAmelCase_ = [ "ffmpeg", "-i", "pipe:0", "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-hide_banner", "-loglevel", "quiet", "pipe:1", ] try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: UpperCAmelCase_ = ffmpeg_process.communicate(__SCREAMING_SNAKE_CASE ) except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to load audio files from filename" ) from error UpperCAmelCase_ = output_stream[0] UpperCAmelCase_ = np.frombuffer(__SCREAMING_SNAKE_CASE , np.floataa ) if audio.shape[0] == 0: raise ValueError("Malformed soundfile" ) return audio def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = "f32le" , ) -> Dict: UpperCAmelCase_ = f'''{sampling_rate}''' UpperCAmelCase_ = "1" if format_for_conversion == "s16le": UpperCAmelCase_ = 2 elif format_for_conversion == "f32le": UpperCAmelCase_ = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) UpperCAmelCase_ = platform.system() if system == "Linux": UpperCAmelCase_ = "alsa" UpperCAmelCase_ = "default" elif system == "Darwin": UpperCAmelCase_ = "avfoundation" UpperCAmelCase_ = ":0" elif system == "Windows": UpperCAmelCase_ = "dshow" UpperCAmelCase_ = "default" UpperCAmelCase_ = [ "ffmpeg", "-f", format_, "-i", input_, "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-fflags", "nobuffer", "-hide_banner", "-loglevel", "quiet", "pipe:1", ] UpperCAmelCase_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample UpperCAmelCase_ = _ffmpeg_stream(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for item in iterator: yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = "f32le" , ) -> int: if stream_chunk_s is not None: UpperCAmelCase_ = stream_chunk_s else: UpperCAmelCase_ = chunk_length_s UpperCAmelCase_ = ffmpeg_microphone(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , format_for_conversion=__SCREAMING_SNAKE_CASE ) if format_for_conversion == "s16le": UpperCAmelCase_ = np.intaa UpperCAmelCase_ = 2 elif format_for_conversion == "f32le": UpperCAmelCase_ = np.floataa UpperCAmelCase_ = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: UpperCAmelCase_ = chunk_length_s / 6 UpperCAmelCase_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(__SCREAMING_SNAKE_CASE , (int, float) ): UpperCAmelCase_ = [stride_length_s, stride_length_s] UpperCAmelCase_ = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample UpperCAmelCase_ = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample UpperCAmelCase_ = datetime.datetime.now() UpperCAmelCase_ = datetime.timedelta(seconds=__SCREAMING_SNAKE_CASE ) for item in chunk_bytes_iter(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=(stride_left, stride_right) , stream=__SCREAMING_SNAKE_CASE ): # Put everything back in numpy scale UpperCAmelCase_ = np.frombuffer(item["raw"] , dtype=__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = ( item["stride"][0] // size_of_sample, item["stride"][1] // size_of_sample, ) UpperCAmelCase_ = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = False ) -> Dict: UpperCAmelCase_ = B"" UpperCAmelCase_ , UpperCAmelCase_ = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) UpperCAmelCase_ = 0 for raw in iterator: acc += raw if stream and len(__SCREAMING_SNAKE_CASE ) < chunk_len: UpperCAmelCase_ = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(__SCREAMING_SNAKE_CASE ) >= chunk_len: # We are flushing the accumulator UpperCAmelCase_ = (_stride_left, stride_right) UpperCAmelCase_ = {"raw": acc[:chunk_len], "stride": stride} if stream: UpperCAmelCase_ = False yield item UpperCAmelCase_ = stride_left UpperCAmelCase_ = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(__SCREAMING_SNAKE_CASE ) > stride_left: UpperCAmelCase_ = {"raw": acc, "stride": (_stride_left, 0)} if stream: UpperCAmelCase_ = False yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ = 2**24 # 16Mo try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdout=subprocess.PIPE , bufsize=__SCREAMING_SNAKE_CASE ) as ffmpeg_process: while True: UpperCAmelCase_ = ffmpeg_process.stdout.read(__SCREAMING_SNAKE_CASE ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to stream audio files from filename" ) from error
23
1
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def snake_case__ ( __SCREAMING_SNAKE_CASE = 3 ) -> qiskit.result.counts.Counts: if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): raise TypeError("number of qubits must be a integer." ) if number_of_qubits <= 0: raise ValueError("number of qubits must be > 0." ) if math.floor(__SCREAMING_SNAKE_CASE ) != number_of_qubits: raise ValueError("number of qubits must be exact integer." ) if number_of_qubits > 10: raise ValueError("number of qubits too large to simulate(>10)." ) UpperCAmelCase_ = QuantumRegister(__SCREAMING_SNAKE_CASE , "qr" ) UpperCAmelCase_ = ClassicalRegister(__SCREAMING_SNAKE_CASE , "cr" ) UpperCAmelCase_ = QuantumCircuit(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = number_of_qubits for i in range(__SCREAMING_SNAKE_CASE ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(__SCREAMING_SNAKE_CASE ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(__SCREAMING_SNAKE_CASE , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # simulate with 10000 shots UpperCAmelCase_ = Aer.get_backend("qasm_simulator" ) UpperCAmelCase_ = execute(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , shots=1_0000 ) return job.result().get_counts(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print( f'''Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}''' )
23
from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class lowerCamelCase ( lowercase__, lowercase__ ): '''simple docstring''' @register_to_config def __init__( self , lowerCAmelCase = 768 , ): super().__init__() UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(torch.ones(1 , lowerCAmelCase ) ) def A__ ( self , lowerCAmelCase = None , lowerCAmelCase = None , ): UpperCAmelCase_ = nn.Parameter(self.mean.to(lowerCAmelCase ).to(lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(self.std.to(lowerCAmelCase ).to(lowerCAmelCase ) ) return self def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (embeds - self.mean) * 1.0 / self.std return embeds def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (embeds * self.std) + self.mean return embeds
23
1
import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=7 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=99 , lowerCAmelCase=64 , lowerCAmelCase=5 , lowerCAmelCase=4 , lowerCAmelCase=37 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=16 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=3 , lowerCAmelCase=4 , lowerCAmelCase=None , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = seq_length UpperCAmelCase_ = is_training UpperCAmelCase_ = use_input_mask UpperCAmelCase_ = use_token_type_ids UpperCAmelCase_ = use_labels UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = type_sequence_label_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = num_labels UpperCAmelCase_ = num_choices UpperCAmelCase_ = scope UpperCAmelCase_ = vocab_size - 1 def A__ ( self ): UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase_ = None if self.use_input_mask: UpperCAmelCase_ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase_ = self.get_config() return config, input_ids, input_mask, token_labels def A__ ( self ): return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCAmelCase , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.prepare_config_and_inputs() UpperCAmelCase_ = True return config, input_ids, input_mask, token_labels def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = GPTNeoXModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase ) UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = True UpperCAmelCase_ = GPTNeoXModel(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = GPTNeoXForCausalLM(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = GPTNeoXForQuestionAnswering(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = GPTNeoXForSequenceClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = GPTNeoXForTokenClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = True UpperCAmelCase_ = GPTNeoXForCausalLM(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() # first forward pass UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase , use_cache=lowerCAmelCase ) UpperCAmelCase_ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids UpperCAmelCase_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCAmelCase_ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and UpperCAmelCase_ = torch.cat([input_ids, next_tokens] , dim=-1 ) UpperCAmelCase_ = torch.cat([input_mask, next_mask] , dim=-1 ) UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase , output_hidden_states=lowerCAmelCase ) UpperCAmelCase_ = output_from_no_past["hidden_states"][0] UpperCAmelCase_ = model( lowerCAmelCase , attention_mask=lowerCAmelCase , past_key_values=lowerCAmelCase , output_hidden_states=lowerCAmelCase , )["hidden_states"][0] # select random slice UpperCAmelCase_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() UpperCAmelCase_ = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCAmelCase_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCAmelCase , lowerCAmelCase , atol=1e-3 ) ) def A__ ( self ): UpperCAmelCase_ = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = config_and_inputs UpperCAmelCase_ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class lowerCamelCase ( lowercase__, lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : List[Any] = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) lowerCAmelCase_ : str = (GPTNeoXForCausalLM,) if is_torch_available() else () lowerCAmelCase_ : Union[str, Any] = ( { 'feature-extraction': GPTNeoXModel, 'question-answering': GPTNeoXForQuestionAnswering, 'text-classification': GPTNeoXForSequenceClassification, 'text-generation': GPTNeoXForCausalLM, 'token-classification': GPTNeoXForTokenClassification, 'zero-shot': GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase_ : Union[str, Any] = False lowerCAmelCase_ : int = False lowerCAmelCase_ : Dict = False lowerCAmelCase_ : Any = False def A__ ( self ): UpperCAmelCase_ = GPTNeoXModelTester(self ) UpperCAmelCase_ = ConfigTester(self , config_class=lowerCAmelCase , hidden_size=64 , num_attention_heads=8 ) def A__ ( self ): self.config_tester.run_common_tests() def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): # This regression test was failing with PyTorch < 1.3 UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_decoder() UpperCAmelCase_ = None self.model_tester.create_and_check_model_as_decoder(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCAmelCase ) @unittest.skip(reason="Feed forward chunking is not implemented" ) def A__ ( self ): pass @parameterized.expand([("linear",), ("dynamic",)] ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = ids_tensor([1, 10] , config.vocab_size ) UpperCAmelCase_ = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights UpperCAmelCase_ = GPTNeoXModel(lowerCAmelCase ) original_model.to(lowerCAmelCase ) original_model.eval() UpperCAmelCase_ = original_model(lowerCAmelCase ).last_hidden_state UpperCAmelCase_ = original_model(lowerCAmelCase ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights UpperCAmelCase_ = {"type": scaling_type, "factor": 10.0} UpperCAmelCase_ = GPTNeoXModel(lowerCAmelCase ) scaled_model.to(lowerCAmelCase ) scaled_model.eval() UpperCAmelCase_ = scaled_model(lowerCAmelCase ).last_hidden_state UpperCAmelCase_ = scaled_model(lowerCAmelCase ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(lowerCAmelCase , lowerCAmelCase , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(lowerCAmelCase , lowerCAmelCase , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(lowerCAmelCase , lowerCAmelCase , atol=1e-5 ) ) @require_torch class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self ): UpperCAmelCase_ = AutoTokenizer.from_pretrained("EleutherAI/pythia-410m-deduped" ) for checkpointing in [True, False]: UpperCAmelCase_ = GPTNeoXForCausalLM.from_pretrained("EleutherAI/pythia-410m-deduped" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(lowerCAmelCase ) UpperCAmelCase_ = tokenizer("My favorite food is" , return_tensors="pt" ).to(lowerCAmelCase ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 UpperCAmelCase_ = "My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure" UpperCAmelCase_ = model.generate(**lowerCAmelCase , do_sample=lowerCAmelCase , max_new_tokens=20 ) UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase )[0] self.assertEqual(lowerCAmelCase , lowerCAmelCase )
23
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_VISION_2_SEQ_MAPPING if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_VISION_2_SEQ_MAPPING SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @add_end_docstrings(lowercase__ ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): super().__init__(*lowerCAmelCase , **lowerCAmelCase ) requires_backends(self , "vision" ) self.check_model_type( TF_MODEL_FOR_VISION_2_SEQ_MAPPING if self.framework == "tf" else MODEL_FOR_VISION_2_SEQ_MAPPING ) def A__ ( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None ): UpperCAmelCase_ = {} UpperCAmelCase_ = {} if prompt is not None: UpperCAmelCase_ = prompt if generate_kwargs is not None: UpperCAmelCase_ = generate_kwargs if max_new_tokens is not None: if "generate_kwargs" not in forward_kwargs: UpperCAmelCase_ = {} if "max_new_tokens" in forward_kwargs["generate_kwargs"]: raise ValueError( "'max_new_tokens' is defined twice, once in 'generate_kwargs' and once as a direct parameter," " please use only one" ) UpperCAmelCase_ = max_new_tokens return preprocess_params, forward_kwargs, {} def __call__( self , lowerCAmelCase , **lowerCAmelCase ): return super().__call__(lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = load_image(lowerCAmelCase ) if prompt is not None: if not isinstance(lowerCAmelCase , lowerCAmelCase ): raise ValueError( f'''Received an invalid text input, got - {type(lowerCAmelCase )} - but expected a single string. ''' "Note also that one single text can be provided for conditional image to text generation." ) UpperCAmelCase_ = self.model.config.model_type if model_type == "git": UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = self.tokenizer(text=lowerCAmelCase , add_special_tokens=lowerCAmelCase ).input_ids UpperCAmelCase_ = [self.tokenizer.cls_token_id] + input_ids UpperCAmelCase_ = torch.tensor(lowerCAmelCase ).unsqueeze(0 ) model_inputs.update({"input_ids": input_ids} ) elif model_type == "pix2struct": UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , header_text=lowerCAmelCase , return_tensors=self.framework ) elif model_type != "vision-encoder-decoder": # vision-encoder-decoder does not support conditional generation UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = self.tokenizer(lowerCAmelCase , return_tensors=self.framework ) model_inputs.update(lowerCAmelCase ) else: raise ValueError(f'''Model type {model_type} does not support conditional text generation''' ) else: UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) if self.model.config.model_type == "git" and prompt is None: UpperCAmelCase_ = None return model_inputs def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): # Git model sets `model_inputs["input_ids"] = None` in `preprocess` (when `prompt=None`). In batch model, the # pipeline will group them into a list of `None`, which fail `_forward`. Avoid this by checking it first. if ( "input_ids" in model_inputs and isinstance(model_inputs["input_ids"] , lowerCAmelCase ) and all(x is None for x in model_inputs["input_ids"] ) ): UpperCAmelCase_ = None if generate_kwargs is None: UpperCAmelCase_ = {} # FIXME: We need to pop here due to a difference in how `generation.py` and `generation.tf_utils.py` # parse inputs. In the Tensorflow version, `generate` raises an error if we don't use `input_ids` whereas # the PyTorch version matches it with `self.model.main_input_name` or `self.model.encoder.main_input_name` # in the `_prepare_model_inputs` method. UpperCAmelCase_ = model_inputs.pop(self.model.main_input_name ) UpperCAmelCase_ = self.model.generate(lowerCAmelCase , **lowerCAmelCase , **lowerCAmelCase ) return model_outputs def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = [] for output_ids in model_outputs: UpperCAmelCase_ = { "generated_text": self.tokenizer.decode( lowerCAmelCase , skip_special_tokens=lowerCAmelCase , ) } records.append(lowerCAmelCase ) return records
23
1
import inspect import os import sys import unittest import accelerate from accelerate.test_utils import execute_subprocess_async, require_tpu class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def A__ ( self ): UpperCAmelCase_ = inspect.getfile(accelerate.test_utils ) UpperCAmelCase_ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_script.py"] ) UpperCAmelCase_ = os.path.sep.join(inspect.getfile(self.__class__ ).split(os.path.sep )[:-1] ) @require_tpu def A__ ( self ): UpperCAmelCase_ = f''' {self.test_dir}/xla_spawn.py --num_cores 8 {self.test_file_path} '''.split() UpperCAmelCase_ = [sys.executable] + distributed_args execute_subprocess_async(lowerCAmelCase , env=os.environ.copy() )
23
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowerCamelCase ( lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : int = TextToVideoSDPipeline lowerCAmelCase_ : Dict = TEXT_TO_IMAGE_PARAMS lowerCAmelCase_ : Optional[Any] = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. lowerCAmelCase_ : Optional[Any] = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def A__ ( self ): torch.manual_seed(0 ) UpperCAmelCase_ = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) UpperCAmelCase_ = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowerCAmelCase , set_alpha_to_one=lowerCAmelCase , ) torch.manual_seed(0 ) UpperCAmelCase_ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) UpperCAmelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="gelu" , projection_dim=512 , ) UpperCAmelCase_ = CLIPTextModel(lowerCAmelCase ) UpperCAmelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCAmelCase_ = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def A__ ( self , lowerCAmelCase , lowerCAmelCase=0 ): if str(lowerCAmelCase ).startswith("mps" ): UpperCAmelCase_ = torch.manual_seed(lowerCAmelCase ) else: UpperCAmelCase_ = torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) UpperCAmelCase_ = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def A__ ( self ): UpperCAmelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ = self.get_dummy_components() UpperCAmelCase_ = TextToVideoSDPipeline(**lowerCAmelCase ) UpperCAmelCase_ = sd_pipe.to(lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase ) UpperCAmelCase_ = self.get_dummy_inputs(lowerCAmelCase ) UpperCAmelCase_ = "np" UpperCAmelCase_ = sd_pipe(**lowerCAmelCase ).frames UpperCAmelCase_ = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) UpperCAmelCase_ = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def A__ ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=3e-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def A__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=1e-2 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def A__ ( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def A__ ( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def A__ ( self ): pass def A__ ( self ): return super().test_progress_bar() @slow @skip_mps class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def A__ ( self ): UpperCAmelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy" ) UpperCAmelCase_ = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) UpperCAmelCase_ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) UpperCAmelCase_ = pipe.to("cuda" ) UpperCAmelCase_ = "Spiderman is surfing" UpperCAmelCase_ = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase_ = pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=25 , output_type="pt" ).frames UpperCAmelCase_ = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2 def A__ ( self ): UpperCAmelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy" ) UpperCAmelCase_ = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) UpperCAmelCase_ = pipe.to("cuda" ) UpperCAmelCase_ = "Spiderman is surfing" UpperCAmelCase_ = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase_ = pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=2 , output_type="pt" ).frames UpperCAmelCase_ = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2
23
1
from typing import List from .keymap import KEYMAP, get_character def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Any: def decorator(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = getattr(__SCREAMING_SNAKE_CASE , "handle_key" , [] ) handle += [key] setattr(__SCREAMING_SNAKE_CASE , "handle_key" , __SCREAMING_SNAKE_CASE ) return func return decorator def snake_case__ ( *__SCREAMING_SNAKE_CASE ) -> str: def decorator(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = getattr(__SCREAMING_SNAKE_CASE , "handle_key" , [] ) handle += keys setattr(__SCREAMING_SNAKE_CASE , "handle_key" , __SCREAMING_SNAKE_CASE ) return func return decorator class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __new__( cls , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = super().__new__(cls , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) if not hasattr(lowerCAmelCase , "key_handler" ): setattr(lowerCAmelCase , "key_handler" , {} ) setattr(lowerCAmelCase , "handle_input" , KeyHandler.handle_input ) for value in attrs.values(): UpperCAmelCase_ = getattr(lowerCAmelCase , "handle_key" , [] ) for key in handled_keys: UpperCAmelCase_ = value return new_cls @staticmethod def A__ ( cls ): UpperCAmelCase_ = get_character() if char != KEYMAP["undefined"]: UpperCAmelCase_ = ord(lowerCAmelCase ) UpperCAmelCase_ = cls.key_handler.get(lowerCAmelCase ) if handler: UpperCAmelCase_ = char return handler(cls ) else: return None def snake_case__ ( cls ) -> str: return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = [1] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = 0, 0, 0 UpperCAmelCase_ = ugly_nums[ia] * 2 UpperCAmelCase_ = ugly_nums[ia] * 3 UpperCAmelCase_ = ugly_nums[ia] * 5 for _ in range(1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ugly_nums.append(__SCREAMING_SNAKE_CASE ) if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(f'''{ugly_numbers(200) = }''')
23
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_plbart": ["PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP", "PLBartConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["PLBartTokenizer"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "PLBART_PRETRAINED_MODEL_ARCHIVE_LIST", "PLBartForCausalLM", "PLBartForConditionalGeneration", "PLBartForSequenceClassification", "PLBartModel", "PLBartPreTrainedModel", ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure)
23
import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=3 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=224 , lowerCAmelCase=1000 , lowerCAmelCase=[3, 3, 6, 4] , lowerCAmelCase=[48, 56, 112, 220] , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = is_training UpperCAmelCase_ = use_labels UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = num_labels UpperCAmelCase_ = image_size UpperCAmelCase_ = layer_depths UpperCAmelCase_ = embed_dims def A__ ( self ): UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase_ = self.get_config() return config, pixel_values, labels def A__ ( self ): return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="gelu" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=lowerCAmelCase , layer_scale_init_value=1e-5 , ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = SwiftFormerModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = SwiftFormerForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) UpperCAmelCase_ = SwiftFormerForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A__ ( self ): ((UpperCAmelCase_) , (UpperCAmelCase_) , (UpperCAmelCase_)) = self.prepare_config_and_inputs() UpperCAmelCase_ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Optional[Any] = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () lowerCAmelCase_ : int = ( {'feature-extraction': SwiftFormerModel, 'image-classification': SwiftFormerForImageClassification} if is_torch_available() else {} ) lowerCAmelCase_ : List[Any] = False lowerCAmelCase_ : Dict = False lowerCAmelCase_ : int = False lowerCAmelCase_ : str = False lowerCAmelCase_ : Optional[Any] = False def A__ ( self ): UpperCAmelCase_ = SwiftFormerModelTester(self ) UpperCAmelCase_ = ConfigTester( self , config_class=lowerCAmelCase , has_text_modality=lowerCAmelCase , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def A__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="SwiftFormer does not use inputs_embeds" ) def A__ ( self ): pass def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase , nn.Linear ) ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase ) @slow def A__ ( self ): for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = SwiftFormerModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) @unittest.skip(reason="SwiftFormer does not output attentions" ) def A__ ( self ): pass def A__ ( self ): def check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = model_class(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCAmelCase_ = model(**self._prepare_for_class(lowerCAmelCase , lowerCAmelCase ) ) UpperCAmelCase_ = outputs.hidden_states UpperCAmelCase_ = 8 self.assertEqual(len(lowerCAmelCase ) , lowerCAmelCase ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(lowerCAmelCase ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = True check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ = True check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): def _config_zero_init(lowerCAmelCase ): UpperCAmelCase_ = copy.deepcopy(lowerCAmelCase ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(lowerCAmelCase , lowerCAmelCase , 1e-1_0 ) if isinstance(getattr(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) , lowerCAmelCase ): UpperCAmelCase_ = _config_zero_init(getattr(lowerCAmelCase , lowerCAmelCase ) ) setattr(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return configs_no_init UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = _config_zero_init(lowerCAmelCase ) for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(config=lowerCAmelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9) / 1e9).round().item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def A__ ( self ): pass def snake_case__ ( ) -> str: UpperCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def A__ ( self ): return ViTImageProcessor.from_pretrained("MBZUAI/swiftformer-xs" ) if is_vision_available() else None @slow def A__ ( self ): UpperCAmelCase_ = SwiftFormerForImageClassification.from_pretrained("MBZUAI/swiftformer-xs" ).to(lowerCAmelCase ) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(images=lowerCAmelCase , return_tensors="pt" ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase_ = model(**lowerCAmelCase ) # verify the logits UpperCAmelCase_ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase ) UpperCAmelCase_ = torch.tensor([[-2.1_7_0_3e0_0, 2.1_1_0_7e0_0, -2.0_8_1_1e0_0]] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase , atol=1e-4 ) )
23
1
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/vocab.json", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/vocab.json", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/vocab.json", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/vocab.json", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/vocab.json", }, "merges_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/merges.txt", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/merges.txt", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/merges.txt", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/merges.txt", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/merges.txt", }, "tokenizer_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/tokenizer.json", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/tokenizer.json", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/tokenizer.json", }, } SCREAMING_SNAKE_CASE = { "gpt2": 1024, "gpt2-medium": 1024, "gpt2-large": 1024, "gpt2-xl": 1024, "distilgpt2": 1024, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = VOCAB_FILES_NAMES lowerCAmelCase_ : str = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : List[Any] = ['input_ids', 'attention_mask'] lowerCAmelCase_ : List[Any] = GPTaTokenizer def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase="<|endoftext|>" , lowerCAmelCase="<|endoftext|>" , lowerCAmelCase="<|endoftext|>" , lowerCAmelCase=False , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , lowerCAmelCase , tokenizer_file=lowerCAmelCase , unk_token=lowerCAmelCase , bos_token=lowerCAmelCase , eos_token=lowerCAmelCase , add_prefix_space=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = kwargs.pop("add_bos_token" , lowerCAmelCase ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowerCAmelCase ) != add_prefix_space: UpperCAmelCase_ = getattr(lowerCAmelCase , pre_tok_state.pop("type" ) ) UpperCAmelCase_ = add_prefix_space UpperCAmelCase_ = pre_tok_class(**lowerCAmelCase ) UpperCAmelCase_ = add_prefix_space def A__ ( self , *lowerCAmelCase , **lowerCAmelCase ): UpperCAmelCase_ = kwargs.get("is_split_into_words" , lowerCAmelCase ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , *lowerCAmelCase , **lowerCAmelCase ): UpperCAmelCase_ = kwargs.get("is_split_into_words" , lowerCAmelCase ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) + [self.eos_token_id] ) if len(lowerCAmelCase ) > self.model_max_length: UpperCAmelCase_ = input_ids[-self.model_max_length :] return input_ids
23
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available SCREAMING_SNAKE_CASE = {"tokenization_herbert": ["HerbertTokenizer"]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["HerbertTokenizerFast"] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = [1] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = 0, 0, 0 UpperCAmelCase_ = ugly_nums[ia] * 2 UpperCAmelCase_ = ugly_nums[ia] * 3 UpperCAmelCase_ = ugly_nums[ia] * 5 for _ in range(1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ugly_nums.append(__SCREAMING_SNAKE_CASE ) if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(f'''{ugly_numbers(200) = }''')
23
import math def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> list[int]: UpperCAmelCase_ = [] UpperCAmelCase_ = 2 UpperCAmelCase_ = int(math.sqrt(__SCREAMING_SNAKE_CASE ) ) # Size of every segment UpperCAmelCase_ = [True] * (end + 1) UpperCAmelCase_ = [] while start <= end: if temp[start] is True: in_prime.append(__SCREAMING_SNAKE_CASE ) for i in range(start * start , end + 1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = False start += 1 prime += in_prime UpperCAmelCase_ = end + 1 UpperCAmelCase_ = min(2 * end , __SCREAMING_SNAKE_CASE ) while low <= n: UpperCAmelCase_ = [True] * (high - low + 1) for each in in_prime: UpperCAmelCase_ = math.floor(low / each ) * each if t < low: t += each for j in range(__SCREAMING_SNAKE_CASE , high + 1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = False for j in range(len(__SCREAMING_SNAKE_CASE ) ): if temp[j] is True: prime.append(j + low ) UpperCAmelCase_ = high + 1 UpperCAmelCase_ = min(high + end , __SCREAMING_SNAKE_CASE ) return prime print(sieve(10**6))
23
1
import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=30 , lowerCAmelCase=2 , lowerCAmelCase=3 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=32 , lowerCAmelCase=5 , lowerCAmelCase=4 , lowerCAmelCase=37 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=10 , lowerCAmelCase=0.02 , lowerCAmelCase=3 , lowerCAmelCase=None , lowerCAmelCase=2 , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = image_size UpperCAmelCase_ = patch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = is_training UpperCAmelCase_ = use_labels UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = type_sequence_label_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = scope UpperCAmelCase_ = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase_ = (image_size // patch_size) ** 2 UpperCAmelCase_ = num_patches + 2 def A__ ( self ): UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase_ = self.get_config() return config, pixel_values, labels def A__ ( self ): return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCAmelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = DeiTModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = DeiTForMaskedImageModeling(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase_ = 1 UpperCAmelCase_ = DeiTForMaskedImageModeling(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = self.type_sequence_label_size UpperCAmelCase_ = DeiTForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase_ = 1 UpperCAmelCase_ = DeiTForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ = model(lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def A__ ( self ): UpperCAmelCase_ = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) = config_and_inputs UpperCAmelCase_ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) lowerCAmelCase_ : Optional[Any] = ( { 'feature-extraction': DeiTModel, 'image-classification': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowerCAmelCase_ : Union[str, Any] = False lowerCAmelCase_ : int = False lowerCAmelCase_ : Dict = False def A__ ( self ): UpperCAmelCase_ = DeiTModelTester(self ) UpperCAmelCase_ = ConfigTester(self , config_class=lowerCAmelCase , has_text_modality=lowerCAmelCase , hidden_size=37 ) def A__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="DeiT does not use inputs_embeds" ) def A__ ( self ): pass def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCAmelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase , nn.Linear ) ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False ): UpperCAmelCase_ = super()._prepare_for_class(lowerCAmelCase , lowerCAmelCase , return_labels=lowerCAmelCase ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def A__ ( self ): if not self.model_tester.is_training: return UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowerCAmelCase ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue UpperCAmelCase_ = model_class(lowerCAmelCase ) model.to(lowerCAmelCase ) model.train() UpperCAmelCase_ = self._prepare_for_class(lowerCAmelCase , lowerCAmelCase , return_labels=lowerCAmelCase ) UpperCAmelCase_ = model(**lowerCAmelCase ).loss loss.backward() def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return UpperCAmelCase_ = False UpperCAmelCase_ = True for model_class in self.all_model_classes: if model_class in get_values(lowerCAmelCase ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue UpperCAmelCase_ = model_class(lowerCAmelCase ) model.gradient_checkpointing_enable() model.to(lowerCAmelCase ) model.train() UpperCAmelCase_ = self._prepare_for_class(lowerCAmelCase , lowerCAmelCase , return_labels=lowerCAmelCase ) UpperCAmelCase_ = model(**lowerCAmelCase ).loss loss.backward() def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = [ {"title": "multi_label_classification", "num_labels": 2, "dtype": torch.float}, {"title": "single_label_classification", "num_labels": 1, "dtype": torch.long}, {"title": "regression", "num_labels": 1, "dtype": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowerCAmelCase ), *get_values(lowerCAmelCase ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f'''Testing {model_class} with {problem_type['title']}''' ): UpperCAmelCase_ = problem_type["title"] UpperCAmelCase_ = problem_type["num_labels"] UpperCAmelCase_ = model_class(lowerCAmelCase ) model.to(lowerCAmelCase ) model.train() UpperCAmelCase_ = self._prepare_for_class(lowerCAmelCase , lowerCAmelCase , return_labels=lowerCAmelCase ) if problem_type["num_labels"] > 1: UpperCAmelCase_ = inputs["labels"].unsqueeze(1 ).repeat(1 , problem_type["num_labels"] ) UpperCAmelCase_ = inputs["labels"].to(problem_type["dtype"] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowerCAmelCase ) as warning_list: UpperCAmelCase_ = model(**lowerCAmelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f'''Something is going wrong in the regression problem: intercepted {w.message}''' ) loss.backward() @slow def A__ ( self ): for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = DeiTModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) def snake_case__ ( ) -> Any: UpperCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def A__ ( self ): return ( DeiTImageProcessor.from_pretrained("facebook/deit-base-distilled-patch16-224" ) if is_vision_available() else None ) @slow def A__ ( self ): UpperCAmelCase_ = DeiTForImageClassificationWithTeacher.from_pretrained("facebook/deit-base-distilled-patch16-224" ).to( lowerCAmelCase ) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(images=lowerCAmelCase , return_tensors="pt" ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase_ = model(**lowerCAmelCase ) # verify the logits UpperCAmelCase_ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase ) UpperCAmelCase_ = torch.tensor([-1.0266, 0.1912, -1.2861] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def A__ ( self ): UpperCAmelCase_ = DeiTModel.from_pretrained( "facebook/deit-base-distilled-patch16-224" , torch_dtype=torch.floataa , device_map="auto" ) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(images=lowerCAmelCase , return_tensors="pt" ) UpperCAmelCase_ = inputs.pixel_values.to(lowerCAmelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): UpperCAmelCase_ = model(lowerCAmelCase )
23
from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : torch.FloatTensor class lowerCamelCase ( lowercase__, lowercase__ ): '''simple docstring''' @register_to_config def __init__( self , lowerCAmelCase = 32 , lowerCAmelCase = 64 , lowerCAmelCase = 20 , lowerCAmelCase = 768 , lowerCAmelCase=77 , lowerCAmelCase=4 , lowerCAmelCase = 0.0 , lowerCAmelCase = "silu" , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = "linear" , lowerCAmelCase = "prd" , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , ): super().__init__() UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = attention_head_dim UpperCAmelCase_ = num_attention_heads * attention_head_dim UpperCAmelCase_ = additional_embeddings UpperCAmelCase_ = time_embed_dim or inner_dim UpperCAmelCase_ = embedding_proj_dim or embedding_dim UpperCAmelCase_ = clip_embed_dim or embedding_dim UpperCAmelCase_ = Timesteps(lowerCAmelCase , lowerCAmelCase , 0 ) UpperCAmelCase_ = TimestepEmbedding(lowerCAmelCase , lowerCAmelCase , out_dim=lowerCAmelCase , act_fn=lowerCAmelCase ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) if embedding_proj_norm_type is None: UpperCAmelCase_ = None elif embedding_proj_norm_type == "layer": UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) else: raise ValueError(f'''unsupported embedding_proj_norm_type: {embedding_proj_norm_type}''' ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) if encoder_hid_proj_type is None: UpperCAmelCase_ = None elif encoder_hid_proj_type == "linear": UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) else: raise ValueError(f'''unsupported encoder_hid_proj_type: {encoder_hid_proj_type}''' ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , num_embeddings + additional_embeddings , lowerCAmelCase ) ) if added_emb_type == "prd": UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , 1 , lowerCAmelCase ) ) elif added_emb_type is None: UpperCAmelCase_ = None else: raise ValueError( f'''`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `\'prd\'` or `None`.''' ) UpperCAmelCase_ = nn.ModuleList( [ BasicTransformerBlock( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , dropout=lowerCAmelCase , activation_fn="gelu" , attention_bias=lowerCAmelCase , ) for d in range(lowerCAmelCase ) ] ) if norm_in_type == "layer": UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) elif norm_in_type is None: UpperCAmelCase_ = None else: raise ValueError(f'''Unsupported norm_in_type: {norm_in_type}.''' ) UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] , -10000.0 ) causal_attention_mask.triu_(1 ) UpperCAmelCase_ = causal_attention_mask[None, ...] self.register_buffer("causal_attention_mask" , lowerCAmelCase , persistent=lowerCAmelCase ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def A__ ( self ): UpperCAmelCase_ = {} def fn_recursive_add_processors(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if hasattr(lowerCAmelCase , "set_processor" ): UpperCAmelCase_ = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f'''{name}.{sub_name}''' , lowerCAmelCase , lowerCAmelCase ) return processors for name, module in self.named_children(): fn_recursive_add_processors(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return processors def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = len(self.attn_processors.keys() ) if isinstance(lowerCAmelCase , lowerCAmelCase ) and len(lowerCAmelCase ) != count: raise ValueError( f'''A dict of processors was passed, but the number of processors {len(lowerCAmelCase )} does not match the''' f''' number of attention layers: {count}. Please make sure to pass {count} processor classes.''' ) def fn_recursive_attn_processor(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if hasattr(lowerCAmelCase , "set_processor" ): if not isinstance(lowerCAmelCase , lowerCAmelCase ): module.set_processor(lowerCAmelCase ) else: module.set_processor(processor.pop(f'''{name}.processor''' ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f'''{name}.{sub_name}''' , lowerCAmelCase , lowerCAmelCase ) for name, module in self.named_children(): fn_recursive_attn_processor(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): self.set_attn_processor(AttnProcessor() ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = True , ): UpperCAmelCase_ = hidden_states.shape[0] UpperCAmelCase_ = timestep if not torch.is_tensor(lowerCAmelCase ): UpperCAmelCase_ = torch.tensor([timesteps] , dtype=torch.long , device=hidden_states.device ) elif torch.is_tensor(lowerCAmelCase ) and len(timesteps.shape ) == 0: UpperCAmelCase_ = timesteps[None].to(hidden_states.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML UpperCAmelCase_ = timesteps * torch.ones(lowerCAmelCase , dtype=timesteps.dtype , device=timesteps.device ) UpperCAmelCase_ = self.time_proj(lowerCAmelCase ) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. UpperCAmelCase_ = timesteps_projected.to(dtype=self.dtype ) UpperCAmelCase_ = self.time_embedding(lowerCAmelCase ) if self.embedding_proj_norm is not None: UpperCAmelCase_ = self.embedding_proj_norm(lowerCAmelCase ) UpperCAmelCase_ = self.embedding_proj(lowerCAmelCase ) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: UpperCAmelCase_ = self.encoder_hidden_states_proj(lowerCAmelCase ) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError("`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set" ) UpperCAmelCase_ = self.proj_in(lowerCAmelCase ) UpperCAmelCase_ = self.positional_embedding.to(hidden_states.dtype ) UpperCAmelCase_ = [] UpperCAmelCase_ = 0 if encoder_hidden_states is not None: additional_embeds.append(lowerCAmelCase ) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape ) == 2: UpperCAmelCase_ = proj_embeddings[:, None, :] if len(hidden_states.shape ) == 2: UpperCAmelCase_ = hidden_states[:, None, :] UpperCAmelCase_ = additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: UpperCAmelCase_ = self.prd_embedding.to(hidden_states.dtype ).expand(lowerCAmelCase , -1 , -1 ) additional_embeds.append(lowerCAmelCase ) UpperCAmelCase_ = torch.cat( lowerCAmelCase , dim=1 , ) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens UpperCAmelCase_ = additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: UpperCAmelCase_ = F.pad( lowerCAmelCase , ( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) , value=0.0 , ) UpperCAmelCase_ = hidden_states + positional_embeddings if attention_mask is not None: UpperCAmelCase_ = (1 - attention_mask.to(hidden_states.dtype )) * -10000.0 UpperCAmelCase_ = F.pad(lowerCAmelCase , (0, self.additional_embeddings) , value=0.0 ) UpperCAmelCase_ = (attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype ) UpperCAmelCase_ = attention_mask.repeat_interleave(self.config.num_attention_heads , dim=0 ) if self.norm_in is not None: UpperCAmelCase_ = self.norm_in(lowerCAmelCase ) for block in self.transformer_blocks: UpperCAmelCase_ = block(lowerCAmelCase , attention_mask=lowerCAmelCase ) UpperCAmelCase_ = self.norm_out(lowerCAmelCase ) if self.prd_embedding is not None: UpperCAmelCase_ = hidden_states[:, -1] else: UpperCAmelCase_ = hidden_states[:, additional_embeddings_len:] UpperCAmelCase_ = self.proj_to_clip_embeddings(lowerCAmelCase ) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (prior_latents * self.clip_std) + self.clip_mean return prior_latents
23
1
import itertools import json import os import unittest from transformers import AddedToken, RobertaTokenizer, RobertaTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCamelCase ( lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = RobertaTokenizer lowerCAmelCase_ : List[str] = RobertaTokenizerFast lowerCAmelCase_ : List[str] = True lowerCAmelCase_ : Optional[int] = {'cls_token': '<s>'} def A__ ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase_ = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] UpperCAmelCase_ = dict(zip(lowerCAmelCase , range(len(lowerCAmelCase ) ) ) ) UpperCAmelCase_ = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] UpperCAmelCase_ = {"unk_token": "<unk>"} UpperCAmelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowerCAmelCase ) ) def A__ ( self , **lowerCAmelCase ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase ) def A__ ( self , **lowerCAmelCase ): kwargs.update(self.special_tokens_map ) return RobertaTokenizerFast.from_pretrained(self.tmpdirname , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = "lower newer" UpperCAmelCase_ = "lower newer" return input_text, output_text def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCAmelCase_ = "lower newer" UpperCAmelCase_ = ["l", "o", "w", "er", "\u0120", "n", "e", "w", "er"] UpperCAmelCase_ = tokenizer.tokenize(lowerCAmelCase ) # , add_prefix_space=True) self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = tokens + [tokenizer.unk_token] UpperCAmelCase_ = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase ) , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.get_tokenizer() self.assertListEqual(tokenizer.encode("Hello world!" , add_special_tokens=lowerCAmelCase ) , [0, 3_1414, 232, 328, 2] ) self.assertListEqual( tokenizer.encode("Hello world! cécé herlolip 418" , add_special_tokens=lowerCAmelCase ) , [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2] , ) @slow def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained("roberta-base" ) UpperCAmelCase_ = tokenizer.encode("sequence builders" , add_special_tokens=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.encode("multi-sequence build" , add_special_tokens=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.encode( "sequence builders" , add_special_tokens=lowerCAmelCase , add_prefix_space=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.encode( "sequence builders" , "multi-sequence build" , add_special_tokens=lowerCAmelCase , add_prefix_space=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase ) UpperCAmelCase_ = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase , lowerCAmelCase ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def A__ ( self ): UpperCAmelCase_ = self.get_tokenizer() UpperCAmelCase_ = "Encode this sequence." UpperCAmelCase_ = tokenizer.byte_encoder[" ".encode("utf-8" )[0]] # Testing encoder arguments UpperCAmelCase_ = tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase , add_prefix_space=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase , add_prefix_space=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(lowerCAmelCase , lowerCAmelCase ) tokenizer.add_special_tokens({"bos_token": "<s>"} ) UpperCAmelCase_ = tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(lowerCAmelCase , lowerCAmelCase ) # Testing spaces after special tokens UpperCAmelCase_ = "<mask>" tokenizer.add_special_tokens( {"mask_token": AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase )} ) # mask token has a left space UpperCAmelCase_ = tokenizer.convert_tokens_to_ids(lowerCAmelCase ) UpperCAmelCase_ = "Encode <mask> sequence" UpperCAmelCase_ = "Encode <mask>sequence" UpperCAmelCase_ = tokenizer.encode(lowerCAmelCase ) UpperCAmelCase_ = encoded.index(lowerCAmelCase ) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = tokenizer.encode(lowerCAmelCase ) UpperCAmelCase_ = encoded.index(lowerCAmelCase ) UpperCAmelCase_ = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): pass def A__ ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = self.tokenizer_class.from_pretrained(lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = "A, <mask> AllenNLP sentence." UpperCAmelCase_ = tokenizer_r.encode_plus(lowerCAmelCase , add_special_tokens=lowerCAmelCase , return_token_type_ids=lowerCAmelCase ) UpperCAmelCase_ = tokenizer_p.encode_plus(lowerCAmelCase , add_special_tokens=lowerCAmelCase , return_token_type_ids=lowerCAmelCase ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["token_type_ids"] ) , sum(tokens_p["token_type_ids"] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) , sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) , ) UpperCAmelCase_ = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) UpperCAmelCase_ = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] , [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual( lowerCAmelCase , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( lowerCAmelCase , ["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) def A__ ( self ): for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=lowerCAmelCase , add_prefix_space=lowerCAmelCase , trim_offsets=lowerCAmelCase ) UpperCAmelCase_ = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) UpperCAmelCase_ = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["add_prefix_space"] , lowerCAmelCase ) self.assertEqual(post_processor_state["add_prefix_space"] , lowerCAmelCase ) self.assertEqual(post_processor_state["trim_offsets"] , lowerCAmelCase ) def A__ ( self ): # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` and # `trim_offsets` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): UpperCAmelCase_ = "hello" # `hello` is a token in the vocabulary of `pretrained_name` UpperCAmelCase_ = f'''{text_of_1_token} {text_of_1_token}''' UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase , use_fast=lowerCAmelCase , add_prefix_space=lowerCAmelCase , trim_offsets=lowerCAmelCase ) UpperCAmelCase_ = tokenizer_r(lowerCAmelCase , return_offsets_mapping=lowerCAmelCase , add_special_tokens=lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowerCAmelCase ) + 1, len(lowerCAmelCase ) + 1 + len(lowerCAmelCase )) , ) UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase , use_fast=lowerCAmelCase , add_prefix_space=lowerCAmelCase , trim_offsets=lowerCAmelCase ) UpperCAmelCase_ = tokenizer_r(lowerCAmelCase , return_offsets_mapping=lowerCAmelCase , add_special_tokens=lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowerCAmelCase ) + 1, len(lowerCAmelCase ) + 1 + len(lowerCAmelCase )) , ) UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase , use_fast=lowerCAmelCase , add_prefix_space=lowerCAmelCase , trim_offsets=lowerCAmelCase ) UpperCAmelCase_ = tokenizer_r(lowerCAmelCase , return_offsets_mapping=lowerCAmelCase , add_special_tokens=lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowerCAmelCase ), len(lowerCAmelCase ) + 1 + len(lowerCAmelCase )) , ) UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase , use_fast=lowerCAmelCase , add_prefix_space=lowerCAmelCase , trim_offsets=lowerCAmelCase ) UpperCAmelCase_ = tokenizer_r(lowerCAmelCase , return_offsets_mapping=lowerCAmelCase , add_special_tokens=lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowerCAmelCase ), len(lowerCAmelCase ) + 1 + len(lowerCAmelCase )) , ) UpperCAmelCase_ = f''' {text}''' # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase , use_fast=lowerCAmelCase , add_prefix_space=lowerCAmelCase , trim_offsets=lowerCAmelCase ) UpperCAmelCase_ = tokenizer_r(lowerCAmelCase , return_offsets_mapping=lowerCAmelCase , add_special_tokens=lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowerCAmelCase ) + 1, 1 + len(lowerCAmelCase ) + 1 + len(lowerCAmelCase )) , ) UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase , use_fast=lowerCAmelCase , add_prefix_space=lowerCAmelCase , trim_offsets=lowerCAmelCase ) UpperCAmelCase_ = tokenizer_r(lowerCAmelCase , return_offsets_mapping=lowerCAmelCase , add_special_tokens=lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowerCAmelCase ), 1 + len(lowerCAmelCase ) + 1 + len(lowerCAmelCase )) , ) UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase , use_fast=lowerCAmelCase , add_prefix_space=lowerCAmelCase , trim_offsets=lowerCAmelCase ) UpperCAmelCase_ = tokenizer_r(lowerCAmelCase , return_offsets_mapping=lowerCAmelCase , add_special_tokens=lowerCAmelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(lowerCAmelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowerCAmelCase ), 1 + len(lowerCAmelCase ) + 1 + len(lowerCAmelCase )) , )
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE = { "configuration_gpt_bigcode": ["GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTBigCodeConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTBigCodeForSequenceClassification", "GPTBigCodeForTokenClassification", "GPTBigCodeForCausalLM", "GPTBigCodeModel", "GPTBigCodePreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
def snake_case__ ( __SCREAMING_SNAKE_CASE = 1000 ) -> int: UpperCAmelCase_ , UpperCAmelCase_ = 1, 1 UpperCAmelCase_ = 2 while True: UpperCAmelCase_ = 0 UpperCAmelCase_ = fa + fa UpperCAmelCase_ , UpperCAmelCase_ = fa, f index += 1 for _ in str(__SCREAMING_SNAKE_CASE ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
23
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "xlm-roberta-base": "https://huggingface.co/xlm-roberta-base/resolve/main/config.json", "xlm-roberta-large": "https://huggingface.co/xlm-roberta-large/resolve/main/config.json", "xlm-roberta-large-finetuned-conll02-dutch": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll02-spanish": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-english": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-german": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json" ), } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = 'xlm-roberta' def __init__( self , lowerCAmelCase=3_0522 , lowerCAmelCase=768 , lowerCAmelCase=12 , lowerCAmelCase=12 , lowerCAmelCase=3072 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=1e-1_2 , lowerCAmelCase=1 , lowerCAmelCase=0 , lowerCAmelCase=2 , lowerCAmelCase="absolute" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__(pad_token_id=lowerCAmelCase , bos_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = position_embedding_type UpperCAmelCase_ = use_cache UpperCAmelCase_ = classifier_dropout class lowerCamelCase ( lowercase__ ): '''simple docstring''' @property def A__ ( self ): if self.task == "multiple-choice": UpperCAmelCase_ = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
23
1
import numpy class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase_ = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase_ = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase_ = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase_ = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase_ = numpy.zeros(output_array.shape ) def A__ ( self ): UpperCAmelCase_ = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase_ = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase_ = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def A__ ( self ): UpperCAmelCase_ = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase_ = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase_ = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): for iteration in range(1 , iterations + 1 ): UpperCAmelCase_ = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase_ = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = input_arr UpperCAmelCase_ = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase_ = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase_ = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> numpy.ndarray: return (value) * (1 - (value)) def snake_case__ ( ) -> int: UpperCAmelCase_ = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase_ = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase_ = TwoHiddenLayerNeuralNetwork( input_array=__SCREAMING_SNAKE_CASE , output_array=__SCREAMING_SNAKE_CASE ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=__SCREAMING_SNAKE_CASE , iterations=10 , give_loss=__SCREAMING_SNAKE_CASE ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = int(__SCREAMING_SNAKE_CASE ) if decimal in (0, 1): # Exit cases for the recursion return str(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ , UpperCAmelCase_ = divmod(__SCREAMING_SNAKE_CASE , 2 ) return binary_recursive(__SCREAMING_SNAKE_CASE ) + str(__SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = str(__SCREAMING_SNAKE_CASE ).strip() if not number: raise ValueError("No input value was provided" ) UpperCAmelCase_ = "-" if number.startswith("-" ) else "" UpperCAmelCase_ = number.lstrip("-" ) if not number.isnumeric(): raise ValueError("Input value is not an integer" ) return f'''{negative}0b{binary_recursive(int(__SCREAMING_SNAKE_CASE ) )}''' if __name__ == "__main__": from doctest import testmod testmod()
23
1
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class lowerCamelCase ( datasets.BeamBasedBuilder ): '''simple docstring''' def A__ ( self ): return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=lowerCAmelCase , ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def A__ ( self , lowerCAmelCase , lowerCAmelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCAmelCase ) class lowerCamelCase ( datasets.BeamBasedBuilder ): '''simple docstring''' def A__ ( self ): return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=lowerCAmelCase , ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ): return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def A__ ( self , lowerCAmelCase , lowerCAmelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(lowerCAmelCase ) def snake_case__ ( ) -> Tuple: return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"] )] def snake_case__ ( ) -> Any: return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"] )] class lowerCamelCase ( lowercase__ ): '''simple docstring''' @require_beam def A__ ( self ): UpperCAmelCase_ = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCAmelCase_ = DummyBeamDataset(cache_dir=lowerCAmelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCAmelCase , builder.name , "default" , "0.0.0" , f'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) UpperCAmelCase_ = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCAmelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCAmelCase ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCAmelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def A__ ( self ): import apache_beam as beam UpperCAmelCase_ = beam.io.parquetio.WriteToParquet UpperCAmelCase_ = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCAmelCase_ = DummyBeamDataset(cache_dir=lowerCAmelCase , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: UpperCAmelCase_ = partial(lowerCAmelCase , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( lowerCAmelCase , builder.name , "default" , "0.0.0" , f'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertTrue( os.path.exists( os.path.join( lowerCAmelCase , builder.name , "default" , "0.0.0" , f'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) UpperCAmelCase_ = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCAmelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCAmelCase ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(lowerCAmelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def A__ ( self ): with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCAmelCase_ = DummyBeamDataset(cache_dir=lowerCAmelCase ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def A__ ( self ): UpperCAmelCase_ = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: UpperCAmelCase_ = NestedBeamDataset(cache_dir=lowerCAmelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(lowerCAmelCase , builder.name , "default" , "0.0.0" , f'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) UpperCAmelCase_ = builder.as_dataset() self.assertEqual(dset["train"].num_rows , lowerCAmelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , lowerCAmelCase ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(lowerCAmelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
23
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self ): UpperCAmelCase_ = AutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" , return_dict=lowerCAmelCase ).to(lowerCAmelCase ) UpperCAmelCase_ = AutoTokenizer.from_pretrained("google/mt5-small" ) UpperCAmelCase_ = tokenizer("Hello there" , return_tensors="pt" ).input_ids UpperCAmelCase_ = tokenizer("Hi I am" , return_tensors="pt" ).input_ids UpperCAmelCase_ = model(input_ids.to(lowerCAmelCase ) , labels=labels.to(lowerCAmelCase ) ).loss UpperCAmelCase_ = -(labels.shape[-1] * loss.item()) UpperCAmelCase_ = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
23
1
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = "https://openaipublic.azureedge.net/jukebox/models/" SCREAMING_SNAKE_CASE = { "jukebox-1b-lyrics": [ "5b/vqvae.pth.tar", "5b/prior_level_0.pth.tar", "5b/prior_level_1.pth.tar", "1b_lyrics/prior_level_2.pth.tar", ], "jukebox-5b-lyrics": [ "5b/vqvae.pth.tar", "5b/prior_level_0.pth.tar", "5b/prior_level_1.pth.tar", "5b_lyrics/prior_level_2.pth.tar", ], } def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Tuple: if key.endswith(".model.1.bias" ) and len(key.split("." ) ) > 10: UpperCAmelCase_ = key.replace(".model.1.bias" , ".conv1d_1.bias" ) elif key.endswith(".model.1.weight" ) and len(key.split("." ) ) > 10: UpperCAmelCase_ = key.replace(".model.1.weight" , ".conv1d_1.weight" ) elif key.endswith(".model.3.bias" ) and len(key.split("." ) ) > 10: UpperCAmelCase_ = key.replace(".model.3.bias" , ".conv1d_2.bias" ) elif key.endswith(".model.3.weight" ) and len(key.split("." ) ) > 10: UpperCAmelCase_ = key.replace(".model.3.weight" , ".conv1d_2.weight" ) if "conditioner_blocks.0." in key: UpperCAmelCase_ = key.replace("conditioner_blocks.0" , "conditioner_blocks" ) if "prime_prior" in key: UpperCAmelCase_ = key.replace("prime_prior" , "encoder" ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: UpperCAmelCase_ = key.replace(".emb." , "." ) if key.endswith("k" ): # replace vqvae.X.k with vqvae.X.codebook return key.replace(".k" , ".codebook" ) if "y_emb." in key: return key.replace("y_emb." , "metadata_embedding." ) if "x_emb.emb." in key: UpperCAmelCase_ = key.replace("0.x_emb.emb" , "embed_tokens" ) if "prime_state_ln" in key: return key.replace("prime_state_ln" , "encoder.final_layer_norm" ) if ".ln" in key: return key.replace(".ln" , ".layer_norm" ) if "_ln" in key: return key.replace("_ln" , "_layer_norm" ) if "prime_state_proj" in key: return key.replace("prime_state_proj" , "encoder.proj_in" ) if "prime_x_out" in key: return key.replace("prime_x_out" , "encoder.lm_head" ) if "prior.x_out" in key: return key.replace("x_out" , "fc_proj_out" ) if "x_emb" in key: return key.replace("x_emb" , "embed_tokens" ) return key def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ = {} import re UpperCAmelCase_ = re.compile(R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) UpperCAmelCase_ = re.compile( R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) UpperCAmelCase_ = re.compile(R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) UpperCAmelCase_ = re.compile(R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) UpperCAmelCase_ = re.compile( R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) UpperCAmelCase_ = re.compile(R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) UpperCAmelCase_ = re.compile(R"conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)" ) UpperCAmelCase_ = re.compile( R"conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) UpperCAmelCase_ = re.compile(R"conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)" ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = re_encoder_block_conv_in.match(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = regex_match.groups() UpperCAmelCase_ = int(groups[2] ) * 2 + int(groups[3] ) UpperCAmelCase_ = f'''encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}''' UpperCAmelCase_ = re_encoder_block_conv_in.sub(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) elif re_encoder_block_resnet.fullmatch(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = re_encoder_block_resnet.match(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = regex_match.groups() UpperCAmelCase_ = int(groups[2] ) * 2 + int(groups[3] ) UpperCAmelCase_ = {"1": 1, "3": 2}[groups[-2]] UpperCAmelCase_ = f'''encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.''' UpperCAmelCase_ = f'''resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}''' UpperCAmelCase_ = prefix + resnet_block UpperCAmelCase_ = re_encoder_block_resnet.sub(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) elif re_encoder_block_proj_out.fullmatch(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = re_encoder_block_proj_out.match(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = regex_match.groups() UpperCAmelCase_ = f'''encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}''' UpperCAmelCase_ = re_encoder_block_proj_out.sub(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = re_decoder_block_conv_out.match(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = regex_match.groups() UpperCAmelCase_ = int(groups[2] ) * 2 + int(groups[3] ) - 2 UpperCAmelCase_ = f'''decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}''' UpperCAmelCase_ = re_decoder_block_conv_out.sub(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) elif re_decoder_block_resnet.fullmatch(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = re_decoder_block_resnet.match(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = regex_match.groups() UpperCAmelCase_ = int(groups[2] ) * 2 + int(groups[3] ) - 2 UpperCAmelCase_ = {"1": 1, "3": 2}[groups[-2]] UpperCAmelCase_ = f'''decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.''' UpperCAmelCase_ = f'''resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}''' UpperCAmelCase_ = prefix + resnet_block UpperCAmelCase_ = re_decoder_block_resnet.sub(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) elif re_decoder_block_proj_in.fullmatch(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = re_decoder_block_proj_in.match(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = regex_match.groups() UpperCAmelCase_ = f'''decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}''' UpperCAmelCase_ = re_decoder_block_proj_in.sub(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = re_prior_cond_conv_out.match(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = regex_match.groups() UpperCAmelCase_ = int(groups[1] ) * 2 + int(groups[2] ) - 2 UpperCAmelCase_ = f'''conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}''' UpperCAmelCase_ = re_prior_cond_conv_out.sub(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) elif re_prior_cond_resnet.fullmatch(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = re_prior_cond_resnet.match(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = regex_match.groups() UpperCAmelCase_ = int(groups[1] ) * 2 + int(groups[2] ) - 2 UpperCAmelCase_ = {"1": 1, "3": 2}[groups[-2]] UpperCAmelCase_ = f'''conditioner_blocks.upsampler.upsample_block.{block_index}.''' UpperCAmelCase_ = f'''resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}''' UpperCAmelCase_ = prefix + resnet_block UpperCAmelCase_ = re_prior_cond_resnet.sub(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) elif re_prior_cond_proj_in.fullmatch(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = re_prior_cond_proj_in.match(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = regex_match.groups() UpperCAmelCase_ = f'''conditioner_blocks.upsampler.proj_in.{groups[-1]}''' UpperCAmelCase_ = re_prior_cond_proj_in.sub(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # keep original key else: UpperCAmelCase_ = original_key UpperCAmelCase_ = replace_key(__SCREAMING_SNAKE_CASE ) if f'''{key_prefix}.{key}''' not in model_state_dict or key is None: print(f'''failed converting {original_key} to {key}, does not match''' ) # handle missmatched shape elif value.shape != model_state_dict[f'''{key_prefix}.{key}'''].shape: UpperCAmelCase_ = model_state_dict[f'''{key_prefix}.{key}'''] print(f'''{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match''' ) UpperCAmelCase_ = original_key UpperCAmelCase_ = original_key UpperCAmelCase_ = value return new_dict @torch.no_grad() def snake_case__ ( __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None ) -> int: for file in MODEL_MAPPING[model_name]: if not os.path.isfile(f'''{pytorch_dump_folder_path}/{file.split('/' )[-1]}''' ): UpperCAmelCase_ = requests.get(f'''{PREFIX}{file}''' , allow_redirects=__SCREAMING_SNAKE_CASE ) os.makedirs(f'''{pytorch_dump_folder_path}/''' , exist_ok=__SCREAMING_SNAKE_CASE ) open(f'''{pytorch_dump_folder_path}/{file.split('/' )[-1]}''' , "wb" ).write(r.content ) UpperCAmelCase_ = MODEL_MAPPING[model_name.split("/" )[-1]] UpperCAmelCase_ = JukeboxConfig.from_pretrained(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = JukeboxModel(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = [] UpperCAmelCase_ = {} for i, dict_name in enumerate(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = torch.load(f'''{pytorch_dump_folder_path}/{dict_name.split('/' )[-1]}''' )["model"] UpperCAmelCase_ = {} for k in old_dic.keys(): if k.endswith(".b" ): UpperCAmelCase_ = old_dic[k] elif k.endswith(".w" ): UpperCAmelCase_ = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: UpperCAmelCase_ = old_dic[k] else: UpperCAmelCase_ = old_dic[k] UpperCAmelCase_ = "vqvae" if i == 0 else f'''priors.{3 - i}''' UpperCAmelCase_ = fix_jukebox_keys(__SCREAMING_SNAKE_CASE , model.state_dict() , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) weight_dict.append(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = weight_dict.pop(0 ) model.vqvae.load_state_dict(__SCREAMING_SNAKE_CASE ) for i in range(len(__SCREAMING_SNAKE_CASE ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(__SCREAMING_SNAKE_CASE ).mkdir(exist_ok=__SCREAMING_SNAKE_CASE ) with open(f'''{pytorch_dump_folder_path}/mapping.json''' , "w" ) as txtfile: json.dump(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__SCREAMING_SNAKE_CASE ) return weight_dict if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="jukebox-5b-lyrics", type=str, help="Name of the model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default="jukebox-5b-lyrics-converted", type=str, help="Path to the output PyTorch model directory.", ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
23
def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ = 0 while b > 0: if b & 1: UpperCAmelCase_ = ((res % c) + (a % c)) % c a += a b >>= 1 return res
23
1
from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar SCREAMING_SNAKE_CASE = TypeVar("T") class lowerCamelCase ( Generic[T] ): '''simple docstring''' lowerCAmelCase_ : deque[T] # Cache store of keys lowerCAmelCase_ : set[T] # References of the keys in cache lowerCAmelCase_ : int = 10 # Maximum capacity of cache def __init__( self , lowerCAmelCase ): UpperCAmelCase_ = deque() UpperCAmelCase_ = set() if not n: UpperCAmelCase_ = sys.maxsize elif n < 0: raise ValueError("n should be an integer greater than 0." ) else: UpperCAmelCase_ = n def A__ ( self , lowerCAmelCase ): if x not in self.key_reference: if len(self.dq_store ) == LRUCache._MAX_CAPACITY: UpperCAmelCase_ = self.dq_store.pop() self.key_reference.remove(lowerCAmelCase ) else: self.dq_store.remove(lowerCAmelCase ) self.dq_store.appendleft(lowerCAmelCase ) self.key_reference.add(lowerCAmelCase ) def A__ ( self ): for k in self.dq_store: print(lowerCAmelCase ) def __repr__( self ): return f'''LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store )}''' if __name__ == "__main__": import doctest doctest.testmod() SCREAMING_SNAKE_CASE = LRUCache(4) lru_cache.refer("A") lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer("A") lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
23
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> List[List[ImageInput]]: if isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(__SCREAMING_SNAKE_CASE ): return [[videos]] raise ValueError(f'''Could not make batched video from {videos}''' ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[Any] = ['pixel_values'] def __init__( self , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = True , lowerCAmelCase = 1 / 255 , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = None , **lowerCAmelCase , ): super().__init__(**lowerCAmelCase ) UpperCAmelCase_ = size if size is not None else {"shortest_edge": 224} UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) UpperCAmelCase_ = crop_size if crop_size is not None else {"height": 224, "width": 224} UpperCAmelCase_ = get_size_dict(lowerCAmelCase , param_name="crop_size" ) UpperCAmelCase_ = do_resize UpperCAmelCase_ = size UpperCAmelCase_ = do_center_crop UpperCAmelCase_ = crop_size UpperCAmelCase_ = resample UpperCAmelCase_ = do_rescale UpperCAmelCase_ = rescale_factor UpperCAmelCase_ = do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) if "shortest_edge" in size: UpperCAmelCase_ = get_resize_output_image_size(lowerCAmelCase , size["shortest_edge"] , default_to_square=lowerCAmelCase ) elif "height" in size and "width" in size: UpperCAmelCase_ = (size["height"], size["width"]) else: raise ValueError(f'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = get_size_dict(lowerCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(lowerCAmelCase , size=(size["height"], size["width"]) , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return rescale(lowerCAmelCase , scale=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return normalize(lowerCAmelCase , mean=lowerCAmelCase , std=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , ): if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. UpperCAmelCase_ = to_numpy_array(lowerCAmelCase ) if do_resize: UpperCAmelCase_ = self.resize(image=lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase ) if do_center_crop: UpperCAmelCase_ = self.center_crop(lowerCAmelCase , size=lowerCAmelCase ) if do_rescale: UpperCAmelCase_ = self.rescale(image=lowerCAmelCase , scale=lowerCAmelCase ) if do_normalize: UpperCAmelCase_ = self.normalize(image=lowerCAmelCase , mean=lowerCAmelCase , std=lowerCAmelCase ) UpperCAmelCase_ = to_channel_dimension_format(lowerCAmelCase , lowerCAmelCase ) return image def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , **lowerCAmelCase , ): UpperCAmelCase_ = do_resize if do_resize is not None else self.do_resize UpperCAmelCase_ = resample if resample is not None else self.resample UpperCAmelCase_ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase_ = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase_ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase_ = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else self.image_mean UpperCAmelCase_ = image_std if image_std is not None else self.image_std UpperCAmelCase_ = size if size is not None else self.size UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) UpperCAmelCase_ = crop_size if crop_size is not None else self.crop_size UpperCAmelCase_ = get_size_dict(lowerCAmelCase , param_name="crop_size" ) if not valid_images(lowerCAmelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) UpperCAmelCase_ = make_batched(lowerCAmelCase ) UpperCAmelCase_ = [ [ self._preprocess_image( image=lowerCAmelCase , do_resize=lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase , do_center_crop=lowerCAmelCase , crop_size=lowerCAmelCase , do_rescale=lowerCAmelCase , rescale_factor=lowerCAmelCase , do_normalize=lowerCAmelCase , image_mean=lowerCAmelCase , image_std=lowerCAmelCase , data_format=lowerCAmelCase , ) for img in video ] for video in videos ] UpperCAmelCase_ = {"pixel_values": videos} return BatchFeature(data=lowerCAmelCase , tensor_type=lowerCAmelCase )
23
1
import gc import threading import time import psutil import torch class lowerCamelCase : '''simple docstring''' def __init__( self ): UpperCAmelCase_ = psutil.Process() UpperCAmelCase_ = False def A__ ( self ): UpperCAmelCase_ = -1 while True: UpperCAmelCase_ = max(self.process.memory_info().rss , self.cpu_memory_peak ) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def A__ ( self ): UpperCAmelCase_ = True UpperCAmelCase_ = threading.Thread(target=self.peak_monitor ) UpperCAmelCase_ = True self.thread.start() def A__ ( self ): UpperCAmelCase_ = False self.thread.join() return self.cpu_memory_peak SCREAMING_SNAKE_CASE = PeakCPUMemory() def snake_case__ ( ) -> List[Any]: # Time UpperCAmelCase_ = {"time": time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem UpperCAmelCase_ = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): UpperCAmelCase_ = torch.cuda.memory_allocated(__SCREAMING_SNAKE_CASE ) torch.cuda.reset_peak_memory_stats() return measures def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: # Time UpperCAmelCase_ = {"time": time.time() - start_measures["time"]} gc.collect() torch.cuda.empty_cache() # CPU mem UpperCAmelCase_ = (psutil.Process().memory_info().rss - start_measures["cpu"]) / 2**20 UpperCAmelCase_ = (cpu_peak_tracker.stop() - start_measures["cpu"]) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): UpperCAmelCase_ = (torch.cuda.memory_allocated(__SCREAMING_SNAKE_CASE ) - start_measures[str(__SCREAMING_SNAKE_CASE )]) / 2**20 UpperCAmelCase_ = (torch.cuda.max_memory_allocated(__SCREAMING_SNAKE_CASE ) - start_measures[str(__SCREAMING_SNAKE_CASE )]) / 2**20 return measures def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: print(f'''{description}:''' ) print(f'''- Time: {measures['time']:.2f}s''' ) for i in range(torch.cuda.device_count() ): print(f'''- GPU {i} allocated: {measures[str(__SCREAMING_SNAKE_CASE )]:.2f}MiB''' ) UpperCAmelCase_ = measures[f'''{i}-peak'''] print(f'''- GPU {i} peak: {peak:.2f}MiB''' ) print(f'''- CPU RAM allocated: {measures['cpu']:.2f}MiB''' ) print(f'''- CPU RAM peak: {measures['cpu-peak']:.2f}MiB''' )
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 1 for i in range(1 , num + 1 ): fact *= i return fact def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 0 while number > 0: UpperCAmelCase_ = number % 10 sum_of_digits += last_digit UpperCAmelCase_ = number // 10 # Removing the last_digit from the given number return sum_of_digits def snake_case__ ( __SCREAMING_SNAKE_CASE = 100 ) -> int: UpperCAmelCase_ = factorial(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = split_and_add(__SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
23
1
import logging from transformers.configuration_utils import PretrainedConfig SCREAMING_SNAKE_CASE = logging.getLogger(__name__) class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = 'masked_bert' def __init__( self , lowerCAmelCase=3_0522 , lowerCAmelCase=768 , lowerCAmelCase=12 , lowerCAmelCase=12 , lowerCAmelCase=3072 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=1e-1_2 , lowerCAmelCase=0 , lowerCAmelCase="topK" , lowerCAmelCase="constant" , lowerCAmelCase=0.0 , **lowerCAmelCase , ): super().__init__(pad_token_id=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = pruning_method UpperCAmelCase_ = mask_init UpperCAmelCase_ = mask_scale
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_xlnet": ["XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["XLNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["XLNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "XLNetForMultipleChoice", "XLNetForQuestionAnswering", "XLNetForQuestionAnsweringSimple", "XLNetForSequenceClassification", "XLNetForTokenClassification", "XLNetLMHeadModel", "XLNetModel", "XLNetPreTrainedModel", "load_tf_weights_in_xlnet", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLNetForMultipleChoice", "TFXLNetForQuestionAnsweringSimple", "TFXLNetForSequenceClassification", "TFXLNetForTokenClassification", "TFXLNetLMHeadModel", "TFXLNetMainLayer", "TFXLNetModel", "TFXLNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
from ...configuration_utils import PretrainedConfig class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Dict = 'bert-generation' def __init__( self , lowerCAmelCase=5_0358 , lowerCAmelCase=1024 , lowerCAmelCase=24 , lowerCAmelCase=16 , lowerCAmelCase=4096 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=0.02 , lowerCAmelCase=1e-1_2 , lowerCAmelCase=0 , lowerCAmelCase=2 , lowerCAmelCase=1 , lowerCAmelCase="absolute" , lowerCAmelCase=True , **lowerCAmelCase , ): super().__init__(pad_token_id=lowerCAmelCase , bos_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = position_embedding_type UpperCAmelCase_ = use_cache
23
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[Any]: # Initialise PyTorch model UpperCAmelCase_ = MobileBertConfig.from_json_file(__SCREAMING_SNAKE_CASE ) print(f'''Building PyTorch model from configuration: {config}''' ) UpperCAmelCase_ = MobileBertForPreTraining(__SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint UpperCAmelCase_ = load_tf_weights_in_mobilebert(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , __SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--mobilebert_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained MobileBERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
23
1
from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run SCREAMING_SNAKE_CASE = True except (ImportError, AttributeError): SCREAMING_SNAKE_CASE = object def snake_case__ ( *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) -> Dict: pass SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = logging.get_logger("transformers-cli/serving") def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) return ServeCommand(__SCREAMING_SNAKE_CASE , args.host , args.port , args.workers ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : dict class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] lowerCAmelCase_ : Optional[List[int]] class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : str class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Any class lowerCamelCase ( lowercase__ ): '''simple docstring''' @staticmethod def A__ ( lowerCAmelCase ): UpperCAmelCase_ = parser.add_parser( "serve" , help="CLI tool to run inference requests through REST and GraphQL endpoints." ) serve_parser.add_argument( "--task" , type=lowerCAmelCase , choices=get_supported_tasks() , help="The task to run the pipeline on" , ) serve_parser.add_argument("--host" , type=lowerCAmelCase , default="localhost" , help="Interface the server will listen on." ) serve_parser.add_argument("--port" , type=lowerCAmelCase , default=8888 , help="Port the serving will listen to." ) serve_parser.add_argument("--workers" , type=lowerCAmelCase , default=1 , help="Number of http workers" ) serve_parser.add_argument("--model" , type=lowerCAmelCase , help="Model's name or path to stored model." ) serve_parser.add_argument("--config" , type=lowerCAmelCase , help="Model's config name or path to stored model." ) serve_parser.add_argument("--tokenizer" , type=lowerCAmelCase , help="Tokenizer name to use." ) serve_parser.add_argument( "--device" , type=lowerCAmelCase , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) serve_parser.set_defaults(func=lowerCAmelCase ) def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = pipeline UpperCAmelCase_ = host UpperCAmelCase_ = port UpperCAmelCase_ = workers if not _serve_dependencies_installed: raise RuntimeError( "Using serve command requires FastAPI and uvicorn. " "Please install transformers with [serving]: pip install \"transformers[serving]\"." "Or install FastAPI and uvicorn separately." ) else: logger.info(f'''Serving model over {host}:{port}''' ) UpperCAmelCase_ = FastAPI( routes=[ APIRoute( "/" , self.model_info , response_model=lowerCAmelCase , response_class=lowerCAmelCase , methods=["GET"] , ), APIRoute( "/tokenize" , self.tokenize , response_model=lowerCAmelCase , response_class=lowerCAmelCase , methods=["POST"] , ), APIRoute( "/detokenize" , self.detokenize , response_model=lowerCAmelCase , response_class=lowerCAmelCase , methods=["POST"] , ), APIRoute( "/forward" , self.forward , response_model=lowerCAmelCase , response_class=lowerCAmelCase , methods=["POST"] , ), ] , timeout=600 , ) def A__ ( self ): run(self._app , host=self.host , port=self.port , workers=self.workers ) def A__ ( self ): return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def A__ ( self , lowerCAmelCase = Body(lowerCAmelCase , embed=lowerCAmelCase ) , lowerCAmelCase = Body(lowerCAmelCase , embed=lowerCAmelCase ) ): try: UpperCAmelCase_ = self._pipeline.tokenizer.tokenize(lowerCAmelCase ) if return_ids: UpperCAmelCase_ = self._pipeline.tokenizer.convert_tokens_to_ids(lowerCAmelCase ) return ServeTokenizeResult(tokens=lowerCAmelCase , tokens_ids=lowerCAmelCase ) else: return ServeTokenizeResult(tokens=lowerCAmelCase ) except Exception as e: raise HTTPException(status_code=500 , detail={"model": "", "error": str(lowerCAmelCase )} ) def A__ ( self , lowerCAmelCase = Body(lowerCAmelCase , embed=lowerCAmelCase ) , lowerCAmelCase = Body(lowerCAmelCase , embed=lowerCAmelCase ) , lowerCAmelCase = Body(lowerCAmelCase , embed=lowerCAmelCase ) , ): try: UpperCAmelCase_ = self._pipeline.tokenizer.decode(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return ServeDeTokenizeResult(model="" , text=lowerCAmelCase ) except Exception as e: raise HTTPException(status_code=500 , detail={"model": "", "error": str(lowerCAmelCase )} ) async def A__ ( self , lowerCAmelCase=Body(lowerCAmelCase , embed=lowerCAmelCase ) ): # Check we don't have empty string if len(lowerCAmelCase ) == 0: return ServeForwardResult(output=[] , attention=[] ) try: # Forward through the model UpperCAmelCase_ = self._pipeline(lowerCAmelCase ) return ServeForwardResult(output=lowerCAmelCase ) except Exception as e: raise HTTPException(500 , {"error": str(lowerCAmelCase )} )
23
import heapq as hq import math from collections.abc import Iterator class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase ): UpperCAmelCase_ = str(id_ ) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = [] UpperCAmelCase_ = {} # {vertex:distance} def __lt__( self , lowerCAmelCase ): return self.key < other.key def __repr__( self ): return self.id def A__ ( self , lowerCAmelCase ): self.neighbors.append(lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = weight def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __SCREAMING_SNAKE_CASE ) graph[b - 1].add_edge(graph[a - 1] , __SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> list: UpperCAmelCase_ = [] for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = graph[:] while q: UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE ) q.remove(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Iterator[tuple]: for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = list(__SCREAMING_SNAKE_CASE ) hq.heapify(__SCREAMING_SNAKE_CASE ) while h: UpperCAmelCase_ = hq.heappop(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] hq.heapify(__SCREAMING_SNAKE_CASE ) for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def snake_case__ ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
23
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE = { "configuration_nllb_moe": [ "NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP", "NllbMoeConfig", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST", "NllbMoeForConditionalGeneration", "NllbMoeModel", "NllbMoePreTrainedModel", "NllbMoeTop2Router", "NllbMoeSparseMLP", ] if TYPE_CHECKING: from .configuration_nllb_moe import ( NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP, NllbMoeConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nllb_moe import ( NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST, NllbMoeForConditionalGeneration, NllbMoeModel, NllbMoePreTrainedModel, NllbMoeSparseMLP, NllbMoeTopaRouter, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_fnet": ["FNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "FNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["FNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["FNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "FNET_PRETRAINED_MODEL_ARCHIVE_LIST", "FNetForMaskedLM", "FNetForMultipleChoice", "FNetForNextSentencePrediction", "FNetForPreTraining", "FNetForQuestionAnswering", "FNetForSequenceClassification", "FNetForTokenClassification", "FNetLayer", "FNetModel", "FNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = { "configuration_roberta": ["ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP", "RobertaConfig", "RobertaOnnxConfig"], "tokenization_roberta": ["RobertaTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["RobertaTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "RobertaForCausalLM", "RobertaForMaskedLM", "RobertaForMultipleChoice", "RobertaForQuestionAnswering", "RobertaForSequenceClassification", "RobertaForTokenClassification", "RobertaModel", "RobertaPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRobertaForCausalLM", "TFRobertaForMaskedLM", "TFRobertaForMultipleChoice", "TFRobertaForQuestionAnswering", "TFRobertaForSequenceClassification", "TFRobertaForTokenClassification", "TFRobertaMainLayer", "TFRobertaModel", "TFRobertaPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "FlaxRobertaForCausalLM", "FlaxRobertaForMaskedLM", "FlaxRobertaForMultipleChoice", "FlaxRobertaForQuestionAnswering", "FlaxRobertaForSequenceClassification", "FlaxRobertaForTokenClassification", "FlaxRobertaModel", "FlaxRobertaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_roberta import ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaConfig, RobertaOnnxConfig from .tokenization_roberta import RobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roberta_fast import RobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta import ( ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaForCausalLM, RobertaForMaskedLM, RobertaForMultipleChoice, RobertaForQuestionAnswering, RobertaForSequenceClassification, RobertaForTokenClassification, RobertaModel, RobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta import ( TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForMultipleChoice, TFRobertaForQuestionAnswering, TFRobertaForSequenceClassification, TFRobertaForTokenClassification, TFRobertaMainLayer, TFRobertaModel, TFRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, FlaxRobertaPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": 512, } SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": {"do_lower_case": True}, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = VOCAB_FILES_NAMES lowerCAmelCase_ : int = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Dict = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : List[str] = RetriBertTokenizer lowerCAmelCase_ : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase="[UNK]" , lowerCAmelCase="[SEP]" , lowerCAmelCase="[PAD]" , lowerCAmelCase="[CLS]" , lowerCAmelCase="[MASK]" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , do_lower_case=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , pad_token=lowerCAmelCase , cls_token=lowerCAmelCase , mask_token=lowerCAmelCase , tokenize_chinese_chars=lowerCAmelCase , strip_accents=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(lowerCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**lowerCAmelCase ) UpperCAmelCase_ = do_lower_case def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
1
def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> float: if density <= 0: raise ValueError("Impossible fluid density" ) if bulk_modulus <= 0: raise ValueError("Impossible bulk modulus" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
23
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-german-cased": ( "https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json" ), "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE = { "distilbert-base-uncased": 512, "distilbert-base-uncased-distilled-squad": 512, "distilbert-base-cased": 512, "distilbert-base-cased-distilled-squad": 512, "distilbert-base-german-cased": 512, "distilbert-base-multilingual-cased": 512, } SCREAMING_SNAKE_CASE = { "distilbert-base-uncased": {"do_lower_case": True}, "distilbert-base-uncased-distilled-squad": {"do_lower_case": True}, "distilbert-base-cased": {"do_lower_case": False}, "distilbert-base-cased-distilled-squad": {"do_lower_case": False}, "distilbert-base-german-cased": {"do_lower_case": False}, "distilbert-base-multilingual-cased": {"do_lower_case": False}, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Any = VOCAB_FILES_NAMES lowerCAmelCase_ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Union[str, Any] = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : int = ['input_ids', 'attention_mask'] lowerCAmelCase_ : str = DistilBertTokenizer def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase="[UNK]" , lowerCAmelCase="[SEP]" , lowerCAmelCase="[PAD]" , lowerCAmelCase="[CLS]" , lowerCAmelCase="[MASK]" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , do_lower_case=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , pad_token=lowerCAmelCase , cls_token=lowerCAmelCase , mask_token=lowerCAmelCase , tokenize_chinese_chars=lowerCAmelCase , strip_accents=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(lowerCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**lowerCAmelCase ) UpperCAmelCase_ = do_lower_case def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
1
def snake_case__ ( __SCREAMING_SNAKE_CASE = 100_0000 ) -> int: UpperCAmelCase_ = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , __SCREAMING_SNAKE_CASE ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
23
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> np.array: UpperCAmelCase_ = f'''{sampling_rate}''' UpperCAmelCase_ = "1" UpperCAmelCase_ = "f32le" UpperCAmelCase_ = [ "ffmpeg", "-i", "pipe:0", "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-hide_banner", "-loglevel", "quiet", "pipe:1", ] try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: UpperCAmelCase_ = ffmpeg_process.communicate(__SCREAMING_SNAKE_CASE ) except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to load audio files from filename" ) from error UpperCAmelCase_ = output_stream[0] UpperCAmelCase_ = np.frombuffer(__SCREAMING_SNAKE_CASE , np.floataa ) if audio.shape[0] == 0: raise ValueError("Malformed soundfile" ) return audio def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = "f32le" , ) -> Dict: UpperCAmelCase_ = f'''{sampling_rate}''' UpperCAmelCase_ = "1" if format_for_conversion == "s16le": UpperCAmelCase_ = 2 elif format_for_conversion == "f32le": UpperCAmelCase_ = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) UpperCAmelCase_ = platform.system() if system == "Linux": UpperCAmelCase_ = "alsa" UpperCAmelCase_ = "default" elif system == "Darwin": UpperCAmelCase_ = "avfoundation" UpperCAmelCase_ = ":0" elif system == "Windows": UpperCAmelCase_ = "dshow" UpperCAmelCase_ = "default" UpperCAmelCase_ = [ "ffmpeg", "-f", format_, "-i", input_, "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-fflags", "nobuffer", "-hide_banner", "-loglevel", "quiet", "pipe:1", ] UpperCAmelCase_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample UpperCAmelCase_ = _ffmpeg_stream(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for item in iterator: yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = "f32le" , ) -> int: if stream_chunk_s is not None: UpperCAmelCase_ = stream_chunk_s else: UpperCAmelCase_ = chunk_length_s UpperCAmelCase_ = ffmpeg_microphone(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , format_for_conversion=__SCREAMING_SNAKE_CASE ) if format_for_conversion == "s16le": UpperCAmelCase_ = np.intaa UpperCAmelCase_ = 2 elif format_for_conversion == "f32le": UpperCAmelCase_ = np.floataa UpperCAmelCase_ = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: UpperCAmelCase_ = chunk_length_s / 6 UpperCAmelCase_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(__SCREAMING_SNAKE_CASE , (int, float) ): UpperCAmelCase_ = [stride_length_s, stride_length_s] UpperCAmelCase_ = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample UpperCAmelCase_ = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample UpperCAmelCase_ = datetime.datetime.now() UpperCAmelCase_ = datetime.timedelta(seconds=__SCREAMING_SNAKE_CASE ) for item in chunk_bytes_iter(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=(stride_left, stride_right) , stream=__SCREAMING_SNAKE_CASE ): # Put everything back in numpy scale UpperCAmelCase_ = np.frombuffer(item["raw"] , dtype=__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = ( item["stride"][0] // size_of_sample, item["stride"][1] // size_of_sample, ) UpperCAmelCase_ = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = False ) -> Dict: UpperCAmelCase_ = B"" UpperCAmelCase_ , UpperCAmelCase_ = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) UpperCAmelCase_ = 0 for raw in iterator: acc += raw if stream and len(__SCREAMING_SNAKE_CASE ) < chunk_len: UpperCAmelCase_ = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(__SCREAMING_SNAKE_CASE ) >= chunk_len: # We are flushing the accumulator UpperCAmelCase_ = (_stride_left, stride_right) UpperCAmelCase_ = {"raw": acc[:chunk_len], "stride": stride} if stream: UpperCAmelCase_ = False yield item UpperCAmelCase_ = stride_left UpperCAmelCase_ = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(__SCREAMING_SNAKE_CASE ) > stride_left: UpperCAmelCase_ = {"raw": acc, "stride": (_stride_left, 0)} if stream: UpperCAmelCase_ = False yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ = 2**24 # 16Mo try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdout=subprocess.PIPE , bufsize=__SCREAMING_SNAKE_CASE ) as ffmpeg_process: while True: UpperCAmelCase_ = ffmpeg_process.stdout.read(__SCREAMING_SNAKE_CASE ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to stream audio files from filename" ) from error
23
1
import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all BART models at https://huggingface.co/models?filter=bart SCREAMING_SNAKE_CASE = { "vocab_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/vocab.json", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/vocab.json", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json", }, "merges_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/merges.txt", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/merges.txt", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt", }, } SCREAMING_SNAKE_CASE = { "facebook/bart-base": 1024, "facebook/bart-large": 1024, "facebook/bart-large-mnli": 1024, "facebook/bart-large-cnn": 1024, "facebook/bart-large-xsum": 1024, "yjernite/bart_eli5": 1024, } @lru_cache() def snake_case__ ( ) -> Tuple: UpperCAmelCase_ = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) UpperCAmelCase_ = bs[:] UpperCAmelCase_ = 0 for b in range(2**8 ): if b not in bs: bs.append(__SCREAMING_SNAKE_CASE ) cs.append(2**8 + n ) n += 1 UpperCAmelCase_ = [chr(__SCREAMING_SNAKE_CASE ) for n in cs] return dict(zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Optional[int]: UpperCAmelCase_ = set() UpperCAmelCase_ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase_ = char return pairs class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : int = VOCAB_FILES_NAMES lowerCAmelCase_ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : str = ['input_ids', 'attention_mask'] def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase="replace" , lowerCAmelCase="<s>" , lowerCAmelCase="</s>" , lowerCAmelCase="</s>" , lowerCAmelCase="<s>" , lowerCAmelCase="<unk>" , lowerCAmelCase="<pad>" , lowerCAmelCase="<mask>" , lowerCAmelCase=False , **lowerCAmelCase , ): UpperCAmelCase_ = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else bos_token UpperCAmelCase_ = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else eos_token UpperCAmelCase_ = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else sep_token UpperCAmelCase_ = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else cls_token UpperCAmelCase_ = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else unk_token UpperCAmelCase_ = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase_ = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else mask_token super().__init__( errors=lowerCAmelCase , bos_token=lowerCAmelCase , eos_token=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , cls_token=lowerCAmelCase , pad_token=lowerCAmelCase , mask_token=lowerCAmelCase , add_prefix_space=lowerCAmelCase , **lowerCAmelCase , ) with open(lowerCAmelCase , encoding="utf-8" ) as vocab_handle: UpperCAmelCase_ = json.load(lowerCAmelCase ) UpperCAmelCase_ = {v: k for k, v in self.encoder.items()} UpperCAmelCase_ = errors # how to handle errors in decoding UpperCAmelCase_ = bytes_to_unicode() UpperCAmelCase_ = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase , encoding="utf-8" ) as merges_handle: UpperCAmelCase_ = merges_handle.read().split("\n" )[1:-1] UpperCAmelCase_ = [tuple(merge.split() ) for merge in bpe_merges] UpperCAmelCase_ = dict(zip(lowerCAmelCase , range(len(lowerCAmelCase ) ) ) ) UpperCAmelCase_ = {} UpperCAmelCase_ = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions UpperCAmelCase_ = re.compile(r"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property def A__ ( self ): return len(self.encoder ) def A__ ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def A__ ( self , lowerCAmelCase ): if token in self.cache: return self.cache[token] UpperCAmelCase_ = tuple(lowerCAmelCase ) UpperCAmelCase_ = get_pairs(lowerCAmelCase ) if not pairs: return token while True: UpperCAmelCase_ = min(lowerCAmelCase , key=lambda lowerCAmelCase : self.bpe_ranks.get(lowerCAmelCase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase_ , UpperCAmelCase_ = bigram UpperCAmelCase_ = [] UpperCAmelCase_ = 0 while i < len(lowerCAmelCase ): try: UpperCAmelCase_ = word.index(lowerCAmelCase , lowerCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase_ = j if word[i] == first and i < len(lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase_ = tuple(lowerCAmelCase ) UpperCAmelCase_ = new_word if len(lowerCAmelCase ) == 1: break else: UpperCAmelCase_ = get_pairs(lowerCAmelCase ) UpperCAmelCase_ = " ".join(lowerCAmelCase ) UpperCAmelCase_ = word return word def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = [] for token in re.findall(self.pat , lowerCAmelCase ): UpperCAmelCase_ = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase ).split(" " ) ) return bpe_tokens def A__ ( self , lowerCAmelCase ): return self.encoder.get(lowerCAmelCase , self.encoder.get(self.unk_token ) ) def A__ ( self , lowerCAmelCase ): return self.decoder.get(lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = "".join(lowerCAmelCase ) UpperCAmelCase_ = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): if not os.path.isdir(lowerCAmelCase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase_ = os.path.join( lowerCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase_ = os.path.join( lowerCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(lowerCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase , ensure_ascii=lowerCAmelCase ) + "\n" ) UpperCAmelCase_ = 0 with open(lowerCAmelCase , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) UpperCAmelCase_ = token_index writer.write(" ".join(lowerCAmelCase ) + "\n" ) index += 1 return vocab_file, merge_file def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] UpperCAmelCase_ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase , token_ids_a=lowerCAmelCase , already_has_special_tokens=lowerCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase )) + [1] return [1] + ([0] * len(lowerCAmelCase )) + [1, 1] + ([0] * len(lowerCAmelCase )) + [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A__ ( self , lowerCAmelCase , lowerCAmelCase=False , **lowerCAmelCase ): UpperCAmelCase_ = kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase ) > 0 and not text[0].isspace()): UpperCAmelCase_ = " " + text return (text, kwargs)
23
from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class lowerCamelCase ( lowercase__, lowercase__ ): '''simple docstring''' @register_to_config def __init__( self , lowerCAmelCase = 768 , ): super().__init__() UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(torch.ones(1 , lowerCAmelCase ) ) def A__ ( self , lowerCAmelCase = None , lowerCAmelCase = None , ): UpperCAmelCase_ = nn.Parameter(self.mean.to(lowerCAmelCase ).to(lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(self.std.to(lowerCAmelCase ).to(lowerCAmelCase ) ) return self def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (embeds - self.mean) * 1.0 / self.std return embeds def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (embeds * self.std) + self.mean return embeds
23
1
import tensorflow as tf from ...tf_utils import shape_list class lowerCamelCase ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=1 , lowerCAmelCase=False , **lowerCAmelCase ): super().__init__(**lowerCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = d_embed UpperCAmelCase_ = d_proj UpperCAmelCase_ = cutoffs + [vocab_size] UpperCAmelCase_ = [0] + self.cutoffs UpperCAmelCase_ = div_val UpperCAmelCase_ = self.cutoffs[0] UpperCAmelCase_ = len(self.cutoffs ) - 1 UpperCAmelCase_ = self.shortlist_size + self.n_clusters UpperCAmelCase_ = keep_order UpperCAmelCase_ = [] UpperCAmelCase_ = [] def A__ ( self , lowerCAmelCase ): if self.n_clusters > 0: UpperCAmelCase_ = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer="zeros" , trainable=lowerCAmelCase , name="cluster_weight" ) UpperCAmelCase_ = self.add_weight( shape=(self.n_clusters,) , initializer="zeros" , trainable=lowerCAmelCase , name="cluster_bias" ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: UpperCAmelCase_ = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer="zeros" , trainable=lowerCAmelCase , name=f'''out_projs_._{i}''' , ) self.out_projs.append(lowerCAmelCase ) else: self.out_projs.append(lowerCAmelCase ) UpperCAmelCase_ = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer="zeros" , trainable=lowerCAmelCase , name=f'''out_layers_._{i}_._weight''' , ) UpperCAmelCase_ = self.add_weight( shape=(self.vocab_size,) , initializer="zeros" , trainable=lowerCAmelCase , name=f'''out_layers_._{i}_._bias''' , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): UpperCAmelCase_ , UpperCAmelCase_ = self.cutoff_ends[i], self.cutoff_ends[i + 1] UpperCAmelCase_ = self.d_embed // (self.div_val**i) UpperCAmelCase_ = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer="zeros" , trainable=lowerCAmelCase , name=f'''out_projs_._{i}''' ) self.out_projs.append(lowerCAmelCase ) UpperCAmelCase_ = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer="zeros" , trainable=lowerCAmelCase , name=f'''out_layers_._{i}_._weight''' , ) UpperCAmelCase_ = self.add_weight( shape=(r_idx - l_idx,) , initializer="zeros" , trainable=lowerCAmelCase , name=f'''out_layers_._{i}_._bias''' , ) self.out_layers.append((weight, bias) ) super().build(lowerCAmelCase ) @staticmethod def A__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = x if proj is not None: UpperCAmelCase_ = tf.einsum("ibd,ed->ibe" , lowerCAmelCase , lowerCAmelCase ) return tf.einsum("ibd,nd->ibn" , lowerCAmelCase , lowerCAmelCase ) + b @staticmethod def A__ ( lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = shape_list(lowerCAmelCase ) UpperCAmelCase_ = tf.range(lp_size[0] , dtype=target.dtype ) UpperCAmelCase_ = tf.stack([r, target] , 1 ) return tf.gather_nd(lowerCAmelCase , lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=True , lowerCAmelCase=False ): UpperCAmelCase_ = 0 if self.n_clusters == 0: UpperCAmelCase_ = self._logit(lowerCAmelCase , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: UpperCAmelCase_ = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=lowerCAmelCase , logits=lowerCAmelCase ) UpperCAmelCase_ = tf.nn.log_softmax(lowerCAmelCase , axis=-1 ) else: UpperCAmelCase_ = shape_list(lowerCAmelCase ) UpperCAmelCase_ = [] UpperCAmelCase_ = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): UpperCAmelCase_ , UpperCAmelCase_ = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: UpperCAmelCase_ = (target >= l_idx) & (target < r_idx) UpperCAmelCase_ = tf.where(lowerCAmelCase ) UpperCAmelCase_ = tf.boolean_mask(lowerCAmelCase , lowerCAmelCase ) - l_idx if self.div_val == 1: UpperCAmelCase_ = self.out_layers[0][0][l_idx:r_idx] UpperCAmelCase_ = self.out_layers[0][1][l_idx:r_idx] else: UpperCAmelCase_ = self.out_layers[i][0] UpperCAmelCase_ = self.out_layers[i][1] if i == 0: UpperCAmelCase_ = tf.concat([cur_W, self.cluster_weight] , 0 ) UpperCAmelCase_ = tf.concat([cur_b, self.cluster_bias] , 0 ) UpperCAmelCase_ = self._logit(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , self.out_projs[0] ) UpperCAmelCase_ = tf.nn.log_softmax(lowerCAmelCase ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: UpperCAmelCase_ = tf.boolean_mask(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = self._gather_logprob(lowerCAmelCase , lowerCAmelCase ) else: UpperCAmelCase_ = self._logit(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , self.out_projs[i] ) UpperCAmelCase_ = tf.nn.log_softmax(lowerCAmelCase ) UpperCAmelCase_ = self.cutoffs[0] + i - 1 # No probability for the head cluster UpperCAmelCase_ = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(lowerCAmelCase ) if target is not None: UpperCAmelCase_ = tf.boolean_mask(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = tf.boolean_mask(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = self._gather_logprob(lowerCAmelCase , lowerCAmelCase ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(lowerCAmelCase , -cur_logprob , shape_list(lowerCAmelCase ) ) UpperCAmelCase_ = tf.concat(lowerCAmelCase , axis=-1 ) if target is not None: if return_mean: UpperCAmelCase_ = tf.reduce_mean(lowerCAmelCase ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(lowerCAmelCase ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(lowerCAmelCase , name=self.name , aggregation="mean" if return_mean else "" ) return out
23
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_VISION_2_SEQ_MAPPING if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_VISION_2_SEQ_MAPPING SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @add_end_docstrings(lowercase__ ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): super().__init__(*lowerCAmelCase , **lowerCAmelCase ) requires_backends(self , "vision" ) self.check_model_type( TF_MODEL_FOR_VISION_2_SEQ_MAPPING if self.framework == "tf" else MODEL_FOR_VISION_2_SEQ_MAPPING ) def A__ ( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None ): UpperCAmelCase_ = {} UpperCAmelCase_ = {} if prompt is not None: UpperCAmelCase_ = prompt if generate_kwargs is not None: UpperCAmelCase_ = generate_kwargs if max_new_tokens is not None: if "generate_kwargs" not in forward_kwargs: UpperCAmelCase_ = {} if "max_new_tokens" in forward_kwargs["generate_kwargs"]: raise ValueError( "'max_new_tokens' is defined twice, once in 'generate_kwargs' and once as a direct parameter," " please use only one" ) UpperCAmelCase_ = max_new_tokens return preprocess_params, forward_kwargs, {} def __call__( self , lowerCAmelCase , **lowerCAmelCase ): return super().__call__(lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = load_image(lowerCAmelCase ) if prompt is not None: if not isinstance(lowerCAmelCase , lowerCAmelCase ): raise ValueError( f'''Received an invalid text input, got - {type(lowerCAmelCase )} - but expected a single string. ''' "Note also that one single text can be provided for conditional image to text generation." ) UpperCAmelCase_ = self.model.config.model_type if model_type == "git": UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = self.tokenizer(text=lowerCAmelCase , add_special_tokens=lowerCAmelCase ).input_ids UpperCAmelCase_ = [self.tokenizer.cls_token_id] + input_ids UpperCAmelCase_ = torch.tensor(lowerCAmelCase ).unsqueeze(0 ) model_inputs.update({"input_ids": input_ids} ) elif model_type == "pix2struct": UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , header_text=lowerCAmelCase , return_tensors=self.framework ) elif model_type != "vision-encoder-decoder": # vision-encoder-decoder does not support conditional generation UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = self.tokenizer(lowerCAmelCase , return_tensors=self.framework ) model_inputs.update(lowerCAmelCase ) else: raise ValueError(f'''Model type {model_type} does not support conditional text generation''' ) else: UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) if self.model.config.model_type == "git" and prompt is None: UpperCAmelCase_ = None return model_inputs def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): # Git model sets `model_inputs["input_ids"] = None` in `preprocess` (when `prompt=None`). In batch model, the # pipeline will group them into a list of `None`, which fail `_forward`. Avoid this by checking it first. if ( "input_ids" in model_inputs and isinstance(model_inputs["input_ids"] , lowerCAmelCase ) and all(x is None for x in model_inputs["input_ids"] ) ): UpperCAmelCase_ = None if generate_kwargs is None: UpperCAmelCase_ = {} # FIXME: We need to pop here due to a difference in how `generation.py` and `generation.tf_utils.py` # parse inputs. In the Tensorflow version, `generate` raises an error if we don't use `input_ids` whereas # the PyTorch version matches it with `self.model.main_input_name` or `self.model.encoder.main_input_name` # in the `_prepare_model_inputs` method. UpperCAmelCase_ = model_inputs.pop(self.model.main_input_name ) UpperCAmelCase_ = self.model.generate(lowerCAmelCase , **lowerCAmelCase , **lowerCAmelCase ) return model_outputs def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = [] for output_ids in model_outputs: UpperCAmelCase_ = { "generated_text": self.tokenizer.decode( lowerCAmelCase , skip_special_tokens=lowerCAmelCase , ) } records.append(lowerCAmelCase ) return records
23
1
import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = inspect.getfile(accelerate.test_utils ) lowerCAmelCase_ : Union[str, Any] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['scripts', 'test_cli.py'] ) lowerCAmelCase_ : Dict = ['accelerate', 'launch'] lowerCAmelCase_ : int = Path.home() / '.cache/huggingface/accelerate' lowerCAmelCase_ : str = 'default_config.yaml' lowerCAmelCase_ : Dict = config_folder / config_file lowerCAmelCase_ : Tuple = config_folder / '_default_config.yaml' lowerCAmelCase_ : Dict = Path('tests/test_configs' ) @classmethod def A__ ( cls ): if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def A__ ( cls ): if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def A__ ( self ): UpperCAmelCase_ = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] , env=os.environ.copy() ) def A__ ( self ): for config in sorted(self.test_config_path.glob("**/*.yaml" ) ): with self.subTest(config_file=lowerCAmelCase ): execute_subprocess_async( self.base_cmd + ["--config_file", str(lowerCAmelCase ), self.test_file_path] , env=os.environ.copy() ) def A__ ( self ): execute_subprocess_async(["accelerate", "test"] , env=os.environ.copy() ) class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Any = 'test-tpu' lowerCAmelCase_ : List[Any] = 'us-central1-a' lowerCAmelCase_ : int = 'ls' lowerCAmelCase_ : Optional[int] = ['accelerate', 'tpu-config'] lowerCAmelCase_ : str = 'cd /usr/share' lowerCAmelCase_ : Optional[int] = 'tests/test_samples/test_command_file.sh' lowerCAmelCase_ : List[str] = 'Running gcloud compute tpus tpu-vm ssh' def A__ ( self ): UpperCAmelCase_ = run_command( self.cmd + ["--command", self.command, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug"] , return_stdout=lowerCAmelCase , ) self.assertIn( f'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all''' , lowerCAmelCase , ) def A__ ( self ): UpperCAmelCase_ = run_command( self.cmd + [ "--config_file", "tests/test_configs/0_12_0.yaml", "--command", self.command, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug", ] , return_stdout=lowerCAmelCase , ) self.assertIn( f'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all''' , lowerCAmelCase , ) def A__ ( self ): UpperCAmelCase_ = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--debug"] , return_stdout=lowerCAmelCase ) self.assertIn( f'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo "hello world"; echo "this is a second command" --worker all''' , lowerCAmelCase , ) def A__ ( self ): UpperCAmelCase_ = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--command", self.command, "--debug"] , return_stdout=lowerCAmelCase , ) self.assertIn( f'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all''' , lowerCAmelCase , ) def A__ ( self ): UpperCAmelCase_ = run_command( self.cmd + [ "--config_file", "tests/test_configs/latest.yaml", "--command", self.command, "--command", "echo \"Hello World\"", "--debug", ] , return_stdout=lowerCAmelCase , ) self.assertIn( f'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo "Hello World" --worker all''' , lowerCAmelCase , ) def A__ ( self ): UpperCAmelCase_ = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--command_file", self.command_file, "--debug"] , return_stdout=lowerCAmelCase , ) self.assertIn( f'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo "hello world"; echo "this is a second command" --worker all''' , lowerCAmelCase , ) def A__ ( self ): UpperCAmelCase_ = run_command( self.cmd + [ "--config_file", "tests/test_configs/0_12_0.yaml", "--command_file", self.command_file, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug", ] , return_stdout=lowerCAmelCase , ) self.assertIn( f'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo "hello world"; echo "this is a second command" --worker all''' , lowerCAmelCase , ) def A__ ( self ): UpperCAmelCase_ = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--install_accelerate", "--debug"] , return_stdout=lowerCAmelCase , ) self.assertIn( f'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo "hello world"; echo "this is a second command" --worker all''' , lowerCAmelCase , ) def A__ ( self ): UpperCAmelCase_ = run_command( self.cmd + [ "--config_file", "tests/test_configs/latest.yaml", "--install_accelerate", "--accelerate_version", "12.0.0", "--debug", ] , return_stdout=lowerCAmelCase , ) self.assertIn( f'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo "hello world"; echo "this is a second command" --worker all''' , lowerCAmelCase , )
23
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowerCamelCase ( lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : int = TextToVideoSDPipeline lowerCAmelCase_ : Dict = TEXT_TO_IMAGE_PARAMS lowerCAmelCase_ : Optional[Any] = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. lowerCAmelCase_ : Optional[Any] = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def A__ ( self ): torch.manual_seed(0 ) UpperCAmelCase_ = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) UpperCAmelCase_ = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowerCAmelCase , set_alpha_to_one=lowerCAmelCase , ) torch.manual_seed(0 ) UpperCAmelCase_ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) UpperCAmelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="gelu" , projection_dim=512 , ) UpperCAmelCase_ = CLIPTextModel(lowerCAmelCase ) UpperCAmelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCAmelCase_ = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def A__ ( self , lowerCAmelCase , lowerCAmelCase=0 ): if str(lowerCAmelCase ).startswith("mps" ): UpperCAmelCase_ = torch.manual_seed(lowerCAmelCase ) else: UpperCAmelCase_ = torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) UpperCAmelCase_ = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def A__ ( self ): UpperCAmelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ = self.get_dummy_components() UpperCAmelCase_ = TextToVideoSDPipeline(**lowerCAmelCase ) UpperCAmelCase_ = sd_pipe.to(lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase ) UpperCAmelCase_ = self.get_dummy_inputs(lowerCAmelCase ) UpperCAmelCase_ = "np" UpperCAmelCase_ = sd_pipe(**lowerCAmelCase ).frames UpperCAmelCase_ = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) UpperCAmelCase_ = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def A__ ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=3e-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def A__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=1e-2 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def A__ ( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def A__ ( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def A__ ( self ): pass def A__ ( self ): return super().test_progress_bar() @slow @skip_mps class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def A__ ( self ): UpperCAmelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy" ) UpperCAmelCase_ = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) UpperCAmelCase_ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) UpperCAmelCase_ = pipe.to("cuda" ) UpperCAmelCase_ = "Spiderman is surfing" UpperCAmelCase_ = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase_ = pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=25 , output_type="pt" ).frames UpperCAmelCase_ = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2 def A__ ( self ): UpperCAmelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy" ) UpperCAmelCase_ = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) UpperCAmelCase_ = pipe.to("cuda" ) UpperCAmelCase_ = "Spiderman is surfing" UpperCAmelCase_ = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase_ = pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=2 , output_type="pt" ).frames UpperCAmelCase_ = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2
23
1
import os def snake_case__ ( ) -> str: UpperCAmelCase_ = os.path.dirname(os.path.realpath(__SCREAMING_SNAKE_CASE ) ) UpperCAmelCase_ = os.path.join(__SCREAMING_SNAKE_CASE , "triangle.txt" ) with open(__SCREAMING_SNAKE_CASE ) as f: UpperCAmelCase_ = f.readlines() UpperCAmelCase_ = [] for line in triangle: UpperCAmelCase_ = [] for number in line.strip().split(" " ): numbers_from_line.append(int(__SCREAMING_SNAKE_CASE ) ) a.append(__SCREAMING_SNAKE_CASE ) for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): for j in range(len(a[i] ) ): UpperCAmelCase_ = a[i - 1][j] if j != len(a[i - 1] ) else 0 UpperCAmelCase_ = a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return max(a[-1] ) if __name__ == "__main__": print(solution())
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = [1] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = 0, 0, 0 UpperCAmelCase_ = ugly_nums[ia] * 2 UpperCAmelCase_ = ugly_nums[ia] * 3 UpperCAmelCase_ = ugly_nums[ia] * 5 for _ in range(1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ugly_nums.append(__SCREAMING_SNAKE_CASE ) if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(f'''{ugly_numbers(200) = }''')
23
1
from ....configuration_utils import PretrainedConfig from ....utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) # TODO: upload to AWS SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/config.json" ), } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Optional[Any] = 'retribert' def __init__( self , lowerCAmelCase=3_0522 , lowerCAmelCase=768 , lowerCAmelCase=8 , lowerCAmelCase=12 , lowerCAmelCase=3072 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=1e-1_2 , lowerCAmelCase=True , lowerCAmelCase=128 , lowerCAmelCase=0 , **lowerCAmelCase , ): super().__init__(pad_token_id=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = share_encoders UpperCAmelCase_ = projection_dim
23
import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=3 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=224 , lowerCAmelCase=1000 , lowerCAmelCase=[3, 3, 6, 4] , lowerCAmelCase=[48, 56, 112, 220] , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = is_training UpperCAmelCase_ = use_labels UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = num_labels UpperCAmelCase_ = image_size UpperCAmelCase_ = layer_depths UpperCAmelCase_ = embed_dims def A__ ( self ): UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase_ = self.get_config() return config, pixel_values, labels def A__ ( self ): return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="gelu" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=lowerCAmelCase , layer_scale_init_value=1e-5 , ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = SwiftFormerModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = SwiftFormerForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) UpperCAmelCase_ = SwiftFormerForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A__ ( self ): ((UpperCAmelCase_) , (UpperCAmelCase_) , (UpperCAmelCase_)) = self.prepare_config_and_inputs() UpperCAmelCase_ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Optional[Any] = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () lowerCAmelCase_ : int = ( {'feature-extraction': SwiftFormerModel, 'image-classification': SwiftFormerForImageClassification} if is_torch_available() else {} ) lowerCAmelCase_ : List[Any] = False lowerCAmelCase_ : Dict = False lowerCAmelCase_ : int = False lowerCAmelCase_ : str = False lowerCAmelCase_ : Optional[Any] = False def A__ ( self ): UpperCAmelCase_ = SwiftFormerModelTester(self ) UpperCAmelCase_ = ConfigTester( self , config_class=lowerCAmelCase , has_text_modality=lowerCAmelCase , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def A__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="SwiftFormer does not use inputs_embeds" ) def A__ ( self ): pass def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase , nn.Linear ) ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase ) @slow def A__ ( self ): for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = SwiftFormerModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) @unittest.skip(reason="SwiftFormer does not output attentions" ) def A__ ( self ): pass def A__ ( self ): def check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = model_class(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCAmelCase_ = model(**self._prepare_for_class(lowerCAmelCase , lowerCAmelCase ) ) UpperCAmelCase_ = outputs.hidden_states UpperCAmelCase_ = 8 self.assertEqual(len(lowerCAmelCase ) , lowerCAmelCase ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(lowerCAmelCase ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = True check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ = True check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): def _config_zero_init(lowerCAmelCase ): UpperCAmelCase_ = copy.deepcopy(lowerCAmelCase ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(lowerCAmelCase , lowerCAmelCase , 1e-1_0 ) if isinstance(getattr(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) , lowerCAmelCase ): UpperCAmelCase_ = _config_zero_init(getattr(lowerCAmelCase , lowerCAmelCase ) ) setattr(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return configs_no_init UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = _config_zero_init(lowerCAmelCase ) for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(config=lowerCAmelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9) / 1e9).round().item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def A__ ( self ): pass def snake_case__ ( ) -> str: UpperCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def A__ ( self ): return ViTImageProcessor.from_pretrained("MBZUAI/swiftformer-xs" ) if is_vision_available() else None @slow def A__ ( self ): UpperCAmelCase_ = SwiftFormerForImageClassification.from_pretrained("MBZUAI/swiftformer-xs" ).to(lowerCAmelCase ) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(images=lowerCAmelCase , return_tensors="pt" ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase_ = model(**lowerCAmelCase ) # verify the logits UpperCAmelCase_ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase ) UpperCAmelCase_ = torch.tensor([[-2.1_7_0_3e0_0, 2.1_1_0_7e0_0, -2.0_8_1_1e0_0]] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase , atol=1e-4 ) )
23
1
import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=7 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=99 , lowerCAmelCase=32 , lowerCAmelCase=5 , lowerCAmelCase=4 , lowerCAmelCase=37 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=16 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=4 , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = seq_length UpperCAmelCase_ = is_training UpperCAmelCase_ = use_attention_mask UpperCAmelCase_ = use_token_type_ids UpperCAmelCase_ = use_labels UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = type_sequence_label_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = num_choices def A__ ( self ): UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase_ = None if self.use_attention_mask: UpperCAmelCase_ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=lowerCAmelCase , ) return config, input_ids, attention_mask def A__ ( self ): UpperCAmelCase_ = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = config_and_inputs UpperCAmelCase_ = {"input_ids": input_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class lowerCamelCase ( lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : List[Any] = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def A__ ( self ): UpperCAmelCase_ = FlaxDistilBertModelTester(self ) @slow def A__ ( self ): for model_class_name in self.all_model_classes: UpperCAmelCase_ = model_class_name.from_pretrained("distilbert-base-uncased" ) UpperCAmelCase_ = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCAmelCase ) @require_flax class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self ): UpperCAmelCase_ = FlaxDistilBertModel.from_pretrained("distilbert-base-uncased" ) UpperCAmelCase_ = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) UpperCAmelCase_ = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase )[0] UpperCAmelCase_ = (1, 11, 768) self.assertEqual(output.shape , lowerCAmelCase ) UpperCAmelCase_ = np.array([[[-0.1639, 0.3299, 0.1648], [-0.1746, 0.3289, 0.1710], [-0.1884, 0.3357, 0.1810]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCAmelCase , atol=1e-4 ) )
23
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available SCREAMING_SNAKE_CASE = {"tokenization_herbert": ["HerbertTokenizer"]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["HerbertTokenizerFast"] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : str = 'EncodecFeatureExtractor' lowerCAmelCase_ : str = ('T5Tokenizer', 'T5TokenizerFast') def __init__( self , lowerCAmelCase , lowerCAmelCase ): super().__init__(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = self.feature_extractor UpperCAmelCase_ = False def A__ ( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True ): return self.tokenizer.get_decoder_prompt_ids(task=lowerCAmelCase , language=lowerCAmelCase , no_timestamps=lowerCAmelCase ) def __call__( self , *lowerCAmelCase , **lowerCAmelCase ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = kwargs.pop("audio" , lowerCAmelCase ) UpperCAmelCase_ = kwargs.pop("sampling_rate" , lowerCAmelCase ) UpperCAmelCase_ = kwargs.pop("text" , lowerCAmelCase ) if len(lowerCAmelCase ) > 0: UpperCAmelCase_ = args[0] UpperCAmelCase_ = args[1:] if audio is None and text is None: raise ValueError("You need to specify either an `audio` or `text` input to process." ) if text is not None: UpperCAmelCase_ = self.tokenizer(lowerCAmelCase , **lowerCAmelCase ) if audio is not None: UpperCAmelCase_ = self.feature_extractor(lowerCAmelCase , *lowerCAmelCase , sampling_rate=lowerCAmelCase , **lowerCAmelCase ) if audio is None: return inputs elif text is None: return audio_inputs else: UpperCAmelCase_ = audio_inputs["input_values"] if "padding_mask" in audio_inputs: UpperCAmelCase_ = audio_inputs["padding_mask"] return inputs def A__ ( self , *lowerCAmelCase , **lowerCAmelCase ): UpperCAmelCase_ = kwargs.pop("audio" , lowerCAmelCase ) UpperCAmelCase_ = kwargs.pop("padding_mask" , lowerCAmelCase ) if len(lowerCAmelCase ) > 0: UpperCAmelCase_ = args[0] UpperCAmelCase_ = args[1:] if audio_values is not None: return self._decode_audio(lowerCAmelCase , padding_mask=lowerCAmelCase ) else: return self.tokenizer.batch_decode(*lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , *lowerCAmelCase , **lowerCAmelCase ): return self.tokenizer.decode(*lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = to_numpy(lowerCAmelCase ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = audio_values.shape if padding_mask is None: return list(lowerCAmelCase ) UpperCAmelCase_ = to_numpy(lowerCAmelCase ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) UpperCAmelCase_ = seq_len - padding_mask.shape[-1] UpperCAmelCase_ = 1 - self.feature_extractor.padding_value UpperCAmelCase_ = np.pad(lowerCAmelCase , ((0, 0), (0, difference)) , "constant" , constant_values=lowerCAmelCase ) UpperCAmelCase_ = audio_values.tolist() for i in range(lowerCAmelCase ): UpperCAmelCase_ = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] UpperCAmelCase_ = sliced_audio.reshape(lowerCAmelCase , -1 ) return audio_values
23
import math def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> list[int]: UpperCAmelCase_ = [] UpperCAmelCase_ = 2 UpperCAmelCase_ = int(math.sqrt(__SCREAMING_SNAKE_CASE ) ) # Size of every segment UpperCAmelCase_ = [True] * (end + 1) UpperCAmelCase_ = [] while start <= end: if temp[start] is True: in_prime.append(__SCREAMING_SNAKE_CASE ) for i in range(start * start , end + 1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = False start += 1 prime += in_prime UpperCAmelCase_ = end + 1 UpperCAmelCase_ = min(2 * end , __SCREAMING_SNAKE_CASE ) while low <= n: UpperCAmelCase_ = [True] * (high - low + 1) for each in in_prime: UpperCAmelCase_ = math.floor(low / each ) * each if t < low: t += each for j in range(__SCREAMING_SNAKE_CASE , high + 1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = False for j in range(len(__SCREAMING_SNAKE_CASE ) ): if temp[j] is True: prime.append(j + low ) UpperCAmelCase_ = high + 1 UpperCAmelCase_ = min(high + end , __SCREAMING_SNAKE_CASE ) return prime print(sieve(10**6))
23
1
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[Any]: UpperCAmelCase_ = ("dense.weight", "attention.self.query", "attention.self.key", "attention.self.value") UpperCAmelCase_ = ( ("layer.", "layer_"), ("word_embeddings.weight", "word_embeddings"), ("position_embeddings.weight", "position_embeddings"), ("token_type_embeddings.weight", "token_type_embeddings"), (".", "/"), ("LayerNorm/weight", "LayerNorm/gamma"), ("LayerNorm/bias", "LayerNorm/beta"), ("weight", "kernel"), ) if not os.path.isdir(__SCREAMING_SNAKE_CASE ): os.makedirs(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = model.state_dict() def to_tf_var_name(__SCREAMING_SNAKE_CASE ): for patt, repl in iter(__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = name.replace(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return f'''bert/{name}''' def create_tf_var(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = tf.dtypes.as_dtype(tensor.dtype ) UpperCAmelCase_ = tf.get_variable(dtype=__SCREAMING_SNAKE_CASE , shape=tensor.shape , name=__SCREAMING_SNAKE_CASE , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(__SCREAMING_SNAKE_CASE ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: UpperCAmelCase_ = to_tf_var_name(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): UpperCAmelCase_ = torch_tensor.T UpperCAmelCase_ = create_tf_var(tensor=__SCREAMING_SNAKE_CASE , name=__SCREAMING_SNAKE_CASE , session=__SCREAMING_SNAKE_CASE ) tf.keras.backend.set_value(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = session.run(__SCREAMING_SNAKE_CASE ) print(f'''Successfully created {tf_name}: {np.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )}''' ) UpperCAmelCase_ = tf.train.Saver(tf.trainable_variables() ) saver.save(__SCREAMING_SNAKE_CASE , os.path.join(__SCREAMING_SNAKE_CASE , model_name.replace("-" , "_" ) + ".ckpt" ) ) def snake_case__ ( __SCREAMING_SNAKE_CASE=None ) -> List[Any]: UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument("--model_name" , type=__SCREAMING_SNAKE_CASE , required=__SCREAMING_SNAKE_CASE , help="model name e.g. bert-base-uncased" ) parser.add_argument( "--cache_dir" , type=__SCREAMING_SNAKE_CASE , default=__SCREAMING_SNAKE_CASE , required=__SCREAMING_SNAKE_CASE , help="Directory containing pytorch model" ) parser.add_argument("--pytorch_model_path" , type=__SCREAMING_SNAKE_CASE , required=__SCREAMING_SNAKE_CASE , help="/path/to/<pytorch-model-name>.bin" ) parser.add_argument("--tf_cache_dir" , type=__SCREAMING_SNAKE_CASE , required=__SCREAMING_SNAKE_CASE , help="Directory in which to save tensorflow model" ) UpperCAmelCase_ = parser.parse_args(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=__SCREAMING_SNAKE_CASE , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
23
from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : torch.FloatTensor class lowerCamelCase ( lowercase__, lowercase__ ): '''simple docstring''' @register_to_config def __init__( self , lowerCAmelCase = 32 , lowerCAmelCase = 64 , lowerCAmelCase = 20 , lowerCAmelCase = 768 , lowerCAmelCase=77 , lowerCAmelCase=4 , lowerCAmelCase = 0.0 , lowerCAmelCase = "silu" , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = "linear" , lowerCAmelCase = "prd" , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , ): super().__init__() UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = attention_head_dim UpperCAmelCase_ = num_attention_heads * attention_head_dim UpperCAmelCase_ = additional_embeddings UpperCAmelCase_ = time_embed_dim or inner_dim UpperCAmelCase_ = embedding_proj_dim or embedding_dim UpperCAmelCase_ = clip_embed_dim or embedding_dim UpperCAmelCase_ = Timesteps(lowerCAmelCase , lowerCAmelCase , 0 ) UpperCAmelCase_ = TimestepEmbedding(lowerCAmelCase , lowerCAmelCase , out_dim=lowerCAmelCase , act_fn=lowerCAmelCase ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) if embedding_proj_norm_type is None: UpperCAmelCase_ = None elif embedding_proj_norm_type == "layer": UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) else: raise ValueError(f'''unsupported embedding_proj_norm_type: {embedding_proj_norm_type}''' ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) if encoder_hid_proj_type is None: UpperCAmelCase_ = None elif encoder_hid_proj_type == "linear": UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) else: raise ValueError(f'''unsupported encoder_hid_proj_type: {encoder_hid_proj_type}''' ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , num_embeddings + additional_embeddings , lowerCAmelCase ) ) if added_emb_type == "prd": UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , 1 , lowerCAmelCase ) ) elif added_emb_type is None: UpperCAmelCase_ = None else: raise ValueError( f'''`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `\'prd\'` or `None`.''' ) UpperCAmelCase_ = nn.ModuleList( [ BasicTransformerBlock( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , dropout=lowerCAmelCase , activation_fn="gelu" , attention_bias=lowerCAmelCase , ) for d in range(lowerCAmelCase ) ] ) if norm_in_type == "layer": UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) elif norm_in_type is None: UpperCAmelCase_ = None else: raise ValueError(f'''Unsupported norm_in_type: {norm_in_type}.''' ) UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] , -10000.0 ) causal_attention_mask.triu_(1 ) UpperCAmelCase_ = causal_attention_mask[None, ...] self.register_buffer("causal_attention_mask" , lowerCAmelCase , persistent=lowerCAmelCase ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def A__ ( self ): UpperCAmelCase_ = {} def fn_recursive_add_processors(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if hasattr(lowerCAmelCase , "set_processor" ): UpperCAmelCase_ = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f'''{name}.{sub_name}''' , lowerCAmelCase , lowerCAmelCase ) return processors for name, module in self.named_children(): fn_recursive_add_processors(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return processors def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = len(self.attn_processors.keys() ) if isinstance(lowerCAmelCase , lowerCAmelCase ) and len(lowerCAmelCase ) != count: raise ValueError( f'''A dict of processors was passed, but the number of processors {len(lowerCAmelCase )} does not match the''' f''' number of attention layers: {count}. Please make sure to pass {count} processor classes.''' ) def fn_recursive_attn_processor(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if hasattr(lowerCAmelCase , "set_processor" ): if not isinstance(lowerCAmelCase , lowerCAmelCase ): module.set_processor(lowerCAmelCase ) else: module.set_processor(processor.pop(f'''{name}.processor''' ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f'''{name}.{sub_name}''' , lowerCAmelCase , lowerCAmelCase ) for name, module in self.named_children(): fn_recursive_attn_processor(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): self.set_attn_processor(AttnProcessor() ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = True , ): UpperCAmelCase_ = hidden_states.shape[0] UpperCAmelCase_ = timestep if not torch.is_tensor(lowerCAmelCase ): UpperCAmelCase_ = torch.tensor([timesteps] , dtype=torch.long , device=hidden_states.device ) elif torch.is_tensor(lowerCAmelCase ) and len(timesteps.shape ) == 0: UpperCAmelCase_ = timesteps[None].to(hidden_states.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML UpperCAmelCase_ = timesteps * torch.ones(lowerCAmelCase , dtype=timesteps.dtype , device=timesteps.device ) UpperCAmelCase_ = self.time_proj(lowerCAmelCase ) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. UpperCAmelCase_ = timesteps_projected.to(dtype=self.dtype ) UpperCAmelCase_ = self.time_embedding(lowerCAmelCase ) if self.embedding_proj_norm is not None: UpperCAmelCase_ = self.embedding_proj_norm(lowerCAmelCase ) UpperCAmelCase_ = self.embedding_proj(lowerCAmelCase ) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: UpperCAmelCase_ = self.encoder_hidden_states_proj(lowerCAmelCase ) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError("`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set" ) UpperCAmelCase_ = self.proj_in(lowerCAmelCase ) UpperCAmelCase_ = self.positional_embedding.to(hidden_states.dtype ) UpperCAmelCase_ = [] UpperCAmelCase_ = 0 if encoder_hidden_states is not None: additional_embeds.append(lowerCAmelCase ) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape ) == 2: UpperCAmelCase_ = proj_embeddings[:, None, :] if len(hidden_states.shape ) == 2: UpperCAmelCase_ = hidden_states[:, None, :] UpperCAmelCase_ = additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: UpperCAmelCase_ = self.prd_embedding.to(hidden_states.dtype ).expand(lowerCAmelCase , -1 , -1 ) additional_embeds.append(lowerCAmelCase ) UpperCAmelCase_ = torch.cat( lowerCAmelCase , dim=1 , ) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens UpperCAmelCase_ = additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: UpperCAmelCase_ = F.pad( lowerCAmelCase , ( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) , value=0.0 , ) UpperCAmelCase_ = hidden_states + positional_embeddings if attention_mask is not None: UpperCAmelCase_ = (1 - attention_mask.to(hidden_states.dtype )) * -10000.0 UpperCAmelCase_ = F.pad(lowerCAmelCase , (0, self.additional_embeddings) , value=0.0 ) UpperCAmelCase_ = (attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype ) UpperCAmelCase_ = attention_mask.repeat_interleave(self.config.num_attention_heads , dim=0 ) if self.norm_in is not None: UpperCAmelCase_ = self.norm_in(lowerCAmelCase ) for block in self.transformer_blocks: UpperCAmelCase_ = block(lowerCAmelCase , attention_mask=lowerCAmelCase ) UpperCAmelCase_ = self.norm_out(lowerCAmelCase ) if self.prd_embedding is not None: UpperCAmelCase_ = hidden_states[:, -1] else: UpperCAmelCase_ = hidden_states[:, additional_embeddings_len:] UpperCAmelCase_ = self.proj_to_clip_embeddings(lowerCAmelCase ) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (prior_latents * self.clip_std) + self.clip_mean return prior_latents
23
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE = {"configuration_sew": ["SEW_PRETRAINED_CONFIG_ARCHIVE_MAP", "SEWConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "SEW_PRETRAINED_MODEL_ARCHIVE_LIST", "SEWForCTC", "SEWForSequenceClassification", "SEWModel", "SEWPreTrainedModel", ] if TYPE_CHECKING: from .configuration_sew import SEW_PRETRAINED_CONFIG_ARCHIVE_MAP, SEWConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_sew import ( SEW_PRETRAINED_MODEL_ARCHIVE_LIST, SEWForCTC, SEWForSequenceClassification, SEWModel, SEWPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE = { "configuration_gpt_bigcode": ["GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTBigCodeConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTBigCodeForSequenceClassification", "GPTBigCodeForTokenClassification", "GPTBigCodeForCausalLM", "GPTBigCodeModel", "GPTBigCodePreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
import math from datetime import datetime, timedelta def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> datetime: UpperCAmelCase_ = year % 19 UpperCAmelCase_ = year % 4 UpperCAmelCase_ = year % 7 UpperCAmelCase_ = math.floor(year / 100 ) UpperCAmelCase_ = math.floor((13 + 8 * leap_day_inhibits) / 25 ) UpperCAmelCase_ = leap_day_inhibits / 4 UpperCAmelCase_ = ( 15 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 30 UpperCAmelCase_ = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 UpperCAmelCase_ = (19 * metonic_cycle + secular_moon_shift) % 30 # PHM -> Paschal Full Moon UpperCAmelCase_ = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 29 and days_from_phm_to_sunday == 6: return datetime(__SCREAMING_SNAKE_CASE , 4 , 19 ) elif days_to_add == 28 and days_from_phm_to_sunday == 6: return datetime(__SCREAMING_SNAKE_CASE , 4 , 18 ) else: return datetime(__SCREAMING_SNAKE_CASE , 3 , 22 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (1994, 2000, 2010, 2021, 2023): SCREAMING_SNAKE_CASE = "will be" if year > datetime.now().year else "was" print(f'''Easter in {year} {tense} {gauss_easter(year)}''')
23
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "xlm-roberta-base": "https://huggingface.co/xlm-roberta-base/resolve/main/config.json", "xlm-roberta-large": "https://huggingface.co/xlm-roberta-large/resolve/main/config.json", "xlm-roberta-large-finetuned-conll02-dutch": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll02-spanish": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-english": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-german": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json" ), } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = 'xlm-roberta' def __init__( self , lowerCAmelCase=3_0522 , lowerCAmelCase=768 , lowerCAmelCase=12 , lowerCAmelCase=12 , lowerCAmelCase=3072 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=1e-1_2 , lowerCAmelCase=1 , lowerCAmelCase=0 , lowerCAmelCase=2 , lowerCAmelCase="absolute" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__(pad_token_id=lowerCAmelCase , bos_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = position_embedding_type UpperCAmelCase_ = use_cache UpperCAmelCase_ = classifier_dropout class lowerCamelCase ( lowercase__ ): '''simple docstring''' @property def A__ ( self ): if self.task == "multiple-choice": UpperCAmelCase_ = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
23
1
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = int(__SCREAMING_SNAKE_CASE ) if decimal in (0, 1): # Exit cases for the recursion return str(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ , UpperCAmelCase_ = divmod(__SCREAMING_SNAKE_CASE , 2 ) return binary_recursive(__SCREAMING_SNAKE_CASE ) + str(__SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = str(__SCREAMING_SNAKE_CASE ).strip() if not number: raise ValueError("No input value was provided" ) UpperCAmelCase_ = "-" if number.startswith("-" ) else "" UpperCAmelCase_ = number.lstrip("-" ) if not number.isnumeric(): raise ValueError("Input value is not an integer" ) return f'''{negative}0b{binary_recursive(int(__SCREAMING_SNAKE_CASE ) )}''' if __name__ == "__main__": from doctest import testmod testmod()
23
1
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import Callable, Dict, List, Tuple import timm import torch import torch.nn as nn from classy_vision.models.regnet import RegNet, RegNetParams, RegNetYaagf, RegNetYaagf, RegNetYaaagf from huggingface_hub import cached_download, hf_hub_url from torch import Tensor from vissl.models.model_helpers import get_trunk_forward_outputs from transformers import AutoImageProcessor, RegNetConfig, RegNetForImageClassification, RegNetModel from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE = logging.get_logger() @dataclass class lowerCamelCase : '''simple docstring''' lowerCAmelCase_ : nn.Module lowerCAmelCase_ : List[nn.Module] = field(default_factory=lowercase__ ) lowerCAmelCase_ : list = field(default_factory=lowercase__ ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = len(list(m.modules() ) ) == 1 or isinstance(lowerCAmelCase , nn.Convad ) or isinstance(lowerCAmelCase , nn.BatchNormad ) if has_not_submodules: self.traced.append(lowerCAmelCase ) def __call__( self , lowerCAmelCase ): for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(lowerCAmelCase ) [x.remove() for x in self.handles] return self @property def A__ ( self ): # check the len of the state_dict keys to see if we have learnable params return list(filter(lambda lowerCAmelCase : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class lowerCamelCase : '''simple docstring''' lowerCAmelCase_ : nn.Module lowerCAmelCase_ : nn.Module lowerCAmelCase_ : int = 1 lowerCAmelCase_ : List = field(default_factory=lowercase__ ) lowerCAmelCase_ : List = field(default_factory=lowercase__ ) lowerCAmelCase_ : bool = True def __call__( self , lowerCAmelCase ): UpperCAmelCase_ = Tracker(self.dest )(lowerCAmelCase ).parametrized UpperCAmelCase_ = Tracker(self.src )(lowerCAmelCase ).parametrized UpperCAmelCase_ = list(filter(lambda lowerCAmelCase : type(lowerCAmelCase ) not in self.src_skip , lowerCAmelCase ) ) UpperCAmelCase_ = list(filter(lambda lowerCAmelCase : type(lowerCAmelCase ) not in self.dest_skip , lowerCAmelCase ) ) if len(lowerCAmelCase ) != len(lowerCAmelCase ) and self.raise_if_mismatch: raise Exception( f'''Numbers of operations are different. Source module has {len(lowerCAmelCase )} operations while''' f''' destination module has {len(lowerCAmelCase )}.''' ) for dest_m, src_m in zip(lowerCAmelCase , lowerCAmelCase ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(f'''Transfered from={src_m} to={dest_m}''' ) class lowerCamelCase ( nn.Module ): '''simple docstring''' def __init__( self , lowerCAmelCase ): super().__init__() UpperCAmelCase_ = [] # - get the stem feature_blocks.append(("conv1", model.stem) ) # - get all the feature blocks for k, v in model.trunk_output.named_children(): assert k.startswith("block" ), f'''Unexpected layer name {k}''' UpperCAmelCase_ = len(lowerCAmelCase ) + 1 feature_blocks.append((f'''res{block_index}''', v) ) UpperCAmelCase_ = nn.ModuleDict(lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): return get_trunk_forward_outputs( lowerCAmelCase , out_feat_keys=lowerCAmelCase , feature_blocks=self._feature_blocks , ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = x.split("-" ) return x_split[0] + x_split[1] + "_" + "".join(x_split[2:] ) def __getitem__( self , lowerCAmelCase ): # default to timm! if x not in self: UpperCAmelCase_ = self.convert_name_to_timm(lowerCAmelCase ) UpperCAmelCase_ = partial(lambda: (timm.create_model(lowerCAmelCase , pretrained=lowerCAmelCase ).eval(), None) ) else: UpperCAmelCase_ = super().__getitem__(lowerCAmelCase ) return val class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __getitem__( self , lowerCAmelCase ): if "seer" in x and "in1k" not in x: UpperCAmelCase_ = RegNetModel else: UpperCAmelCase_ = RegNetForImageClassification return val def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Tuple: for from_key, to_key in keys: UpperCAmelCase_ = from_state_dict[from_key].clone() print(f'''Copied key={from_key} to={to_key}''' ) return to_state_dict def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = True , ) -> Optional[Any]: print(f'''Converting {name}...''' ) with torch.no_grad(): UpperCAmelCase_ , UpperCAmelCase_ = from_model_func() UpperCAmelCase_ = our_model_func(__SCREAMING_SNAKE_CASE ).eval() UpperCAmelCase_ = ModuleTransfer(src=__SCREAMING_SNAKE_CASE , dest=__SCREAMING_SNAKE_CASE , raise_if_mismatch=__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = torch.randn((1, 3, 224, 224) ) module_transfer(__SCREAMING_SNAKE_CASE ) if from_state_dict is not None: UpperCAmelCase_ = [] # for seer - in1k finetuned we have to manually copy the head if "seer" in name and "in1k" in name: UpperCAmelCase_ = [("0.clf.0.weight", "classifier.1.weight"), ("0.clf.0.bias", "classifier.1.bias")] UpperCAmelCase_ = manually_copy_vissl_head(__SCREAMING_SNAKE_CASE , our_model.state_dict() , __SCREAMING_SNAKE_CASE ) our_model.load_state_dict(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = our_model(__SCREAMING_SNAKE_CASE , output_hidden_states=__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = ( our_outputs.logits if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) else our_outputs.last_hidden_state ) UpperCAmelCase_ = from_model(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = from_output[-1] if type(__SCREAMING_SNAKE_CASE ) is list else from_output # now since I don't want to use any config files, vissl seer model doesn't actually have an head, so let's just check the last hidden state if "seer" in name and "in1k" in name: UpperCAmelCase_ = our_outputs.hidden_states[-1] assert torch.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ), "The model logits don't match the original one." if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / name , commit_message="Add model" , use_temp_dir=__SCREAMING_SNAKE_CASE , ) UpperCAmelCase_ = 224 if "seer" not in name else 384 # we can use the convnext one UpperCAmelCase_ = AutoImageProcessor.from_pretrained("facebook/convnext-base-224-22k-1k" , size=__SCREAMING_SNAKE_CASE ) image_processor.push_to_hub( repo_path_or_name=save_directory / name , commit_message="Add image processor" , use_temp_dir=__SCREAMING_SNAKE_CASE , ) print(f'''Pushed {name}''' ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = True ) -> int: UpperCAmelCase_ = "imagenet-1k-id2label.json" UpperCAmelCase_ = 1000 UpperCAmelCase_ = (1, num_labels) UpperCAmelCase_ = "huggingface/label-files" UpperCAmelCase_ = num_labels UpperCAmelCase_ = json.load(open(cached_download(hf_hub_url(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , repo_type="dataset" ) ) , "r" ) ) UpperCAmelCase_ = {int(__SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} UpperCAmelCase_ = idalabel UpperCAmelCase_ = {v: k for k, v in idalabel.items()} UpperCAmelCase_ = partial(__SCREAMING_SNAKE_CASE , num_labels=__SCREAMING_SNAKE_CASE , idalabel=__SCREAMING_SNAKE_CASE , labelaid=__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = { "regnet-x-002": ImageNetPreTrainedConfig( depths=[1, 1, 4, 7] , hidden_sizes=[24, 56, 152, 368] , groups_width=8 , layer_type="x" ), "regnet-x-004": ImageNetPreTrainedConfig( depths=[1, 2, 7, 12] , hidden_sizes=[32, 64, 160, 384] , groups_width=16 , layer_type="x" ), "regnet-x-006": ImageNetPreTrainedConfig( depths=[1, 3, 5, 7] , hidden_sizes=[48, 96, 240, 528] , groups_width=24 , layer_type="x" ), "regnet-x-008": ImageNetPreTrainedConfig( depths=[1, 3, 7, 5] , hidden_sizes=[64, 128, 288, 672] , groups_width=16 , layer_type="x" ), "regnet-x-016": ImageNetPreTrainedConfig( depths=[2, 4, 10, 2] , hidden_sizes=[72, 168, 408, 912] , groups_width=24 , layer_type="x" ), "regnet-x-032": ImageNetPreTrainedConfig( depths=[2, 6, 15, 2] , hidden_sizes=[96, 192, 432, 1008] , groups_width=48 , layer_type="x" ), "regnet-x-040": ImageNetPreTrainedConfig( depths=[2, 5, 14, 2] , hidden_sizes=[80, 240, 560, 1360] , groups_width=40 , layer_type="x" ), "regnet-x-064": ImageNetPreTrainedConfig( depths=[2, 4, 10, 1] , hidden_sizes=[168, 392, 784, 1624] , groups_width=56 , layer_type="x" ), "regnet-x-080": ImageNetPreTrainedConfig( depths=[2, 5, 15, 1] , hidden_sizes=[80, 240, 720, 1920] , groups_width=120 , layer_type="x" ), "regnet-x-120": ImageNetPreTrainedConfig( depths=[2, 5, 11, 1] , hidden_sizes=[224, 448, 896, 2240] , groups_width=112 , layer_type="x" ), "regnet-x-160": ImageNetPreTrainedConfig( depths=[2, 6, 13, 1] , hidden_sizes=[256, 512, 896, 2048] , groups_width=128 , layer_type="x" ), "regnet-x-320": ImageNetPreTrainedConfig( depths=[2, 7, 13, 1] , hidden_sizes=[336, 672, 1344, 2520] , groups_width=168 , layer_type="x" ), # y variant "regnet-y-002": ImageNetPreTrainedConfig(depths=[1, 1, 4, 7] , hidden_sizes=[24, 56, 152, 368] , groups_width=8 ), "regnet-y-004": ImageNetPreTrainedConfig( depths=[1, 3, 6, 6] , hidden_sizes=[48, 104, 208, 440] , groups_width=8 ), "regnet-y-006": ImageNetPreTrainedConfig( depths=[1, 3, 7, 4] , hidden_sizes=[48, 112, 256, 608] , groups_width=16 ), "regnet-y-008": ImageNetPreTrainedConfig( depths=[1, 3, 8, 2] , hidden_sizes=[64, 128, 320, 768] , groups_width=16 ), "regnet-y-016": ImageNetPreTrainedConfig( depths=[2, 6, 17, 2] , hidden_sizes=[48, 120, 336, 888] , groups_width=24 ), "regnet-y-032": ImageNetPreTrainedConfig( depths=[2, 5, 13, 1] , hidden_sizes=[72, 216, 576, 1512] , groups_width=24 ), "regnet-y-040": ImageNetPreTrainedConfig( depths=[2, 6, 12, 2] , hidden_sizes=[128, 192, 512, 1088] , groups_width=64 ), "regnet-y-064": ImageNetPreTrainedConfig( depths=[2, 7, 14, 2] , hidden_sizes=[144, 288, 576, 1296] , groups_width=72 ), "regnet-y-080": ImageNetPreTrainedConfig( depths=[2, 4, 10, 1] , hidden_sizes=[168, 448, 896, 2016] , groups_width=56 ), "regnet-y-120": ImageNetPreTrainedConfig( depths=[2, 5, 11, 1] , hidden_sizes=[224, 448, 896, 2240] , groups_width=112 ), "regnet-y-160": ImageNetPreTrainedConfig( depths=[2, 4, 11, 1] , hidden_sizes=[224, 448, 1232, 3024] , groups_width=112 ), "regnet-y-320": ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] , hidden_sizes=[232, 696, 1392, 3712] , groups_width=232 ), # models created by SEER -> https://arxiv.org/abs/2202.08360 "regnet-y-320-seer": RegNetConfig(depths=[2, 5, 12, 1] , hidden_sizes=[232, 696, 1392, 3712] , groups_width=232 ), "regnet-y-640-seer": RegNetConfig(depths=[2, 5, 12, 1] , hidden_sizes=[328, 984, 1968, 4920] , groups_width=328 ), "regnet-y-1280-seer": RegNetConfig( depths=[2, 7, 17, 1] , hidden_sizes=[528, 1056, 2904, 7392] , groups_width=264 ), "regnet-y-2560-seer": RegNetConfig( depths=[3, 7, 16, 1] , hidden_sizes=[640, 1696, 2544, 5088] , groups_width=640 ), "regnet-y-10b-seer": ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] , hidden_sizes=[2020, 4040, 1_1110, 2_8280] , groups_width=1010 ), # finetuned on imagenet "regnet-y-320-seer-in1k": ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] , hidden_sizes=[232, 696, 1392, 3712] , groups_width=232 ), "regnet-y-640-seer-in1k": ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] , hidden_sizes=[328, 984, 1968, 4920] , groups_width=328 ), "regnet-y-1280-seer-in1k": ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] , hidden_sizes=[528, 1056, 2904, 7392] , groups_width=264 ), "regnet-y-2560-seer-in1k": ImageNetPreTrainedConfig( depths=[3, 7, 16, 1] , hidden_sizes=[640, 1696, 2544, 5088] , groups_width=640 ), "regnet-y-10b-seer-in1k": ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] , hidden_sizes=[2020, 4040, 1_1110, 2_8280] , groups_width=1010 ), } UpperCAmelCase_ = NameToOurModelFuncMap() UpperCAmelCase_ = NameToFromModelFuncMap() # add seer weights logic def load_using_classy_vision(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Tuple[nn.Module, Dict]: UpperCAmelCase_ = torch.hub.load_state_dict_from_url(__SCREAMING_SNAKE_CASE , model_dir=str(__SCREAMING_SNAKE_CASE ) , map_location="cpu" ) UpperCAmelCase_ = model_func() # check if we have a head, if yes add it UpperCAmelCase_ = files["classy_state_dict"]["base_model"]["model"] UpperCAmelCase_ = model_state_dict["trunk"] model.load_state_dict(__SCREAMING_SNAKE_CASE ) return model.eval(), model_state_dict["heads"] # pretrained UpperCAmelCase_ = partial( __SCREAMING_SNAKE_CASE , "https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet32d/seer_regnet32gf_model_iteration244000.torch" , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) UpperCAmelCase_ = partial( __SCREAMING_SNAKE_CASE , "https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet64/seer_regnet64gf_model_final_checkpoint_phase0.torch" , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) UpperCAmelCase_ = partial( __SCREAMING_SNAKE_CASE , "https://dl.fbaipublicfiles.com/vissl/model_zoo/swav_ig1b_regnet128Gf_cnstant_bs32_node16_sinkhorn10_proto16k_syncBN64_warmup8k/model_final_checkpoint_phase0.torch" , lambda: FakeRegNetVisslWrapper(RegNetYaaagf() ) , ) UpperCAmelCase_ = partial( __SCREAMING_SNAKE_CASE , "https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet10B/model_iteration124500_conso.torch" , lambda: FakeRegNetVisslWrapper( RegNet(RegNetParams(depth=27 , group_width=1010 , w_a=1744 , w_a=620.83 , w_m=2.52 ) ) ) , ) # IN1K finetuned UpperCAmelCase_ = partial( __SCREAMING_SNAKE_CASE , "https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet32_finetuned_in1k_model_final_checkpoint_phase78.torch" , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) UpperCAmelCase_ = partial( __SCREAMING_SNAKE_CASE , "https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet64_finetuned_in1k_model_final_checkpoint_phase78.torch" , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) UpperCAmelCase_ = partial( __SCREAMING_SNAKE_CASE , "https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet128_finetuned_in1k_model_final_checkpoint_phase78.torch" , lambda: FakeRegNetVisslWrapper(RegNetYaaagf() ) , ) UpperCAmelCase_ = partial( __SCREAMING_SNAKE_CASE , "https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_10b_finetuned_in1k_model_phase28_conso.torch" , lambda: FakeRegNetVisslWrapper( RegNet(RegNetParams(depth=27 , group_width=1010 , w_a=1744 , w_a=620.83 , w_m=2.52 ) ) ) , ) if model_name: convert_weight_and_push( __SCREAMING_SNAKE_CASE , names_to_from_model_map[model_name] , names_to_ours_model_map[model_name] , names_to_config[model_name] , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , ) else: for model_name, config in names_to_config.items(): convert_weight_and_push( __SCREAMING_SNAKE_CASE , names_to_from_model_map[model_name] , names_to_ours_model_map[model_name] , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , ) return config, expected_shape if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default=None, type=str, help=( "The name of the model you wish to convert, it must be one of the supported regnet* architecture," " currently: regnetx-*, regnety-*. If `None`, all of them will the converted." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=Path, required=True, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", default=True, type=bool, required=False, help="If True, push model and image processor to the hub.", ) SCREAMING_SNAKE_CASE = parser.parse_args() SCREAMING_SNAKE_CASE = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
23
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self ): UpperCAmelCase_ = AutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" , return_dict=lowerCAmelCase ).to(lowerCAmelCase ) UpperCAmelCase_ = AutoTokenizer.from_pretrained("google/mt5-small" ) UpperCAmelCase_ = tokenizer("Hello there" , return_tensors="pt" ).input_ids UpperCAmelCase_ = tokenizer("Hi I am" , return_tensors="pt" ).input_ids UpperCAmelCase_ = model(input_ids.to(lowerCAmelCase ) , labels=labels.to(lowerCAmelCase ) ).loss UpperCAmelCase_ = -(labels.shape[-1] * loss.item()) UpperCAmelCase_ = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
23
1
import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class lowerCamelCase ( enum.Enum ): '''simple docstring''' lowerCAmelCase_ : str = 0 lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : List[str] = 2 @add_end_docstrings(lowercase__ ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : str = '\n In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The\n voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western\n Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision\n and denounces one of the men as a horse thief. Although his father initially slaps him for making such an\n accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of\n the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop,\n begging for his blessing. <eod> </s> <eos>\n ' def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): super().__init__(*lowerCAmelCase , **lowerCAmelCase ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. UpperCAmelCase_ = None if self.model.config.prefix is not None: UpperCAmelCase_ = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. UpperCAmelCase_ = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self._sanitize_parameters(prefix=lowerCAmelCase , **self._forward_params ) UpperCAmelCase_ = {**self._preprocess_params, **preprocess_params} UpperCAmelCase_ = {**self._forward_params, **forward_params} def A__ ( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , **lowerCAmelCase , ): UpperCAmelCase_ = {} if prefix is not None: UpperCAmelCase_ = prefix if prefix: UpperCAmelCase_ = self.tokenizer( lowerCAmelCase , padding=lowerCAmelCase , add_special_tokens=lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = prefix_inputs["input_ids"].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f'''{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected''' " [None, 'hole']" ) UpperCAmelCase_ = handle_long_generation preprocess_params.update(lowerCAmelCase ) UpperCAmelCase_ = generate_kwargs UpperCAmelCase_ = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_full_text`" ) if return_tensors is not None: raise ValueError("`return_full_text` is mutually exclusive with `return_tensors`" ) UpperCAmelCase_ = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_tensors`" ) UpperCAmelCase_ = ReturnType.TENSORS if return_type is not None: UpperCAmelCase_ = return_type if clean_up_tokenization_spaces is not None: UpperCAmelCase_ = clean_up_tokenization_spaces if stop_sequence is not None: UpperCAmelCase_ = self.tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) if len(lowerCAmelCase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) UpperCAmelCase_ = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def A__ ( self , *lowerCAmelCase , **lowerCAmelCase ): # Parse arguments if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"add_space_before_punct_symbol": True} ) return super()._parse_and_tokenize(*lowerCAmelCase , **lowerCAmelCase ) def __call__( self , lowerCAmelCase , **lowerCAmelCase ): return super().__call__(lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase="" , lowerCAmelCase=None , **lowerCAmelCase ): UpperCAmelCase_ = self.tokenizer( prefix + prompt_text , padding=lowerCAmelCase , add_special_tokens=lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = prompt_text if handle_long_generation == "hole": UpperCAmelCase_ = inputs["input_ids"].shape[-1] if "max_new_tokens" in generate_kwargs: UpperCAmelCase_ = generate_kwargs["max_new_tokens"] else: UpperCAmelCase_ = generate_kwargs.get("max_length" , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError("We cannot infer how many new tokens are expected" ) if cur_len + new_tokens > self.tokenizer.model_max_length: UpperCAmelCase_ = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( "We cannot use `hole` to handle this generation the number of desired tokens exceeds the" " models max length" ) UpperCAmelCase_ = inputs["input_ids"][:, -keep_length:] if "attention_mask" in inputs: UpperCAmelCase_ = inputs["attention_mask"][:, -keep_length:] return inputs def A__ ( self , lowerCAmelCase , **lowerCAmelCase ): UpperCAmelCase_ = model_inputs["input_ids"] UpperCAmelCase_ = model_inputs.get("attention_mask" , lowerCAmelCase ) # Allow empty prompts if input_ids.shape[1] == 0: UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = 1 else: UpperCAmelCase_ = input_ids.shape[0] UpperCAmelCase_ = model_inputs.pop("prompt_text" ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. UpperCAmelCase_ = generate_kwargs.pop("prefix_length" , 0 ) if prefix_length > 0: UpperCAmelCase_ = "max_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].max_new_tokens is not None ) if not has_max_new_tokens: UpperCAmelCase_ = generate_kwargs.get("max_length" ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length UpperCAmelCase_ = "min_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL UpperCAmelCase_ = self.model.generate(input_ids=lowerCAmelCase , attention_mask=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = generated_sequence.shape[0] if self.framework == "pt": UpperCAmelCase_ = generated_sequence.reshape(lowerCAmelCase , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": UpperCAmelCase_ = tf.reshape(lowerCAmelCase , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def A__ ( self , lowerCAmelCase , lowerCAmelCase=ReturnType.FULL_TEXT , lowerCAmelCase=True ): UpperCAmelCase_ = model_outputs["generated_sequence"][0] UpperCAmelCase_ = model_outputs["input_ids"] UpperCAmelCase_ = model_outputs["prompt_text"] UpperCAmelCase_ = generated_sequence.numpy().tolist() UpperCAmelCase_ = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: UpperCAmelCase_ = {"generated_token_ids": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text UpperCAmelCase_ = self.tokenizer.decode( lowerCAmelCase , skip_special_tokens=lowerCAmelCase , clean_up_tokenization_spaces=lowerCAmelCase , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: UpperCAmelCase_ = 0 else: UpperCAmelCase_ = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=lowerCAmelCase , clean_up_tokenization_spaces=lowerCAmelCase , ) ) if return_type == ReturnType.FULL_TEXT: UpperCAmelCase_ = prompt_text + text[prompt_length:] else: UpperCAmelCase_ = text[prompt_length:] UpperCAmelCase_ = {"generated_text": all_text} records.append(lowerCAmelCase ) return records
23
def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ = 0 while b > 0: if b & 1: UpperCAmelCase_ = ((res % c) + (a % c)) % c a += a b >>= 1 return res
23
1
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : torch.FloatTensor lowerCAmelCase_ : torch.FloatTensor class lowerCamelCase ( lowercase__, lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = 1 @register_to_config def __init__( self , lowerCAmelCase = 2000 , lowerCAmelCase = 0.15 , lowerCAmelCase = 0.01 , lowerCAmelCase = 1348.0 , lowerCAmelCase = 1e-5 , lowerCAmelCase = 1 , ): # standard deviation of the initial noise distribution UpperCAmelCase_ = sigma_max # setable values UpperCAmelCase_ = None self.set_sigmas(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): return sample def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None ): UpperCAmelCase_ = sampling_eps if sampling_eps is not None else self.config.sampling_eps UpperCAmelCase_ = torch.linspace(1 , lowerCAmelCase , lowerCAmelCase , device=lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None ): UpperCAmelCase_ = sigma_min if sigma_min is not None else self.config.sigma_min UpperCAmelCase_ = sigma_max if sigma_max is not None else self.config.sigma_max UpperCAmelCase_ = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) UpperCAmelCase_ = torch.exp(torch.linspace(math.log(lowerCAmelCase ) , math.log(lowerCAmelCase ) , lowerCAmelCase ) ) UpperCAmelCase_ = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ): return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = True , ): if self.timesteps is None: raise ValueError( "`self.timesteps` is not set, you need to run 'set_timesteps' after creating the scheduler" ) UpperCAmelCase_ = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) UpperCAmelCase_ = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda UpperCAmelCase_ = timesteps.to(self.discrete_sigmas.device ) UpperCAmelCase_ = self.discrete_sigmas[timesteps].to(sample.device ) UpperCAmelCase_ = self.get_adjacent_sigma(lowerCAmelCase , lowerCAmelCase ).to(sample.device ) UpperCAmelCase_ = torch.zeros_like(lowerCAmelCase ) UpperCAmelCase_ = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods UpperCAmelCase_ = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): UpperCAmelCase_ = diffusion.unsqueeze(-1 ) UpperCAmelCase_ = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of UpperCAmelCase_ = randn_tensor( sample.shape , layout=sample.layout , generator=lowerCAmelCase , device=sample.device , dtype=sample.dtype ) UpperCAmelCase_ = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? UpperCAmelCase_ = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=lowerCAmelCase , prev_sample_mean=lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = True , ): if self.timesteps is None: raise ValueError( "`self.timesteps` is not set, you need to run 'set_timesteps' after creating the scheduler" ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction UpperCAmelCase_ = randn_tensor(sample.shape , layout=sample.layout , generator=lowerCAmelCase ).to(sample.device ) # compute step size from the model_output, the noise, and the snr UpperCAmelCase_ = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() UpperCAmelCase_ = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() UpperCAmelCase_ = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 UpperCAmelCase_ = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term UpperCAmelCase_ = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): UpperCAmelCase_ = step_size.unsqueeze(-1 ) UpperCAmelCase_ = sample + step_size * model_output UpperCAmelCase_ = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , ): # Make sure sigmas and timesteps have the same device and dtype as original_samples UpperCAmelCase_ = timesteps.to(original_samples.device ) UpperCAmelCase_ = self.discrete_sigmas.to(original_samples.device )[timesteps] UpperCAmelCase_ = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(lowerCAmelCase ) * sigmas[:, None, None, None] ) UpperCAmelCase_ = noise + original_samples return noisy_samples def __len__( self ): return self.config.num_train_timesteps
23
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> List[List[ImageInput]]: if isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(__SCREAMING_SNAKE_CASE ): return [[videos]] raise ValueError(f'''Could not make batched video from {videos}''' ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[Any] = ['pixel_values'] def __init__( self , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = True , lowerCAmelCase = 1 / 255 , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = None , **lowerCAmelCase , ): super().__init__(**lowerCAmelCase ) UpperCAmelCase_ = size if size is not None else {"shortest_edge": 224} UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) UpperCAmelCase_ = crop_size if crop_size is not None else {"height": 224, "width": 224} UpperCAmelCase_ = get_size_dict(lowerCAmelCase , param_name="crop_size" ) UpperCAmelCase_ = do_resize UpperCAmelCase_ = size UpperCAmelCase_ = do_center_crop UpperCAmelCase_ = crop_size UpperCAmelCase_ = resample UpperCAmelCase_ = do_rescale UpperCAmelCase_ = rescale_factor UpperCAmelCase_ = do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) if "shortest_edge" in size: UpperCAmelCase_ = get_resize_output_image_size(lowerCAmelCase , size["shortest_edge"] , default_to_square=lowerCAmelCase ) elif "height" in size and "width" in size: UpperCAmelCase_ = (size["height"], size["width"]) else: raise ValueError(f'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = get_size_dict(lowerCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(lowerCAmelCase , size=(size["height"], size["width"]) , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return rescale(lowerCAmelCase , scale=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return normalize(lowerCAmelCase , mean=lowerCAmelCase , std=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , ): if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. UpperCAmelCase_ = to_numpy_array(lowerCAmelCase ) if do_resize: UpperCAmelCase_ = self.resize(image=lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase ) if do_center_crop: UpperCAmelCase_ = self.center_crop(lowerCAmelCase , size=lowerCAmelCase ) if do_rescale: UpperCAmelCase_ = self.rescale(image=lowerCAmelCase , scale=lowerCAmelCase ) if do_normalize: UpperCAmelCase_ = self.normalize(image=lowerCAmelCase , mean=lowerCAmelCase , std=lowerCAmelCase ) UpperCAmelCase_ = to_channel_dimension_format(lowerCAmelCase , lowerCAmelCase ) return image def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , **lowerCAmelCase , ): UpperCAmelCase_ = do_resize if do_resize is not None else self.do_resize UpperCAmelCase_ = resample if resample is not None else self.resample UpperCAmelCase_ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase_ = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase_ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase_ = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else self.image_mean UpperCAmelCase_ = image_std if image_std is not None else self.image_std UpperCAmelCase_ = size if size is not None else self.size UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) UpperCAmelCase_ = crop_size if crop_size is not None else self.crop_size UpperCAmelCase_ = get_size_dict(lowerCAmelCase , param_name="crop_size" ) if not valid_images(lowerCAmelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) UpperCAmelCase_ = make_batched(lowerCAmelCase ) UpperCAmelCase_ = [ [ self._preprocess_image( image=lowerCAmelCase , do_resize=lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase , do_center_crop=lowerCAmelCase , crop_size=lowerCAmelCase , do_rescale=lowerCAmelCase , rescale_factor=lowerCAmelCase , do_normalize=lowerCAmelCase , image_mean=lowerCAmelCase , image_std=lowerCAmelCase , data_format=lowerCAmelCase , ) for img in video ] for video in videos ] UpperCAmelCase_ = {"pixel_values": videos} return BatchFeature(data=lowerCAmelCase , tensor_type=lowerCAmelCase )
23
1
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase=2 , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase=10 , lowerCAmelCase=3 , lowerCAmelCase=32 * 8 , lowerCAmelCase=32 * 8 , lowerCAmelCase=4 , lowerCAmelCase=64 , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = is_training UpperCAmelCase_ = use_auxiliary_loss UpperCAmelCase_ = num_queries UpperCAmelCase_ = num_channels UpperCAmelCase_ = min_size UpperCAmelCase_ = max_size UpperCAmelCase_ = num_labels UpperCAmelCase_ = hidden_dim UpperCAmelCase_ = hidden_dim def A__ ( self ): UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( lowerCAmelCase ) UpperCAmelCase_ = torch.ones([self.batch_size, self.min_size, self.max_size] , device=lowerCAmelCase ) UpperCAmelCase_ = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=lowerCAmelCase ) > 0.5 ).float() UpperCAmelCase_ = (torch.rand((self.batch_size, self.num_labels) , device=lowerCAmelCase ) > 0.5).long() UpperCAmelCase_ = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def A__ ( self ): UpperCAmelCase_ = MaskaFormerConfig( hidden_size=self.hidden_dim , ) UpperCAmelCase_ = self.num_queries UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = [1, 1, 1, 1] UpperCAmelCase_ = self.num_channels UpperCAmelCase_ = 64 UpperCAmelCase_ = 128 UpperCAmelCase_ = self.hidden_dim UpperCAmelCase_ = self.hidden_dim UpperCAmelCase_ = self.hidden_dim return config def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.prepare_config_and_inputs() UpperCAmelCase_ = {"pixel_values": pixel_values, "pixel_mask": pixel_mask} return config, inputs_dict def A__ ( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = output.encoder_hidden_states UpperCAmelCase_ = output.pixel_decoder_hidden_states UpperCAmelCase_ = output.transformer_decoder_hidden_states self.parent.assertTrue(len(lowerCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowerCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(lowerCAmelCase ) , config.decoder_layers ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False ): with torch.no_grad(): UpperCAmelCase_ = MaskaFormerModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(pixel_values=lowerCAmelCase , pixel_mask=lowerCAmelCase ) UpperCAmelCase_ = model(lowerCAmelCase , output_hidden_states=lowerCAmelCase ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.hidden_dim) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(lowerCAmelCase , lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = MaskaFormerForUniversalSegmentation(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() def comm_check_on_output(lowerCAmelCase ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): UpperCAmelCase_ = model(pixel_values=lowerCAmelCase , pixel_mask=lowerCAmelCase ) UpperCAmelCase_ = model(lowerCAmelCase ) comm_check_on_output(lowerCAmelCase ) UpperCAmelCase_ = model( pixel_values=lowerCAmelCase , pixel_mask=lowerCAmelCase , mask_labels=lowerCAmelCase , class_labels=lowerCAmelCase ) comm_check_on_output(lowerCAmelCase ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () lowerCAmelCase_ : str = {'feature-extraction': MaskaFormerModel} if is_torch_available() else {} lowerCAmelCase_ : Optional[Any] = False lowerCAmelCase_ : Dict = False lowerCAmelCase_ : int = False lowerCAmelCase_ : Optional[int] = False def A__ ( self ): UpperCAmelCase_ = MaskaFormerModelTester(self ) UpperCAmelCase_ = ConfigTester(self , config_class=lowerCAmelCase , has_text_modality=lowerCAmelCase ) def A__ ( self ): self.config_tester.run_common_tests() def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(lowerCAmelCase , **lowerCAmelCase , output_hidden_states=lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*lowerCAmelCase ) @unittest.skip(reason="Mask2Former does not use inputs_embeds" ) def A__ ( self ): pass @unittest.skip(reason="Mask2Former does not have a get_input_embeddings method" ) def A__ ( self ): pass @unittest.skip(reason="Mask2Former is not a generative model" ) def A__ ( self ): pass @unittest.skip(reason="Mask2Former does not use token embeddings" ) def A__ ( self ): pass @require_torch_multi_gpu @unittest.skip( reason="Mask2Former has some layers using `add_module` which doesn't work well with `nn.DataParallel`" ) def A__ ( self ): pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def A__ ( self ): pass def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase ) @slow def A__ ( self ): for model_name in ["facebook/mask2former-swin-small-coco-instance"]: UpperCAmelCase_ = MaskaFormerModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = (self.model_tester.min_size,) * 2 UpperCAmelCase_ = { "pixel_values": torch.randn((2, 3, *size) , device=lowerCAmelCase ), "mask_labels": torch.randn((2, 10, *size) , device=lowerCAmelCase ), "class_labels": torch.zeros(2 , 10 , device=lowerCAmelCase ).long(), } UpperCAmelCase_ = self.model_tester.get_config() UpperCAmelCase_ = MaskaFormerForUniversalSegmentation(lowerCAmelCase ).to(lowerCAmelCase ) UpperCAmelCase_ = model(**lowerCAmelCase ) self.assertTrue(outputs.loss is not None ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(lowerCAmelCase , **lowerCAmelCase , output_hidden_states=lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ).to(lowerCAmelCase ) UpperCAmelCase_ = model(**lowerCAmelCase , output_attentions=lowerCAmelCase ) self.assertTrue(outputs.attentions is not None ) def A__ ( self ): if not self.model_tester.is_training: return UpperCAmelCase_ = self.all_model_classes[1] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() UpperCAmelCase_ = model_class(lowerCAmelCase ) model.to(lowerCAmelCase ) model.train() UpperCAmelCase_ = model(lowerCAmelCase , mask_labels=lowerCAmelCase , class_labels=lowerCAmelCase ).loss loss.backward() def A__ ( self ): UpperCAmelCase_ = self.all_model_classes[1] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() UpperCAmelCase_ = True UpperCAmelCase_ = True UpperCAmelCase_ = model_class(lowerCAmelCase ).to(lowerCAmelCase ) model.train() UpperCAmelCase_ = model(lowerCAmelCase , mask_labels=lowerCAmelCase , class_labels=lowerCAmelCase ) UpperCAmelCase_ = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() UpperCAmelCase_ = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() UpperCAmelCase_ = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() UpperCAmelCase_ = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=lowerCAmelCase ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) SCREAMING_SNAKE_CASE = 1e-4 def snake_case__ ( ) -> Union[str, Any]: UpperCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @slow class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def A__ ( self ): return "facebook/mask2former-swin-small-coco-instance" @cached_property def A__ ( self ): return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def A__ ( self ): UpperCAmelCase_ = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(lowerCAmelCase ) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(lowerCAmelCase , return_tensors="pt" ).to(lowerCAmelCase ) UpperCAmelCase_ = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(lowerCAmelCase , (1, 3, 384, 384) ) with torch.no_grad(): UpperCAmelCase_ = model(**lowerCAmelCase ) UpperCAmelCase_ = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(lowerCAmelCase ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , lowerCAmelCase , atol=lowerCAmelCase ) ) UpperCAmelCase_ = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(lowerCAmelCase ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , lowerCAmelCase , atol=lowerCAmelCase ) ) UpperCAmelCase_ = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(lowerCAmelCase ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , lowerCAmelCase , atol=lowerCAmelCase ) ) def A__ ( self ): UpperCAmelCase_ = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(lowerCAmelCase ).eval() UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(lowerCAmelCase , return_tensors="pt" ).to(lowerCAmelCase ) UpperCAmelCase_ = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(lowerCAmelCase , (1, 3, 384, 384) ) with torch.no_grad(): UpperCAmelCase_ = model(**lowerCAmelCase ) # masks_queries_logits UpperCAmelCase_ = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) UpperCAmelCase_ = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] UpperCAmelCase_ = torch.tensor(lowerCAmelCase ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , lowerCAmelCase , atol=lowerCAmelCase ) ) # class_queries_logits UpperCAmelCase_ = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape , (1, model.config.num_queries, model.config.num_labels + 1) ) UpperCAmelCase_ = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , lowerCAmelCase , atol=lowerCAmelCase ) ) def A__ ( self ): UpperCAmelCase_ = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(lowerCAmelCase ).eval() UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="pt" , ) UpperCAmelCase_ = inputs["pixel_values"].to(lowerCAmelCase ) UpperCAmelCase_ = [el.to(lowerCAmelCase ) for el in inputs["mask_labels"]] UpperCAmelCase_ = [el.to(lowerCAmelCase ) for el in inputs["class_labels"]] with torch.no_grad(): UpperCAmelCase_ = model(**lowerCAmelCase ) self.assertTrue(outputs.loss is not None )
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 1 for i in range(1 , num + 1 ): fact *= i return fact def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 0 while number > 0: UpperCAmelCase_ = number % 10 sum_of_digits += last_digit UpperCAmelCase_ = number // 10 # Removing the last_digit from the given number return sum_of_digits def snake_case__ ( __SCREAMING_SNAKE_CASE = 100 ) -> int: UpperCAmelCase_ = factorial(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = split_and_add(__SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
23
1
import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder SCREAMING_SNAKE_CASE = "__DUMMY_TRANSFORMERS_USER__" SCREAMING_SNAKE_CASE = "Dummy User" SCREAMING_SNAKE_CASE = "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" SCREAMING_SNAKE_CASE = "https://hub-ci.huggingface.co" SCREAMING_SNAKE_CASE = CI_HUB_ENDPOINT + "/datasets/{repo_id}/resolve/{revision}/{path}" SCREAMING_SNAKE_CASE = CI_HUB_ENDPOINT + "/{repo_id}/resolve/{revision}/{filename}" SCREAMING_SNAKE_CASE = Path("~/.huggingface/hub_ci_token").expanduser() @pytest.fixture def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Dict: monkeypatch.setattr( "huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE" , __SCREAMING_SNAKE_CASE ) @pytest.fixture def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Dict: monkeypatch.setattr("datasets.config.HF_ENDPOINT" , __SCREAMING_SNAKE_CASE ) monkeypatch.setattr("datasets.config.HUB_DATASETS_URL" , __SCREAMING_SNAKE_CASE ) @pytest.fixture def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Any: monkeypatch.setattr("huggingface_hub.hf_api.HfFolder.path_token" , __SCREAMING_SNAKE_CASE ) @pytest.fixture def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> int: HfFolder.save_token(__SCREAMING_SNAKE_CASE ) yield HfFolder.delete_token() @pytest.fixture(scope="session" ) def snake_case__ ( ) -> Union[str, Any]: return HfApi(endpoint=__SCREAMING_SNAKE_CASE ) @pytest.fixture(scope="session" ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ = HfFolder.get_token() HfFolder.save_token(__SCREAMING_SNAKE_CASE ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(__SCREAMING_SNAKE_CASE ) @pytest.fixture def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Tuple: def _cleanup_repo(__SCREAMING_SNAKE_CASE ): hf_api.delete_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type="dataset" ) return _cleanup_repo @pytest.fixture def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Tuple: @contextmanager def _temporary_repo(__SCREAMING_SNAKE_CASE ): try: yield repo_id finally: cleanup_repo(__SCREAMING_SNAKE_CASE ) return _temporary_repo @pytest.fixture(scope="session" ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Any: UpperCAmelCase_ = f'''repo_txt_data-{int(time.time() * 10E3 )}''' UpperCAmelCase_ = f'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type="dataset" , private=__SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=__SCREAMING_SNAKE_CASE , path_or_fileobj=str(__SCREAMING_SNAKE_CASE ) , path_in_repo="data/text_data.txt" , repo_id=__SCREAMING_SNAKE_CASE , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="session" ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Any: UpperCAmelCase_ = f'''repo_zipped_txt_data-{int(time.time() * 10E3 )}''' UpperCAmelCase_ = f'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type="dataset" , private=__SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=__SCREAMING_SNAKE_CASE , path_or_fileobj=str(__SCREAMING_SNAKE_CASE ) , path_in_repo="data.zip" , repo_id=__SCREAMING_SNAKE_CASE , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Dict: return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="session" ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = f'''repo_zipped_img_data-{int(time.time() * 10E3 )}''' UpperCAmelCase_ = f'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type="dataset" , private=__SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=__SCREAMING_SNAKE_CASE , path_or_fileobj=str(__SCREAMING_SNAKE_CASE ) , path_in_repo="data.zip" , repo_id=__SCREAMING_SNAKE_CASE , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[Any]: return hf_private_dataset_repo_zipped_img_data_
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_xlnet": ["XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["XLNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["XLNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "XLNetForMultipleChoice", "XLNetForQuestionAnswering", "XLNetForQuestionAnsweringSimple", "XLNetForSequenceClassification", "XLNetForTokenClassification", "XLNetLMHeadModel", "XLNetModel", "XLNetPreTrainedModel", "load_tf_weights_in_xlnet", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLNetForMultipleChoice", "TFXLNetForQuestionAnsweringSimple", "TFXLNetForSequenceClassification", "TFXLNetForTokenClassification", "TFXLNetLMHeadModel", "TFXLNetMainLayer", "TFXLNetModel", "TFXLNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
23
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[Any]: # Initialise PyTorch model UpperCAmelCase_ = MobileBertConfig.from_json_file(__SCREAMING_SNAKE_CASE ) print(f'''Building PyTorch model from configuration: {config}''' ) UpperCAmelCase_ = MobileBertForPreTraining(__SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint UpperCAmelCase_ = load_tf_weights_in_mobilebert(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , __SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--mobilebert_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained MobileBERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
23
1
from abc import ABC, abstractmethod from argparse import ArgumentParser class lowerCamelCase ( lowercase__ ): '''simple docstring''' @staticmethod @abstractmethod def A__ ( lowerCAmelCase ): raise NotImplementedError() @abstractmethod def A__ ( self ): raise NotImplementedError()
23
import heapq as hq import math from collections.abc import Iterator class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase ): UpperCAmelCase_ = str(id_ ) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = [] UpperCAmelCase_ = {} # {vertex:distance} def __lt__( self , lowerCAmelCase ): return self.key < other.key def __repr__( self ): return self.id def A__ ( self , lowerCAmelCase ): self.neighbors.append(lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = weight def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __SCREAMING_SNAKE_CASE ) graph[b - 1].add_edge(graph[a - 1] , __SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> list: UpperCAmelCase_ = [] for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = graph[:] while q: UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE ) q.remove(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Iterator[tuple]: for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = list(__SCREAMING_SNAKE_CASE ) hq.heapify(__SCREAMING_SNAKE_CASE ) while h: UpperCAmelCase_ = hq.heappop(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] hq.heapify(__SCREAMING_SNAKE_CASE ) for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def snake_case__ ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
23
1
def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: UpperCAmelCase_ = [1] for i in range(2 , __SCREAMING_SNAKE_CASE ): factorials.append(factorials[-1] * i ) assert 0 <= k < factorials[-1] * n, "k out of bounds" UpperCAmelCase_ = [] UpperCAmelCase_ = list(range(__SCREAMING_SNAKE_CASE ) ) # Find permutation while factorials: UpperCAmelCase_ = factorials.pop() UpperCAmelCase_ , UpperCAmelCase_ = divmod(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) permutation.append(elements[number] ) elements.remove(elements[number] ) permutation.append(elements[0] ) return permutation if __name__ == "__main__": import doctest doctest.testmod()
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_fnet": ["FNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "FNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["FNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["FNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "FNET_PRETRAINED_MODEL_ARCHIVE_LIST", "FNetForMaskedLM", "FNetForMultipleChoice", "FNetForNextSentencePrediction", "FNetForPreTraining", "FNetForQuestionAnswering", "FNetForSequenceClassification", "FNetForTokenClassification", "FNetLayer", "FNetModel", "FNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Any: # initialize config if "resnet-50" in model_name: UpperCAmelCase_ = ResNetConfig.from_pretrained("microsoft/resnet-50" ) elif "resnet-101" in model_name: UpperCAmelCase_ = ResNetConfig.from_pretrained("microsoft/resnet-101" ) else: raise ValueError("Model name should include either resnet50 or resnet101" ) UpperCAmelCase_ = DetrConfig(use_timm_backbone=__SCREAMING_SNAKE_CASE , backbone_config=__SCREAMING_SNAKE_CASE ) # set label attributes UpperCAmelCase_ = "panoptic" in model_name if is_panoptic: UpperCAmelCase_ = 250 else: UpperCAmelCase_ = 91 UpperCAmelCase_ = "huggingface/label-files" UpperCAmelCase_ = "coco-detection-id2label.json" UpperCAmelCase_ = json.load(open(hf_hub_download(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , repo_type="dataset" ) , "r" ) ) UpperCAmelCase_ = {int(__SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} UpperCAmelCase_ = idalabel UpperCAmelCase_ = {v: k for k, v in idalabel.items()} return config, is_panoptic def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Tuple: # here we list all keys to be renamed (original name on the left, our name on the right) UpperCAmelCase_ = [] # stem # fmt: off rename_keys.append(("backbone.0.body.conv1.weight", "backbone.conv_encoder.model.embedder.embedder.convolution.weight") ) rename_keys.append(("backbone.0.body.bn1.weight", "backbone.conv_encoder.model.embedder.embedder.normalization.weight") ) rename_keys.append(("backbone.0.body.bn1.bias", "backbone.conv_encoder.model.embedder.embedder.normalization.bias") ) rename_keys.append(("backbone.0.body.bn1.running_mean", "backbone.conv_encoder.model.embedder.embedder.normalization.running_mean") ) rename_keys.append(("backbone.0.body.bn1.running_var", "backbone.conv_encoder.model.embedder.embedder.normalization.running_var") ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( f'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight''', f'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight''', ) ) rename_keys.append( ( f'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight''', f'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight''', ) ) rename_keys.append( ( f'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias''', f'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias''', ) ) rename_keys.append( ( f'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean''', f'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean''', ) ) rename_keys.append( ( f'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var''', f'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var''', ) ) # 3 convs for i in range(3 ): rename_keys.append( ( f'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight''', f'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight''', ) ) rename_keys.append( ( f'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight''', f'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight''', ) ) rename_keys.append( ( f'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias''', f'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias''', ) ) rename_keys.append( ( f'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean''', f'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean''', ) ) rename_keys.append( ( f'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var''', f'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var''', ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( f'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', f'''encoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (f'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', f'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.weight''', f'''encoder.layers.{i}.fc1.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear1.bias''', f'''encoder.layers.{i}.fc1.bias''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.weight''', f'''encoder.layers.{i}.fc2.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.linear2.bias''', f'''encoder.layers.{i}.fc2.bias''') ) rename_keys.append( (f'''transformer.encoder.layers.{i}.norm1.weight''', f'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (f'''transformer.encoder.layers.{i}.norm1.bias''', f'''encoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (f'''transformer.encoder.layers.{i}.norm2.weight''', f'''encoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((f'''transformer.encoder.layers.{i}.norm2.bias''', f'''encoder.layers.{i}.final_layer_norm.bias''') ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( f'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', f'''decoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (f'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', f'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', f'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( f'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', f'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.weight''', f'''decoder.layers.{i}.fc1.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear1.bias''', f'''decoder.layers.{i}.fc1.bias''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.weight''', f'''decoder.layers.{i}.fc2.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.linear2.bias''', f'''decoder.layers.{i}.fc2.bias''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm1.weight''', f'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm1.bias''', f'''decoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.weight''', f'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm2.bias''', f'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append( (f'''transformer.decoder.layers.{i}.norm3.weight''', f'''decoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((f'''transformer.decoder.layers.{i}.norm3.bias''', f'''decoder.layers.{i}.final_layer_norm.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ] ) return rename_keys def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ = state_dict.pop(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = val def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=False ) -> Optional[Any]: UpperCAmelCase_ = "" if is_panoptic: UpperCAmelCase_ = "detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) UpperCAmelCase_ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) UpperCAmelCase_ = state_dict.pop(f'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ = in_proj_weight[:256, :] UpperCAmelCase_ = in_proj_bias[:256] UpperCAmelCase_ = in_proj_weight[256:512, :] UpperCAmelCase_ = in_proj_bias[256:512] UpperCAmelCase_ = in_proj_weight[-256:, :] UpperCAmelCase_ = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention UpperCAmelCase_ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) UpperCAmelCase_ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ = in_proj_weight[:256, :] UpperCAmelCase_ = in_proj_bias[:256] UpperCAmelCase_ = in_proj_weight[256:512, :] UpperCAmelCase_ = in_proj_bias[256:512] UpperCAmelCase_ = in_proj_weight[-256:, :] UpperCAmelCase_ = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention UpperCAmelCase_ = state_dict.pop( f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) UpperCAmelCase_ = state_dict.pop(f'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict UpperCAmelCase_ = in_proj_weight_cross_attn[:256, :] UpperCAmelCase_ = in_proj_bias_cross_attn[:256] UpperCAmelCase_ = in_proj_weight_cross_attn[256:512, :] UpperCAmelCase_ = in_proj_bias_cross_attn[256:512] UpperCAmelCase_ = in_proj_weight_cross_attn[-256:, :] UpperCAmelCase_ = in_proj_bias_cross_attn[-256:] def snake_case__ ( ) -> str: UpperCAmelCase_ = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ = Image.open(requests.get(__SCREAMING_SNAKE_CASE , stream=__SCREAMING_SNAKE_CASE ).raw ) return im @torch.no_grad() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=False ) -> str: UpperCAmelCase_ , UpperCAmelCase_ = get_detr_config(__SCREAMING_SNAKE_CASE ) # load original model from torch hub UpperCAmelCase_ = { "detr-resnet-50": "detr_resnet50", "detr-resnet-101": "detr_resnet101", } logger.info(f'''Converting model {model_name}...''' ) UpperCAmelCase_ = torch.hub.load("facebookresearch/detr" , model_name_to_original_name[model_name] , pretrained=__SCREAMING_SNAKE_CASE ).eval() UpperCAmelCase_ = detr.state_dict() # rename keys for src, dest in create_rename_keys(__SCREAMING_SNAKE_CASE ): if is_panoptic: UpperCAmelCase_ = "detr." + src rename_key(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # query, key and value matrices need special treatment read_in_q_k_v(__SCREAMING_SNAKE_CASE , is_panoptic=__SCREAMING_SNAKE_CASE ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them UpperCAmelCase_ = "detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): UpperCAmelCase_ = state_dict.pop(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = val elif "class_labels_classifier" in key or "bbox_predictor" in key: UpperCAmelCase_ = state_dict.pop(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: UpperCAmelCase_ = state_dict.pop(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): UpperCAmelCase_ = state_dict.pop(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = val # finally, create HuggingFace model and load state dict UpperCAmelCase_ = DetrForSegmentation(__SCREAMING_SNAKE_CASE ) if is_panoptic else DetrForObjectDetection(__SCREAMING_SNAKE_CASE ) model.load_state_dict(__SCREAMING_SNAKE_CASE ) model.eval() # verify our conversion on an image UpperCAmelCase_ = "coco_panoptic" if is_panoptic else "coco_detection" UpperCAmelCase_ = DetrImageProcessor(format=__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = processor(images=prepare_img() , return_tensors="pt" ) UpperCAmelCase_ = encoding["pixel_values"] UpperCAmelCase_ = detr(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = model(__SCREAMING_SNAKE_CASE ) assert torch.allclose(outputs.logits , original_outputs["pred_logits"] , atol=1E-3 ) assert torch.allclose(outputs.pred_boxes , original_outputs["pred_boxes"] , atol=1E-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["pred_masks"] , atol=1E-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__SCREAMING_SNAKE_CASE ).mkdir(exist_ok=__SCREAMING_SNAKE_CASE ) model.save_pretrained(__SCREAMING_SNAKE_CASE ) processor.save_pretrained(__SCREAMING_SNAKE_CASE ) if push_to_hub: # Upload model and image processor to the hub logger.info("Uploading PyTorch model and image processor to the hub..." ) model.push_to_hub(f'''nielsr/{model_name}''' ) processor.push_to_hub(f'''nielsr/{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument( "--model_name", default="detr-resnet-50", type=str, choices=["detr-resnet-50", "detr-resnet-101"], help="Name of the DETR model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument("--push_to_hub", action="store_true", help="Whether to push the model to the hub or not.") SCREAMING_SNAKE_CASE = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
23
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": 512, } SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": {"do_lower_case": True}, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = VOCAB_FILES_NAMES lowerCAmelCase_ : int = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Dict = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : List[str] = RetriBertTokenizer lowerCAmelCase_ : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase="[UNK]" , lowerCAmelCase="[SEP]" , lowerCAmelCase="[PAD]" , lowerCAmelCase="[CLS]" , lowerCAmelCase="[MASK]" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , do_lower_case=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , pad_token=lowerCAmelCase , cls_token=lowerCAmelCase , mask_token=lowerCAmelCase , tokenize_chinese_chars=lowerCAmelCase , strip_accents=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(lowerCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**lowerCAmelCase ) UpperCAmelCase_ = do_lower_case def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
1
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> bool: UpperCAmelCase_ = 0 for ch in input_str: UpperCAmelCase_ = ord(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = pow(2 , __SCREAMING_SNAKE_CASE ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
23
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-german-cased": ( "https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json" ), "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE = { "distilbert-base-uncased": 512, "distilbert-base-uncased-distilled-squad": 512, "distilbert-base-cased": 512, "distilbert-base-cased-distilled-squad": 512, "distilbert-base-german-cased": 512, "distilbert-base-multilingual-cased": 512, } SCREAMING_SNAKE_CASE = { "distilbert-base-uncased": {"do_lower_case": True}, "distilbert-base-uncased-distilled-squad": {"do_lower_case": True}, "distilbert-base-cased": {"do_lower_case": False}, "distilbert-base-cased-distilled-squad": {"do_lower_case": False}, "distilbert-base-german-cased": {"do_lower_case": False}, "distilbert-base-multilingual-cased": {"do_lower_case": False}, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Any = VOCAB_FILES_NAMES lowerCAmelCase_ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Union[str, Any] = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : int = ['input_ids', 'attention_mask'] lowerCAmelCase_ : str = DistilBertTokenizer def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase="[UNK]" , lowerCAmelCase="[SEP]" , lowerCAmelCase="[PAD]" , lowerCAmelCase="[CLS]" , lowerCAmelCase="[MASK]" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , do_lower_case=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , pad_token=lowerCAmelCase , cls_token=lowerCAmelCase , mask_token=lowerCAmelCase , tokenize_chinese_chars=lowerCAmelCase , strip_accents=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(lowerCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**lowerCAmelCase ) UpperCAmelCase_ = do_lower_case def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
1
from __future__ import annotations import typing from collections.abc import Iterable import numpy as np SCREAMING_SNAKE_CASE = typing.Union[Iterable[float], Iterable[int], np.ndarray] # noqa: UP007 SCREAMING_SNAKE_CASE = typing.Union[np.floataa, int, float] # noqa: UP007 def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> VectorOut: return np.sqrt(np.sum((np.asarray(__SCREAMING_SNAKE_CASE ) - np.asarray(__SCREAMING_SNAKE_CASE )) ** 2 ) ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> VectorOut: return sum((va - va) ** 2 for va, va in zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) ** (1 / 2) if __name__ == "__main__": def snake_case__ ( ) -> None: from timeit import timeit print("Without Numpy" ) print( timeit( "euclidean_distance_no_np([1, 2, 3], [4, 5, 6])" , number=1_0000 , globals=globals() , ) ) print("With Numpy" ) print( timeit( "euclidean_distance([1, 2, 3], [4, 5, 6])" , number=1_0000 , globals=globals() , ) ) benchmark()
23
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> np.array: UpperCAmelCase_ = f'''{sampling_rate}''' UpperCAmelCase_ = "1" UpperCAmelCase_ = "f32le" UpperCAmelCase_ = [ "ffmpeg", "-i", "pipe:0", "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-hide_banner", "-loglevel", "quiet", "pipe:1", ] try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: UpperCAmelCase_ = ffmpeg_process.communicate(__SCREAMING_SNAKE_CASE ) except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to load audio files from filename" ) from error UpperCAmelCase_ = output_stream[0] UpperCAmelCase_ = np.frombuffer(__SCREAMING_SNAKE_CASE , np.floataa ) if audio.shape[0] == 0: raise ValueError("Malformed soundfile" ) return audio def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = "f32le" , ) -> Dict: UpperCAmelCase_ = f'''{sampling_rate}''' UpperCAmelCase_ = "1" if format_for_conversion == "s16le": UpperCAmelCase_ = 2 elif format_for_conversion == "f32le": UpperCAmelCase_ = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) UpperCAmelCase_ = platform.system() if system == "Linux": UpperCAmelCase_ = "alsa" UpperCAmelCase_ = "default" elif system == "Darwin": UpperCAmelCase_ = "avfoundation" UpperCAmelCase_ = ":0" elif system == "Windows": UpperCAmelCase_ = "dshow" UpperCAmelCase_ = "default" UpperCAmelCase_ = [ "ffmpeg", "-f", format_, "-i", input_, "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-fflags", "nobuffer", "-hide_banner", "-loglevel", "quiet", "pipe:1", ] UpperCAmelCase_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample UpperCAmelCase_ = _ffmpeg_stream(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for item in iterator: yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = "f32le" , ) -> int: if stream_chunk_s is not None: UpperCAmelCase_ = stream_chunk_s else: UpperCAmelCase_ = chunk_length_s UpperCAmelCase_ = ffmpeg_microphone(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , format_for_conversion=__SCREAMING_SNAKE_CASE ) if format_for_conversion == "s16le": UpperCAmelCase_ = np.intaa UpperCAmelCase_ = 2 elif format_for_conversion == "f32le": UpperCAmelCase_ = np.floataa UpperCAmelCase_ = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: UpperCAmelCase_ = chunk_length_s / 6 UpperCAmelCase_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(__SCREAMING_SNAKE_CASE , (int, float) ): UpperCAmelCase_ = [stride_length_s, stride_length_s] UpperCAmelCase_ = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample UpperCAmelCase_ = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample UpperCAmelCase_ = datetime.datetime.now() UpperCAmelCase_ = datetime.timedelta(seconds=__SCREAMING_SNAKE_CASE ) for item in chunk_bytes_iter(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=(stride_left, stride_right) , stream=__SCREAMING_SNAKE_CASE ): # Put everything back in numpy scale UpperCAmelCase_ = np.frombuffer(item["raw"] , dtype=__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = ( item["stride"][0] // size_of_sample, item["stride"][1] // size_of_sample, ) UpperCAmelCase_ = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = False ) -> Dict: UpperCAmelCase_ = B"" UpperCAmelCase_ , UpperCAmelCase_ = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) UpperCAmelCase_ = 0 for raw in iterator: acc += raw if stream and len(__SCREAMING_SNAKE_CASE ) < chunk_len: UpperCAmelCase_ = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(__SCREAMING_SNAKE_CASE ) >= chunk_len: # We are flushing the accumulator UpperCAmelCase_ = (_stride_left, stride_right) UpperCAmelCase_ = {"raw": acc[:chunk_len], "stride": stride} if stream: UpperCAmelCase_ = False yield item UpperCAmelCase_ = stride_left UpperCAmelCase_ = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(__SCREAMING_SNAKE_CASE ) > stride_left: UpperCAmelCase_ = {"raw": acc, "stride": (_stride_left, 0)} if stream: UpperCAmelCase_ = False yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ = 2**24 # 16Mo try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdout=subprocess.PIPE , bufsize=__SCREAMING_SNAKE_CASE ) as ffmpeg_process: while True: UpperCAmelCase_ = ffmpeg_process.stdout.read(__SCREAMING_SNAKE_CASE ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to stream audio files from filename" ) from error
23
1
import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_led import LEDTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json", }, "merges_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt", }, "tokenizer_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json", }, } SCREAMING_SNAKE_CASE = { "allenai/led-base-16384": 16384, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = VOCAB_FILES_NAMES lowerCAmelCase_ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : str = LEDTokenizer lowerCAmelCase_ : List[Any] = ['input_ids', 'attention_mask'] def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase="replace" , lowerCAmelCase="<s>" , lowerCAmelCase="</s>" , lowerCAmelCase="</s>" , lowerCAmelCase="<s>" , lowerCAmelCase="<unk>" , lowerCAmelCase="<pad>" , lowerCAmelCase="<mask>" , lowerCAmelCase=False , lowerCAmelCase=True , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , lowerCAmelCase , tokenizer_file=lowerCAmelCase , errors=lowerCAmelCase , bos_token=lowerCAmelCase , eos_token=lowerCAmelCase , sep_token=lowerCAmelCase , cls_token=lowerCAmelCase , unk_token=lowerCAmelCase , pad_token=lowerCAmelCase , mask_token=lowerCAmelCase , add_prefix_space=lowerCAmelCase , trim_offsets=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowerCAmelCase ) != add_prefix_space: UpperCAmelCase_ = getattr(lowerCAmelCase , pre_tok_state.pop("type" ) ) UpperCAmelCase_ = add_prefix_space UpperCAmelCase_ = pre_tok_class(**lowerCAmelCase ) UpperCAmelCase_ = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` UpperCAmelCase_ = "post_processor" UpperCAmelCase_ = getattr(self.backend_tokenizer , lowerCAmelCase , lowerCAmelCase ) if tokenizer_component_instance: UpperCAmelCase_ = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: UpperCAmelCase_ = tuple(state["sep"] ) if "cls" in state: UpperCAmelCase_ = tuple(state["cls"] ) UpperCAmelCase_ = False if state.get("add_prefix_space" , lowerCAmelCase ) != add_prefix_space: UpperCAmelCase_ = add_prefix_space UpperCAmelCase_ = True if state.get("trim_offsets" , lowerCAmelCase ) != trim_offsets: UpperCAmelCase_ = trim_offsets UpperCAmelCase_ = True if changes_to_apply: UpperCAmelCase_ = getattr(lowerCAmelCase , state.pop("type" ) ) UpperCAmelCase_ = component_class(**lowerCAmelCase ) setattr(self.backend_tokenizer , lowerCAmelCase , lowerCAmelCase ) @property # Copied from transformers.models.bart.tokenization_bart_fast.BartTokenizerFast.mask_token with BART->LED def A__ ( self ): if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else value UpperCAmelCase_ = value def A__ ( self , *lowerCAmelCase , **lowerCAmelCase ): UpperCAmelCase_ = kwargs.get("is_split_into_words" , lowerCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , *lowerCAmelCase , **lowerCAmelCase ): UpperCAmelCase_ = kwargs.get("is_split_into_words" , lowerCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = PaddingStrategy.DO_NOT_PAD , lowerCAmelCase = None , lowerCAmelCase = None , ): UpperCAmelCase_ = super()._pad( encoded_inputs=lowerCAmelCase , max_length=lowerCAmelCase , padding_strategy=lowerCAmelCase , pad_to_multiple_of=lowerCAmelCase , return_attention_mask=lowerCAmelCase , ) # Load from model defaults if return_attention_mask is None: UpperCAmelCase_ = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: UpperCAmelCase_ = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. UpperCAmelCase_ = len(encoded_inputs["global_attention_mask"] ) != len(lowerCAmelCase ) if needs_to_be_padded: UpperCAmelCase_ = len(lowerCAmelCase ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` UpperCAmelCase_ = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": UpperCAmelCase_ = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
23
from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class lowerCamelCase ( lowercase__, lowercase__ ): '''simple docstring''' @register_to_config def __init__( self , lowerCAmelCase = 768 , ): super().__init__() UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(torch.ones(1 , lowerCAmelCase ) ) def A__ ( self , lowerCAmelCase = None , lowerCAmelCase = None , ): UpperCAmelCase_ = nn.Parameter(self.mean.to(lowerCAmelCase ).to(lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(self.std.to(lowerCAmelCase ).to(lowerCAmelCase ) ) return self def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (embeds - self.mean) * 1.0 / self.std return embeds def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (embeds * self.std) + self.mean return embeds
23
1
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def A__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = FlaxStableDiffusionPipeline.from_pretrained( "stabilityai/stable-diffusion-2" , revision="bf16" , dtype=jnp.bfloataa , ) UpperCAmelCase_ = "A painting of a squirrel eating a burger" UpperCAmelCase_ = jax.device_count() UpperCAmelCase_ = num_samples * [prompt] UpperCAmelCase_ = sd_pipe.prepare_inputs(lowerCAmelCase ) UpperCAmelCase_ = replicate(lowerCAmelCase ) UpperCAmelCase_ = shard(lowerCAmelCase ) UpperCAmelCase_ = jax.random.PRNGKey(0 ) UpperCAmelCase_ = jax.random.split(lowerCAmelCase , jax.device_count() ) UpperCAmelCase_ = sd_pipe(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , num_inference_steps=25 , jit=lowerCAmelCase )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) UpperCAmelCase_ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) UpperCAmelCase_ = images[0, 253:256, 253:256, -1] UpperCAmelCase_ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) UpperCAmelCase_ = jnp.array([0.4238, 0.4414, 0.4395, 0.4453, 0.4629, 0.4590, 0.4531, 0.45508, 0.4512] ) print(f'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def A__ ( self ): UpperCAmelCase_ = "stabilityai/stable-diffusion-2" UpperCAmelCase_ , UpperCAmelCase_ = FlaxDPMSolverMultistepScheduler.from_pretrained(lowerCAmelCase , subfolder="scheduler" ) UpperCAmelCase_ , UpperCAmelCase_ = FlaxStableDiffusionPipeline.from_pretrained( lowerCAmelCase , scheduler=lowerCAmelCase , revision="bf16" , dtype=jnp.bfloataa , ) UpperCAmelCase_ = scheduler_params UpperCAmelCase_ = "A painting of a squirrel eating a burger" UpperCAmelCase_ = jax.device_count() UpperCAmelCase_ = num_samples * [prompt] UpperCAmelCase_ = sd_pipe.prepare_inputs(lowerCAmelCase ) UpperCAmelCase_ = replicate(lowerCAmelCase ) UpperCAmelCase_ = shard(lowerCAmelCase ) UpperCAmelCase_ = jax.random.PRNGKey(0 ) UpperCAmelCase_ = jax.random.split(lowerCAmelCase , jax.device_count() ) UpperCAmelCase_ = sd_pipe(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , num_inference_steps=25 , jit=lowerCAmelCase )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) UpperCAmelCase_ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) UpperCAmelCase_ = images[0, 253:256, 253:256, -1] UpperCAmelCase_ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) UpperCAmelCase_ = jnp.array([0.4336, 0.42969, 0.4453, 0.4199, 0.4297, 0.4531, 0.4434, 0.4434, 0.4297] ) print(f'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
23
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_VISION_2_SEQ_MAPPING if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_VISION_2_SEQ_MAPPING SCREAMING_SNAKE_CASE = logging.get_logger(__name__) @add_end_docstrings(lowercase__ ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' def __init__( self , *lowerCAmelCase , **lowerCAmelCase ): super().__init__(*lowerCAmelCase , **lowerCAmelCase ) requires_backends(self , "vision" ) self.check_model_type( TF_MODEL_FOR_VISION_2_SEQ_MAPPING if self.framework == "tf" else MODEL_FOR_VISION_2_SEQ_MAPPING ) def A__ ( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None ): UpperCAmelCase_ = {} UpperCAmelCase_ = {} if prompt is not None: UpperCAmelCase_ = prompt if generate_kwargs is not None: UpperCAmelCase_ = generate_kwargs if max_new_tokens is not None: if "generate_kwargs" not in forward_kwargs: UpperCAmelCase_ = {} if "max_new_tokens" in forward_kwargs["generate_kwargs"]: raise ValueError( "'max_new_tokens' is defined twice, once in 'generate_kwargs' and once as a direct parameter," " please use only one" ) UpperCAmelCase_ = max_new_tokens return preprocess_params, forward_kwargs, {} def __call__( self , lowerCAmelCase , **lowerCAmelCase ): return super().__call__(lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = load_image(lowerCAmelCase ) if prompt is not None: if not isinstance(lowerCAmelCase , lowerCAmelCase ): raise ValueError( f'''Received an invalid text input, got - {type(lowerCAmelCase )} - but expected a single string. ''' "Note also that one single text can be provided for conditional image to text generation." ) UpperCAmelCase_ = self.model.config.model_type if model_type == "git": UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = self.tokenizer(text=lowerCAmelCase , add_special_tokens=lowerCAmelCase ).input_ids UpperCAmelCase_ = [self.tokenizer.cls_token_id] + input_ids UpperCAmelCase_ = torch.tensor(lowerCAmelCase ).unsqueeze(0 ) model_inputs.update({"input_ids": input_ids} ) elif model_type == "pix2struct": UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , header_text=lowerCAmelCase , return_tensors=self.framework ) elif model_type != "vision-encoder-decoder": # vision-encoder-decoder does not support conditional generation UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) UpperCAmelCase_ = self.tokenizer(lowerCAmelCase , return_tensors=self.framework ) model_inputs.update(lowerCAmelCase ) else: raise ValueError(f'''Model type {model_type} does not support conditional text generation''' ) else: UpperCAmelCase_ = self.image_processor(images=lowerCAmelCase , return_tensors=self.framework ) if self.model.config.model_type == "git" and prompt is None: UpperCAmelCase_ = None return model_inputs def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): # Git model sets `model_inputs["input_ids"] = None` in `preprocess` (when `prompt=None`). In batch model, the # pipeline will group them into a list of `None`, which fail `_forward`. Avoid this by checking it first. if ( "input_ids" in model_inputs and isinstance(model_inputs["input_ids"] , lowerCAmelCase ) and all(x is None for x in model_inputs["input_ids"] ) ): UpperCAmelCase_ = None if generate_kwargs is None: UpperCAmelCase_ = {} # FIXME: We need to pop here due to a difference in how `generation.py` and `generation.tf_utils.py` # parse inputs. In the Tensorflow version, `generate` raises an error if we don't use `input_ids` whereas # the PyTorch version matches it with `self.model.main_input_name` or `self.model.encoder.main_input_name` # in the `_prepare_model_inputs` method. UpperCAmelCase_ = model_inputs.pop(self.model.main_input_name ) UpperCAmelCase_ = self.model.generate(lowerCAmelCase , **lowerCAmelCase , **lowerCAmelCase ) return model_outputs def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = [] for output_ids in model_outputs: UpperCAmelCase_ = { "generated_text": self.tokenizer.decode( lowerCAmelCase , skip_special_tokens=lowerCAmelCase , ) } records.append(lowerCAmelCase ) return records
23
1
import argparse import torch from transformers import GPTaLMHeadModel, RobertaForMaskedLM if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser( description=( "Extraction some layers of the full RobertaForMaskedLM or GPT2LMHeadModel for Transfer Learned" " Distillation" ) ) parser.add_argument("--model_type", default="roberta", choices=["roberta", "gpt2"]) parser.add_argument("--model_name", default="roberta-large", type=str) parser.add_argument("--dump_checkpoint", default="serialization_dir/tf_roberta_048131723.pth", type=str) parser.add_argument("--vocab_transform", action="store_true") SCREAMING_SNAKE_CASE = parser.parse_args() if args.model_type == "roberta": SCREAMING_SNAKE_CASE = RobertaForMaskedLM.from_pretrained(args.model_name) SCREAMING_SNAKE_CASE = "roberta" elif args.model_type == "gpt2": SCREAMING_SNAKE_CASE = GPTaLMHeadModel.from_pretrained(args.model_name) SCREAMING_SNAKE_CASE = "transformer" SCREAMING_SNAKE_CASE = model.state_dict() SCREAMING_SNAKE_CASE = {} # Embeddings # if args.model_type == "gpt2": for param_name in ["wte.weight", "wpe.weight"]: SCREAMING_SNAKE_CASE = state_dict[f'''{prefix}.{param_name}'''] else: for w in ["word_embeddings", "position_embeddings", "token_type_embeddings"]: SCREAMING_SNAKE_CASE = f'''{prefix}.embeddings.{w}.weight''' SCREAMING_SNAKE_CASE = state_dict[param_name] for w in ["weight", "bias"]: SCREAMING_SNAKE_CASE = f'''{prefix}.embeddings.LayerNorm.{w}''' SCREAMING_SNAKE_CASE = state_dict[param_name] # Transformer Blocks # SCREAMING_SNAKE_CASE = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: if args.model_type == "gpt2": for layer in ["ln_1", "attn.c_attn", "attn.c_proj", "ln_2", "mlp.c_fc", "mlp.c_proj"]: for w in ["weight", "bias"]: SCREAMING_SNAKE_CASE = state_dict[ f'''{prefix}.h.{teacher_idx}.{layer}.{w}''' ] SCREAMING_SNAKE_CASE = state_dict[f'''{prefix}.h.{teacher_idx}.attn.bias'''] else: for layer in [ "attention.self.query", "attention.self.key", "attention.self.value", "attention.output.dense", "attention.output.LayerNorm", "intermediate.dense", "output.dense", "output.LayerNorm", ]: for w in ["weight", "bias"]: SCREAMING_SNAKE_CASE = state_dict[ f'''{prefix}.encoder.layer.{teacher_idx}.{layer}.{w}''' ] std_idx += 1 # Language Modeling Head ###s if args.model_type == "roberta": for layer in ["lm_head.decoder.weight", "lm_head.bias"]: SCREAMING_SNAKE_CASE = state_dict[f'''{layer}'''] if args.vocab_transform: for w in ["weight", "bias"]: SCREAMING_SNAKE_CASE = state_dict[f'''lm_head.dense.{w}'''] SCREAMING_SNAKE_CASE = state_dict[f'''lm_head.layer_norm.{w}'''] elif args.model_type == "gpt2": for w in ["weight", "bias"]: SCREAMING_SNAKE_CASE = state_dict[f'''{prefix}.ln_f.{w}'''] SCREAMING_SNAKE_CASE = state_dict["lm_head.weight"] print(f'''N layers selected for distillation: {std_idx}''') print(f'''Number of params transferred for distillation: {len(compressed_sd.keys())}''') print(f'''Save transferred checkpoint to {args.dump_checkpoint}.''') torch.save(compressed_sd, args.dump_checkpoint)
23
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowerCamelCase ( lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : int = TextToVideoSDPipeline lowerCAmelCase_ : Dict = TEXT_TO_IMAGE_PARAMS lowerCAmelCase_ : Optional[Any] = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. lowerCAmelCase_ : Optional[Any] = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def A__ ( self ): torch.manual_seed(0 ) UpperCAmelCase_ = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) UpperCAmelCase_ = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowerCAmelCase , set_alpha_to_one=lowerCAmelCase , ) torch.manual_seed(0 ) UpperCAmelCase_ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) UpperCAmelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="gelu" , projection_dim=512 , ) UpperCAmelCase_ = CLIPTextModel(lowerCAmelCase ) UpperCAmelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCAmelCase_ = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def A__ ( self , lowerCAmelCase , lowerCAmelCase=0 ): if str(lowerCAmelCase ).startswith("mps" ): UpperCAmelCase_ = torch.manual_seed(lowerCAmelCase ) else: UpperCAmelCase_ = torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) UpperCAmelCase_ = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def A__ ( self ): UpperCAmelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ = self.get_dummy_components() UpperCAmelCase_ = TextToVideoSDPipeline(**lowerCAmelCase ) UpperCAmelCase_ = sd_pipe.to(lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase ) UpperCAmelCase_ = self.get_dummy_inputs(lowerCAmelCase ) UpperCAmelCase_ = "np" UpperCAmelCase_ = sd_pipe(**lowerCAmelCase ).frames UpperCAmelCase_ = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) UpperCAmelCase_ = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def A__ ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=3e-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def A__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=1e-2 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def A__ ( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def A__ ( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def A__ ( self ): pass def A__ ( self ): return super().test_progress_bar() @slow @skip_mps class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def A__ ( self ): UpperCAmelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy" ) UpperCAmelCase_ = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) UpperCAmelCase_ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) UpperCAmelCase_ = pipe.to("cuda" ) UpperCAmelCase_ = "Spiderman is surfing" UpperCAmelCase_ = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase_ = pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=25 , output_type="pt" ).frames UpperCAmelCase_ = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2 def A__ ( self ): UpperCAmelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy" ) UpperCAmelCase_ = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) UpperCAmelCase_ = pipe.to("cuda" ) UpperCAmelCase_ = "Spiderman is surfing" UpperCAmelCase_ = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase_ = pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=2 , output_type="pt" ).frames UpperCAmelCase_ = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2
23
1
from bisect import bisect from itertools import accumulate def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ = sorted(zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) , key=lambda __SCREAMING_SNAKE_CASE : x[0] / x[1] , reverse=__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ , UpperCAmelCase_ = [i[0] for i in r], [i[1] for i in r] UpperCAmelCase_ = list(accumulate(__SCREAMING_SNAKE_CASE ) ) UpperCAmelCase_ = bisect(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return ( 0 if k == 0 else sum(vl[:k] ) + (w - acc[k - 1]) * (vl[k]) / (wt[k]) if k != n else sum(vl[:k] ) ) if __name__ == "__main__": import doctest doctest.testmod()
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = [1] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = 0, 0, 0 UpperCAmelCase_ = ugly_nums[ia] * 2 UpperCAmelCase_ = ugly_nums[ia] * 3 UpperCAmelCase_ = ugly_nums[ia] * 5 for _ in range(1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ugly_nums.append(__SCREAMING_SNAKE_CASE ) if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 UpperCAmelCase_ = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(f'''{ugly_numbers(200) = }''')
23
1
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=() , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE="no" , __SCREAMING_SNAKE_CASE="29500" ) -> List[str]: UpperCAmelCase_ = False UpperCAmelCase_ = False if any(key.startswith("KAGGLE" ) for key in os.environ.keys() ): UpperCAmelCase_ = True elif "IPython" in sys.modules: UpperCAmelCase_ = "google.colab" in str(sys.modules["IPython"].get_ipython() ) try: UpperCAmelCase_ = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( f'''Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.''' ) if (in_colab or in_kaggle) and (os.environ.get("TPU_NAME" , __SCREAMING_SNAKE_CASE ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside " "your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if num_processes is None: UpperCAmelCase_ = 8 UpperCAmelCase_ = PrepareForLaunch(__SCREAMING_SNAKE_CASE , distributed_type="TPU" ) print(f'''Launching a training on {num_processes} TPU cores.''' ) xmp.spawn(__SCREAMING_SNAKE_CASE , args=__SCREAMING_SNAKE_CASE , nprocs=__SCREAMING_SNAKE_CASE , start_method="fork" ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on one CPU." ) function(*__SCREAMING_SNAKE_CASE ) else: if num_processes is None: raise ValueError( "You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call." ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( "To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized " "inside your training function. Restart your notebook and make sure no cells initializes an " "`Accelerator`." ) if torch.cuda.is_initialized(): raise ValueError( "To launch a multi-GPU training from your notebook, you need to avoid running any instruction " "using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA " "function." ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__SCREAMING_SNAKE_CASE , master_addr="127.0.01" , master_port=__SCREAMING_SNAKE_CASE , mixed_precision=__SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = PrepareForLaunch(__SCREAMING_SNAKE_CASE , distributed_type="MULTI_GPU" ) print(f'''Launching training on {num_processes} GPUs.''' ) try: start_processes(__SCREAMING_SNAKE_CASE , args=__SCREAMING_SNAKE_CASE , nprocs=__SCREAMING_SNAKE_CASE , start_method="fork" ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( "CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. " "This likely stems from an outside import causing issues once the `notebook_launcher()` is called. " "Please review your imports and test them when running the `notebook_launcher()` to identify " "which one is problematic." ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): UpperCAmelCase_ = "1" print("Launching training on MPS." ) elif torch.cuda.is_available(): print("Launching training on one GPU." ) else: print("Launching training on CPU." ) function(*__SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=() , __SCREAMING_SNAKE_CASE=2 ) -> Optional[Any]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=__SCREAMING_SNAKE_CASE , master_addr="127.0.01" , master_port="29500" , accelerate_mixed_precision="no" , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu="yes" , ): UpperCAmelCase_ = PrepareForLaunch(__SCREAMING_SNAKE_CASE , debug=__SCREAMING_SNAKE_CASE ) start_processes(__SCREAMING_SNAKE_CASE , args=__SCREAMING_SNAKE_CASE , nprocs=__SCREAMING_SNAKE_CASE , start_method="fork" )
23
import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=3 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=224 , lowerCAmelCase=1000 , lowerCAmelCase=[3, 3, 6, 4] , lowerCAmelCase=[48, 56, 112, 220] , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = is_training UpperCAmelCase_ = use_labels UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = num_labels UpperCAmelCase_ = image_size UpperCAmelCase_ = layer_depths UpperCAmelCase_ = embed_dims def A__ ( self ): UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase_ = self.get_config() return config, pixel_values, labels def A__ ( self ): return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="gelu" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=lowerCAmelCase , layer_scale_init_value=1e-5 , ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = SwiftFormerModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = SwiftFormerForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase , labels=lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) UpperCAmelCase_ = SwiftFormerForImageClassification(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A__ ( self ): ((UpperCAmelCase_) , (UpperCAmelCase_) , (UpperCAmelCase_)) = self.prepare_config_and_inputs() UpperCAmelCase_ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Optional[Any] = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () lowerCAmelCase_ : int = ( {'feature-extraction': SwiftFormerModel, 'image-classification': SwiftFormerForImageClassification} if is_torch_available() else {} ) lowerCAmelCase_ : List[Any] = False lowerCAmelCase_ : Dict = False lowerCAmelCase_ : int = False lowerCAmelCase_ : str = False lowerCAmelCase_ : Optional[Any] = False def A__ ( self ): UpperCAmelCase_ = SwiftFormerModelTester(self ) UpperCAmelCase_ = ConfigTester( self , config_class=lowerCAmelCase , has_text_modality=lowerCAmelCase , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def A__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="SwiftFormer does not use inputs_embeds" ) def A__ ( self ): pass def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase , nn.Linear ) ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase ) @slow def A__ ( self ): for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = SwiftFormerModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) @unittest.skip(reason="SwiftFormer does not output attentions" ) def A__ ( self ): pass def A__ ( self ): def check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = model_class(lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCAmelCase_ = model(**self._prepare_for_class(lowerCAmelCase , lowerCAmelCase ) ) UpperCAmelCase_ = outputs.hidden_states UpperCAmelCase_ = 8 self.assertEqual(len(lowerCAmelCase ) , lowerCAmelCase ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(lowerCAmelCase ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = True check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ = True check_hidden_states_output(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): def _config_zero_init(lowerCAmelCase ): UpperCAmelCase_ = copy.deepcopy(lowerCAmelCase ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(lowerCAmelCase , lowerCAmelCase , 1e-1_0 ) if isinstance(getattr(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) , lowerCAmelCase ): UpperCAmelCase_ = _config_zero_init(getattr(lowerCAmelCase , lowerCAmelCase ) ) setattr(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return configs_no_init UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = _config_zero_init(lowerCAmelCase ) for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(config=lowerCAmelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9) / 1e9).round().item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def A__ ( self ): pass def snake_case__ ( ) -> str: UpperCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def A__ ( self ): return ViTImageProcessor.from_pretrained("MBZUAI/swiftformer-xs" ) if is_vision_available() else None @slow def A__ ( self ): UpperCAmelCase_ = SwiftFormerForImageClassification.from_pretrained("MBZUAI/swiftformer-xs" ).to(lowerCAmelCase ) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(images=lowerCAmelCase , return_tensors="pt" ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase_ = model(**lowerCAmelCase ) # verify the logits UpperCAmelCase_ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase ) UpperCAmelCase_ = torch.tensor([[-2.1_7_0_3e0_0, 2.1_1_0_7e0_0, -2.0_8_1_1e0_0]] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase , atol=1e-4 ) )
23
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pathlib import Path import torch from ...utils import is_npu_available, is_xpu_available from .config_args import ClusterConfig, default_json_config_file from .config_utils import SubcommandHelpFormatter SCREAMING_SNAKE_CASE = "Create a default config file for Accelerate with only a few flags set." def snake_case__ ( __SCREAMING_SNAKE_CASE="no" , __SCREAMING_SNAKE_CASE = default_json_config_file , __SCREAMING_SNAKE_CASE = False ) -> Optional[Any]: UpperCAmelCase_ = Path(__SCREAMING_SNAKE_CASE ) path.parent.mkdir(parents=__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) if path.exists(): print( f'''Configuration already exists at {save_location}, will not override. Run `accelerate config` manually or pass a different `save_location`.''' ) return False UpperCAmelCase_ = mixed_precision.lower() if mixed_precision not in ["no", "fp16", "bf16", "fp8"]: raise ValueError( f'''`mixed_precision` should be one of \'no\', \'fp16\', \'bf16\', or \'fp8\'. Received {mixed_precision}''' ) UpperCAmelCase_ = { "compute_environment": "LOCAL_MACHINE", "mixed_precision": mixed_precision, } if torch.cuda.is_available(): UpperCAmelCase_ = torch.cuda.device_count() UpperCAmelCase_ = num_gpus UpperCAmelCase_ = False if num_gpus > 1: UpperCAmelCase_ = "MULTI_GPU" else: UpperCAmelCase_ = "NO" elif is_xpu_available() and use_xpu: UpperCAmelCase_ = torch.xpu.device_count() UpperCAmelCase_ = num_xpus UpperCAmelCase_ = False if num_xpus > 1: UpperCAmelCase_ = "MULTI_XPU" else: UpperCAmelCase_ = "NO" elif is_npu_available(): UpperCAmelCase_ = torch.npu.device_count() UpperCAmelCase_ = num_npus UpperCAmelCase_ = False if num_npus > 1: UpperCAmelCase_ = "MULTI_NPU" else: UpperCAmelCase_ = "NO" else: UpperCAmelCase_ = 0 UpperCAmelCase_ = True UpperCAmelCase_ = 1 UpperCAmelCase_ = "NO" UpperCAmelCase_ = ClusterConfig(**__SCREAMING_SNAKE_CASE ) config.to_json_file(__SCREAMING_SNAKE_CASE ) return path def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = parser.add_parser("default" , parents=__SCREAMING_SNAKE_CASE , help=__SCREAMING_SNAKE_CASE , formatter_class=__SCREAMING_SNAKE_CASE ) parser.add_argument( "--config_file" , default=__SCREAMING_SNAKE_CASE , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , dest="save_location" , ) parser.add_argument( "--mixed_precision" , choices=["no", "fp16", "bf16"] , type=__SCREAMING_SNAKE_CASE , help="Whether or not to use mixed precision training. " "Choose between FP16 and BF16 (bfloat16) training. " "BF16 training is only supported on Nvidia Ampere GPUs and PyTorch 1.10 or later." , default="no" , ) parser.set_defaults(func=__SCREAMING_SNAKE_CASE ) return parser def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ = write_basic_config(args.mixed_precision , args.save_location ) if config_file: print(f'''accelerate configuration saved at {config_file}''' )
23
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available SCREAMING_SNAKE_CASE = {"tokenization_herbert": ["HerbertTokenizer"]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["HerbertTokenizerFast"] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
import unittest from transformers import AutoConfig, AutoTokenizer, BertConfig, TensorType, is_flax_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, slow if is_flax_available(): import jax from transformers.models.auto.modeling_flax_auto import FlaxAutoModel from transformers.models.bert.modeling_flax_bert import FlaxBertModel from transformers.models.roberta.modeling_flax_roberta import FlaxRobertaModel @require_flax class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self ): for model_name in ["bert-base-cased", "bert-large-uncased"]: with self.subTest(lowerCAmelCase ): UpperCAmelCase_ = AutoConfig.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) self.assertIsInstance(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = FlaxAutoModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) self.assertIsInstance(lowerCAmelCase , lowerCAmelCase ) @slow def A__ ( self ): for model_name in ["roberta-base", "roberta-large"]: with self.subTest(lowerCAmelCase ): UpperCAmelCase_ = AutoConfig.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) self.assertIsInstance(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = FlaxAutoModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) self.assertIsInstance(lowerCAmelCase , lowerCAmelCase ) @slow def A__ ( self ): for model_name in ["bert-base-cased", "bert-large-uncased"]: UpperCAmelCase_ = AutoTokenizer.from_pretrained(lowerCAmelCase ) UpperCAmelCase_ = FlaxBertModel.from_pretrained(lowerCAmelCase ) UpperCAmelCase_ = tokenizer("Do you support jax jitted function?" , return_tensors=TensorType.JAX ) @jax.jit def eval(**lowerCAmelCase ): return model(**lowerCAmelCase ) eval(**lowerCAmelCase ).block_until_ready() @slow def A__ ( self ): for model_name in ["roberta-base", "roberta-large"]: UpperCAmelCase_ = AutoTokenizer.from_pretrained(lowerCAmelCase ) UpperCAmelCase_ = FlaxRobertaModel.from_pretrained(lowerCAmelCase ) UpperCAmelCase_ = tokenizer("Do you support jax jitted function?" , return_tensors=TensorType.JAX ) @jax.jit def eval(**lowerCAmelCase ): return model(**lowerCAmelCase ) eval(**lowerCAmelCase ).block_until_ready() def A__ ( self ): with self.assertRaisesRegex( lowerCAmelCase , "bert-base is not a local folder and is not a valid model identifier" ): UpperCAmelCase_ = FlaxAutoModel.from_pretrained("bert-base" ) def A__ ( self ): with self.assertRaisesRegex( lowerCAmelCase , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): UpperCAmelCase_ = FlaxAutoModel.from_pretrained(lowerCAmelCase , revision="aaaaaa" ) def A__ ( self ): with self.assertRaisesRegex( lowerCAmelCase , "hf-internal-testing/config-no-model does not appear to have a file named flax_model.msgpack" , ): UpperCAmelCase_ = FlaxAutoModel.from_pretrained("hf-internal-testing/config-no-model" ) def A__ ( self ): with self.assertRaisesRegex(lowerCAmelCase , "Use `from_pt=True` to load this model" ): UpperCAmelCase_ = FlaxAutoModel.from_pretrained("hf-internal-testing/tiny-bert-pt-only" )
23
import math def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> list[int]: UpperCAmelCase_ = [] UpperCAmelCase_ = 2 UpperCAmelCase_ = int(math.sqrt(__SCREAMING_SNAKE_CASE ) ) # Size of every segment UpperCAmelCase_ = [True] * (end + 1) UpperCAmelCase_ = [] while start <= end: if temp[start] is True: in_prime.append(__SCREAMING_SNAKE_CASE ) for i in range(start * start , end + 1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = False start += 1 prime += in_prime UpperCAmelCase_ = end + 1 UpperCAmelCase_ = min(2 * end , __SCREAMING_SNAKE_CASE ) while low <= n: UpperCAmelCase_ = [True] * (high - low + 1) for each in in_prime: UpperCAmelCase_ = math.floor(low / each ) * each if t < low: t += each for j in range(__SCREAMING_SNAKE_CASE , high + 1 , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = False for j in range(len(__SCREAMING_SNAKE_CASE ) ): if temp[j] is True: prime.append(j + low ) UpperCAmelCase_ = high + 1 UpperCAmelCase_ = min(high + end , __SCREAMING_SNAKE_CASE ) return prime print(sieve(10**6))
23
1
import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ = SwinConfig(image_size=192 ) if "base" in model_name: UpperCAmelCase_ = 6 UpperCAmelCase_ = 128 UpperCAmelCase_ = (2, 2, 18, 2) UpperCAmelCase_ = (4, 8, 16, 32) elif "large" in model_name: UpperCAmelCase_ = 12 UpperCAmelCase_ = 192 UpperCAmelCase_ = (2, 2, 18, 2) UpperCAmelCase_ = (6, 12, 24, 48) else: raise ValueError("Model not supported, only supports base and large variants" ) UpperCAmelCase_ = window_size UpperCAmelCase_ = embed_dim UpperCAmelCase_ = depths UpperCAmelCase_ = num_heads return config def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Optional[Any]: if "encoder.mask_token" in name: UpperCAmelCase_ = name.replace("encoder.mask_token" , "embeddings.mask_token" ) if "encoder.patch_embed.proj" in name: UpperCAmelCase_ = name.replace("encoder.patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "encoder.patch_embed.norm" in name: UpperCAmelCase_ = name.replace("encoder.patch_embed.norm" , "embeddings.norm" ) if "attn.proj" in name: UpperCAmelCase_ = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: UpperCAmelCase_ = name.replace("attn" , "attention.self" ) if "norm1" in name: UpperCAmelCase_ = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: UpperCAmelCase_ = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: UpperCAmelCase_ = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: UpperCAmelCase_ = name.replace("mlp.fc2" , "output.dense" ) if name == "encoder.norm.weight": UpperCAmelCase_ = "layernorm.weight" if name == "encoder.norm.bias": UpperCAmelCase_ = "layernorm.bias" if "decoder" in name: pass else: UpperCAmelCase_ = "swin." + name return name def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[str]: for key in orig_state_dict.copy().keys(): UpperCAmelCase_ = orig_state_dict.pop(__SCREAMING_SNAKE_CASE ) if "attn_mask" in key: pass elif "qkv" in key: UpperCAmelCase_ = key.split("." ) UpperCAmelCase_ = int(key_split[2] ) UpperCAmelCase_ = int(key_split[4] ) UpperCAmelCase_ = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: UpperCAmelCase_ = val[:dim, :] UpperCAmelCase_ = val[ dim : dim * 2, : ] UpperCAmelCase_ = val[-dim:, :] else: UpperCAmelCase_ = val[ :dim ] UpperCAmelCase_ = val[ dim : dim * 2 ] UpperCAmelCase_ = val[ -dim: ] else: UpperCAmelCase_ = val return orig_state_dict def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ = torch.load(__SCREAMING_SNAKE_CASE , map_location="cpu" )["model"] UpperCAmelCase_ = get_swin_config(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = SwinForMaskedImageModeling(__SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase_ = convert_state_dict(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) model.load_state_dict(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ = ViTImageProcessor(size={"height": 192, "width": 192} ) UpperCAmelCase_ = Image.open(requests.get(__SCREAMING_SNAKE_CASE , stream=__SCREAMING_SNAKE_CASE ).raw ) UpperCAmelCase_ = image_processor(images=__SCREAMING_SNAKE_CASE , return_tensors="pt" ) with torch.no_grad(): UpperCAmelCase_ = model(**__SCREAMING_SNAKE_CASE ).logits print(outputs.keys() ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__SCREAMING_SNAKE_CASE ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__SCREAMING_SNAKE_CASE ) if push_to_hub: print(f'''Pushing model and image processor for {model_name} to hub''' ) model.push_to_hub(f'''microsoft/{model_name}''' ) image_processor.push_to_hub(f'''microsoft/{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="swin-base-simmim-window6-192", type=str, choices=["swin-base-simmim-window6-192", "swin-large-simmim-window12-192"], help="Name of the Swin SimMIM model you'd like to convert.", ) parser.add_argument( "--checkpoint_path", default="/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth", type=str, help="Path to the original PyTorch checkpoint (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
23
from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : torch.FloatTensor class lowerCamelCase ( lowercase__, lowercase__ ): '''simple docstring''' @register_to_config def __init__( self , lowerCAmelCase = 32 , lowerCAmelCase = 64 , lowerCAmelCase = 20 , lowerCAmelCase = 768 , lowerCAmelCase=77 , lowerCAmelCase=4 , lowerCAmelCase = 0.0 , lowerCAmelCase = "silu" , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = "linear" , lowerCAmelCase = "prd" , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , ): super().__init__() UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = attention_head_dim UpperCAmelCase_ = num_attention_heads * attention_head_dim UpperCAmelCase_ = additional_embeddings UpperCAmelCase_ = time_embed_dim or inner_dim UpperCAmelCase_ = embedding_proj_dim or embedding_dim UpperCAmelCase_ = clip_embed_dim or embedding_dim UpperCAmelCase_ = Timesteps(lowerCAmelCase , lowerCAmelCase , 0 ) UpperCAmelCase_ = TimestepEmbedding(lowerCAmelCase , lowerCAmelCase , out_dim=lowerCAmelCase , act_fn=lowerCAmelCase ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) if embedding_proj_norm_type is None: UpperCAmelCase_ = None elif embedding_proj_norm_type == "layer": UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) else: raise ValueError(f'''unsupported embedding_proj_norm_type: {embedding_proj_norm_type}''' ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) if encoder_hid_proj_type is None: UpperCAmelCase_ = None elif encoder_hid_proj_type == "linear": UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) else: raise ValueError(f'''unsupported encoder_hid_proj_type: {encoder_hid_proj_type}''' ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , num_embeddings + additional_embeddings , lowerCAmelCase ) ) if added_emb_type == "prd": UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , 1 , lowerCAmelCase ) ) elif added_emb_type is None: UpperCAmelCase_ = None else: raise ValueError( f'''`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `\'prd\'` or `None`.''' ) UpperCAmelCase_ = nn.ModuleList( [ BasicTransformerBlock( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , dropout=lowerCAmelCase , activation_fn="gelu" , attention_bias=lowerCAmelCase , ) for d in range(lowerCAmelCase ) ] ) if norm_in_type == "layer": UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) elif norm_in_type is None: UpperCAmelCase_ = None else: raise ValueError(f'''Unsupported norm_in_type: {norm_in_type}.''' ) UpperCAmelCase_ = nn.LayerNorm(lowerCAmelCase ) UpperCAmelCase_ = nn.Linear(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] , -10000.0 ) causal_attention_mask.triu_(1 ) UpperCAmelCase_ = causal_attention_mask[None, ...] self.register_buffer("causal_attention_mask" , lowerCAmelCase , persistent=lowerCAmelCase ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) UpperCAmelCase_ = nn.Parameter(torch.zeros(1 , lowerCAmelCase ) ) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def A__ ( self ): UpperCAmelCase_ = {} def fn_recursive_add_processors(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if hasattr(lowerCAmelCase , "set_processor" ): UpperCAmelCase_ = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f'''{name}.{sub_name}''' , lowerCAmelCase , lowerCAmelCase ) return processors for name, module in self.named_children(): fn_recursive_add_processors(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return processors def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = len(self.attn_processors.keys() ) if isinstance(lowerCAmelCase , lowerCAmelCase ) and len(lowerCAmelCase ) != count: raise ValueError( f'''A dict of processors was passed, but the number of processors {len(lowerCAmelCase )} does not match the''' f''' number of attention layers: {count}. Please make sure to pass {count} processor classes.''' ) def fn_recursive_attn_processor(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if hasattr(lowerCAmelCase , "set_processor" ): if not isinstance(lowerCAmelCase , lowerCAmelCase ): module.set_processor(lowerCAmelCase ) else: module.set_processor(processor.pop(f'''{name}.processor''' ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f'''{name}.{sub_name}''' , lowerCAmelCase , lowerCAmelCase ) for name, module in self.named_children(): fn_recursive_attn_processor(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): self.set_attn_processor(AttnProcessor() ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = True , ): UpperCAmelCase_ = hidden_states.shape[0] UpperCAmelCase_ = timestep if not torch.is_tensor(lowerCAmelCase ): UpperCAmelCase_ = torch.tensor([timesteps] , dtype=torch.long , device=hidden_states.device ) elif torch.is_tensor(lowerCAmelCase ) and len(timesteps.shape ) == 0: UpperCAmelCase_ = timesteps[None].to(hidden_states.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML UpperCAmelCase_ = timesteps * torch.ones(lowerCAmelCase , dtype=timesteps.dtype , device=timesteps.device ) UpperCAmelCase_ = self.time_proj(lowerCAmelCase ) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. UpperCAmelCase_ = timesteps_projected.to(dtype=self.dtype ) UpperCAmelCase_ = self.time_embedding(lowerCAmelCase ) if self.embedding_proj_norm is not None: UpperCAmelCase_ = self.embedding_proj_norm(lowerCAmelCase ) UpperCAmelCase_ = self.embedding_proj(lowerCAmelCase ) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: UpperCAmelCase_ = self.encoder_hidden_states_proj(lowerCAmelCase ) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError("`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set" ) UpperCAmelCase_ = self.proj_in(lowerCAmelCase ) UpperCAmelCase_ = self.positional_embedding.to(hidden_states.dtype ) UpperCAmelCase_ = [] UpperCAmelCase_ = 0 if encoder_hidden_states is not None: additional_embeds.append(lowerCAmelCase ) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape ) == 2: UpperCAmelCase_ = proj_embeddings[:, None, :] if len(hidden_states.shape ) == 2: UpperCAmelCase_ = hidden_states[:, None, :] UpperCAmelCase_ = additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: UpperCAmelCase_ = self.prd_embedding.to(hidden_states.dtype ).expand(lowerCAmelCase , -1 , -1 ) additional_embeds.append(lowerCAmelCase ) UpperCAmelCase_ = torch.cat( lowerCAmelCase , dim=1 , ) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens UpperCAmelCase_ = additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: UpperCAmelCase_ = F.pad( lowerCAmelCase , ( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) , value=0.0 , ) UpperCAmelCase_ = hidden_states + positional_embeddings if attention_mask is not None: UpperCAmelCase_ = (1 - attention_mask.to(hidden_states.dtype )) * -10000.0 UpperCAmelCase_ = F.pad(lowerCAmelCase , (0, self.additional_embeddings) , value=0.0 ) UpperCAmelCase_ = (attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype ) UpperCAmelCase_ = attention_mask.repeat_interleave(self.config.num_attention_heads , dim=0 ) if self.norm_in is not None: UpperCAmelCase_ = self.norm_in(lowerCAmelCase ) for block in self.transformer_blocks: UpperCAmelCase_ = block(lowerCAmelCase , attention_mask=lowerCAmelCase ) UpperCAmelCase_ = self.norm_out(lowerCAmelCase ) if self.prd_embedding is not None: UpperCAmelCase_ = hidden_states[:, -1] else: UpperCAmelCase_ = hidden_states[:, additional_embeddings_len:] UpperCAmelCase_ = self.proj_to_clip_embeddings(lowerCAmelCase ) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=lowerCAmelCase ) def A__ ( self , lowerCAmelCase ): UpperCAmelCase_ = (prior_latents * self.clip_std) + self.clip_mean return prior_latents
23
1
from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class lowerCamelCase : '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = MBartConfig lowerCAmelCase_ : Dict = {} lowerCAmelCase_ : List[str] = 'gelu' def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=7 , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase=99 , lowerCAmelCase=32 , lowerCAmelCase=2 , lowerCAmelCase=4 , lowerCAmelCase=37 , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=20 , lowerCAmelCase=2 , lowerCAmelCase=1 , lowerCAmelCase=0 , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = seq_length UpperCAmelCase_ = is_training UpperCAmelCase_ = use_labels UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = eos_token_id UpperCAmelCase_ = pad_token_id UpperCAmelCase_ = bos_token_id def A__ ( self ): UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCAmelCase_ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCAmelCase_ = tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase_ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) UpperCAmelCase_ = prepare_mbart_inputs_dict(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return config, inputs_dict def A__ ( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = TFMBartModel(config=lowerCAmelCase ).get_decoder() UpperCAmelCase_ = inputs_dict["input_ids"] UpperCAmelCase_ = input_ids[:1, :] UpperCAmelCase_ = inputs_dict["attention_mask"][:1, :] UpperCAmelCase_ = inputs_dict["head_mask"] UpperCAmelCase_ = 1 # first forward pass UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase , head_mask=lowerCAmelCase , use_cache=lowerCAmelCase ) UpperCAmelCase_ , UpperCAmelCase_ = outputs.to_tuple() UpperCAmelCase_ = past_key_values[1] def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , ) -> Union[str, Any]: if attention_mask is None: UpperCAmelCase_ = tf.cast(tf.math.not_equal(__SCREAMING_SNAKE_CASE , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: UpperCAmelCase_ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: UpperCAmelCase_ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: UpperCAmelCase_ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: UpperCAmelCase_ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : int = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () lowerCAmelCase_ : Any = (TFMBartForConditionalGeneration,) if is_tf_available() else () lowerCAmelCase_ : Any = ( { 'conversational': TFMBartForConditionalGeneration, 'feature-extraction': TFMBartModel, 'summarization': TFMBartForConditionalGeneration, 'text2text-generation': TFMBartForConditionalGeneration, 'translation': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) lowerCAmelCase_ : str = True lowerCAmelCase_ : Union[str, Any] = False lowerCAmelCase_ : Dict = False def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def A__ ( self ): UpperCAmelCase_ = TFMBartModelTester(self ) UpperCAmelCase_ = ConfigTester(self , config_class=lowerCAmelCase ) def A__ ( self ): self.config_tester.run_common_tests() def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase ) @require_sentencepiece @require_tokenizers @require_tf class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : List[Any] = [ ' UN Chief Says There Is No Military Solution in Syria', ] lowerCAmelCase_ : Optional[int] = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', ] lowerCAmelCase_ : int = 'facebook/mbart-large-en-ro' @cached_property def A__ ( self ): return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def A__ ( self ): UpperCAmelCase_ = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def A__ ( self , **lowerCAmelCase ): UpperCAmelCase_ = self.translate_src_text(**lowerCAmelCase ) self.assertListEqual(self.expected_text , lowerCAmelCase ) def A__ ( self , **lowerCAmelCase ): UpperCAmelCase_ = self.tokenizer(self.src_text , **lowerCAmelCase , return_tensors="tf" ) UpperCAmelCase_ = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 ) UpperCAmelCase_ = self.tokenizer.batch_decode(lowerCAmelCase , skip_special_tokens=lowerCAmelCase ) return generated_words @slow def A__ ( self ): self._assert_generated_batch_equal_expected()
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE = { "configuration_gpt_bigcode": ["GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTBigCodeConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTBigCodeForSequenceClassification", "GPTBigCodeForTokenClassification", "GPTBigCodeForCausalLM", "GPTBigCodeModel", "GPTBigCodePreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record SCREAMING_SNAKE_CASE = "\\n@article{wang2019superglue,\n title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems},\n author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R},\n journal={arXiv preprint arXiv:1905.00537},\n year={2019}\n}\n" SCREAMING_SNAKE_CASE = "\\nSuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after\nGLUE with a new set of more difficult language understanding tasks, improved\nresources, and a new public leaderboard.\n" SCREAMING_SNAKE_CASE = "\nCompute SuperGLUE evaluation metric associated to each SuperGLUE dataset.\nArgs:\n predictions: list of predictions to score. Depending on the SuperGlUE subset:\n - for 'record': list of question-answer dictionaries with the following keys:\n - 'idx': index of the question as specified by the dataset\n - 'prediction_text': the predicted answer text\n - for 'multirc': list of question-answer dictionaries with the following keys:\n - 'idx': index of the question-answer pair as specified by the dataset\n - 'prediction': the predicted answer label\n - otherwise: list of predicted labels\n references: list of reference labels. Depending on the SuperGLUE subset:\n - for 'record': list of question-answers dictionaries with the following keys:\n - 'idx': index of the question as specified by the dataset\n - 'answers': list of possible answers\n - otherwise: list of reference labels\nReturns: depending on the SuperGLUE subset:\n - for 'record':\n - 'exact_match': Exact match between answer and gold answer\n - 'f1': F1 score\n - for 'multirc':\n - 'exact_match': Exact match between answer and gold answer\n - 'f1_m': Per-question macro-F1 score\n - 'f1_a': Average F1 score over all answers\n - for 'axb':\n 'matthews_correlation': Matthew Correlation\n - for 'cb':\n - 'accuracy': Accuracy\n - 'f1': F1 score\n - for all others:\n - 'accuracy': Accuracy\nExamples:\n\n >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"]\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0}\n\n >>> super_glue_metric = datasets.load_metric('super_glue', 'cb')\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0, 'f1': 1.0}\n\n >>> super_glue_metric = datasets.load_metric('super_glue', 'record')\n >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}]\n >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 1.0, 'f1': 1.0}\n\n >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc')\n >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0}\n\n >>> super_glue_metric = datasets.load_metric('super_glue', 'axb')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'matthews_correlation': 1.0}\n" def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[str]: return float((preds == labels).mean() ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE="binary" ) -> Union[str, Any]: UpperCAmelCase_ = simple_accuracy(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = float(fa_score(y_true=__SCREAMING_SNAKE_CASE , y_pred=__SCREAMING_SNAKE_CASE , average=__SCREAMING_SNAKE_CASE ) ) return { "accuracy": acc, "f1": fa, } def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[str]: UpperCAmelCase_ = {} for id_pred, label in zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): UpperCAmelCase_ = f'''{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}''' UpperCAmelCase_ = id_pred["prediction"] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase_ = [(pred, label)] UpperCAmelCase_ , UpperCAmelCase_ = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase_ , UpperCAmelCase_ = zip(*__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = fa_score(y_true=__SCREAMING_SNAKE_CASE , y_pred=__SCREAMING_SNAKE_CASE , average="macro" ) fas.append(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = int(sum(pred == label for pred, label in preds_labels ) == len(__SCREAMING_SNAKE_CASE ) ) ems.append(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = float(sum(__SCREAMING_SNAKE_CASE ) / len(__SCREAMING_SNAKE_CASE ) ) UpperCAmelCase_ = sum(__SCREAMING_SNAKE_CASE ) / len(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = float(fa_score(y_true=__SCREAMING_SNAKE_CASE , y_pred=[id_pred["prediction"] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase ( datasets.Metric ): '''simple docstring''' def A__ ( self ): if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="numpy" if not self.config_name == "record" and not self.config_name == "multirc" else None , ) def A__ ( self ): if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "prediction_text": datasets.Value("string" ), }, "references": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "answers": datasets.Sequence(datasets.Value("string" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("int64" ), "paragraph": datasets.Value("int64" ), "question": datasets.Value("int64" ), }, "prediction": datasets.Value("int64" ), }, "references": datasets.Value("int64" ), } else: return { "predictions": datasets.Value("int64" ), "references": datasets.Value("int64" ), } def A__ ( self , lowerCAmelCase , lowerCAmelCase ): if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(lowerCAmelCase , lowerCAmelCase )} elif self.config_name == "cb": return acc_and_fa(lowerCAmelCase , lowerCAmelCase , fa_avg="macro" ) elif self.config_name == "record": UpperCAmelCase_ = [ { "qas": [ {"id": ref["idx"]["query"], "answers": [{"text": ans} for ans in ref["answers"]]} for ref in references ] } ] UpperCAmelCase_ = {pred["idx"]["query"]: pred["prediction_text"] for pred in predictions} return evaluate_record(lowerCAmelCase , lowerCAmelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(lowerCAmelCase , lowerCAmelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(lowerCAmelCase , lowerCAmelCase )} else: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" )
23
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "xlm-roberta-base": "https://huggingface.co/xlm-roberta-base/resolve/main/config.json", "xlm-roberta-large": "https://huggingface.co/xlm-roberta-large/resolve/main/config.json", "xlm-roberta-large-finetuned-conll02-dutch": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll02-spanish": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-english": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json" ), "xlm-roberta-large-finetuned-conll03-german": ( "https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json" ), } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = 'xlm-roberta' def __init__( self , lowerCAmelCase=3_0522 , lowerCAmelCase=768 , lowerCAmelCase=12 , lowerCAmelCase=12 , lowerCAmelCase=3072 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=2 , lowerCAmelCase=0.02 , lowerCAmelCase=1e-1_2 , lowerCAmelCase=1 , lowerCAmelCase=0 , lowerCAmelCase=2 , lowerCAmelCase="absolute" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__(pad_token_id=lowerCAmelCase , bos_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = position_embedding_type UpperCAmelCase_ = use_cache UpperCAmelCase_ = classifier_dropout class lowerCamelCase ( lowercase__ ): '''simple docstring''' @property def A__ ( self ): if self.task == "multiple-choice": UpperCAmelCase_ = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
23
1
from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_herbert import HerbertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "allegro/herbert-base-cased": "https://huggingface.co/allegro/herbert-base-cased/resolve/main/vocab.json" }, "merges_file": { "allegro/herbert-base-cased": "https://huggingface.co/allegro/herbert-base-cased/resolve/main/merges.txt" }, } SCREAMING_SNAKE_CASE = {"allegro/herbert-base-cased": 514} SCREAMING_SNAKE_CASE = {} class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Optional[Any] = VOCAB_FILES_NAMES lowerCAmelCase_ : Dict = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : int = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : List[Any] = HerbertTokenizer def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase="<s>" , lowerCAmelCase="<unk>" , lowerCAmelCase="<pad>" , lowerCAmelCase="<mask>" , lowerCAmelCase="</s>" , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , lowerCAmelCase , tokenizer_file=lowerCAmelCase , cls_token=lowerCAmelCase , unk_token=lowerCAmelCase , pad_token=lowerCAmelCase , mask_token=lowerCAmelCase , sep_token=lowerCAmelCase , **lowerCAmelCase , ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.cls_token_id] UpperCAmelCase_ = [self.sep_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase , token_ids_a=lowerCAmelCase , already_has_special_tokens=lowerCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase )) + [1] return [1] + ([0] * len(lowerCAmelCase )) + [1] + ([0] * len(lowerCAmelCase )) + [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = int(__SCREAMING_SNAKE_CASE ) if decimal in (0, 1): # Exit cases for the recursion return str(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ , UpperCAmelCase_ = divmod(__SCREAMING_SNAKE_CASE , 2 ) return binary_recursive(__SCREAMING_SNAKE_CASE ) + str(__SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = str(__SCREAMING_SNAKE_CASE ).strip() if not number: raise ValueError("No input value was provided" ) UpperCAmelCase_ = "-" if number.startswith("-" ) else "" UpperCAmelCase_ = number.lstrip("-" ) if not number.isnumeric(): raise ValueError("Input value is not an integer" ) return f'''{negative}0b{binary_recursive(int(__SCREAMING_SNAKE_CASE ) )}''' if __name__ == "__main__": from doctest import testmod testmod()
23
1
import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase=14 , lowerCAmelCase=7 , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase=True , lowerCAmelCase=99 , lowerCAmelCase=32 , lowerCAmelCase=4 , lowerCAmelCase=4 , lowerCAmelCase=4 , lowerCAmelCase=37 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=512 , lowerCAmelCase=0.02 , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = seq_length UpperCAmelCase_ = is_training UpperCAmelCase_ = use_input_mask UpperCAmelCase_ = use_token_type_ids UpperCAmelCase_ = use_labels UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = rotary_dim UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = initializer_range UpperCAmelCase_ = None UpperCAmelCase_ = vocab_size - 1 UpperCAmelCase_ = vocab_size - 1 UpperCAmelCase_ = vocab_size - 1 def A__ ( self ): UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase_ = None if self.use_input_mask: UpperCAmelCase_ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=lowerCAmelCase , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def A__ ( self ): UpperCAmelCase_ = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = config_and_inputs UpperCAmelCase_ = {"input_ids": input_ids, "attention_mask": attention_mask} return config, inputs_dict def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = 20 UpperCAmelCase_ = model_class_name(lowerCAmelCase ) UpperCAmelCase_ = model.init_cache(input_ids.shape[0] , lowerCAmelCase ) UpperCAmelCase_ = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype="i4" ) UpperCAmelCase_ = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) UpperCAmelCase_ = model( input_ids[:, :-1] , attention_mask=lowerCAmelCase , past_key_values=lowerCAmelCase , position_ids=lowerCAmelCase , ) UpperCAmelCase_ = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="i4" ) UpperCAmelCase_ = model( input_ids[:, -1:] , attention_mask=lowerCAmelCase , past_key_values=outputs_cache.past_key_values , position_ids=lowerCAmelCase , ) UpperCAmelCase_ = model(lowerCAmelCase ) UpperCAmelCase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f'''Max diff is {diff}''' ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = 20 UpperCAmelCase_ = model_class_name(lowerCAmelCase ) UpperCAmelCase_ = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]) )] , axis=-1 , ) UpperCAmelCase_ = model.init_cache(input_ids.shape[0] , lowerCAmelCase ) UpperCAmelCase_ = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) UpperCAmelCase_ = model( input_ids[:, :-1] , attention_mask=lowerCAmelCase , past_key_values=lowerCAmelCase , position_ids=lowerCAmelCase , ) UpperCAmelCase_ = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="i4" ) UpperCAmelCase_ = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=lowerCAmelCase , position_ids=lowerCAmelCase , ) UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase ) UpperCAmelCase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f'''Max diff is {diff}''' ) @require_flax class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () lowerCAmelCase_ : Dict = (FlaxGPTJForCausalLM,) if is_flax_available() else () def A__ ( self ): UpperCAmelCase_ = FlaxGPTJModelTester(self ) def A__ ( self ): for model_class_name in self.all_model_classes: UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): for model_class_name in self.all_model_classes: UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) @tooslow def A__ ( self ): UpperCAmelCase_ = GPTaTokenizer.from_pretrained("gpt2" , pad_token="<|endoftext|>" , padding_side="left" ) UpperCAmelCase_ = tokenizer(["Hello this is a long string", "Hey"] , return_tensors="np" , padding=lowerCAmelCase , truncation=lowerCAmelCase ) UpperCAmelCase_ = FlaxGPTJForCausalLM.from_pretrained("EleutherAI/gpt-j-6B" ) UpperCAmelCase_ = False UpperCAmelCase_ = model.config.eos_token_id UpperCAmelCase_ = jax.jit(model.generate ) UpperCAmelCase_ = jit_generate( inputs["input_ids"] , attention_mask=inputs["attention_mask"] , pad_token_id=tokenizer.pad_token_id ).sequences UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase , skip_special_tokens=lowerCAmelCase ) UpperCAmelCase_ = [ "Hello this is a long string of text.\n\nI'm trying to get the text of the", "Hey, I'm a little late to the party. I'm going to", ] self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) @is_pt_flax_cross_test def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs UpperCAmelCase_ = self._prepare_for_class(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class UpperCAmelCase_ = model_class.__name__[4:] # Skip the "Flax" at the beginning UpperCAmelCase_ = getattr(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ , UpperCAmelCase_ = pt_inputs["input_ids"].shape UpperCAmelCase_ = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCAmelCase ): UpperCAmelCase_ = 0 UpperCAmelCase_ = 1 UpperCAmelCase_ = 0 UpperCAmelCase_ = 1 UpperCAmelCase_ = pt_model_class(lowerCAmelCase ).eval() UpperCAmelCase_ = model_class(lowerCAmelCase , dtype=jnp.floataa ) UpperCAmelCase_ = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , lowerCAmelCase ) UpperCAmelCase_ = fx_state with torch.no_grad(): UpperCAmelCase_ = pt_model(**lowerCAmelCase ).to_tuple() UpperCAmelCase_ = fx_model(**lowerCAmelCase ).to_tuple() self.assertEqual(len(lowerCAmelCase ) , len(lowerCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(lowerCAmelCase , lowerCAmelCase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4e-2 ) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(lowerCAmelCase ) UpperCAmelCase_ = model_class.from_pretrained(lowerCAmelCase , from_pt=lowerCAmelCase ) UpperCAmelCase_ = fx_model_loaded(**lowerCAmelCase ).to_tuple() self.assertEqual( len(lowerCAmelCase ) , len(lowerCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output_loaded, pt_output in zip(lowerCAmelCase , lowerCAmelCase ): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4e-2 ) @is_pt_flax_cross_test def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs UpperCAmelCase_ = self._prepare_for_class(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class UpperCAmelCase_ = model_class.__name__[4:] # Skip the "Flax" at the beginning UpperCAmelCase_ = getattr(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = pt_model_class(lowerCAmelCase ).eval() UpperCAmelCase_ = model_class(lowerCAmelCase , dtype=jnp.floataa ) UpperCAmelCase_ = load_flax_weights_in_pytorch_model(lowerCAmelCase , fx_model.params ) UpperCAmelCase_ , UpperCAmelCase_ = pt_inputs["input_ids"].shape UpperCAmelCase_ = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCAmelCase ): UpperCAmelCase_ = 0 UpperCAmelCase_ = 1 UpperCAmelCase_ = 0 UpperCAmelCase_ = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): UpperCAmelCase_ = pt_model(**lowerCAmelCase ).to_tuple() UpperCAmelCase_ = fx_model(**lowerCAmelCase ).to_tuple() self.assertEqual(len(lowerCAmelCase ) , len(lowerCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(lowerCAmelCase , lowerCAmelCase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4e-2 ) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(lowerCAmelCase ) UpperCAmelCase_ = pt_model_class.from_pretrained(lowerCAmelCase , from_flax=lowerCAmelCase ) with torch.no_grad(): UpperCAmelCase_ = pt_model_loaded(**lowerCAmelCase ).to_tuple() self.assertEqual( len(lowerCAmelCase ) , len(lowerCAmelCase ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(lowerCAmelCase , lowerCAmelCase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4e-2 ) @tooslow def A__ ( self ): for model_class_name in self.all_model_classes: UpperCAmelCase_ = model_class_name.from_pretrained("EleutherAI/gpt-j-6B" ) UpperCAmelCase_ = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCAmelCase )
23
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self ): UpperCAmelCase_ = AutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" , return_dict=lowerCAmelCase ).to(lowerCAmelCase ) UpperCAmelCase_ = AutoTokenizer.from_pretrained("google/mt5-small" ) UpperCAmelCase_ = tokenizer("Hello there" , return_tensors="pt" ).input_ids UpperCAmelCase_ = tokenizer("Hi I am" , return_tensors="pt" ).input_ids UpperCAmelCase_ = model(input_ids.to(lowerCAmelCase ) , labels=labels.to(lowerCAmelCase ) ).loss UpperCAmelCase_ = -(labels.shape[-1] * loss.item()) UpperCAmelCase_ = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
23
1
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> np.array: UpperCAmelCase_ = f'''{sampling_rate}''' UpperCAmelCase_ = "1" UpperCAmelCase_ = "f32le" UpperCAmelCase_ = [ "ffmpeg", "-i", "pipe:0", "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-hide_banner", "-loglevel", "quiet", "pipe:1", ] try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: UpperCAmelCase_ = ffmpeg_process.communicate(__SCREAMING_SNAKE_CASE ) except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to load audio files from filename" ) from error UpperCAmelCase_ = output_stream[0] UpperCAmelCase_ = np.frombuffer(__SCREAMING_SNAKE_CASE , np.floataa ) if audio.shape[0] == 0: raise ValueError("Malformed soundfile" ) return audio def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = "f32le" , ) -> Dict: UpperCAmelCase_ = f'''{sampling_rate}''' UpperCAmelCase_ = "1" if format_for_conversion == "s16le": UpperCAmelCase_ = 2 elif format_for_conversion == "f32le": UpperCAmelCase_ = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) UpperCAmelCase_ = platform.system() if system == "Linux": UpperCAmelCase_ = "alsa" UpperCAmelCase_ = "default" elif system == "Darwin": UpperCAmelCase_ = "avfoundation" UpperCAmelCase_ = ":0" elif system == "Windows": UpperCAmelCase_ = "dshow" UpperCAmelCase_ = "default" UpperCAmelCase_ = [ "ffmpeg", "-f", format_, "-i", input_, "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-fflags", "nobuffer", "-hide_banner", "-loglevel", "quiet", "pipe:1", ] UpperCAmelCase_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample UpperCAmelCase_ = _ffmpeg_stream(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for item in iterator: yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = "f32le" , ) -> int: if stream_chunk_s is not None: UpperCAmelCase_ = stream_chunk_s else: UpperCAmelCase_ = chunk_length_s UpperCAmelCase_ = ffmpeg_microphone(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , format_for_conversion=__SCREAMING_SNAKE_CASE ) if format_for_conversion == "s16le": UpperCAmelCase_ = np.intaa UpperCAmelCase_ = 2 elif format_for_conversion == "f32le": UpperCAmelCase_ = np.floataa UpperCAmelCase_ = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: UpperCAmelCase_ = chunk_length_s / 6 UpperCAmelCase_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(__SCREAMING_SNAKE_CASE , (int, float) ): UpperCAmelCase_ = [stride_length_s, stride_length_s] UpperCAmelCase_ = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample UpperCAmelCase_ = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample UpperCAmelCase_ = datetime.datetime.now() UpperCAmelCase_ = datetime.timedelta(seconds=__SCREAMING_SNAKE_CASE ) for item in chunk_bytes_iter(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=(stride_left, stride_right) , stream=__SCREAMING_SNAKE_CASE ): # Put everything back in numpy scale UpperCAmelCase_ = np.frombuffer(item["raw"] , dtype=__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = ( item["stride"][0] // size_of_sample, item["stride"][1] // size_of_sample, ) UpperCAmelCase_ = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = False ) -> Dict: UpperCAmelCase_ = B"" UpperCAmelCase_ , UpperCAmelCase_ = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) UpperCAmelCase_ = 0 for raw in iterator: acc += raw if stream and len(__SCREAMING_SNAKE_CASE ) < chunk_len: UpperCAmelCase_ = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(__SCREAMING_SNAKE_CASE ) >= chunk_len: # We are flushing the accumulator UpperCAmelCase_ = (_stride_left, stride_right) UpperCAmelCase_ = {"raw": acc[:chunk_len], "stride": stride} if stream: UpperCAmelCase_ = False yield item UpperCAmelCase_ = stride_left UpperCAmelCase_ = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(__SCREAMING_SNAKE_CASE ) > stride_left: UpperCAmelCase_ = {"raw": acc, "stride": (_stride_left, 0)} if stream: UpperCAmelCase_ = False yield item def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[Any]: UpperCAmelCase_ = 2**24 # 16Mo try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdout=subprocess.PIPE , bufsize=__SCREAMING_SNAKE_CASE ) as ffmpeg_process: while True: UpperCAmelCase_ = ffmpeg_process.stdout.read(__SCREAMING_SNAKE_CASE ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to stream audio files from filename" ) from error
23
def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ = 0 while b > 0: if b & 1: UpperCAmelCase_ = ((res % c) + (a % c)) % c a += a b >>= 1 return res
23
1
import os import time import numpy as np import onnxruntime as ort SCREAMING_SNAKE_CASE = "1" SCREAMING_SNAKE_CASE = "0" SCREAMING_SNAKE_CASE = "1" SCREAMING_SNAKE_CASE = ort.SessionOptions() SCREAMING_SNAKE_CASE = ort.GraphOptimizationLevel.ORT_DISABLE_ALL print("Create inference session...") SCREAMING_SNAKE_CASE = ["TensorrtExecutionProvider", "CUDAExecutionProvider"] SCREAMING_SNAKE_CASE = ort.InferenceSession("model.onnx", sess_options=sess_opt, providers=execution_provider) SCREAMING_SNAKE_CASE = ort.RunOptions() SCREAMING_SNAKE_CASE = 128 SCREAMING_SNAKE_CASE = 1 SCREAMING_SNAKE_CASE = np.ones((batch, sequence), dtype=np.intaa) SCREAMING_SNAKE_CASE = np.ones((batch, sequence), dtype=np.intaa) SCREAMING_SNAKE_CASE = np.ones((batch, sequence), dtype=np.intaa) print("Warm up phase...") sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print("Start inference...") SCREAMING_SNAKE_CASE = time.time() SCREAMING_SNAKE_CASE = 2000 SCREAMING_SNAKE_CASE = {} for iter in range(max_iters): SCREAMING_SNAKE_CASE = sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print("Average Inference Time = {:.3f} ms".format((time.time() - start_time) * 1000 / max_iters))
23
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> List[List[ImageInput]]: if isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(__SCREAMING_SNAKE_CASE ): return [[videos]] raise ValueError(f'''Could not make batched video from {videos}''' ) class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[Any] = ['pixel_values'] def __init__( self , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = True , lowerCAmelCase = 1 / 255 , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = None , **lowerCAmelCase , ): super().__init__(**lowerCAmelCase ) UpperCAmelCase_ = size if size is not None else {"shortest_edge": 224} UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) UpperCAmelCase_ = crop_size if crop_size is not None else {"height": 224, "width": 224} UpperCAmelCase_ = get_size_dict(lowerCAmelCase , param_name="crop_size" ) UpperCAmelCase_ = do_resize UpperCAmelCase_ = size UpperCAmelCase_ = do_center_crop UpperCAmelCase_ = crop_size UpperCAmelCase_ = resample UpperCAmelCase_ = do_rescale UpperCAmelCase_ = rescale_factor UpperCAmelCase_ = do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = PILImageResampling.BILINEAR , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) if "shortest_edge" in size: UpperCAmelCase_ = get_resize_output_image_size(lowerCAmelCase , size["shortest_edge"] , default_to_square=lowerCAmelCase ) elif "height" in size and "width" in size: UpperCAmelCase_ = (size["height"], size["width"]) else: raise ValueError(f'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): UpperCAmelCase_ = get_size_dict(lowerCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(lowerCAmelCase , size=(size["height"], size["width"]) , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return rescale(lowerCAmelCase , scale=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return normalize(lowerCAmelCase , mean=lowerCAmelCase , std=lowerCAmelCase , data_format=lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , ): if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. UpperCAmelCase_ = to_numpy_array(lowerCAmelCase ) if do_resize: UpperCAmelCase_ = self.resize(image=lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase ) if do_center_crop: UpperCAmelCase_ = self.center_crop(lowerCAmelCase , size=lowerCAmelCase ) if do_rescale: UpperCAmelCase_ = self.rescale(image=lowerCAmelCase , scale=lowerCAmelCase ) if do_normalize: UpperCAmelCase_ = self.normalize(image=lowerCAmelCase , mean=lowerCAmelCase , std=lowerCAmelCase ) UpperCAmelCase_ = to_channel_dimension_format(lowerCAmelCase , lowerCAmelCase ) return image def A__ ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , **lowerCAmelCase , ): UpperCAmelCase_ = do_resize if do_resize is not None else self.do_resize UpperCAmelCase_ = resample if resample is not None else self.resample UpperCAmelCase_ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase_ = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase_ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase_ = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase_ = image_mean if image_mean is not None else self.image_mean UpperCAmelCase_ = image_std if image_std is not None else self.image_std UpperCAmelCase_ = size if size is not None else self.size UpperCAmelCase_ = get_size_dict(lowerCAmelCase , default_to_square=lowerCAmelCase ) UpperCAmelCase_ = crop_size if crop_size is not None else self.crop_size UpperCAmelCase_ = get_size_dict(lowerCAmelCase , param_name="crop_size" ) if not valid_images(lowerCAmelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) UpperCAmelCase_ = make_batched(lowerCAmelCase ) UpperCAmelCase_ = [ [ self._preprocess_image( image=lowerCAmelCase , do_resize=lowerCAmelCase , size=lowerCAmelCase , resample=lowerCAmelCase , do_center_crop=lowerCAmelCase , crop_size=lowerCAmelCase , do_rescale=lowerCAmelCase , rescale_factor=lowerCAmelCase , do_normalize=lowerCAmelCase , image_mean=lowerCAmelCase , image_std=lowerCAmelCase , data_format=lowerCAmelCase , ) for img in video ] for video in videos ] UpperCAmelCase_ = {"pixel_values": videos} return BatchFeature(data=lowerCAmelCase , tensor_type=lowerCAmelCase )
23
1
import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html SCREAMING_SNAKE_CASE = "platform" import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , ) -> Union[str, Any]: if attention_mask is None: UpperCAmelCase_ = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: UpperCAmelCase_ = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: UpperCAmelCase_ = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: UpperCAmelCase_ = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: UpperCAmelCase_ = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=7 , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase=99 , lowerCAmelCase=16 , lowerCAmelCase=2 , lowerCAmelCase=4 , lowerCAmelCase=4 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=32 , lowerCAmelCase=2 , lowerCAmelCase=1 , lowerCAmelCase=0 , lowerCAmelCase=0.02 , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = seq_length UpperCAmelCase_ = is_training UpperCAmelCase_ = use_labels UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = eos_token_id UpperCAmelCase_ = pad_token_id UpperCAmelCase_ = bos_token_id UpperCAmelCase_ = initializer_range def A__ ( self ): UpperCAmelCase_ = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) UpperCAmelCase_ = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) UpperCAmelCase_ = shift_tokens_right(lowerCAmelCase , 1 , 2 ) UpperCAmelCase_ = BlenderbotConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=lowerCAmelCase , ) UpperCAmelCase_ = prepare_blenderbot_inputs_dict(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) return config, inputs_dict def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.prepare_config_and_inputs() return config, inputs_dict def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = 20 UpperCAmelCase_ = model_class_name(lowerCAmelCase ) UpperCAmelCase_ = model.encode(inputs_dict["input_ids"] ) UpperCAmelCase_ , UpperCAmelCase_ = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) UpperCAmelCase_ = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="i4" ) UpperCAmelCase_ = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) UpperCAmelCase_ = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase , decoder_attention_mask=lowerCAmelCase , past_key_values=lowerCAmelCase , decoder_position_ids=lowerCAmelCase , ) UpperCAmelCase_ = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) UpperCAmelCase_ = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase , decoder_attention_mask=lowerCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCAmelCase , ) UpperCAmelCase_ = model.decode(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f'''Max diff is {diff}''' ) def A__ ( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = 20 UpperCAmelCase_ = model_class_name(lowerCAmelCase ) UpperCAmelCase_ = model.encode(inputs_dict["input_ids"] ) UpperCAmelCase_ , UpperCAmelCase_ = ( inputs_dict["decoder_input_ids"], inputs_dict["decoder_attention_mask"], ) UpperCAmelCase_ = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) UpperCAmelCase_ = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) UpperCAmelCase_ = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase , decoder_attention_mask=lowerCAmelCase , past_key_values=lowerCAmelCase , decoder_position_ids=lowerCAmelCase , ) UpperCAmelCase_ = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="i4" ) UpperCAmelCase_ = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCAmelCase , decoder_position_ids=lowerCAmelCase , ) UpperCAmelCase_ = model.decode(lowerCAmelCase , lowerCAmelCase , decoder_attention_mask=lowerCAmelCase ) UpperCAmelCase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f'''Max diff is {diff}''' ) @require_flax class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Dict = 99 def A__ ( self ): UpperCAmelCase_ = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) UpperCAmelCase_ = input_ids.shape[0] UpperCAmelCase_ = BlenderbotConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = self._get_config_and_data() UpperCAmelCase_ = FlaxBlenderbotForConditionalGeneration(lowerCAmelCase ) UpperCAmelCase_ = lm_model(input_ids=lowerCAmelCase ) UpperCAmelCase_ = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["logits"].shape , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = BlenderbotConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) UpperCAmelCase_ = FlaxBlenderbotForConditionalGeneration(lowerCAmelCase ) UpperCAmelCase_ = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) UpperCAmelCase_ = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) UpperCAmelCase_ = lm_model(input_ids=lowerCAmelCase , decoder_input_ids=lowerCAmelCase ) UpperCAmelCase_ = (*summary.shape, config.vocab_size) self.assertEqual(outputs["logits"].shape , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) UpperCAmelCase_ = shift_tokens_right(lowerCAmelCase , 1 , 2 ) UpperCAmelCase_ = np.equal(lowerCAmelCase , 1 ).astype(np.floataa ).sum() UpperCAmelCase_ = np.equal(lowerCAmelCase , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(lowerCAmelCase , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class lowerCamelCase ( lowercase__, unittest.TestCase, lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Dict = True lowerCAmelCase_ : List[str] = ( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) lowerCAmelCase_ : Optional[int] = (FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def A__ ( self ): UpperCAmelCase_ = FlaxBlenderbotModelTester(self ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase_ = self._prepare_for_class(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = model_class(lowerCAmelCase ) @jax.jit def encode_jitted(lowerCAmelCase , lowerCAmelCase=None , **lowerCAmelCase ): return model.encode(input_ids=lowerCAmelCase , attention_mask=lowerCAmelCase ) with self.subTest("JIT Enabled" ): UpperCAmelCase_ = encode_jitted(**lowerCAmelCase ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): UpperCAmelCase_ = encode_jitted(**lowerCAmelCase ).to_tuple() self.assertEqual(len(lowerCAmelCase ) , len(lowerCAmelCase ) ) for jitted_output, output in zip(lowerCAmelCase , lowerCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = model.encode(inputs_dict["input_ids"] , inputs_dict["attention_mask"] ) UpperCAmelCase_ = { "decoder_input_ids": inputs_dict["decoder_input_ids"], "decoder_attention_mask": inputs_dict["decoder_attention_mask"], "encoder_outputs": encoder_outputs, } @jax.jit def decode_jitted(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): return model.decode( decoder_input_ids=lowerCAmelCase , decoder_attention_mask=lowerCAmelCase , encoder_outputs=lowerCAmelCase , ) with self.subTest("JIT Enabled" ): UpperCAmelCase_ = decode_jitted(**lowerCAmelCase ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): UpperCAmelCase_ = decode_jitted(**lowerCAmelCase ).to_tuple() self.assertEqual(len(lowerCAmelCase ) , len(lowerCAmelCase ) ) for jitted_output, output in zip(lowerCAmelCase , lowerCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def A__ ( self ): for model_class_name in self.all_model_classes: UpperCAmelCase_ = model_class_name.from_pretrained("facebook/blenderbot-400M-distill" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids UpperCAmelCase_ = np.ones((1, 1) ) * model.config.eos_token_id UpperCAmelCase_ = model(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) @unittest.skipUnless(jax_device != "cpu" , "3B test too slow on CPU." ) @slow def A__ ( self ): UpperCAmelCase_ = {"num_beams": 1, "early_stopping": True, "min_length": 15, "max_length": 25} UpperCAmelCase_ = {"skip_special_tokens": True, "clean_up_tokenization_spaces": True} UpperCAmelCase_ = FlaxBlenderbotForConditionalGeneration.from_pretrained("facebook/blenderbot-3B" , from_pt=lowerCAmelCase ) UpperCAmelCase_ = BlenderbotTokenizer.from_pretrained("facebook/blenderbot-3B" ) UpperCAmelCase_ = ["Sam"] UpperCAmelCase_ = tokenizer(lowerCAmelCase , return_tensors="jax" ) UpperCAmelCase_ = model.generate(**lowerCAmelCase , **lowerCAmelCase ) UpperCAmelCase_ = "Sam is a great name. It means \"sun\" in Gaelic." UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase , **lowerCAmelCase ) assert generated_txt[0].strip() == tgt_text
23
def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 1 for i in range(1 , num + 1 ): fact *= i return fact def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ = 0 while number > 0: UpperCAmelCase_ = number % 10 sum_of_digits += last_digit UpperCAmelCase_ = number // 10 # Removing the last_digit from the given number return sum_of_digits def snake_case__ ( __SCREAMING_SNAKE_CASE = 100 ) -> int: UpperCAmelCase_ = factorial(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = split_and_add(__SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
23
1
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None ) -> List[str]: if attention_mask is None: UpperCAmelCase_ = tf.cast(tf.math.not_equal(__SCREAMING_SNAKE_CASE , config.pad_token_id ) , tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class lowerCamelCase : '''simple docstring''' lowerCAmelCase_ : Any = OPTConfig lowerCAmelCase_ : int = {} lowerCAmelCase_ : Union[str, Any] = 'gelu' def __init__( self , lowerCAmelCase , lowerCAmelCase=13 , lowerCAmelCase=7 , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase=99 , lowerCAmelCase=16 , lowerCAmelCase=2 , lowerCAmelCase=4 , lowerCAmelCase=4 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=20 , lowerCAmelCase=2 , lowerCAmelCase=1 , lowerCAmelCase=0 , lowerCAmelCase=16 , lowerCAmelCase=16 , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = seq_length UpperCAmelCase_ = is_training UpperCAmelCase_ = use_labels UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = eos_token_id UpperCAmelCase_ = pad_token_id UpperCAmelCase_ = bos_token_id UpperCAmelCase_ = embed_dim UpperCAmelCase_ = word_embed_proj_dim UpperCAmelCase_ = False def A__ ( self ): UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCAmelCase_ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCAmelCase_ = tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCAmelCase_ = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=lowerCAmelCase , **self.config_updates , ) UpperCAmelCase_ = prepare_opt_inputs_dict(lowerCAmelCase , lowerCAmelCase ) return config, inputs_dict def A__ ( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = TFOPTModel(config=lowerCAmelCase ) UpperCAmelCase_ = inputs_dict["input_ids"] UpperCAmelCase_ = input_ids[:1, :] UpperCAmelCase_ = inputs_dict["attention_mask"][:1, :] UpperCAmelCase_ = 1 # first forward pass UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase , use_cache=lowerCAmelCase ) UpperCAmelCase_ , UpperCAmelCase_ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCAmelCase_ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and UpperCAmelCase_ = tf.concat([input_ids, next_tokens] , axis=-1 ) UpperCAmelCase_ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase )[0] UpperCAmelCase_ = model(lowerCAmelCase , attention_mask=lowerCAmelCase , past_key_values=lowerCAmelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice UpperCAmelCase_ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) UpperCAmelCase_ = output_from_no_past[:, -3:, random_slice_idx] UpperCAmelCase_ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase , lowerCAmelCase , rtol=1e-3 ) @require_tf class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Tuple = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () lowerCAmelCase_ : Optional[int] = (TFOPTForCausalLM,) if is_tf_available() else () lowerCAmelCase_ : Tuple = ( {'feature-extraction': TFOPTModel, 'text-generation': TFOPTForCausalLM} if is_tf_available() else {} ) lowerCAmelCase_ : List[str] = False lowerCAmelCase_ : Tuple = False lowerCAmelCase_ : Union[str, Any] = False lowerCAmelCase_ : Tuple = 10 def A__ ( self ): UpperCAmelCase_ = TFOPTModelTester(self ) UpperCAmelCase_ = ConfigTester(self , config_class=lowerCAmelCase ) def A__ ( self ): self.config_tester.run_common_tests() def A__ ( self ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(lowerCAmelCase , lowerCAmelCase ): if hasattr(lowerCAmelCase , "weight" ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(lowerCAmelCase , "weight" ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings UpperCAmelCase_ = model_class(config=lowerCAmelCase ) UpperCAmelCase_ = _get_word_embedding_weight(lowerCAmelCase , model.get_input_embeddings() ) UpperCAmelCase_ = _get_word_embedding_weight(lowerCAmelCase , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(lowerCAmelCase ) UpperCAmelCase_ = _get_word_embedding_weight(lowerCAmelCase , model.get_input_embeddings() ) UpperCAmelCase_ = _get_word_embedding_weight(lowerCAmelCase , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. UpperCAmelCase_ = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , lowerCAmelCase ) # check that weights remain the same after resizing UpperCAmelCase_ = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: UpperCAmelCase_ = False self.assertTrue(lowerCAmelCase ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , lowerCAmelCase ) UpperCAmelCase_ = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: UpperCAmelCase_ = False self.assertTrue(lowerCAmelCase ) def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> Optional[int]: return tf.constant(__SCREAMING_SNAKE_CASE , dtype=tf.intaa ) @require_tf class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : List[Any] = 99 def A__ ( self ): UpperCAmelCase_ = tf.ones((4, 1) , dtype=tf.intaa ) * 2 UpperCAmelCase_ = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) UpperCAmelCase_ = input_ids.shape[0] UpperCAmelCase_ = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def A__ ( self ): UpperCAmelCase_ = TFOPTModel.from_pretrained("facebook/opt-350m" ) UpperCAmelCase_ = _long_tensor([[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2]] ) UpperCAmelCase_ = tf.not_equal(lowerCAmelCase , model.config.pad_token_id ) with tf.GradientTape(): UpperCAmelCase_ = model(input_ids=lowerCAmelCase , attention_mask=lowerCAmelCase ).last_hidden_state UpperCAmelCase_ = (1, 11, 512) self.assertEqual(output.shape , lowerCAmelCase ) UpperCAmelCase_ = tf.constant( [[-0.2873, -1.9218, -0.3033], [-1.2710, -0.1338, -0.1902], [0.4095, 0.1214, -1.3121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase , atol=4e-3 ) ) UpperCAmelCase_ = tf.function(lowerCAmelCase , jit_compile=lowerCAmelCase ) UpperCAmelCase_ = xla_generate(lowerCAmelCase , lowerCAmelCase )[0] self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase , atol=4e-2 ) ) @require_tf @slow class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def A__ ( self ): super().setUp() UpperCAmelCase_ = "facebook/opt-350m" def A__ ( self ): UpperCAmelCase_ = TFOPTForCausalLM.from_pretrained(self.path_model ) UpperCAmelCase_ = GPTaTokenizer.from_pretrained(self.path_model ) UpperCAmelCase_ = [ "Today is a beautiful day and I want to", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False UpperCAmelCase_ = tokenizer(lowerCAmelCase , return_tensors="tf" , padding=lowerCAmelCase , add_special_tokens=lowerCAmelCase ) UpperCAmelCase_ = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) UpperCAmelCase_ = tf.constant( [ [1.3851, -13.8923, -10.5229, -10.7533, -0.2309, -10.2384, -0.5365, -9.0947, -5.1670], [-4.7073, -10.6276, -3.9415, -21.5242, -0.2822, -0.2822, -0.2822, -0.2822, -0.2822], [0.6247, -3.4229, -8.9179, -1.4297, -14.1650, 1.4146, -9.0218, -0.2703, -0.2703], [6.4783, -1.9913, -10.7926, -2.3336, 1.5092, -0.9974, -6.8213, 1.3477, 1.3477], ] ) self.assertTrue(np.allclose(lowerCAmelCase , lowerCAmelCase , atol=1e-4 ) ) UpperCAmelCase_ = tf.function(lowerCAmelCase , jit_compile=lowerCAmelCase ) UpperCAmelCase_ = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(lowerCAmelCase , lowerCAmelCase , atol=1e-4 ) ) @require_tf @slow class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @property def A__ ( self ): return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def A__ ( self ): UpperCAmelCase_ = "facebook/opt-125m" UpperCAmelCase_ = [ "Today is a beautiful day and I want to", "In the city of New York, the city", "Paris is the capital of France and the capital", "Computers and mobile phones have taken over the", ] UpperCAmelCase_ = [] UpperCAmelCase_ = GPTaTokenizer.from_pretrained(lowerCAmelCase ) UpperCAmelCase_ = TFOPTForCausalLM.from_pretrained(lowerCAmelCase ) for prompt in self.prompts: UpperCAmelCase_ = tokenizer(lowerCAmelCase , return_tensors="tf" ).input_ids UpperCAmelCase_ = model.generate(lowerCAmelCase , max_length=10 ) UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase , skip_special_tokens=lowerCAmelCase ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = "facebook/opt-350m" UpperCAmelCase_ = GPTaTokenizer.from_pretrained(lowerCAmelCase ) UpperCAmelCase_ = TFOPTForCausalLM.from_pretrained(lowerCAmelCase ) UpperCAmelCase_ = "left" # use different length sentences to test batching UpperCAmelCase_ = [ "Hello, my dog is a little", "Today, I", ] UpperCAmelCase_ = tokenizer(lowerCAmelCase , return_tensors="tf" , padding=lowerCAmelCase ) UpperCAmelCase_ = inputs["input_ids"] UpperCAmelCase_ = model.generate(input_ids=lowerCAmelCase , attention_mask=inputs["attention_mask"] ) UpperCAmelCase_ = tokenizer(sentences[0] , return_tensors="tf" ).input_ids UpperCAmelCase_ = model.generate(input_ids=lowerCAmelCase ) UpperCAmelCase_ = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs["attention_mask"][-1] , tf.intaa ) ) UpperCAmelCase_ = tokenizer(sentences[1] , return_tensors="tf" ).input_ids UpperCAmelCase_ = model.generate(input_ids=lowerCAmelCase , max_length=model.config.max_length - num_paddings ) UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase , skip_special_tokens=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCAmelCase ) UpperCAmelCase_ = [ "Hello, my dog is a little bit of a dork.\nI'm a little bit", "Today, I was in the middle of a conversation with a friend about the", ] self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) self.assertListEqual(lowerCAmelCase , [non_padded_sentence, padded_sentence] ) def A__ ( self ): UpperCAmelCase_ = "facebook/opt-350m" UpperCAmelCase_ = [ "Today is a beautiful day and I want to", "In the city of San Francisco, the city", "Paris is the capital of France and the capital", "Computers and mobile phones have taken over the", ] UpperCAmelCase_ = [] UpperCAmelCase_ = GPTaTokenizer.from_pretrained(lowerCAmelCase ) UpperCAmelCase_ = TFOPTForCausalLM.from_pretrained(lowerCAmelCase ) for prompt in self.prompts: UpperCAmelCase_ = tokenizer(lowerCAmelCase , return_tensors="tf" ).input_ids UpperCAmelCase_ = model.generate(lowerCAmelCase , max_length=10 ) UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase , skip_special_tokens=lowerCAmelCase ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase , lowerCAmelCase )
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_xlnet": ["XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["XLNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["XLNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "XLNetForMultipleChoice", "XLNetForQuestionAnswering", "XLNetForQuestionAnsweringSimple", "XLNetForSequenceClassification", "XLNetForTokenClassification", "XLNetLMHeadModel", "XLNetModel", "XLNetPreTrainedModel", "load_tf_weights_in_xlnet", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLNetForMultipleChoice", "TFXLNetForQuestionAnsweringSimple", "TFXLNetForSequenceClassification", "TFXLNetForTokenClassification", "TFXLNetLMHeadModel", "TFXLNetMainLayer", "TFXLNetModel", "TFXLNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def A__ ( self ): debug_launcher(test_script.main ) def A__ ( self ): debug_launcher(test_ops.main )
23
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[Any]: # Initialise PyTorch model UpperCAmelCase_ = MobileBertConfig.from_json_file(__SCREAMING_SNAKE_CASE ) print(f'''Building PyTorch model from configuration: {config}''' ) UpperCAmelCase_ = MobileBertForPreTraining(__SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint UpperCAmelCase_ = load_tf_weights_in_mobilebert(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , __SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--mobilebert_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained MobileBERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
23
1
import itertools import string from collections.abc import Generator, Iterable def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Generator[tuple[str, ...], None, None]: UpperCAmelCase_ = iter(__SCREAMING_SNAKE_CASE ) while True: UpperCAmelCase_ = tuple(itertools.islice(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) if not chunk: return yield chunk def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = "".join([c.upper() for c in dirty if c in string.ascii_letters] ) UpperCAmelCase_ = "" if len(__SCREAMING_SNAKE_CASE ) < 2: return dirty for i in range(len(__SCREAMING_SNAKE_CASE ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(__SCREAMING_SNAKE_CASE ) & 1: clean += "X" return clean def snake_case__ ( __SCREAMING_SNAKE_CASE ) -> list[str]: # I and J are used interchangeably to allow # us to use a 5x5 table (25 letters) UpperCAmelCase_ = "ABCDEFGHIKLMNOPQRSTUVWXYZ" # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler UpperCAmelCase_ = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(__SCREAMING_SNAKE_CASE ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(__SCREAMING_SNAKE_CASE ) return table def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = generate_table(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = prepare_input(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = "" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(__SCREAMING_SNAKE_CASE , 2 ): UpperCAmelCase_ , UpperCAmelCase_ = divmod(table.index(__SCREAMING_SNAKE_CASE ) , 5 ) UpperCAmelCase_ , UpperCAmelCase_ = divmod(table.index(__SCREAMING_SNAKE_CASE ) , 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ = generate_table(__SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = "" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(__SCREAMING_SNAKE_CASE , 2 ): UpperCAmelCase_ , UpperCAmelCase_ = divmod(table.index(__SCREAMING_SNAKE_CASE ) , 5 ) UpperCAmelCase_ , UpperCAmelCase_ = divmod(table.index(__SCREAMING_SNAKE_CASE ) , 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
23
import heapq as hq import math from collections.abc import Iterator class lowerCamelCase : '''simple docstring''' def __init__( self , lowerCAmelCase ): UpperCAmelCase_ = str(id_ ) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = [] UpperCAmelCase_ = {} # {vertex:distance} def __lt__( self , lowerCAmelCase ): return self.key < other.key def __repr__( self ): return self.id def A__ ( self , lowerCAmelCase ): self.neighbors.append(lowerCAmelCase ) def A__ ( self , lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = weight def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __SCREAMING_SNAKE_CASE ) graph[b - 1].add_edge(graph[a - 1] , __SCREAMING_SNAKE_CASE ) def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> list: UpperCAmelCase_ = [] for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = graph[:] while q: UpperCAmelCase_ = min(__SCREAMING_SNAKE_CASE ) q.remove(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def snake_case__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Iterator[tuple]: for u in graph: UpperCAmelCase_ = math.inf UpperCAmelCase_ = None UpperCAmelCase_ = 0 UpperCAmelCase_ = list(__SCREAMING_SNAKE_CASE ) hq.heapify(__SCREAMING_SNAKE_CASE ) while h: UpperCAmelCase_ = hq.heappop(__SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): UpperCAmelCase_ = u UpperCAmelCase_ = u.edges[v.id] hq.heapify(__SCREAMING_SNAKE_CASE ) for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def snake_case__ ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
23
1
import gc import math import unittest import torch from diffusers import UNetaDModel from diffusers.utils import floats_tensor, logging, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin SCREAMING_SNAKE_CASE = logging.get_logger(__name__) enable_full_determinism() class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = UNetaDModel lowerCAmelCase_ : str = 'sample' @property def A__ ( self ): UpperCAmelCase_ = 4 UpperCAmelCase_ = 3 UpperCAmelCase_ = (32, 32) UpperCAmelCase_ = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCAmelCase ) UpperCAmelCase_ = torch.tensor([10] ).to(lowerCAmelCase ) return {"sample": noise, "timestep": time_step} @property def A__ ( self ): return (3, 32, 32) @property def A__ ( self ): return (3, 32, 32) def A__ ( self ): UpperCAmelCase_ = { "block_out_channels": (32, 64), "down_block_types": ("DownBlock2D", "AttnDownBlock2D"), "up_block_types": ("AttnUpBlock2D", "UpBlock2D"), "attention_head_dim": 3, "out_channels": 3, "in_channels": 3, "layers_per_block": 2, "sample_size": 32, } UpperCAmelCase_ = self.dummy_input return init_dict, inputs_dict class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : str = UNetaDModel lowerCAmelCase_ : Tuple = 'sample' @property def A__ ( self ): UpperCAmelCase_ = 4 UpperCAmelCase_ = 4 UpperCAmelCase_ = (32, 32) UpperCAmelCase_ = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCAmelCase ) UpperCAmelCase_ = torch.tensor([10] ).to(lowerCAmelCase ) return {"sample": noise, "timestep": time_step} @property def A__ ( self ): return (4, 32, 32) @property def A__ ( self ): return (4, 32, 32) def A__ ( self ): UpperCAmelCase_ = { "sample_size": 32, "in_channels": 4, "out_channels": 4, "layers_per_block": 2, "block_out_channels": (32, 64), "attention_head_dim": 32, "down_block_types": ("DownBlock2D", "DownBlock2D"), "up_block_types": ("UpBlock2D", "UpBlock2D"), } UpperCAmelCase_ = self.dummy_input return init_dict, inputs_dict def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" , output_loading_info=lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCAmelCase ) UpperCAmelCase_ = model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != "cuda" , "This test is supposed to run on GPU" ) def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" , output_loading_info=lowerCAmelCase ) model.to(lowerCAmelCase ) UpperCAmelCase_ = model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != "cuda" , "This test is supposed to run on GPU" ) def A__ ( self ): # by defautl model loading will use accelerate as `low_cpu_mem_usage=True` UpperCAmelCase_ , UpperCAmelCase_ = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" , output_loading_info=lowerCAmelCase ) model_accelerate.to(lowerCAmelCase ) model_accelerate.eval() UpperCAmelCase_ = torch.randn( 1 , model_accelerate.config.in_channels , model_accelerate.config.sample_size , model_accelerate.config.sample_size , generator=torch.manual_seed(0 ) , ) UpperCAmelCase_ = noise.to(lowerCAmelCase ) UpperCAmelCase_ = torch.tensor([10] * noise.shape[0] ).to(lowerCAmelCase ) UpperCAmelCase_ = model_accelerate(lowerCAmelCase , lowerCAmelCase )["sample"] # two models don't need to stay in the device at the same time del model_accelerate torch.cuda.empty_cache() gc.collect() UpperCAmelCase_ , UpperCAmelCase_ = UNetaDModel.from_pretrained( "fusing/unet-ldm-dummy-update" , output_loading_info=lowerCAmelCase , low_cpu_mem_usage=lowerCAmelCase ) model_normal_load.to(lowerCAmelCase ) model_normal_load.eval() UpperCAmelCase_ = model_normal_load(lowerCAmelCase , lowerCAmelCase )["sample"] assert torch_all_close(lowerCAmelCase , lowerCAmelCase , rtol=1e-3 ) def A__ ( self ): UpperCAmelCase_ = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" ) model.eval() model.to(lowerCAmelCase ) UpperCAmelCase_ = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) UpperCAmelCase_ = noise.to(lowerCAmelCase ) UpperCAmelCase_ = torch.tensor([10] * noise.shape[0] ).to(lowerCAmelCase ) with torch.no_grad(): UpperCAmelCase_ = model(lowerCAmelCase , lowerCAmelCase ).sample UpperCAmelCase_ = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off UpperCAmelCase_ = torch.tensor([-13.3258, -20.1100, -15.9873, -17.6617, -23.0596, -17.9419, -13.3675, -16.1889, -12.3800] ) # fmt: on self.assertTrue(torch_all_close(lowerCAmelCase , lowerCAmelCase , rtol=1e-3 ) ) class lowerCamelCase ( lowercase__, lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Optional[int] = UNetaDModel lowerCAmelCase_ : str = 'sample' @property def A__ ( self , lowerCAmelCase=(32, 32) ): UpperCAmelCase_ = 4 UpperCAmelCase_ = 3 UpperCAmelCase_ = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCAmelCase ) UpperCAmelCase_ = torch.tensor(batch_size * [10] ).to(dtype=torch.intaa , device=lowerCAmelCase ) return {"sample": noise, "timestep": time_step} @property def A__ ( self ): return (3, 32, 32) @property def A__ ( self ): return (3, 32, 32) def A__ ( self ): UpperCAmelCase_ = { "block_out_channels": [32, 64, 64, 64], "in_channels": 3, "layers_per_block": 1, "out_channels": 3, "time_embedding_type": "fourier", "norm_eps": 1e-6, "mid_block_scale_factor": math.sqrt(2.0 ), "norm_num_groups": None, "down_block_types": [ "SkipDownBlock2D", "AttnSkipDownBlock2D", "SkipDownBlock2D", "SkipDownBlock2D", ], "up_block_types": [ "SkipUpBlock2D", "SkipUpBlock2D", "AttnSkipUpBlock2D", "SkipUpBlock2D", ], } UpperCAmelCase_ = self.dummy_input return init_dict, inputs_dict @slow def A__ ( self ): UpperCAmelCase_ , UpperCAmelCase_ = UNetaDModel.from_pretrained("google/ncsnpp-celebahq-256" , output_loading_info=lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCAmelCase ) UpperCAmelCase_ = self.dummy_input UpperCAmelCase_ = floats_tensor((4, 3) + (256, 256) ).to(lowerCAmelCase ) UpperCAmelCase_ = noise UpperCAmelCase_ = model(**lowerCAmelCase ) assert image is not None, "Make sure output is not None" @slow def A__ ( self ): UpperCAmelCase_ = UNetaDModel.from_pretrained("google/ncsnpp-celebahq-256" ) model.to(lowerCAmelCase ) UpperCAmelCase_ = 4 UpperCAmelCase_ = 3 UpperCAmelCase_ = (256, 256) UpperCAmelCase_ = torch.ones((batch_size, num_channels) + sizes ).to(lowerCAmelCase ) UpperCAmelCase_ = torch.tensor(batch_size * [1e-4] ).to(lowerCAmelCase ) with torch.no_grad(): UpperCAmelCase_ = model(lowerCAmelCase , lowerCAmelCase ).sample UpperCAmelCase_ = output[0, -3:, -3:, -1].flatten().cpu() # fmt: off UpperCAmelCase_ = torch.tensor([-4842.8691, -6499.6631, -3800.1953, -7978.2686, -10980.7129, -20028.8535, 8148.2822, 2342.2905, 567.7608] ) # fmt: on self.assertTrue(torch_all_close(lowerCAmelCase , lowerCAmelCase , rtol=1e-2 ) ) def A__ ( self ): UpperCAmelCase_ = UNetaDModel.from_pretrained("fusing/ncsnpp-ffhq-ve-dummy-update" ) model.to(lowerCAmelCase ) UpperCAmelCase_ = 4 UpperCAmelCase_ = 3 UpperCAmelCase_ = (32, 32) UpperCAmelCase_ = torch.ones((batch_size, num_channels) + sizes ).to(lowerCAmelCase ) UpperCAmelCase_ = torch.tensor(batch_size * [1e-4] ).to(lowerCAmelCase ) with torch.no_grad(): UpperCAmelCase_ = model(lowerCAmelCase , lowerCAmelCase ).sample UpperCAmelCase_ = output[0, -3:, -3:, -1].flatten().cpu() # fmt: off UpperCAmelCase_ = torch.tensor([-0.0325, -0.0900, -0.0869, -0.0332, -0.0725, -0.0270, -0.0101, 0.0227, 0.0256] ) # fmt: on self.assertTrue(torch_all_close(lowerCAmelCase , lowerCAmelCase , rtol=1e-2 ) ) def A__ ( self ): # not required for this model pass
23
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE = {"configuration_fnet": ["FNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "FNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["FNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = ["FNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "FNET_PRETRAINED_MODEL_ARCHIVE_LIST", "FNetForMaskedLM", "FNetForMultipleChoice", "FNetForNextSentencePrediction", "FNetForPreTraining", "FNetForQuestionAnswering", "FNetForSequenceClassification", "FNetForTokenClassification", "FNetLayer", "FNetModel", "FNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE = { "configuration_pegasus_x": ["PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP", "PegasusXConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE = [ "PEGASUS_X_PRETRAINED_MODEL_ARCHIVE_LIST", "PegasusXForConditionalGeneration", "PegasusXModel", "PegasusXPreTrainedModel", ] if TYPE_CHECKING: from .configuration_pegasus_x import PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP, PegasusXConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pegasus_x import ( PEGASUS_X_PRETRAINED_MODEL_ARCHIVE_LIST, PegasusXForConditionalGeneration, PegasusXModel, PegasusXPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": 512, } SCREAMING_SNAKE_CASE = { "yjernite/retribert-base-uncased": {"do_lower_case": True}, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : List[str] = VOCAB_FILES_NAMES lowerCAmelCase_ : int = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Dict = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : List[str] = RetriBertTokenizer lowerCAmelCase_ : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase="[UNK]" , lowerCAmelCase="[SEP]" , lowerCAmelCase="[PAD]" , lowerCAmelCase="[CLS]" , lowerCAmelCase="[MASK]" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , do_lower_case=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , pad_token=lowerCAmelCase , cls_token=lowerCAmelCase , mask_token=lowerCAmelCase , tokenize_chinese_chars=lowerCAmelCase , strip_accents=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(lowerCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**lowerCAmelCase ) UpperCAmelCase_ = do_lower_case def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
1
import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizerOutput from transformers.testing_utils import require_phonemizer from ...test_tokenization_common import TokenizerTesterMixin @require_phonemizer class lowerCamelCase ( lowercase__, unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ : Tuple = WavaVecaPhonemeCTCTokenizer lowerCAmelCase_ : List[Any] = False def A__ ( self ): super().setUp() UpperCAmelCase_ = ( "<s> <pad> </s> <unk> n s t ə l a i k d m ɛ ɾ e ɪ p o ɐ z ð f j v b ɹ ʁ ʊ iː r w ʌ u ɡ æ aɪ ʃ h ɔ ɑː " "ŋ ɚ eɪ β uː y ɑ̃ oʊ ᵻ eː θ aʊ ts oː ɔ̃ ɣ ɜ ɑ dʒ əl x ɜː ç ʒ tʃ ɔː ɑːɹ ɛ̃ ʎ ɔːɹ ʋ aː ɕ œ ø oːɹ ɲ yː " "ʔ iə i5 s. tɕ ?? nʲ ɛː œ̃ ɭ ɔø ʑ tʲ ɨ ɛɹ ts. rʲ ɪɹ ɭʲ i.5 ɔɪ q sʲ u5 ʊɹ iɜ a5 iɛ5 øː ʕ ja əɜ th ɑ5 " "oɪ dʲ ə5 tɕh ts.h mʲ ɯ dʑ vʲ e̞ tʃʲ ei5 o5 onɡ5 ɑu5 iɑ5 ai5 aɪɚ kh ə1 ʐ i2 ʉ ħ t[ aɪə ʲ ju ə2 u2 oɜ " "pː iɛɜ ou5 y5 uɜ tː uo5 d[ uoɜ tsh ɑɜ ɵ i̪5 uei5 ɟ aɜ ɑɨ i.ɜ eʊ o2 ɐ̃ ä pʲ kʲ n̩ ɒ ph ɑu2 uɨ əɪ ɫ ɬ " "yɜ bʲ ɑ2 s̪ aiɜ χ ɐ̃ʊ̃ 1 ə4 yæɜ a2 ɨː t̪ iouɜ ũ onɡɜ aɨ iɛ2 ɔɨ ɑuɜ o̞ ei2 iou2 c kː y2 ɖ oe dˤ yɛɜ " "əʊ S ɡʲ onɡ2 u\" eiɜ ʈ ɯᵝ iou5 dZ r̝̊ i.2 tS s^ ʝ yə5 iɑɜ uə5 pf ɨu iɑ2 ou2 ər2 fʲ ai2 r̝ uəɜ ɳ əɨ " "ua5 uɪ ɽ bː yu5 uo2 yɛ5 l̩ ɻ ərɜ ʂ i̪2 ouɜ uaɜ a. a.ː yæ5 dː r̩ ee ɪu ər5 i̪ ɜ æi u: i.ː t^ o1 ɪ^ " "ai ueiɜ æː ɛɪ eə i. ɴ ie ua2 ɑ1 o4 tʃː o: ɑ: u1 N i̪1 au yæ2 u. qː yəɜ y: kʰ tʃʰ iʊ sx õ uo tʰ " "uai5 bʰ u.ː uə2 ʊə d^ s̪ː yiɜ dʰ r. oe: i1 ɟː yu2 nʲʲ i̪4 uei2 tsʲ ɸ ĩ ɑ4 t̪ː eɑ u4 e: tsː ʈʰ ɡʰ " "ɯɯ dʒʲ ʂʲ X ɵː uaiɜ tɕʲ ã t^ː ẽː yɛ2 cː i.1 ɛʊ dˤdˤ dʒː i4 ɡː yi ɕʲ ɟʰ pʰ dʑʲ yuɜ ua1 ua4 æiː ɐɐ " "ui iou1 ʊː a1 iou4 cʰ iɛ1 yə2 ɖʰ ẽ ʒʲ ää ər4 iːː ɪː iɑ1 ər1 œː øi ɪuː cʰcʰ əː1 iː1 ũ kʰː o̞o̞ xʲ " "ou1 iɛ4 e̞e̞ y1 dzː dʲʲ dʰː ɯᵝɯᵝ lː uo1 i.4 i: yɛ5ʲ a4" ).split(" " ) UpperCAmelCase_ = dict(zip(lowerCAmelCase , range(len(lowerCAmelCase ) ) ) ) UpperCAmelCase_ = {"pad_token": "<pad>", "unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>"} UpperCAmelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase ) + "\n" ) def A__ ( self , lowerCAmelCase , lowerCAmelCase=False , lowerCAmelCase=20 , lowerCAmelCase=5 ): UpperCAmelCase_ = [(i, tokenizer.decode([i] , clean_up_tokenization_spaces=lowerCAmelCase )) for i in range(len(lowerCAmelCase ) )] UpperCAmelCase_ = list(filter(lambda lowerCAmelCase : [t[0]] == tokenizer.encode(t[1] , do_phonemize=lowerCAmelCase ) , lowerCAmelCase ) ) if max_length is not None and len(lowerCAmelCase ) > max_length: UpperCAmelCase_ = toks[:max_length] if min_length is not None and len(lowerCAmelCase ) < min_length and len(lowerCAmelCase ) > 0: while len(lowerCAmelCase ) < min_length: UpperCAmelCase_ = toks + toks # toks_str = [t[1] for t in toks] UpperCAmelCase_ = [t[0] for t in toks] # Ensure consistency UpperCAmelCase_ = tokenizer.decode(lowerCAmelCase , clean_up_tokenization_spaces=lowerCAmelCase ) if " " not in output_txt and len(lowerCAmelCase ) > 1: UpperCAmelCase_ = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=lowerCAmelCase ) + " " + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=lowerCAmelCase ) ) if with_prefix_space: UpperCAmelCase_ = " " + output_txt UpperCAmelCase_ = tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) return output_txt, output_ids def A__ ( self , **lowerCAmelCase ): kwargs.update(self.special_tokens_map ) return WavaVecaPhonemeCTCTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) # check adding a single token tokenizer.add_tokens("xxx" ) UpperCAmelCase_ = tokenizer("m xxx ɪ" , do_phonemize=lowerCAmelCase ).input_ids self.assertEqual(lowerCAmelCase , [13, 392, 17] ) # xxx should be last token tokenizer.add_tokens(["aaa", "bbb", "ccc"] ) UpperCAmelCase_ = tokenizer("m aaa ɪ ccc" , do_phonemize=lowerCAmelCase ).input_ids self.assertEqual(lowerCAmelCase , [13, 393, 17, 395] ) # aaa and ccc should be after xxx and 2 after aaa UpperCAmelCase_ = tokenizer("maɪ c" , do_phonemize=lowerCAmelCase ).input_ids self.assertEqual(lowerCAmelCase , [3, 200] ) # mai should be <unk> (=3) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) UpperCAmelCase_ = "Hello how are you" UpperCAmelCase_ = tokenizer.phonemize(lowerCAmelCase , phonemizer_lang="en-us" ) self.assertEqual(lowerCAmelCase , "h ə l oʊ h aʊ ɑːɹ j uː" ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) UpperCAmelCase_ = "Hello how are you" UpperCAmelCase_ = tokenizer.phonemize(lowerCAmelCase , phonemizer_lang="en-us" ) self.assertEqual(tokenizer(lowerCAmelCase ).input_ids , tokenizer(lowerCAmelCase , do_phonemize=lowerCAmelCase ).input_ids ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) UpperCAmelCase_ = "Hello how are you" UpperCAmelCase_ = tokenizer.phonemize(lowerCAmelCase , phonemizer_lang="en-us" ) UpperCAmelCase_ = tokenizer.decode(tokenizer(lowerCAmelCase ).input_ids ) self.assertEqual(lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) UpperCAmelCase_ = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98], [24, 22, 5, 24, 22, 5, 77], ] UpperCAmelCase_ = tokenizer.decode(sample_ids[0] ) UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase ) self.assertEqual(lowerCAmelCase , batch_tokens[0] ) self.assertEqual(lowerCAmelCase , ["k s ɾ ɾ l ɭʲ", "j ð s j ð s oːɹ"] ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft" , word_delimiter_token="|" ) tokenizer.add_tokens("|" ) UpperCAmelCase_ = "Hello how are you" UpperCAmelCase_ = tokenizer.phonemize(lowerCAmelCase , phonemizer_lang="en-us" ) self.assertEqual(lowerCAmelCase , "h ə l oʊ | h aʊ | ɑːɹ | j uː |" ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft" , word_delimiter_token="|" ) tokenizer.add_tokens("|" ) UpperCAmelCase_ = "Hello how are you" UpperCAmelCase_ = tokenizer.phonemize(lowerCAmelCase , phonemizer_lang="en-us" ) self.assertEqual(tokenizer(lowerCAmelCase ).input_ids , tokenizer(lowerCAmelCase , do_phonemize=lowerCAmelCase ).input_ids ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft" , word_delimiter_token="|" ) tokenizer.add_tokens("|" ) # fmt: off UpperCAmelCase_ = [ [11, 5, 15, tokenizer.pad_token_id, tokenizer.word_delimiter_token_id, 15, 8, tokenizer.word_delimiter_token_id, 98], [tokenizer.word_delimiter_token_id, 24, 22, tokenizer.word_delimiter_token_id, 5, 24, 22, 5, 77], ] # fmt: on # decode with word_del_token filter UpperCAmelCase_ = tokenizer.decode(sample_ids[0] ) UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase ) self.assertEqual(lowerCAmelCase , batch_tokens[0] ) self.assertEqual(lowerCAmelCase , ["k s ɾ ɾ l ɭʲ", "j ð s j ð s oːɹ"] ) # decode with no word_del_token filter UpperCAmelCase_ = tokenizer.decode(sample_ids[0] , filter_word_delimiter_token=lowerCAmelCase ) UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase , filter_word_delimiter_token=lowerCAmelCase ) self.assertEqual(lowerCAmelCase , batch_tokens[0] ) self.assertEqual(lowerCAmelCase , ["k s ɾ | ɾ l | ɭʲ", "| j ð | s j ð s oːɹ"] ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft" , word_delimiter_token="|" ) tokenizer.add_tokens("|" ) UpperCAmelCase_ = "Hello how are you" UpperCAmelCase_ = tokenizer.phonemize(lowerCAmelCase , phonemizer_lang="en-us" ) UpperCAmelCase_ = tokenizer.decode(tokenizer(lowerCAmelCase ).input_ids , filter_word_delimiter_token=lowerCAmelCase ) self.assertEqual(lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft" , word_delimiter_token="|" ) tokenizer.add_tokens("|" ) UpperCAmelCase_ = "Hello how are you" UpperCAmelCase_ = tokenizer.phonemize(lowerCAmelCase , phonemizer_lang="en-us" ) UpperCAmelCase_ = tokenizer.decode(tokenizer(lowerCAmelCase ).input_ids , filter_word_delimiter_token=lowerCAmelCase ) self.assertEqual(" ".join([p.strip() for p in phonemes.split(" |" )] ).strip() , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained( "facebook/wav2vec2-lv-60-espeak-cv-ft" , word_delimiter_token=lowerCAmelCase ) UpperCAmelCase_ = "Hello how are you" UpperCAmelCase_ = tokenizer(lowerCAmelCase , phonemizer_lang="en-us" ).input_ids UpperCAmelCase_ = tokenizer(lowerCAmelCase , phonemizer_lang="fr-fr" ).input_ids self.assertNotEqual(lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase_ = tokenizer.decode(lowerCAmelCase ) UpperCAmelCase_ = tokenizer.decode(lowerCAmelCase ) self.assertEqual(lowerCAmelCase , "h ə l oʊ h aʊ ɑːɹ j uː" ) self.assertEqual(lowerCAmelCase , "ɛ l o h aʊ a ʁ j u" ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) UpperCAmelCase_ = "Hello how Are you" UpperCAmelCase_ = "hello how are you" UpperCAmelCase_ = tokenizer(lowerCAmelCase ).input_ids UpperCAmelCase_ = tokenizer(lowerCAmelCase ).input_ids self.assertEqual(lowerCAmelCase , lowerCAmelCase ) def A__ ( self ): UpperCAmelCase_ = self.tokenizer_class.from_pretrained("facebook/wav2vec2-lv-60-espeak-cv-ft" ) tokenizer.add_tokens(["!", "?"] ) tokenizer.add_special_tokens({"cls_token": "$$$"} ) # fmt: off UpperCAmelCase_ = [ [11, 5, 15, tokenizer.pad_token_id, 15, 8, 98, 392, 392, 393, 392, 392, 393, 394, 394], [24, 22, 5, 24, 22, 5, 77, tokenizer.pad_token_id, 394, 394], ] # fmt: on UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase ) self.assertEqual(lowerCAmelCase , ["k s ɾ ɾ l ɭʲ!?!? $$$", "j ð s j ð s oːɹ $$$"] ) @staticmethod def A__ ( lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase_ = [d[key] for d in offsets] return retrieved_list def A__ ( self ): UpperCAmelCase_ = self.get_tokenizer(word_delimiter_token="|" ) tokenizer.add_tokens("|" ) # fmt: off # ksssɾɾ|ɾɾ<pad>ɾɾ|<pad>ɾlll|ɭʲ -> k s ɾ ɾ | ɾ l | ɭʲ" UpperCAmelCase_ = [11, 5, 5, 5, 15, 15, tokenizer.pad_token_id, 15, 15, tokenizer.word_delimiter_token_id, tokenizer.pad_token_id, 15, 8, 8, 8, tokenizer.word_delimiter_token_id, 98] # fmt: on UpperCAmelCase_ = tokenizer.decode(lowerCAmelCase , output_char_offsets=lowerCAmelCase , filter_word_delimiter_token=lowerCAmelCase ) # check Wav2Vec2CTCTokenizerOutput keys for char self.assertEqual(len(outputs.keys() ) , 2 ) self.assertTrue("text" in outputs ) self.assertTrue("char_offsets" in outputs ) self.assertTrue(isinstance(lowerCAmelCase , lowerCAmelCase ) ) # check that order of chars is correct and identical for both outputs self.assertEqual(" ".join(self.get_from_offsets(outputs["char_offsets"] , "char" ) ) , outputs.text ) self.assertListEqual( self.get_from_offsets(outputs["char_offsets"] , "char" ) , ["k", "s", "ɾ", "ɾ", "|", "ɾ", "l", "|", "ɭʲ"] ) # check that offsets are actually correct for char # 0-1 is 11, 1-4 is 5, 4-6 is first 15, 6-7 is <pad> (thus not shown), 7-9 is second 15, 9-10 is word_delimiter_token, # 10-11 is <pad> (thus not shown), 11-12 is third 15, 12-15 is 8, 15-16 is word_delimiter_token, 16-17 is 98 self.assertListEqual( self.get_from_offsets(outputs["char_offsets"] , "start_offset" ) , [0, 1, 4, 7, 9, 11, 12, 15, 16] ) self.assertListEqual( self.get_from_offsets(outputs["char_offsets"] , "end_offset" ) , [1, 4, 6, 9, 10, 12, 15, 16, 17] ) def A__ ( self ): UpperCAmelCase_ = self.get_tokenizer(word_delimiter_token="|" ) def check_list_tuples_equal(lowerCAmelCase , lowerCAmelCase ): self.assertTrue(isinstance(lowerCAmelCase , lowerCAmelCase ) ) self.assertTrue(isinstance(outputs_list[0] , lowerCAmelCase ) ) # transform list to ModelOutput UpperCAmelCase_ = WavaVecaPhonemeCTCTokenizerOutput( {k: [d[k] for d in outputs_list] for k in outputs_list[0]} ) self.assertListEqual(outputs_batch["text"] , outputs_batch_a["text"] ) def recursive_check(lowerCAmelCase , lowerCAmelCase ): if isinstance(lowerCAmelCase , lowerCAmelCase ): [recursive_check(lowerCAmelCase , lowerCAmelCase ) for la, la in zip(lowerCAmelCase , lowerCAmelCase )] self.assertEqual(lowerCAmelCase , lowerCAmelCase ) if "char_offsets" in outputs_batch: recursive_check(outputs_batch["char_offsets"] , outputs_batch_a["char_offsets"] ) # fmt: off UpperCAmelCase_ = [ [11, 5, 15, tokenizer.pad_token_id, 15, 4, 8, 98, 32, 32, 32, 32, 4, 33, tokenizer.word_delimiter_token_id, 32, 32, 33, 34, 34], [24, 22, 5, tokenizer.word_delimiter_token_id, tokenizer.word_delimiter_token_id, 24, 22, 22, 22, 4, 5, 77, tokenizer.pad_token_id, 22, 22, 4, 34, 34, 34, 34], ] # fmt: on # We assume that `decode` works as expected. All we will check now is # the output type is correct and the output is identical to `decode` # char UpperCAmelCase_ = tokenizer.batch_decode(lowerCAmelCase , output_char_offsets=lowerCAmelCase ) UpperCAmelCase_ = [tokenizer.decode(lowerCAmelCase , output_char_offsets=lowerCAmelCase ) for ids in sample_ids] check_list_tuples_equal(lowerCAmelCase , lowerCAmelCase ) @unittest.skip("Wav2Vec2PhonemeTokenizer always lower cases letters to correctly map to phonemes" ) def A__ ( self ): pass @unittest.skip("Wav2Vec2PhonemeTokenizer always puts spaces between phonemes" ) def A__ ( self ): pass @unittest.skip("encodes to text to ids, but decodes ids to phonemes -> not possible to have internal consistency" ) def A__ ( self ): pass @unittest.skip("Wav2Vec2PhonemeModel has no max model length => no testing" ) def A__ ( self ): pass def A__ ( self ): UpperCAmelCase_ = self.get_tokenizers(do_lower_case=lowerCAmelCase ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): UpperCAmelCase_ = tokenizer.vocab_size UpperCAmelCase_ = len(lowerCAmelCase ) self.assertNotEqual(lowerCAmelCase , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) UpperCAmelCase_ = ["aaaaa bbbbbb", "cccccccccdddddddd"] UpperCAmelCase_ = tokenizer.add_tokens(lowerCAmelCase ) UpperCAmelCase_ = tokenizer.vocab_size UpperCAmelCase_ = len(lowerCAmelCase ) self.assertNotEqual(lowerCAmelCase , 0 ) self.assertEqual(lowerCAmelCase , lowerCAmelCase ) self.assertEqual(lowerCAmelCase , len(lowerCAmelCase ) ) self.assertEqual(lowerCAmelCase , all_size + len(lowerCAmelCase ) ) UpperCAmelCase_ = tokenizer.encode("aaaaa bbbbbb low cccccccccdddddddd l" , add_special_tokens=lowerCAmelCase ) self.assertGreaterEqual(len(lowerCAmelCase ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) UpperCAmelCase_ = {"eos_token": ">>>>|||<||<<|<<", "pad_token": "<<<<<|||>|>>>>|>"} UpperCAmelCase_ = tokenizer.add_special_tokens(lowerCAmelCase ) UpperCAmelCase_ = tokenizer.vocab_size UpperCAmelCase_ = len(lowerCAmelCase ) self.assertNotEqual(lowerCAmelCase , 0 ) self.assertEqual(lowerCAmelCase , lowerCAmelCase ) self.assertEqual(lowerCAmelCase , len(lowerCAmelCase ) ) self.assertEqual(lowerCAmelCase , all_size_a + len(lowerCAmelCase ) ) UpperCAmelCase_ = tokenizer.encode( ">>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l" , add_special_tokens=lowerCAmelCase ) self.assertGreaterEqual(len(lowerCAmelCase ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) @unittest.skip("The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode." ) def A__ ( self ): pass @unittest.skip("The tokenizer shouldn't be used to encode input IDs (except for labels), only to decode." ) def A__ ( self ): pass def A__ ( self ): # The default common tokenizer tests assumes that the output of `convert_tokens_to_string` is a string which # is not the case for Wav2Vec2PhonemeCTCTokenizer. UpperCAmelCase_ = self.get_tokenizers(fast=lowerCAmelCase , do_lower_case=lowerCAmelCase ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): UpperCAmelCase_ = ["ð", "ɪ", "s", "ɪ", "z", "ɐ", "t", "ɛ", "k", "s", "t"] UpperCAmelCase_ = tokenizer.convert_tokens_to_string(lowerCAmelCase ) self.assertIsInstance(output["text"] , lowerCAmelCase )
23
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} SCREAMING_SNAKE_CASE = { "vocab_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-german-cased": ( "https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json" ), "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json" ), }, } SCREAMING_SNAKE_CASE = { "distilbert-base-uncased": 512, "distilbert-base-uncased-distilled-squad": 512, "distilbert-base-cased": 512, "distilbert-base-cased-distilled-squad": 512, "distilbert-base-german-cased": 512, "distilbert-base-multilingual-cased": 512, } SCREAMING_SNAKE_CASE = { "distilbert-base-uncased": {"do_lower_case": True}, "distilbert-base-uncased-distilled-squad": {"do_lower_case": True}, "distilbert-base-cased": {"do_lower_case": False}, "distilbert-base-cased-distilled-squad": {"do_lower_case": False}, "distilbert-base-german-cased": {"do_lower_case": False}, "distilbert-base-multilingual-cased": {"do_lower_case": False}, } class lowerCamelCase ( lowercase__ ): '''simple docstring''' lowerCAmelCase_ : Any = VOCAB_FILES_NAMES lowerCAmelCase_ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Union[str, Any] = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : int = ['input_ids', 'attention_mask'] lowerCAmelCase_ : str = DistilBertTokenizer def __init__( self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase="[UNK]" , lowerCAmelCase="[SEP]" , lowerCAmelCase="[PAD]" , lowerCAmelCase="[CLS]" , lowerCAmelCase="[MASK]" , lowerCAmelCase=True , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , do_lower_case=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , pad_token=lowerCAmelCase , cls_token=lowerCAmelCase , mask_token=lowerCAmelCase , tokenize_chinese_chars=lowerCAmelCase , strip_accents=lowerCAmelCase , **lowerCAmelCase , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase_ = getattr(lowerCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase_ = do_lower_case UpperCAmelCase_ = strip_accents UpperCAmelCase_ = tokenize_chinese_chars UpperCAmelCase_ = normalizer_class(**lowerCAmelCase ) UpperCAmelCase_ = do_lower_case def A__ ( self , lowerCAmelCase , lowerCAmelCase=None ): UpperCAmelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCAmelCase , lowerCAmelCase = None ): UpperCAmelCase_ = self._tokenizer.model.save(lowerCAmelCase , name=lowerCAmelCase ) return tuple(lowerCAmelCase )
23
1