Searched refs:self (Results 1 – 7 of 7) sorted by relevance
| /honee/tests/junit-xml/junit_xml/ |
| H A D | __init__.py | 84 self, argument 98 self.name = name 105 self.test_cases = test_cases 106 self.timestamp = timestamp 107 self.hostname = hostname 108 self.id = id 109 self.package = package 110 self.file = file 111 self.log = log 112 self.url = url [all …]
|
| /honee/tests/ |
| H A D | smartsim_regression_framework.py | 88 def __init__(self, directory_path: Path): argument 89 self.exp: Experiment 90 self.database = None 91 self.directory_path: Path = directory_path 92 self.original_path: Path 94 def setup(self): argument 96 self.original_path = Path(os.getcwd()) 98 if self.directory_path.exists() and self.directory_path.is_dir(): 99 shutil.rmtree(self.directory_path) 100 self.directory_path.mkdir() [all …]
|
| H A D | junit_common.py | 26 def __init__(self, message): argument 33 def __init__(self, option_strings, dest, type, default, **kwargs): argument 37 self.enum_type = type 38 if isinstance(default, self.enum_type): 41 default = self.enum_type(default.lower()) 43 default = [self.enum_type(v.lower()) for v in default] 49 def __call__(self, parser, namespace, values, option_string=None): argument 50 if isinstance(values, self.enum_type): 53 values = self.enum_type(values.lower()) 55 values = [self.enum_type(v.lower()) for v in values] [all …]
|
| H A D | junit.py | 51 def __init__(self, has_torch: bool): argument 52 self.has_torch: bool = has_torch 53 self.csv_rtol = 1e-9 54 self.csv_comment_diff_fn = diff_csv_comment_function 55 self.csv_comment_str = '#' 57 def get_source_path(self, test: str) -> Path: argument 72 def get_run_path(self, test: str) -> Path: argument 83 def get_output_path(self, test: str, output_file: str) -> Path: argument 95 def check_pre_skip(self, test: str, spec: TestSpec, resource: str, nproc: int) -> Optional[str]: argument 110 if condition == 'torch' and not self.has_torch: [all …]
|
| /honee/tests/createPyTorchModel/ |
| H A D | update_weights.py | 22 def __init__(self, inputDim=6, outputDim=6, numNeurons=20, numLayers=1): argument 24 self.ndIn = inputDim 25 self.ndOut = outputDim 26 self.nNeurons = numNeurons 27 self.nLayers = numLayers 28 self.net = nn.Sequential( 29 nn.Linear(self.ndIn, self.nNeurons), 31 nn.Linear(self.nNeurons, self.ndOut)) 34 def forward(self, x): argument 35 return self.net(x)
|
| /honee/ |
| H A D | Makefile | 309 CEED_BACKENDS ?= /cpu/self
|
| /honee/doc/ |
| H A D | runtime_options.md | 16 - `/cpu/self/opt/blocked`
|