/dports/www/qt5-webengine/qtwebengine-everywhere-src-5.15.2/src/3rdparty/chromium/third_party/webrtc/modules/audio_processing/test/py_quality_assessment/quality_assessment/ |
H A D | export.py | 121 for tab_index, score_name in enumerate(score_names): 126 self._FormatName(score_name))) 133 for tab_index, score_name in enumerate(score_names): 155 def _BuildScoreTab(self, score_name, anchor_data): argument 159 self._scores_data_frame.eval_score_name == score_name] 185 score_name, apm_config[0], test_data_gen_info[0], 198 score_name, apm_configs, test_data_gen_configs, 203 def _BuildScoreTableCell(self, score_name, test_data_gen, argument 239 score_name, apm_config[0], 253 score_name, apm_config[0], test_data_gen_info[0], [all …]
|
H A D | collect_data.py | 110 test_data_gen_name, score_name, args): argument 134 (score_name, args.eval_scores), 167 score_name) = _GetScoreDescriptors(score_filepath) 176 score_name, 185 score_name) 208 score_name,
|
/dports/net-im/tg_owt/tg_owt-d578c76/src/modules/audio_processing/test/py_quality_assessment/quality_assessment/ |
H A D | export.py | 125 for tab_index, score_name in enumerate(score_names): 130 self._FormatName(score_name))) 138 for tab_index, score_name in enumerate(score_names): 161 def _BuildScoreTab(self, score_name, anchor_data): argument 165 self._scores_data_frame.eval_score_name == score_name] 193 score_name, apm_config[0], test_data_gen_info[0], 199 self._BuildScoreTableCell(score_name, 216 def _BuildScoreTableCell(self, score_name, test_data_gen, argument 350 def _SliceDataForScoreTableCell(self, score_name, apm_config, argument 397 def _ScoreStatsInspectorDialogId(cls, score_name, apm_config, argument [all …]
|
H A D | collect_data.py | 126 test_data_gen_name, score_name, args): argument 150 (score_name, args.eval_scores), 179 score_name) = _GetScoreDescriptors(score_filepath) 183 echo_simulator_name, test_data_gen_name, score_name, 187 test_data_gen_name, score_name) 210 score_name,
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/webrtc/modules/audio_processing/test/py_quality_assessment/quality_assessment/ |
H A D | export.py | 125 for tab_index, score_name in enumerate(score_names): 130 self._FormatName(score_name))) 138 for tab_index, score_name in enumerate(score_names): 161 def _BuildScoreTab(self, score_name, anchor_data): argument 165 self._scores_data_frame.eval_score_name == score_name] 193 score_name, apm_config[0], test_data_gen_info[0], 199 self._BuildScoreTableCell(score_name, 216 def _BuildScoreTableCell(self, score_name, test_data_gen, argument 350 def _SliceDataForScoreTableCell(self, score_name, apm_config, argument 397 def _ScoreStatsInspectorDialogId(cls, score_name, apm_config, argument [all …]
|
H A D | collect_data.py | 126 test_data_gen_name, score_name, args): argument 150 (score_name, args.eval_scores), 179 score_name) = _GetScoreDescriptors(score_filepath) 183 echo_simulator_name, test_data_gen_name, score_name, 187 test_data_gen_name, score_name) 210 score_name,
|
/dports/www/firefox/firefox-99.0/third_party/libwebrtc/modules/audio_processing/test/py_quality_assessment/quality_assessment/ |
H A D | export.py | 121 for tab_index, score_name in enumerate(score_names): 126 self._FormatName(score_name))) 133 for tab_index, score_name in enumerate(score_names): 155 def _BuildScoreTab(self, score_name, anchor_data): argument 159 self._scores_data_frame.eval_score_name == score_name] 185 score_name, apm_config[0], test_data_gen_info[0], 198 score_name, apm_configs, test_data_gen_configs, 203 def _BuildScoreTableCell(self, score_name, test_data_gen, argument 239 score_name, apm_config[0], 253 score_name, apm_config[0], test_data_gen_info[0], [all …]
|
H A D | collect_data.py | 110 test_data_gen_name, score_name, args): argument 134 (score_name, args.eval_scores), 167 score_name) = _GetScoreDescriptors(score_filepath) 176 score_name, 185 score_name) 208 score_name,
|
/dports/mail/thunderbird/thunderbird-91.8.0/third_party/libwebrtc/webrtc/modules/audio_processing/test/py_quality_assessment/quality_assessment/ |
H A D | export.py | 121 for tab_index, score_name in enumerate(score_names): 126 self._FormatName(score_name))) 133 for tab_index, score_name in enumerate(score_names): 155 def _BuildScoreTab(self, score_name, anchor_data): argument 159 self._scores_data_frame.eval_score_name == score_name] 185 score_name, apm_config[0], test_data_gen_info[0], 198 score_name, apm_configs, test_data_gen_configs, 203 def _BuildScoreTableCell(self, score_name, test_data_gen, argument 239 score_name, apm_config[0], 253 score_name, apm_config[0], test_data_gen_info[0], [all …]
|
H A D | collect_data.py | 110 test_data_gen_name, score_name, args): argument 134 (score_name, args.eval_scores), 167 score_name) = _GetScoreDescriptors(score_filepath) 176 score_name, 185 score_name) 208 score_name,
|
/dports/www/firefox-esr/firefox-91.8.0/third_party/libwebrtc/webrtc/modules/audio_processing/test/py_quality_assessment/quality_assessment/ |
H A D | export.py | 121 for tab_index, score_name in enumerate(score_names): 126 self._FormatName(score_name))) 133 for tab_index, score_name in enumerate(score_names): 155 def _BuildScoreTab(self, score_name, anchor_data): argument 159 self._scores_data_frame.eval_score_name == score_name] 185 score_name, apm_config[0], test_data_gen_info[0], 198 score_name, apm_configs, test_data_gen_configs, 203 def _BuildScoreTableCell(self, score_name, test_data_gen, argument 239 score_name, apm_config[0], 253 score_name, apm_config[0], test_data_gen_info[0], [all …]
|
H A D | collect_data.py | 110 test_data_gen_name, score_name, args): argument 134 (score_name, args.eval_scores), 167 score_name) = _GetScoreDescriptors(score_filepath) 176 score_name, 185 score_name) 208 score_name,
|
/dports/lang/spidermonkey78/firefox-78.9.0/media/webrtc/trunk/webrtc/modules/audio_processing/test/py_quality_assessment/quality_assessment/ |
H A D | export.py | 121 for tab_index, score_name in enumerate(score_names): 126 self._FormatName(score_name))) 133 for tab_index, score_name in enumerate(score_names): 155 def _BuildScoreTab(self, score_name, anchor_data): argument 159 self._scores_data_frame.eval_score_name == score_name] 185 score_name, apm_config[0], test_data_gen_info[0], 198 score_name, apm_configs, test_data_gen_configs, 203 def _BuildScoreTableCell(self, score_name, test_data_gen, argument 239 score_name, apm_config[0], 253 score_name, apm_config[0], test_data_gen_info[0], [all …]
|
H A D | collect_data.py | 110 test_data_gen_name, score_name, args): argument 134 (score_name, args.eval_scores), 167 score_name) = _GetScoreDescriptors(score_filepath) 176 score_name, 185 score_name) 208 score_name,
|
/dports/www/qt5-webengine/qtwebengine-everywhere-src-5.15.2/src/3rdparty/chromium/third_party/webrtc/modules/audio_processing/test/py_quality_assessment/ |
H A D | apm_quality_assessment_optimize.py | 81 for score_name in score_names: 82 scores = data_frame[data_frame.eval_score_name == score_name].score 83 normalization_constants[score_name] = max(scores) 94 for score_name in score_names: 96 score_name].score 97 scores[score_name] = sum(data_cell_scores) / len(data_cell_scores) 98 scores[score_name] /= normalization_constants[score_name]
|
/dports/net-im/tg_owt/tg_owt-d578c76/src/modules/audio_processing/test/py_quality_assessment/ |
H A D | apm_quality_assessment_optimize.py | 92 for score_name in score_names: 93 scores = data_frame[data_frame.eval_score_name == score_name].score 94 normalization_constants[score_name] = max(scores) 105 for score_name in score_names: 107 score_name].score 108 scores[score_name] = sum(data_cell_scores) / len(data_cell_scores) 109 scores[score_name] /= normalization_constants[score_name]
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/webrtc/modules/audio_processing/test/py_quality_assessment/ |
H A D | apm_quality_assessment_optimize.py | 92 for score_name in score_names: 93 scores = data_frame[data_frame.eval_score_name == score_name].score 94 normalization_constants[score_name] = max(scores) 105 for score_name in score_names: 107 score_name].score 108 scores[score_name] = sum(data_cell_scores) / len(data_cell_scores) 109 scores[score_name] /= normalization_constants[score_name]
|
/dports/www/firefox/firefox-99.0/third_party/libwebrtc/modules/audio_processing/test/py_quality_assessment/ |
H A D | apm_quality_assessment_optimize.py | 81 for score_name in score_names: 82 scores = data_frame[data_frame.eval_score_name == score_name].score 83 normalization_constants[score_name] = max(scores) 94 for score_name in score_names: 96 score_name].score 97 scores[score_name] = sum(data_cell_scores) / len(data_cell_scores) 98 scores[score_name] /= normalization_constants[score_name]
|
/dports/mail/thunderbird/thunderbird-91.8.0/third_party/libwebrtc/webrtc/modules/audio_processing/test/py_quality_assessment/ |
H A D | apm_quality_assessment_optimize.py | 81 for score_name in score_names: 82 scores = data_frame[data_frame.eval_score_name == score_name].score 83 normalization_constants[score_name] = max(scores) 94 for score_name in score_names: 96 score_name].score 97 scores[score_name] = sum(data_cell_scores) / len(data_cell_scores) 98 scores[score_name] /= normalization_constants[score_name]
|
/dports/www/firefox-esr/firefox-91.8.0/third_party/libwebrtc/webrtc/modules/audio_processing/test/py_quality_assessment/ |
H A D | apm_quality_assessment_optimize.py | 81 for score_name in score_names: 82 scores = data_frame[data_frame.eval_score_name == score_name].score 83 normalization_constants[score_name] = max(scores) 94 for score_name in score_names: 96 score_name].score 97 scores[score_name] = sum(data_cell_scores) / len(data_cell_scores) 98 scores[score_name] /= normalization_constants[score_name]
|
/dports/lang/spidermonkey78/firefox-78.9.0/media/webrtc/trunk/webrtc/modules/audio_processing/test/py_quality_assessment/ |
H A D | apm_quality_assessment_optimize.py | 81 for score_name in score_names: 82 scores = data_frame[data_frame.eval_score_name == score_name].score 83 normalization_constants[score_name] = max(scores) 94 for score_name in score_names: 96 score_name].score 97 scores[score_name] = sum(data_cell_scores) / len(data_cell_scores) 98 scores[score_name] /= normalization_constants[score_name]
|
/dports/biology/ncbi-cxx-toolkit/ncbi_cxx--25_2_0/include/algo/align/util/ |
H A D | score_lookup.hpp | 91 string HelpText(const string &score_name); 93 IScore::EComplexity Complexity(const string &score_name); 96 const string &score_name); 100 const string &score_name); 108 const string &score_name);
|
/dports/games/xbomb/xbomb-2.2b/ |
H A D | hiscore.c | 73 struct score_name struct 84 static void decrypt_score(struct score_name* sn); argument 85 static void encrypt_score(struct score_name* sn); 222 struct score_name sn; in load_high_scores() 263 struct score_name sn; in save_high_scores() 295 static void decrypt_score(struct score_name* sn) in decrypt_score() 300 for(i=0;i<sizeof(struct score_name);i++) in decrypt_score() 313 static void encrypt_score(struct score_name* sn) in encrypt_score() 318 for(i=0;i<sizeof(struct score_name);i++) in encrypt_score()
|
/dports/science/py-scikit-learn/scikit-learn-1.0.2/examples/calibration/ |
H A D | plot_calibration_curve.py | 160 score_name = metric.__name__.replace("_", " ").replace("score", "").capitalize() variable 161 scores[score_name].append(metric(y_test, y_prob[:, 1])) 164 score_name = metric.__name__.replace("_", " ").replace("score", "").capitalize() variable 165 scores[score_name].append(metric(y_test, y_pred)) 305 score_name = metric.__name__.replace("_", " ").replace("score", "").capitalize() variable 306 scores[score_name].append(metric(y_test, y_prob[:, 1])) 309 score_name = metric.__name__.replace("_", " ").replace("score", "").capitalize() variable 310 scores[score_name].append(metric(y_test, y_pred))
|
/dports/games/neverball/neverball-1.6.0/ball/ |
H A D | util.c | 57 static int score_name[4]; variable 82 score_name[j] = gui_label(ld, s, GUI_SML, gui_yel, gui_wht); in gui_scores() 85 gui_set_trunc(score_name[j], TRUNC_TAIL); in gui_scores() 86 gui_set_fill (score_name[j]); in gui_scores() 101 score_name[j] = gui_label(kd, s, GUI_SML, gui_yel, gui_wht); in gui_scores() 104 gui_set_trunc(score_name[j], TRUNC_TAIL); in gui_scores() 105 gui_set_fill (score_name[j]); in gui_scores() 127 gui_set_label(score_name[j], ""); in gui_set_scores() 140 set_score_color(score_name[j], j, gui_grn, gui_red); in gui_set_scores() 142 gui_set_color(score_name[j], gui_yel, gui_wht); in gui_set_scores() [all …]
|