physbo.search.discrete package
Module contents
- class physbo.search.discrete.History[source]
Bases:
object- export_all_sequence_best_fx()[source]
- Export all fx and actions at each sequence.
(The total number of data is total_num_research.)
- Returns:
best_fx (numpy.ndarray)
best_actions (numpy.ndarray)
- export_sequence_best_fx()[source]
Export fx and actions at each sequence. (The total number of data is num_runs.)
- Returns:
best_fx (numpy.ndarray)
best_actions (numpy.ndarray)
- load(filename)[source]
Load the information of the history.
- Parameters:
filename (str) – The name of the file which stores the information of the history
- save(filename)[source]
Save the information of the history.
- Parameters:
filename (str) – The name of the file which stores the information of the history
- property time_get_action
- property time_run_simulator
- property time_total
- property time_update_predictor
- write(t, action, time_total=None, time_update_predictor=None, time_get_action=None, time_run_simulator=None)[source]
Overwrite fx and chosen_actions by t and action.
- Parameters:
t (numpy.ndarray) – N dimensional array. The negative energy of each search candidate (value of the objective function to be optimized).
action (numpy.ndarray) – N dimensional array. The indexes of actions of each search candidate.
time_total (numpy.ndarray) – N dimenstional array. The total elapsed time in each step. If None (default), filled by 0.0.
time_update_predictor (numpy.ndarray) – N dimenstional array. The elapsed time for updating predictor (e.g., learning hyperparemters) in each step. If None (default), filled by 0.0.
time_get_action (numpy.ndarray) – N dimenstional array. The elapsed time for getting next action in each step. If None (default), filled by 0.0.
time_run_simulator (numpy.ndarray) – N dimenstional array. The elapsed time for running the simulator in each step. If None (default), filled by 0.0.
- class physbo.search.discrete.Policy(test_X, config=None, initial_data=None, comm=None)[source]
Bases:
objectSingle objective Bayesian optimization with discrete search space
- Parameters:
test_X (numpy.ndarray or physbo.Variable) – The set of candidates. Each row vector represents the feature vector of each search candidate.
config (SetConfig object (physbo.misc.SetConfig))
initial_data (tuple[np.ndarray, np.ndarray]) – The initial training datasets. The first elements is the array of actions and the second is the array of value of objective functions
comm (MPI.Comm, optional) – MPI Communicator
- actions: ndarray
Array of action indices that have not been searched yet.
Initially contains all candidate indices, and decreases as actions are evaluated.
- bayes_search(training=None, max_num_probes=None, num_search_each_probe=1, predictor=None, is_disp=True, simulator=None, score='TS', interval=0, num_rand_basis=0)[source]
Performing Bayesian optimization.
- Parameters:
training (physbo.Variable) – Training dataset.
max_num_probes (int) – Maximum number of searching process by Bayesian optimization.
num_search_each_probe (int) – Number of searching by Bayesian optimization at each process.
predictor (predictor object) – Base class is defined in physbo.predictor. If None, blm_predictor is defined.
is_disp (bool) – If true, process messages are outputted.
simulator (callable) – Callable (function or object with
__call__) Here, action is an integer which represents the index of the candidate.score (str) – The type of aquision funciton. TS (Thompson Sampling), EI (Expected Improvement) and PI (Probability of Improvement) are available.
interval (int) – The interval number of learning the hyper parameter. If you set the negative value to interval, the hyper parameter learning is not performed. If you set zero to interval, the hyper parameter learning is performed only at the first step.
num_rand_basis (int) – The number of basis function. If you choose 0, ordinary Gaussian process run.
- Returns:
history
- Return type:
history object (physbo.search.discrete.results.history)
- config: SetConfig
Configuration object containing settings for learning and search.
- get_permutation_importance(n_perm: int, split_features_parallel=False)[source]
Calculating permutation importance of model
- Parameters:
n_perm (int) – number of permutations
split_features_parallel (bool) – If true, split features in parallel.
- Returns:
importance_mean (numpy.ndarray) – importance_mean
importance_std (numpy.ndarray) – importance_std
- get_post_fcov(xs, diag=True)[source]
Calculate covariance of predictor (post distribution)
- Parameters:
xs (physbo.Variable or np.ndarray) – input parameters to calculate covariance shape is (num_points, num_parameters)
diag (bool) – If true, only variances (diagonal elements) are returned.
- Returns:
fcov – Covariance matrix of the post distribution. Returned shape is (num_points) if diag=true, (num_points, num_points) if diag=false.
- Return type:
numpy.ndarray
- get_post_fmean(xs)[source]
Calculate mean value of predictor (post distribution)
- Parameters:
xs (physbo.Variable or np.ndarray) – input parameters to calculate mean value shape is (num_points, num_parameters)
- Returns:
fmean – Mean value of the post distribution. Returned shape is (num_points).
- Return type:
numpy.ndarray
- get_score(mode, *, actions=None, xs=None, predictor=None, training=None, parallel=True, alpha=1)[source]
Calcualte score (acquisition function)
- Parameters:
mode (str) – The type of aquisition funciton. TS, EI and PI are available. These functions are defined in score.py.
actions (array of int) – actions to calculate score
xs (physbo.Variable or np.ndarray) – input parameters to calculate score
predictor (predictor object) – predictor used to calculate score. If not given, self.predictor will be used.
training (physbo.Variable) – Training dataset. If not given, self.training will be used.
parallel (bool) – Calculate scores in parallel by MPI (default: True)
alpha (float) – Tuning parameter which is used if mode = TS. In TS, multi variation is tuned as np.random.multivariate_normal(mean, cov*alpha**2, size).
- Returns:
f – Score defined in each mode.
- Return type:
float or list of float
- Raises:
RuntimeError – If both actions and xs are given
Notes
When neither actions nor xs are given, scores for actions not yet searched will be calculated.
When parallel is True, it is assumed that the function receives the same input (actions or xs) for all the ranks. If you want to split the input array itself, set parallel be False and merge results by yourself.
- history: History
History object storing search results including chosen actions,
objective function values, and timing information.
- load(file_history, file_training=None, file_predictor=None)[source]
Loading files about history, training and predictor.
- Parameters:
file_history (str) – The name of the file that stores the information of the history.
file_training (str) – The name of the file that stores the training dataset.
file_predictor (str) – The name of the file that stores the predictor dataset.
- mpicomm: Any | None
MPI communicator for parallel computation.
If None, MPI is not used (single process).
- mpirank: int
Rank of the current MPI process. Set to 0 if MPI is not used.
- mpisize: int
Number of MPI processes. Set to 1 if MPI is not used.
- new_data: Variable | None
New data that has been added to training but not yet to predictor.
Set to None after predictor is updated.
- predictor: Any | None
Predictor used for Bayesian optimization.
Base class is defined in physbo.predictor. If None, predictor is not yet initialized.
- random_search(max_num_probes, num_search_each_probe=1, simulator=None, is_disp=True)[source]
Performing random search.
- Parameters:
max_num_probes (int) – Maximum number of random search process.
num_search_each_probe (int) – Number of search at each random search process.
simulator (callable) – Callable (function or object with
__call__) from action to t Here, action is an integer which represents the index of the candidate.is_disp (bool) – If true, process messages are outputted.
- Returns:
history
- Return type:
history object (physbo.search.discrete.results.history)
- save(file_history, file_training=None, file_predictor=None)[source]
Saving history, training and predictor into the corresponding files.
- Parameters:
file_history (str) – The name of the file that stores the information of the history.
file_training (str) – The name of the file that stores the training dataset.
file_predictor (str) – The name of the file that stores the predictor dataset.
- seed: int | None
Seed parameter for np.random. Set by set_seed() method.
- set_seed(seed)[source]
Setting a seed parameter for np.random.
- Parameters:
seed (int) – seed number
-------
- test: Variable
The set of candidate search points.
Each row of X represents the feature vector of each search candidate.
- training: Variable
Training dataset containing pairs of input (X) and output (t).
Stores all evaluated data points.
- write(action, t, X=None, time_total=None, time_update_predictor=None, time_get_action=None, time_run_simulator=None)[source]
Writing history (update history, not output to a file).
- Parameters:
action (numpy.ndarray) – Indexes of actions.
t (numpy.ndarray) – N dimensional array (1D) or N x k dimensional array (2D). The negative energy of each search candidate (value of the objective function to be optimized). Will be normalized to (N, 1) shape internally.
X (numpy.ndarray) – N x d dimensional matrix. Each row of X denotes the d-dimensional feature vector of each search candidate.
time_total (numpy.ndarray) – N dimenstional array. The total elapsed time in each step. If None (default), filled by 0.0.
time_update_predictor (numpy.ndarray) – N dimenstional array. The elapsed time for updating predictor (e.g., learning hyperparemters) in each step. If None (default), filled by 0.0.
time_get_action (numpy.ndarray) – N dimenstional array. The elapsed time for getting next action in each step. If None (default), filled by 0.0.
time_run_simulator (numpy.ndarray) – N dimenstional array. The elapsed time for running the simulator in each step. If None (default), filled by 0.0.
Note
In this function, (X, t, Z) is added to the training data set (self.training), but not yet to predictor (self.predictor). self.newdata means such data. To add (X, t, Z) to the predictor, call self._update_predictor() after self.write().