Type tf.keras.models
Namespace tensorflow
Methods
- clone_model
- clone_model
- clone_model
- clone_model
- clone_model_dyn
- load_model
- load_model_dyn
- model_from_config
- model_from_config
- model_from_config_dyn
- model_from_json
- model_from_json_dyn
- model_from_yaml
- model_from_yaml_dyn
- save_model
- save_model
- save_model
- save_model
- save_model
- save_model
- save_model_dyn
Properties
Public static methods
object clone_model(Model model, IGraphNodeBase input_tensors, object clone_function)
Clone any `Model` instance. Model cloning is similar to calling a model on new inputs,
except that it creates new layers (and thus new weights) instead
of sharing the weights of the existing layers.
Parameters
-
Modelmodel - Instance of `Model` (could be a functional model or a Sequential model).
-
IGraphNodeBaseinput_tensors - optional list of input tensors or InputLayer objects to build the model upon. If not provided, placeholders will be created.
-
objectclone_function - Callable to be used to clone each layer in the target model (except `InputLayer` instances). It takes as argument the layer instance to be cloned, and returns the corresponding layer instance to be used in the model copy. If unspecified, this callable defaults to the following serialization/deserialization function: `lambda layer: layer.__class__.from_config(layer.get_config())`. By passing a custom callable, you can customize your copy of the model, e.g. by wrapping certain layers of interest (you might want to replace all `LSTM` instances with equivalent `Bidirectional(LSTM(...))` instances, for example).
Returns
-
object - An instance of `Model` reproducing the behavior of the original model, on top of new inputs tensors, using newly instantiated weights. The cloned model might behave differently from the original model if a custom clone_function modifies the layer.
object clone_model(Model model, ValueTuple<IEnumerable<object>, object> input_tensors, object clone_function)
Clone any `Model` instance. Model cloning is similar to calling a model on new inputs,
except that it creates new layers (and thus new weights) instead
of sharing the weights of the existing layers.
Parameters
-
Modelmodel - Instance of `Model` (could be a functional model or a Sequential model).
-
ValueTuple<IEnumerable<object>, object>input_tensors - optional list of input tensors or InputLayer objects to build the model upon. If not provided, placeholders will be created.
-
objectclone_function - Callable to be used to clone each layer in the target model (except `InputLayer` instances). It takes as argument the layer instance to be cloned, and returns the corresponding layer instance to be used in the model copy. If unspecified, this callable defaults to the following serialization/deserialization function: `lambda layer: layer.__class__.from_config(layer.get_config())`. By passing a custom callable, you can customize your copy of the model, e.g. by wrapping certain layers of interest (you might want to replace all `LSTM` instances with equivalent `Bidirectional(LSTM(...))` instances, for example).
Returns
-
object - An instance of `Model` reproducing the behavior of the original model, on top of new inputs tensors, using newly instantiated weights. The cloned model might behave differently from the original model if a custom clone_function modifies the layer.
object clone_model(Model model, IEnumerable<IGraphNodeBase> input_tensors, object clone_function)
Clone any `Model` instance. Model cloning is similar to calling a model on new inputs,
except that it creates new layers (and thus new weights) instead
of sharing the weights of the existing layers.
Parameters
-
Modelmodel - Instance of `Model` (could be a functional model or a Sequential model).
-
IEnumerable<IGraphNodeBase>input_tensors - optional list of input tensors or InputLayer objects to build the model upon. If not provided, placeholders will be created.
-
objectclone_function - Callable to be used to clone each layer in the target model (except `InputLayer` instances). It takes as argument the layer instance to be cloned, and returns the corresponding layer instance to be used in the model copy. If unspecified, this callable defaults to the following serialization/deserialization function: `lambda layer: layer.__class__.from_config(layer.get_config())`. By passing a custom callable, you can customize your copy of the model, e.g. by wrapping certain layers of interest (you might want to replace all `LSTM` instances with equivalent `Bidirectional(LSTM(...))` instances, for example).
Returns
-
object - An instance of `Model` reproducing the behavior of the original model, on top of new inputs tensors, using newly instantiated weights. The cloned model might behave differently from the original model if a custom clone_function modifies the layer.
object clone_model(Model model, IDictionary<object, object> input_tensors, object clone_function)
Clone any `Model` instance. Model cloning is similar to calling a model on new inputs,
except that it creates new layers (and thus new weights) instead
of sharing the weights of the existing layers.
Parameters
-
Modelmodel - Instance of `Model` (could be a functional model or a Sequential model).
-
IDictionary<object, object>input_tensors - optional list of input tensors or InputLayer objects to build the model upon. If not provided, placeholders will be created.
-
objectclone_function - Callable to be used to clone each layer in the target model (except `InputLayer` instances). It takes as argument the layer instance to be cloned, and returns the corresponding layer instance to be used in the model copy. If unspecified, this callable defaults to the following serialization/deserialization function: `lambda layer: layer.__class__.from_config(layer.get_config())`. By passing a custom callable, you can customize your copy of the model, e.g. by wrapping certain layers of interest (you might want to replace all `LSTM` instances with equivalent `Bidirectional(LSTM(...))` instances, for example).
Returns
-
object - An instance of `Model` reproducing the behavior of the original model, on top of new inputs tensors, using newly instantiated weights. The cloned model might behave differently from the original model if a custom clone_function modifies the layer.
object clone_model_dyn(object model, object input_tensors, object clone_function)
Clone any `Model` instance. Model cloning is similar to calling a model on new inputs,
except that it creates new layers (and thus new weights) instead
of sharing the weights of the existing layers.
Parameters
-
objectmodel - Instance of `Model` (could be a functional model or a Sequential model).
-
objectinput_tensors - optional list of input tensors or InputLayer objects to build the model upon. If not provided, placeholders will be created.
-
objectclone_function - Callable to be used to clone each layer in the target model (except `InputLayer` instances). It takes as argument the layer instance to be cloned, and returns the corresponding layer instance to be used in the model copy. If unspecified, this callable defaults to the following serialization/deserialization function: `lambda layer: layer.__class__.from_config(layer.get_config())`. By passing a custom callable, you can customize your copy of the model, e.g. by wrapping certain layers of interest (you might want to replace all `LSTM` instances with equivalent `Bidirectional(LSTM(...))` instances, for example).
Returns
-
object - An instance of `Model` reproducing the behavior of the original model, on top of new inputs tensors, using newly instantiated weights. The cloned model might behave differently from the original model if a custom clone_function modifies the layer.
RevivedModel load_model(object filepath, object custom_objects, bool compile)
Loads a model saved via `save_model`.
Parameters
-
objectfilepath - One of the following: - String, path to the saved model - `h5py.File` object from which to load the model
-
objectcustom_objects - Optional dictionary mapping names (strings) to custom classes or functions to be considered during deserialization.
-
boolcompile - Boolean, whether to compile the model after loading.
Returns
-
RevivedModel - A Keras model instance. If an optimizer was found as part of the saved model, the model is already compiled. Otherwise, the model is uncompiled and a warning will be displayed. When `compile` is set to False, the compilation is omitted without any warning.
object load_model_dyn(object filepath, object custom_objects, ImplicitContainer<T> compile)
Loads a model saved via `save_model`.
Parameters
-
objectfilepath - One of the following: - String, path to the saved model - `h5py.File` object from which to load the model
-
objectcustom_objects - Optional dictionary mapping names (strings) to custom classes or functions to be considered during deserialization.
-
ImplicitContainer<T>compile - Boolean, whether to compile the model after loading.
Returns
-
object - A Keras model instance. If an optimizer was found as part of the saved model, the model is already compiled. Otherwise, the model is uncompiled and a warning will be displayed. When `compile` is set to False, the compilation is omitted without any warning.
object model_from_config(IEnumerable<object> config, IDictionary<object, object> custom_objects)
Instantiates a Keras model from its config.
Parameters
-
IEnumerable<object>config - Configuration dictionary.
-
IDictionary<object, object>custom_objects - Optional dictionary mapping names (strings) to custom classes or functions to be considered during deserialization.
Returns
-
object - A Keras model instance (uncompiled).
object model_from_config(object config, IDictionary<object, object> custom_objects)
Instantiates a Keras model from its config.
Parameters
-
objectconfig - Configuration dictionary.
-
IDictionary<object, object>custom_objects - Optional dictionary mapping names (strings) to custom classes or functions to be considered during deserialization.
Returns
-
object - A Keras model instance (uncompiled).
object model_from_config_dyn(object config, object custom_objects)
Instantiates a Keras model from its config.
Parameters
-
objectconfig - Configuration dictionary.
-
objectcustom_objects - Optional dictionary mapping names (strings) to custom classes or functions to be considered during deserialization.
Returns
-
object - A Keras model instance (uncompiled).
object model_from_json(string json_string, object custom_objects)
Parses a JSON model configuration file and returns a model instance.
Parameters
-
stringjson_string - JSON string encoding a model configuration.
-
objectcustom_objects - Optional dictionary mapping names (strings) to custom classes or functions to be considered during deserialization.
Returns
-
object - A Keras model instance (uncompiled).
object model_from_json_dyn(object json_string, object custom_objects)
Parses a JSON model configuration file and returns a model instance.
Parameters
-
objectjson_string - JSON string encoding a model configuration.
-
objectcustom_objects - Optional dictionary mapping names (strings) to custom classes or functions to be considered during deserialization.
Returns
-
object - A Keras model instance (uncompiled).
object model_from_yaml(object yaml_string, object custom_objects)
Parses a yaml model configuration file and returns a model instance.
Parameters
-
objectyaml_string - YAML string encoding a model configuration.
-
objectcustom_objects - Optional dictionary mapping names (strings) to custom classes or functions to be considered during deserialization.
Returns
-
object - A Keras model instance (uncompiled).
object model_from_yaml_dyn(object yaml_string, object custom_objects)
Parses a yaml model configuration file and returns a model instance.
Parameters
-
objectyaml_string - YAML string encoding a model configuration.
-
objectcustom_objects - Optional dictionary mapping names (strings) to custom classes or functions to be considered during deserialization.
Returns
-
object - A Keras model instance (uncompiled).
void save_model(AutoTrackable model, IGraphNodeBase filepath, bool overwrite, bool include_optimizer, string save_format, object signatures)
Saves a model as a TensorFlow SavedModel or HDF5 file. The saved model contains:
- the model's configuration (topology)
- the model's weights
- the model's optimizer's state (if any) Thus the saved model can be reinstantiated in
the exact same state, without any of the code
used for model definition or training. _SavedModel serialization_ (not yet added) The SavedModel serialization path uses
tf.saved_model.save to save the model
and all trackable objects attached to the model (e.g. layers and variables).
`@tf.function`-decorated methods are also saved. Additional trackable objects
and functions are added to the SavedModel to allow the model to be
loaded back as a Keras Model object.
Parameters
-
AutoTrackablemodel - Keras model instance to be saved.
-
IGraphNodeBasefilepath - One of the following: - String, path where to save the model - `h5py.File` object where to save the model
-
booloverwrite - Whether we should overwrite any existing model at the target location, or instead ask the user with a manual prompt.
-
boolinclude_optimizer - If True, save optimizer's state together.
-
stringsave_format - Either 'tf' or 'h5', indicating whether to save the model to Tensorflow SavedModel or HDF5. Defaults to 'tf' in TF 2.X, and 'h5' in TF 1.X.
-
objectsignatures - Signatures to save with the SavedModel. Applicable to the 'tf'
format only. Please see the `signatures` argument in
tf.saved_model.savefor details.
void save_model(AutoTrackable model, Byte[] filepath, bool overwrite, bool include_optimizer, string save_format, object signatures)
Saves a model as a TensorFlow SavedModel or HDF5 file. The saved model contains:
- the model's configuration (topology)
- the model's weights
- the model's optimizer's state (if any) Thus the saved model can be reinstantiated in
the exact same state, without any of the code
used for model definition or training. _SavedModel serialization_ (not yet added) The SavedModel serialization path uses
tf.saved_model.save to save the model
and all trackable objects attached to the model (e.g. layers and variables).
`@tf.function`-decorated methods are also saved. Additional trackable objects
and functions are added to the SavedModel to allow the model to be
loaded back as a Keras Model object.
Parameters
-
AutoTrackablemodel - Keras model instance to be saved.
-
Byte[]filepath - One of the following: - String, path where to save the model - `h5py.File` object where to save the model
-
booloverwrite - Whether we should overwrite any existing model at the target location, or instead ask the user with a manual prompt.
-
boolinclude_optimizer - If True, save optimizer's state together.
-
stringsave_format - Either 'tf' or 'h5', indicating whether to save the model to Tensorflow SavedModel or HDF5. Defaults to 'tf' in TF 2.X, and 'h5' in TF 1.X.
-
objectsignatures - Signatures to save with the SavedModel. Applicable to the 'tf'
format only. Please see the `signatures` argument in
tf.saved_model.savefor details.
void save_model(Checkpoint model, string filepath, bool overwrite, bool include_optimizer, string save_format, object signatures)
Saves a model as a TensorFlow SavedModel or HDF5 file. The saved model contains:
- the model's configuration (topology)
- the model's weights
- the model's optimizer's state (if any) Thus the saved model can be reinstantiated in
the exact same state, without any of the code
used for model definition or training. _SavedModel serialization_ (not yet added) The SavedModel serialization path uses
tf.saved_model.save to save the model
and all trackable objects attached to the model (e.g. layers and variables).
`@tf.function`-decorated methods are also saved. Additional trackable objects
and functions are added to the SavedModel to allow the model to be
loaded back as a Keras Model object.
Parameters
-
Checkpointmodel - Keras model instance to be saved.
-
stringfilepath - One of the following: - String, path where to save the model - `h5py.File` object where to save the model
-
booloverwrite - Whether we should overwrite any existing model at the target location, or instead ask the user with a manual prompt.
-
boolinclude_optimizer - If True, save optimizer's state together.
-
stringsave_format - Either 'tf' or 'h5', indicating whether to save the model to Tensorflow SavedModel or HDF5. Defaults to 'tf' in TF 2.X, and 'h5' in TF 1.X.
-
objectsignatures - Signatures to save with the SavedModel. Applicable to the 'tf'
format only. Please see the `signatures` argument in
tf.saved_model.savefor details.
void save_model(Checkpoint model, IGraphNodeBase filepath, bool overwrite, bool include_optimizer, string save_format, object signatures)
Saves a model as a TensorFlow SavedModel or HDF5 file. The saved model contains:
- the model's configuration (topology)
- the model's weights
- the model's optimizer's state (if any) Thus the saved model can be reinstantiated in
the exact same state, without any of the code
used for model definition or training. _SavedModel serialization_ (not yet added) The SavedModel serialization path uses
tf.saved_model.save to save the model
and all trackable objects attached to the model (e.g. layers and variables).
`@tf.function`-decorated methods are also saved. Additional trackable objects
and functions are added to the SavedModel to allow the model to be
loaded back as a Keras Model object.
Parameters
-
Checkpointmodel - Keras model instance to be saved.
-
IGraphNodeBasefilepath - One of the following: - String, path where to save the model - `h5py.File` object where to save the model
-
booloverwrite - Whether we should overwrite any existing model at the target location, or instead ask the user with a manual prompt.
-
boolinclude_optimizer - If True, save optimizer's state together.
-
stringsave_format - Either 'tf' or 'h5', indicating whether to save the model to Tensorflow SavedModel or HDF5. Defaults to 'tf' in TF 2.X, and 'h5' in TF 1.X.
-
objectsignatures - Signatures to save with the SavedModel. Applicable to the 'tf'
format only. Please see the `signatures` argument in
tf.saved_model.savefor details.
void save_model(Checkpoint model, Byte[] filepath, bool overwrite, bool include_optimizer, string save_format, object signatures)
Saves a model as a TensorFlow SavedModel or HDF5 file. The saved model contains:
- the model's configuration (topology)
- the model's weights
- the model's optimizer's state (if any) Thus the saved model can be reinstantiated in
the exact same state, without any of the code
used for model definition or training. _SavedModel serialization_ (not yet added) The SavedModel serialization path uses
tf.saved_model.save to save the model
and all trackable objects attached to the model (e.g. layers and variables).
`@tf.function`-decorated methods are also saved. Additional trackable objects
and functions are added to the SavedModel to allow the model to be
loaded back as a Keras Model object.
Parameters
-
Checkpointmodel - Keras model instance to be saved.
-
Byte[]filepath - One of the following: - String, path where to save the model - `h5py.File` object where to save the model
-
booloverwrite - Whether we should overwrite any existing model at the target location, or instead ask the user with a manual prompt.
-
boolinclude_optimizer - If True, save optimizer's state together.
-
stringsave_format - Either 'tf' or 'h5', indicating whether to save the model to Tensorflow SavedModel or HDF5. Defaults to 'tf' in TF 2.X, and 'h5' in TF 1.X.
-
objectsignatures - Signatures to save with the SavedModel. Applicable to the 'tf'
format only. Please see the `signatures` argument in
tf.saved_model.savefor details.
void save_model(AutoTrackable model, string filepath, bool overwrite, bool include_optimizer, string save_format, object signatures)
Saves a model as a TensorFlow SavedModel or HDF5 file. The saved model contains:
- the model's configuration (topology)
- the model's weights
- the model's optimizer's state (if any) Thus the saved model can be reinstantiated in
the exact same state, without any of the code
used for model definition or training. _SavedModel serialization_ (not yet added) The SavedModel serialization path uses
tf.saved_model.save to save the model
and all trackable objects attached to the model (e.g. layers and variables).
`@tf.function`-decorated methods are also saved. Additional trackable objects
and functions are added to the SavedModel to allow the model to be
loaded back as a Keras Model object.
Parameters
-
AutoTrackablemodel - Keras model instance to be saved.
-
stringfilepath - One of the following: - String, path where to save the model - `h5py.File` object where to save the model
-
booloverwrite - Whether we should overwrite any existing model at the target location, or instead ask the user with a manual prompt.
-
boolinclude_optimizer - If True, save optimizer's state together.
-
stringsave_format - Either 'tf' or 'h5', indicating whether to save the model to Tensorflow SavedModel or HDF5. Defaults to 'tf' in TF 2.X, and 'h5' in TF 1.X.
-
objectsignatures - Signatures to save with the SavedModel. Applicable to the 'tf'
format only. Please see the `signatures` argument in
tf.saved_model.savefor details.
object save_model_dyn(object model, object filepath, ImplicitContainer<T> overwrite, ImplicitContainer<T> include_optimizer, object save_format, object signatures)
Saves a model as a TensorFlow SavedModel or HDF5 file. The saved model contains:
- the model's configuration (topology)
- the model's weights
- the model's optimizer's state (if any) Thus the saved model can be reinstantiated in
the exact same state, without any of the code
used for model definition or training. _SavedModel serialization_ (not yet added) The SavedModel serialization path uses
tf.saved_model.save to save the model
and all trackable objects attached to the model (e.g. layers and variables).
`@tf.function`-decorated methods are also saved. Additional trackable objects
and functions are added to the SavedModel to allow the model to be
loaded back as a Keras Model object.
Parameters
-
objectmodel - Keras model instance to be saved.
-
objectfilepath - One of the following: - String, path where to save the model - `h5py.File` object where to save the model
-
ImplicitContainer<T>overwrite - Whether we should overwrite any existing model at the target location, or instead ask the user with a manual prompt.
-
ImplicitContainer<T>include_optimizer - If True, save optimizer's state together.
-
objectsave_format - Either 'tf' or 'h5', indicating whether to save the model to Tensorflow SavedModel or HDF5. Defaults to 'tf' in TF 2.X, and 'h5' in TF 1.X.
-
objectsignatures - Signatures to save with the SavedModel. Applicable to the 'tf'
format only. Please see the `signatures` argument in
tf.saved_model.savefor details.