Package inference

Class ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder

java.lang.Object
com.google.protobuf.AbstractMessageLite.Builder
com.google.protobuf.AbstractMessage.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
inference.ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder
All Implemented Interfaces:
com.google.protobuf.Message.Builder, com.google.protobuf.MessageLite.Builder, com.google.protobuf.MessageLiteOrBuilder, com.google.protobuf.MessageOrBuilder, ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder, Cloneable
Enclosing class:
ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators

public static final class ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder extends com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder> implements ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
@@
@@  .. cpp:var:: message ExecutionAccelerators
@@
@@     Specify the preferred execution accelerators to be used to execute
@@     the model. Currently only recognized by ONNX Runtime backend and
@@     TensorFlow backend.
@@
@@     For ONNX Runtime backend, it will deploy the model with the execution
@@     accelerators by priority, the priority is determined based on the
@@     order that they are set, i.e. the provider at the front has highest
@@     priority. Overall, the priority will be in the following order:
@@         <gpu_execution_accelerator> (if instance is on GPU)
@@         CUDA Execution Provider     (if instance is on GPU)
@@         <cpu_execution_accelerator>
@@         Default CPU Execution Provider
@@
 
Protobuf type inference.ModelOptimizationPolicy.ExecutionAccelerators
  • Method Details

    • getDescriptor

      public static final com.google.protobuf.Descriptors.Descriptor getDescriptor()
    • internalGetFieldAccessorTable

      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
      Specified by:
      internalGetFieldAccessorTable in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • clear

      Specified by:
      clear in interface com.google.protobuf.Message.Builder
      Specified by:
      clear in interface com.google.protobuf.MessageLite.Builder
      Overrides:
      clear in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • getDescriptorForType

      public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()
      Specified by:
      getDescriptorForType in interface com.google.protobuf.Message.Builder
      Specified by:
      getDescriptorForType in interface com.google.protobuf.MessageOrBuilder
      Overrides:
      getDescriptorForType in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • getDefaultInstanceForType

      Specified by:
      getDefaultInstanceForType in interface com.google.protobuf.MessageLiteOrBuilder
      Specified by:
      getDefaultInstanceForType in interface com.google.protobuf.MessageOrBuilder
    • build

      Specified by:
      build in interface com.google.protobuf.Message.Builder
      Specified by:
      build in interface com.google.protobuf.MessageLite.Builder
    • buildPartial

      Specified by:
      buildPartial in interface com.google.protobuf.Message.Builder
      Specified by:
      buildPartial in interface com.google.protobuf.MessageLite.Builder
    • clone

      Specified by:
      clone in interface com.google.protobuf.Message.Builder
      Specified by:
      clone in interface com.google.protobuf.MessageLite.Builder
      Overrides:
      clone in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • setField

      public ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder setField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value)
      Specified by:
      setField in interface com.google.protobuf.Message.Builder
      Overrides:
      setField in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • clearField

      public ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder clearField(com.google.protobuf.Descriptors.FieldDescriptor field)
      Specified by:
      clearField in interface com.google.protobuf.Message.Builder
      Overrides:
      clearField in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • clearOneof

      public ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof)
      Specified by:
      clearOneof in interface com.google.protobuf.Message.Builder
      Overrides:
      clearOneof in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • setRepeatedField

      public ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder setRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, int index, Object value)
      Specified by:
      setRepeatedField in interface com.google.protobuf.Message.Builder
      Overrides:
      setRepeatedField in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • addRepeatedField

      public ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder addRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value)
      Specified by:
      addRepeatedField in interface com.google.protobuf.Message.Builder
      Overrides:
      addRepeatedField in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • mergeFrom

      public ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder mergeFrom(com.google.protobuf.Message other)
      Specified by:
      mergeFrom in interface com.google.protobuf.Message.Builder
      Overrides:
      mergeFrom in class com.google.protobuf.AbstractMessage.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • mergeFrom

    • isInitialized

      public final boolean isInitialized()
      Specified by:
      isInitialized in interface com.google.protobuf.MessageLiteOrBuilder
      Overrides:
      isInitialized in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • mergeFrom

      public ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder mergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws IOException
      Specified by:
      mergeFrom in interface com.google.protobuf.Message.Builder
      Specified by:
      mergeFrom in interface com.google.protobuf.MessageLite.Builder
      Overrides:
      mergeFrom in class com.google.protobuf.AbstractMessage.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
      Throws:
      IOException
    • getGpuExecutionAcceleratorList

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
      Specified by:
      getGpuExecutionAcceleratorList in interface ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
    • getGpuExecutionAcceleratorCount

      public int getGpuExecutionAcceleratorCount()
      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
      Specified by:
      getGpuExecutionAcceleratorCount in interface ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
    • getGpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
      Specified by:
      getGpuExecutionAccelerator in interface ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
    • setGpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • setGpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • addGpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • addGpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • addGpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • addGpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • addAllGpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • clearGpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • removeGpuExecutionAccelerator

      public ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder removeGpuExecutionAccelerator(int index)
      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • getGpuExecutionAcceleratorBuilder

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • getGpuExecutionAcceleratorOrBuilder

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
      Specified by:
      getGpuExecutionAcceleratorOrBuilder in interface ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
    • getGpuExecutionAcceleratorOrBuilderList

      public List<? extends ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.AcceleratorOrBuilder> getGpuExecutionAcceleratorOrBuilderList()
      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
      Specified by:
      getGpuExecutionAcceleratorOrBuilderList in interface ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
    • addGpuExecutionAcceleratorBuilder

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • addGpuExecutionAcceleratorBuilder

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • getGpuExecutionAcceleratorBuilderList

      @@    .. cpp:var:: Accelerator gpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on GPU.
      @@
      @@       For ONNX Runtime backend, possible value is "tensorrt" as name,
      @@       and no parameters are required.
      @@
      @@       For TensorFlow backend, possible values are "tensorrt",
      @@       "auto_mixed_precision", "gpu_io".
      @@
      @@       For "tensorrt", the following parameters can be specified:
      @@         "precision_mode": The precision used for optimization.
      @@         Allowed values are "FP32" and "FP16". Default value is "FP32".
      @@
      @@         "max_cached_engines": The maximum number of cached TensorRT
      @@         engines in dynamic TensorRT ops. Default value is 100.
      @@
      @@         "minimum_segment_size": The smallest model subgraph that will
      @@         be considered for optimization by TensorRT. Default value is 3.
      @@
      @@         "max_workspace_size_bytes": The maximum GPU memory the model
      @@         can use temporarily during execution. Default value is 1GB.
      @@
      @@       For "auto_mixed_precision", no parameters are required. If set,
      @@       the model will try to use FP16 for better performance.
      @@       This optimization can not be set with "tensorrt".
      @@
      @@       For "gpu_io", no parameters are required. If set, the model will
      @@       be executed using TensorFlow Callable API to set input and output
      @@       tensors in GPU memory if possible, which can reduce data transfer
      @@       overhead if the model is used in ensemble. However, the Callable
      @@       object will be created on model creation and it will request all
      @@       outputs for every model execution, which may impact the
      @@       performance if a request does not require all outputs. This
      @@       optimization will only take affect if the model instance is
      @@       created with KIND_GPU.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator gpu_execution_accelerator = 1;
    • getCpuExecutionAcceleratorList

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
      Specified by:
      getCpuExecutionAcceleratorList in interface ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
    • getCpuExecutionAcceleratorCount

      public int getCpuExecutionAcceleratorCount()
      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
      Specified by:
      getCpuExecutionAcceleratorCount in interface ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
    • getCpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
      Specified by:
      getCpuExecutionAccelerator in interface ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
    • setCpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • setCpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • addCpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • addCpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • addCpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • addCpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • addAllCpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • clearCpuExecutionAccelerator

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • removeCpuExecutionAccelerator

      public ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder removeCpuExecutionAccelerator(int index)
      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • getCpuExecutionAcceleratorBuilder

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • getCpuExecutionAcceleratorOrBuilder

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
      Specified by:
      getCpuExecutionAcceleratorOrBuilder in interface ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
    • getCpuExecutionAcceleratorOrBuilderList

      public List<? extends ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.AcceleratorOrBuilder> getCpuExecutionAcceleratorOrBuilderList()
      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
      Specified by:
      getCpuExecutionAcceleratorOrBuilderList in interface ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAcceleratorsOrBuilder
    • addCpuExecutionAcceleratorBuilder

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • addCpuExecutionAcceleratorBuilder

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • getCpuExecutionAcceleratorBuilderList

      @@    .. cpp:var:: Accelerator cpu_execution_accelerator (repeated)
      @@
      @@       The preferred execution provider to be used if the model instance
      @@       is deployed on CPU.
      @@
      @@       For ONNX Runtime backend, possible value is "openvino" as name,
      @@       and no parameters are required.
      @@
       
      repeated .inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator cpu_execution_accelerator = 2;
    • setUnknownFields

      public final ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder setUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields)
      Specified by:
      setUnknownFields in interface com.google.protobuf.Message.Builder
      Overrides:
      setUnknownFields in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>
    • mergeUnknownFields

      public final ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder mergeUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields)
      Specified by:
      mergeUnknownFields in interface com.google.protobuf.Message.Builder
      Overrides:
      mergeUnknownFields in class com.google.protobuf.GeneratedMessageV3.Builder<ModelConfigOuterClass.ModelOptimizationPolicy.ExecutionAccelerators.Builder>