Package org.tensorflow.framework
Class GPUOptions.Builder
java.lang.Object
com.google.protobuf.AbstractMessageLite.Builder
com.google.protobuf.AbstractMessage.Builder<GPUOptions.Builder>
com.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
org.tensorflow.framework.GPUOptions.Builder
- All Implemented Interfaces:
com.google.protobuf.Message.Builder,com.google.protobuf.MessageLite.Builder,com.google.protobuf.MessageLiteOrBuilder,com.google.protobuf.MessageOrBuilder,Cloneable,GPUOptionsOrBuilder
- Enclosing class:
GPUOptions
public static final class GPUOptions.Builder
extends com.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
implements GPUOptionsOrBuilder
Protobuf type
tensorflow.GPUOptions-
Method Summary
Modifier and TypeMethodDescriptionaddRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value) build()clear()The type of GPU allocation strategy to use.If true, the allocator does not pre-allocate the entire specified GPU memory region, instead starting small and growing as needed.Delay deletion of up to this many bytes to reduce the number of interactions with gpu driver code.Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.clearField(com.google.protobuf.Descriptors.FieldDescriptor field) Force all tensors to be gpu_compatible.clearOneof(com.google.protobuf.Descriptors.OneofDescriptor oneof) Fraction of the total GPU memory to allocate for each process. 1 means to allocate all of the GPU memory, 0.5 means the process allocates up to ~50% of the total GPU memory.In the event polling loop sleep this many microseconds between PollEvents calls, when the queue is not empty.This field is deprecated and ignored.A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices.clone()The type of GPU allocation strategy to use.com.google.protobuf.ByteStringThe type of GPU allocation strategy to use.booleanIf true, the allocator does not pre-allocate the entire specified GPU memory region, instead starting small and growing as needed.longDelay deletion of up to this many bytes to reduce the number of interactions with gpu driver code.static final com.google.protobuf.Descriptors.Descriptorcom.google.protobuf.Descriptors.DescriptorEverything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.booleanForce all tensors to be gpu_compatible.doubleFraction of the total GPU memory to allocate for each process. 1 means to allocate all of the GPU memory, 0.5 means the process allocates up to ~50% of the total GPU memory.intIn the event polling loop sleep this many microseconds between PollEvents calls, when the queue is not empty.intThis field is deprecated and ignored.A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices.com.google.protobuf.ByteStringA comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices.booleanEverything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTablefinal booleanEverything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.mergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) mergeFrom(com.google.protobuf.Message other) mergeFrom(GPUOptions other) final GPUOptions.BuildermergeUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields) setAllocatorType(String value) The type of GPU allocation strategy to use.setAllocatorTypeBytes(com.google.protobuf.ByteString value) The type of GPU allocation strategy to use.setAllowGrowth(boolean value) If true, the allocator does not pre-allocate the entire specified GPU memory region, instead starting small and growing as needed.setDeferredDeletionBytes(long value) Delay deletion of up to this many bytes to reduce the number of interactions with gpu driver code.Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.setExperimental(GPUOptions.Experimental.Builder builderForValue) Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.setForceGpuCompatible(boolean value) Force all tensors to be gpu_compatible.setPerProcessGpuMemoryFraction(double value) Fraction of the total GPU memory to allocate for each process. 1 means to allocate all of the GPU memory, 0.5 means the process allocates up to ~50% of the total GPU memory.setPollingActiveDelayUsecs(int value) In the event polling loop sleep this many microseconds between PollEvents calls, when the queue is not empty.setPollingInactiveDelayMsecs(int value) This field is deprecated and ignored.setRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, int index, Object value) final GPUOptions.BuildersetUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields) setVisibleDeviceList(String value) A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices.setVisibleDeviceListBytes(com.google.protobuf.ByteString value) A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices.Methods inherited from class com.google.protobuf.GeneratedMessageV3.Builder
getAllFields, getField, getFieldBuilder, getOneofFieldDescriptor, getParentForChildren, getRepeatedField, getRepeatedFieldBuilder, getRepeatedFieldCount, getUnknownFields, getUnknownFieldSetBuilder, hasField, hasOneof, internalGetMapField, internalGetMapFieldReflection, internalGetMutableMapField, internalGetMutableMapFieldReflection, isClean, markClean, mergeUnknownLengthDelimitedField, mergeUnknownVarintField, newBuilderForField, onBuilt, onChanged, parseUnknownField, setUnknownFieldSetBuilder, setUnknownFieldsProto3Methods inherited from class com.google.protobuf.AbstractMessage.Builder
findInitializationErrors, getInitializationErrorString, internalMergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, newUninitializedMessageException, toStringMethods inherited from class com.google.protobuf.AbstractMessageLite.Builder
addAll, addAll, mergeDelimitedFrom, mergeDelimitedFrom, mergeFrom, newUninitializedMessageExceptionMethods inherited from class java.lang.Object
equals, finalize, getClass, hashCode, notify, notifyAll, wait, wait, waitMethods inherited from interface com.google.protobuf.Message.Builder
mergeDelimitedFrom, mergeDelimitedFromMethods inherited from interface com.google.protobuf.MessageLite.Builder
mergeFromMethods inherited from interface com.google.protobuf.MessageOrBuilder
findInitializationErrors, getAllFields, getField, getInitializationErrorString, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, getUnknownFields, hasField, hasOneof
-
Method Details
-
getDescriptor
public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() -
internalGetFieldAccessorTable
protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()- Specified by:
internalGetFieldAccessorTablein classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
clear
- Specified by:
clearin interfacecom.google.protobuf.Message.Builder- Specified by:
clearin interfacecom.google.protobuf.MessageLite.Builder- Overrides:
clearin classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
getDescriptorForType
public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()- Specified by:
getDescriptorForTypein interfacecom.google.protobuf.Message.Builder- Specified by:
getDescriptorForTypein interfacecom.google.protobuf.MessageOrBuilder- Overrides:
getDescriptorForTypein classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
getDefaultInstanceForType
- Specified by:
getDefaultInstanceForTypein interfacecom.google.protobuf.MessageLiteOrBuilder- Specified by:
getDefaultInstanceForTypein interfacecom.google.protobuf.MessageOrBuilder
-
build
- Specified by:
buildin interfacecom.google.protobuf.Message.Builder- Specified by:
buildin interfacecom.google.protobuf.MessageLite.Builder
-
buildPartial
- Specified by:
buildPartialin interfacecom.google.protobuf.Message.Builder- Specified by:
buildPartialin interfacecom.google.protobuf.MessageLite.Builder
-
clone
- Specified by:
clonein interfacecom.google.protobuf.Message.Builder- Specified by:
clonein interfacecom.google.protobuf.MessageLite.Builder- Overrides:
clonein classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
setField
public GPUOptions.Builder setField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value) - Specified by:
setFieldin interfacecom.google.protobuf.Message.Builder- Overrides:
setFieldin classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
clearField
- Specified by:
clearFieldin interfacecom.google.protobuf.Message.Builder- Overrides:
clearFieldin classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
clearOneof
- Specified by:
clearOneofin interfacecom.google.protobuf.Message.Builder- Overrides:
clearOneofin classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
setRepeatedField
public GPUOptions.Builder setRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, int index, Object value) - Specified by:
setRepeatedFieldin interfacecom.google.protobuf.Message.Builder- Overrides:
setRepeatedFieldin classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
addRepeatedField
public GPUOptions.Builder addRepeatedField(com.google.protobuf.Descriptors.FieldDescriptor field, Object value) - Specified by:
addRepeatedFieldin interfacecom.google.protobuf.Message.Builder- Overrides:
addRepeatedFieldin classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
mergeFrom
- Specified by:
mergeFromin interfacecom.google.protobuf.Message.Builder- Overrides:
mergeFromin classcom.google.protobuf.AbstractMessage.Builder<GPUOptions.Builder>
-
mergeFrom
-
isInitialized
public final boolean isInitialized()- Specified by:
isInitializedin interfacecom.google.protobuf.MessageLiteOrBuilder- Overrides:
isInitializedin classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
mergeFrom
public GPUOptions.Builder mergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws IOException - Specified by:
mergeFromin interfacecom.google.protobuf.Message.Builder- Specified by:
mergeFromin interfacecom.google.protobuf.MessageLite.Builder- Overrides:
mergeFromin classcom.google.protobuf.AbstractMessage.Builder<GPUOptions.Builder>- Throws:
IOException
-
getPerProcessGpuMemoryFraction
public double getPerProcessGpuMemoryFraction()Fraction of the total GPU memory to allocate for each process. 1 means to allocate all of the GPU memory, 0.5 means the process allocates up to ~50% of the total GPU memory. GPU memory is pre-allocated unless the allow_growth option is enabled. If greater than 1.0, uses CUDA unified memory to potentially oversubscribe the amount of memory available on the GPU device by using host memory as a swap space. Accessing memory not available on the device will be significantly slower as that would require memory transfer between the host and the device. Options to reduce the memory requirement should be considered before enabling this option as this may come with a negative performance impact. Oversubscription using the unified memory requires Pascal class or newer GPUs and it is currently only supported on the Linux operating system. See https://docs.nvidia.com/cuda/cuda-c-programming-guide/index.html#um-requirements for the detailed requirements.
double per_process_gpu_memory_fraction = 1;- Specified by:
getPerProcessGpuMemoryFractionin interfaceGPUOptionsOrBuilder- Returns:
- The perProcessGpuMemoryFraction.
-
setPerProcessGpuMemoryFraction
Fraction of the total GPU memory to allocate for each process. 1 means to allocate all of the GPU memory, 0.5 means the process allocates up to ~50% of the total GPU memory. GPU memory is pre-allocated unless the allow_growth option is enabled. If greater than 1.0, uses CUDA unified memory to potentially oversubscribe the amount of memory available on the GPU device by using host memory as a swap space. Accessing memory not available on the device will be significantly slower as that would require memory transfer between the host and the device. Options to reduce the memory requirement should be considered before enabling this option as this may come with a negative performance impact. Oversubscription using the unified memory requires Pascal class or newer GPUs and it is currently only supported on the Linux operating system. See https://docs.nvidia.com/cuda/cuda-c-programming-guide/index.html#um-requirements for the detailed requirements.
double per_process_gpu_memory_fraction = 1;- Parameters:
value- The perProcessGpuMemoryFraction to set.- Returns:
- This builder for chaining.
-
clearPerProcessGpuMemoryFraction
Fraction of the total GPU memory to allocate for each process. 1 means to allocate all of the GPU memory, 0.5 means the process allocates up to ~50% of the total GPU memory. GPU memory is pre-allocated unless the allow_growth option is enabled. If greater than 1.0, uses CUDA unified memory to potentially oversubscribe the amount of memory available on the GPU device by using host memory as a swap space. Accessing memory not available on the device will be significantly slower as that would require memory transfer between the host and the device. Options to reduce the memory requirement should be considered before enabling this option as this may come with a negative performance impact. Oversubscription using the unified memory requires Pascal class or newer GPUs and it is currently only supported on the Linux operating system. See https://docs.nvidia.com/cuda/cuda-c-programming-guide/index.html#um-requirements for the detailed requirements.
double per_process_gpu_memory_fraction = 1;- Returns:
- This builder for chaining.
-
getAllowGrowth
public boolean getAllowGrowth()If true, the allocator does not pre-allocate the entire specified GPU memory region, instead starting small and growing as needed.
bool allow_growth = 4;- Specified by:
getAllowGrowthin interfaceGPUOptionsOrBuilder- Returns:
- The allowGrowth.
-
setAllowGrowth
If true, the allocator does not pre-allocate the entire specified GPU memory region, instead starting small and growing as needed.
bool allow_growth = 4;- Parameters:
value- The allowGrowth to set.- Returns:
- This builder for chaining.
-
clearAllowGrowth
If true, the allocator does not pre-allocate the entire specified GPU memory region, instead starting small and growing as needed.
bool allow_growth = 4;- Returns:
- This builder for chaining.
-
getAllocatorType
The type of GPU allocation strategy to use. Allowed values: "": The empty string (default) uses a system-chosen default which may change over time. "BFC": A "Best-fit with coalescing" algorithm, simplified from a version of dlmalloc.string allocator_type = 2;- Specified by:
getAllocatorTypein interfaceGPUOptionsOrBuilder- Returns:
- The allocatorType.
-
getAllocatorTypeBytes
public com.google.protobuf.ByteString getAllocatorTypeBytes()The type of GPU allocation strategy to use. Allowed values: "": The empty string (default) uses a system-chosen default which may change over time. "BFC": A "Best-fit with coalescing" algorithm, simplified from a version of dlmalloc.string allocator_type = 2;- Specified by:
getAllocatorTypeBytesin interfaceGPUOptionsOrBuilder- Returns:
- The bytes for allocatorType.
-
setAllocatorType
The type of GPU allocation strategy to use. Allowed values: "": The empty string (default) uses a system-chosen default which may change over time. "BFC": A "Best-fit with coalescing" algorithm, simplified from a version of dlmalloc.string allocator_type = 2;- Parameters:
value- The allocatorType to set.- Returns:
- This builder for chaining.
-
clearAllocatorType
The type of GPU allocation strategy to use. Allowed values: "": The empty string (default) uses a system-chosen default which may change over time. "BFC": A "Best-fit with coalescing" algorithm, simplified from a version of dlmalloc.string allocator_type = 2;- Returns:
- This builder for chaining.
-
setAllocatorTypeBytes
The type of GPU allocation strategy to use. Allowed values: "": The empty string (default) uses a system-chosen default which may change over time. "BFC": A "Best-fit with coalescing" algorithm, simplified from a version of dlmalloc.string allocator_type = 2;- Parameters:
value- The bytes for allocatorType to set.- Returns:
- This builder for chaining.
-
getDeferredDeletionBytes
public long getDeferredDeletionBytes()Delay deletion of up to this many bytes to reduce the number of interactions with gpu driver code. If 0, the system chooses a reasonable default (several MBs).
int64 deferred_deletion_bytes = 3;- Specified by:
getDeferredDeletionBytesin interfaceGPUOptionsOrBuilder- Returns:
- The deferredDeletionBytes.
-
setDeferredDeletionBytes
Delay deletion of up to this many bytes to reduce the number of interactions with gpu driver code. If 0, the system chooses a reasonable default (several MBs).
int64 deferred_deletion_bytes = 3;- Parameters:
value- The deferredDeletionBytes to set.- Returns:
- This builder for chaining.
-
clearDeferredDeletionBytes
Delay deletion of up to this many bytes to reduce the number of interactions with gpu driver code. If 0, the system chooses a reasonable default (several MBs).
int64 deferred_deletion_bytes = 3;- Returns:
- This builder for chaining.
-
getVisibleDeviceList
A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices. For example, if TensorFlow can see 8 GPU devices in the process, and one wanted to map visible GPU devices 5 and 3 as "/device:GPU:0", and "/device:GPU:1", then one would specify this field as "5,3". This field is similar in spirit to the CUDA_VISIBLE_DEVICES environment variable, except it applies to the visible GPU devices in the process. NOTE: 1. The GPU driver provides the process with the visible GPUs in an order which is not guaranteed to have any correlation to the *physical* GPU id in the machine. This field is used for remapping "visible" to "virtual", which means this operates only after the process starts. Users are required to use vendor specific mechanisms (e.g., CUDA_VISIBLE_DEVICES) to control the physical to visible device mapping prior to invoking TensorFlow. 2. In the code, the ids in this list are also called "platform GPU id"s, and the 'virtual' ids of GPU devices (i.e. the ids in the device name "/device:GPU:<id>") are also called "TF GPU id"s. Please refer to third_party/tensorflow/core/common_runtime/gpu/gpu_id.h for more information. 3. The visible_device_list is also used for PluggableDevice. And different types of PluggableDevices share this field. In that case, the pluggable_device_type is used to distinguish them, making the visible_device_list a list of <pluggable_device_type>:<device_index>, e.g. "PluggableDeviceA:0,PluggableDeviceA:1,PluggableDeviceB:0".string visible_device_list = 5;- Specified by:
getVisibleDeviceListin interfaceGPUOptionsOrBuilder- Returns:
- The visibleDeviceList.
-
getVisibleDeviceListBytes
public com.google.protobuf.ByteString getVisibleDeviceListBytes()A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices. For example, if TensorFlow can see 8 GPU devices in the process, and one wanted to map visible GPU devices 5 and 3 as "/device:GPU:0", and "/device:GPU:1", then one would specify this field as "5,3". This field is similar in spirit to the CUDA_VISIBLE_DEVICES environment variable, except it applies to the visible GPU devices in the process. NOTE: 1. The GPU driver provides the process with the visible GPUs in an order which is not guaranteed to have any correlation to the *physical* GPU id in the machine. This field is used for remapping "visible" to "virtual", which means this operates only after the process starts. Users are required to use vendor specific mechanisms (e.g., CUDA_VISIBLE_DEVICES) to control the physical to visible device mapping prior to invoking TensorFlow. 2. In the code, the ids in this list are also called "platform GPU id"s, and the 'virtual' ids of GPU devices (i.e. the ids in the device name "/device:GPU:<id>") are also called "TF GPU id"s. Please refer to third_party/tensorflow/core/common_runtime/gpu/gpu_id.h for more information. 3. The visible_device_list is also used for PluggableDevice. And different types of PluggableDevices share this field. In that case, the pluggable_device_type is used to distinguish them, making the visible_device_list a list of <pluggable_device_type>:<device_index>, e.g. "PluggableDeviceA:0,PluggableDeviceA:1,PluggableDeviceB:0".string visible_device_list = 5;- Specified by:
getVisibleDeviceListBytesin interfaceGPUOptionsOrBuilder- Returns:
- The bytes for visibleDeviceList.
-
setVisibleDeviceList
A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices. For example, if TensorFlow can see 8 GPU devices in the process, and one wanted to map visible GPU devices 5 and 3 as "/device:GPU:0", and "/device:GPU:1", then one would specify this field as "5,3". This field is similar in spirit to the CUDA_VISIBLE_DEVICES environment variable, except it applies to the visible GPU devices in the process. NOTE: 1. The GPU driver provides the process with the visible GPUs in an order which is not guaranteed to have any correlation to the *physical* GPU id in the machine. This field is used for remapping "visible" to "virtual", which means this operates only after the process starts. Users are required to use vendor specific mechanisms (e.g., CUDA_VISIBLE_DEVICES) to control the physical to visible device mapping prior to invoking TensorFlow. 2. In the code, the ids in this list are also called "platform GPU id"s, and the 'virtual' ids of GPU devices (i.e. the ids in the device name "/device:GPU:<id>") are also called "TF GPU id"s. Please refer to third_party/tensorflow/core/common_runtime/gpu/gpu_id.h for more information. 3. The visible_device_list is also used for PluggableDevice. And different types of PluggableDevices share this field. In that case, the pluggable_device_type is used to distinguish them, making the visible_device_list a list of <pluggable_device_type>:<device_index>, e.g. "PluggableDeviceA:0,PluggableDeviceA:1,PluggableDeviceB:0".string visible_device_list = 5;- Parameters:
value- The visibleDeviceList to set.- Returns:
- This builder for chaining.
-
clearVisibleDeviceList
A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices. For example, if TensorFlow can see 8 GPU devices in the process, and one wanted to map visible GPU devices 5 and 3 as "/device:GPU:0", and "/device:GPU:1", then one would specify this field as "5,3". This field is similar in spirit to the CUDA_VISIBLE_DEVICES environment variable, except it applies to the visible GPU devices in the process. NOTE: 1. The GPU driver provides the process with the visible GPUs in an order which is not guaranteed to have any correlation to the *physical* GPU id in the machine. This field is used for remapping "visible" to "virtual", which means this operates only after the process starts. Users are required to use vendor specific mechanisms (e.g., CUDA_VISIBLE_DEVICES) to control the physical to visible device mapping prior to invoking TensorFlow. 2. In the code, the ids in this list are also called "platform GPU id"s, and the 'virtual' ids of GPU devices (i.e. the ids in the device name "/device:GPU:<id>") are also called "TF GPU id"s. Please refer to third_party/tensorflow/core/common_runtime/gpu/gpu_id.h for more information. 3. The visible_device_list is also used for PluggableDevice. And different types of PluggableDevices share this field. In that case, the pluggable_device_type is used to distinguish them, making the visible_device_list a list of <pluggable_device_type>:<device_index>, e.g. "PluggableDeviceA:0,PluggableDeviceA:1,PluggableDeviceB:0".string visible_device_list = 5;- Returns:
- This builder for chaining.
-
setVisibleDeviceListBytes
A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices. For example, if TensorFlow can see 8 GPU devices in the process, and one wanted to map visible GPU devices 5 and 3 as "/device:GPU:0", and "/device:GPU:1", then one would specify this field as "5,3". This field is similar in spirit to the CUDA_VISIBLE_DEVICES environment variable, except it applies to the visible GPU devices in the process. NOTE: 1. The GPU driver provides the process with the visible GPUs in an order which is not guaranteed to have any correlation to the *physical* GPU id in the machine. This field is used for remapping "visible" to "virtual", which means this operates only after the process starts. Users are required to use vendor specific mechanisms (e.g., CUDA_VISIBLE_DEVICES) to control the physical to visible device mapping prior to invoking TensorFlow. 2. In the code, the ids in this list are also called "platform GPU id"s, and the 'virtual' ids of GPU devices (i.e. the ids in the device name "/device:GPU:<id>") are also called "TF GPU id"s. Please refer to third_party/tensorflow/core/common_runtime/gpu/gpu_id.h for more information. 3. The visible_device_list is also used for PluggableDevice. And different types of PluggableDevices share this field. In that case, the pluggable_device_type is used to distinguish them, making the visible_device_list a list of <pluggable_device_type>:<device_index>, e.g. "PluggableDeviceA:0,PluggableDeviceA:1,PluggableDeviceB:0".string visible_device_list = 5;- Parameters:
value- The bytes for visibleDeviceList to set.- Returns:
- This builder for chaining.
-
getPollingActiveDelayUsecs
public int getPollingActiveDelayUsecs()In the event polling loop sleep this many microseconds between PollEvents calls, when the queue is not empty. If value is not set or set to 0, gets set to a non-zero default.
int32 polling_active_delay_usecs = 6;- Specified by:
getPollingActiveDelayUsecsin interfaceGPUOptionsOrBuilder- Returns:
- The pollingActiveDelayUsecs.
-
setPollingActiveDelayUsecs
In the event polling loop sleep this many microseconds between PollEvents calls, when the queue is not empty. If value is not set or set to 0, gets set to a non-zero default.
int32 polling_active_delay_usecs = 6;- Parameters:
value- The pollingActiveDelayUsecs to set.- Returns:
- This builder for chaining.
-
clearPollingActiveDelayUsecs
In the event polling loop sleep this many microseconds between PollEvents calls, when the queue is not empty. If value is not set or set to 0, gets set to a non-zero default.
int32 polling_active_delay_usecs = 6;- Returns:
- This builder for chaining.
-
getPollingInactiveDelayMsecs
public int getPollingInactiveDelayMsecs()This field is deprecated and ignored.
int32 polling_inactive_delay_msecs = 7;- Specified by:
getPollingInactiveDelayMsecsin interfaceGPUOptionsOrBuilder- Returns:
- The pollingInactiveDelayMsecs.
-
setPollingInactiveDelayMsecs
This field is deprecated and ignored.
int32 polling_inactive_delay_msecs = 7;- Parameters:
value- The pollingInactiveDelayMsecs to set.- Returns:
- This builder for chaining.
-
clearPollingInactiveDelayMsecs
This field is deprecated and ignored.
int32 polling_inactive_delay_msecs = 7;- Returns:
- This builder for chaining.
-
getForceGpuCompatible
public boolean getForceGpuCompatible()Force all tensors to be gpu_compatible. On a GPU-enabled TensorFlow, enabling this option forces all CPU tensors to be allocated with Cuda pinned memory. Normally, TensorFlow will infer which tensors should be allocated as the pinned memory. But in case where the inference is incomplete, this option can significantly speed up the cross-device memory copy performance as long as it fits the memory. Note that this option is not something that should be enabled by default for unknown or very large models, since all Cuda pinned memory is unpageable, having too much pinned memory might negatively impact the overall host system performance.
bool force_gpu_compatible = 8;- Specified by:
getForceGpuCompatiblein interfaceGPUOptionsOrBuilder- Returns:
- The forceGpuCompatible.
-
setForceGpuCompatible
Force all tensors to be gpu_compatible. On a GPU-enabled TensorFlow, enabling this option forces all CPU tensors to be allocated with Cuda pinned memory. Normally, TensorFlow will infer which tensors should be allocated as the pinned memory. But in case where the inference is incomplete, this option can significantly speed up the cross-device memory copy performance as long as it fits the memory. Note that this option is not something that should be enabled by default for unknown or very large models, since all Cuda pinned memory is unpageable, having too much pinned memory might negatively impact the overall host system performance.
bool force_gpu_compatible = 8;- Parameters:
value- The forceGpuCompatible to set.- Returns:
- This builder for chaining.
-
clearForceGpuCompatible
Force all tensors to be gpu_compatible. On a GPU-enabled TensorFlow, enabling this option forces all CPU tensors to be allocated with Cuda pinned memory. Normally, TensorFlow will infer which tensors should be allocated as the pinned memory. But in case where the inference is incomplete, this option can significantly speed up the cross-device memory copy performance as long as it fits the memory. Note that this option is not something that should be enabled by default for unknown or very large models, since all Cuda pinned memory is unpageable, having too much pinned memory might negatively impact the overall host system performance.
bool force_gpu_compatible = 8;- Returns:
- This builder for chaining.
-
hasExperimental
public boolean hasExperimental()Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;- Specified by:
hasExperimentalin interfaceGPUOptionsOrBuilder- Returns:
- Whether the experimental field is set.
-
getExperimental
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;- Specified by:
getExperimentalin interfaceGPUOptionsOrBuilder- Returns:
- The experimental.
-
setExperimental
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9; -
setExperimental
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9; -
mergeExperimental
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9; -
clearExperimental
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9; -
getExperimentalBuilder
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9; -
getExperimentalOrBuilder
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;- Specified by:
getExperimentalOrBuilderin interfaceGPUOptionsOrBuilder
-
setUnknownFields
- Specified by:
setUnknownFieldsin interfacecom.google.protobuf.Message.Builder- Overrides:
setUnknownFieldsin classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-
mergeUnknownFields
public final GPUOptions.Builder mergeUnknownFields(com.google.protobuf.UnknownFieldSet unknownFields) - Specified by:
mergeUnknownFieldsin interfacecom.google.protobuf.Message.Builder- Overrides:
mergeUnknownFieldsin classcom.google.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
-