public static final class GPUOptions.Builder extends org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder> implements GPUOptionsOrBuilder
tensorflow.GPUOptions
Modifier and Type | Method and Description |
---|---|
GPUOptions.Builder |
addRepeatedField(org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field,
Object value) |
GPUOptions |
build() |
GPUOptions |
buildPartial() |
GPUOptions.Builder |
clear() |
GPUOptions.Builder |
clearAllocatorType()
The type of GPU allocation strategy to use.
|
GPUOptions.Builder |
clearAllowGrowth()
If true, the allocator does not pre-allocate the entire specified
GPU memory region, instead starting small and growing as needed.
|
GPUOptions.Builder |
clearDeferredDeletionBytes()
Delay deletion of up to this many bytes to reduce the number of
interactions with gpu driver code.
|
GPUOptions.Builder |
clearExperimental()
Everything inside experimental is subject to change and is not subject
to API stability guarantees in
https://www.tensorflow.org/guide/version_compat.
|
GPUOptions.Builder |
clearField(org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field) |
GPUOptions.Builder |
clearForceGpuCompatible()
Force all tensors to be gpu_compatible.
|
GPUOptions.Builder |
clearOneof(org.nd4j.shade.protobuf.Descriptors.OneofDescriptor oneof) |
GPUOptions.Builder |
clearPerProcessGpuMemoryFraction()
Fraction of the available GPU memory to allocate for each process.
|
GPUOptions.Builder |
clearPollingActiveDelayUsecs()
In the event polling loop sleep this many microseconds between
PollEvents calls, when the queue is not empty.
|
GPUOptions.Builder |
clearPollingInactiveDelayMsecs()
This field is deprecated and ignored.
|
GPUOptions.Builder |
clearVisibleDeviceList()
A comma-separated list of GPU ids that determines the 'visible'
to 'virtual' mapping of GPU devices.
|
GPUOptions.Builder |
clone() |
String |
getAllocatorType()
The type of GPU allocation strategy to use.
|
org.nd4j.shade.protobuf.ByteString |
getAllocatorTypeBytes()
The type of GPU allocation strategy to use.
|
boolean |
getAllowGrowth()
If true, the allocator does not pre-allocate the entire specified
GPU memory region, instead starting small and growing as needed.
|
GPUOptions |
getDefaultInstanceForType() |
long |
getDeferredDeletionBytes()
Delay deletion of up to this many bytes to reduce the number of
interactions with gpu driver code.
|
static org.nd4j.shade.protobuf.Descriptors.Descriptor |
getDescriptor() |
org.nd4j.shade.protobuf.Descriptors.Descriptor |
getDescriptorForType() |
GPUOptions.Experimental |
getExperimental()
Everything inside experimental is subject to change and is not subject
to API stability guarantees in
https://www.tensorflow.org/guide/version_compat.
|
GPUOptions.Experimental.Builder |
getExperimentalBuilder()
Everything inside experimental is subject to change and is not subject
to API stability guarantees in
https://www.tensorflow.org/guide/version_compat.
|
GPUOptions.ExperimentalOrBuilder |
getExperimentalOrBuilder()
Everything inside experimental is subject to change and is not subject
to API stability guarantees in
https://www.tensorflow.org/guide/version_compat.
|
boolean |
getForceGpuCompatible()
Force all tensors to be gpu_compatible.
|
double |
getPerProcessGpuMemoryFraction()
Fraction of the available GPU memory to allocate for each process.
|
int |
getPollingActiveDelayUsecs()
In the event polling loop sleep this many microseconds between
PollEvents calls, when the queue is not empty.
|
int |
getPollingInactiveDelayMsecs()
This field is deprecated and ignored.
|
String |
getVisibleDeviceList()
A comma-separated list of GPU ids that determines the 'visible'
to 'virtual' mapping of GPU devices.
|
org.nd4j.shade.protobuf.ByteString |
getVisibleDeviceListBytes()
A comma-separated list of GPU ids that determines the 'visible'
to 'virtual' mapping of GPU devices.
|
boolean |
hasExperimental()
Everything inside experimental is subject to change and is not subject
to API stability guarantees in
https://www.tensorflow.org/guide/version_compat.
|
protected org.nd4j.shade.protobuf.GeneratedMessageV3.FieldAccessorTable |
internalGetFieldAccessorTable() |
boolean |
isInitialized() |
GPUOptions.Builder |
mergeExperimental(GPUOptions.Experimental value)
Everything inside experimental is subject to change and is not subject
to API stability guarantees in
https://www.tensorflow.org/guide/version_compat.
|
GPUOptions.Builder |
mergeFrom(org.nd4j.shade.protobuf.CodedInputStream input,
org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry) |
GPUOptions.Builder |
mergeFrom(GPUOptions other) |
GPUOptions.Builder |
mergeFrom(org.nd4j.shade.protobuf.Message other) |
GPUOptions.Builder |
mergeUnknownFields(org.nd4j.shade.protobuf.UnknownFieldSet unknownFields) |
GPUOptions.Builder |
setAllocatorType(String value)
The type of GPU allocation strategy to use.
|
GPUOptions.Builder |
setAllocatorTypeBytes(org.nd4j.shade.protobuf.ByteString value)
The type of GPU allocation strategy to use.
|
GPUOptions.Builder |
setAllowGrowth(boolean value)
If true, the allocator does not pre-allocate the entire specified
GPU memory region, instead starting small and growing as needed.
|
GPUOptions.Builder |
setDeferredDeletionBytes(long value)
Delay deletion of up to this many bytes to reduce the number of
interactions with gpu driver code.
|
GPUOptions.Builder |
setExperimental(GPUOptions.Experimental.Builder builderForValue)
Everything inside experimental is subject to change and is not subject
to API stability guarantees in
https://www.tensorflow.org/guide/version_compat.
|
GPUOptions.Builder |
setExperimental(GPUOptions.Experimental value)
Everything inside experimental is subject to change and is not subject
to API stability guarantees in
https://www.tensorflow.org/guide/version_compat.
|
GPUOptions.Builder |
setField(org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field,
Object value) |
GPUOptions.Builder |
setForceGpuCompatible(boolean value)
Force all tensors to be gpu_compatible.
|
GPUOptions.Builder |
setPerProcessGpuMemoryFraction(double value)
Fraction of the available GPU memory to allocate for each process.
|
GPUOptions.Builder |
setPollingActiveDelayUsecs(int value)
In the event polling loop sleep this many microseconds between
PollEvents calls, when the queue is not empty.
|
GPUOptions.Builder |
setPollingInactiveDelayMsecs(int value)
This field is deprecated and ignored.
|
GPUOptions.Builder |
setRepeatedField(org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field,
int index,
Object value) |
GPUOptions.Builder |
setUnknownFields(org.nd4j.shade.protobuf.UnknownFieldSet unknownFields) |
GPUOptions.Builder |
setVisibleDeviceList(String value)
A comma-separated list of GPU ids that determines the 'visible'
to 'virtual' mapping of GPU devices.
|
GPUOptions.Builder |
setVisibleDeviceListBytes(org.nd4j.shade.protobuf.ByteString value)
A comma-separated list of GPU ids that determines the 'visible'
to 'virtual' mapping of GPU devices.
|
getAllFields, getField, getFieldBuilder, getOneofFieldDescriptor, getParentForChildren, getRepeatedField, getRepeatedFieldBuilder, getRepeatedFieldCount, getUnknownFields, hasField, hasOneof, internalGetMapField, internalGetMutableMapField, isClean, markClean, newBuilderForField, onBuilt, onChanged, setUnknownFieldsProto3
findInitializationErrors, getInitializationErrorString, internalMergeFrom, mergeDelimitedFrom, mergeDelimitedFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, newUninitializedMessageException, toString
addAll, addAll, mergeFrom, newUninitializedMessageException
equals, finalize, getClass, hashCode, notify, notifyAll, wait, wait, wait
public static final org.nd4j.shade.protobuf.Descriptors.Descriptor getDescriptor()
protected org.nd4j.shade.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
internalGetFieldAccessorTable
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public GPUOptions.Builder clear()
clear
in interface org.nd4j.shade.protobuf.Message.Builder
clear
in interface org.nd4j.shade.protobuf.MessageLite.Builder
clear
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public org.nd4j.shade.protobuf.Descriptors.Descriptor getDescriptorForType()
getDescriptorForType
in interface org.nd4j.shade.protobuf.Message.Builder
getDescriptorForType
in interface org.nd4j.shade.protobuf.MessageOrBuilder
getDescriptorForType
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public GPUOptions getDefaultInstanceForType()
getDefaultInstanceForType
in interface org.nd4j.shade.protobuf.MessageLiteOrBuilder
getDefaultInstanceForType
in interface org.nd4j.shade.protobuf.MessageOrBuilder
public GPUOptions build()
build
in interface org.nd4j.shade.protobuf.Message.Builder
build
in interface org.nd4j.shade.protobuf.MessageLite.Builder
public GPUOptions buildPartial()
buildPartial
in interface org.nd4j.shade.protobuf.Message.Builder
buildPartial
in interface org.nd4j.shade.protobuf.MessageLite.Builder
public GPUOptions.Builder clone()
clone
in interface org.nd4j.shade.protobuf.Message.Builder
clone
in interface org.nd4j.shade.protobuf.MessageLite.Builder
clone
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public GPUOptions.Builder setField(org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field, Object value)
setField
in interface org.nd4j.shade.protobuf.Message.Builder
setField
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public GPUOptions.Builder clearField(org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field)
clearField
in interface org.nd4j.shade.protobuf.Message.Builder
clearField
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public GPUOptions.Builder clearOneof(org.nd4j.shade.protobuf.Descriptors.OneofDescriptor oneof)
clearOneof
in interface org.nd4j.shade.protobuf.Message.Builder
clearOneof
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public GPUOptions.Builder setRepeatedField(org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field, int index, Object value)
setRepeatedField
in interface org.nd4j.shade.protobuf.Message.Builder
setRepeatedField
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public GPUOptions.Builder addRepeatedField(org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field, Object value)
addRepeatedField
in interface org.nd4j.shade.protobuf.Message.Builder
addRepeatedField
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public GPUOptions.Builder mergeFrom(org.nd4j.shade.protobuf.Message other)
mergeFrom
in interface org.nd4j.shade.protobuf.Message.Builder
mergeFrom
in class org.nd4j.shade.protobuf.AbstractMessage.Builder<GPUOptions.Builder>
public GPUOptions.Builder mergeFrom(GPUOptions other)
public final boolean isInitialized()
isInitialized
in interface org.nd4j.shade.protobuf.MessageLiteOrBuilder
isInitialized
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public GPUOptions.Builder mergeFrom(org.nd4j.shade.protobuf.CodedInputStream input, org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry) throws IOException
mergeFrom
in interface org.nd4j.shade.protobuf.Message.Builder
mergeFrom
in interface org.nd4j.shade.protobuf.MessageLite.Builder
mergeFrom
in class org.nd4j.shade.protobuf.AbstractMessage.Builder<GPUOptions.Builder>
IOException
public double getPerProcessGpuMemoryFraction()
Fraction of the available GPU memory to allocate for each process. 1 means to allocate all of the GPU memory, 0.5 means the process allocates up to ~50% of the available GPU memory. GPU memory is pre-allocated unless the allow_growth option is enabled. If greater than 1.0, uses CUDA unified memory to potentially oversubscribe the amount of memory available on the GPU device by using host memory as a swap space. Accessing memory not available on the device will be significantly slower as that would require memory transfer between the host and the device. Options to reduce the memory requirement should be considered before enabling this option as this may come with a negative performance impact. Oversubscription using the unified memory requires Pascal class or newer GPUs and it is currently only supported on the Linux operating system. See https://docs.nvidia.com/cuda/cuda-c-programming-guide/index.html#um-requirements for the detailed requirements.
double per_process_gpu_memory_fraction = 1;
getPerProcessGpuMemoryFraction
in interface GPUOptionsOrBuilder
public GPUOptions.Builder setPerProcessGpuMemoryFraction(double value)
Fraction of the available GPU memory to allocate for each process. 1 means to allocate all of the GPU memory, 0.5 means the process allocates up to ~50% of the available GPU memory. GPU memory is pre-allocated unless the allow_growth option is enabled. If greater than 1.0, uses CUDA unified memory to potentially oversubscribe the amount of memory available on the GPU device by using host memory as a swap space. Accessing memory not available on the device will be significantly slower as that would require memory transfer between the host and the device. Options to reduce the memory requirement should be considered before enabling this option as this may come with a negative performance impact. Oversubscription using the unified memory requires Pascal class or newer GPUs and it is currently only supported on the Linux operating system. See https://docs.nvidia.com/cuda/cuda-c-programming-guide/index.html#um-requirements for the detailed requirements.
double per_process_gpu_memory_fraction = 1;
public GPUOptions.Builder clearPerProcessGpuMemoryFraction()
Fraction of the available GPU memory to allocate for each process. 1 means to allocate all of the GPU memory, 0.5 means the process allocates up to ~50% of the available GPU memory. GPU memory is pre-allocated unless the allow_growth option is enabled. If greater than 1.0, uses CUDA unified memory to potentially oversubscribe the amount of memory available on the GPU device by using host memory as a swap space. Accessing memory not available on the device will be significantly slower as that would require memory transfer between the host and the device. Options to reduce the memory requirement should be considered before enabling this option as this may come with a negative performance impact. Oversubscription using the unified memory requires Pascal class or newer GPUs and it is currently only supported on the Linux operating system. See https://docs.nvidia.com/cuda/cuda-c-programming-guide/index.html#um-requirements for the detailed requirements.
double per_process_gpu_memory_fraction = 1;
public boolean getAllowGrowth()
If true, the allocator does not pre-allocate the entire specified GPU memory region, instead starting small and growing as needed.
bool allow_growth = 4;
getAllowGrowth
in interface GPUOptionsOrBuilder
public GPUOptions.Builder setAllowGrowth(boolean value)
If true, the allocator does not pre-allocate the entire specified GPU memory region, instead starting small and growing as needed.
bool allow_growth = 4;
public GPUOptions.Builder clearAllowGrowth()
If true, the allocator does not pre-allocate the entire specified GPU memory region, instead starting small and growing as needed.
bool allow_growth = 4;
public String getAllocatorType()
The type of GPU allocation strategy to use. Allowed values: "": The empty string (default) uses a system-chosen default which may change over time. "BFC": A "Best-fit with coalescing" algorithm, simplified from a version of dlmalloc.
string allocator_type = 2;
getAllocatorType
in interface GPUOptionsOrBuilder
public org.nd4j.shade.protobuf.ByteString getAllocatorTypeBytes()
The type of GPU allocation strategy to use. Allowed values: "": The empty string (default) uses a system-chosen default which may change over time. "BFC": A "Best-fit with coalescing" algorithm, simplified from a version of dlmalloc.
string allocator_type = 2;
getAllocatorTypeBytes
in interface GPUOptionsOrBuilder
public GPUOptions.Builder setAllocatorType(String value)
The type of GPU allocation strategy to use. Allowed values: "": The empty string (default) uses a system-chosen default which may change over time. "BFC": A "Best-fit with coalescing" algorithm, simplified from a version of dlmalloc.
string allocator_type = 2;
public GPUOptions.Builder clearAllocatorType()
The type of GPU allocation strategy to use. Allowed values: "": The empty string (default) uses a system-chosen default which may change over time. "BFC": A "Best-fit with coalescing" algorithm, simplified from a version of dlmalloc.
string allocator_type = 2;
public GPUOptions.Builder setAllocatorTypeBytes(org.nd4j.shade.protobuf.ByteString value)
The type of GPU allocation strategy to use. Allowed values: "": The empty string (default) uses a system-chosen default which may change over time. "BFC": A "Best-fit with coalescing" algorithm, simplified from a version of dlmalloc.
string allocator_type = 2;
public long getDeferredDeletionBytes()
Delay deletion of up to this many bytes to reduce the number of interactions with gpu driver code. If 0, the system chooses a reasonable default (several MBs).
int64 deferred_deletion_bytes = 3;
getDeferredDeletionBytes
in interface GPUOptionsOrBuilder
public GPUOptions.Builder setDeferredDeletionBytes(long value)
Delay deletion of up to this many bytes to reduce the number of interactions with gpu driver code. If 0, the system chooses a reasonable default (several MBs).
int64 deferred_deletion_bytes = 3;
public GPUOptions.Builder clearDeferredDeletionBytes()
Delay deletion of up to this many bytes to reduce the number of interactions with gpu driver code. If 0, the system chooses a reasonable default (several MBs).
int64 deferred_deletion_bytes = 3;
public String getVisibleDeviceList()
A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices. For example, if TensorFlow can see 8 GPU devices in the process, and one wanted to map visible GPU devices 5 and 3 as "/device:GPU:0", and "/device:GPU:1", then one would specify this field as "5,3". This field is similar in spirit to the CUDA_VISIBLE_DEVICES environment variable, except it applies to the visible GPU devices in the process. NOTE: 1. The GPU driver provides the process with the visible GPUs in an order which is not guaranteed to have any correlation to the *physical* GPU id in the machine. This field is used for remapping "visible" to "virtual", which means this operates only after the process starts. Users are required to use vendor specific mechanisms (e.g., CUDA_VISIBLE_DEVICES) to control the physical to visible device mapping prior to invoking TensorFlow. 2. In the code, the ids in this list are also called "CUDA GPU id"s, and the 'virtual' ids of GPU devices (i.e. the ids in the device name "/device:GPU:<id>") are also called "TF GPU id"s. Please refer to third_party/tensorflow/core/common_runtime/gpu/gpu_id.h for more information.
string visible_device_list = 5;
getVisibleDeviceList
in interface GPUOptionsOrBuilder
public org.nd4j.shade.protobuf.ByteString getVisibleDeviceListBytes()
A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices. For example, if TensorFlow can see 8 GPU devices in the process, and one wanted to map visible GPU devices 5 and 3 as "/device:GPU:0", and "/device:GPU:1", then one would specify this field as "5,3". This field is similar in spirit to the CUDA_VISIBLE_DEVICES environment variable, except it applies to the visible GPU devices in the process. NOTE: 1. The GPU driver provides the process with the visible GPUs in an order which is not guaranteed to have any correlation to the *physical* GPU id in the machine. This field is used for remapping "visible" to "virtual", which means this operates only after the process starts. Users are required to use vendor specific mechanisms (e.g., CUDA_VISIBLE_DEVICES) to control the physical to visible device mapping prior to invoking TensorFlow. 2. In the code, the ids in this list are also called "CUDA GPU id"s, and the 'virtual' ids of GPU devices (i.e. the ids in the device name "/device:GPU:<id>") are also called "TF GPU id"s. Please refer to third_party/tensorflow/core/common_runtime/gpu/gpu_id.h for more information.
string visible_device_list = 5;
getVisibleDeviceListBytes
in interface GPUOptionsOrBuilder
public GPUOptions.Builder setVisibleDeviceList(String value)
A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices. For example, if TensorFlow can see 8 GPU devices in the process, and one wanted to map visible GPU devices 5 and 3 as "/device:GPU:0", and "/device:GPU:1", then one would specify this field as "5,3". This field is similar in spirit to the CUDA_VISIBLE_DEVICES environment variable, except it applies to the visible GPU devices in the process. NOTE: 1. The GPU driver provides the process with the visible GPUs in an order which is not guaranteed to have any correlation to the *physical* GPU id in the machine. This field is used for remapping "visible" to "virtual", which means this operates only after the process starts. Users are required to use vendor specific mechanisms (e.g., CUDA_VISIBLE_DEVICES) to control the physical to visible device mapping prior to invoking TensorFlow. 2. In the code, the ids in this list are also called "CUDA GPU id"s, and the 'virtual' ids of GPU devices (i.e. the ids in the device name "/device:GPU:<id>") are also called "TF GPU id"s. Please refer to third_party/tensorflow/core/common_runtime/gpu/gpu_id.h for more information.
string visible_device_list = 5;
public GPUOptions.Builder clearVisibleDeviceList()
A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices. For example, if TensorFlow can see 8 GPU devices in the process, and one wanted to map visible GPU devices 5 and 3 as "/device:GPU:0", and "/device:GPU:1", then one would specify this field as "5,3". This field is similar in spirit to the CUDA_VISIBLE_DEVICES environment variable, except it applies to the visible GPU devices in the process. NOTE: 1. The GPU driver provides the process with the visible GPUs in an order which is not guaranteed to have any correlation to the *physical* GPU id in the machine. This field is used for remapping "visible" to "virtual", which means this operates only after the process starts. Users are required to use vendor specific mechanisms (e.g., CUDA_VISIBLE_DEVICES) to control the physical to visible device mapping prior to invoking TensorFlow. 2. In the code, the ids in this list are also called "CUDA GPU id"s, and the 'virtual' ids of GPU devices (i.e. the ids in the device name "/device:GPU:<id>") are also called "TF GPU id"s. Please refer to third_party/tensorflow/core/common_runtime/gpu/gpu_id.h for more information.
string visible_device_list = 5;
public GPUOptions.Builder setVisibleDeviceListBytes(org.nd4j.shade.protobuf.ByteString value)
A comma-separated list of GPU ids that determines the 'visible' to 'virtual' mapping of GPU devices. For example, if TensorFlow can see 8 GPU devices in the process, and one wanted to map visible GPU devices 5 and 3 as "/device:GPU:0", and "/device:GPU:1", then one would specify this field as "5,3". This field is similar in spirit to the CUDA_VISIBLE_DEVICES environment variable, except it applies to the visible GPU devices in the process. NOTE: 1. The GPU driver provides the process with the visible GPUs in an order which is not guaranteed to have any correlation to the *physical* GPU id in the machine. This field is used for remapping "visible" to "virtual", which means this operates only after the process starts. Users are required to use vendor specific mechanisms (e.g., CUDA_VISIBLE_DEVICES) to control the physical to visible device mapping prior to invoking TensorFlow. 2. In the code, the ids in this list are also called "CUDA GPU id"s, and the 'virtual' ids of GPU devices (i.e. the ids in the device name "/device:GPU:<id>") are also called "TF GPU id"s. Please refer to third_party/tensorflow/core/common_runtime/gpu/gpu_id.h for more information.
string visible_device_list = 5;
public int getPollingActiveDelayUsecs()
In the event polling loop sleep this many microseconds between PollEvents calls, when the queue is not empty. If value is not set or set to 0, gets set to a non-zero default.
int32 polling_active_delay_usecs = 6;
getPollingActiveDelayUsecs
in interface GPUOptionsOrBuilder
public GPUOptions.Builder setPollingActiveDelayUsecs(int value)
In the event polling loop sleep this many microseconds between PollEvents calls, when the queue is not empty. If value is not set or set to 0, gets set to a non-zero default.
int32 polling_active_delay_usecs = 6;
public GPUOptions.Builder clearPollingActiveDelayUsecs()
In the event polling loop sleep this many microseconds between PollEvents calls, when the queue is not empty. If value is not set or set to 0, gets set to a non-zero default.
int32 polling_active_delay_usecs = 6;
public int getPollingInactiveDelayMsecs()
This field is deprecated and ignored.
int32 polling_inactive_delay_msecs = 7;
getPollingInactiveDelayMsecs
in interface GPUOptionsOrBuilder
public GPUOptions.Builder setPollingInactiveDelayMsecs(int value)
This field is deprecated and ignored.
int32 polling_inactive_delay_msecs = 7;
public GPUOptions.Builder clearPollingInactiveDelayMsecs()
This field is deprecated and ignored.
int32 polling_inactive_delay_msecs = 7;
public boolean getForceGpuCompatible()
Force all tensors to be gpu_compatible. On a GPU-enabled TensorFlow, enabling this option forces all CPU tensors to be allocated with Cuda pinned memory. Normally, TensorFlow will infer which tensors should be allocated as the pinned memory. But in case where the inference is incomplete, this option can significantly speed up the cross-device memory copy performance as long as it fits the memory. Note that this option is not something that should be enabled by default for unknown or very large models, since all Cuda pinned memory is unpageable, having too much pinned memory might negatively impact the overall host system performance.
bool force_gpu_compatible = 8;
getForceGpuCompatible
in interface GPUOptionsOrBuilder
public GPUOptions.Builder setForceGpuCompatible(boolean value)
Force all tensors to be gpu_compatible. On a GPU-enabled TensorFlow, enabling this option forces all CPU tensors to be allocated with Cuda pinned memory. Normally, TensorFlow will infer which tensors should be allocated as the pinned memory. But in case where the inference is incomplete, this option can significantly speed up the cross-device memory copy performance as long as it fits the memory. Note that this option is not something that should be enabled by default for unknown or very large models, since all Cuda pinned memory is unpageable, having too much pinned memory might negatively impact the overall host system performance.
bool force_gpu_compatible = 8;
public GPUOptions.Builder clearForceGpuCompatible()
Force all tensors to be gpu_compatible. On a GPU-enabled TensorFlow, enabling this option forces all CPU tensors to be allocated with Cuda pinned memory. Normally, TensorFlow will infer which tensors should be allocated as the pinned memory. But in case where the inference is incomplete, this option can significantly speed up the cross-device memory copy performance as long as it fits the memory. Note that this option is not something that should be enabled by default for unknown or very large models, since all Cuda pinned memory is unpageable, having too much pinned memory might negatively impact the overall host system performance.
bool force_gpu_compatible = 8;
public boolean hasExperimental()
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;
hasExperimental
in interface GPUOptionsOrBuilder
public GPUOptions.Experimental getExperimental()
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;
getExperimental
in interface GPUOptionsOrBuilder
public GPUOptions.Builder setExperimental(GPUOptions.Experimental value)
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;
public GPUOptions.Builder setExperimental(GPUOptions.Experimental.Builder builderForValue)
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;
public GPUOptions.Builder mergeExperimental(GPUOptions.Experimental value)
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;
public GPUOptions.Builder clearExperimental()
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;
public GPUOptions.Experimental.Builder getExperimentalBuilder()
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;
public GPUOptions.ExperimentalOrBuilder getExperimentalOrBuilder()
Everything inside experimental is subject to change and is not subject to API stability guarantees in https://www.tensorflow.org/guide/version_compat.
.tensorflow.GPUOptions.Experimental experimental = 9;
getExperimentalOrBuilder
in interface GPUOptionsOrBuilder
public final GPUOptions.Builder setUnknownFields(org.nd4j.shade.protobuf.UnknownFieldSet unknownFields)
setUnknownFields
in interface org.nd4j.shade.protobuf.Message.Builder
setUnknownFields
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
public final GPUOptions.Builder mergeUnknownFields(org.nd4j.shade.protobuf.UnknownFieldSet unknownFields)
mergeUnknownFields
in interface org.nd4j.shade.protobuf.Message.Builder
mergeUnknownFields
in class org.nd4j.shade.protobuf.GeneratedMessageV3.Builder<GPUOptions.Builder>
Copyright © 2020. All rights reserved.