ConfigProtoOrBuilder

antarmuka publik ConfigProtoOrBuilder
Subkelas Tidak Langsung yang Diketahui

Metode Publik

boolean abstrak
berisiDeviceCount (kunci string)
 Map from device type name (e.g., "CPU" or "GPU" ) to maximum
 number of devices of that type to use.
boolean abstrak
dapatkanAllowSoftPlacement ()
 Whether soft placement is allowed.
ClusterDef abstrak
dapatkanClusterDef ()
 Optional list of all workers to use in this session.
abstrak ClusterDefOrBuilder
dapatkanClusterDefOrBuilder ()
 Optional list of all workers to use in this session.
Peta abstrak<String, Integer>
dapatkan Jumlah Perangkat ()
Gunakan getDeviceCountMap() sebagai gantinya.
abstrak ke dalam
dapatkanDeviceCountCount ()
 Map from device type name (e.g., "CPU" or "GPU" ) to maximum
 number of devices of that type to use.
Peta abstrak<String, Integer>
dapatkanDeviceCountMap ()
 Map from device type name (e.g., "CPU" or "GPU" ) to maximum
 number of devices of that type to use.
abstrak ke dalam
getDeviceCountOrDefault (kunci string, int defaultValue)
 Map from device type name (e.g., "CPU" or "GPU" ) to maximum
 number of devices of that type to use.
abstrak ke dalam
getDeviceCountOrThrow (kunci string)
 Map from device type name (e.g., "CPU" or "GPU" ) to maximum
 number of devices of that type to use.
Tali abstrak
getDeviceFilters (indeks int)
 When any filters are present sessions will ignore all devices which do not
 match the filters.
abstrak com.google.protobuf.ByteString
getDeviceFiltersBytes (indeks int)
 When any filters are present sessions will ignore all devices which do not
 match the filters.
abstrak ke dalam
dapatkanDeviceFiltersCount ()
 When any filters are present sessions will ignore all devices which do not
 match the filters.
Daftar abstrak<String>
dapatkanDaftarFilterPerangkat ()
 When any filters are present sessions will ignore all devices which do not
 match the filters.
abstrak ConfigProto.Eksperimental
dapatkan Eksperimental ()
.tensorflow.ConfigProto.Experimental experimental = 16;
abstrak ConfigProto.ExperimentalOrBuilder
dapatkanExperimentalOrBuilder ()
.tensorflow.ConfigProto.Experimental experimental = 16;
Opsi GPU abstrak
dapatkanGpuOptions ()
 Options that apply to all GPUs.
abstrak GPUOptionsOrBuilder
dapatkanGpuOptionsOrBuilder ()
 Options that apply to all GPUs.
GraphOptions abstrak
dapatkanGraphOptions ()
 Options that apply to all graphs.
abstrak GraphOptionsOrBuilder
dapatkanGraphOptionsOrBuilder ()
 Options that apply to all graphs.
abstrak ke dalam
dapatkanInterOpParallelismThreads ()
 Nodes that perform blocking operations are enqueued on a pool of
 inter_op_parallelism_threads available in each process.
abstrak ke dalam
dapatkanIntraOpParallelismThreads ()
 The execution of an individual op (for some op types) can be
 parallelized on a pool of intra_op_parallelism_threads.
boolean abstrak
dapatkanIsolateSessionState ()
 If true, any resources such as Variables used in the session will not be
 shared with other sessions.
boolean abstrak
dapatkanLogDevicePlacement ()
 Whether device placements should be logged.
abstrak panjang
dapatkanOperationTimeoutInMs ()
 Global timeout for all blocking operations in this session.
abstrak ke dalam
dapatkanPeriode Penempatan ()
 Assignment of Nodes to Devices is recomputed every placement_period
 steps until the system warms up (at which point the recomputation
 typically slows down automatically).
Opsi RPCO abstrak
dapatkanRpcOptions ()
 Options that apply when this session uses the distributed runtime.
abstrak RPCOptionsOrBuilder
dapatkanRpcOptionsOrBuilder ()
 Options that apply when this session uses the distributed runtime.
abstrak ThreadPoolOptionProto
getSessionInterOpThreadPool (indeks int)
 This option is experimental - it may be replaced with a different mechanism
 in the future.
abstrak ke dalam
getSessionInterOpThreadPoolCount ()
 This option is experimental - it may be replaced with a different mechanism
 in the future.
Daftar abstrak< ThreadPoolOptionProto >
dapatkanSessionInterOpThreadPoolList ()
 This option is experimental - it may be replaced with a different mechanism
 in the future.
abstrak ThreadPoolOptionProtoOrBuilder
getSessionInterOpThreadPoolOrBuilder (indeks int)
 This option is experimental - it may be replaced with a different mechanism
 in the future.
Daftar abstrak<? memperluas ThreadPoolOptionProtoOrBuilder >
getSessionInterOpThreadPoolOrBuilderList ()
 This option is experimental - it may be replaced with a different mechanism
 in the future.
boolean abstrak
dapatkanShareClusterDevicesInSession ()
 When true, WorkerSessions are created with device attributes from the
 full cluster.
boolean abstrak
dapatkanUsePerSessionThreads ()
 If true, use a new set of threads for this session rather than the global
 pool of threads.
boolean abstrak
hasClusterDef ()
 Optional list of all workers to use in this session.
boolean abstrak
memiliki Eksperimental ()
.tensorflow.ConfigProto.Experimental experimental = 16;
boolean abstrak
memilikiGpuOptions ()
 Options that apply to all GPUs.
boolean abstrak
hasGraphOptions ()
 Options that apply to all graphs.
boolean abstrak
memilikiRpcOptions ()
 Options that apply when this session uses the distributed runtime.

Metode Publik

boolean abstrak publik berisiDeviceCount (kunci string)

 Map from device type name (e.g., "CPU" or "GPU" ) to maximum
 number of devices of that type to use.  If a particular device
 type is not found in the map, the system picks an appropriate
 number.
 
map<string, int32> device_count = 1;

boolean abstrak publik getAllowSoftPlacement ()

 Whether soft placement is allowed. If allow_soft_placement is true,
 an op will be placed on CPU if
   1. there's no GPU implementation for the OP
 or
   2. no GPU devices are known or registered
 or
   3. need to co-locate with reftype input(s) which are from CPU.
 
bool allow_soft_placement = 7;

abstrak publik ClusterDef getClusterDef ()

 Optional list of all workers to use in this session.
 
.tensorflow.ClusterDef cluster_def = 14;

abstrak publik ClusterDefOrBuilder getClusterDefOrBuilder ()

 Optional list of all workers to use in this session.
 
.tensorflow.ClusterDef cluster_def = 14;

peta abstrak publik<String, Integer> getDeviceCount ()

Gunakan getDeviceCountMap() sebagai gantinya.

abstrak publik int getDeviceCountCount ()

 Map from device type name (e.g., "CPU" or "GPU" ) to maximum
 number of devices of that type to use.  If a particular device
 type is not found in the map, the system picks an appropriate
 number.
 
map<string, int32> device_count = 1;

Peta abstrak publik<String, Integer> getDeviceCountMap ()

 Map from device type name (e.g., "CPU" or "GPU" ) to maximum
 number of devices of that type to use.  If a particular device
 type is not found in the map, the system picks an appropriate
 number.
 
map<string, int32> device_count = 1;

abstrak publik int getDeviceCountOrDefault (kunci string, int defaultValue)

 Map from device type name (e.g., "CPU" or "GPU" ) to maximum
 number of devices of that type to use.  If a particular device
 type is not found in the map, the system picks an appropriate
 number.
 
map<string, int32> device_count = 1;

abstrak publik int getDeviceCountOrThrow (kunci string)

 Map from device type name (e.g., "CPU" or "GPU" ) to maximum
 number of devices of that type to use.  If a particular device
 type is not found in the map, the system picks an appropriate
 number.
 
map<string, int32> device_count = 1;

String abstrak publik getDeviceFilters (indeks int)

 When any filters are present sessions will ignore all devices which do not
 match the filters. Each filter can be partially specified, e.g. "/job:ps"
 "/job:worker/replica:3", etc.
 
repeated string device_filters = 4;

abstrak publik com.google.protobuf.ByteString getDeviceFiltersBytes (indeks int)

 When any filters are present sessions will ignore all devices which do not
 match the filters. Each filter can be partially specified, e.g. "/job:ps"
 "/job:worker/replica:3", etc.
 
repeated string device_filters = 4;

abstrak publik int getDeviceFiltersCount ()

 When any filters are present sessions will ignore all devices which do not
 match the filters. Each filter can be partially specified, e.g. "/job:ps"
 "/job:worker/replica:3", etc.
 
repeated string device_filters = 4;

Daftar abstrak publik<String> getDeviceFiltersList ()

 When any filters are present sessions will ignore all devices which do not
 match the filters. Each filter can be partially specified, e.g. "/job:ps"
 "/job:worker/replica:3", etc.
 
repeated string device_filters = 4;

abstrak publik ConfigProto.Experimental getExperimental ()

.tensorflow.ConfigProto.Experimental experimental = 16;

abstrak publik ConfigProto.ExperimentalOrBuilder getExperimentalOrBuilder ()

.tensorflow.ConfigProto.Experimental experimental = 16;

GPUOptions abstrak publik getGpuOptions ()

 Options that apply to all GPUs.
 
.tensorflow.GPUOptions gpu_options = 6;

GPUOptionsOrBuilder abstrak publik getGpuOptionsOrBuilder ()

 Options that apply to all GPUs.
 
.tensorflow.GPUOptions gpu_options = 6;

GraphOptions abstrak publik getGraphOptions ()

 Options that apply to all graphs.
 
.tensorflow.GraphOptions graph_options = 10;

abstrak publik GraphOptionsOrBuilder getGraphOptionsOrBuilder ()

 Options that apply to all graphs.
 
.tensorflow.GraphOptions graph_options = 10;

abstrak publik int getInterOpParallelismThreads ()

 Nodes that perform blocking operations are enqueued on a pool of
 inter_op_parallelism_threads available in each process.
 0 means the system picks an appropriate number.
 Negative means all operations are performed in caller's thread.
 Note that the first Session created in the process sets the
 number of threads for all future sessions unless use_per_session_threads is
 true or session_inter_op_thread_pool is configured.
 
int32 inter_op_parallelism_threads = 5;

abstrak publik int getIntraOpParallelismThreads ()

 The execution of an individual op (for some op types) can be
 parallelized on a pool of intra_op_parallelism_threads.
 0 means the system picks an appropriate number.
 If you create an ordinary session, e.g., from Python or C++,
 then there is exactly one intra op thread pool per process.
 The first session created determines the number of threads in this pool.
 All subsequent sessions reuse/share this one global pool.
 There are notable exceptions to the default behavior describe above:
 1. There is an environment variable  for overriding this thread pool,
    named TF_OVERRIDE_GLOBAL_THREADPOOL.
 2. When connecting to a server, such as a remote `tf.train.Server`
    instance, then this option will be ignored altogether.
 
int32 intra_op_parallelism_threads = 2;

boolean abstrak publik getIsolateSessionState ()

 If true, any resources such as Variables used in the session will not be
 shared with other sessions. However, when clusterspec propagation is
 enabled, this field is ignored and sessions are always isolated.
 
bool isolate_session_state = 15;

boolean abstrak publik getLogDevicePlacement ()

 Whether device placements should be logged.
 
bool log_device_placement = 8;

abstrak publik getOperationTimeoutInMs panjang ()

 Global timeout for all blocking operations in this session.  If non-zero,
 and not overridden on a per-operation basis, this value will be used as the
 deadline for all blocking operations.
 
int64 operation_timeout_in_ms = 11;

abstrak publik int getPlacementPeriod ()

 Assignment of Nodes to Devices is recomputed every placement_period
 steps until the system warms up (at which point the recomputation
 typically slows down automatically).
 
int32 placement_period = 3;

RPCOptions abstrak publik getRpcOptions ()

 Options that apply when this session uses the distributed runtime.
 
.tensorflow.RPCOptions rpc_options = 13;

abstrak publik RPCOptionsOrBuilder getRpcOptionsOrBuilder ()

 Options that apply when this session uses the distributed runtime.
 
.tensorflow.RPCOptions rpc_options = 13;

abstrak publik ThreadPoolOptionProto getSessionInterOpThreadPool (int indeks)

 This option is experimental - it may be replaced with a different mechanism
 in the future.
 Configures session thread pools. If this is configured, then RunOptions for
 a Run call can select the thread pool to use.
 The intended use is for when some session invocations need to run in a
 background pool limited to a small number of threads:
 - For example, a session may be configured to have one large pool (for
 regular compute) and one small pool (for periodic, low priority work);
 using the small pool is currently the mechanism for limiting the inter-op
 parallelism of the low priority work.  Note that it does not limit the
 parallelism of work spawned by a single op kernel implementation.
 - Using this setting is normally not needed in training, but may help some
 serving use cases.
 - It is also generally recommended to set the global_name field of this
 proto, to avoid creating multiple large pools. It is typically better to
 run the non-low-priority work, even across sessions, in a single large
 pool.
 
repeated .tensorflow.ThreadPoolOptionProto session_inter_op_thread_pool = 12;

abstrak publik int getSessionInterOpThreadPoolCount ()

 This option is experimental - it may be replaced with a different mechanism
 in the future.
 Configures session thread pools. If this is configured, then RunOptions for
 a Run call can select the thread pool to use.
 The intended use is for when some session invocations need to run in a
 background pool limited to a small number of threads:
 - For example, a session may be configured to have one large pool (for
 regular compute) and one small pool (for periodic, low priority work);
 using the small pool is currently the mechanism for limiting the inter-op
 parallelism of the low priority work.  Note that it does not limit the
 parallelism of work spawned by a single op kernel implementation.
 - Using this setting is normally not needed in training, but may help some
 serving use cases.
 - It is also generally recommended to set the global_name field of this
 proto, to avoid creating multiple large pools. It is typically better to
 run the non-low-priority work, even across sessions, in a single large
 pool.
 
repeated .tensorflow.ThreadPoolOptionProto session_inter_op_thread_pool = 12;

Daftar abstrak publik< ThreadPoolOptionProto > getSessionInterOpThreadPoolList ()

 This option is experimental - it may be replaced with a different mechanism
 in the future.
 Configures session thread pools. If this is configured, then RunOptions for
 a Run call can select the thread pool to use.
 The intended use is for when some session invocations need to run in a
 background pool limited to a small number of threads:
 - For example, a session may be configured to have one large pool (for
 regular compute) and one small pool (for periodic, low priority work);
 using the small pool is currently the mechanism for limiting the inter-op
 parallelism of the low priority work.  Note that it does not limit the
 parallelism of work spawned by a single op kernel implementation.
 - Using this setting is normally not needed in training, but may help some
 serving use cases.
 - It is also generally recommended to set the global_name field of this
 proto, to avoid creating multiple large pools. It is typically better to
 run the non-low-priority work, even across sessions, in a single large
 pool.
 
repeated .tensorflow.ThreadPoolOptionProto session_inter_op_thread_pool = 12;

abstrak publik ThreadPoolOptionProtoOrBuilder getSessionInterOpThreadPoolOrBuilder (int indeks)

 This option is experimental - it may be replaced with a different mechanism
 in the future.
 Configures session thread pools. If this is configured, then RunOptions for
 a Run call can select the thread pool to use.
 The intended use is for when some session invocations need to run in a
 background pool limited to a small number of threads:
 - For example, a session may be configured to have one large pool (for
 regular compute) and one small pool (for periodic, low priority work);
 using the small pool is currently the mechanism for limiting the inter-op
 parallelism of the low priority work.  Note that it does not limit the
 parallelism of work spawned by a single op kernel implementation.
 - Using this setting is normally not needed in training, but may help some
 serving use cases.
 - It is also generally recommended to set the global_name field of this
 proto, to avoid creating multiple large pools. It is typically better to
 run the non-low-priority work, even across sessions, in a single large
 pool.
 
repeated .tensorflow.ThreadPoolOptionProto session_inter_op_thread_pool = 12;

Daftar abstrak publik<? memperluas ThreadPoolOptionProtoOrBuilder > getSessionInterOpThreadPoolOrBuilderList ()

 This option is experimental - it may be replaced with a different mechanism
 in the future.
 Configures session thread pools. If this is configured, then RunOptions for
 a Run call can select the thread pool to use.
 The intended use is for when some session invocations need to run in a
 background pool limited to a small number of threads:
 - For example, a session may be configured to have one large pool (for
 regular compute) and one small pool (for periodic, low priority work);
 using the small pool is currently the mechanism for limiting the inter-op
 parallelism of the low priority work.  Note that it does not limit the
 parallelism of work spawned by a single op kernel implementation.
 - Using this setting is normally not needed in training, but may help some
 serving use cases.
 - It is also generally recommended to set the global_name field of this
 proto, to avoid creating multiple large pools. It is typically better to
 run the non-low-priority work, even across sessions, in a single large
 pool.
 
repeated .tensorflow.ThreadPoolOptionProto session_inter_op_thread_pool = 12;

boolean abstrak publik getShareClusterDevicesInSession ()

 When true, WorkerSessions are created with device attributes from the
 full cluster.
 This is helpful when a worker wants to partition a graph
 (for example during a PartitionedCallOp).
 
bool share_cluster_devices_in_session = 17;

boolean abstrak publik getUsePerSessionThreads ()

 If true, use a new set of threads for this session rather than the global
 pool of threads. Only supported by direct sessions.
 If false, use the global threads created by the first session, or the
 per-session thread pools configured by session_inter_op_thread_pool.
 This option is deprecated. The same effect can be achieved by setting
 session_inter_op_thread_pool to have one element, whose num_threads equals
 inter_op_parallelism_threads.
 
bool use_per_session_threads = 9;

boolean abstrak publik hasClusterDef ()

 Optional list of all workers to use in this session.
 
.tensorflow.ClusterDef cluster_def = 14;

boolean abstrak publik hasExperimental ()

.tensorflow.ConfigProto.Experimental experimental = 16;

boolean abstrak publik hasGpuOptions ()

 Options that apply to all GPUs.
 
.tensorflow.GPUOptions gpu_options = 6;

boolean abstrak publik hasGraphOptions ()

 Options that apply to all graphs.
 
.tensorflow.GraphOptions graph_options = 10;

boolean abstrak publik hasRpcOptions ()

 Options that apply when this session uses the distributed runtime.
 
.tensorflow.RPCOptions rpc_options = 13;