o
    "jR                     @   s2  d Z ddlmZ ddlmZ ddlmZ ddlm	Z
 e
 Ze dZeee  eede  ejdkrd	e_d
e_de_de_de_de_de_de_de_de_de_de_de_de_de_de_de_de_de_de_de_de _de _d e!_d!e!_d"e"_d#e"_d$e#_d%e#_d&e$_d'e$_d(e%_d)e%_d*e&_d+e&_d,e'_d-e'_d.e(_d/e(_d0e)_d1e)_d2e*_d3e*_d4e+_d5e+_d6e,_d7e,_d8e-_d9e-_d:e._d;e._d<e/_d=e/_d>e0_d?e0_d@e1_dAe1_dBe2_dCe2_dDe3_dEe3_dFe4_dGe4_dHe5_dIe5_dJe6_dKe6_dLe7_dMe7_dNe8_dOe8_dPe9_dQe9_d	S d	S )RzGenerated protocol buffer code.    )builder)
descriptor)descriptor_pool)symbol_databases%  
1paddle/fluid/framework/distributed_strategy.protopaddle.fleet"W
RefinedOpsPattern
main_ops (	
num (:0
pre_ops (	
suf_ops (	"
RecomputeConfig
checkpoints (	
enable_offload (:false
checkpoint_shape (
enable_tuning (:false=
refined_ops_patterns (2.paddle.fleet.RefinedOpsPattern"
ShardingConfig7
sharding_segment_strategy (	:segment_broadcast_MB 
segment_broadcast_MB (:32
segment_anchors (	
sharding_degree (:8
	mp_degree (:1
	dp_degree (:1
	hybrid_dp (:false"
gradient_merge_acc_step (:1
optimize_offload	 (:false'
pp_allreduce_in_optimize
 (:false
	pp_degree (:1
optimize_cast (:false(
_dp_as_optimizer_sharding (:false
stage (:1
enable_tuning (:false
use_calc_stream (:false"
MpConfig

sync_param (:true
	sync_grad (:false
sync_moment (:false
	sync_mode (	:	broadcast!
mp_async_allreduce (:false!
mp_skip_c_identity (:false-
mp_fused_linear_param_grad_add (:false!
need_broadcast_data (:true"
PpConfig
dp_comm_overlap (:false
delay_scale_loss (:false
enable_timer (:false$
sharding_comm_overlap (:false
	profiling (:false"
DygraphShardingConfig
tensor_fusion (:false
accumulate_steps (:1
comm_overlap (:false
split_param (:false
fuse_optimizer (:true"
HybridConfig
	dp_degree (:-1
	mp_degree (:1
	pp_degree (:1
sharding_degree (:1

sep_degree (:1*

mp_configs (2.paddle.fleet.MpConfig*

pp_configs (2.paddle.fleet.PpConfig=
sharding_configs (2#.paddle.fleet.DygraphShardingConfig"
	AMPConfig 
init_loss_scaling (:32768 
incr_every_n_steps (:1000"
decr_every_n_nan_or_inf (:2

incr_ratio (:2

decr_ratio (:0.8&
use_dynamic_loss_scaling (:true
custom_white_list (	
custom_black_list (	
custom_black_varnames	 (	
use_pure_fp16
 (:false
use_fp16_guard (:true!
use_optimizer_fp16 (:false
use_pure_bf16 (:false";
LocalSGDConfig
k_steps (:1

begin_step (:1"H
AdaptiveLocalSGDConfig
init_k_steps (:1

begin_step (:1"<
GradientMergeConfig
k_steps (:1
avg (:true"S
	DGCConfig
rampup_begin_step (:0
rampup_step (:1
sparsity ("

LarsConfig

lars_coeff (:0.001!
lars_weight_decay (:0.0005
epsilon (:0!
exclude_from_weight_decay (	"P

LambConfig
lamb_weight_decay (:0.01!
exclude_from_weight_decay (	"
BuildStrategy*
enable_sequential_execution (:false'
fuse_elewise_add_act_ops (:false
fuse_bn_act_ops (:false'
fuse_relu_depthwise_conv (:false!
fuse_broadcast_ops (:false%
fuse_all_optimizer_ops (:false
enable_inplace (:false/
!enable_backward_optimizer_op_deps (:true$
cache_runtime_context	 (:false!
fuse_bn_add_act_ops
 (:true!
enable_auto_fusion (:false
enable_addto (:false
fix_op_run_order (:false'
allow_cuda_graph_capture (:false
reduce_strategy (:0!
fuse_gemm_epilogue (:false
debug_graphviz_path (	
fused_attention (:false 
fused_feedforward (:false"
ExecutionStrategy
num_threads (:1(
num_iteration_per_drop_scope (:10 
num_iteration_per_run (:1!
use_thread_barrier (:false"Q
GradientScaleConfig
scale_strategy (	:avg
scale_gradient (:false"
AsyncConfig
k_steps (:-1
max_merge_var_num (:1
send_queue_size (:16&
independent_recv_thread (:false(
min_send_grad_num_before_recv (:1
thread_pool_size (:1
send_wait_times (:1&
runtime_split_send_recv (:false
launch_barrier	 (:true&
heter_worker_device_guard
 (	:cpu
lr_decay_steps (:10

use_ps_gpu (:0"
TrainerDescConfig
dump_fields_path (	
dump_fields (	

dump_param (	
stat_var_names (	
trainer (	
device_worker (	
local_sparse (	
remote_sparse (	"
PipelineConfig
micro_batch_size (:1
accumulate_steps (:1
schedule_mode (	:1F1B
p2p_cache_shape (:true&
enable_partial_send_recv (:true"W
TensorParallelConfig!
tensor_parallel_degree (:1
tensor_init_seed (:-1"
	QatConfig"
channel_wise_abs_max (:true
weight_bits (:8
activation_bits (:8
not_quant_pattern (	
algo (	"
TableParameter
table_id (

table_name (	
table_class (	
	shard_num (:1000%
type (2.paddle.fleet.TableType6
accessor (2$.paddle.fleet.TableAccessorParameter
compress_in_save (:false'
enable_sparse_table_cache
 (:true(
sparse_table_cache_rate (:0.00055'
sparse_table_cache_file_num (:16
enable_revert (:false
shard_merge_rate (:1"
TableAccessorParameter
accessor_class (	
fea_dim (:11

embedx_dim (:8
embedx_threshold (:10>
ctr_accessor_param (2".paddle.fleet.CtrAccessorParameterK
table_accessor_save_param (2(.paddle.fleet.TableAccessorSaveParameter3
embed_sgd_param
 (2.paddle.fleet.SGDParameter4
embedx_sgd_param (2.paddle.fleet.SGDParameter8
graph_sgd_param (2.paddle.fleet.GraphSGDParameter"S
GraphSGDParameter
nodeid_slot (:9008#
feature_learning_rate (:0.05"
SGDParameter
name (	8
naive (2).paddle.fleet.SparseNaiveSGDRuleParameter<
adagrad (2+.paddle.fleet.SparseAdagradSGDRuleParameter2
adam (2$.paddle.fleet.SparseAdamSGDParameter"p
SparseNaiveSGDRuleParameter
learning_rate (:0.05
initial_range (:0.0001
weight_bounds ("
SparseAdagradSGDRuleParameter
learning_rate (:0.05
initial_g2sum (:3
initial_range (:0.0001
weight_bounds ("
SparseAdamSGDParameter
learning_rate (:0.001
initial_range (:0.0001
beta1_decay_rate (:0.9
beta2_decay_rate (:0.999
ada_epsilon (:1e-08
weight_bounds ("
CtrAccessorParameter
nonclk_coeff (:0.1
click_coeff (:1
base_threshold (:1.5
delta_threshold (:0.25
delta_keep_days (:16#
show_click_decay_rate (:0.98
delete_threshold (:0.8$
delete_after_unseen_days (:30"
ssd_unseenday_threshold	 (:1

show_scale
 (:true
	zero_init (:true
load_filter_slots (
save_filter_slots ("S
TableAccessorSaveParameter
param (
	converter (	
deconverter (	"R
FsClientParameter
uri (	
user (	
passwd (	

hadoop_bin (	"
DistributedStrategy,
mode (2.paddle.fleet.Mode:
COLLECTIVE
amp (:false
	recompute (:false
localsgd (:false
dgc (:false
gradient_merge (:false
lars (:false
lamb (:false
pipeline	 (:false
elastic
 (:false
auto (:false
a_sync (:true!
sync_nccl_allreduce (:true
nccl_comm_num (:1)
use_hierarchical_allreduce (:false.
#hierarchical_allreduce_inter_nranks (:1
sync_batch_norm (:false!
fuse_all_reduce_ops (:true 
fuse_grad_size_in_MB (:32$
fuse_grad_size_in_TFLOPS (:50&
cudnn_exhaustive_search (:false&
conv_workspace_size_limit (:5121
"cudnn_batchnorm_spatial_persistent (:false 
adaptive_localsgd (:false
fp16_allreduce (:false
sharding (:false"
last_comm_group_size_MB (:1%
find_unused_parameters (:false
tensor_parallel (:false(
without_graph_optimization (:true 
fuse_grad_size_in_num (:8$
calc_comm_same_stream  (:false
asp! (:false
fuse_grad_merge" (:false
	semi_auto# (:false

adam_d2sum$ (:false
auto_search% (:false
heter_ccl_mode& (:false
is_fl_ps_mode' (:false
with_coordinator( (:false
qat) (:false

split_data* (:true8
recompute_configse (2.paddle.fleet.RecomputeConfig,
amp_configsf (2.paddle.fleet.AMPConfig6
localsgd_configsg (2.paddle.fleet.LocalSGDConfigA
gradient_merge_configsh (2!.paddle.fleet.GradientMergeConfig,
dgc_configsi (2.paddle.fleet.DGCConfig6
pipeline_configsj (2.paddle.fleet.PipelineConfig1
a_sync_configsk (2.paddle.fleet.AsyncConfig.
lars_configsl (2.paddle.fleet.LarsConfig.
lamb_configsm (2.paddle.fleet.LambConfigG
adaptive_localsgd_configsn (2$.paddle.fleet.AdaptiveLocalSGDConfig6
sharding_configso (2.paddle.fleet.ShardingConfig2
hybrid_configsp (2.paddle.fleet.HybridConfigC
tensor_parallel_configsq (2".paddle.fleet.TensorParallelConfig=
trainer_desc_configsr (2.paddle.fleet.TrainerDescConfig:
downpour_table_params (2.paddle.fleet.TableParameter8
fs_client_paramt (2.paddle.fleet.FsClientParameter,
qat_configsu (2.paddle.fleet.QatConfig4
build_strategy (2.paddle.fleet.BuildStrategy<
execution_strategy (2.paddle.fleet.ExecutionStrategyB
gradient_scale_configs (2!.paddle.fleet.GradientScaleConfig"
DistributedJobInfo

worker_num (

server_num (

worker_ips (	
server_endpoints (	
origin_startup (	
origin_main (	
distributed_main (	
optimizer_name (	3
strategye (2!.paddle.fleet.DistributedStrategy*7
Mode

COLLECTIVE
PS
PIPELINE	
HETER*4
	TableType
PS_SPARSE_TABLE 
PS_DENSE_TABLEz.paddle.base.framework.distributed_strategy_pb2FNi2%  ii%  ik%  i%  C         iY  i\  i^  ia  iq  it  i  i  i  i  i  i  i  i  i  i  i	  i	  iK	  iM	  i	  i	  i$
  i&
  iv
  iy
  i1  i4  i  i  i!  i$  i  i  is  iv  i$  i&  i}  i  i  i  i  i  iY  i[  i  i  iy  i{  i  i  iz  i}  iE  iH  i  i  i6  i8  i  i  i/$  i2$  i0%  ):__doc__Zgoogle.protobuf.internalr   Z_builderZgoogle.protobufr   Z_descriptorr   Z_descriptor_poolr   Z_symbol_databaseZDefaultZ_sym_dbZAddSerializedFileZ
DESCRIPTORZBuildMessageAndEnumDescriptorsglobalsZBuildTopDescriptorsAndMessagesZ_USE_C_DESCRIPTORS_optionsZ_MODEZ_serialized_startZ_serialized_endZ
_TABLETYPEZ_REFINEDOPSPATTERNZ_RECOMPUTECONFIGZ_SHARDINGCONFIGZ	_MPCONFIGZ	_PPCONFIGZ_DYGRAPHSHARDINGCONFIGZ_HYBRIDCONFIGZ
_AMPCONFIGZ_LOCALSGDCONFIGZ_ADAPTIVELOCALSGDCONFIGZ_GRADIENTMERGECONFIGZ
_DGCCONFIGZ_LARSCONFIGZ_LAMBCONFIGZ_BUILDSTRATEGYZ_EXECUTIONSTRATEGYZ_GRADIENTSCALECONFIGZ_ASYNCCONFIGZ_TRAINERDESCCONFIGZ_PIPELINECONFIGZ_TENSORPARALLELCONFIGZ
_QATCONFIGZ_TABLEPARAMETERZ_TABLEACCESSORPARAMETERZ_GRAPHSGDPARAMETERZ_SGDPARAMETERZ_SPARSENAIVESGDRULEPARAMETERZ_SPARSEADAGRADSGDRULEPARAMETERZ_SPARSEADAMSGDPARAMETERZ_CTRACCESSORPARAMETERZ_TABLEACCESSORSAVEPARAMETERZ_FSCLIENTPARAMETERZ_DISTRIBUTEDSTRATEGYZ_DISTRIBUTEDJOBINFO r   r   k/var/www/html/Deteccion_Ine/venv/lib/python3.10/site-packages/paddle/base/proto/distributed_strategy_pb2.py<module>   s   
