o
    &j[                     @   s  d Z ddlmZ ddlmZ ddlmZ ddlm	Z
 e Ze dZe
ee  e
ede  ejdkrd	e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_d	e_d
e_de _!de _"de#_!de#_"de$_!de$_"de%_!de%_"de&_!de&_"de'_!de'_"de(_!de(_"de)_!de)_"de*_!de*_"de+_!de+_"de,_!de,_"de-_!de-_"de._!de._"d e/_!d!e/_"d"e0_!d!e0_"d#e1_!d$e1_"d%e2_!d&e2_"d'e3_!d(e3_"d)e4_!d*e4_"d+e5_!d,e5_"d-e6_!d.e6_"d/e7_!d0e7_"d1e8_!d0e8_"d2e9_!d3e9_"d4e:_!d5e:_"d6e_!d5e_"d6e_!d5e_"d7e;_!d8e;_"d9e<_!d8e<_"d:e_!d8e_"d;e=_!d<e=_"d=e>_!d,e>_"d>e?_!d?e?_"d@e@_!d?e@_"dAeA_!dBeA_"dCe_!dBe_"dDeB_!dEeB_"dFeC_!dGeC_"dHeD_!dGeD_"dIeE_!dJeE_"dKeF_!dLeF_"dMeG_!dNeG_"dOeH_!dPeH_"dQeI_!dReI_"dSeJ_!dTeJ_"dUeK_!dTeK_"dVe_!dWe_"dXe_!dTe_"dYeL_!dZeL_"d[eM_!d\eM_"d]eN_!d^eN_"d_e_!d\e_"d`eO_!daeO_"dbeP_!dceP_"ddeQ_!deeQ_"dfeR_!deeR_"d:e_!d8e_"dgeS_!dheS_"dieT_!djeT_"dke_!dle_"dme_!dne_"doe_!dpe_"d	S d	S )qzGenerated protocol buffer code.    )
descriptor)descriptor_pool)symbol_database)builders   
model_config.protxt	inference"
ModelRateLimiter7
	resources (2$.inference.ModelRateLimiter.Resource
priority (7
Resource
name (	
global (
count ("
ModelInstanceGroup
name (	0
kind (2".inference.ModelInstanceGroup.Kind
count (1
rate_limiter (2.inference.ModelRateLimiter
gpus (H
secondary_devices (2-.inference.ModelInstanceGroup.SecondaryDevice
profile (	
passive (
host_policy	 (	
SecondaryDeviceO
kind (2A.inference.ModelInstanceGroup.SecondaryDevice.SecondaryDeviceKind
	device_id ("%
SecondaryDeviceKind

KIND_NVDLA "A
Kind
	KIND_AUTO 
KIND_GPU
KIND_CPU

KIND_MODEL"#
ModelTensorReshape
shape ("

ModelInput
name (	&
	data_type (2.inference.DataType,
format (2.inference.ModelInput.Format
dims (.
reshape (2.inference.ModelTensorReshape
is_shape_tensor (
allow_ragged_batch (
optional (";
Format
FORMAT_NONE 
FORMAT_NHWC
FORMAT_NCHW"
ModelOutput
name (	&
	data_type (2.inference.DataType
dims (.
reshape (2.inference.ModelTensorReshape
label_filename (	
is_shape_tensor ("

BatchInput(
kind (2.inference.BatchInput.Kind
target_name (	&
	data_type (2.inference.DataType
source_input (	"
Kind
BATCH_ELEMENT_COUNT #
BATCH_ACCUMULATED_ELEMENT_COUNT-
)BATCH_ACCUMULATED_ELEMENT_COUNT_WITH_ZERO$
 BATCH_MAX_ELEMENT_COUNT_AS_SHAPE
BATCH_ITEM_SHAPE
BATCH_ITEM_SHAPE_FLATTEN"
BatchOutput
target_name (	)
kind (2.inference.BatchOutput.Kind
source_input (	"*
Kind"
BATCH_SCATTER_WITH_INPUT_SHAPE "
ModelVersionPolicy6
latest (2$.inference.ModelVersionPolicy.LatestH 0
all (2!.inference.ModelVersionPolicy.AllH :
specific (2&.inference.ModelVersionPolicy.SpecificH 
Latest
num_versions (
All
Specific
versions (B
policy_choice"
ModelOptimizationPolicy7
graph (2(.inference.ModelOptimizationPolicy.GraphB
priority (20.inference.ModelOptimizationPolicy.ModelPriority5
cuda (2'.inference.ModelOptimizationPolicy.CudaX
execution_accelerators (28.inference.ModelOptimizationPolicy.ExecutionAcceleratorsR
input_pinned_memory (25.inference.ModelOptimizationPolicy.PinnedMemoryBufferS
output_pinned_memory (25.inference.ModelOptimizationPolicy.PinnedMemoryBuffer&
gather_kernel_buffer_threshold (
eager_batching (
Graph
level (
Cuda
graphs (
busy_wait_events (E

graph_spec (21.inference.ModelOptimizationPolicy.Cuda.GraphSpec
output_copy_stream (
	GraphSpec

batch_size (K
input (2<.inference.ModelOptimizationPolicy.Cuda.GraphSpec.InputEntryW
graph_lower_bound (2<.inference.ModelOptimizationPolicy.Cuda.GraphSpec.LowerBound
Shape
dim (

LowerBound

batch_size (V
input (2G.inference.ModelOptimizationPolicy.Cuda.GraphSpec.LowerBound.InputEntrye

InputEntry
key (	F
value (27.inference.ModelOptimizationPolicy.Cuda.GraphSpec.Shape:8e

InputEntry
key (	F
value (27.inference.ModelOptimizationPolicy.Cuda.GraphSpec.Shape:8
ExecutionAcceleratorsg
gpu_execution_accelerator (2D.inference.ModelOptimizationPolicy.ExecutionAccelerators.Acceleratorg
cpu_execution_accelerator (2D.inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator
Accelerator
name (	h

parameters (2T.inference.ModelOptimizationPolicy.ExecutionAccelerators.Accelerator.ParametersEntry1
ParametersEntry
key (	
value (	:8$
PinnedMemoryBuffer
enable ("I
ModelPriority
PRIORITY_DEFAULT 
PRIORITY_MAX
PRIORITY_MIN"
ModelQueuePolicyA
timeout_action (2).inference.ModelQueuePolicy.TimeoutAction$
default_timeout_microseconds (
allow_timeout_override (
max_queue_size ("&
TimeoutAction

REJECT 	
DELAY"
ModelDynamicBatching
preferred_batch_size ($
max_queue_delay_microseconds (
preserve_ordering (
priority_levels (
default_priority_level (9
default_queue_policy (2.inference.ModelQueuePolicyW
priority_queue_policy (28.inference.ModelDynamicBatching.PriorityQueuePolicyEntryW
PriorityQueuePolicyEntry
key (*
value (2.inference.ModelQueuePolicy:8"	
ModelSequenceBatchingA
direct (2/.inference.ModelSequenceBatching.StrategyDirectH A
oldest (2/.inference.ModelSequenceBatching.StrategyOldestH &
max_sequence_idle_microseconds (D
control_input (2-.inference.ModelSequenceBatching.ControlInput5
state (2&.inference.ModelSequenceBatching.State
Control;
kind (2-.inference.ModelSequenceBatching.Control.Kind
int32_false_true (
fp32_false_true (
bool_false_true (&
	data_type (2.inference.DataType"u
Kind
CONTROL_SEQUENCE_START 
CONTROL_SEQUENCE_READY
CONTROL_SEQUENCE_END
CONTROL_SEQUENCE_CORRIDW
ControlInput
name (	9
control (2(.inference.ModelSequenceBatching.Control
InitialState&
	data_type (2.inference.DataType
dims (
	zero_data (H 
	data_file (	H 
name (	B

state_data
State

input_name (	
output_name (	&
	data_type (2.inference.DataType
dims (D
initial_state (2-.inference.ModelSequenceBatching.InitialStateX
StrategyDirect$
max_queue_delay_microseconds ( 
minimum_slot_utilization (u
StrategyOldest
max_candidate_sequences (
preferred_batch_size ($
max_queue_delay_microseconds (B
strategy_choice"
ModelEnsembling-
step (2.inference.ModelEnsembling.Step
Step

model_name (	
model_version (@
	input_map (2-.inference.ModelEnsembling.Step.InputMapEntryB

output_map (2..inference.ModelEnsembling.Step.OutputMapEntry/
InputMapEntry
key (	
value (	:80
OutputMapEntry
key (	
value (	:8"&
ModelParameter
string_value (	"
ModelWarmup
name (	

batch_size (2
inputs (2".inference.ModelWarmup.InputsEntry
count (
Input&
	data_type (2.inference.DataType
dims (
	zero_data (H 
random_data (H 
input_data_file (	H B
input_data_typeK
InputsEntry
key (	+
value (2.inference.ModelWarmup.Input:8".
ModelOperations
op_library_filename (	"+
ModelTransactionPolicy
	decoupled ("
ModelRepositoryAgents6
agents (2&.inference.ModelRepositoryAgents.Agent
Agent
name (	J

parameters (26.inference.ModelRepositoryAgents.Agent.ParametersEntry1
ParametersEntry
key (	
value (	:8"$
ModelResponseCache
enable ("

ModelConfig
name (	
platform (	
backend (	5
version_policy (2.inference.ModelVersionPolicy
max_batch_size ($
input (2.inference.ModelInput&
output (2.inference.ModelOutput*
batch_input (2.inference.BatchInput,
batch_output (2.inference.BatchOutput8
optimization (2".inference.ModelOptimizationPolicy;
dynamic_batching (2.inference.ModelDynamicBatchingH =
sequence_batching (2 .inference.ModelSequenceBatchingH 9
ensemble_scheduling (2.inference.ModelEnsemblingH 5
instance_group (2.inference.ModelInstanceGroup
default_model_filename (	H
cc_model_filenames	 (2,.inference.ModelConfig.CcModelFilenamesEntry;
metric_tags
 (2&.inference.ModelConfig.MetricTagsEntry:

parameters (2&.inference.ModelConfig.ParametersEntry,
model_warmup (2.inference.ModelWarmup4
model_operations (2.inference.ModelOperationsC
model_transaction_policy (2!.inference.ModelTransactionPolicyA
model_repository_agents (2 .inference.ModelRepositoryAgents5
response_cache (2.inference.ModelResponseCache7
CcModelFilenamesEntry
key (	
value (	:81
MetricTagsEntry
key (	
value (	:8L
ParametersEntry
key (	(
value (2.inference.ModelParameter:8B
scheduling_choice*
DataType
TYPE_INVALID 
	TYPE_BOOL

TYPE_UINT8
TYPE_UINT16
TYPE_UINT32
TYPE_UINT64
	TYPE_INT8

TYPE_INT16

TYPE_INT32

TYPE_INT64	
	TYPE_FP16

	TYPE_FP32
	TYPE_FP64
TYPE_STRING
	TYPE_BF16bproto3zmodel_config.protxt_pb2FNs   8i  i   #            i  i  i  i[  i  i  i  i  i  i  i   i  i  i.  ia  i1  i  i  i  i  i  i  i  i  i  i  i  i  i	  i	  i
  i  i
  i^  ir  iu  iT  i  i  ib  i  i1  id  i  i  i  i  i  i  iO  i  iR  iA  i  i  iS  i  i!  i$  i  i  i]  i_  i  i  i.  iD  i  i  i@  io  iq  i  i  i  i%  iA  i  i  i'  iU  iW  i  i  ik  i  im  i  i  i  i  i0  i2  ic  ie  i  )U__doc__Zgoogle.protobufr   Z_descriptorr   Z_descriptor_poolr   Z_symbol_databaseZgoogle.protobuf.internalr   Z_builderZDefaultZ_sym_dbZAddSerializedFileZ
DESCRIPTORZBuildMessageAndEnumDescriptorsglobalsZBuildTopDescriptorsAndMessagesZ_USE_C_DESCRIPTORS_optionsZ=_MODELOPTIMIZATIONPOLICY_CUDA_GRAPHSPEC_LOWERBOUND_INPUTENTRYZ_serialized_optionsZ2_MODELOPTIMIZATIONPOLICY_CUDA_GRAPHSPEC_INPUTENTRYZJ_MODELOPTIMIZATIONPOLICY_EXECUTIONACCELERATORS_ACCELERATOR_PARAMETERSENTRYZ._MODELDYNAMICBATCHING_PRIORITYQUEUEPOLICYENTRYZ#_MODELENSEMBLING_STEP_INPUTMAPENTRYZ$_MODELENSEMBLING_STEP_OUTPUTMAPENTRYZ_MODELWARMUP_INPUTSENTRYZ,_MODELREPOSITORYAGENTS_AGENT_PARAMETERSENTRYZ"_MODELCONFIG_CCMODELFILENAMESENTRYZ_MODELCONFIG_METRICTAGSENTRYZ_MODELCONFIG_PARAMETERSENTRYZ	_DATATYPEZ_serialized_startZ_serialized_endZ_MODELRATELIMITERZ_MODELRATELIMITER_RESOURCEZ_MODELINSTANCEGROUPZ#_MODELINSTANCEGROUP_SECONDARYDEVICEZ7_MODELINSTANCEGROUP_SECONDARYDEVICE_SECONDARYDEVICEKINDZ_MODELINSTANCEGROUP_KINDZ_MODELTENSORRESHAPEZ_MODELINPUTZ_MODELINPUT_FORMATZ_MODELOUTPUTZ_BATCHINPUTZ_BATCHINPUT_KINDZ_BATCHOUTPUTZ_BATCHOUTPUT_KINDZ_MODELVERSIONPOLICYZ_MODELVERSIONPOLICY_LATESTZ_MODELVERSIONPOLICY_ALLZ_MODELVERSIONPOLICY_SPECIFICZ_MODELOPTIMIZATIONPOLICYZ_MODELOPTIMIZATIONPOLICY_GRAPHZ_MODELOPTIMIZATIONPOLICY_CUDAZ'_MODELOPTIMIZATIONPOLICY_CUDA_GRAPHSPECZ-_MODELOPTIMIZATIONPOLICY_CUDA_GRAPHSPEC_SHAPEZ2_MODELOPTIMIZATIONPOLICY_CUDA_GRAPHSPEC_LOWERBOUNDZ._MODELOPTIMIZATIONPOLICY_EXECUTIONACCELERATORSZ:_MODELOPTIMIZATIONPOLICY_EXECUTIONACCELERATORS_ACCELERATORZ+_MODELOPTIMIZATIONPOLICY_PINNEDMEMORYBUFFERZ&_MODELOPTIMIZATIONPOLICY_MODELPRIORITYZ_MODELQUEUEPOLICYZ_MODELQUEUEPOLICY_TIMEOUTACTIONZ_MODELDYNAMICBATCHINGZ_MODELSEQUENCEBATCHINGZ_MODELSEQUENCEBATCHING_CONTROLZ#_MODELSEQUENCEBATCHING_CONTROL_KINDZ#_MODELSEQUENCEBATCHING_CONTROLINPUTZ#_MODELSEQUENCEBATCHING_INITIALSTATEZ_MODELSEQUENCEBATCHING_STATEZ%_MODELSEQUENCEBATCHING_STRATEGYDIRECTZ%_MODELSEQUENCEBATCHING_STRATEGYOLDESTZ_MODELENSEMBLINGZ_MODELENSEMBLING_STEPZ_MODELPARAMETERZ_MODELWARMUPZ_MODELWARMUP_INPUTZ_MODELOPERATIONSZ_MODELTRANSACTIONPOLICYZ_MODELREPOSITORYAGENTSZ_MODELREPOSITORYAGENTS_AGENTZ_MODELRESPONSECACHEZ_MODELCONFIG r   r   t/var/www/html/Deteccion_Ine/venv/lib/python3.10/site-packages/visualdl/component/inference/proto/model_config_pb2.py<module>   sF  
 