identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
cpmant/modeling_cpmant.py:CpmAntLayerNorm
[ -0.000021966368876746856, 0.049015820026397705, 0.04162987321615219, 0.04453948885202408, 0, 0.01991967111825943, 0.008057394996285439, -0.00788953248411417, 0.0037769039627164602, 0.049015820026397705, 0.023612642660737038, 0.00956815667450428, 0.004364422056823969, -0.002825683681294322,...
[ "ModelLayerNorm", "Module", "Parameter", "True", "__init__", "class", "config", "def", "dim", "dim_norm", "empty", "eps", "float32", "forward", "hidden_size", "hidden_states", "if", "keepdim", "mean", "nn", "pow", "return", "rsqrt", "self", "size", "super", "to", ...
cpmant/modeling_cpmant.py:CpmAntAttention
[ -0.0000913355324883014, 0.0346812829375267, 0.037813786417245865, -0.02136814594268799, -0.00038981481338851154, 0.016781264916062355, 0.02494814805686474, -0.01946626789867878, 0.0034261750988662243, 0.008222820237278938, 0.015215014107525349, 0.03311503306031227, -0.0008530477061867714, ...
[ "Dropout", "False", "Linear", "ModelAttention", "Module", "None", "Softmax", "__init__", "attention_mask", "attention_out", "attn_weights", "batch_size", "cache_position", "class", "config", "contiguous", "def", "device", "dim", "dim_head", "dim_model", "dropout", "dropou...
cpmant/modeling_cpmant.py:CpmAntSelfAttentionBlock
[ 0.00007607758743688464, 0.03066888637840748, 0.032683633267879486, -0.010409512557089329, 0.0000926922366488725, 0.02663939818739891, 0.023169560357928276, -0.012760047800838947, 0.009626001119613647, 0.0013781411107629538, 0.01981165260076523, 0.03313135355710983, 0.003693697974085808, -0...
[ "Dropout", "False", "ModelAttention", "ModelLayerNorm", "ModelSelfAttentionBlock", "Module", "None", "__init__", "attention_mask", "attn_weights", "cache_position", "class", "config", "def", "dropout", "dropout_p", "else", "forward", "hidden_states", "if", "is", "layer_idx"...
cpmant/modeling_cpmant.py:CpmAntDenseGatedACT
[ -0.00017635135736782104, 0.011571498587727547, 0.013737591914832592, 0.005443734582513571, -0.0007552825263701379, 0.027019163593649864, 0.034429483115673065, -0.031009336933493614, 0.002436855109408498, -0.007182309404015541, 0.025423096492886543, -0.030553316697478294, 0.001147174392826855...
[ "GELU", "Linear", "ModelDenseGatedACT", "Module", "__init__", "act", "class", "config", "def", "dim_ff", "forward", "gate_score", "hidden_size", "hidden_states", "nn", "return", "self", "super", "torch", "w_0", "w_1" ]
cpmant/modeling_cpmant.py:CpmAntFeedForward
[ -0.0002584679168649018, 0.04066561907529831, 0.024123672395944595, 0.003446238813921809, -0.0011990038910880685, 0.016656821593642235, 0.03561113402247429, -0.03974661976099014, -0.002498523099347949, -0.011430025100708008, 0.03078640066087246, -0.01585269905626774, 0.004078049212694168, -...
[ "Dropout", "Linear", "ModelDenseGatedACT", "ModelFeedForward", "Module", "None", "__init__", "class", "config", "def", "dim_ff", "dropout", "dropout_p", "else", "forward", "hidden_size", "hidden_states", "if", "is", "nn", "not", "return", "self", "super", "torch", "...
cpmant/modeling_cpmant.py:CpmAntFFNBlock
[ -0.0001340094459010288, 0.05361797660589218, 0.037032756954431534, 0.017153209075331688, -0.0005715369479730725, 0.02362826094031334, 0.03589678183197975, -0.028399351984262466, 0.0056230719201266766, -0.007440630346536636, 0.02226509153842926, 0.00309552950784564, 0.006787445396184921, 0....
[ "Dropout", "ModelFFNBlock", "ModelFeedForward", "ModelLayerNorm", "Module", "None", "__init__", "class", "config", "def", "dropout", "dropout_p", "else", "ffn", "forward", "hidden_states", "if", "is", "layernorm_before_ffn", "ln_outputs", "nn", "not", "outputs", "return...
cpmant/modeling_cpmant.py:CpmAntTransformerBlock
[ -0.00007616254879394546, 0.02682330645620823, 0.03178223595023155, -0.03042980097234249, 0, 0.045531999319791794, 0.02716141566634178, -0.029753582552075386, 0.01019961852580309, -0.004451767075806856, 0.0054097422398626804, 0.013918816111981869, 0.0038319008890539408, 0.009241643361747265...
[ "False", "ModelFFNBlock", "ModelSelfAttentionBlock", "ModelTransformerBlock", "Module", "None", "__init__", "attention_mask", "attn_weights", "cache_position", "class", "config", "def", "ffn", "forward", "hidden_states", "layer_idx", "nn", "output_attentions", "past_key_values"...
cpmant/modeling_cpmant.py:CpmAntEncoder
[ -0.00002259437314933166, 0.02751687541604042, 0.030324719846248627, 0.013589967042207718, 0.00009651965228840709, 0.048294924199581146, 0.013814594596624374, -0.03638966381549835, 0.010445181280374527, 0.0012845888268202543, 0.007749650627374649, 0.009883612394332886, -0.00004804046329809353...
[ "ModelEncoder", "ModelLayerNorm", "ModelTransformerBlock", "Module", "ModuleList", "None", "__init__", "all_hidden_states", "all_self_attns", "attention_mask", "attn_weights", "cache_position", "class", "config", "def", "else", "enumerate", "for", "forward", "hidden_states", ...
cpmant/modeling_cpmant.py:CpmAntIntermediate
[ -0.00025367451598867774, 0.02240910567343235, 0.04047359153628349, 0.012862369418144226, -0.0009432404185645282, 0.03635763004422188, 0.03452831506729126, -0.018864808604121208, -0.001329111517407, -0.003772961674258113, 0.023209432139992714, -0.02103712037205696, -0.0013719861162826419, 0...
[ "ACT2FN", "Linear", "ModelIntermediate", "Module", "__init__", "class", "config", "def", "dense", "else", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "intermediate_act_fn", "intermediate_size", "isinstance", "nn", "return", "self", "str", "super" ]
cpmant/modeling_cpmant.py:CpmAntSegmentPositionEmbedding
[ -0.00031115286401472986, 0.011453303508460522, 0.006100754253566265, -0.02969801239669323, -0.0012805828591808677, 0.025784321129322052, 0.022331062704324722, -0.0024460572749376297, 0.003741028718650341, 0.0172662865370512, 0.02520877681672573, 0.039136916399002075, 0.0013453314313665032, ...
[ "AssertionError", "F", "ModelSegmentPositionEmbedding", "Module", "None", "Parameter", "__init__", "_position_bucket", "_segment_relative_position_bucket", "abs", "absolute_position_bucket", "and", "arange", "batch", "be", "but", "class", "config", "contiguous", "def", "devic...
cpmant/modeling_cpmant.py:CpmAntOutput
[ -0.0002591986849438399, 0.03935529664158821, 0.0507957898080349, 0.030660521239042282, -0.0012513039400801063, 0.04988054931163788, 0.03798243775963783, -0.023681821301579475, 0.002259497530758381, 0.014758236706256866, 0.011040075682103634, 0.00267421524040401, 0.0013013561256229877, 0.00...
[ "Dropout", "LayerNorm", "Linear", "ModelOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "eps", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "intermediate_size", "layer_norm_eps", "nn", "return", "self", "su...
cpmant/modeling_cpmant.py:CpmAntPreTrainedModel
[ -0.00021195212320890278, 0.05084000900387764, -0.0020233411341905594, 0.011969060637056828, -0.0011470350436866283, 0.04012485221028328, 0.01949246972799301, 0.005842041689902544, 0.0035907181445509195, 0.008606324903666973, 0.0009119284222833812, 0.013279957696795464, 0.00009439884161110967...
[ "Model", "ModelConfig", "ModelLayerNorm", "ModelPreTrainedModel", "ModelSegmentPositionEmbedding", "PreTrainedModel", "_init_weights", "base_model_prefix", "class", "config", "def", "elif", "if", "init", "init_std", "isinstance", "mean", "module", "no_grad", "normal_", "ones_...
cpmant/modeling_cpmant.py:CpmAntModel
[ 0.000011867956345668063, 0.04413085803389549, -0.014224921353161335, 0.015344994142651558, -0.00010150657180929556, 0.05017925053834915, 0.02542564645409584, 0.0007770503289066255, 0.00624440424144268, 0.007784503977745771, 0.028561849147081375, 0.010416674427688122, -0.0008050521137192845, ...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "ModelSegmentPositionEmbedding", "None", "__init__", "_prepare_attention_mask", "all_attentions", "all_hidden_states", "and", "arange", "attention", "attention_mask", "attent...
cpmant/modeling_cpmant.py:CpmAntForCausalLM
[ -0.00012092290126020089, 0.04374254494905472, 0.008636348880827427, 0.01345924474298954, -0.0006554371793754399, 0.023777998983860016, 0.026806330308318138, -0.0035891321022063494, 0.0022011473774909973, 0.020861830562353134, 0.02624552696943283, -0.01345924474298954, -0.00006177583418320864...
[ "CausalLMOutputWithPast", "CrossEntropyLoss", "GenerationMixin", "Linear", "Model", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_position", "class", "config", "def", "e...
mistral3/modeling_mistral3.py:Mistral3RMSNorm
[ -0.00009884802420856431, 0.04179859161376953, 0.03230918198823929, 0.053095508366823196, -0.00046952810953371227, 0.03931327164173126, 0.022028988227248192, -0.02937198430299759, 0.007964326068758965, 0.042476408183574677, 0.019995542243123055, 0.006975846365094185, 0.0028524715453386307, ...
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", "to", "torch", "tuple", ...
mistral3/modeling_mistral3.py:Mistral3PatchMerger
[ -0.00016565353143960238, 0.02254282869398594, 0.01355917751789093, 0.011327213607728481, -0.0003016637929249555, 0.0029852509032934904, 0.0032502966932952404, -0.04687123000621796, 0.00714228255674243, 0.004938218742609024, -0.010267031379044056, 0.005831004120409489, -0.001401952002197504, ...
[ "Linear", "ModelPatchMerger", "Module", "__init__", "append", "cat", "class", "config", "d", "def", "dim", "enumerate", "for", "forward", "functional", "grid", "h", "hidden_size", "image_features", "image_grid", "image_index", "image_size", "image_sizes", "image_tokens"...
mistral3/modeling_mistral3.py:Mistral3MultiModalProjector
[ -0.000226370437303558, 0.03870934247970581, 0.028296303004026413, 0.049348752945661545, -0.0005057964590378106, 0.0319182313978672, 0.0168645977973938, -0.038482971489429474, -0.00023875005717854947, 0.017090966925024986, 0.021052449941635132, 0.0042444453574717045, -0.0011035559000447392, ...
[ "ACT2FN", "Linear", "ModelMultiModalProjector", "ModelPatchMerger", "ModelRMSNorm", "Module", "__init__", "act", "class", "config", "def", "else", "eps", "forward", "hidden_size", "hidden_states", "if", "image_features", "image_sizes", "int", "isinstance", "len", "linear_...
mistral3/modeling_mistral3.py:Mistral3CausalLMOutputWithPast
[ -0.00020158913685008883, 0.02637176588177681, 0.02978190779685974, 0.009434727020561695, -0.0011793408775702119, 0.03387407958507538, 0.03773890808224678, -0.030463937669992447, 0.01784641109406948, -0.004205842036753893, 0.020006166771054268, 0.018642110750079155, -0.0008596400148235261, ...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "image_hidden_states", "logits", "loss", "past_key_values", "r" ]
mistral3/modeling_mistral3.py:Mistral3ModelOutputWithPast
[ -0.000123380494187586, 0.02538112923502922, 0.014777458272874355, 0.01703355833888054, -0.0007931602885946631, 0.02910369634628296, 0.04331712797284126, -0.038353707641363144, 0.016920752823352814, -0.007952754385769367, 0.02763723023235798, 0.026170765981078148, -0.002354804892092943, 0.0...
[ "BaseModelOutputWithPast", "ModelModelOutputWithPast", "None", "class", "image_hidden_states", "r" ]
mistral3/modeling_mistral3.py:Mistral3PreTrainedModel
[ -0.00040332425851374865, 0.03276687115430832, 0.01826929859817028, 0.019447963684797287, -0.0021068628411740065, 0.03158820793032646, 0.028523679822683334, -0.03323833644390106, -0.003683326533064246, 0.01744423434138298, 0.015440504997968674, 0.008898916654288769, -0.004272658843547106, 0...
[ "ModelConfig", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "base_model_prefix", "class", "config", "image", "input_modalities", ...
mistral3/modeling_mistral3.py:Mistral3Model
[ -0.0000704072808730416, 0.03337826579809189, -0.003476902609691024, 0.015910306945443153, -0.00020079112437088042, 0.034713394939899445, 0.01613282784819603, -0.038496267050504684, 0.00023642937594559044, 0.024143612012267113, 0.030930524691939354, 0.009512805379927158, 0.0007440571789629757...
[ "AutoModel", "False", "Image", "ModelModel", "ModelModelOutputWithPast", "ModelMultiModalProjector", "ModelPreTrainedModel", "None", "Obtains", "True", "ValueError", "You", "__init__", "_checkpoint_conversion_mapping", "all", "and", "apply", "as_tensor", "attention_mask", "atte...
mistral3/modeling_mistral3.py:Mistral3ForConditionalGeneration
[ -0.0002581247827038169, 0.036887090653181076, 0.00534636527299881, 0.0255720317363739, -0.0009052046807482839, 0.03122956119477749, 0.023648472502827644, -0.03915010392665863, -0.00282876449637115, 0.021498611196875572, 0.03168216347694397, -0.0004632101918105036, 0.002786333207041025, -0....
[ "False", "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_checkpoint_conversion_mapping", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_posit...
persimmon/modeling_persimmon.py:PersimmonRotaryEmbedding
[ -0.00032534197089262307, 0.04792523384094238, 0.004478450398892164, -0.0062523819506168365, -0.0016866890946403146, 0.03722348436713219, 0.03954995051026344, 0.010934398509562016, -0.003867752617225051, 0.019891299307346344, 0.0026172762736678123, -0.0038095908239483833, -0.00119958491995930...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
persimmon/modeling_persimmon.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
persimmon/modeling_persimmon.py:apply_rotary_pos_emb
[ -0.0001444592053303495, 0.027112245559692383, 0.028019767254590988, 0.0028360087890177965, -0.0005707467789761722, 0.021667107939720154, 0.046510547399520874, -0.002183726755902171, 0.01293220091611147, 0.03652779385447502, 0.007884104736149311, 0.0017654155381023884, -0.0007834474672563374,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
persimmon/modeling_persimmon.py:PersimmonMLP
[ -0.00026337814051657915, 0.017429541796445847, 0.04173916578292847, 0.017085537314414978, -0.000946011277846992, 0.041509829461574554, 0.034171074628829956, -0.021442921832203865, -0.0021213586442172527, -0.009345444850623608, 0.033024393022060394, -0.03073103353381157, -0.000820593151729553...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "__init__", "act", "class", "config", "def", "dense_4h_to_h", "dense_h_to_4h", "forward", "hidden_act", "hidden_size", "hidden_states", "intermediate_size", "nn", "return", "self", "super" ]
persimmon/modeling_persimmon.py:eager_attention_forward
[ 0.000020310135369072668, 0.02814454957842827, 0.029840005561709404, -0.014467881061136723, 0, 0.038430310785770416, 0.06013213098049164, -0.026562124490737915, 0.02305818535387516, 0.01695454865694046, 0.025205761194229126, 0.027353337034583092, 0.0022464776411652565, -0.010907425545156002...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling",...
persimmon/modeling_persimmon.py:PersimmonAttention
[ -0.00014542994904331863, 0.041291527450084686, 0.02459440752863884, -0.004907599650323391, -0.0006205011159181595, 0.021322675049304962, 0.028430232778191566, -0.017486849799752235, -0.000243264643358998, 0.02527131699025631, 0.0037512113340198994, 0.026173865422606468, -0.000528836157172918...
[ "ALL_ATTENTION_FUNCTIONS", "Dropout", "False", "Instantiating", "LayerNorm", "Linear", "ModelAttention", "Module", "None", "Please", "Tensor", "True", "ValueError", "_", "__class__", "__init__", "__name__", "_attn_implementation", "_split_heads", "a", "and", "apply_rotary_p...
persimmon/modeling_persimmon.py:PersimmonDecoderLayer
[ -0.0001449557312298566, 0.04434938356280327, 0.014933976344764233, -0.005006276071071625, -0.0008555923704989254, 0.03552475944161415, 0.04412310943007469, -0.03190440312027931, 0.0065053305588662624, -0.006788170896470547, -0.004440594930201769, 0.02160901017487049, -0.002276865765452385, ...
[ "Dropout", "False", "GradientCheckpointingLayer", "LayerNorm", "ModelAttention", "ModelDecoderLayer", "ModelMLP", "None", "Tensor", "__init__", "attention_mask", "cache_position", "class", "config", "def", "dropout", "eps", "forward", "hidden_dropout", "hidden_size", "hidden_...
persimmon/modeling_persimmon.py:PersimmonPreTrainedModel
[ -0.00039423315320163965, 0.03270674869418144, 0.01693742349743843, 0.012557055801153183, -0.0017375461757183075, 0.02359558455646038, 0.026399020105600357, -0.03083779290318489, -0.003971533849835396, 0.006512147840112448, 0.009286381304264069, -0.0019565646070986986, -0.003577300813049078, ...
[ "ModelConfig", "ModelDecoderLayer", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_no_split_modules", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_sdpa", "base_model_prefix", "class", "config", "model", ...
persimmon/modeling_persimmon.py:PersimmonModel
[ 0.000022689519028062932, 0.04691494628787041, 0.009159584529697895, -0.014130334369838238, 0.00009686679550213739, 0.039095789194107056, 0.025691518560051918, -0.022117044776678085, 0.012901609763503075, 0.004160908982157707, 0.01228724792599678, 0.011896289885044098, 0.00038223114097490907,...
[ "AttentionMaskConverter", "BaseModelOutputWithPast", "BlockMask", "DynamicCache", "Embedding", "False", "LayerNorm", "ModelDecoderLayer", "ModelModel", "ModelPreTrainedModel", "ModelRotaryEmbedding", "ModuleList", "None", "Setting", "Tensor", "True", "ValueError", "You", "__init_...
persimmon/modeling_persimmon.py:PersimmonForCausalLM
[ -0.0002664232742972672, 0.039103832095861435, 0.015232306905090809, -0.004774305038154125, -0.001278831739909947, 0.0259176567196846, 0.038194440305233, -0.013015665113925934, 0.005342674907296896, 0.020006611943244934, 0.027054395526647568, 0.008298196829855442, 0.0006216543260961771, -0....
[ "CausalLMOutputWithPast", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_position", "can_return_tuple", "class", "config", "def", "else", "em...
persimmon/modeling_persimmon.py:PersimmonForSequenceClassification
[ -0.00028286955785006285, 0.024906935170292854, -0.010723819956183434, 0.02444569580256939, -0.0012539949966594577, 0.03390110656619072, 0.031594909727573395, -0.003718743799254298, 0.010262579657137394, -0.012972362339496613, 0.027559062466025352, 0.010320235043764114, -0.005650184582918882,...
[ "GenericForSequenceClassification", "ModelForSequenceClassification", "ModelPreTrainedModel", "class" ]
persimmon/modeling_persimmon.py:PersimmonForTokenClassification
[ -0.00024937751004472375, 0.02962389588356018, -0.005080842413008213, -0.02101229690015316, -0.001105154980905354, 0.04087638109922409, 0.048454590141773224, -0.0025978318881243467, 0.005597538314759731, -0.01906033605337143, 0.039498526602983475, 0.0252606850117445, -0.006171644665300846, ...
[ "GenericForTokenClassification", "ModelForTokenClassification", "ModelPreTrainedModel", "class" ]
sam3_video/modeling_sam3_video.py:_load_cv_utils_kernel_once
[ -0.0003534568240866065, 0.046185024082660675, -0.01602337509393692, -0.006715679541230202, -0.0016641925321891904, 0.033696215599775314, 0.011958622373640537, -0.05608181655406952, 0.012960083782672882, 0.014197182841598988, -0.00022183098189998418, -0.01089825201779604, -0.00050073047168552...
[ "Exception", "Failed", "False", "Install", "NMS", "None", "_load_cv_utils_kernel_once", "and", "as", "be", "better", "community", "cuda", "cv", "cv_utils", "cv_utils_kernel", "def", "e", "except", "f", "filling", "for", "get_kernel", "global", "hole", "if", "insta...
sam3_video/modeling_sam3_video.py:Sam3VideoInferenceCache
[ -0.00024386259610764682, 0.02069651335477829, -0.0219405647367239, 0.015720302239060402, -0.0010956146288663149, -0.004834841005504131, 0.022732235491275787, -0.07057171314954758, -0.0012864635791629553, 0.022053660824894905, -0.0017459148075431585, 0.011196473613381386, 0.001809531124308705...
[ "ModelInferenceCache", "None", "Tensor", "True", "__init__", "_vision_features", "and", "cache_vision_features", "cached", "class", "clear", "clear_all", "cpu", "def", "elif", "else", "features", "for", "frame_idx", "get_vision_features", "if", "in", "inference_device", ...
sam3_video/modeling_sam3_video.py:Sam3VideoInferenceSession
[ -0.00004972903843736276, 0.027480222284793854, -0.04980089142918587, 0.020862536504864693, -0.00021206038945820183, 0.021086864173412323, 0.05204417556524277, -0.06281193345785141, 0.007234589196741581, -0.0021170987747609615, 0.030957311391830444, 0.026022087782621384, -0.003729458665475249...
[ "All", "Cannot", "False", "ModelInferenceCache", "ModelInferenceSession", "None", "OrderedDict", "RuntimeError", "Tensor", "True", "__init__", "_map_keys", "_obj_id_to_idx", "_obj_idx_to_id", "add_mask_inputs", "add_new_frame", "add_prompt", "and", "append", "as", "cache", ...
sam3_video/modeling_sam3_video.py:Sam3VideoSegmentationOutput
[ -0.00029969969182275236, 0.02934890426695347, -0.035357341170310974, -0.01975851319730282, -0.001466000801883638, 0.04136577621102333, 0.042521245777606964, -0.059622183442115784, 0.0174475759267807, -0.01421226467937231, 0.01143913995474577, 0.05546249449253082, -0.0034808493219316006, -0...
[ "FloatTensor", "ModelOutput", "ModelSegmentationOutput", "None", "class", "float", "frame_idx", "obj_id_to_mask", "obj_id_to_score", "obj_id_to_tracker_score", "object_ids", "r", "removed_obj_ids", "suppressed_obj_ids", "torch" ]
sam3_video/modeling_sam3_video.py:Sam3VideoPreTrainedModel
[ -0.0002458768431097269, 0.04196297749876976, 0.0024944026954472065, 0.020069250836968422, -0.0012329475721344352, 0.030103877186775208, 0.04127880185842514, -0.047436412423849106, -0.0015607833629474044, 0.02554268203675747, 0.022007757797837257, 0.02029731124639511, -0.0032070891465991735, ...
[ "Model", "ModelConfig", "ModelPreTrainedModel", "PreTrainedModel", "True", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "base_model_prefix", "class", "config_class", "input_modalities", "main_input_name", "pixel_values", "text", "vid...
sam3_video/modeling_sam3_video.py:Sam3VideoModel
[ -0.00035009559360332787, 0.04481223598122597, -0.019021861255168915, -0.011669853702187538, -0.001502493629232049, 0.0737534761428833, 0.01937195658683777, -0.047613002359867096, 0.007060261443257332, -0.010269471444189548, 0.026257170364260674, 0.028824537992477417, -0.001765065360814333, ...
[ "ALWAYS_OCCLUDED", "Any", "AutoModel", "False", "H_mask", "ModelModel", "ModelPreTrainedModel", "ModelSegmentationOutput", "ModelVisionNeck", "Modele", "NEVER_OCCLUDED", "NO_OBJ_LOGIT", "None", "Propagate", "Removing", "Suppressing", "Tensor", "True", "ValueError", "W_mask", ...
sam3_video/modeling_sam3_video.py:fast_diag_box_iou
[ 0.00019801610324066132, -0.03486517071723938, 0.02419918030500412, -0.026263564825057983, 0.0012400646228343248, 0.0337182879447937, 0.01232896652072668, -0.007684099953621626, 0.03142452612519264, -0.01640039123594761, 0.028901390731334686, -0.011468805372714996, 0.00003180801650159992, -...
[ "Model_diag_box_iou", "area1", "area2", "box1_XY", "box1_xy", "box2_XY", "box2_xy", "boxes1", "boxes2", "clamp", "def", "inter", "iou", "lt", "max", "min", "prod", "rb", "return", "torch", "union" ]
sam3_video/modeling_sam3_video.py:mask_iou
[ 0.000197233326616697, 0.03488519787788391, 0.020196693018078804, -0.009467200376093388, 0.0011762278154492378, 0.02260652557015419, 0.01790161430835724, -0.02134423330426216, 0.02719668298959732, -0.02260652557015419, 0.058065492659807205, 0, -0.0028401599265635014, -0.002954914001747966, ...
[ "H", "M", "Model_iou", "N", "W", "_", "clamp", "def", "dim", "float", "gt_Models", "gt_flat", "intersection", "ious", "min", "pred_Models", "pred_flat", "return", "shape", "sum", "union", "view" ]
sam3_video/modeling_sam3_video.py:nms_masks
[ -0.00006891156226629391, 0.0420246385037899, -0.019776301458477974, -0.0323612205684185, -0.0002949590270873159, 0.027978971600532532, -0.0014607495395466685, -0.04067625477910042, 0.007416112814098597, -0.027978971600532532, 0.028990259394049644, 0.007079016882926226, -0.0022051699925214052...
[ "Exception", "Failed", "Model", "Model_masks", "True", "_load_cv_utils_kernel_once", "as", "cumsum", "cv_utils_kernel", "def", "dim", "e", "except", "f", "generic_Model", "if", "iou_threshold", "ious", "is_valid", "isin", "keep", "kept_inds", "kernels", "library", "lo...
sam3_video/modeling_sam3_video.py:fill_holes_in_mask_scores
[ -0.0000685307095409371, 0.003508772235363722, -0.02829655073583126, -0.040747031569480896, -0.000583616376388818, 0.012507075443863869, 0.02840973623096943, -0.04685908928513527, 0.02263724058866501, -0.0029003964737057686, -0.0014501982368528843, 0.015506509691476822, -0.0006402094732038677...
[ "Model_holes", "Model_holes_in_mask_scores", "True", "_", "_get_connected_components_with_padding", "areas_bg", "areas_fg", "bg_area_thresh", "clamp_", "def", "dim", "dtype", "fg_area_thresh", "floor_divide_", "if", "int32", "keepdim", "mask", "mask_bg", "mask_fg", "max", "...
sam3_video/modeling_sam3_video.py:_get_connected_components_with_padding
[ -0.0002181246381951496, -0.0034045251086354256, 0.0015455689281225204, -0.033959779888391495, -0.0011324675288051367, 0.01880323514342308, 0.0026495466008782387, -0.05766325443983078, 0.011680797673761845, -0.006979988422244787, 0.019600948318839073, 0.018005521968007088, -0.0041595036163926...
[ "Exception", "F", "Failed", "H", "Hole", "True", "W", "_", "_get_connected_components_with_padding", "_load_cv_utils_kernel_once", "and", "as", "be", "cc_2d", "components", "compute", "connected", "constant", "contiguous", "counts", "cv_utils_kernel", "def", "dtype", "e...
autoformer/modeling_autoformer.py:AutoFormerDecoderOutput
[ -0.0002448121376801282, 0.05737244710326195, 0.012557422742247581, 0.004349015653133392, -0.001360867521725595, 0.04792557656764984, 0.08340893685817719, -0.028686223551630974, 0.014112699776887894, -0.002260912209749222, 0.0021313056349754333, 0.032257601618766785, -0.003312164219096303, ...
[ "ModelDecoderOutput", "ModelOutput", "None", "attentions", "class", "cross_attentions", "hidden_states", "last_hidden_state", "past_key_values", "r", "trend" ]
autoformer/modeling_autoformer.py:AutoformerModelOutput
[ -0.0002737418399192393, 0.03204220160841942, 0.006166395265609026, 0.01659739948809147, -0.001311079366132617, 0.06777271628379822, 0.07468830049037933, -0.008586849085986614, 0.010776784271001816, 0.0029535305220633745, 0.014522724784910679, 0.03088960610330105, -0.0014047279255464673, 0....
[ "ModelModelOutput", "ModelOutput", "None", "class", "cross_attentions", "decoder_attentions", "decoder_hidden_states", "encoder_attentions", "encoder_hidden_states", "encoder_last_hidden_state", "last_hidden_state", "loc", "past_key_values", "r", "scale", "static_features", "trend" ]
autoformer/modeling_autoformer.py:AutoformerFeatureEmbedder
[ -0.00017565416055731475, 0.015252782963216305, 0.014348914846777916, 0.01920720748603344, -0.0003001126169692725, 0.016382619738578796, 0.008869211189448833, -0.02383953519165516, 0.011524325236678123, -0.0036437204107642174, 0.002132565015926957, -0.014800848439335823, 0.0031494172289967537...
[ "Embedding", "ModelFeatureEmbedder", "Module", "ModuleList", "__init__", "c", "cardinalities", "cat", "cat_feature_slice", "cat_feature_slices", "chunk", "class", "d", "def", "dim", "else", "embed", "embedders", "embedding_dims", "features", "for", "forward", "if", "in"...
autoformer/modeling_autoformer.py:AutoformerStdScaler
[ -0.0002436667709844187, 0.061923377215862274, -0.014057972468435764, 0.022993605583906174, -0.0008928519673645496, 0.037108492106199265, 0.008366485126316547, -0.0015224727103486657, 0.008594145067036152, 0.02766062505543232, 0.028457432985305786, 0.02811594493687153, 0.0031445464119315147, ...
[ "ModelStdScaler", "Module", "True", "__init__", "clamp_min", "class", "config", "data", "def", "denominator", "dim", "else", "forward", "hasattr", "if", "keepdim", "loc", "minimum_scale", "nn", "observed_indicator", "return", "scale", "scaling_dim", "self", "sqrt", ...
autoformer/modeling_autoformer.py:AutoformerMeanScaler
[ -0.0002414414193481207, 0.05749146267771721, -0.0013563326792791486, 0.029995545744895935, -0.001086486387066543, 0.03158621862530708, 0.002783677540719509, 0.005822998937219381, 0.002556438557803631, 0.030222784727811813, 0.01590672880411148, 0.007072813343256712, 0.002954106777906418, 0....
[ "ModelMeanScaler", "Module", "None", "True", "__init__", "abs", "batch_observations", "batch_sum", "clamp", "class", "config", "data", "def", "default_scale", "dim", "else", "forward", "hasattr", "if", "is", "keepdim", "min", "minimum_scale", "nn", "not", "num_obser...
autoformer/modeling_autoformer.py:AutoformerNOPScaler
[ -0.00023304640490096062, 0.0414431206882, -0.010189997963607311, 0.04121541231870651, -0.0010531562147662044, 0.033473290503025055, 0.017192063853144646, -0.008823741227388382, 0.010588490404188633, 0.03324558213353157, 0.003586424048990011, -0.0070874569937586784, 0.0005479258834384382, 0...
[ "False", "ModelNOPScaler", "Module", "None", "True", "__init__", "class", "config", "data", "def", "dim", "else", "forward", "hasattr", "if", "keepdim", "loc", "mean", "nn", "observed_indicator", "ones_like", "requires_grad", "return", "scale", "scaling_dim", "self"...
autoformer/modeling_autoformer.py:weighted_average
[ -0.00007938854105304927, 0.033500220626592636, 0.03528689965605736, 0.005471702665090561, -0.0003873462846968323, 0.03484022989869118, 0.004159610718488693, -0.008933392353355885, 0.004103776998817921, -0.0012492790119722486, 0.01133424136787653, -0.015521768480539322, -0.0021775143686681986...
[ "Model_average", "Model_tensor", "None", "clamp", "def", "dim", "else", "if", "input_tensor", "is", "min", "not", "return", "sum", "sum_weights", "torch", "weights", "where", "zeros_like" ]
autoformer/modeling_autoformer.py:nll
[ 0.00003988773823948577, 0.028024237602949142, 0.02416665107011795, 0.009019947610795498, -0.00017018768994603306, 0.06262906640768051, 0.04243346303701401, 0.025414694100618362, 0.01667839288711548, -0.020082145929336548, 0.025301234796643257, -0.028931906446814537, 0.001205496140755713, 0...
[ "Model", "def", "input", "log_prob", "return", "target" ]
autoformer/modeling_autoformer.py:AutoformerSinusoidalPositionalEmbedding
[ -0.00016163897817023098, 0.040929801762104034, 0.012087784707546234, 0.006043892353773117, -0.0008890143944881856, -0.002825168427079916, 0.04025513306260109, -0.009838894940912724, 0.010457339696586132, -0.011300673708319664, 0.019115567207336426, 0.015067564323544502, 0.0025862236507236958...
[ "Embedding", "False", "FloatTensor", "ModelSinusoidalPositionalEmbedding", "None", "True", "__init__", "_freeze", "arange", "array", "class", "cos", "create_weight", "def", "device", "dim", "dtype", "else", "embedding_dim", "empty", "for", "forward", "if", "in", "inpu...
autoformer/modeling_autoformer.py:AutoformerValueEmbedding
[ -0.0001237391261383891, 0.02477611042559147, 0.0004772795073222369, 0.02398417890071869, -0.0004808149242307991, 0.040954116731882095, 0.042537979781627655, -0.02353164739906788, 0.0033939876593649387, 0.0266993697732687, -0.0042990511283278465, -0.03009335696697235, 0.0014000198571011424, ...
[ "Linear", "ModelValueEmbedding", "Module", "__init__", "class", "d_model", "def", "feature_size", "forward", "in_features", "nn", "out_features", "return", "self", "super", "value_projection", "x" ]
autoformer/modeling_autoformer.py:AutoformerSeriesDecompositionLayer
[ -0.0003124829090666026, 0.018513273447752, 0.029255542904138565, 0.010228011757135391, -0.0010856549488380551, 0.03725510463118553, 0.02674139477312565, -0.0037997921463102102, 0.011656505055725574, 0.010056592524051666, 0.02022746577858925, -0.00665677897632122, -0.0026141428388655186, -0...
[ "AvgPool1d", "ModelSeriesDecompositionLayer", "Module", "__init__", "avg", "cat", "class", "config", "def", "dim", "end", "forward", "front", "kernel_size", "moving_average", "nn", "num_of_pads", "padding", "permute", "repeat", "return", "self", "stride", "super", "to...
autoformer/modeling_autoformer.py:AutoformerLayernorm
[ 0.000033383348636562005, 0.043121688067913055, 0.037312548607587814, 0.016645418480038643, -0.00018939051369670779, 0.028263697400689125, -0.007931709289550781, 0.007261424325406551, 0.015193133614957333, 0.018656274303793907, 0.0010124101536348462, 0.016533704474568367, 0.006311853416264057...
[ "LayerNorm", "ModelLayernorm", "Module", "__init__", "bias", "class", "config", "d_model", "def", "dim", "forward", "layernorm", "mean", "nn", "repeat", "return", "self", "shape", "super", "torch", "unsqueeze", "x", "x_hat" ]
autoformer/modeling_autoformer.py:AutoformerAttention
[ -0.00033794186310842633, 0.03888488560914993, 0.05384061112999916, -0.019097311422228813, -0.0013445772929117084, 0.02576986514031887, 0.02680526115000248, -0.029681362211704254, -0.0009347328450530767, 0.007938038557767868, 0.036814093589782715, 0.027380481362342834, -0.0019413682166486979,...
[ "Attention", "EncoderDecoderCache", "False", "Linear", "ModelAttention", "Module", "None", "True", "ValueError", "_", "__init__", "and", "arange", "attention_mask", "attn_output", "attn_weights", "attn_weights_reshaped", "autocorrelation_factor", "autocorrelations", "autocorrel...
autoformer/modeling_autoformer.py:AutoformerEncoderLayer
[ -0.0002092567301588133, 0.032005637884140015, 0.03109767846763134, 0.013789664022624493, -0.0006987046799622476, 0.029735736548900604, 0.022699035704135895, -0.012200730852782726, 0.005646384786814451, 0.004624928347766399, 0.006781336851418018, 0.027806317433714867, 0.00067387759918347, 0...
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelAttention", "ModelEncoderLayer", "ModelLayernorm", "ModelSeriesDecompositionLayer", "_", "__init__", "activation_dropout", "activation_fn", "activation_function", "and", "any", "attention_dropout", "atten...
autoformer/modeling_autoformer.py:AutoformerDecoderLayer
[ -0.00019368017092347145, 0.044148463755846024, 0.015621764585375786, -0.003396035870537162, -0.0005164804751984775, 0.03282834589481354, 0.05320456251502037, -0.030790723860263824, 0.0021932730451226234, -0.006876972503960133, 0, 0.02818709798157215, -0.0035658376291394234, 0.0069052726030...
[ "ACT2FN", "Conv1d", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelAttention", "ModelDecoderLayer", "ModelLayernorm", "ModelSeriesDecompositionLayer", "None", "True", "__init__", "activation_dropout", "activation_fn", "activation_function", "attention_dropout", ...
autoformer/modeling_autoformer.py:AutoformerPreTrainedModel
[ -0.00025913555873557925, 0.04611903056502342, -0.0006283149705268443, 0.0053673009388148785, -0.0010649405885487795, 0.0308974776417017, 0.022491546347737312, -0.01238171011209488, -0.00047567347064614296, 0.014028818346560001, 0.00021476302936207503, -0.0018742955289781094, -0.0006034663529...
[ "Conv1d", "Embedding", "False", "LayerNorm", "Linear", "ModelConfig", "ModelPreTrainedModel", "ModelSinusoidalPositionalEmbedding", "None", "PreTrainedModel", "Tensor", "True", "_attn_implementation", "_init_weights", "_is_hf_initialized", "_prepare_4d_attention_mask", "_prepare_4d_a...
autoformer/modeling_autoformer.py:AutoformerEncoder
[ -0.0001503712119301781, 0.0387214720249176, 0.01992684043943882, 0.012793936766684055, -0.000789006648119539, 0.043250299990177155, 0.02139870822429657, -0.02117226831614971, 0.0027880591806024313, -0.010982406325638294, 0.007642395794391632, 0.0036513670347630978, 0.0009482232271693647, -...
[ "BaseModelOutput", "False", "LayerNorm", "ModelEncoder", "ModelEncoderLayer", "ModelPreTrainedModel", "ModelSinusoidalPositionalEmbedding", "ModelValueEmbedding", "ModuleList", "None", "True", "_", "__init__", "_update_full_mask", "all_attentions", "attention_mask", "attentions", "...
autoformer/modeling_autoformer.py:AutoformerDecoder
[ -0.00021881390421185642, 0.06127503886818886, 0.004401285666972399, -0.0009038354037329555, -0.0009860022692009807, 0.029037052765488625, 0.07042057067155838, -0.030180243775248528, 0.006459029391407967, -0.010460197925567627, 0.0005251533584669232, 0.007545060943812132, -0.00374395074322819...
[ "DynamicCache", "EncoderDecoderCache", "False", "LayerNorm", "Linear", "ModelDecoder", "ModelDecoderLayer", "ModelDecoderOutput", "ModelPreTrainedModel", "ModelSinusoidalPositionalEmbedding", "ModelValueEmbedding", "ModuleList", "None", "Setting", "True", "__init__", "_prepare_4d_att...
autoformer/modeling_autoformer.py:AutoformerModel
[ -0.0003119048778899014, 0.037399906665086746, -0.005879586562514305, 0.014799349009990692, -0.0011759172193706036, 0.04566000774502754, 0.022600555792450905, -0.021109148859977722, 0.004789711907505989, -0.013078494928777218, 0.010726659558713436, 0.004990478046238422, 0.0012834706576541066,...
[ "BaseModelOutput", "ModelDecoder", "ModelDecoderOutput", "ModelEncoder", "ModelFeatureEmbedder", "ModelMeanScaler", "ModelModel", "ModelModelOutput", "ModelNOPScaler", "ModelPreTrainedModel", "ModelSeriesDecompositionLayer", "ModelStdScaler", "None", "True", "ValueError", "_", "__ini...
autoformer/modeling_autoformer.py:AutoformerForPrediction
[ -0.0005060589173808694, 0.03095889836549759, -0.0038698622956871986, 0.02286195568740368, -0.002217728877440095, 0.025124337524175644, 0.05072496458888054, -0.022981029003858566, -0.01143097784370184, 0.0017116699600592256, 0.015479449182748795, -0.013693359680473804, -0.002455874113366008, ...
[ "False", "ModelForPrediction", "ModelModel", "ModelPreTrainedModel", "NegativeBinomialOutput", "None", "NormalOutput", "SampleTSPredictionOutput", "Seq2SeqTSPredictionOutput", "StudentTOutput", "True", "_", "__init__", "auto_docstring", "cat", "class", "config", "context_length", ...
dia/modeling_dia.py:DiaPreTrainedModel
[ -0.0003079982416238636, 0.0348128043115139, 0.016138385981321335, 0.017867499962449074, -0.0012680160580202937, 0.03504335507750511, 0.03389061242341995, -0.029164370149374008, -0.0025216229259967804, 0.003155631013214588, 0.008933749981224537, 0.0038616852834820747, -0.005158519838005304, ...
[ "ModelConfig", "ModelDecoderLayer", "ModelEncoderLayer", "ModelMultiChannelEmbedding", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_init_weights", "_no_split_modules", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "arange", "base_model...
dia/modeling_dia.py:DiaMultiChannelEmbedding
[ -0.00019506823446135968, 0.015704317018389702, 0.018867775797843933, 0.009151436388492584, -0.00029480899684131145, 0.021353352814912796, 0.035701900720596313, -0.05716823413968086, 0.006722351536154747, -0.03005286678671837, 0.03750959038734436, 0.023725947365164757, -0.005225357133895159, ...
[ "Embedding", "False", "ModelMultiChannelEmbedding", "Module", "__init__", "arange", "audio_codes", "class", "config", "def", "device", "dim", "dtype", "embed", "embeds", "forward", "hidden_size", "long", "nn", "num_channels", "offsets", "persistent", "register_buffer", ...
dia/modeling_dia.py:DiaMLP
[ -0.00040280685061588883, 0.012889819219708443, 0.02554420940577984, 0.019540730863809586, -0.0015155838336795568, 0.05179470777511597, 0.028369374573230743, -0.020247023552656174, -0.004149462562054396, -0.003914032131433487, 0.040729474276304245, -0.05014669522643089, 0.00019404622435104102...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "__init__", "activation_fn", "chunk", "class", "config", "def", "dim", "down_proj", "forward", "gate", "gate_up_proj", "hidden_act", "hidden_size", "hidden_states", "intermeModelte_size", "nn", "return", "self", "super", "up_sta...
dia/modeling_dia.py:DiaRMSNorm
[ -0.00009884802420856431, 0.04179859161376953, 0.03230918198823929, 0.053095508366823196, -0.00046952810953371227, 0.03931327164173126, 0.022028988227248192, -0.02937198430299759, 0.007964326068758965, 0.042476408183574677, 0.019995542243123055, 0.006975846365094185, 0.0028524715453386307, ...
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", "to", "torch", "tuple", ...
dia/modeling_dia.py:DiaRotaryEmbedding
[ -0.00029968636226840317, 0.04852752387523651, 0.0020364229567348957, -0.005228263325989246, -0.0013792794197797775, 0.04136393964290619, 0.040670689195394516, 0.0062681385315954685, -0.0026430170983076096, 0.021144136786460876, 0.0021808501332998276, -0.004043960478156805, -0.001364836702123...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
dia/modeling_dia.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
dia/modeling_dia.py:apply_rotary_pos_emb
[ -0.0001444592053303495, 0.027112245559692383, 0.028019767254590988, 0.0028360087890177965, -0.0005707467789761722, 0.021667107939720154, 0.046510547399520874, -0.002183726755902171, 0.01293220091611147, 0.03652779385447502, 0.007884104736149311, 0.0017654155381023884, -0.0007834474672563374,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
dia/modeling_dia.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
dia/modeling_dia.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
dia/modeling_dia.py:DiaSelfAttention
[ -0.00006637266051257029, 0.04073435440659523, 0.023517901077866554, -0.0058794752694666386, -0.0004712019581347704, 0.020029598847031593, 0.03308259695768356, -0.010858743451535702, 0.005119925830513239, 0.017103927209973335, 0.011533898301422596, 0.01969202235341072, -0.00003120833935099654...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelSelfAttention", "Module", "None", "Tensor", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs", "cache_position", ...
dia/modeling_dia.py:DiaCrossAttention
[ -0.00019110992434434593, 0.03200644627213478, 0.043894555419683456, -0.003915074281394482, -0.0008180219447240233, 0.01851801574230194, 0.02697686292231083, -0.028348568826913834, -0.0009644799865782261, 0.02149004302918911, 0.01977541297674179, 0.02571946755051613, -0.00016878399765118957, ...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelCrossAttention", "Module", "None", "True", "__init__", "_attn_implementation", "and", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "class", "config", "contiguous", "cross_attention...
dia/modeling_dia.py:DiaEncoderLayer
[ -0.00011289615940768272, 0.0235952977091074, 0.01953103579580784, 0.01795049011707306, -0.00019315826648380607, 0.046287424862384796, 0.03025617077946663, -0.005164999049156904, 0.006124616600573063, 0.010894479230046272, 0.008297868072986603, 0.03115933947265148, -0.00034927247907035053, ...
[ "False", "GradientCheckpointingLayer", "ModelEncoderLayer", "ModelMLP", "ModelRMSNorm", "ModelSelfAttention", "None", "Tensor", "__init__", "attention_mask", "class", "config", "def", "eps", "forward", "hidden_size", "hidden_states", "is_causal", "kwargs", "layer_idx", "mlp",...
dia/modeling_dia.py:DiaEncoder
[ -0.00006470664811786264, 0.036510396748781204, 0.012620877474546432, 0.012508191168308258, -0.00046131000271998346, 0.057019323110580444, 0.027833543717861176, -0.009916404262185097, 0.011494013480842113, 0.01408580131828785, 0.01949474960565567, 0.027608171105384827, -0.0020142695866525173,...
[ "BaseModelOutput", "Embedding", "False", "ModelEncoder", "ModelEncoderLayer", "ModelPreTrainedModel", "ModelRMSNorm", "ModelRotaryEmbedding", "ModuleList", "None", "__init__", "all_attentions", "arange", "attention_mask", "attentions", "auto_docstring", "can_return_tuple", "class",...
dia/modeling_dia.py:DiaDecoderLayer
[ -0.0001779330341378227, 0.047363825142383575, 0.005920478142797947, 0.003994198050349951, -0.0007081911317072809, 0.03761911392211914, 0.04079180955886841, -0.03082047961652279, 0.007365188095718622, 0.001671331119723618, 0.004305802285671234, 0.024361776188015938, -0.001869624713435769, -...
[ "EncoderDecoderCache", "GradientCheckpointingLayer", "ModelCrossAttention", "ModelDecoderLayer", "ModelMLP", "ModelRMSNorm", "ModelSelfAttention", "None", "Tensor", "True", "__init__", "attention_mask", "cache_position", "class", "config", "cross_attention", "cross_attn_weights", "...
dia/modeling_dia.py:DiaDecoder
[ -0.00014194523100741208, 0.06364788860082626, -0.003949780482798815, -0.014557762071490288, -0.0009310196619480848, 0.03385525941848755, 0.04762306809425354, -0.0365636833012104, 0.009874450974166393, -0.0012413596268743277, 0.013316403143107891, 0.023021576926112175, -0.0036112279631197453,...
[ "BaseModelOutputWithPastAndCrossAttentions", "False", "ModelDecoder", "ModelDecoderLayer", "ModelMultiChannelEmbedding", "ModelPreTrainedModel", "ModelRMSNorm", "ModelRotaryEmbedding", "ModuleList", "None", "__init__", "all_cross_attentions", "all_hidden_states", "all_self_attns", "and",...
dia/modeling_dia.py:DiaModel
[ -0.0000969369793892838, 0.05322498083114624, 0.005417943466454744, -0.004042403772473335, -0.0004140656383242458, 0.03975030034780502, 0.026949357241392136, -0.025377310812473297, 0.015046724118292332, -0.005979388486593962, 0.025601889938116074, 0, -0.0009334022179245949, 0.01420455705374...
[ "BaseModelOutput", "DynamicCache", "EncoderDecoderCache", "False", "ModelDecoder", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "Neither", "None", "Seq2SeqModelOutput", "Setting", "True", "ValueError", "You", "__init__", "and", "attention_mask", "attentions", "auto_doc...
dia/modeling_dia.py:DiaForConditionalGeneration
[ -0.0002146933547919616, 0.04387872666120529, -0.006022016052156687, 0.009669152088463306, -0.0008729096152819693, 0.026915302500128746, 0.03551010414958, -0.04161693528294563, -0.0009117841254919767, -0.022504810243844986, 0.042069293558597565, 0.008312078192830086, -0.004947665613144636, ...
[ "ForMaskedLM", "Linear", "ModelForConditionalGeneration", "ModelGenerationMixin", "ModelModel", "ModelPreTrainedModel", "None", "Seq2SeqLMOutput", "__init__", "attention_mask", "audio", "audio_logits", "auto_docstring", "base_model_prefix", "batch_size", "cache_position", "can_return...
florence2/modeling_florence2.py:drop_path
[ 0, 0.017753854393959045, 0.03775503486394882, -0.02123720571398735, 0, 0.043598074465990067, 0.0339345820248127, -0.027304979041218758, 0.014832334592938423, 0.0017486985307186842, 0.020113544538617134, -0.04944111406803131, 0.0016574009787291288, -0.018315685912966728, -0.05034004524350...
[ "False", "Model_path", "Model_prob", "def", "device", "div", "dtype", "floor_", "if", "input", "keep_prob", "ndim", "not", "or", "output", "rand", "random_tensor", "return", "shape", "torch", "training" ]
florence2/modeling_florence2.py:Florence2VisionDropPath
[ -0.00006749371095793322, 0.024459021165966988, 0.030068887397646904, -0.01430516131222248, -0.00048034486826509237, 0.030966466292738914, 0.03388359770178795, -0.027937138453125954, 0.008134307339787483, 0.00541352154687047, 0.015258838422596455, -0.025805387645959854, 0.0003488635993562639,...
[ "ModelVisionDropPath", "Module", "None", "__init__", "class", "def", "drop_path", "drop_prob", "extra_repr", "f", "forward", "hidden_states", "nn", "p", "return", "self", "super", "training" ]
florence2/modeling_florence2.py:Florence2VisionLearnedAbsolutePositionEmbedding2D
[ -0.00027497709379531443, 0.005285274237394333, 0.027997668832540512, 0.028911879286170006, -0.001271322718821466, 0.030626021325588226, 0.027083460241556168, -0.008227886632084846, 0.008742129430174828, 0.012456106022000313, 0.014398801140487194, 0.021941030398011208, 0.0013070340501144528, ...
[ "Embedding", "ModelVisionLearnedAbsolutePositionEmbedding2D", "Module", "None", "__init__", "arange", "cat", "class", "column_embeddings", "config", "def", "device", "dim", "embed_dim", "embedding_dim", "forward", "height", "height_values", "max_position_embeddings", "nn", "n...
florence2/modeling_florence2.py:Florence2VisionPositionalEmbeddingCosine1D
[ -0.0001842452329583466, 0.036334216594696045, 0.02132660523056984, 0.03362607583403587, -0.0006876137340441346, 0.0031030774116516113, 0.03768828511238098, -0.0007546120323240757, 0.015458967536687851, 0.019521178677678108, 0.023809067904949188, 0.019069822505116463, 0.0011354442685842514, ...
[ "ModelVisionPositionalEmbeddingCosine1D", "Module", "__init__", "arange", "class", "config", "cos", "cosine", "def", "dtype", "emb", "embed_dim", "empty", "exp", "float", "forward", "get_sinusoid_embeddings", "got", "half_dim", "if", "int64", "len_seq", "log", "math", ...
florence2/modeling_florence2.py:Florence2VisionMLP
[ -0.0002199504233431071, 0.033647049218416214, 0.02403360605239868, 0.03959822654724121, -0.0006580630433745682, 0.051271695643663406, 0.02414805255830288, -0.031815916299819946, 0.00044168904423713684, -0.002045717788860202, 0.051958367228507996, -0.021973583847284317, 0.00164515757933259, ...
[ "ACT2FN", "Linear", "ModelVisionMLP", "Module", "__init__", "activation_fn", "activation_function", "class", "config", "def", "embed_dim", "fc1", "fc2", "forward", "hidden_states", "int", "mlp_ratio", "nn", "return", "self", "stage_idx", "super" ]
florence2/modeling_florence2.py:Florence2VisionConvEmbed
[ -0.00008720046753296629, 0.016009652987122536, 0.004397017415612936, 0.03427418693900108, 0, 0.017362581565976143, 0.004256087355315685, -0.017249837517738342, 0.011218031868338585, 0.025818385183811188, 0.04148980602622032, 0.010597939603030682, 0.0004016506427433342, -0.01837727800011635...
[ "Conv2d", "LayerNorm", "ModelVisionConvEmbed", "Module", "__init__", "and", "class", "config", "conv", "def", "dim_norm", "else", "embed_dim", "forward", "hidden_states", "if", "in_channels", "kernel_size", "nn", "norm", "not", "padding", "patch_padding", "patch_prenorm...
florence2/modeling_florence2.py:eager_attention_forward
[ 0.00002095530362566933, 0.025862814858555794, 0.0269921962171793, -0.01146321278065443, 0.00008955634984886274, 0.03704368323087692, 0.060083046555519104, -0.023942869156599045, 0.021458230912685394, 0.014569009654223919, 0.023152301087975502, 0.026879258453845978, 0.003063444746658206, -0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shape", "size", "softmax", "tor...
florence2/modeling_florence2.py:Florence2VisionChannelAttention
[ -0.00021182477939873934, 0.016860540956258774, 0.025974346324801445, -0.0038733675610274076, -0.0003791485505644232, 0.01617700606584549, 0.03144263103604317, -0.038505829870700836, 0.004927151370793581, 0.016860540956258774, 0.05058162286877632, 0.01651877351105213, -0.0006087737274356186, ...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelVisionChannelAttention", "Module", "None", "_", "__init__", "_attn_implementation", "attention_interface", "attention_mask", "batch_size", "class", "config", "def", "dim", "eager_attention_forward", "embed_dim", "forward", "ge...
florence2/modeling_florence2.py:Florence2VisionChannelBlock
[ -0.00014660789747722447, -0.007193560712039471, 0.006226837169378996, -0.024111222475767136, 0, 0.03457458317279816, 0.016832362860441208, -0.04344569519162178, 0.009724101983010769, 0.007847520522773266, 0.03684922680258751, -0.0015638174954801798, 0.0010306979529559612, 0.000646851782221...
[ "Conv2d", "Identity", "LayerNorm", "ModelVisionChannelAttention", "ModelVisionChannelBlock", "ModelVisionDropPath", "ModelVisionMLP", "Module", "__init__", "batch_size", "channel_attn", "class", "config", "conv1", "conv2", "def", "dim_in", "drop_path1", "drop_path2", "drop_path...
florence2/modeling_florence2.py:Florence2VisionWindowAttention
[ -0.0001072902960004285, 0.01726454123854637, 0.03789230436086655, 0.0039517865516245365, -0.00030479204724542797, 0.01793718710541725, 0.04327346384525299, -0.04349767789244652, 0.004540350288152695, 0.021300408989191055, 0.03161429241299629, 0.02410309389233589, -0.0005535303498618305, -0...
[ "ALL_ATTENTION_FUNCTIONS", "F", "False", "Linear", "ModelVisionWindowAttention", "Module", "None", "_", "__init__", "_attn_implementation", "attention_interface", "attention_mask", "batch_size", "class", "config", "contiguous", "def", "dim", "eager_attention_forward", "embed_di...
florence2/modeling_florence2.py:Florence2VisionSpatialBlock
[ -0.000046070334065007046, 0.0021263230592012405, 0.005131526384502649, -0.029938628897070885, 0, 0.03742328658699989, 0.01916525885462761, -0.04490794241428375, 0.013324958272278309, 0.0025232366751879454, 0.03197989985346794, -0.0013608467997983098, 0.0028067464008927345, -0.0128146400675...
[ "Conv2d", "Identity", "LayerNorm", "ModelVisionDropPath", "ModelVisionMLP", "ModelVisionSpatialBlock", "ModelVisionWindowAttention", "Module", "__init__", "batch_size", "class", "config", "conv1", "conv2", "def", "drop_path1", "drop_path2", "drop_path_rate", "else", "embed_dim"...
florence2/modeling_florence2.py:Florence2VisionBlock
[ -0.00011034178169211373, -0.0067215426824986935, 0.0022263338323682547, -0.018831664696335793, 0, 0.04469684138894081, 0.02132742665708065, -0.05195724219083786, 0.01264897920191288, 0.0062677678652107716, 0.031537365168333054, -0.011174210347235203, 0.0027226503007113934, -0.0249576270580...
[ "ModelVisionBlock", "ModelVisionChannelBlock", "ModelVisionSpatialBlock", "Module", "__init__", "channel_block", "channel_drop_path_rate", "class", "config", "def", "drop_path_rate", "forward", "hidden_states", "nn", "return", "self", "spatial_block", "spatial_drop_path_rate", "s...
florence2/modeling_florence2.py:Florence2VisionPreTrainedModel
[ -0.0002202051255153492, 0.0279589481651783, 0.004290448501706123, 0.03432358801364899, -0.0009873714298009872, 0.03182319179177284, 0.04978056624531746, -0.040688224136829376, 0.0034806616604328156, 0.026481442153453827, 0.024094702675938606, 0.016707176342606544, -0.0034096278250217438, 0...
[ "ModelVisionBlock", "ModelVisionChannelAttention", "ModelVisionConfig", "ModelVisionPreTrainedModel", "ModelVisionWindowAttention", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "attentions", "cl...
florence2/modeling_florence2.py:Florence2VisionBackbone
[ -0.00023536886146757752, 0.001330807339400053, 0.0029872378800064325, 0.014610566198825836, -0.0006689431029371917, 0.038508471101522446, 0.004247257485985756, -0.007361913099884987, 0.003397806081920862, -0.0019962110091000795, 0.02299182116985321, -0.009343966841697693, -0.0004601195687428...
[ "BaseModelOutputWithPooling", "Expected", "ModelVisionBackbone", "ModelVisionBlock", "ModelVisionConvEmbed", "ModelVisionPreTrainedModel", "ModuleList", "ValueError", "__init__", "append", "block", "block_idx", "blocks", "channel_drop_path_rate", "check_model_inputs", "class", "confi...
florence2/modeling_florence2.py:Florence2MultiModalProjector
[ -0.00032657457632012665, 0.03403513878583908, 0.02695400081574917, 0.033121444284915924, -0.0009422479197382927, 0.03472040966153145, 0.02147183194756508, -0.028895603492856026, 0.004682686645537615, 0.028552968055009842, 0.029466662555933, 0.015418603084981441, -0.001277745352126658, -0.0...
[ "LayerNorm", "Linear", "ModelMultiModalProjector", "ModelVisionLearnedAbsolutePositionEmbedding2D", "ModelVisionPositionalEmbeddingCosine1D", "Module", "__init__", "cat", "class", "config", "def", "dim", "embed_dim", "flatten", "forward", "image_features", "image_position_embed", "...
florence2/modeling_florence2.py:Florence2Seq2SeqModelOutput
[ -0.00009063803008757532, 0.034917641431093216, 0.03153851255774498, 0.013065955601632595, -0.0005174287362024188, 0.027708835899829865, 0.04167589172720909, -0.04618139564990997, 0.018923107534646988, -0.008504134602844715, 0.024780260398983955, 0.01255908701568842, -0.002844098024070263, ...
[ "ModelSeq2SeqModelOutput", "None", "Seq2SeqModelOutput", "class", "image_hidden_states", "r" ]
florence2/modeling_florence2.py:Florence2Seq2SeqLMOutput
[ -0.00008733053982723504, 0.030782479792833328, 0.017750481143593788, 0.0016851723194122314, -0.0005827887798659503, 0.03415282443165779, 0.03617503121495247, -0.0451626181602478, 0.02190724015235901, -0.018986275419592857, 0.03662441298365593, 0.027973860502243042, -0.0012638792395591736, ...
[ "ModelSeq2SeqLMOutput", "None", "Seq2SeqLMOutput", "class", "image_hidden_states", "r" ]
florence2/modeling_florence2.py:Florence2PreTrainedModel
[ -0.0003231237642467022, 0.05406817048788071, 0.00982007384300232, 0.03026893362402916, -0.0015307762660086155, 0.022874996066093445, 0.03096211515367031, -0.012130679562687874, 0.00023286572832148522, 0.014556815847754478, 0.011668558232486248, 0.0006859610439278185, -0.002281723078340292, ...
[ "False", "ModelConfig", "ModelPreTrainedModel", "ModelVisionPositionalEmbeddingCosine1D", "PreTrainedModel", "True", "_can_compile_fullgraph", "_init_weights", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "ba...