identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
visual_bert/modeling_visual_bert.py:VisualBertForPreTrainingOutput
[ -0.0002650046953931451, 0.018440058454871178, 0.029822809621691704, 0.016277335584163666, -0.001131160999648273, 0.047352246940135956, 0.03938432037830353, -0.004866126459091902, 0.013147078454494476, -0.01639116369187832, 0.01081361435353756, 0.017870919778943062, -0.0014370724093168974, ...
[ "ModelForPreTrainingOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "loss", "prediction_logits", "r", "seq_relationship_logits" ]
visual_bert/modeling_visual_bert.py:VisualBertModel
[ 0.00002271621451654937, 0.030866093933582306, 0.01867622323334217, 0.008778943680226803, 0, 0.04719381034374237, -0.007269188761711121, -0.02158389985561371, 0.005927184596657753, -0.00048228271771222353, 0.0380234494805336, 0.008275691419839859, 0.0023485070560127497, 0.021136565133929253...
[ "BaseModelOutputWithPooling", "ModelEmbeddings", "ModelEncoder", "ModelLayer", "ModelModel", "ModelPooler", "ModelPreTrainedModel", "Model_attention_mask", "Model_embedding_output", "Model_embeds", "Model_input_shape", "Model_token_type_ids", "None", "True", "__init__", "add_pooling_la...
visual_bert/modeling_visual_bert.py:VisualBertForPreTraining
[ 0.00001091428566724062, 0.05226694047451019, 0.02301969565451145, 0.01757058873772621, -0.00009296148346038535, 0.032249815762043, -0.0046150595881044865, -0.006422161124646664, 0.004364845342934132, -0.0034195922780781984, 0.03047051467001438, 0.037810128182172775, 0.0014665324706584215, ...
[ "CrossEntropyLoss", "Found", "Model", "ModelForPreTraining", "ModelForPreTrainingOutput", "ModelModel", "ModelPreTrainedModel", "ModelPreTrainingHeads", "Model_attention_mask", "Model_embeds", "Model_token_type_ids", "None", "The", "ValueError", "__init__", "_tied_weights_keys", "and...
visual_bert/modeling_visual_bert.py:VisualBertForMultipleChoice
[ -0.00019843553309328854, 0.05578872933983803, 0.020523903891444206, 0.03379073366522789, -0.0008008290897123516, 0.047851309180259705, 0.034017518162727356, -0.02381226234138012, 0.0025796617846935987, 0.005045931786298752, 0.047170959413051605, 0.001162265194579959, -0.0013890486443415284, ...
[ "CrossEntropyLoss", "Dropout", "Linear", "Model", "ModelForMultipleChoice", "ModelModel", "ModelPreTrainedModel", "Model_attention_mask", "Model_embeds", "Model_token_type_ids", "MultipleChoiceModelOutput", "None", "_", "__init__", "attention_mask", "attentions", "auto_docstring", ...
visual_bert/modeling_visual_bert.py:VisualBertForQuestionAnswering
[ -0.0001688133634161204, 0.048618245869874954, 0.015643371269106865, 0.004839316476136446, -0.0005767789552919567, 0.040290120989084244, 0.0290358979254961, 0.009566090069711208, 0.006133552175015211, 0.019919976592063904, 0.03578843176364899, 0.025322003290057182, 0.00042906729504466057, 0...
[ "Dropout", "KLDivLoss", "Linear", "LogSoftmax", "Model", "ModelForQuestionAnswering", "ModelModel", "ModelPreTrainedModel", "Model_attention_mask", "Model_embeds", "Model_token_type_ids", "None", "SequenceClassifierOutput", "__init__", "attention_mask", "attentions", "auto_docstring"...
visual_bert/modeling_visual_bert.py:VisualBertForVisualReasoning
[ -0.00015001701831351966, 0.05037748068571091, 0.011069491505622864, -0.012368462048470974, -0.0004941737279295921, 0.06777239590883255, 0.013159140013158321, 0.0015107595827430487, 0.0030215191654860973, 0.020331718027591705, 0.03998571261763573, 0.008019733242690563, 0.0019060985650867224, ...
[ "CrossEntropyLoss", "Dropout", "Linear", "Model", "ModelForModelReasoning", "ModelModel", "ModelPreTrainedModel", "Model_attention_mask", "Model_embeds", "Model_token_type_ids", "None", "SequenceClassifierOutput", "__init__", "attention_mask", "attentions", "auto_docstring", "class",...
visual_bert/modeling_visual_bert.py:VisualBertRegionToPhraseAttention
[ -0.00027977090212516487, 0.0223531611263752, 0.04698725789785385, -0.0140847722068429, -0.0011475953506305814, 0.012773234397172928, 0.05018056556582451, -0.02406386099755764, -0.001568142673932016, 0.021782927215099335, 0.02372172102332115, 0.02782740443944931, -0.00048826259444467723, -0...
[ "Dropout", "Linear", "ModelRegionToPhraseAttention", "Module", "The", "ValueError", "_", "__init__", "a", "all_head_size", "attention", "attention_head_size", "attention_mask", "attention_probs_dropout_prob", "attention_scores", "batch_size", "class", "config", "def", "dropout"...
visual_bert/modeling_visual_bert.py:VisualBertForRegionToPhraseAlignment
[ -0.0002897057856898755, 0.0412026010453701, 0.02426375448703766, -0.01831226795911789, -0.0012375087244436145, 0.03845576196908951, 0.03502221032977104, -0.03502221032977104, -0.0067812614142894745, -0.014993168413639069, 0.04349163547158241, 0.05264776945114136, -0.001216048956848681, -0....
[ "Dropout", "KLDivLoss", "LogSoftmax", "Model", "ModelForRegionToPhraseAlignment", "ModelModel", "ModelPreTrainedModel", "ModelPreTrainingHeads", "ModelRegionToPhraseAttention", "Model_attention_mask", "Model_embeds", "Model_features", "Model_token_type_ids", "None", "SequenceClassifierOu...
blenderbot_small/modeling_blenderbot_small.py:shift_tokens_right
[ -0.0001357423752779141, 0.026739485561847687, 0.00459761219099164, -0.04174518957734108, -0.0005782272783108056, 0.05189942196011543, 0.04039129242300987, -0.057089366018772125, 0.006572046782821417, 0.007784913759678602, 0.020759768784046173, -0.009872172959148884, -0.0006698974757455289, ...
[ "Model_tokens_right", "Modeled_input_ids", "None", "clone", "decoder_start_token_id", "def", "if", "input_ids", "is", "masked_fill_", "new_zeros", "pad_token_id", "return", "shape" ]
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallLearnedPositionalEmbedding
[ -0.000267448223894462, 0.013325742445886135, -0.006404397543519735, 0.004796118475496769, -0.0016226388979703188, 0.02791513316333294, 0.02676636166870594, -0.006576713174581528, 0.004939714912325144, -0.0008328589028678834, 0.009936868213117123, 0.005313065368682146, 0.0017446958227083087, ...
[ "Embedding", "ModelLearnedPositionalEmbedding", "None", "__init__", "arange", "class", "def", "device", "dtype", "embedding_dim", "forward", "if", "input_ids_shape", "is", "long", "nn", "num_embeddings", "past_key_values_length", "position_ids", "return", "self", "seq_len",...
blenderbot_small/modeling_blenderbot_small.py:eager_attention_forward
[ 0.00002095530362566933, 0.025862814858555794, 0.0269921962171793, -0.01146321278065443, 0.00008955634984886274, 0.03704368323087692, 0.060083046555519104, -0.023942869156599045, 0.021458230912685394, 0.014569009654223919, 0.023152301087975502, 0.026879258453845978, 0.003063444746658206, -0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shape", "size", "softmax", "tor...
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallAttention
[ -0.0001932329178089276, 0.04201360046863556, 0.04201360046863556, -0.006776387337595224, -0.0007764610345475376, 0.02010328322649002, 0.029138466343283653, -0.028234947472810745, 0.0020611512009054422, 0.014908052049577236, 0.008188134990632534, 0.009261062368750572, -0.0016235094517469406, ...
[ "ALL_ATTENTION_FUNCTIONS", "EncoderDecoderCache", "False", "Instantiating", "Linear", "ModelAttention", "Module", "None", "Please", "True", "ValueError", "__class__", "__init__", "__name__", "_attn_implementation", "a", "and", "attention_interface", "attention_mask", "attn_outp...
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallEncoderLayer
[ -0.00012346176663413644, 0.036347147077322006, 0.032960765063762665, 0.01275536511093378, -0.0005785065586678684, 0.036798663437366486, 0.02641376480460167, -0.011005735024809837, 0.0042894147336483, 0.010667096823453903, 0.0035698090214282274, 0.022462988272309303, 0.0015027060871943831, ...
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelAttention", "ModelEncoderLayer", "None", "__init__", "activation_dropout", "activation_fn", "activation_function", "and", "any", "attention_dropout", "attention_mask", "attn_weights", "clamp", "clamp_va...
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallDecoderLayer
[ -0.00013334836694411933, 0.051318809390068054, 0.022494366392493248, -0.0071213324554264545, -0.0007629998726770282, 0.0316503643989563, 0.041823696345090866, -0.03504147753119469, 0.007347406353801489, -0.005990962032228708, -0.00477581424638629, 0.02170310728251934, -0.0005157314008101821,...
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelAttention", "ModelDecoderLayer", "None", "True", "__init__", "activation_dropout", "activation_fn", "activation_function", "attention_dropout", "attention_mask", "cache_position", "class", "config", "cr...
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallPreTrainedModel
[ -0.00028302345890551805, 0.03372073173522949, -0.001559299067594111, 0.01925271563231945, -0.0011534540681168437, 0.02175898663699627, 0.03918895870447159, -0.02654368430376053, -0.0022784278262406588, 0.0019224234856665134, 0.006892244331538677, -0.020391929894685745, -0.002634432166814804,...
[ "ModelConfig", "ModelForConditionalGeneration", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_init_weights", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "attention_mask", "base_model_prefix", "class", "config", "decoder_input_ids", ...
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallEncoder
[ -0.00020457802747841924, 0.04398339241743088, 0.014850062318146229, 0.011279245838522911, -0.0010698279365897179, 0.04194292426109314, 0.01099584810435772, -0.009635536931455135, 0.0027347919531166553, -0.012356159277260303, 0.02199169620871544, 0.0133197121322155, -0.00028871186077594757, ...
[ "BaseModelOutput", "Embedding", "False", "LayerNorm", "ModelEncoder", "ModelEncoderLayer", "ModelLearnedPositionalEmbedding", "ModelPreTrainedModel", "ModuleList", "None", "True", "_", "__init__", "all_attentions", "and", "attention_mask", "attentions", "class", "config", "crea...
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallDecoder
[ -0.00018115821876563132, 0.0700194239616394, 0.003353202948346734, -0.01750485599040985, -0.001250346889719367, 0.026143616065382957, 0.048649862408638, -0.030462997034192085, 0.0058823139406740665, -0.014435823075473309, 0.0013355978298932314, 0.011196288280189037, -0.00268540414981544, 0...
[ "BaseModelOutputWithPastAndCrossAttentions", "DynamicCache", "Embedding", "EncoderDecoderCache", "False", "LayerNorm", "ModelDecoder", "ModelDecoderLayer", "ModelLearnedPositionalEmbedding", "ModelPreTrainedModel", "ModuleList", "None", "Setting", "True", "ValueError", "You", "__init...
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallModel
[ 0.000020982022761018015, 0.05237112566828728, 0.012197548523545265, -0.018576083704829216, 0.00009004784806165844, 0.03715216740965843, 0.021149877458810806, -0.03782359138131142, 0.011470171622931957, -0.009008281864225864, 0.030214112251996994, -0.0010421071201562881, -0.001692549791187048...
[ "BaseModelOutput", "Embedding", "ModelDecoder", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "None", "Seq2SeqModelOutput", "__init__", "_tied_weights_keys", "and", "attention_mask", "attentions", "auto_docstring", "cache_position", "class", "config", "cross_attentions", ...
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallForConditionalGeneration
[ -0.0003165907401125878, 0.037318356335163116, 0.006897027604281902, -0.0013665271690115333, -0.0014595369575545192, 0.039378877729177475, 0.014194712042808533, -0.023123644292354584, -0.005208543501794338, -0.0046647945418953896, 0.02861837111413479, 0.004893741570413113, 0.00125920830760151...
[ "CrossEntropyLoss", "False", "GenerationMixin", "Linear", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "Seq2SeqLMOutput", "True", "__init__", "_keys_to_ignore_on_load_missing", "_resize_final_logits_bias", "_tied_weights_keys", "and", "attention_mask",...
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallDecoderWrapper
[ -0.00007960084622027352, 0.058670200407505035, -0.005794242024421692, -0.009013265371322632, -0.0005808238056488335, 0.02396073006093502, 0.03000689670443535, -0.02955903299152851, -0.0018194479634985328, -0.014107719995081425, -0.008117537014186382, 0.00895728264003992, -0.00405876850709319...
[ "ModelDecoder", "ModelDecoderWrapper", "ModelPreTrainedModel", "__init__", "args", "class", "config", "decoder", "def", "forward", "kwargs", "post_init", "return", "self", "super" ]
blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallForCausalLM
[ -0.00017654977273195982, 0.051267948001623154, 0.01765146479010582, -0.010287318378686905, -0.0008045751019380987, 0.01787632331252098, 0.027882568538188934, -0.02383509837090969, 0.002881011227145791, 0.010343533009290695, 0.019337909296154976, -0.0021923792082816362, -0.0000619241764070466...
[ "CausalLMOutputWithCrossAttentions", "CrossEntropyLoss", "False", "GenerationMixin", "Linear", "ModelDecoderWrapper", "ModelForCausalLM", "ModelPreTrainedModel", "None", "True", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_position", "cla...
dots1/modeling_dots1.py:Dots1RMSNorm
[ -0.00009835156379267573, 0.04154934734106064, 0.0313878208398819, 0.048549506813287735, -0.00046573654981330037, 0.03861379250884056, 0.02461347170174122, -0.03003295138478279, 0.008129219524562359, 0.04358164966106415, 0.01964561454951763, 0.004770271480083466, 0.0023992490023374557, 0.01...
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "dtype", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "input_dtype", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", ...
dots1/modeling_dots1.py:Dots1RotaryEmbedding
[ -0.00029968636226840317, 0.04852752387523651, 0.0020364229567348957, -0.005228263325989246, -0.0013792794197797775, 0.04136393964290619, 0.040670689195394516, 0.0062681385315954685, -0.0026430170983076096, 0.021144136786460876, 0.0021808501332998276, -0.004043960478156805, -0.001364836702123...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
dots1/modeling_dots1.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
dots1/modeling_dots1.py:apply_rotary_pos_emb
[ -0.0001444592053303495, 0.027112245559692383, 0.028019767254590988, 0.0028360087890177965, -0.0005707467789761722, 0.021667107939720154, 0.046510547399520874, -0.002183726755902171, 0.01293220091611147, 0.03652779385447502, 0.007884104736149311, 0.0017654155381023884, -0.0007834474672563374,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
dots1/modeling_dots1.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
dots1/modeling_dots1.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
dots1/modeling_dots1.py:Dots1Attention
[ -0.0000575865815335419, 0.03601139411330223, 0.024420227855443954, 0.003826210740953684, -0.00039211625698953867, 0.0337606817483902, 0.03466096892952919, -0.009790598414838314, 0.0034745370503515005, 0.018005697056651115, 0.021381765604019165, 0.031509969383478165, 0.00016264912846963853, ...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelAttention", "ModelRMSNorm", "Module", "None", "Tensor", "True", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs", "cache...
dots1/modeling_dots1.py:Dots1MLP
[ -0.00024263472005259246, 0.02898676134645939, 0.02358049899339676, 0.02829659916460514, -0.0008483228739351034, 0.05774346739053726, 0.029906975105404854, -0.005061180330812931, -0.001998591236770153, -0.004831126891076565, 0.028066545724868774, -0.04946153610944748, -0.0024299416691064835, ...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "None", "__init__", "act_fn", "class", "config", "def", "down_proj", "else", "forward", "gate_proj", "hidden_act", "hidden_size", "if", "intermediate_size", "is", "nn", "return", "self", "super", "up_proj", "x" ]
dots1/modeling_dots1.py:Dots1TopkRouter
[ -0.00040611581061966717, 0.03441784903407097, 0.0062308176420629025, -0.007833028212189674, -0.0017876513302326202, 0.06978515535593033, 0.07168407738208771, -0.02883978560566902, -0.006349499803036451, 0.013470434583723545, 0.019463887438178062, -0.0206507109105587, 0.00022252919734455645, ...
[ "F", "ModelTopkRouter", "Module", "Parameter", "__init__", "class", "config", "def", "e_score_correction_bias", "empty", "float32", "forward", "hidden_size", "hidden_states", "linear", "n_routed_experts", "nn", "register_buffer", "return", "router_logits", "self", "super", ...
dots1/modeling_dots1.py:Dots1NaiveMoe
[ -0.00035851384745910764, 0.041463855654001236, 0.00232943007722497, -0.0018635441083461046, -0.0012811865890398622, 0.054275721311569214, 0.05334394797682762, -0.01584012433886528, -0.005095628555864096, -0.012753630056977272, 0.014093051664531231, -0.030515534803271294, -0.00012921057350467...
[ "ACT2FN", "ModelNaiveMoe", "Module", "None", "Parameter", "__init__", "act_fn", "chunk", "class", "config", "continue", "current_hidden_states", "current_state", "def", "dim", "down_proj", "dtype", "empty", "expert_hit", "expert_idx", "expert_mask", "final_hidden_states", ...
dots1/modeling_dots1.py:Dots1MoE
[ -0.0003974373103119433, 0.037682946771383286, -0.01301239151507616, -0.01348342839628458, -0.0015455895336344838, 0.06924241036176682, 0.05252060294151306, -0.02602478303015232, -0.008419782854616642, -0.020843379199504852, 0.03108843043446541, -0.02908652275800705, -0.0008316743769682944, ...
[ "False", "ModelMLP", "ModelMoE", "ModelNaiveMoe", "ModelTopkRouter", "Module", "True", "__init__", "bool", "class", "config", "def", "denominator", "dim", "e_score_correction_bias", "expand", "experts", "forward", "gate", "gather", "group_idx", "group_mask", "group_scores...
dots1/modeling_dots1.py:Dots1DecoderLayer
[ -0.00024360170937143266, 0.046885326504707336, 0.009616044349968433, -0.00440972438082099, -0.0007930391002446413, 0.036870986223220825, 0.04142295941710472, -0.038691774010658264, 0.0039545269683003426, -0.00557616725564003, -0.00440972438082099, 0.01627330482006073, -0.0012731300666928291,...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelDecoderLayer", "ModelMLP", "ModelMoE", "ModelRMSNorm", "None", "Tensor", "_", "__init__", "attention_mask", "attention_type", "cache_position", "class", "config", "def", "else", "eps", "first_k_dense_replace", "fo...
dots1/modeling_dots1.py:Dots1PreTrainedModel
[ -0.00037101644556969404, 0.05074922740459442, 0.0014767908724024892, -0.005674369167536497, -0.0014840657822787762, 0.041670236736536026, 0.03398801386356354, -0.018972761929035187, -0.005121481604874134, -0.0031136281322687864, 0.008613401092588902, -0.009660976938903332, -0.002226098673418...
[ "ModelAttention", "ModelConfig", "ModelDecoderLayer", "ModelNaiveMoe", "ModelPreTrainedModel", "ModelTopkRouter", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_init_weights", "_keep_in_fp32_modules_strict", "_no_split_modules", "_skip_keys_device_placement",...
dots1/modeling_dots1.py:Dots1Model
[ -0.00008759692718740553, 0.04394812509417534, 0.0015212813159450889, -0.006958452984690666, -0.0006585175869986415, 0.04146900027990341, 0.040116749703884125, -0.020734500139951706, 0.014085937291383743, -0.0015564961358904839, 0.021523313596844673, 0.005887921899557114, -0.00134520709980279...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelDecoderLayer", "ModelModel", "ModelPreTrainedModel", "ModelRMSNorm", "ModelRotaryEmbedding", "ModuleList", "None", "ValueError", "You", "__init__", "and", "arange", "attention_mask", "attention_type", "auto_do...
dots1/modeling_dots1.py:Dots1ForCausalLM
[ -0.00028262686100788414, 0.03526614233851433, 0.008361488580703735, -0.001301150070503354, -0.0012087186332792044, 0.02741658128798008, 0.0389065183699131, -0.007394513580948114, 0.003640376031398773, 0.026734011247754097, 0.025823917239904404, 0.0026734010316431522, 0.0009385344455949962, ...
[ "CausalLMOutputWithPast", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_pp_plan", "_tied_weights_keys", "_tp_plan", "attention_mask", "attentions", "auto_docstring", "cache_position", "can_return_tuple", "class", "colwi...
depth_anything/modeling_depth_anything.py:DepthAnythingReassembleLayer
[ -0.0003730056341737509, 0.010099845007061958, 0.03167678788304329, 0.013772516511380672, -0.001162055996246636, 0.03420174866914749, 0.00350051443092525, -0.06886257976293564, 0.0009109945385716856, 0.009583375416696072, -0.022609880194067955, -0.029610909521579742, -0.0016283130971714854, ...
[ "Conv2d", "ConvTranspose2d", "Identity", "ModelReassembleLayer", "Module", "__init__", "channels", "class", "config", "def", "elif", "factor", "forward", "hidden_state", "if", "in_channels", "int", "kernel_size", "nn", "out_channels", "padding", "projection", "reassemble_...
depth_anything/modeling_depth_anything.py:DepthAnythingReassembleStage
[ -0.0002945549495052546, -0.005138741806149483, 0.0007523572421632707, 0.025211066007614136, -0.0009085068595595658, 0.0235076155513525, 0.0001357436995021999, -0.06632100045681, 0.0021293128374964, 0.003917935770004988, -0.027028078213334084, -0.02055496722459793, -0.00299523351714015, 0.0...
[ "ModelReassembleLayer", "ModelReassembleStage", "Module", "ModuleList", "None", "_", "__init__", "append", "batch_size", "channels", "class", "config", "contiguous", "def", "enumerate", "factor", "for", "forward", "hidden_state", "hidden_states", "i", "in", "layers", "n...
depth_anything/modeling_depth_anything.py:DepthAnythingPreActResidualLayer
[ 0, 0.017853641882538795, 0.02892741933465004, -0.011808263137936592, 0.00046611562720499933, 0.04384312033653259, -0.008022839203476906, -0.021469568833708763, 0.013672725297510624, -0.014802702702581882, 0.02158256620168686, -0.03322133421897888, 0.0038984217680990696, -0.0126557452604174...
[ "Conv2d", "ModelPreActResidualLayer", "Module", "ReLU", "__init__", "activation1", "activation2", "class", "config", "convolution1", "convolution2", "def", "forward", "fusion_hidden_size", "hidden_state", "kernel_size", "nn", "padding", "residual", "return", "self", "stride...
depth_anything/modeling_depth_anything.py:DepthAnythingFeatureFusionLayer
[ -0.0001829315588111058, 0.021951787173748016, 0.03174566105008125, -0.0003781758132390678, -0.00039048853795975447, 0.03422227501869202, 0.015197391621768475, -0.05290943756699562, 0.004812507424503565, -0.0026454718317836523, 0.0057975235395133495, -0.012664493173360825, 0.00205446220934391...
[ "Conv2d", "False", "ModelFeatureFusionLayer", "ModelPreActResidualLayer", "Module", "None", "True", "__init__", "align_corners", "bilinear", "class", "config", "def", "else", "forward", "functional", "fusion_hidden_size", "hidden_state", "if", "interpolate", "is", "kernel_s...
depth_anything/modeling_depth_anything.py:DepthAnythingFeatureFusionStage
[ -0.0001391211844747886, 0.006159830838441849, 0.009239746257662773, 0.026095284149050713, -0.00019774457905441523, 0.011871674098074436, 0.014671597629785538, -0.06406224519014359, 0.004871866200119257, 0.011983671225607395, 0.012879646383225918, -0.007335798814892769, 0.0026039285585284233,...
[ "ModelFeatureFusionLayer", "ModelFeatureFusionStage", "Module", "ModuleList", "None", "_", "__init__", "append", "class", "config", "def", "else", "enumerate", "for", "forward", "fused_hidden_state", "fused_hidden_states", "hidden_state", "hidden_states", "idx", "if", "in",...
depth_anything/modeling_depth_anything.py:DepthAnythingPreTrainedModel
[ -0.0002876944199670106, 0.03110673651099205, 0.012236841022968292, 0.03225036710500717, -0.001365208299830556, 0.02218642085790634, 0.026074763387441635, -0.0007219164399430156, -0.005117744207382202, 0.02104279212653637, 0.016125183552503586, -0.01160784438252449, -0.004174249246716499, 0...
[ "Model", "ModelConfig", "ModelPreTrainedModel", "PreTrainedModel", "True", "base_model_prefix", "class", "config", "image", "input_modalities", "main_input_name", "pixel_values", "supports_gradient_checkpointing" ]
depth_anything/modeling_depth_anything.py:DepthAnythingNeck
[ -0.00022688345052301884, -0.003222096711397171, 0.007007005158811808, 0.0016813997644931078, -0.00040803843876346946, 0.016321538016200066, 0.014520540833473206, -0.0706891417503357, 0.003137674881145358, 0.008667299523949623, 0.0006507509970106184, -0.009230110794305801, 0.00289848004467785...
[ "ModelFeatureFusionStage", "ModelNeck", "ModelReassembleStage", "Module", "ModuleList", "None", "The", "TypeError", "ValueError", "__init__", "a", "be", "channel", "class", "config", "convs", "def", "enumerate", "equal", "feature", "features", "for", "forward", "fusion_...
depth_anything/modeling_depth_anything.py:DepthAnythingDepthEstimationHead
[ -0.00019547012925613672, 0.0292090754956007, 0.02468053624033928, 0.004670055583119392, -0.0007323496392928064, 0.050493206828832626, 0.03486974909901619, -0.03690759092569351, 0.00517951650545001, 0.022642694413661957, 0.007472089026123285, -0.010698673315346241, 0.005462550092488527, 0.0...
[ "Conv2d", "Model", "ModelModelEstimationHead", "Model_estimation_type", "Module", "ReLU", "Sigmoid", "True", "Unknown", "ValueError", "__init__", "activation1", "activation2", "align_corners", "bilinear", "class", "config", "conv1", "conv2", "conv3", "def", "dim", "elif",...
depth_anything/modeling_depth_anything.py:DepthAnythingForDepthEstimation
[ -0.00013594729534815997, 0.027280675247311592, 0.004462579730898142, 0.013303538784384727, -0.0004841477784793824, 0.061521850526332855, 0.04086487740278244, -0.017850318923592567, 0.00592203950509429, 0.01257380936294794, -0.0026803542859852314, 0.015492728911340237, 0.0024558219593018293, ...
[ "DPTViTEmbeddings", "ModelEstimatorOutput", "ModelForModelEstimation", "ModelModelEstimationHead", "ModelNeck", "ModelPreTrainedModel", "None", "_", "__init__", "_no_split_modules", "attentions", "auto_docstring", "backbone", "class", "config", "def", "else", "feature_maps", "for...
swiftformer/modeling_swiftformer.py:SwiftFormerPatchEmbedding
[ -0.000032579442631686106, 0.0015990342944860458, 0.019723771139979362, 0.008227629587054253, 0.00027824603603221476, 0.018258575350046158, 0.0037475163117051125, 0.005691716447472572, 0.011158018372952938, -0.013074042275547981, 0.007325971964746714, -0.004733704961836338, -0.001267956686206...
[ "BatchNorm2d", "Conv2d", "ModelPatchEmbedding", "Module", "ReLU", "Sequential", "__init__", "batch_norm_eps", "class", "config", "def", "embed_dims", "eps", "forward", "in_chs", "kernel_size", "nn", "num_channels", "out_chs", "padding", "patch_embedding", "return", "self"...
swiftformer/modeling_swiftformer.py:drop_path
[ 0, 0.017753854393959045, 0.03775503486394882, -0.02123720571398735, 0, 0.043598074465990067, 0.0339345820248127, -0.027304979041218758, 0.014832334592938423, 0.0017486985307186842, 0.020113544538617134, -0.04944111406803131, 0.0016574009787291288, -0.018315685912966728, -0.05034004524350...
[ "False", "Model_path", "Model_prob", "def", "device", "div", "dtype", "floor_", "if", "input", "keep_prob", "ndim", "not", "or", "output", "rand", "random_tensor", "return", "shape", "torch", "training" ]
swiftformer/modeling_swiftformer.py:SwiftFormerDropPath
[ -0.00005618787690764293, 0.026857804507017136, 0.04180378094315529, -0.021239018067717552, -0.0005759257473982871, 0.03303847089409828, 0.02753205969929695, -0.017642993479967117, 0.006096384488046169, -0.00803486630320549, 0.01017000526189804, -0.03438698127865791, 0.0002458219532854855, ...
[ "ModelDropPath", "Module", "__init__", "class", "config", "def", "drop_path", "drop_path_rate", "drop_prob", "extra_repr", "f", "forward", "hidden_states", "nn", "p", "return", "self", "super", "training" ]
swiftformer/modeling_swiftformer.py:SwiftFormerEmbeddings
[ 0.000022186150090419687, 0.002627899870276451, 0.02102319896221161, 0.020571086555719376, 0.0005686718504875898, 0.030291490256786346, 0.011133253574371338, -0.0007240853155963123, 0.014071979559957981, -0.008251040242612362, 0.01808447204530239, 0.011415823362767696, -0.00048743304796516895...
[ "BatchNorm2d", "Conv2d", "Iterable", "ModelEmbeddings", "Module", "__init__", "abc", "batch_norm_eps", "class", "collections", "config", "def", "down_pad", "down_patch_size", "down_stride", "else", "embed_dim", "embed_dims", "eps", "forward", "if", "in_chans", "index", ...
swiftformer/modeling_swiftformer.py:SwiftFormerConvEncoder
[ 0, 0.014406394213438034, 0.024826830253005028, -0.008882993832230568, 0, 0.04282058775424957, 0.009224647656083107, -0.03644305095076561, 0.015260527841746807, 0.0035446563269943, 0.02459906041622162, -0.009053820744156837, 0.0007473672740161419, -0.0015943835023790598, -0.02551013603806...
[ "BatchNorm2d", "Conv2d", "Dropout", "GELU", "ModelConvEncoder", "Module", "Parameter", "True", "__init__", "act", "batch_norm_eps", "class", "config", "def", "depth_wise_conv", "dim", "drop_conv_encoder_rate", "drop_path", "eps", "forward", "groups", "hidden_dim", "input"...
swiftformer/modeling_swiftformer.py:SwiftFormerMlp
[ -0.00011673776316456497, 0.02320605143904686, 0.027507662773132324, 0.01313122920691967, -0.00009109082748182118, 0.0536569207906723, 0.0036507081240415573, -0.018791241571307182, 0.007046715822070837, -0.006565614603459835, 0.03848808631300926, -0.030111268162727356, 0.0026743561029434204, ...
[ "ACT2CLS", "BatchNorm2d", "Conv2d", "Dropout", "ModelMlp", "Module", "__init__", "act", "act_layer", "batch_norm_eps", "class", "config", "def", "drop", "drop_mlp_rate", "eps", "fc1", "fc2", "forward", "hidden_act", "hidden_features", "in_features", "int", "mlp_ratio", ...
swiftformer/modeling_swiftformer.py:SwiftFormerEfficientAdditiveAttention
[ -0.00002216685788880568, 0.039104972034692764, 0.05146573856472969, 0.007416460197418928, -0.00009481270535616204, 0.022249380126595497, 0.038880232721567154, -0.0029075895436108112, 0.008708721958100796, 0.006910792551934719, 0.043824538588523865, 0.03034006431698799, 0.002064809901639819, ...
[ "Linear", "ModelEfficientAdditiveAttention", "Module", "Parameter", "__init__", "class", "config", "def", "dim", "final", "forward", "functional", "global_queries", "key", "nn", "normalize", "out", "proj", "query", "query_weight", "randn", "repeat", "return", "scale_fac...
swiftformer/modeling_swiftformer.py:SwiftFormerLocalRepresentation
[ -0.00016435969155281782, 0.001807956607080996, 0.01532118208706379, -0.006431466434150934, -0.0004019666521344334, 0.06265677511692047, 0.014577990397810936, -0.03704524412751198, 0.012062571942806244, 0.01737925037741661, 0.012405583634972572, -0.007717759348452091, 0.0023296200670301914, ...
[ "BatchNorm2d", "Conv2d", "GELU", "Identity", "ModelLocalRepresentation", "Module", "Parameter", "True", "__init__", "act", "batch_norm_eps", "class", "config", "def", "depth_wise_conv", "dim", "drop_path", "eps", "forward", "groups", "input", "kernel_size", "layer_scale",...
swiftformer/modeling_swiftformer.py:SwiftFormerEncoderBlock
[ -0.0002884801651816815, 0.008539012633264065, 0.02480929344892502, 0.006346563342958689, -0.0011250725947320461, 0.05492662265896797, 0.003981025889515877, -0.04315663129091263, 0.0019328170455992222, 0.016962632536888123, 0.007500484120100737, -0.0016731849173083901, 0.0013702807482331991, ...
[ "Identity", "ModelDropPath", "ModelEfficientAdditiveAttention", "ModelEncoderBlock", "ModelLocalRepresentation", "ModelMlp", "Module", "Parameter", "True", "__init__", "attn", "batch_size", "channels", "class", "config", "def", "dim", "drop_path", "else", "forward", "height",...
swiftformer/modeling_swiftformer.py:SwiftFormerStage
[ -0.00006884575122967362, -0.028470367193222046, -0.0024431415367871523, 0.015139004215598106, 0.00029480105149559677, 0.048580385744571686, 0.011975630186498165, -0.021013842895627022, 0.0018358867382630706, -0.00909470021724701, 0.009377144277095795, -0.0327635183930397, -0.0009956154972314...
[ "GradientCheckpointingLayer", "ModelEncoderBlock", "ModelStage", "ModuleList", "__init__", "append", "block", "block_dpr", "block_idx", "blocks", "class", "config", "def", "depth", "depths", "dim", "drop_path", "drop_path_rate", "else", "embed_dims", "for", "forward", "if...
swiftformer/modeling_swiftformer.py:SwiftFormerEncoder
[ -0.00019828318909276277, 0.018323855474591255, 0.012462498620152473, 0.03619246184825897, -0.0005939604016020894, 0.03801346570253372, 0.009218834340572357, -0.03118469938635826, 0.0032721171155571938, -0.002717279829084873, 0.02651837468147278, -0.0036704617086797953, -0.001671625068411231,...
[ "BaseModelOutputWithNoAttention", "False", "ModelEmbeddings", "ModelEncoder", "ModelStage", "Module", "ModuleList", "None", "__init__", "all_hidden_states", "append", "block", "break", "class", "config", "def", "depths", "downsamples", "else", "embed_dims", "for", "forward"...
swiftformer/modeling_swiftformer.py:SwiftFormerPreTrainedModel
[ -0.00012081753084203228, 0.041911426931619644, 0.0008965010638348758, 0.02308490313589573, -0.0004114799958188087, 0.027343282476067543, 0.0005533092771656811, -0.005519084632396698, -0.0019330804934725165, 0.02073158696293831, 0.005379006266593933, -0.005266943946480751, -0.0013587594730779...
[ "BatchNorm2d", "Conv2d", "LayerNorm", "Linear", "Model", "ModelConfig", "ModelConvEncoder", "ModelEfficientAdditiveAttention", "ModelEncoderBlock", "ModelLocalRepresentation", "ModelPreTrainedModel", "None", "PreTrainedModel", "True", "_init_weights", "_no_split_modules", "base_model...
swiftformer/modeling_swiftformer.py:SwiftFormerModel
[ 0.00006303300324361771, 0.027118198573589325, 0.008684547618031502, 0.031152311712503433, 0.0005357805057428777, 0.0225237924605608, 0.02723025716841221, -0.017144976183772087, 0.011990278027951717, 0.008012195117771626, 0.0156882144510746, 0.01288674771785736, -0.0004902567015960813, 0.00...
[ "BaseModelOutputWithNoAttention", "ModelEncoder", "ModelModel", "ModelPatchEmbedding", "ModelPreTrainedModel", "None", "__init__", "auto_docstring", "class", "config", "def", "else", "embedding_output", "encoder", "encoder_outputs", "forward", "hidden_states", "if", "is", "kwar...
swiftformer/modeling_swiftformer.py:SwiftFormerForImageClassification
[ -0.00016941585636232048, 0.035545893013477325, 0.01279875636100769, 0.02179700881242752, -0.000384242128347978, 0.009389480575919151, 0.0143077801913023, 0.005896370392292738, 0.002487094607204199, 0.005756645929068327, 0.04404113441705704, 0.013637103140354156, 0, -0.0019281969871371984, ...
[ "BatchNorm2d", "Identity", "ImageClassifierOutputWithNoAttention", "Linear", "Model", "ModelForImageClassification", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "auto_docstring", "batch_norm_eps", "class", "cls_out", "config", "def", "dist_head", "distillation_out", ...
moshi/modeling_moshi.py:MoshiConditionalGenerationGenerateOutput
[ -0.00010677255340851843, 0.02440917305648327, -0.0006221808143891394, 0.02002227120101452, -0.0004147872095927596, 0.0204722099006176, 0.05151797831058502, -0.030595829710364342, 0.008211380802094936, -0.01462300680577755, 0.0321706160902977, 0.029021043330430984, -0.0074802301824092865, -...
[ "ModelConditionalGenerationGenerateOutput", "ModelOutput", "None", "attentions", "audio_codes", "audio_sequences", "beam_indices", "class", "hidden_states", "logits", "past_key_values", "r", "scores", "sequences", "sequences_scores" ]
moshi/modeling_moshi.py:MoshiCausalLMOutputWithPast
[ -0.00019186620193067938, 0.020465727895498276, 0.028310924768447876, -0.00026470431475900114, -0.0012720018858090043, 0.031608179211616516, 0.04252323508262634, -0.019556140527129173, 0.01864655315876007, -0.01193834189325571, 0.012904779054224491, 0.025809558108448982, -0.001570460386574268...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "logits", "loss", "past_key_values", "r" ]
moshi/modeling_moshi.py:MoshiConditionalGenerationOutputWithPast
[ -0.00028117818874306977, 0.020860914140939713, 0.01868312619626522, 0.00922694243490696, -0.001296643284149468, 0.03461536020040512, 0.06281197816133499, -0.05020373687148094, 0.009800044819712639, -0.01369713805615902, 0.027279656380414963, 0.030947508290410042, -0.004069024231284857, -0....
[ "ModelConditionalGenerationOutputWithPast", "ModelOutput", "None", "attentions", "audio_logits", "class", "depth_attentions", "depth_hidden_states", "depth_loss", "depth_past_key_values", "hidden_states", "last_hidden_state", "logits", "loss", "past_key_values", "r" ]
moshi/modeling_moshi.py:MoshiUnconditionalInput
[ -0.0000717238217475824, 0.04442980885505676, 0.016887860372662544, 0.023801684379577637, -0.0007012996356934309, 0.032868992537260056, 0.05780408903956413, -0.011390806175768375, 0.01586778834462166, -0.02731526456773281, 0.038536060601472855, -0.0018276292830705643, -0.005638732109218836, ...
[ "ModelOutput", "ModelUnconditionalInput", "Model_audio_codes", "None", "attention_mask", "class", "input_ids", "r", "user_audio_codes" ]
moshi/modeling_moshi.py:MoshiRMSNorm
[ -0.00004397882003104314, 0.04096085578203201, 0.022177699953317642, 0.05272861570119858, -0.00037658243672922254, 0.0538601316511631, 0.018443699926137924, -0.006534500978887081, 0.010466516949236393, 0.04322388768196106, 0.014709699898958206, 0.013125577941536903, 0.003168242983520031, 0....
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "_norm", "class", "def", "dim", "eps", "extra_repr", "f", "float", "forward", "keepdim", "mean", "nn", "ones", "output", "pow", "return", "rsqrt", "self", "shape", "super", "torch", "tuple", "type_as", ...
moshi/modeling_moshi.py:MoshiFlexibleLinear
[ -0.00007256858953041956, 0.012272448278963566, 0.03242628276348114, 0.014862047508358955, -0.00010335526167182252, 0.0452616885304451, 0.03242628276348114, -0.023756757378578186, -0.004616241902112961, -0.002054790500551462, -0.002420712262392044, -0.03220110386610031, -0.0012947996146976948...
[ "ModelFlexibleLinear", "Module", "None", "Parameter", "__init__", "class", "def", "else", "forward", "if", "index_select", "input_size", "is", "layer_idx", "matmul", "nn", "not", "num_layers", "output_size", "randn", "return", "selected_weights", "self", "squeeze", "s...
moshi/modeling_moshi.py:MoshiLinear
[ -0.00011919748067157343, -0.007063554599881172, 0.022264324128627777, 0.03480919823050499, -0.0005121076828800142, 0.04701501876115799, 0.037295568734407425, -0.03684350103139877, 0.0026417693588882685, 0.0059616402722895145, 0.013053448870778084, -0.03684350103139877, 0.00020307718659751117...
[ "False", "Linear", "ModelFlexibleLinear", "ModelLinear", "Module", "None", "__init__", "class", "def", "else", "forward", "if", "input_dim", "layer_idx", "nn", "not", "num_codebooks", "num_layers", "output_dim", "self", "super", "use_flexible_linear", "x" ]
moshi/modeling_moshi.py:MoshiRotaryEmbedding
[ -0.00029968636226840317, 0.04852752387523651, 0.0020364229567348957, -0.005228263325989246, -0.0013792794197797775, 0.04136393964290619, 0.040670689195394516, 0.0062681385315954685, -0.0026430170983076096, 0.021144136786460876, 0.0021808501332998276, -0.004043960478156805, -0.001364836702123...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
moshi/modeling_moshi.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
moshi/modeling_moshi.py:apply_rotary_pos_emb
[ -0.0001444592053303495, 0.027112245559692383, 0.028019767254590988, 0.0028360087890177965, -0.0005707467789761722, 0.021667107939720154, 0.046510547399520874, -0.002183726755902171, 0.01293220091611147, 0.03652779385447502, 0.007884104736149311, 0.0017654155381023884, -0.0007834474672563374,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
moshi/modeling_moshi.py:MoshiGatingMLP
[ -0.00024472110089845955, 0.002743734512478113, 0.016233762726187706, 0.01806291937828064, -0.0005215953569859266, 0.04892992973327637, 0.03292481228709221, -0.05167366564273834, 0.002457928843796253, -0.01806291937828064, 0.05281689018011093, -0.036811769008636475, 0.00225786492228508, 0.0...
[ "ACT2FN", "False", "Linear", "ModelFlexibleLinear", "ModelGatingMLP", "Module", "None", "_", "__init__", "activation_fn", "batch_size", "class", "config", "def", "else", "fc1", "fc2", "ffn_dim", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "is", "lay...
moshi/modeling_moshi.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
moshi/modeling_moshi.py:MoshiAttention
[ -0.00010913627920672297, 0.023207301273941994, 0.024108557030558586, -0.010477082803845406, -0.0004136617062613368, 0.030192023143172264, 0.03897925466299057, -0.02061619609594345, -0.0007252281648106873, 0.019602283835411072, 0.013462488539516926, 0.020841509103775024, -0.000255238061072304...
[ "False", "Instantiating", "ModelAttention", "ModelLinear", "ModelRotaryEmbedding", "Module", "None", "Please", "True", "ValueError", "_", "__class__", "__init__", "__name__", "a", "and", "apply_rotary_pos_emb", "attention_dropout", "attention_mask", "attn_output", "attn_weigh...
moshi/modeling_moshi.py:MoshiFlashAttention2
[ 0, 0.024354619905352592, 0.04848477244377136, -0.01279459334909916, 0, 0.02121208794414997, 0.03232318162918091, -0.05050497129559517, 0.010830511339008808, 0.021324321627616882, 0.013692459091544151, 0.0205386895686388, 0.0026094235945492983, -0.004012339748442173, 0.010549928061664104,...
[ "False", "ModelAttention", "ModelFlashAttention2", "None", "StaticCache", "The", "ValueError", "We", "_", "__init__", "_flash_attention_forward", "_flash_attn_uses_top_left_mask", "_is_quantized", "an", "and", "apply_rotary_pos_emb", "args", "at", "attention_dropout", "attentio...
moshi/modeling_moshi.py:MoshiSdpaAttention
[ 0.000032642928999848664, 0.060571253299713135, 0.03163165599107742, -0.004234379157423973, 0.00009332810441264883, 0.02108776941895485, 0.04374590516090393, -0.012170335277915001, 0.009814786724746227, 0.010039124637842178, 0.010768222622573376, 0.03881046921014786, 0.0004416653828229755, ...
[ "False", "If", "ModelAttention", "ModelSdpaAttention", "None", "The", "True", "_", "__class__", "__name__", "and", "apply_rotary_pos_emb", "attention", "attention_dropout", "attention_mask", "attn_implementation", "attn_mask", "attn_output", "be", "bsz", "cache_kwargs", "ca...
moshi/modeling_moshi.py:MoshiDecoderLayer
[ -0.00024899491108953953, 0.02648302912712097, 0.011349868960678577, 0.003625652752816677, -0.0011177901178598404, 0.044711604714393616, 0.04769238084554672, -0.04196012392640114, 0.003339039860293269, -0.006735402159392834, 0.0012897578999400139, 0.013356159441173077, -0.0010174756171181798,...
[ "False", "GradientCheckpointingLayer", "ModelDecoderLayer", "ModelGatingMLP", "ModelRMSNorm", "Model_ATTENTION_CLASSES", "None", "True", "__init__", "_attn_implementation", "attention_mask", "cache_position", "class", "config", "def", "else", "eps", "forward", "hidden_size", "h...
moshi/modeling_moshi.py:MoshiPreTrainedModel
[ -0.00021302353707142174, 0.04706671088933945, 0.010012989863753319, 0.016744887456297874, -0.0011101973941549659, 0.02636188454926014, 0.028624705970287323, -0.03552631661295891, -0.0018951139645650983, 0.011370684020221233, 0.007354173809289932, -0.0007000607438385487, -0.003125523682683706...
[ "MimiTransformerLayer", "ModelConfig", "ModelDecoderLayer", "ModelFlexibleLinear", "ModelPreTrainedModel", "PreTrainedModel", "True", "_init_weights", "_no_split_modules", "_supports_flash_attn", "_supports_sdpa", "audio", "base_model_prefix", "class", "config", "def", "if", "init"...
moshi/modeling_moshi.py:MoshiDepthDecoder
[ -0.00012850548955611885, 0.03334805369377136, 0.02208182029426098, -0.022532468661665916, -0.0006970982649363577, 0.026700975373387337, 0.061738964170217514, -0.05903507024049759, 0.004900812171399593, -0.016561364755034447, 0.02501104027032852, 0.02681363746523857, -0.004055844619870186, ...
[ "Attention", "AttentionMaskConverter", "BlockMask", "CausalLMOutputWithPast", "CrossEntropyLoss", "DynamicCache", "Embedding", "False", "Flash", "GenerationMixin", "Make", "Model", "ModelDecoderLayer", "ModelDepthConfig", "ModelDepthDecoder", "ModelFlexibleLinear", "ModelPreTrainedMo...
moshi/modeling_moshi.py:MoshiModel
[ 0, 0.03807422146201134, 0.011478257365524769, -0.00517921382561326, -0.00009711025631986558, 0.034714728593826294, 0.03202713653445244, -0.03113127499818802, 0.01511770486831665, -0.001126828952692449, 0.01903711073100567, 0.016909433528780937, 0.0016797450371086597, 0.023180481046438217, ...
[ "Attention", "AttentionMaskConverter", "BaseModelOutputWithPast", "BlockMask", "DynamicCache", "Embedding", "False", "Flash", "Make", "Model", "ModelDecoderLayer", "ModelModel", "ModelPreTrainedModel", "ModelRMSNorm", "ModuleList", "None", "Setting", "StaticCache", "Tensor", "T...
moshi/modeling_moshi.py:MoshiForCausalLM
[ -0.00024000955454539508, 0.046759508550167084, 0.019313709810376167, -0.00047296000411733985, -0.0010235700756311417, 0.024960992857813835, 0.025525720790028572, -0.011181620880961418, 0.00022942089708521962, 0.01965254731476307, 0.02518688514828682, 0.002950705587863922, -0.0000966214938671...
[ "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "attention_mask", "attentions", "auto_docstring", "cache_position", "class", "config", "contiguous", "def", "device", "else", "float", "forw...
moshi/modeling_moshi.py:MoshiForConditionalGeneration
[ -0.00022083797375671566, 0.041261084377765656, 0.011797022074460983, -0.0042172931134700775, -0.0010899422923102975, 0.029179107397794724, 0.05243121087551117, -0.043312739580869675, 0.007408758159726858, -0.01732509583234787, 0.03465019166469574, -0.003205712651833892, -0.004958168603479862...
[ "Any", "At", "AutoModel", "Check", "Embedding", "False", "GenerationMixin", "Make", "Model", "ModelConditionalGenerationGenerateOutput", "ModelConditionalGenerationOutputWithPast", "ModelConfig", "ModelDepthDecoder", "ModelForCausalLM", "ModelForConditionalGeneration", "ModelPreTrained...
luke/modeling_luke.py:BaseLukeModelOutputWithPooling
[ -0.00019867275841534138, 0.0184109415858984, 0.0393533892929554, 0.02450956590473652, -0.0010787660721689463, 0.06075610592961311, 0.0317588746547699, -0.001186642679385841, 0.01599450595676899, 0.003653421299532056, 0.015764368698000908, 0.012024646624922752, -0.002488353755325079, 0.0237...
[ "ModelModelModelOutputWithPooling", "ModelModelOutputWithPooling", "None", "class", "entity_hidden_states", "entity_last_hidden_state", "r" ]
luke/modeling_luke.py:BaseLukeModelOutput
[ -0.00017150506027974188, 0.03165803104639053, 0.02553439699113369, 0.016868876293301582, -0.0011337387841194868, 0.08134034276008606, 0.043212056159973145, -0.004968231078237295, 0.017908738926053047, 0.007972277700901031, 0.014615842141211033, 0.018024278804659843, -0.0010976324556395411, ...
[ "ModelModelModelOutput", "ModelModelOutput", "None", "class", "entity_hidden_states", "entity_last_hidden_state", "r" ]
luke/modeling_luke.py:LukeMaskedLMOutput
[ -0.00023732711269985884, 0.02603817544877529, 0.032205112278461456, -0.002055645454674959, -0.001106336945667863, 0.07811452448368073, 0.03197670727968216, -0.007765771355479956, 0.008565189316868782, -0.001969993580132723, 0.023068908601999283, 0.03563118726015091, -0.0035402781795710325, ...
[ "ModelMaskedLMOutput", "ModelOutput", "None", "attentions", "class", "entity_hidden_states", "entity_logits", "hidden_states", "logits", "loss", "mep_loss", "mlm_loss", "r" ]
luke/modeling_luke.py:EntityClassificationOutput
[ -0.00015837239334359765, 0.01959594339132309, 0.01959594339132309, 0.008390217088162899, -0.0008129782509058714, 0.027366748079657555, 0.048426758497953415, -0.004166953731328249, 0.014640647917985916, -0.018694980069994926, 0.03311038762331009, 0.027254128828644753, -0.0020553215872496367, ...
[ "ModelClassificationOutput", "ModelOutput", "Model_hidden_states", "None", "attentions", "class", "hidden_states", "logits", "loss", "r" ]
luke/modeling_luke.py:EntityPairClassificationOutput
[ -0.0002146075712516904, 0.019183434545993805, 0.030875112861394882, -0.009364694356918335, -0.0009577528107911348, 0.030421067029237747, 0.05039907991886139, 0, 0.014415953308343887, -0.011918701231479645, 0.02746976912021637, 0.025880608707666397, -0.0020857728086411953, -0.01827534288167...
[ "ModelOutput", "ModelPairClassificationOutput", "Model_hidden_states", "None", "attentions", "class", "hidden_states", "logits", "loss", "r" ]
luke/modeling_luke.py:EntitySpanClassificationOutput
[ -0.00014456479402724653, 0.019632603973150253, 0.017714476212859154, -0.0027220493648201227, -0.0007051941356621683, 0.042650140821933746, 0.04716338589787483, -0.011395937763154507, 0.013426896184682846, -0.016586165875196457, 0.03745991364121437, 0.04738904535770416, -0.0011776741594076157...
[ "ModelOutput", "ModelSpanClassificationOutput", "Model_hidden_states", "None", "attentions", "class", "hidden_states", "logits", "loss", "r" ]
luke/modeling_luke.py:LukeSequenceClassifierOutput
[ -0.00030442987917922437, 0.020744210109114647, 0.01764976978302002, 0.022692561149597168, -0.0013466544914990664, 0.052032437175512314, 0.04080076515674591, -0.014039590023458004, 0.006532707251608372, 0.0071917083114385605, 0.026130829006433487, 0.03025674819946289, -0.004899530205875635, ...
[ "ModelOutput", "ModelSequenceClassifierOutput", "None", "attentions", "class", "entity_hidden_states", "hidden_states", "logits", "loss", "r" ]
luke/modeling_luke.py:LukeTokenClassifierOutput
[ -0.00029148123576305807, 0.029248882085084915, 0.03109133057296276, -0.01658204384148121, -0.0014610046055167913, 0.055503781884908676, 0.05227949842810631, -0.007197067141532898, 0.007657679729163647, 0.007887985557317734, 0.030630718916654587, 0.03823082149028778, -0.004145510960370302, ...
[ "ModelOutput", "ModelTokenClassifierOutput", "None", "attentions", "class", "entity_hidden_states", "hidden_states", "logits", "loss", "r" ]
luke/modeling_luke.py:LukeQuestionAnsweringModelOutput
[ -0.00016167385911103338, 0.032568007707595825, 0.04251934215426445, 0.009781710803508759, -0.0009894793620333076, 0.05654167756438255, 0.05428001284599304, 0.010573294013738632, 0.013400377705693245, 0.01786717027425766, 0.01594475284218788, 0.02883625589311123, -0.0030391153413802385, -0....
[ "ModelOutput", "ModelQuestionAnsweringModelOutput", "None", "attentions", "class", "end_logits", "entity_hidden_states", "hidden_states", "loss", "r", "start_logits" ]
luke/modeling_luke.py:LukeMultipleChoiceModelOutput
[ -0.0003017971757799387, 0.03145919367671013, 0.024635324254631996, 0.0183897502720356, -0.0014746497618034482, 0.05967994034290314, 0.044181663542985916, -0.014630838297307491, 0.0091948751360178, 0.011970685794949532, 0.025444936007261276, 0.014399521052837372, -0.005985342897474766, -0.0...
[ "ModelMultipleChoiceModelOutput", "ModelOutput", "None", "attentions", "class", "entity_hidden_states", "hidden_states", "logits", "loss", "r" ]
luke/modeling_luke.py:LukeEmbeddings
[ -0.00022853288101032376, 0.014285963028669357, 0.0055839973501861095, -0.007256362121552229, -0.0011196340201422572, 0.0333339124917984, 0.02471698261797428, -0.007426432799547911, 0.0038832875434309244, -0.014739485457539558, 0.021995848044753075, 0.01723385974764824, -0.0013180500827729702...
[ "Dropout", "Embedding", "LayerNorm", "ModelEmbeddings", "Module", "None", "__init__", "arange", "class", "config", "create_position_ids_from_input_ids", "create_position_ids_from_inputs_embeds", "def", "device", "dropout", "dtype", "else", "embeddings", "eps", "expand", "forw...
luke/modeling_luke.py:LukeEntityEmbeddings
[ -0.0002992511144839227, 0.03299694135785103, 0.00784542690962553, -0.008768417872488499, -0.0014854392502456903, 0.04407284036278725, 0.025151515379548073, 0.011652766726911068, -0.0016801328165456653, 0.023074785247445107, 0.016036976128816605, 0.025382263585925102, -0.002754552522674203, ...
[ "Dropout", "Embedding", "LayerNorm", "Linear", "ModelEntityEmbeddings", "Module", "None", "__init__", "clamp", "class", "config", "def", "dim", "dropout", "embeddings", "entity_emb_size", "entity_embedding_dense", "entity_embeddings", "entity_ids", "entity_vocab_size", "eps",...
luke/modeling_luke.py:LukeSelfAttention
[ -0.0002379106153966859, 0.039088357239961624, 0.045451581478118896, -0.009260759688913822, -0.0009658460621722043, 0.03181610628962517, 0.01795337349176407, -0.01016979105770588, 0.0007172827608883381, 0.018748776987195015, 0.021589500829577446, 0.03227062150835991, 0.000025189230655087158, ...
[ "Dropout", "False", "Linear", "ModelSelfAttention", "Module", "None", "The", "ValueError", "__init__", "a", "all_head_size", "and", "attention", "attention_head_size", "attention_mask", "attention_probs", "attention_probs_dropout_prob", "attention_scores", "cat", "class", "co...
luke/modeling_luke.py:LukeSelfOutput
[ -0.00012168083776487038, 0.04875698313117027, 0.03859927877783775, 0.020879726856946945, -0.0006630723946727812, 0.05620596557855606, 0.02370131015777588, -0.01952536590397358, 0.0035551965702325106, 0.017155233770608902, 0.01365646906197071, 0.003512872848659754, 0.0036962758749723434, -0...
[ "Dropout", "LayerNorm", "Linear", "ModelSelfOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "eps", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "layer_norm_eps", "nn", "return", "self", "super" ]
luke/modeling_luke.py:LukeAttention
[ -0.00010630641918396577, 0.03058813512325287, 0.056003276258707047, -0.014169503934681416, -0.0005447106086649001, 0.05307941138744354, 0.031262874603271484, -0.029913395643234253, 0.005116765387356281, -0.002642724895849824, 0.021928993985056877, 0.04250850901007652, 0.0014619329012930393, ...
[ "False", "ModelAttention", "ModelSelfAttention", "ModelSelfOutput", "Module", "None", "__init__", "attention_mask", "attention_output", "cat", "class", "concat_hidden_states", "concat_self_outputs", "config", "def", "dim", "else", "entity_attention_output", "entity_hidden_states"...
luke/modeling_luke.py:LukeIntermediate
[ -0.00025367451598867774, 0.02240910567343235, 0.04047359153628349, 0.012862369418144226, -0.0009432404185645282, 0.03635763004422188, 0.03452831506729126, -0.018864808604121208, -0.001329111517407, -0.003772961674258113, 0.023209432139992714, -0.02103712037205696, -0.0013719861162826419, 0...
[ "ACT2FN", "Linear", "ModelIntermediate", "Module", "__init__", "class", "config", "def", "dense", "else", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "intermediate_act_fn", "intermediate_size", "isinstance", "nn", "return", "self", "str", "super" ]
luke/modeling_luke.py:LukeOutput
[ -0.0002591986849438399, 0.03935529664158821, 0.0507957898080349, 0.030660521239042282, -0.0012513039400801063, 0.04988054931163788, 0.03798243775963783, -0.023681821301579475, 0.002259497530758381, 0.014758236706256866, 0.011040075682103634, 0.00267421524040401, 0.0013013561256229877, 0.00...
[ "Dropout", "LayerNorm", "Linear", "ModelOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "eps", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "intermediate_size", "layer_norm_eps", "nn", "return", "self", "su...
luke/modeling_luke.py:LukeLayer
[ -0.00026279655867256224, 0.015682563185691833, 0.03409252688288689, 0.00037643834366463125, -0.0008771723369136453, 0.05113879218697548, 0.019660023972392082, -0.007443535141646862, -0.0011435202322900295, 0.005397983826696873, 0.011534638702869415, 0.019660023972392082, 0.001832473441027104...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelIntermediate", "ModelLayer", "ModelOutput", "None", "__init__", "apply_chunking_to_forward", "attention", "attention_mask", "attention_output", "cat", "chunk_size_feed_forward", "class", "concat_attention_output", "config...
luke/modeling_luke.py:LukeEncoder
[ -0.00026698975125327706, 0.021101148799061775, 0.03096364252269268, 0.007855590432882309, -0.0010894614970311522, 0.06009240448474884, 0.012672157026827335, -0.031193003058433533, 0.003039024071767926, 0.017202023416757584, 0.01777542382478714, 0.011984076350927353, -0.0014550044434145093, ...
[ "BaseModelModelOutput", "False", "ModelEncoder", "ModelLayer", "Module", "ModuleList", "None", "True", "_", "__init__", "all_entity_hidden_states", "all_self_attentions", "all_word_hidden_states", "attention_mask", "attentions", "class", "config", "def", "else", "entity_hidden_...
luke/modeling_luke.py:LukePooler
[ -0.0002915378427132964, 0.011618588119745255, 0.03685896843671799, 0.022893769666552544, -0.001151842763647437, 0.03296703100204468, 0.037087906152009964, -0.0066105760633945465, -0.0011303798528388143, 0.0015954095870256424, 0.014022434130311012, -0.016025640070438385, -0.001244848710484802...
[ "Linear", "ModelPooler", "Module", "Tanh", "__init__", "activation", "class", "config", "def", "dense", "first_token_tensor", "forward", "hidden_size", "hidden_states", "nn", "pooled_output", "return", "self", "super" ]