| 0
				 stringclasses 12
				values | 1
				 float64 0 55.9k | 
|---|---|
| 
	megatron.core.transformer.mlp.forward.activation | 0.336544 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc2 | 7.055808 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp | 10.999552 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 1.777888 | 
| 
	megatron.core.transformer.attention.forward.qkv | 0.029792 | 
| 
	megatron.core.transformer.attention.forward.adjust_key_value | 0.002944 | 
| 
	megatron.core.transformer.attention.forward.rotary_pos_emb | 0.002944 | 
| 
	megatron.core.transformer.attention.forward.core_attention | 3.3504 | 
| 
	megatron.core.transformer.attention.forward.linear_proj | 2.982944 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attention | 6.386656 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.02352 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc1 | 0.037312 | 
| 
	megatron.core.transformer.mlp.forward.activation | 0.008544 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc2 | 1.616576 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp | 1.673856 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.025568 | 
| 
	megatron.core.transformer.attention.forward.qkv | 0.02912 | 
| 
	megatron.core.transformer.attention.forward.adjust_key_value | 0.002944 | 
| 
	megatron.core.transformer.attention.forward.rotary_pos_emb | 0.002944 | 
| 
	megatron.core.transformer.attention.forward.core_attention | 2.944672 | 
| 
	megatron.core.transformer.attention.forward.linear_proj | 5.2744 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attention | 8.271712 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.023328 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc1 | 0.037312 | 
| 
	megatron.core.transformer.mlp.forward.activation | 0.008832 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc2 | 1.51664 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp | 1.574528 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.025632 | 
| 
	megatron.core.transformer.attention.forward.qkv | 0.071104 | 
| 
	megatron.core.transformer.attention.forward.adjust_key_value | 0.002944 | 
| 
	megatron.core.transformer.attention.forward.rotary_pos_emb | 0.002944 | 
| 
	megatron.core.transformer.attention.forward.core_attention | 39.779808 | 
| 
	megatron.core.transformer.attention.forward.linear_proj | 2.902944 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attention | 42.777534 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.064864 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc1 | 0.11856 | 
| 
	megatron.core.transformer.mlp.forward.activation | 0.017696 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc2 | 0.288352 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp | 0.435968 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.064576 | 
| 
	megatron.core.transformer.attention.forward.qkv | 0.0736 | 
| 
	megatron.core.transformer.attention.forward.adjust_key_value | 0.002912 | 
| 
	megatron.core.transformer.attention.forward.rotary_pos_emb | 0.002976 | 
| 
	megatron.core.transformer.attention.forward.core_attention | 4.52384 | 
| 
	megatron.core.transformer.attention.forward.linear_proj | 2.472896 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attention | 7.093984 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.064768 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc1 | 0.118752 | 
| 
	megatron.core.transformer.mlp.forward.activation | 0.017536 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc2 | 0.293792 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp | 0.442304 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.064832 | 
| 
	megatron.core.transformer.attention.forward.qkv | 0.240416 | 
| 
	megatron.core.transformer.attention.forward.adjust_key_value | 0.002944 | 
| 
	megatron.core.transformer.attention.forward.rotary_pos_emb | 0.002944 | 
| 
	megatron.core.transformer.attention.forward.core_attention | 17.291807 | 
| 
	megatron.core.transformer.attention.forward.linear_proj | 0.69632 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attention | 18.252129 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.233152 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc1 | 0.427968 | 
| 
	megatron.core.transformer.mlp.forward.activation | 0.047776 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc2 | 0.948512 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp | 1.435648 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.232416 | 
| 
	megatron.core.transformer.attention.forward.qkv | 0.244032 | 
| 
	megatron.core.transformer.attention.forward.adjust_key_value | 0.003104 | 
| 
	megatron.core.transformer.attention.forward.rotary_pos_emb | 0.003072 | 
| 
	megatron.core.transformer.attention.forward.core_attention | 237.877594 | 
| 
	megatron.core.transformer.attention.forward.linear_proj | 0.852032 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attention | 238.997757 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.232704 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc1 | 0.434816 | 
| 
	megatron.core.transformer.mlp.forward.activation | 0.048864 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc2 | 0.947072 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp | 1.442464 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.232768 | 
| 
	megatron.core.transformer.attention.forward.qkv | 0.918816 | 
| 
	megatron.core.transformer.attention.forward.adjust_key_value | 0.002976 | 
| 
	megatron.core.transformer.attention.forward.rotary_pos_emb | 0.00288 | 
| 
	megatron.core.transformer.attention.forward.core_attention | 118.55085 | 
| 
	megatron.core.transformer.attention.forward.linear_proj | 3.06144 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attention | 122.554077 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.899232 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc1 | 1.805568 | 
| 
	megatron.core.transformer.mlp.forward.activation | 0.170976 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc2 | 3.497248 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp | 5.484864 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.894496 | 
| 
	megatron.core.transformer.attention.forward.qkv | 0.940384 | 
| 
	megatron.core.transformer.attention.forward.adjust_key_value | 0.00304 | 
| 
	megatron.core.transformer.attention.forward.rotary_pos_emb | 0.00288 | 
| 
	megatron.core.transformer.attention.forward.core_attention | 82.459389 | 
| 
	megatron.core.transformer.attention.forward.linear_proj | 3.145088 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attention | 86.56794 | 
| 
	megatron.core.transformer.transformer_layer._forward_attention.self_attn_bda | 0.89344 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc1 | 1.804992 | 
| 
	megatron.core.transformer.mlp.forward.activation | 0.169888 | 
| 
	megatron.core.transformer.mlp.forward.linear_fc2 | 3.509728 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp | 5.495744 | 
| 
	megatron.core.transformer.transformer_layer._forward_mlp.mlp_bda | 0.893472 | 
			Subsets and Splits
				
	
				
			
				
No community queries yet
The top public SQL queries from the community will appear here once available.
