Update modeling_phi3_small.py
Browse files- modeling_phi3_small.py +0 -3
modeling_phi3_small.py
CHANGED
|
@@ -215,9 +215,6 @@ class Phi3SmallSelfAttention(nn.Module):
|
|
| 215 |
f"Layer {layer_idx + 1} is using dense attention since it is divisible by "
|
| 216 |
f"{self.config.dense_attention_every_n_layers}"
|
| 217 |
)
|
| 218 |
-
# use warnings to allow the modeling use different flash attention implementation later
|
| 219 |
-
if not is_flash_attention_available:
|
| 220 |
-
logger.warning_once("Flash Attention is not available, but is needed for dense attention")
|
| 221 |
else:
|
| 222 |
# BlockSparse related Parameters
|
| 223 |
self.blocksparse_params = BlockSparseParams.from_config(config)
|
|
|
|
| 215 |
f"Layer {layer_idx + 1} is using dense attention since it is divisible by "
|
| 216 |
f"{self.config.dense_attention_every_n_layers}"
|
| 217 |
)
|
|
|
|
|
|
|
|
|
|
| 218 |
else:
|
| 219 |
# BlockSparse related Parameters
|
| 220 |
self.blocksparse_params = BlockSparseParams.from_config(config)
|