GeoFLARE: added GALE_FA, an alternate attention to GALE, for GeoTransolver#1405
GeoFLARE: added GALE_FA, an alternate attention to GALE, for GeoTransolver#1405mnabian merged 33 commits intoNVIDIA:mainfrom
Conversation
Greptile OverviewGreptile SummaryThis PR adds GAFLARE (Geometry-Aware FLARE), an alternative attention mechanism to GALE for the GeoTransolver model. Key Changes:
Critical Issues Found:
Minor Issues:
Important Files Changed
|
Additional Comments (1)
ATTENTION_CLASSES = {
"GALE": GALE,
"GAFLARE": GAFLARE,
}
if attention_type not in ATTENTION_CLASSES:
raise ValueError(f"Unknown attention_type: {attention_type}")
self.Attn = ATTENTION_CLASSES[attention_type](
hidden_dim,
heads=num_heads,
# ... rest of arguments
) |
coreyjadams
left a comment
There was a problem hiding this comment.
Hi @dakhare-creator - thanks for opening this. Overall it seems reasonable. Good to put it in experimental. I made a couple comments in the code, but also:
- it'd be unusual for use to publish a "GeoFlare" model without also making available "Flare". Can you add standard FLARE attention to the experimental/nn folder so we can do that too?
- We can also probably easily create a flare.py itself for the model.
- How much overlap in the math and attention blocks is there between PhysicsAttention and FLARE Attention? PhysicsAttention can be used on 2d and 3d data, could this? I strongly suspect yes. Can we make this modular so we can reuse as much as possible between attention layers?
Once we're ready, let's also invite the FLARE authors to take a look.
update main
update geoflare
vpuri3
left a comment
There was a problem hiding this comment.
Hi @dakhare-creator, I've left several comments in the code. Implementing FLARE straight on a geom problem can be error prone. I'd recommend first validating FLARE on one of the paper's benchmark problems
https://github.com/vpuri3/FLARE.py/blob/master/pdebench/models/flare.py
and then exposing the relevant layers to the geometry solvers interface.
|
/blossom-ci |
|
Hi @dakhare-creator - with the geometry encoded FLARE, does this also enable standard FLARE attention and model? |
Hi @coreyjadams, Yes if the context is None, the attention mechanism will be standard FLARE. |
|
/blossom-ci |
|
Hi all, is there any update on this? |
|
/blossom-ci |
|
/blossom-ci |
1 similar comment
|
/blossom-ci |
|
/blossom-ci |
PhysicsNeMo Pull Request
Description
Checklist
Dependencies
Review Process
All PRs are reviewed by the PhysicsNeMo team before merging.
Depending on which files are changed, GitHub may automatically assign a maintainer for review.
We are also testing AI-based code review tools (e.g., Greptile), which may add automated comments with a confidence score.
This score reflects the AI’s assessment of merge readiness and is not a qualitative judgment of your work, nor is
it an indication that the PR will be accepted / rejected.
AI-generated feedback should be reviewed critically for usefulness.
You are not required to respond to every AI comment, but they are intended to help both authors and reviewers.
Please react to Greptile comments with 👍 or 👎 to provide feedback on their accuracy.