We present an approach to modifying Transformer architectures by integrating graph-aware relational reasoning into the attention mechanism.
-
lamm-mit/Llama-3.2-3B-Instruct-Sparse-GIN-orca-math-word-problems
Updated • 10 • 1 -
lamm-mit/Llama-3.2-3B-Instruct-Sparse-GIN-logic
Updated • 6 • 1 -
lamm-mit/Llama-3.2-3B-Instruct-Sparse-GIN-bio
Updated • 8 -
Graph-Aware Isomorphic Attention for Adaptive Dynamics in Transformers
Paper • 2501.02393 • Published • 6