Solution in this Paper
– Approximates floating-point multiplication using integer addition
– Linear O(n) complexity vs O(m^2) for standard floating-point multiplication
– Replaces tensor multiplications in attention mechanisms and linear transformations
– Implements L-Mul-based attention mechanism in transformer models