"For example, HyperAttention makes the inference time of ChatGLM2 50% faster on 32k context length while perplexity increases from 5.6 to 6.3."
"when half of all attention layers are patched (i.e., 14 layers), we verify that most of the tasks do not degrade more than 13%."
According to the paper, for most tasks it reduces benchmark scores substantially. Perhaps to the point where a smaller model would yield better inference time and higher benchmarks.
However, summarization benchmarks see almost no degredation, great!
Personally, I have found that Mistral 7B (with its native 8K context, and decent results stretched out even more) is performing much better than llama 13B tunes for storytelling, where that long context is really important.
And I think the optimized backends should implement that sliding 16k context soon...
Anyway, point is a huge context really helps certain types of queries, and VRAM usage is reasonable with a 7B model.
ML researchers are playing scientists: tweak a few parameters in an LLM, re-train it on a largish dataset (need access to $$$ GPUs), find metrics on which the tweaked LLM makes a barely noticeable improvement and make the other metrics where it actually gets worse look insignificant, write a paper, upload to arxiv, and update your resume.
So true, if they were real scientists then they'd never publish negative results and only pass them along in their network to most efficiently gate keep the field!
Also publish yet another "SOTA framework" thats barebones and won't be maintained for very long!
The researchers who published the negative CFG LLM paper made an earnest effort to pull it into the popular frameworks. That really stands out in my memory, that is incredibly rare.
This paper presents formal results, apparently. Also, in the case of formal results, peer review by experts in the exact sub-field makes sense/increases trust, sure, but why not share on archive instead of waiting for a year (or however long the process takes in the particular instance)?
"when half of all attention layers are patched (i.e., 14 layers), we verify that most of the tasks do not degrade more than 13%."
According to the paper, for most tasks it reduces benchmark scores substantially. Perhaps to the point where a smaller model would yield better inference time and higher benchmarks.
However, summarization benchmarks see almost no degredation, great!