Extend context windows of transformer models using RoPE, YaRN, ALiBi, and position interpolation techniques. Use when processing long documents (32k-128k+ tokens), extending pre-trained models beyond original context limits, or implementing efficient positional encodings.
/plugin marketplace add zechenzhangAGI/AI-research-SKILLs/plugin install long-context@zechenzhangAGI/AI-research-SKILLs