Official implementation for "RIFLEx: A Free Lunch for Length Extrapolation in Video Diffusion Transformers" (ICML 2025)
-
Updated
May 13, 2025 - Python
Official implementation for "RIFLEx: A Free Lunch for Length Extrapolation in Video Diffusion Transformers" (ICML 2025)
[ECCV 2024] Official PyTorch implementation of RoPE-ViT "Rotary Position Embedding for Vision Transformer"
Experimental playground for benchmarking language model (LM) architectures, layers, and tricks on smaller datasets. Designed for flexible experimentation and exploration.
[NeurIPS'23] DropPos: Pre-Training Vision Transformers by Reconstructing Dropped Positions
Official source code for the paper: "It’s Just a Matter of Time: Detecting Depression with Time-Enriched Multimodal Transformers"
[CVPR 2023] An official Pytorch implementation of "Masked Jigsaw Puzzle: A Versatile Position Embedding for Vision Transformers".
HoPE: Hybrid of Position Embedding for Length Generalization in Vision-Language Models
Word and Position Embedding visualization of all pre-trained transformer models like BERT
Add a description, image, and links to the position-embedding topic page so that developers can more easily learn about it.
To associate your repository with the position-embedding topic, visit your repo's landing page and select "manage topics."