Jian Ren
I am a Lead Research Scientist working in the Creative Vision
team at Snap Inc. |
![]() |
SnapFusion: Text-to-Image Diffusion Model on Mobile Devices within Two Seconds
|
![]() |
LightSpeed: Light and Fast Neural Light Fields on Mobile Devices
|
![]() |
Rethinking Vision Transformers for
MobileNet Size and Speed
|
![]() |
COMCAT: Towards Efficient
Compression and Customization of
Attention-Based Vision Models
|
![]() |
Real-Time Neural Light Field on Mobile Devices
|
![]() |
SINE: SINgle Image Editing with Text-to-Image Diffusion Models
|
![]() |
Unsupervised
Volumetric
Animation
|
![]() |
Make-A-Story: Visual Memory
Conditioned Consistent Story Generation
|
![]() |
Invertible Neural Skinning
|
![]() |
3D Generation on ImageNet
|
![]() |
Discrete Contrastive Diffusion for
Cross-Modal and Conditional Generation
|
![]() |
EfficientFormer: Vision
Transformers at
MobileNet Speed
|
![]() |
Layer Freezing & Data Sieving: Missing
Pieces
of a
Generic Framework for Sparse Training
|
![]() |
R2L: Distilling Neural Radiance Field to
Neural Light Field for Efficient Novel View Synthesis |
![]() |
Cross-Modal 3D Shape Generation
and
Manipulation |
![]() |
Show Me What and Tell Me How: Video
Synthesis
via Multimodal Conditioning |
![]() |
Are Multimodal Transformers Robust to
Missing Modality?
|
![]() |
In&Out: Diverse Image Outpainting via
GAN
Inversion
|
![]() |
F8Net: Fixed-Point 8-bit Only
Multiplication for Network
Quantization |
![]() |
Lottery Ticket Implies Accuracy Degradation, Is
It
a Desirable
Phenomenon? |
![]() |
Flow Guided Transformable Bottleneck Networks
for
Motion Retargeting |
![]() |
Motion Representations
for
Articulated
Animation |
![]() |
Teachers Do More Than Teach: Compressing
Image-to-Image Models |
![]() |
A Good Image Generator Is What You Need
for
High-Resolution Video Synthesis |
![]() |
SMIL: Multimodal learning with severely missing
modality |