realitydefender - Multimodal AI PhD Intern (Spring 2026)
Upload My Resume
Drop here or click to browse · Tap to choose · PDF, DOCX, DOC, RTF, TXT
Requirements
• PhD student in a relevant technical field, preferably three or more years into the program • Experience in multi-modal learning, such as in audio-visual classification and audio-language reasoning. • Proficient in Python and in building deep learning models with PyTorch. • Published peer-reviewed research papers in reputable AI and speech venues, e.g. CVPR, NeurIPS, ACL, Interspeech. • Excited about Reality Defender's mission to build a best-in-class and comprehensive deepfake and AI-generated content detection platform. • Available to start in Spring 2026, for a minimum duration of 3 months.
Responsibilities
• Investigate and propose new methods for detecting generative multi-modal content, spanning audio and vision modalities. • Perform research on multi-modal deepfake detection and reasoning tasks. • Collaborate with researchers in the team. • Write up results of research for internal reports and submission to academic journals/workshops. • Independently implement and evaluate ideas on modern deep learning stack - Python, PyTorch, and GPU-enabled cloud compute, like AWS/GCP.
Benefits
• $5K – $7K per month
No credit card. Takes 10 seconds.