Jathushan Rajasegaran

I am a Ph.D. student at BAIR advised by Prof. Jitendra Malik. I am broadly interested in Computer Vision and Deep Learning, with a focus on developing models for understanding long videos. Before coming to Berkeley, I was working with Prof. Salman Khan at Inception Institute. I completed my undergraduate study at University of Moratuwa, with a major in Electronic and Telecommunication Engineering. My Bachelor's Thesis was advised by Dr. Ranga Rodigo.

Email  /  Bio  /  Google Scholar  /  Github /  Photos /  Art




Cars don't run like cheetahs, Planes don't fly like birds and Machines won't think in a way same as humans. They will do better. --Richard Feynman

We want AI agents that can discover like we can, not which contain what we have discovered. --Richard Sutton

Machines should be able to understand the world outside our window. That world may change and evolve, but the machines should perceive trees, cars and spaceships not pixels. --Max Wertheimer, Stan Lee

Jathushan2
Research

My research interests lie in the general area of computer vision and deep learning, particularly in learning from videos, deep neural architectures and meta/continual learning.

Blog 1: Research Statement (as of June 2023)



Humans in 4D: Reconstructing and Tracking Humans with Transformers.
Shubham Goel, Georgios Pavlakos, Jathushan Rajasegaran, Angjoo Kanazawa,
Jitendra Malik
International Conference on Computer Vision (ICCV) , 2023  
project page/ arxiv/ code/ demo/

A fully "transformerized" deisgn for Human Mesh Recovery achieves improved precision and remarkable robustness for 3D human reconstruction and tracking!.



On the Benefits of 3D Pose and Tracking for Human Action Recognition.
Jathushan Rajasegaran, Georgios Pavlakos, Angjoo Kanazawa,
Christoph Feichtenhofer, Jitendra Malik
Computer Vision and Pattern Recognition (CVPR) , 2023  
project page/ paper/ arxiv/ code/ demo/ poster

Using 3D human reconstruction and tracking to recognize atomic actions in video.



Tracking People by Predicting 3D Appearance, Location and Pose.
Jathushan Rajasegaran, Georgios Pavlakos, Angjoo Kanazawa, Jitendra Malik
Computer Vision and Pattern Recognition (CVPR) , 2022
(Oral Presentation) (Best paper finalist - Top 0.4%)
paper/ arxiv/ project page/ video/ results/ poster/ code

Performing monocular tracking of people by predicting their appearance, pose and location and in 3D.



Tracking People with 3D Representations.
Jathushan Rajasegaran, Georgios Pavlakos, Angjoo Kanazawa, Jitendra Malik
Neural Information Processing Systems (NeurIPS), 2021
paper/ arxiv/ project page/ video/ code/ poster

Performing monocular tracking of people by lifting them to 3D and then using 3D representations of their appearance, pose and location.

iTAML: An Incremental Task-Agnostic Meta-learning Approach.
Jathushan Rajasegaran, Salman Khan, Munawar Hayat, Fahad Shahbaz Khan
Computer Vision and Pattern Recognition (CVPR) , 2020
paper/ arxiv/ slides/ video/ code

By learning generic represenatations from past tasks, we can easily adapt to new tasks as well as remember old tasks.

Random Path Selection for Incremental Learning.
Jathushan Rajasegaran, Munawar Hayat, Salman Khan,
Fahad Shahbaz Khan, Ling Shao
Neural Information Processing Systems (NeurIPS), 2019
paper/ arxiv/ poster/ code

We increase the width of a ResNet like model by adding extra skip connections when new tasks are introduced.

DeepCaps: Going Deeper with Capsule Networks
Jathushan Rajasegaran, Vinoj Jayasundara, Sandaru Jayasekara,
Hirunima Jayasekara, Suranga Seneviratne, Ranga Rodrigo
Computer Vision and Pattern Recognition (CVPR) , 2019   (Oral Presentation)
paper/ poster/ video/ code

Capsule Networks are cool, but they are shallow. We can increase the depth by 3D convolutions and skip connections.

TextCaps: Handwritten Character Recognition with Very Small Datasets
Vinoj Jayasundara, Sandaru Jayasekara, Hirunima Jayasekara, Jathushan Rajasegaran,
Suranga Seneviratne, Ranga Rodrigo
Winter Conference on Applications of Computer Vision (WACV) , 2019
paper/ arxiv/ poster/ code

Capsule Networks can capture actual variations that are present in human hand writing, so we generate more data and retrain the capsule networks.

A Multi-modal Neural Embeddings Approach for Detecting Mobile Counterfeit Apps: A Case Study on Google Play Store
Jathushan Rajasegaran, Naveen Karunanayake, Ashanie Gunathillake,
Suranga Seneviratne, Guillaume Jourjon
International World Wide Web Conference (WWW), 2019
paper/ arxiv/ poster/

We use content and style representations detect counterfeit apps in playstore.


Website source from Jon Barron here