Vicente

Computer Vision Reading Group

This semester we will start a Computer Vision seminar. The overall purpose of this seminar is to bring together people with interests in Computer Vision theory and techniques and to examine some current research papers. This seminar is organized by the students and collaborators of the Vision, Language and Learning (VISLANG) research group. Any student who has some exposure to Computer Vision and Machine Learning should be able to follow the group.

The seminar will focus on topics related to the main research projects in Computer Vision in our group. These include work at the intersection of language and vision, computer vision for images on the web, general works on deep learning and machine learning methods that can be potentially applied to vision and language problems,methods for learning about sense of place, visual recognition of objects in complex environments, buidling systems that can recognize objects and places efficiently. Other topics will depend on the interests of the participants and new members of the reading group.

Schedule - Spring 2022 Meeting on Wednesdays from 2:30pm to 3:30pm CST - First two weeks will be online
DatePresenter(s)Topic(s)
April 27 Ziyan
Conditional Prompt Learning for Vision-Language Models. Kaiyang Zhou, Jingkang Yang, Chen Change Loy, Ziwei Liu.
CVPR 2022.

April 20 Peikun
VOS: Learning What You Don't Know by Virtual Outlier Synthesis. Xuefeng Du, Zhaoning Wang, Mu Cai, Yixuan Li.
ICLR 2022.

April 13 Aman
Make-A-Scene: Scene-Based Text-to-Image Generation with Human Priors. Oran Gafni, Adam Polyak, Oron Ashual, Shelly Sheynin, Devi Parikh, Yaniv Taigman. arXiv 2022.
Hierarchical Text-Conditional Image Generation with CLIP Latents. Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, Mark Chen.
OpenAI Preprint 2022.

April 6 Paola
CNN Filter DB: An Empirical Investigation of Trained Convolutional Filters. Paul Gavrikov, Janis Keuper.
CVPR 2022.

March 30 Diego
MOTSynth: How Can Synthetic Data Help Pedestrian Detection and Tracking?. Matteo Fabbri, Guillem Braso, Gianluca Maugeri, Orcun Cetintas, Riccardo Gasparini, Aljosa Osep, Simone Calderara, Laura Leal-Taixe, Rita Cucchiara.
ICCV 2021.

March 23 Jaspreet
HyperTransformer: Model Generation for Supervised and Semi-Supervised Few-Shot Learning. Andrey Zhmoginov, Mark Sandler, Max Vladymyrov.
Arxiv 2021.

March 16 Ziyan
All Tokens Matter: Token Labeling for Training Better Vision Transformers. Zihang Jiang, Qibin Hou, Li Yuan, Daquan Zhou, Yujun Shi, Xiaojie Jin, Anran Wang, Jiashi Feng.
NeurIPS 2021.

March 9 Peikun
Resolution-robust Large Mask Inpainting with Fourier Convolutions. Roman Suvorov, Elizaveta Logacheva, Anton Mashikhin, Anastasia Remizova, Arsenii Ashukha, Aleksei Silvestrov, Naejin Kong, Harshith Goka, Kiwoong Park, Victor Lempitsky.
WACV 2022.

March 2 Aman
How Do Vision Transformers Work?. Namuk Park, Songkuk Kim.
ICLR 2022.

February 23 Paola
Learning to See by Looking at Noise. Manel Baradad, Jonas Wulff, Tongzhou Wang, Phillip Isola, Antonio Torralba.
NeurIPS 2021.

February 16 Jaspreet
VirTex: Learning Visual Representations from Textual Annotations. Karan Desai, Justin Johnson.
CVPR 2021.

February 9 Ziyan
Efficient Training of Visual Transformers with Small Datasets. Yahui Liu, Enver Sangineto, Wei Bi, Nicu Sebe, Bruno Lepri, Marco De Nadai.
NeurIPS 2021.

February 2 Peikun Guo
Real-Time High-Resolution Background Matting. Shanchuan Lin, Andrey Ryabtsev, Soumyadip Sengupta, Brian Curless, Steve Seitz, Ira Kemelmacher-Shlizerman.
CVPR 2021.

January 26 Aman
Explaining in Style: Training a GAN to explain a classifier in StyleSpace. Oran Lang, Yossi Gandelsman, Michal Yarom, Yoav Wald, Gal Elidan, Avinatan Hassidim, William T. Freeman, Phillip Isola, Amir Globerson, Michal Irani, Inbar Mosseri.
ICCV 2021.

January 19 Paola
IA-RED2: Interpretability-Aware Redundancy Reduction for Vision Transformers. Bowen Pan, Rameswar Panda, Yifan Jiang, Zhangyang Wang, Rogerio Feris, Aude Oliva.
NeurIPS 2021.

January 12 Srishti Yadav
RegionCLIP: Region-based Language-Image Pretraining. Yiwu Zhong, Jianwei Yang, Pengchuan Zhang, Chunyuan Li, Noel Codella, Liunian Harold Li, Luowei Zhou, Xiyang Dai, Lu Yuan, Yin Li, Jianfeng Gao.
Arxiv preprint.

Schedule - Spring 2021 Meeting on Mondays from 1pm to 2:30pm - Online
DatePresenter(s)Topic(s)
May 17 Paola
Open World Compositional Zero-Shot Learning. Massimiliano Mancini, Muhammad Ferjad Naeem, Yongqin Xian, Zeynep Akata.
CVPR 2021.

May 3 Fuwen
Towards Streaming Perception. Mengtian Li, Yu-Xiong Wang, Deva Ramanan.
ECCV 2020.

April 26 Leticia
Differentiable Patch Selection for Image Recognition. Jean-Baptiste Cordonnier, Aravindh Mahendran, Alexey Dosovitskiy, Dirk Weissenborn, Jakob Uszkoreit, Thomas Unterthiner.
CVPR 2021.

April 19 Aman
Towards Unsupervised Image Captioning with Shared Multimodal Embeddings. Iro Laina, Christian Rupprecht, Nassir Navab.
ICCV 2019.

April 12 Ziyan
Decoupling the Role of Data, Attention, and Losses in Multimodal Transformers. Lisa Anne Hendricks, John Mellor, Rosalia Schneider, Jean-Baptiste Alayrac, Aida Nematzadeh.
pre-print of MIT Press Publication version.

April 5 Aman
Barlow Twins: Self-Supervised Learning via Redundancy Reduction. Jure Zbontar, Li Jing, Ishan Misra, Yann LeCun, Stéphane Deny.
arxiv 2103.03230.

On Mutual Information in Contrastive Learning for Visual Representations. Mike Wu, Chengxu Zhuang, Milan Mosse, Daniel Yamins, Noah Goodman.
arxiv 2005.13149.

March 29 Paola
Finetuning Pretrained Transformers into RNNs. Jungo Kasai, Hao Peng, Yizhe Zhang, Dani Yogatama, Gabriel Ilharco, Nikolaos Pappas, Yi Mao, Weizhu Chen, Noah A. Smith.
arxiv 2103.13076.

March 22 Fuwen
Zero-Shot Text-to-Image Generation. Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, Ilya Sutskever.
arXiv 2102.12092.

March 8 Ziyan
Data Parameters: A New Family of Parameters for Learning a Differentiable Curriculum. Shreyas Saxena, Oncel Tuzel, Dennis DeCoste.
NeurIPS 2019.

March 1 Tianlu & Jaspreet
Image Representations Learned With Unsupervised Pre-Training Contain Human-like Biases. Ryan Steed, Aylin Caliskan.
arXiv 2010.15052.

Feb 22 Leticia
Layer-Wise Data-Free CNN Compression. Maxwell Horton, Yanzi Jin, Ali Farhadi, Mohammad Rastegari.
arxiv 2011.09058.

Feb 15 Tianlu
Learning from others' mistakes: Avoiding dataset biases without modeling them. Victor Sanh, Thomas Wolf, Yonatan Belinkov, Alexander M. Rush.
ICLR 2021.

Feb 1 Aman
WaveNet: A Generative Model for Raw Audio. Aaron van den Oord, Sander Dieleman, Heiga Zen, Karen Simonyan, Oriol Vinyals, Alex Graves, Nal Kalchbrenner, Andrew Senior, Koray Kavukcuoglu.
1609.03499.
Wave-Tacotron: Spectrogram-free end-to-end text-to-speech synthesis. Ron J. Weiss, RJ Skerry-Ryan, Eric Battenberg, Soroosh Mariooryad, Diederik P. Kingma.
ICASSP 2020.

Jan 18 Paola
Learning Transferable Visual Models From Natural Language Supervision. Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, Ilya Sutskever.
OpenAI preprint.

Jan 11 Fuwen
When Do Curricula Work? Xiaoxia Wu, Ethan Dyer, Behnam Neyshabur.
ICLR 2021.

Jan 4 Ziyan
Globetrotter: Unsupervised Multilingual Translation from Visual Alignment. Dídac Surís, Dave Epstein, Carl Vondrick.
arxiv 2012.04631.


Department of Computer Science @ Rice University