DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING
UNIVERSITY OF CALIFORNIA, SAN DIEGO

CSE 252C: Selected Topics in Vision and Learning

Spring 2018

Instructor: Ben Ochoa
Email: bochoa at ucsd.edu
Office hours: M 8:00 PM-9:00 PM (primary) and W 8:00 PM-9:00 PM (secondary), EBU3B 4208, and at other times by appointment

TA: Zak Murez
Email: zmurez at eng.ucsd.edu
Office hours: MW 3:30 PM-4:30 PM, EBU3B 4127 or 4148 (directly across the hall from 4127)

Note: when emailing the instructor or TA with questions about the class, please put "CSE 252C" in the subject line.

Class section ID: 938675 (must be taken for 4 units)
Lecture: MW 5:00 PM-6:20 PM, CENTR 216
Class discussion: Piazza

This course is dedicated to recent research on computer vision, including learning-based methods. Students enrolled in this course are required to review multiple papers (~2 per week), present a paper in class, and complete a project inspired by the paper.

Each paper review must answer the questions listed in How to Read an Engineering Research Paper provided by Professor Bill Griswold. Additionally, you must post at least one question or comment about the paper on Piazza. Paper reviews and posting of Piazza questions are due 24 hours prior to the presentation of the review. Paper presenters (see below) must respond to questions/comments prior to presenting the paper in class.

When presenting a paper in class, follow the presentation guidelines provided by Professor Charles Elkan. One week prior to your presentation date, send a draft of your slides to the instructor and TA for review. The instructor and TA will provide you with comments to incorporate into your slides prior to your presentation in class. Immediately after your presentation, send the slides (PDF, one slide per page) to the instructor for publishing on the class website. You will be receive talk feedback from the instructor and TA.

All projects will follow specific guidelines, including preparation of a project proposal, draft project report, and final project report. The project should be at the frontier of current research, but need not necessarily advance the state of the field. For example, replicating the results of an innovative paper would be a good project. Projects must be closely inspired by one or two specific high quality papers and should have an experimental aspect. Project reports will be evaluated using these grading criteria.

Prerequisites: CSE 252A or CSE 252B.

Academic Integrity Policy: Integrity of scholarship is essential for an academic community. The University expects that both faculty and students will honor this principle and in so doing protect the validity of University intellectual work. For students, this means that all academic work will be done by the individual to whom it is assigned, without unauthorized aid of any kind.

Collaboration Policy: It is expected that you complete your academic assignments on your own and in your own words and code. The assignments have been developed by the instructor to facilitate your learning and to provide a method for fairly evaluating your knowledge and abilities (not the knowledge and abilities of others). So, to facilitate learning, you are authorized to discuss assignments with others; however, to ensure fair evaluations, you are not authorized to use the answers developed by another, copy the work completed by others in the past or present, or write your academic assignments in collaboration with another person.

If the work you submit is determined to be other than your own, you will be reported to the Academic Integrity Office for violating UCSD's Policy on Integrity of Scholarship. In accordance with the CSE department academic integrity guidelines, students found committing an academic integrity violation will receive an F in the course.


Grading: Course grades will be weighted as follows.

Class participation: 10%
Paper reviews: 20%
Paper presentation: 20%
Project presentation: 10%
Project report: 40%

Papers:

Date Presenters Paper Slides
April 11 Yujie Li, Xingwei Liu, and Gautam Nain DynamicFusion: Reconstruction and tracking of non-rigid scenes in real-time (Richard A. Newcombe, Dieter Fox, and Steven M. Seitz) [pdf] pdf
April 16 Marcus Loo Vergara and Asbjoern Fintland Lystrup VNect: Real-time 3D Human Pose Estimation with a Single RGB Camera (Dushyant Mehta, Srinath Sridhar, Oleksandr Sotnychenko, Helge Rhodin, Mohammad Shafiei, Hans-Peter Seidel, Weipeng Xu, Dan Casas, and Christian Theobalt) [pdf] pdf
April 18 Yiting Ethan Li, Haakon Hukkelaas, and Kaushik Ram Ramasamy Dynamic Routing Between Capsules (Sara Sabour, Nicholas Frosst, and Geoffrey E. Hinton) [pdf] pdf
April 23 Varun Syal and Sparsh Gupta Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks (Han Zhang, Tao Xu, Hongsheng Li, Shaoting Zhang, Xiaogang Wang, Xiaolei Huang, and Dimitris Metaxas) [pdf] pdf
April 25 Kamran Alipour, Jeffrey Wang, and Menghe Zhang A Closer Look at Spatiotemporal Convolutions for Action Recognition (Du Tran, Heng Wang, Lorenzo Torresani, Jamie Ray, Yann LeCun, and Manohar Paluri) [pdf] pdf
April 30 Chih-Hui Ho, Xingyu Gu, and Yuan Qi Unsupervised Domain Adaptation by Backpropagation (Yaroslav Ganin and Victor Lempitsky) [pdf] pdf
May 2 Alexandr Kuznetsov and Karen Lucknavalai Arbitrary Style Transfer in Real-Time with Adaptive Instance Normalization (Xun Huang and Serge Belongie) [pdf] pdf
May 7 Ajitesh Gupta Artistic Style Transfer for Videos (Manuel Ruder, Alexey Dosovitskiy, and Thomas Brox) [pdf] pdf
May 9 Yunhan Ma, Jingyao Zhan, and Jiageng Zhang Mask R-CNN (Kaiming He, Georgia Gkioxari, Piotr Dollar, and Ross Girshick) [pdf] pdf
May 14 Patrick Hayes Transparency by Design: Closing the Gap Between Performance and Interpretability in Visual Reasoning (David Mascharka, Philip Tran, Ryan Soklaski, and Arjun Majumdar) [pdf] pdf
May 16

Group meetings with instructor and TA

May 21

Group meetings with instructor and TA

May 23

Group meetings with instructor and TA

May 28 No meeting (Memorial Day Observance)
May 30

Group meetings with instructor and TA

June 4

Project presentations

Ajitesh Gupta: Video Colorization

Yunhan Ma, Jingyao Zhan, and Jiageng Zhang: Segmentation-Based Object Tracking

Patrick Hayes: Evaluation of learning using visual question and answering

Yujie Li, Xingwei Liu, and Gautam Nain: Reconstruction and tracking of non-rigid scenes in real-time

Alexandr Kuznetsov and Karen Lucknavalai: Arbitrary Style Transfer in Real-time with Adaptive Instance Normalization

June 6

Project presentations

Marcus Loo Vergara and Asbjoern Fintland Lystrup: Real-time 3D Human Pose Estimation with a Single RGB Camera

Yiting Ethan Li, Haakon Hukkelaas, and Kaushik Ram Ramasamy: Experiments with Capsules on Various Datasets

Kamran Alipour, Jeffrey Wang, and Menghe Zhang: A Closer Look at Spatiotemporal Convolutions for Action Recognition

Varun Syal and Sparsh Gupta: Text to Image Synthesis Using StackGAN

Chih-Hui Ho, Xingyu Gu, and Yuan Qi: Unsupervised Domain Adaptation by Backpropagation

Projects:

Group members Project
Kamran Alipour, Jeffrey Wang, and Menghe Zhang A Closer Look at Spatiotemporal Convolutions for Action Recognition
Yiting Ethan Li, Haakon Hukkelaas, and Kaushik Ram Ramasamy Experiments with Capsules on Various Datasets
Yujie Li, Xingwei Liu, and Gautam Nain Reconstruction and tracking of non-rigid scenes in real-time
Alexandr Kuznetsov and Karen Lucknavalai Arbitrary Style Transfer in Real-time with Adaptive Instance Normalization
Marcus Loo Vergara and Asbjoern Fintland Lystrup Real-time 3D Human Pose Estimation with a Single RGB Camera
Chih-Hui Ho, Xingyu Gu, and Yuan Qi Unsupervised Domain Adaptation by Backpropagation
Varun Syal and Sparsh Gupta Text to Image Synthesis Using StackGAN
Ajitesh Gupta Video Colorization
Yunhan Ma, Jingyao Zhan, and Jiageng Zhang Segmentation-Based Object Tracking
Patrick Hayes Evaluation of learning using visual question and answering

Last update: May 21, 2018