2022/2023

Degree | Type | Year | Semester |
---|---|---|---|

4314099 Computer Vision | OB | 0 | 1 |

- Name:
- Maria Vanrell Martorell
- Email:
- maria.vanrell@uab.cat

- Principal working language:
- english (eng)

- Oriol Ramos Terrades

- Coloma Ballester
- Juan F. Garamendi
- Karim Lekadir
- Pablo Arias

Degree in Engineering, Mathematics, Physics or similar.

**Module Coordinator:** Dra. Coloma Ballester

The aim of this module is to learn about the optimization algorithms and inference techniques that are behind many tasks in computer vision. The main concepts will include energy formulation and minimization, numerical techniques for variational problems, gradient descent optimization algorithms and tools useful for deep learning strategies. convex optimization, and graphical models. These techniques will be applied in the project in the context of image segmentation and inpainting.

- Accept responsibilities for information and knowledge management.
- Choose the most suitable software tools and training sets for developing solutions to problems in computer vision.
- Conceptualise alternatives to complex solutions for vision problems and create prototypes to show the validity of the system proposed.
- Continue the learning process, to a large extent autonomously.
- Identify concepts and apply the most appropriate fundamental techniques for solving basic problems in computer vision.
- Plan, develop, evaluate and manage solutions for projects in the different areas of computer vision.
- Solve problems in new or little-known situations within broader (or multidisciplinary) contexts related to the field of study.
- Understand, analyse and synthesise advanced knowledge in the area, and put forward innovative ideas.
- Use acquired knowledge as a basis for originality in the application of ideas, often in a research context.
- Work in multidisciplinary teams.

- Accept responsibilities for information and knowledge management.
- Choose optimisation and inference techniques and train them to resolve a particular project.
- Continue the learning process, to a large extent autonomously.
- Identify the basic concepts of graphical models and inference algorithms.
- Identify the basic optimisation techniques and their associated algorithms.
- Identify the best representations that can be defined for solving both optimisation and inference problems with graphical models.
- Solve problems in new or little-known situations within broader (or multidisciplinary) contexts related to the field of study.
- Understand, analyse and synthesise advanced knowledge in the area, and put forward innovative ideas.
- Use acquired knowledge as a basis for originality in the application of ideas, often in a research context.
- Use optimisation and inference techniques to plan, develop, evaluate and manage a solution to a particular problem.
- Work in multidisciplinary teams.

- Introduction to optimization problems and energy minimization methods. Examples and overview of a variational formulation.
- Review of numerical linear algebra: least squares methods, singular value decomposition, pseudoinverse, iterative methods. Applications.
- Numerical techniques for variational problems: Gateaux derivative, Euler-Lagrange equation and gradient methods. Applications: denoising, image inpainting and Poisson editing. The Backpropagation strategy for gradient computation. Gradient descent optimization algorithms useful for deep learning strategies.
- Convex optimization. Constrained and unconstrained optimization. Duality principles and methods. Non-convex problems and convex relaxation. Applications: Total Variation restoration, disparity computation, optical flow computation.
- Segmentation with variational models. The Mumford and Shah Functional. Explicit and implicit Shape Representations. Level sets formulation.
- Bayesian networks and MRFs. Inference types. Main Inference algorithms. Examples: stereo, denoising.
- Inference algorithms. Belief propagation: message passing, loopy belief propagation. Example: inference for segmentation.
- Sampling methods: Particle-based methods, Markov Chain Monte Carlo, Gibbs Sampling.

**Supervised sessions: ***(Some of these sessions could be on-line synchronous)*

**Lecture Sessions**, where the lecturers will explain general contents about the topics. Some of them will be used to solve the problems.

**Directed sessions:**

**Project Sessions**, where the problems and goals of the projects will be presented and discussed, students will interact with the project coordinator about problems and ideas on solving the project (approx. 1 hour/week)**Presentation Session**, where the students give an oral presentation about how they have solved the project and a demo of the results.**Exam Session**, where the students are evaluated individually. Knowledge achievements and problem-solving skills

**Autonomous work:**

- Student will autonomously study and work with the materials derived from the lectures.
- Student will work in
**groups**to solve the problems of the projects with deliverables:- Code
- Reports
- Oral presentations

**Annotation**: Within the schedule set by the centre or degree programme, 15 minutes of one class will be reserved for students to evaluate their lecturers and their courses or modules through questionnaires.

Title | Hours | ECTS | Learning Outcomes |
---|---|---|---|

Type: Directed | |||

Lecture sessions | 20 | 0.8 | 8, 4, 6, 5, 3, 2, 9, 10 |

Type: Supervised | |||

Project follow-up sessions | 8 | 0.32 | 1, 8, 4, 6, 5, 7, 3, 2, 11, 10 |

Type: Autonomous | |||

Homework | 113 | 4.52 | 1, 4, 6, 5, 7, 3, 2, 11 |

The final marks for this module will be computed with the following formula:

** ****Final Mark** = 0.4 x **Exam** + 0.55 x **Project **+ 0.05 x **Attendance**

where

**Exam:** is the mark obtained in the Module Exam **(must be >=3)**. This mark can be increased by getting extra points given by delivered exercises in specific lectures, but only if Exam Mark is greater than 3.

**Attendance**: is the mark derived from the control of attendance at lectures **(minimum 70%)**

**Project:** is the mark provided by the project coordinator based on the weekly follow-up of the project and deliverables **(must be >=5)**. All accordingly with specific criteria such as:

- Participation in discussion sessions and in team work (inter-member evaluations)
- Delivery of mandatory and optional exercises.
- Code development (style, comments, etc.)
- Report (justification of the decisions in your project development)
- Presentation (Talk and demonstrations on your project)

Only those students that fail (FinalMark < 5.0) can do a retake exam.

Title | Weighting | Hours | ECTS | Learning Outcomes |
---|---|---|---|---|

Exam | 0,4 | 2.5 | 0.1 | 8, 4, 6, 5, 7, 9 |

Project | 0,55 | 6 | 0.24 | 1, 4, 6, 5, 7, 3, 2, 11, 10 |

Session attendance | 0,05 | 0.5 | 0.02 | 1, 8, 3, 9, 11 |

Journal articles:

- Xavier Bresson and Tony F. Chan. “
*Fast Dual Minimization of the Vectorial Total Variation Norm and Applications to Color Image Processing. Inverse Problems and Imaging”*. American Institue of Mathematical Sciences. Vol 2, No. 4, pp 455-484 2008. - Chan, T. F., & Vese, L. a. “
*Active contours without edges”*. IEEE Transactions on Image Processing : A Publication of the IEEE Signal Processing Society,*10*(2), pp 266–77, 2001. - Daphne Koller and Nir Friedman, "
*Probablistic Graphical Models. Principles and techniques*", 2009. - Patrick Pérez, Michel Gangnet, and Andrew Blake. “
*Poisson image editing”*. In ACM SIGGRAPH 2003 Papers (SIGGRAPH '03). ACM, New York, NY, USA, 313-318 2003. - L.I. Rudin, S. Osher, and E. Fatemi. “Nonlinear Total Variation based Noise Removal Algorithms”. Physical D Nonlinear Phenomena, 60, pp 259-268, November 1992.
- Ruder, Sebastian. "An overview of gradient descent optimization algorithms." arXiv preprint arXiv:1609.04747(2016).

Books:

- S.P. Boyd, L. Vandenberghe, "
*Convex optimization"*, Cambridge University Press, 2004. - Tony F. Chan and Jianhong Shen. “
*Image Processing and Analysis: Variational, PDE, Wavelet and Stochastic Methods”*. Society for Industrial and Applied Mathematics, 2005. - J. Nocedal, S.J. Wright, “Numerical optimization”, Springer Verlag, 1999.
- Aubert Gilles, Pierre Kornprobst. “
*Mathematical Problems in Image Processing: Partial Differential Equations and the Calculus of Variations”*. Springer-Verlag New York. - Joe D. Hoffman. “
*Numerical Methods for Engineers and Scientists*“ - Daphne Koller and Nir Friedman, "
*Probablistic Graphical Models. Principles and techniques*", 2009. - Sebastian Nowozin and Christoph H. Lampert, "
*Structured Learning and Prediction in Computer Vision*", Foundations and Trends in Computer Graphics and Vision: Vol. 6: No. 3-4, pp 185-365, 2011. - C. Pozrikidis. “
*Numerical Computation in Science and Engineering”*.

Matlab Programming Tools with image processing and optimization toolboxes