12th Indian Conference on Computer Vision, Graphics
and Image Processing
December 19 - 21, IIT Jodhpur, India
Abstract: In this tutorial, I will give an overview of the progress towards the goal of single-view 3D inference i.e. being able to infer the 3D structure underlying a single input image. I will briefly summarize the different approaches proposed over the past decades and then focus on more recent learning based methods that tackle this task. Towards building more scalable approaches, we will categorize these learning based methods in context of the supervision required, and I will show how incorporating geometry in the learning process can relax the need of strong supervision. I will finally highlight some exciting future directions and open research questions.
Bio: Shubham Tulsiani is a research scientist at Facebook AI Research (FAIR) and will be joining the CMU School of Computer Science as an Assistant Professor in Fall 2021. He received a PhD. in Computer Science from UC Berkeley under the supervision of Jitendra Malik in 2018. He is interested in building perception systems that can infer the spatial and physical structure of the world they observe. His work was awarded the Best Student Paper Award at CVPR 2015.
Title of the talk: Computational Neuroimaging in Pathology: Basics and Scope
Date&Time: 20 December, 11:00 AM (IST).
Session Chair: Arnav Bhavsar
With neurological and neuro-psychiatric disorders on the rise each year, there has been an increased impetus to understand and quantify the complex information conveyed through brain magnetic resonance images (MRI) for reliable prognosis, diagnosis and for developing personalized treatment strategies for better efficacy and outcomes. Quantitative analysis of brain MRI can unravel the brain structure, function and architecture that can facilitate deeper insights into the neurobiological underpinnings of the pathology under consideration. My talk will focus on computational neuroimaging that includes pre-processing and analysis of these images that is crucial for precise quantification to overcome the limitations of subjective visual interpretation, consequently improving the workflow and reliability in radiology and supporting neurology based decision making.
Bio: Dr. Madhura Ingalhalikar, has cutting-edge experience in the field of medical image analysis for the past 13 years. She is well versed with MRI protocols, expert in processing pipelines, multi-modal MRI analysis, machine learning and deep learning. Her work on Diffusion MRI at UPenn produced multiple high impact journal papers including a paper in PNAS which was commended by the neuro-imaging community and received great media attention. Currently, she is a Professor and heads the Symbiosis Centre for Medical Image analysis (SCMIA), Pune, and focuses on using multi-modality imaging and advanced state of the art deep learning techniques in prediction and classification problems of several neurological disorders with a special emphasis on brain tumors and movement disorders.
Title of the talk: Entering the statistical domain: Use of deformable statistical shape modeling in clinical practices.
Date&Time: 20 December, 12:00 Noon (IST).
Session Chair: Arnav Bhavsar
Abstract: Medicine is evolving rapidly with emerging applications leveraging the fields of machine learning, statistics, image analysis, and biomechanics. Statistical inferences in terms of statistical shape modeling have allowed distributed diagnosis on the organs' anatomical shape (geometry), biological and population variability, and its representation in imaging modalities. Variation in size is also an important determinant for variation in many other organismal traits. The interface of statistical shape modeling of the human joints has been recently recognized within the research community. Once developed, Statistical Shape Models (SSMs) and Statistical Appearance Models (SAMs) have various applications in biomechanics including automatic image segmentation, subject-specific biomechanics models, population classifications based on bone morphometry, relationship between form or shape and function etc. The techniques involved are emerging and have promising applications towards modeling patient-specificity. This tutorial would focus on explaining the science and mathematics that goes into building statistical shape models for solving clinical problems, validations conducted on such methods, or applications of such methods to solve specific clinical problems.
Bio: Dr. Bhushan Borotikar has been a translational researcher in the field of medical sciences since 2005. His research spans the domains of medicalimaging, computational anatomy, human movement sciences, machine learning, and orthopedic biomechanics. Dr. Bhushan Borotikar is currently engaged as a research faculty(Associate Professor) at the Symbiosis Centre for Medical Image Analysis,Symbiosis International University, Pune, India. He also has adjunct appointments at the Laboratory for the Treatment of Medical Information (LaTIM),INSERM, U1101, Brest, France as a Senior Research Scientist and at the Division of Biomedical Engineering, University of Cape Town, Cape Town, South Africa, asan honorary Associate Professor. Dr Borotikar finished his PhD from Cleveland State University, Ohio, USA and Cleveland Clinic, Ohio, USA in 2009.He then worked as a post-doctoral fellow at the National Institutes of Health(NIH), Bethesda, USA till 2014. From 2014 till 2020, he was at the Frenchlaboratory of excellence- LaTIM. Dr Borotikar's current research interests are focused on developing clinical tools and procedures using emerging engineering technologies in three domains: 1) Musculoskeletal disorders and sports injuries in adults and children, 2) Computer assisted orthopaedic surgeries, joint implants and medical device designs, and 3) Advances in medical image acquisition,analysis, and synthesis. Dr Borotikar has published extensively in peer-reviewed journals, international conferences, has written book chapter(s) on his research, and has a provisional patent application on his name.
Title: Marriage of Computer Vision, Speech and Natural Language
Date&Time: 19 December, 11:00 AM (IST).
Session Chair: Aditya Nigam
Abstract: Speech generally is considered to have three parts to it: vision, aural, and the social construct. In recent years, although the field has been moving at a dramatic pace, progress is being made in silos. The primary reason for this being that speech is considered "spoken text" by practitioners and researchers alike. Most open-source datasets due to their distance from real-world conditions help in spreading this false impression. In this condition, it is not surprising that common and important features of speech like intonation and disfluency do not get captured by this intent. This tutorial aims to provide an appreciation of the "full-stack" of speech - aural, vision and the textual (or social construct) parts with a special emphasis on aspects that may have significance for current and future research.
Bio: Rajiv Ratn Shah currently works as an Assistant Professor in the Department of Computer Science and Engineering (joint appointment with the Department of Human-centered Design) at IIIT-Delhi. He is also the director of MIDAS lab at IIIT Delhi. He received his Ph.D. in computer science from the National University of Singapore, Singapore. Before joining IIIT-Delhi, he worked as a Research Fellow in Living Analytics Research Center (LARC) at the Singapore Management University, Singapore. Prior to completing his Ph.D., he received his M.Tech. and M.C.A. degrees in Computer Applications from the Delhi Technological University, Delhi and Jawaharlal Nehru University, Delhi, respectively. He has also received his B.Sc. in Mathematics (Honors) from the Banaras Hindu University, Varanasi. Dr. Shah is the recipient of several awards, including the prestigious Heidelberg Laureate Forum (HLF) and European Research Consortium for Informatics and Mathematics (ERCIM) fellowships. He is involved in organizing and reviewing of many top-tier international conferences and journals. His research interests include multimedia content processing, natural language processing, image processing, speech processing, multimodal computing, data science, and social media computing.
Yaman Kumar is a PhD student at IIIT Delhi and University of Buffalo, USA. Recently, he has received a Google PhD fellowship. Prior to joining his PhD program, Yaman had done his BTech from NSIT, Delhi and was working with Adobe systems. He is also the recipient of several awards at top conferences such as the best student paper award at AAAI 2019.