12th Indian Conference on Computer Vision, Graphics
and Image Processing
December 19 - 21, IIT Jodhpur, India
Department of Ophthalmology and the Department of Genetics, Genomics and Informatics,
University of Tennessee Health Science Center (UTHSC)
Title of the Talk: Mining ophthalmic data using Artificial Intelligence.
Date&Time: 19 December, 9:15 AM IST.
Session Chair: Arnav Bhavsar
Bio:Siamak Yousefi is Assistant Professor at the Department of Ophthalmology and the Department of Genetics, Genomics, and Informatics of the University of Tennessee Health Science Center (UTHSC) in Memphis. He received his PhD in Electrical Engineering from the University of Texas at Dallas in 2012 and completed two postdoctoral trainings at the University of California Los Angeles (UCLA) working on Brain Computer Interface (BCI) and University of California San Diego (UCSD) working on computational ophthalmology. He is the director of the Data Mining and Machine Learning (DM2L) laboratory at UTHSC. He has published more than 100 peer-reviewed journal articles, conference papers, and abstracts, with over 50 in applications of Artificial Intelligence (AI) in vision and ophthalmology. He has been an invited guest speaker, moderator, and co-organizer of numerous Ophthalmology venues including Association for Research in Vision and Ophthalmology (ARVO), The Glaucoma Foundation, Asia-Pacific Glaucoma Congress (APGC), International Society for Eye Research (ISER), and Iranian Society of Ophthalmology (IRSO). He has been a member of several National Institute of Health (NIH) grant review panels. He is an Editorial Board Member of the Translational Vision Science and Technology (TVST) journal.
His lab is working on developing deep learning, manifold learning, conventional machine learning, unsupervised machine learning, and statistical approaches to screen, diagnose, and monitor different ocular conditions such as glaucoma, macular degeneration, keratoconus, keratoplasty, and uveitis from imaging and visual field data.
Title of the Talk: Towards Neural Architecture Search: Challenges and Solutions
Date&Time: 20 December, 3:15 PM (IST).
Session Chair: Abhinav Dhall
In recent years, a large number of related algorithms for Neural Architecture Search (NAS) have emerged. They have made various improvements to the NAS algorithm, and the related research work is complicated and rich. In order to reduce the difficulty for beginners to conduct NAS-related research, in this tutorial, we will provide a new perspective: starting with an overview of the characteristics of the earliest NAS algorithms, summarizing the problems in these early NAS algorithms, and then giving solutions for subsequent related research work. In addition, we will conduct a detailed and comprehensive analysis, comparison and summary of these works. Finally, we will give possible future research directions.
Bio: Dr Xiaojun Chang is a Senior Lecturer at Faculty of Information Technology, Monash University Clayton Campus, Australia. He is also affiliated with Monash University Centre for Data Science. He is ARC Discovery Early Career Researcher Award (DECRA) Fellow between 2019-2021 (awarded in 2018). Before joining Monash, Dr Chang was a Postdoc Research Associate in School of Computer Science, Carnegie Mellon University, working with Prof. Alex Hauptmann. He has spent most of his time working on exploring multiple signals (visual, acoustic, textual) for automatic content analysis in unconstrained or surveillance videos. Dr Chang's system has achieved top performance in various international competitions, such as TRECVID MED, TRECVID SIN, and TRECVID AVS. Dr. Chang received his Ph.D. degree in Centre for Artificial Intelligence & Faculty of Engineering and Information Technology, University of Technology Sydney, under the supervision of Prof. Yi Yang. During his PhD study, he was sequentially mentored by Prof. Feiping Nie and Yaoliang Yu. His research focus in this period was mainly on developing machine learning algorithms and apply them to multimedia analysis and computer vision.
Title of the Talk: Trajectory Forecasting: Multi-object to Multi-camera
Date&Time: 21 December, 3:15 PM (IST).
Session Chair: Abhinav Dhall
Abstract: Predicting future events in videos is a core task in computer vision. Trajectory forecasting is the problem of predicting future locations of an object in a video with wide applications in surveillance, autonomous vehicle navigation and mobile robotics. Pedestrians are a particularly challenging class of objects to predict, as they exhibit highly dynamic motion and may change speed or direction rapidly. This talk will focus on our recent work on pedestrian trajectory forecasting in videos introducing new perspectives to this popular task. In contrast to existing works which primarily consider a birds-eye perspective, we formulate the problem from an object-level perspective and call for the prediction of full object bounding boxes, rather than trajectories alone. Next, we introduce the task of multi-camera trajectory forecasting, where the future trajectory of an object is predicted in a network of cameras. The talk will discuss relevant literature, databases and state-of-the-art in the area.
Bio: Tanaya Guha is an Assistant Professor of Computer Science, University of Warwick, UK, where she is a member of the Warwick Machine Learning Group. Prior to joining Warwick, she was an Assistant Professor in IIT Kanpur and a Postdoctoral Researcher in University of Southern California. She holds a PhD in Electrical & Computer Engineering from the University of British Columbia (UBC), Vancouver, Canada. Her research focuses on building machine intelligence capabilities to understand, recognize, and predict human behaviour combining machine learning and signal processing. She regularly serves in the Program Committees of INTERSPEECH, ICME, ACM MM and ACII. She was a recipient of Warwick Global Research Priority award, ICME'20 Outstanding Area Chair award, and has won prestigious scholarships from Mensa Canada, Amazon and Google at the doctoral level.
Title of the Talk: Overview of Multimodal Embeddings and its Application in Vision-Language Tasks and Beyond
Date&Time: 19 December, 3:15 PM.
Session Chair: Abhinav Dhall
Abstract: We have witnessed significant progress in learning tasks involving vision and language such as Visual Question Answering and Phrase Grounding, in the last few years. This growth has been fueled by advances in deep learning in vision and language domains and availability of large scale multimodal data. In this talk I will focus on building learning models that can jointly understand and reason about multiple modalities. I will begin this talk with an introduction to multimodal embeddings, where the key idea is to align the two modalities by embedding them in a common vector space. This alignment ensures that similar entities in both modalities are closer e.g. word "cow" and images of concept "cow". I will then discuss some of our past work on using these embeddings to solve problems such as such as zero-shot recognition and phrase grounding. I will then cover application of multimodal embeddings to tasks other than vision-language such as social media analysis, visual localization, etc. I will then briefly discuss recent works on multimodal transformer that have shown large performance improvements in vision-language tasks by relying on the ability of transformers to learn strong correlations through multi-head attention, large capacity, large scale data and pre-training strategies. I hope this talk will provide a basic introduction to related concepts and encourage researchers to work in this field.
Bio: Available at http://ksikka.com/tutorial_icgvip20.html
Title: MRI diffusion Brain imaging of of the human Connectome and clinical assessment of brain connectivity disorders and neurosurgical planning.
Date&Time: 20 December, 9:00 AM (IST).
Session Chair: Arnav Bhavsar
Dr Walter Schneider, Professor of Psychology, Neurosurgery, Radiology & Bioengineering at the University of Pittsburgh & Medical Center and Senior Scientist at the Learning Research and Development Center. His research includes basic and actionable neuroscience based on diffusion imaging of white matter fiber tracts with High Definition Fiber Tracking (HDFT) and behavioral assessment. HDFT technology is now being used in neurosurgery for both presurgical planning and operating room real time surgical guidance. HDFT is used in diagnostic assessment of Traumatic Brain Injury (TBI) for visualizing and quantifying fiber breaks where other MRI imaging methods could not. He uses HDFT and MRI to localize tasks that can be used in targeted cognitive therapy to regrow damaged tissue. He has over 200 publications and published the 4th and 9th most cited papers in the history of psychology with over 50,000 citations; first functional neuroimaging paper in Nature helping to spark the modern era of brain imaging, developed a major model of brain executive and control systems (top downloaded paper in Cognitive Science 2003), and received the 2010 Editor's choice award for best imaging methods paper from NeuroImage. His group has developed brain tractographic imaging for mapping the brain Connectome, co-developed E-Prime software used by over 10,000 laboratories in 58 countries, he developed the Integrated Functional Imaging Systems (IFIS) (now sold by Phillips) that has been installed by over 150 brain imaging centers around the world. His technology was the basis of the Pittsburgh based Psychology Software Tools Inc. spinoff company that employs forty people in high technology jobs in Pittsburgh. He develops advanced technology for MRI based imaging, patient assessment, data visualization, mobile computing, and physical MRI phantom engineering. His recent work in diffusion fiber tracking identifies brain networks, quantifies tract integrity, and maps brain areas. His technology is used in clinical neurosurgery and TBI assessments on over a hundred patients per year and has produce improved medical outcomes and helped patients to understand and better deal with their brain pathology and rehabilitation impact. He leads a program to do TBI imaging across 7 university and 8 DoD/VA hospital systems. His work was highlighted by First Lady Obama as the most promising new technology for returning TBI war wounded and has appeared in major media reports including 60 Minutes, Discovery Channel, Scientific American, U.S. Medicine as well as traditional news media including AP, CNN, and Fox news. He is committed to doing collaborative international programs to advance basic and medical brain imaging.
Sudhir Pathak's Bio: Dr. Sudhir Pathak's research involves the development and designing of computational and mathematical models for use in diffusion MRI reconstruction, as well as the development of anisotropic metrics derived from diffusion models and their application in fiber tractography for use in biological tissue and textile phantoms. He is part of a project that uses a textile-based hollow fiber phantom to validate diffusion models. We are using different specialized textiles to mimic axonal features e.g. Axonal diameter, packing density, and crossing angle to validate micro-structural models. He has proposed a novel diffusion metric that can relate to the number of axons. These validation techniques and computational and mathematical models are further used to identify TBI lesions and understanding brain connectivity. He also proposed a new reconstruction algorithm that can be used to parse the geometrical information of both phantom and biological tissues. It combines diffusion spectrum images with popular constrained spherical deconvolution techniques to estimate underlying fiber crossing. This method can further be used for the segmentation of multi-tissue compartments. At the University of Pittsburgh, he is a the computational and mathematical lead in the development and designing of the High Definition Fiber Tracking (HDFT) pipeline. HDFT is used on a daily basis in TBI and other Neuro-surgical projects at the University of Pittsburgh. He is also adding the metrics described above into the HDFT pipeline to further improve the identification of lesions in patient populations with Huntington's disease and Amyotrophic lateral sclerosis. He is also part of an NIH funded grant related to Aphasia.