Mathematics

Is this a compressed sensing application? - Intrim Report

Speaker: 
Abhishek Agarwal
Dimitrios Karslidis
Byong Kwon
Shant Mahserejian
Kevin Palmowski
Xuping Xie
Date: 
Mon, Aug 10, 2015
Location: 
Institute for Mathematics and its Applications
Conference: 
PIMS-IMA Math Modeling in Industry XIX
Abstract: 

Cyber Optics designs and manufacturers some of the most capable 3D measurement systems in the world. These systems are based on Phase Profilometry, and sample a large amount of data, to construct a single 3D Height Image. The ratio of data collected to data output can be as high as 30 to 1. This process of sensing massive amounts of data and outputting a small fraction of the sensed data, is a common problem in sensor design. In the past decade a new branch of mathematics and sensor design has emerged called compressed sensing that specifically addresses this problem. In compressed sensing the sensor designer exploits these massive compressive ratios to “sense” a more limited data set.

My team will determine the feasibility of compressed sensing as applied to the Cyber Optics Phase Profilometry sensor. I (the presenter) have no experience with compressed sensing, but I have 4 years of experience working with phase profilometry, and 20+ years working as a mathematician in industry. We will explore the literature of compressed sensing, and present a “go – no-go” analysis of phase profilometry to my management. That is, we will attempt to address the question, “can compressed sensing be used to reduce the data collection requirements of our sensors by at least an order of magnitude”?

Approximate Mapping of Temperatures from Coarser to Finer Grid using Temporal Derivatives - Intrim Report

Speaker: 
Ilona Ambartsumyan
Cuiyu He
Eldar Khattatov
Sewoong Kim
Lidia Mrad
Minho Song
Date: 
Mon, Aug 10, 2015
Location: 
Institute for Mathematics and its Applications
Conference: 
PIMS-IMA Math Modeling in Industry XIX
Abstract: 

In many practical situations encountered in industries, there is incomplete knowledge of material properties, boundary conditions, and sources for a given material/manufacturing process. However, process monitors such as thermocouples are typically used to measure temperature evolution in certain locations to bridge the resulting gaps. Spatial gradients of temperature are needed to predict required quantities such as internal stresses developed during the process. The temperature measurements are typically performed on a coarse grid. Computation of stresses needs temperatures on a much finer grid for a more precise estimation of spatial gradients. Usually bilinear and/or weighted interpolation techniques are used to improve the spatial resolution of temperatures. However, in the cases where there are strong exothermic and/or endothermic reactions occurring during the process, such interpolation techniques are error-prone. Using more thermocouples to measure temperature on finer grid would be an easy solution. However, such measurement is intrusive as well as perturbing in addition to increasing the cost of data acquisition. The mapping of temperatures from coarser grid to finer grid is an ill-posed problem. In addition to the spatial distribution temperatures, the thermocouple measurements also contain valuable temporal information in the form of derivatives. This additional information of temporal derivatives helps in improving the conditioning of the apparently ill-posed problem. The objective of this exercise is to develop an algorithm/procedure for mapping temperatures from coarse grid to a finer grid recognizing as well as the valuable temporal information in the corresponding derivatives.

Deducing Rock Properties from Spectral Seismic Data

Speaker: 
Jiajun Han
Date: 
Wed, Aug 5, 2015
Location: 
Institute for Mathematics and its Applications
Conference: 
PIMS-IMA Math Modeling in Industry XIX
Abstract: 

Seismic Data in Exploration Geoscience

The recovery and production of hydrocarbon resources begins with an exploration of the earth’s subsurface, often through the use of seismic data collection and analysis. In a typical seismic data survey, a series of seismic sources (e.g. dynamite explosions) are initiated on the surface of the earth. These create vibrational waves that travel into the earth, bounce off geological structures in the subsurface, and reflect back to the surface where the vibrations are recorded as data on geophones. Computer analysis of the recorded data can produce highly accurate images of these geological structures which can indicate the presence of reservoirs that could contain hydrocarbon fluids. High quality images with an accurate analysis by a team of geoscientists can lead to the successful discovery of valuable oil and gas resources. Spectral analysis of the seismic data may reveal additional information beyond the geological image. For instance, selective attenuation of various seismic frequencies is a result of varying rock properties, such as density, elasticity, porosity, pore size, or fluid content. In principle this information is present in the raw data, and the challenge is to find effective algorithms to reveal these rock properties.

Spectral Analysis

Through the Fourier transform, frequency content of a seismic signal can be observed. The Short Time Fourier transform is an example of a time-frequency method that decomposes a signal into individual frequency bands that evolve over time. Such time-frequency methods have been successfully used to analyze complex signals with rich frequency content, including recordings of music, animal sounds, radio-telescope data, amongst others. These time-frequency methods show promise in extracting detailed information about seismic events, as shown in Figure 1, for instance.

Figure 1: Sample time-frequency analysis of a large seismic event (earthquake). From Hotovec, Prejean, Vidale, Gomberg, in J. of Volcanology and Geothermal Research, V. 259, 2013.

Problem Description

Are existing time-frequency analytical techniques effective in providing robust estimation of physical rock parameters that are important to a successful, economically viable identification of oil and gas resources? Can they accurate measure frequency-dependent energy attenuation, amplitude-versus-offset effects, or other physical phenomena which are a result of rock and fluid properties?

Using both synthetic and real seismic data, the goal is to evaluate the effectiveness of existing time-frequency methods such as Gabor and Stockwell transforms, discrete and continuous wavelet transforms, basis and matching pursuit, autoregressive methods, empirical mode decomposition, and others. Specifically, we would like to determine whether these methods can be utilized to extract rock parameters, and whether there are modifications that can make them particularly effective for seismic data.

The source data will include both land-based seismic surveys as well as subsurface microseismic event recordings, as examples of the breadth of data that is available for realistic analysis.

Figure 2: (a). Seismic data set from a sedimentary basin in Canada. The erosional surface and channels are highlighted by arrows. The same frequency attribute are extract from short time Fourier transform (b), continuous wavelet transform (c) and empirical mode decomposition (d).

Class: 

Deep Learning for Image Anomaly Detection

Speaker: 
Jesse Berwald
Date: 
Wed, Aug 5, 2015
Location: 
Institute for Mathematics and its Applications
Conference: 
PIMS-IMA Math Modeling in Industry XIX
Abstract: 

The machine learning community has witnessed significant advances recently in the realm of image recognition [1,2]. Advances in computing power – primarily through the use of GPUs – has enabled a resurgence of neural networks with far more layers than was previously possible. For instance, the winning team, GoogLeNet [1,3], at the ImageNet 2014 competition triumphed with a 43.9% mean average precision, while the previous year’s winner, University of Amsterdam, won with 22.6% mean average precision.

Neural networks mimic the neurons in the brain. As in the human brain, multiple layers of computational “neurons” are designed to react to a variety of stimuli. For instance, a typical scheme to construct a neural network could involve building a layer of neurons that detects edges in an image. An additional layer could then be added which would be trained (optimized) to detect larger regions or shapes. The combination of these two layers could then identify and separate different objects present in a photograph. Adding further layers would allow the network to use the shapes to decipher the types of objects recorded in the image.

Goal of this project

An issue facing industries that deal with large numbers of digital photographs, such as magazines and retailers, is photo accuracy. Nearly all photos used in such contexts undergo some amount of editing (“Photoshopping”). Given the volume of photographs, mistakes occur [4]. Many of these images fall within a very narrow scope. An example would be the images used within a specific category of apparel on a retailer’s website. Detecting anomalies automatically in such cases would enable retailers such as Target to filter out mistakes before they enter production. By training a modern deep convolution neural network [1,5] on a collection of correct images within a narrow category, we would like to construct a network which will learn to recognize well-edited images. This amounts to learning a distribution of correct images so that poorly-edited images may be flagged as anomalies or outliers.

Keywords: neural networks, deep learning, image processing, machine learning

Prerequisites: Programming experience in Python. Experience in a Linux environment is a plus.

Class: 

Fast and Somewhat Accurate Algorithms

Speaker: 
Chai Wah Wu
Date: 
Wed, Aug 5, 2015
Location: 
Institute for Mathematics and its Applications
Conference: 
PIMS-IMA Math Modeling in Industry XIX
Abstract: 

In applications such as image processing, computer vision or image compression, often times accuracy and precision are less important than processing speed as the input data is noisy and the decision making process is robust against minor perturbations. For instance, the human visual system (HVS) makes pattern recognition decisions even though the data is blurry, noisy or incomplete and lossy image compression is based on the premise that we cannot distinguish minor differences in images. In this project we study the tradeoff between accuracy and system complexity as measured by processing speed and hardware complexity.

Knowledge of linear algebra, computer science, and familiarity with software tools such as Matlab or Python is desirable. Familiarity with image processing algorithms is not required.

Fig. 1: error diffusion halftoning using Shiau-Fan error diffusion

Fig. 2: error diffusion halftoning using a single lookup table

References:
1. Wu, C. W., "Locally connected processor arrays for matrix multiplication and linear transforms," Proceedings of 2011 IEEE International Symposium on Circuits and Systems (ISCAS), pp.2169,2172, 15-18 May 2011.

2. Wu, C. W., Stanich, M., Li, H., Qiao, Y., Ernst, L., "Fast Error Diffusion and Digital Halftoning Algorithms Using Look-up Tables," Proceedings of NIP22: International Conference on Digital Printing Technologies, Denver, Colorado, pp. 240-243, September 2006.

Class: 

Disconnecting the G_2 Moduli Space

Author: 
Johannes Nordstrom
Date: 
Tue, Jul 7, 2015
Location: 
PIMS, University of British Columbia
Conference: 
PIMS Symposium on the Geometry and Topology of Manifolds
Abstract: 
Little is currently known about the global properties of the $G_2$ moduli space of a closed 7-manifold, ie the space of Riemannian metrics with holonomy $G_2$ modulo diffeomorphisms. A holonomy $G_2$ metric has an associated $G_2$-structure, and I will define a Z/48 valued homotopy invariant of a $G_2$-structure in terms of the signature and Euler characteristic of a Spin(7)-coboundary. I will describe examples of manifolds with holonomy $G_2$ metrics where the invariant is amenable to computation in terms of eta invariants, and which are candidates for having a disconnected moduli space. This is joint work in progress with Diarmuid Crowley and Sebastian Goette.
Class: 

Universal torsion, L^2-invariants, polytopes and the Thurston norm

Author: 
Wolfgang Lück
Date: 
Thu, Jul 2, 2015
Location: 
PIMS, University of British Columbia
Conference: 
PIMS Symposium on the Geometry and Topology of Manifolds
Abstract: 
We introduce universal torsion which is defined for $L^2$-acyclic manifolds with torsion free fundamental group and takes values in certain $K_1$-groups of a skew field associated to the integral group ring. It encompasses well-know invariants such as the Alexander polynomial and $L^2$-torsion. We discuss also twisted $L^2$-torsion and higher order Alexander polynomials which can also be derived from the universal invariant and assign certain polytopes to the universal torsion. This gives especially in dimension 3 interesting invariants which recover for instance the Thurston norm.
Class: 

Introduction to the Farrell-Jones Conjecture

Author: 
Wolfgang Lück
Date: 
Tue, Jul 7, 2015
Location: 
PIMS, University of British Columbia
Conference: 
PIMS Symposium on the Geometry and Topology of Manifolds
Abstract: 

The Farrell-Jones Conjecture identifies the algebraic K- and L-groups for group rings with certain equivariant homology groups. We will give some details of its formulation, its status and indicate some ideas of proofs for certain classes of groups. We will try to convince the audience about its significance by considering special cases and presenting the surprizing large range of its applications to prominent problems in topology, geometry, and group theory.

Class: 

Decision problems, curvature and topology

Author: 
Martin Birdson
Date: 
Tue, Jul 7, 2015
Location: 
PIMS, University of British Columbia
Conference: 
PIMS Symposium on the Geometry and Topology of Manifolds
Abstract: 

I shall discuss a range of problems in which groups mediate between topological/geometric constructions and algorithmic problems elsewhere in mathematics, with impact in both directions. I shall begin with a discussion of sphere recognition in different dimensions. I'll explain why there is no algorithm that can determine if a compact homology sphere of dimension 5 or more has a non-trivial finite-sheeted covering. I'll sketch how ideas coming from the study of CAT(0) cube complexes were used by Henry Wilton and me to settle isomorphism problems for profinite groups, and to settle a conjecture in combinatorics concerning the extension problem for sets of partial permutations.

Class: 

Pages