Matrix and Tensor Decompositions in Signal Processing. Gérard Favier. Читать онлайн. Newlib. NEWLIB.NET

Автор: Gérard Favier
Издательство: John Wiley & Sons Limited
Серия:
Жанр произведения: Программы
Год издания: 0
isbn: 9781119700968
Скачать книгу
347

      365 348

      366 349

      367 350

      368  351

      369  352

      370  353

      371  354

      372  355

      373  356

      374  357

       Matrices and Tensors with Signal Processing Set

      coordinated by

      Gérard Favier

      Volume 2

      Matrix and Tensor Decompositions in Signal Processing

      Gérard Favier

      First published 2021 in Great Britain and the United States by ISTE Ltd and John Wiley & Sons, Inc.

      Apart from any fair dealing for the purposes of research or private study, or criticism or review, as permitted under the Copyright, Designs and Patents Act 1988, this publication may only be reproduced, stored or transmitted, in any form or by any means, with the prior permission in writing of the publishers, or in the case of reprographic reproduction in accordance with the terms and licenses issued by the CLA. Enquiries concerning reproduction outside these terms should be sent to the publishers at the undermentioned address:

      ISTE Ltd

      27-37 St George’s Road

      London SW19 4EU

      UK

       www.iste.co.uk

      John Wiley & Sons, Inc.

      111 River Street

      Hoboken, NJ 07030

      USA

       www.wiley.com

      © ISTE Ltd 2021

      The rights of Gérard Favier to be identified as the author of this work have been asserted by him in accordance with the Copyright, Designs and Patents Act 1988.

      Library of Congress Control Number: 2021938218

      British Library Cataloguing-in-Publication Data

      A CIP record for this book is available from the British Library

      ISBN 978-1-78630-155-0

      Introduction

      The first book of this series was dedicated to introducing matrices and tensors (of order greater than two) from the perspective of their algebraic structure, presenting their similarities, differences and connections with representations of linear, bilinear and multilinear mappings. This second volume will now study tensor operations and decompositions in greater depth.

      In this introduction, we will motivate the use of tensors by answering five questions that prospective users might and should ask:

       – What are the advantages of tensor approaches?

       – For what uses?

       – In what fields of application?

       – With what tensor decompositions?

       – With what cost functions and optimization algorithms?

      Although our answers are necessarily incomplete, our aim is to:

       – present the advantages of tensor approaches over matrix approaches;

       – show a few examples of how tensor tools can be used;

       – give an overview of the extensive diversity of problems that can be solved using tensors, including a few example applications;

       – introduce the three most widely used tensor decompositions, presenting some of their properties and comparing their parametric complexity;

       – state a few problems based on tensor models in terms of the cost functions to be optimized;

       – describe various types of tensor-based processing, with a brief glimpse of the optimization methods that can be used.

      In most applications, a tensor χ of order N is viewed as an array of real or complex numbers. The current element of the tensor is denoted xi1,… ,iN, where each index

is associated with the nth mode, and In is its dimension, i.e. the number of elements for the nth mode. The order of the tensor is the number N of indices, i.e. the number of modes. Tensors are written with calligraphic letters1. An Nth-order tensor with entries
is written
where
= ℝ or ℂ, depending on whether the tensor is real-valued or complex-valued, and I1 × · · · × IN represents the size of χ.

      In general, a mode (also called a way) can have one of the following interpretations: (i) as a source of information (user, patient, client, trial, etc.); (ii) as a type of entity attached to the data (items/products, types of music, types of film, etc.); (iii) as a tag that characterizes an item, a piece of music, a film, etc.; (iv) as a recording modality that captures diversity in various domains (space, time, frequency, wavelength, polarization, color, etc.). Thus, a digital image in color can be represented as a three-dimensional tensor (of pixels) with two spatial modes, one for the rows (width) and one for the columns (height), and one channel mode (RGB colors). For example, a color image can be represented as a tensor of size 1024 × 768 × 3, where the third mode corresponds to the intensity of the three RGB colors (red, green, blue). For a volumetric image, there are three spatial modes (width × height × depth), and the points of the image are called voxels. In the context of hyperspectral imagery, in addition to the two spatial dimensions, there is a third dimension corresponding to the emission wavelength within a spectral band.

      Tensor approaches benefit from the following advantages over matrix approaches:

       – the essential uniqueness property2, satisfied by some tensor decompositions, such as PARAFAC (parallel factors) (Harshman 1970) under certain mild conditions; for matrix decompositions, this property requires certain restrictive conditions on the factor matrices, such as orthogonality, non-negativity, or a specific structure (triangular, Vandermonde, Toeplitz, etc.);

       – the ability to solve certain problems, such as the identification of communication channels, directly from measured signals, without requiring the calculation of high-order statistics of these signals or the use of long pilot sequences. The resulting deterministic and semi-blind processings can be performed with signal recordings that are shorter than those required by statistical methods, based on the estimation of high-order