The document discusses tensors and their applications in data science. It describes how tensors can be used to efficiently represent large collections of documents by reducing their dimensionality through techniques like sparse matrix representation and singular value decomposition. This achieves significant data compression. The document also provides a brief history of tensors, noting they were first introduced in 1898 to study the properties of crystals, building on earlier work using tensors to study manifolds.
Mathematics was invented by humans to describe patterns and quantities in the real world. Some key points:
- Early humans developed counting as a practical tool for tasks like tracking food supplies and trade goods. Counting led to the development of basic arithmetic operations and the first written number systems.
- Properties of numbers, geometry, algebra, calculus, etc. were conceptualized by mathematicians over thousands of years through observing patterns and designing logical systems to model physical phenomena. Different cultures developed unique systems for writing and representing numbers.
- While mathematics describes inherent patterns in nature, the specific symbols, notations, definitions, and branches we use today are all human constructs. The rules and structures of mathematics have evolved significantly over the course of history
Educational innovation in university levels: Crystallography and X - ray diffraction .
A set of advanced seminars ( Bachelor and Master ) on the characterization of materials is offered . The information is accessible ( Spanish and English) both online and offline, from computers or mobile devices. In this way it is intended among other things to encourage independent learning of students.
x1 t10 04 maximum & minimum problems (13)Nigel Simmons
Maximum/minimum problems involve finding the vertex of a quadratic function. It is important to read the question carefully to determine if you need to find the x-value or y-value of the vertex. To find the maximum/minimum, you set the derivative of the function equal to 0 and solve for the x-value of the vertex, then substitute this back into the original function to find the y-value at the vertex.
Calculo de centros de masas de varias figurasJaimeCruz978742
1) The document discusses finding the center of mass of a solid of revolution by approximating it as a discrete one-dimensional problem using thin slices.
2) Each slice is treated as a bead with mass proportional to its volume, using the solid's uniform density.
3) Taking the limit of this approximation as the number of slices goes to infinity yields an integral expression for the x-coordinate of the center of mass in terms of the solid's cross-sectional functions.
Fractals are geometric shapes that exhibit self-similarity and complex patterns at every scale. Koch's snowflake is a famous fractal where the perimeter tends towards infinity as more iterations are done, even as the area approaches a limit. The Mandelbrot set is another well-known fractal that maps the behavior of values of c under a complex iterative function, resulting in diverse patterns when zooming in. Fractals are found throughout nature in shapes like clouds, coastlines, and Romanesco broccoli. They can also be generated through computer programs and used in applications like diagnosing skin cancer.
These are the slides from a talk that I gave at Durham University in 2013. Tatami coverings (also known as tatami tilings) are a restricted exact covering by monomino 1x1 tiles and domino 2x2 tiles of grid-regions.
The restriction is that no four tiles can meet.
I discovered and studied the resulting structure during my PhD.
1) The document discusses whether dice rolls and other mechanical randomizers can truly produce random outcomes from a dynamics perspective.
2) It analyzes the equations of motion for different dice shapes and coin tossing, showing that outcomes are theoretically predictable if initial conditions can be reproduced precisely.
3) However, in reality small uncertainties in initial conditions mean mechanical randomizers can approximate random processes, even if they are deterministic based on their underlying dynamics.
Mathematics was invented by humans to describe patterns and quantities in the real world. Some key points:
- Early humans developed counting as a practical tool for tasks like tracking food supplies and trade goods. Counting led to the development of basic arithmetic operations and the first written number systems.
- Properties of numbers, geometry, algebra, calculus, etc. were conceptualized by mathematicians over thousands of years through observing patterns and designing logical systems to model physical phenomena. Different cultures developed unique systems for writing and representing numbers.
- While mathematics describes inherent patterns in nature, the specific symbols, notations, definitions, and branches we use today are all human constructs. The rules and structures of mathematics have evolved significantly over the course of history
Educational innovation in university levels: Crystallography and X - ray diffraction .
A set of advanced seminars ( Bachelor and Master ) on the characterization of materials is offered . The information is accessible ( Spanish and English) both online and offline, from computers or mobile devices. In this way it is intended among other things to encourage independent learning of students.
x1 t10 04 maximum & minimum problems (13)Nigel Simmons
Maximum/minimum problems involve finding the vertex of a quadratic function. It is important to read the question carefully to determine if you need to find the x-value or y-value of the vertex. To find the maximum/minimum, you set the derivative of the function equal to 0 and solve for the x-value of the vertex, then substitute this back into the original function to find the y-value at the vertex.
Calculo de centros de masas de varias figurasJaimeCruz978742
1) The document discusses finding the center of mass of a solid of revolution by approximating it as a discrete one-dimensional problem using thin slices.
2) Each slice is treated as a bead with mass proportional to its volume, using the solid's uniform density.
3) Taking the limit of this approximation as the number of slices goes to infinity yields an integral expression for the x-coordinate of the center of mass in terms of the solid's cross-sectional functions.
Fractals are geometric shapes that exhibit self-similarity and complex patterns at every scale. Koch's snowflake is a famous fractal where the perimeter tends towards infinity as more iterations are done, even as the area approaches a limit. The Mandelbrot set is another well-known fractal that maps the behavior of values of c under a complex iterative function, resulting in diverse patterns when zooming in. Fractals are found throughout nature in shapes like clouds, coastlines, and Romanesco broccoli. They can also be generated through computer programs and used in applications like diagnosing skin cancer.
These are the slides from a talk that I gave at Durham University in 2013. Tatami coverings (also known as tatami tilings) are a restricted exact covering by monomino 1x1 tiles and domino 2x2 tiles of grid-regions.
The restriction is that no four tiles can meet.
I discovered and studied the resulting structure during my PhD.
1) The document discusses whether dice rolls and other mechanical randomizers can truly produce random outcomes from a dynamics perspective.
2) It analyzes the equations of motion for different dice shapes and coin tossing, showing that outcomes are theoretically predictable if initial conditions can be reproduced precisely.
3) However, in reality small uncertainties in initial conditions mean mechanical randomizers can approximate random processes, even if they are deterministic based on their underlying dynamics.
Fractals are irregular patterns that are self-similar across different scales. They are a branch of mathematics concerned with shapes found in nature that have non-integer dimensions. Some early contributors to fractal concepts included Leibiz, Cantor, and Hausdorff, but it was Mandelbrot who coined the term "fractal" and brought greater attention to the field. One of the most basic and important fractals is the Mandelbrot set, which is a set of complex numbers that fulfill certain recursive conditions. Fractals can also be classified based on their level of self-similarity, from exact to statistical self-similarity. Fractals are found throughout nature and have applications in fields like astrophysics
The document discusses the concept of parity in problem solving. It provides examples of problems involving parity, such as determining whether gears or dominoes can be arranged in certain ways. It also presents problems involving sums of odd and even numbers, and discusses how considering parity can help determine whether certain arrangements or solutions are possible. The document aims to illustrate how thinking about the parity of variables can help in solving mathematical problems.
The document provides an introduction to graph theory concepts. It defines graphs as pairs of vertices and edges, and introduces basic graph terminology like order, size, adjacency, and isomorphism. Graphs can be represented geometrically by drawing vertices as points and edges as lines between them. Both simple graphs and multigraphs are discussed.
This document provides an overview of calculus concepts including:
1) It introduces the four major concepts of calculus - limits, derivatives, definite integrals, and indefinite integrals.
2) It explains that the foundation of calculus is the concept of a limit, and covers limits by definition, delta-epsilon definition, graphs, numerical tables, and algebra.
3) It discusses differential calculus and the mathematics of change, including the derivative, related rates, and rules for finding derivatives.
4) It covers integral calculus including areas, volumes, the differences between indefinite and definite integrals, and methods for solving differential equations.
This document summarizes research on generalized Cantor sets and functions where the standard construction is modified. It introduces Cantor sets defined by an arbitrary base where the intervals removed at each stage are not all the same length. It also defines irregular or transcendental Cantor sets generated by transcendental numbers like e. The key findings are:
1) There exists a unique probability measure for generalized Cantor sets that generates the cumulative distribution function.
2) The Holder exponent of generalized Cantor sets is shown to be logn/s where n is the base and s is the number of subintervals.
3) Lower and upper densities are defined for the measure on generalized Cantor functions and their properties are
The Pythagorean theorem states that for any right triangle, the square of the hypotenuse is equal to the sum of the squares of the other two sides. The document provides an example proof of the theorem using geometric shapes and explains how to use the theorem to solve for unknown sides of right triangles. It also gives examples of using the Pythagorean theorem and taking square roots to solve for variables.
This document discusses collision detection in game development. It explains that collision detection is an important part of many video games and defines different types of collisions like point-box, box-box, etc. It focuses on point-box collision and provides a detailed explanation of how to code point-box collision detection to check if a bullet has collided with an enemy ship by comparing the position of the bullet and boundaries of the enemy. The document includes pseudocode for the collision detection algorithm.
The document discusses calculating the volumes of solids of revolution using Cavalieri's principle and the fundamental theorem of calculus. It provides an example of finding the volume of a solid generated by revolving a semi-circular base of radius r around its diameter, which is calculated to be 2r^3/3. The document also describes approximating a solid of revolution as cylindrical shells and calculating volume as 2πrhΔx.
Archimedes was an ancient Greek mathematician and inventor. Some of his key contributions included developing the law of buoyancy and hydrostatics, approximating pi, and inventing machines like the screw pump and compound pulley. He is considered one of the most brilliant mathematicians of antiquity.
This document contains slides from a presentation introducing the theory and applications of large deviations. It begins with an example using coin tosses to illustrate basic large deviation principles. It then discusses how large deviations can be used to study distributions like the binomial distribution. Applications discussed include risk management, information theory, and hydrodynamic limits in physics. Transformation techniques like Cramér's theorem and the contraction principle are also mentioned for applying large deviations to transformed sequences.
FRACTAL GEOMETRY AND ITS APPLICATIONS BY MILAN A JOSHIMILANJOSHIJI
This document provides an introduction to fractal geometry and its applications. It discusses how fractals were discovered by Mandelbrot as a way to describe irregular patterns in nature. Some key areas where fractals are applied that are mentioned include astronomy (galaxies and Saturn's rings), biology (bacteria cultures, plants), data compression, diffusion, economy, special effects in movies like Star Trek, weather patterns, antennas, and understanding global warming. The document also provides mathematical definitions of fractals and the Mandelbrot set, and describes how to graph the Mandelbrot set. Examples of fractals discussed in more depth include the Sierpinski triangle, Koch curve, broccoli, veins, and the Lorenz
This document provides an introduction to fractal geometry and its applications. It discusses how fractals were discovered by Mandelbrot as a way to describe irregular patterns in nature. Some key areas where fractals are applied that are mentioned include astronomy (galaxies and Saturn's rings), biology (bacteria cultures, plants), data compression, special effects in movies like Star Trek, weather patterns, heartbeats, antennas, and modeling global warming. The document also provides mathematical definitions of fractals and the Mandelbrot set, and describes how to generate fractal landscapes and the Mandelbrot set using iterated functions.
This document provides an introduction to fractal geometry and its applications. It discusses how fractals were discovered by Benoit Mandelbrot and have self-similar patterns seen at different scales. Examples of fractals in nature are given like coastlines, mountains, trees. The key fractals discussed are the Mandelbrot set and how it is defined mathematically. Applications of fractals mentioned include use in special effects in movies like Star Trek, modeling weather patterns, landscapes, and more. Areas where fractals are found include astronomy, biology, chemistry, art, music and others.
This document provides an introduction to fractal geometry and its applications. It discusses how fractals were discovered by Benoit Mandelbrot and have self-similar patterns seen at different scales. Examples of fractals in nature are given like coastlines, mountains, trees. The key fractals discussed are the Mandelbrot set and how it is defined mathematically. Applications of fractals mentioned include use in special effects in movies like Star Trek, modeling weather patterns, landscapes, human anatomy.
The document describes research into the maximum edge coloring problem, which involves coloring the edges of a graph such that each vertex sees at most two colors. The goal is to maximize the number of colors used. The problem is known to be NP-complete. The authors present a fixed-parameter tractable algorithm that runs in time O*(20k) by reducing the problem into smaller subproblems involving color palettes, vertex covers, and independent sets. They also discuss some open problems regarding improving the running time and determining whether the problem admits a polynomial kernel.
Dependent Types and Dynamics of Natural LanguageDaisuke BEKKI
The document discusses dependent type semantics (DTS) as a framework for natural language semantics. DTS takes a proof-theoretic approach and uses dependent types to provide unified treatments of anaphora and general inferences. The key aspects of DTS are that it uses dependent functions and products to represent anaphora and other context-dependent phenomena compositionally, while maintaining a correspondence to natural language syntax. Underspecified terms are used for lexical items to retrieve contexts during type checking and semantic composition. Examples show how DTS can provide representations of E-type and donkey anaphora through dependent types.
This document introduces genetic algorithms. It defines an algorithm and discusses time complexity analysis using Big O notation. It then provides examples of algorithms with different time complexities like O(n), O(n^2), O(log n), and O(n!). Genetic algorithms are introduced as a metaheuristic to solve NP-hard problems by mimicking biological evolution. The key concepts of genetic algorithms like encoding solutions, fitness functions, crossover and mutation operators are explained. An example of using genetic algorithms to solve the 8 queens problem is presented. Finally, advantages and disadvantages of genetic algorithms are summarized.
This document provides an introduction to data analysis techniques using Python. It discusses key Python libraries for data analysis like NumPy, Pandas, SciPy, Scikit-Learn and libraries for data visualization like matplotlib and Seaborn. It covers essential concepts in data analysis like Series, DataFrames and how to perform data cleaning, transformation, aggregation and visualization on data frames. It also discusses statistical analysis, machine learning techniques and how big data and data analytics can work together. The document is intended as an overview and hands-on guide to getting started with data analysis in Python.
More Related Content
Similar to Tensor models and other dreams by PhD Andres Mendez-Vazquez
Fractals are irregular patterns that are self-similar across different scales. They are a branch of mathematics concerned with shapes found in nature that have non-integer dimensions. Some early contributors to fractal concepts included Leibiz, Cantor, and Hausdorff, but it was Mandelbrot who coined the term "fractal" and brought greater attention to the field. One of the most basic and important fractals is the Mandelbrot set, which is a set of complex numbers that fulfill certain recursive conditions. Fractals can also be classified based on their level of self-similarity, from exact to statistical self-similarity. Fractals are found throughout nature and have applications in fields like astrophysics
The document discusses the concept of parity in problem solving. It provides examples of problems involving parity, such as determining whether gears or dominoes can be arranged in certain ways. It also presents problems involving sums of odd and even numbers, and discusses how considering parity can help determine whether certain arrangements or solutions are possible. The document aims to illustrate how thinking about the parity of variables can help in solving mathematical problems.
The document provides an introduction to graph theory concepts. It defines graphs as pairs of vertices and edges, and introduces basic graph terminology like order, size, adjacency, and isomorphism. Graphs can be represented geometrically by drawing vertices as points and edges as lines between them. Both simple graphs and multigraphs are discussed.
This document provides an overview of calculus concepts including:
1) It introduces the four major concepts of calculus - limits, derivatives, definite integrals, and indefinite integrals.
2) It explains that the foundation of calculus is the concept of a limit, and covers limits by definition, delta-epsilon definition, graphs, numerical tables, and algebra.
3) It discusses differential calculus and the mathematics of change, including the derivative, related rates, and rules for finding derivatives.
4) It covers integral calculus including areas, volumes, the differences between indefinite and definite integrals, and methods for solving differential equations.
This document summarizes research on generalized Cantor sets and functions where the standard construction is modified. It introduces Cantor sets defined by an arbitrary base where the intervals removed at each stage are not all the same length. It also defines irregular or transcendental Cantor sets generated by transcendental numbers like e. The key findings are:
1) There exists a unique probability measure for generalized Cantor sets that generates the cumulative distribution function.
2) The Holder exponent of generalized Cantor sets is shown to be logn/s where n is the base and s is the number of subintervals.
3) Lower and upper densities are defined for the measure on generalized Cantor functions and their properties are
The Pythagorean theorem states that for any right triangle, the square of the hypotenuse is equal to the sum of the squares of the other two sides. The document provides an example proof of the theorem using geometric shapes and explains how to use the theorem to solve for unknown sides of right triangles. It also gives examples of using the Pythagorean theorem and taking square roots to solve for variables.
This document discusses collision detection in game development. It explains that collision detection is an important part of many video games and defines different types of collisions like point-box, box-box, etc. It focuses on point-box collision and provides a detailed explanation of how to code point-box collision detection to check if a bullet has collided with an enemy ship by comparing the position of the bullet and boundaries of the enemy. The document includes pseudocode for the collision detection algorithm.
The document discusses calculating the volumes of solids of revolution using Cavalieri's principle and the fundamental theorem of calculus. It provides an example of finding the volume of a solid generated by revolving a semi-circular base of radius r around its diameter, which is calculated to be 2r^3/3. The document also describes approximating a solid of revolution as cylindrical shells and calculating volume as 2πrhΔx.
Archimedes was an ancient Greek mathematician and inventor. Some of his key contributions included developing the law of buoyancy and hydrostatics, approximating pi, and inventing machines like the screw pump and compound pulley. He is considered one of the most brilliant mathematicians of antiquity.
This document contains slides from a presentation introducing the theory and applications of large deviations. It begins with an example using coin tosses to illustrate basic large deviation principles. It then discusses how large deviations can be used to study distributions like the binomial distribution. Applications discussed include risk management, information theory, and hydrodynamic limits in physics. Transformation techniques like Cramér's theorem and the contraction principle are also mentioned for applying large deviations to transformed sequences.
FRACTAL GEOMETRY AND ITS APPLICATIONS BY MILAN A JOSHIMILANJOSHIJI
This document provides an introduction to fractal geometry and its applications. It discusses how fractals were discovered by Mandelbrot as a way to describe irregular patterns in nature. Some key areas where fractals are applied that are mentioned include astronomy (galaxies and Saturn's rings), biology (bacteria cultures, plants), data compression, diffusion, economy, special effects in movies like Star Trek, weather patterns, antennas, and understanding global warming. The document also provides mathematical definitions of fractals and the Mandelbrot set, and describes how to graph the Mandelbrot set. Examples of fractals discussed in more depth include the Sierpinski triangle, Koch curve, broccoli, veins, and the Lorenz
This document provides an introduction to fractal geometry and its applications. It discusses how fractals were discovered by Mandelbrot as a way to describe irregular patterns in nature. Some key areas where fractals are applied that are mentioned include astronomy (galaxies and Saturn's rings), biology (bacteria cultures, plants), data compression, special effects in movies like Star Trek, weather patterns, heartbeats, antennas, and modeling global warming. The document also provides mathematical definitions of fractals and the Mandelbrot set, and describes how to generate fractal landscapes and the Mandelbrot set using iterated functions.
This document provides an introduction to fractal geometry and its applications. It discusses how fractals were discovered by Benoit Mandelbrot and have self-similar patterns seen at different scales. Examples of fractals in nature are given like coastlines, mountains, trees. The key fractals discussed are the Mandelbrot set and how it is defined mathematically. Applications of fractals mentioned include use in special effects in movies like Star Trek, modeling weather patterns, landscapes, and more. Areas where fractals are found include astronomy, biology, chemistry, art, music and others.
This document provides an introduction to fractal geometry and its applications. It discusses how fractals were discovered by Benoit Mandelbrot and have self-similar patterns seen at different scales. Examples of fractals in nature are given like coastlines, mountains, trees. The key fractals discussed are the Mandelbrot set and how it is defined mathematically. Applications of fractals mentioned include use in special effects in movies like Star Trek, modeling weather patterns, landscapes, human anatomy.
The document describes research into the maximum edge coloring problem, which involves coloring the edges of a graph such that each vertex sees at most two colors. The goal is to maximize the number of colors used. The problem is known to be NP-complete. The authors present a fixed-parameter tractable algorithm that runs in time O*(20k) by reducing the problem into smaller subproblems involving color palettes, vertex covers, and independent sets. They also discuss some open problems regarding improving the running time and determining whether the problem admits a polynomial kernel.
Dependent Types and Dynamics of Natural LanguageDaisuke BEKKI
The document discusses dependent type semantics (DTS) as a framework for natural language semantics. DTS takes a proof-theoretic approach and uses dependent types to provide unified treatments of anaphora and general inferences. The key aspects of DTS are that it uses dependent functions and products to represent anaphora and other context-dependent phenomena compositionally, while maintaining a correspondence to natural language syntax. Underspecified terms are used for lexical items to retrieve contexts during type checking and semantic composition. Examples show how DTS can provide representations of E-type and donkey anaphora through dependent types.
This document introduces genetic algorithms. It defines an algorithm and discusses time complexity analysis using Big O notation. It then provides examples of algorithms with different time complexities like O(n), O(n^2), O(log n), and O(n!). Genetic algorithms are introduced as a metaheuristic to solve NP-hard problems by mimicking biological evolution. The key concepts of genetic algorithms like encoding solutions, fitness functions, crossover and mutation operators are explained. An example of using genetic algorithms to solve the 8 queens problem is presented. Finally, advantages and disadvantages of genetic algorithms are summarized.
This document provides an introduction to data analysis techniques using Python. It discusses key Python libraries for data analysis like NumPy, Pandas, SciPy, Scikit-Learn and libraries for data visualization like matplotlib and Seaborn. It covers essential concepts in data analysis like Series, DataFrames and how to perform data cleaning, transformation, aggregation and visualization on data frames. It also discusses statistical analysis, machine learning techniques and how big data and data analytics can work together. The document is intended as an overview and hands-on guide to getting started with data analysis in Python.
Apache Spark es un motor de cómputo unificado y conjunto de librerías para el procesamiento paralelo de datos de forma eficiente. Spark soporta múltiples lenguajes de programación y puede ejecutarse desde una laptop hasta en un gran cluster. La presentación introduce conceptos clave de Spark como transformaciones, acciones, RDDs, DataFrames y ejemplos básicos de su uso.
Metodos de kernel en machine learning by MC Luis Ricardo Peña LlamasDataLab Community
Este documento describe los métodos de kernel en machine learning. Explica cómo los kernels permiten clasificar datos no linealmente separables mapeando los datos a un espacio de características de dimensión más alta donde son linealmente separables. También resume brevemente la historia de los kernels y define formalmente qué es una función kernel válida de acuerdo con el teorema de Mercer.
Este documento discute el problema de la maldición de la dimensionalidad en machine learning. Explica que a medida que aumenta el número de variables, se hace más difícil encontrar el modelo óptimo que minimice el error. Luego resume métodos para reducir la dimensionalidad como selección de características, extracción de características y casos de éxito al aplicar estas técnicas. Finalmente, ofrece recomendaciones sobre cuándo usar reducción de dimensionalidad y qué algoritmos seleccionar dependiendo del conocimiento del problema y su dimensionalidad.
Nueva introducción de DataLab Community del 2017. Somos una comunidad abierta de Ciencia de Datos. Generamos colaboración entre profesionales y aprendices, compartiendo conocimientos, desarrollando habilidades y vinculando para impulsar la Ciencia de Datos.
El documento describe las diferentes profesiones relacionadas con la ciencia de datos, incluyendo analistas de datos, ingenieros de datos, visualizadores de datos, gerentes de datos y científicos de datos. Explica que los analistas de datos se enfocan en convertir los datos en información e información en conocimientos mediante modelos descriptivos y diagnósticos. Los ingenieros de datos se encargan de construir y mantener la infraestructura de datos. Los visualizadores de datos comunican los hallazgos de una manera comprensible. Los gerentes de datos impulsan
Presentación realizada en Campus Party 2016 sobre el Arte de la Ciencia de Datos. La presentación se divide en dos, por un lado está el tema de la comparativa con las artes liberales y por el otro lado está el arte de analizar datos.
DataLab Community genera colaboración entre profesionales y aprendices en Ciencia de Datos. Compartimos conocimiento y desarrollamos habilidades para impulsar la Ciencia de Datos en nuestra región.
Cómo fue que surgió lo que llamamos Big Data.
Varias perspectivas sobre qué es Data Science.
Qué estudia exactamente la Ciencia de Datos.
Introducción al Arte de la Ciencia de Datos.
Or: Beyond linear.
Abstract: Equivariant neural networks are neural networks that incorporate symmetries. The nonlinear activation functions in these networks result in interesting nonlinear equivariant maps between simple representations, and motivate the key player of this talk: piecewise linear representation theory.
Disclaimer: No one is perfect, so please mind that there might be mistakes and typos.
dtubbenhauer@gmail.com
Corrected slides: dtubbenhauer.com/talks.html
The binding of cosmological structures by massless topological defectsSérgio Sacani
Assuming spherical symmetry and weak field, it is shown that if one solves the Poisson equation or the Einstein field
equations sourced by a topological defect, i.e. a singularity of a very specific form, the result is a localized gravitational
field capable of driving flat rotation (i.e. Keplerian circular orbits at a constant speed for all radii) of test masses on a thin
spherical shell without any underlying mass. Moreover, a large-scale structure which exploits this solution by assembling
concentrically a number of such topological defects can establish a flat stellar or galactic rotation curve, and can also deflect
light in the same manner as an equipotential (isothermal) sphere. Thus, the need for dark matter or modified gravity theory is
mitigated, at least in part.
Immersive Learning That Works: Research Grounding and Paths ForwardLeonel Morgado
We will metaverse into the essence of immersive learning, into its three dimensions and conceptual models. This approach encompasses elements from teaching methodologies to social involvement, through organizational concerns and technologies. Challenging the perception of learning as knowledge transfer, we introduce a 'Uses, Practices & Strategies' model operationalized by the 'Immersive Learning Brain' and ‘Immersion Cube’ frameworks. This approach offers a comprehensive guide through the intricacies of immersive educational experiences and spotlighting research frontiers, along the immersion dimensions of system, narrative, and agency. Our discourse extends to stakeholders beyond the academic sphere, addressing the interests of technologists, instructional designers, and policymakers. We span various contexts, from formal education to organizational transformation to the new horizon of an AI-pervasive society. This keynote aims to unite the iLRN community in a collaborative journey towards a future where immersive learning research and practice coalesce, paving the way for innovative educational research and practice landscapes.
The ability to recreate computational results with minimal effort and actionable metrics provides a solid foundation for scientific research and software development. When people can replicate an analysis at the touch of a button using open-source software, open data, and methods to assess and compare proposals, it significantly eases verification of results, engagement with a diverse range of contributors, and progress. However, we have yet to fully achieve this; there are still many sociotechnical frictions.
Inspired by David Donoho's vision, this talk aims to revisit the three crucial pillars of frictionless reproducibility (data sharing, code sharing, and competitive challenges) with the perspective of deep software variability.
Our observation is that multiple layers — hardware, operating systems, third-party libraries, software versions, input data, compile-time options, and parameters — are subject to variability that exacerbates frictions but is also essential for achieving robust, generalizable results and fostering innovation. I will first review the literature, providing evidence of how the complex variability interactions across these layers affect qualitative and quantitative software properties, thereby complicating the reproduction and replication of scientific studies in various fields.
I will then present some software engineering and AI techniques that can support the strategic exploration of variability spaces. These include the use of abstractions and models (e.g., feature models), sampling strategies (e.g., uniform, random), cost-effective measurements (e.g., incremental build of software configurations), and dimensionality reduction methods (e.g., transfer learning, feature selection, software debloating).
I will finally argue that deep variability is both the problem and solution of frictionless reproducibility, calling the software science community to develop new methods and tools to manage variability and foster reproducibility in software systems.
Exposé invité Journées Nationales du GDR GPL 2024
Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...Travis Hills MN
Travis Hills of Minnesota developed a method to convert waste into high-value dry fertilizer, significantly enriching soil quality. By providing farmers with a valuable resource derived from waste, Travis Hills helps enhance farm profitability while promoting environmental stewardship. Travis Hills' sustainable practices lead to cost savings and increased revenue for farmers by improving resource efficiency and reducing waste.
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...Sérgio Sacani
Context. With a mass exceeding several 104 M⊙ and a rich and dense population of massive stars, supermassive young star clusters
represent the most massive star-forming environment that is dominated by the feedback from massive stars and gravitational interactions
among stars.
Aims. In this paper we present the Extended Westerlund 1 and 2 Open Clusters Survey (EWOCS) project, which aims to investigate
the influence of the starburst environment on the formation of stars and planets, and on the evolution of both low and high mass stars.
The primary targets of this project are Westerlund 1 and 2, the closest supermassive star clusters to the Sun.
Methods. The project is based primarily on recent observations conducted with the Chandra and JWST observatories. Specifically,
the Chandra survey of Westerlund 1 consists of 36 new ACIS-I observations, nearly co-pointed, for a total exposure time of 1 Msec.
Additionally, we included 8 archival Chandra/ACIS-S observations. This paper presents the resulting catalog of X-ray sources within
and around Westerlund 1. Sources were detected by combining various existing methods, and photon extraction and source validation
were carried out using the ACIS-Extract software.
Results. The EWOCS X-ray catalog comprises 5963 validated sources out of the 9420 initially provided to ACIS-Extract, reaching a
photon flux threshold of approximately 2 × 10−8 photons cm−2
s
−1
. The X-ray sources exhibit a highly concentrated spatial distribution,
with 1075 sources located within the central 1 arcmin. We have successfully detected X-ray emissions from 126 out of the 166 known
massive stars of the cluster, and we have collected over 71 000 photons from the magnetar CXO J164710.20-455217.
ESPP presentation to EU Waste Water Network, 4th June 2024 “EU policies driving nutrient removal and recycling
and the revised UWWTD (Urban Waste Water Treatment Directive)”
Phenomics assisted breeding in crop improvementIshaGoswami9
As the population is increasing and will reach about 9 billion upto 2050. Also due to climate change, it is difficult to meet the food requirement of such a large population. Facing the challenges presented by resource shortages, climate
change, and increasing global population, crop yield and quality need to be improved in a sustainable way over the coming decades. Genetic improvement by breeding is the best way to increase crop productivity. With the rapid progression of functional
genomics, an increasing number of crop genomes have been sequenced and dozens of genes influencing key agronomic traits have been identified. However, current genome sequence information has not been adequately exploited for understanding
the complex characteristics of multiple gene, owing to a lack of crop phenotypic data. Efficient, automatic, and accurate technologies and platforms that can capture phenotypic data that can
be linked to genomics information for crop improvement at all growth stages have become as important as genotyping. Thus,
high-throughput phenotyping has become the major bottleneck restricting crop breeding. Plant phenomics has been defined as the high-throughput, accurate acquisition and analysis of multi-dimensional phenotypes
during crop growing stages at the organism level, including the cell, tissue, organ, individual plant, plot, and field levels. With the rapid development of novel sensors, imaging technology,
and analysis methods, numerous infrastructure platforms have been developed for phenotyping.
When I was asked to give a companion lecture in support of ‘The Philosophy of Science’ (https://shorturl.at/4pUXz) I decided not to walk through the detail of the many methodologies in order of use. Instead, I chose to employ a long standing, and ongoing, scientific development as an exemplar. And so, I chose the ever evolving story of Thermodynamics as a scientific investigation at its best.
Conducted over a period of >200 years, Thermodynamics R&D, and application, benefitted from the highest levels of professionalism, collaboration, and technical thoroughness. New layers of application, methodology, and practice were made possible by the progressive advance of technology. In turn, this has seen measurement and modelling accuracy continually improved at a micro and macro level.
Perhaps most importantly, Thermodynamics rapidly became a primary tool in the advance of applied science/engineering/technology, spanning micro-tech, to aerospace and cosmology. I can think of no better a story to illustrate the breadth of scientific methodologies and applications at their best.
Tensor models and other dreams by PhD Andres Mendez-Vazquez
1. Tensor Models and Other Dreams...
Andres Mendez-Vazquez
January 26, 2018
1 / 64
2. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
2 / 64
3. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
3 / 64
4. Tensors are this way...
As words defining an important moment in life
Without you
All the stars we steal from the night sky
Will never be enough
Never be enough
These hands could hold the world
but it’ll
Never be enough...
- Justin Paul / Benj Pasek, Greatest Showman
4 / 64
5. Tensors are like such words...
They represent generalizations that represent our dreams...
In Data Sciences...
5 / 64
6. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
6 / 64
8. Then, we have an Opportunity or a Terrible Problem
How do you represent them in a easy way to handle them?
After all we want to
Search them
Compare them
Rank them
What about using vectors?
word 1 word 2 word 3 word 4
· · ·
word d
counter counter counter counter counter
x1 x2 x3 x4 · · · xd
8 / 64
9. Then, we have an Opportunity or a Terrible Problem
How do you represent them in a easy way to handle them?
After all we want to
Search them
Compare them
Rank them
What about using vectors?
word 1 word 2 word 3 word 4
· · ·
word d
counter counter counter counter counter
x1 x2 x3 x4 · · · xd
8 / 64
10. Then, we have an Opportunity or a Terrible Problem
How do you represent them in a easy way to handle them?
After all we want to
Search them
Compare them
Rank them
What about using vectors?
word 1 word 2 word 3 word 4
· · ·
word d
counter counter counter counter counter
x1 x2 x3 x4 · · · xd
8 / 64
11. Then, we have an Opportunity or a Terrible Problem
How do you represent them in a easy way to handle them?
After all we want to
Search them
Compare them
Rank them
What about using vectors?
word 1 word 2 word 3 word 4
· · ·
word d
counter counter counter counter counter
x1 x2 x3 x4 · · · xd
8 / 64
12. Then, we have an Opportunity or a Terrible Problem
How do you represent them in a easy way to handle them?
After all we want to
Search them
Compare them
Rank them
What about using vectors?
word 1 word 2 word 3 word 4
· · ·
word d
counter counter counter counter counter
x1 x2 x3 x4 · · · xd
8 / 64
13. The Matrix at the Center of Everything!!!
The Vector/Matrix Representation
They are basically a N × d matrix like this
A =
(x1)1 · · · (x1)j · · · (x1)d
...
...
(xi)1 (xi)j (xi)d
...
...
(xN )1 · · · (xN )j · · · (xN )d
A is a matrix with...
N represents the thousands of documents...
d represents the thousands of words in a dictionary.....
9 / 64
14. The Matrix at the Center of Everything!!!
The Vector/Matrix Representation
They are basically a N × d matrix like this
A =
(x1)1 · · · (x1)j · · · (x1)d
...
...
(xi)1 (xi)j (xi)d
...
...
(xN )1 · · · (xN )j · · · (xN )d
A is a matrix with...
N represents the thousands of documents...
d represents the thousands of words in a dictionary.....
9 / 64
15. A Small Problem
The matrix alone consumes... so much...
You have 2 bytes per memory cell
If we have N = 106
, d = 50, 000
We have
2 × N × d = 100 Gigabytes
10 / 64
16. A Small Problem
The matrix alone consumes... so much...
You have 2 bytes per memory cell
If we have N = 106
, d = 50, 000
We have
2 × N × d = 100 Gigabytes
10 / 64
18. We have a trick!!!
Something Notable
The Matrix is Highly SPARSE
12 / 64
19. Therefore
If you are smart enough
You start represent the matrix information using sparse techniques
5x5 Matrix
Numeric Elements
Empty Elements
Sparse Matrix
13 / 64
20. Then
If you are quite smart....
You discover that few of the eigenvalues provide some information...
Every Matrix has a Singular Value Decomposition
A = UΣV T
The columns of U are an orthonormal basis for the column space.
The columns of V are an orthonormal basis for the row space.
The Σ is diagonal and the entries on its diagonal σi = Σii are positive
real numbers, called the singular values of A.
14 / 64
21. Then
If you are quite smart....
You discover that few of the eigenvalues provide some information...
Every Matrix has a Singular Value Decomposition
A = UΣV T
The columns of U are an orthonormal basis for the column space.
The columns of V are an orthonormal basis for the row space.
The Σ is diagonal and the entries on its diagonal σi = Σii are positive
real numbers, called the singular values of A.
14 / 64
22. Then
If you are quite smart....
You discover that few of the eigenvalues provide some information...
Every Matrix has a Singular Value Decomposition
A = UΣV T
The columns of U are an orthonormal basis for the column space.
The columns of V are an orthonormal basis for the row space.
The Σ is diagonal and the entries on its diagonal σi = Σii are positive
real numbers, called the singular values of A.
14 / 64
23. Then
If you are quite smart....
You discover that few of the eigenvalues provide some information...
Every Matrix has a Singular Value Decomposition
A = UΣV T
The columns of U are an orthonormal basis for the column space.
The columns of V are an orthonormal basis for the row space.
The Σ is diagonal and the entries on its diagonal σi = Σii are positive
real numbers, called the singular values of A.
14 / 64
24. How much compression can we get?
The Matrix Sparse Representation
It Achieves 90% Compression - We go from 100 Gigabytes to 10
Gigabytes
From 50,000 dimensions/words we go to 300 dimensions
Using the Singular Value Decomposition
Making Possible to go from 100 Gigabytes to
2 × N × 300 = 0.6 Gigabytes
15 / 64
25. How much compression can we get?
The Matrix Sparse Representation
It Achieves 90% Compression - We go from 100 Gigabytes to 10
Gigabytes
From 50,000 dimensions/words we go to 300 dimensions
Using the Singular Value Decomposition
Making Possible to go from 100 Gigabytes to
2 × N × 300 = 0.6 Gigabytes
15 / 64
26. How much compression can we get?
The Matrix Sparse Representation
It Achieves 90% Compression - We go from 100 Gigabytes to 10
Gigabytes
From 50,000 dimensions/words we go to 300 dimensions
Using the Singular Value Decomposition
Making Possible to go from 100 Gigabytes to
2 × N × 300 = 0.6 Gigabytes
15 / 64
27. IMAGINE!!!!
We have a crazy moment!!!
All the stars we steal from the night sky
Will never be enough
Never be enough
Towers of gold are still too little
These hands could hold the world
but it’ll
Never be enough
Never be enough
For me
16 / 64
28. Then
You go ambitious!!! You add a new dimension representing feelings!!!
Feeling
Dim
ensionality
17 / 64
29. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
18 / 64
30. They have a somewhat short history!!!
First Most
They are abstract entities invariant under coordinate transformations.
They were mentioned first by Woldemar Wright in 1898
A German physicist, who taught at the Georg August University of
Göttingen.
He mentioned the tensors in a study about the physical properties of
crystals.
But Before That
The Great Riemann introduced the concept of topological manifold...
the beginning of the dream...
Through a quadratic linear element to study its properties...
ds2
= gijdxi
dxj
19 / 64
31. They have a somewhat short history!!!
First Most
They are abstract entities invariant under coordinate transformations.
They were mentioned first by Woldemar Wright in 1898
A German physicist, who taught at the Georg August University of
Göttingen.
He mentioned the tensors in a study about the physical properties of
crystals.
But Before That
The Great Riemann introduced the concept of topological manifold...
the beginning of the dream...
Through a quadratic linear element to study its properties...
ds2
= gijdxi
dxj
19 / 64
32. They have a somewhat short history!!!
First Most
They are abstract entities invariant under coordinate transformations.
They were mentioned first by Woldemar Wright in 1898
A German physicist, who taught at the Georg August University of
Göttingen.
He mentioned the tensors in a study about the physical properties of
crystals.
But Before That
The Great Riemann introduced the concept of topological manifold...
the beginning of the dream...
Through a quadratic linear element to study its properties...
ds2
= gijdxi
dxj
19 / 64
33. They have a somewhat short history!!!
First Most
They are abstract entities invariant under coordinate transformations.
They were mentioned first by Woldemar Wright in 1898
A German physicist, who taught at the Georg August University of
Göttingen.
He mentioned the tensors in a study about the physical properties of
crystals.
But Before That
The Great Riemann introduced the concept of topological manifold...
the beginning of the dream...
Through a quadratic linear element to study its properties...
ds2
= gijdxi
dxj
19 / 64
34. They have a somewhat short history!!!
First Most
They are abstract entities invariant under coordinate transformations.
They were mentioned first by Woldemar Wright in 1898
A German physicist, who taught at the Georg August University of
Göttingen.
He mentioned the tensors in a study about the physical properties of
crystals.
But Before That
The Great Riemann introduced the concept of topological manifold...
the beginning of the dream...
Through a quadratic linear element to study its properties...
ds2
= gijdxi
dxj
19 / 64
35. Then
Gregorio Ricci-Curbastro and Tullio Levi-Civita
They wrote a paper in the Mathematische Annalen , Vol. 54 (1901) ,
entitled "Méthodes de calcul differéntiel absolu"
A Monster Came Around
20 / 64
36. Then
Gregorio Ricci-Curbastro and Tullio Levi-Civita
They wrote a paper in the Mathematische Annalen , Vol. 54 (1901) ,
entitled "Méthodes de calcul differéntiel absolu"
A Monster Came Around
20 / 64
37. “Every Genius has stood in the Shoulder of Giants” -
Newton
Einstein adopted the concepts at the paper
And the Theory of General Relativity was born
He renamed the entire field from “calcul absolu”
TENSOR CALCULUS
21 / 64
38. “Every Genius has stood in the Shoulder of Giants” -
Newton
Einstein adopted the concepts at the paper
And the Theory of General Relativity was born
He renamed the entire field from “calcul absolu”
TENSOR CALCULUS
21 / 64
39. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
22 / 64
41. We define
A Coordinate System
We define vectors in terms of a base
v = vxe1 + vye2 =
vx
vy
∈ R2
v = v 2 = v2
x + v2
y
1
2
Note: This is important vectors are always the same thing no
matter the coordinate thing
24 / 64
42. Therefore
Imagine to represent the new basis in terms of an old basis
e1 · v = vx = e1 · vxe1 + e1 · vye2
e2 · v = vy = e2 · vxe1 + e2 · vye2
Where
ei · ej = Projection of ei onto ej
25 / 64
43. Therefore
Imagine to represent the new basis in terms of an old basis
e1 · v = vx = e1 · vxe1 + e1 · vye2
e2 · v = vy = e2 · vxe1 + e2 · vye2
Where
ei · ej = Projection of ei onto ej
25 / 64
44. Using a Little bit of Notation
We need a notation that is both more compact
Let the indices i, j represent the numbers 1, 2 corresponding to the
coordinates x, y
Write components of v as vi and v i in the two coordinate system
Then define
aij
= ei · ej
Note: This define the “ROTATION”
In fact are individually just the cosines of the angle
between one axis and another
26 / 64
45. Using a Little bit of Notation
We need a notation that is both more compact
Let the indices i, j represent the numbers 1, 2 corresponding to the
coordinates x, y
Write components of v as vi and v i in the two coordinate system
Then define
aij
= ei · ej
Note: This define the “ROTATION”
In fact are individually just the cosines of the angle
between one axis and another
26 / 64
46. Therefore
We can rewrite the entire transformation
v i
=
2
j=1
aij
vj
We will agree that whenever an index appears twice, we have a sum
v i
= aij
vj
27 / 64
47. Therefore
We can rewrite the entire transformation
v i
=
2
j=1
aij
vj
We will agree that whenever an index appears twice, we have a sum
v i
= aij
vj
27 / 64
48. We have then...
We can do the following
v 1
v 2 =
a11 a12
a21 a22
v1
v2
Then, we compress our notation more
v = av
28 / 64
49. We have then...
We can do the following
v 1
v 2 =
a11 a12
a21 a22
v1
v2
Then, we compress our notation more
v = av
28 / 64
50. Then, we can redefine our dot product
The Basis of Projecting into other vectors
v · w = vi
wi
= v
i
w
i
= aij
aik
vj
wk
Using the Kronecker Delta
δij
=
0 if i = j
1 if i = j
Therefore, we have
aij
aik
= δjk
29 / 64
51. Then, we can redefine our dot product
The Basis of Projecting into other vectors
v · w = vi
wi
= v
i
w
i
= aij
aik
vj
wk
Using the Kronecker Delta
δij
=
0 if i = j
1 if i = j
Therefore, we have
aij
aik
= δjk
29 / 64
52. Then, we can redefine our dot product
The Basis of Projecting into other vectors
v · w = vi
wi
= v
i
w
i
= aij
aik
vj
wk
Using the Kronecker Delta
δij
=
0 if i = j
1 if i = j
Therefore, we have
aij
aik
= δjk
29 / 64
53. Proving the Invariance of the dot product
Therefore
v
i
· w
i
= δjk
vj
wk
= vj
· wj
30 / 64
54. Then, we have
A scalar is a number K
It has the same value in different coordinate systems.
A vector is a set of numbers vi
They Transform according to
v i
= aij
vj
A (Second Rank) Tensor is a set of numbers Tij
They transform according to
T ij
= aik
ajl
Tkl
31 / 64
55. Then, we have
A scalar is a number K
It has the same value in different coordinate systems.
A vector is a set of numbers vi
They Transform according to
v i
= aij
vj
A (Second Rank) Tensor is a set of numbers Tij
They transform according to
T ij
= aik
ajl
Tkl
31 / 64
56. Then, we have
A scalar is a number K
It has the same value in different coordinate systems.
A vector is a set of numbers vi
They Transform according to
v i
= aij
vj
A (Second Rank) Tensor is a set of numbers Tij
They transform according to
T ij
= aik
ajl
Tkl
31 / 64
57. Then you can go higher
For Example, tensors in Rank 3
32 / 64
58. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
33 / 64
59. Once, we have an idea of Tensor
Do we have similar decompositions that the ones in SVD?
We have them......!!!
A Little Bit of History
Tensor decompositions originated with Hitchcock in 1927
An American mathematician and physicist known for his formulation of
the transportation problem in 1941.
A multiway model is attributed to Cattell in 1944
A British and American psychologist, known for his psychometric
research into intrapersonal psychological structure.
But it is until Ledyard R. Tucker
“Some mathematical notes on three-mode factor analysis,”
Psychometrika, 31 (1966), pp. 279–311.
34 / 64
60. Once, we have an idea of Tensor
Do we have similar decompositions that the ones in SVD?
We have them......!!!
A Little Bit of History
Tensor decompositions originated with Hitchcock in 1927
An American mathematician and physicist known for his formulation of
the transportation problem in 1941.
A multiway model is attributed to Cattell in 1944
A British and American psychologist, known for his psychometric
research into intrapersonal psychological structure.
But it is until Ledyard R. Tucker
“Some mathematical notes on three-mode factor analysis,”
Psychometrika, 31 (1966), pp. 279–311.
34 / 64
61. Once, we have an idea of Tensor
Do we have similar decompositions that the ones in SVD?
We have them......!!!
A Little Bit of History
Tensor decompositions originated with Hitchcock in 1927
An American mathematician and physicist known for his formulation of
the transportation problem in 1941.
A multiway model is attributed to Cattell in 1944
A British and American psychologist, known for his psychometric
research into intrapersonal psychological structure.
But it is until Ledyard R. Tucker
“Some mathematical notes on three-mode factor analysis,”
Psychometrika, 31 (1966), pp. 279–311.
34 / 64
62. Once, we have an idea of Tensor
Do we have similar decompositions that the ones in SVD?
We have them......!!!
A Little Bit of History
Tensor decompositions originated with Hitchcock in 1927
An American mathematician and physicist known for his formulation of
the transportation problem in 1941.
A multiway model is attributed to Cattell in 1944
A British and American psychologist, known for his psychometric
research into intrapersonal psychological structure.
But it is until Ledyard R. Tucker
“Some mathematical notes on three-mode factor analysis,”
Psychometrika, 31 (1966), pp. 279–311.
34 / 64
63. The Dream has been expanding beyond Physics
In the last ten years
1 Signal Processing
2 Numerical Linear Algebra
3 Computer Vision
4 Data Mining
5 Graph analysis
6 Neurosciences
7 etc
And we are going further
The Dream of Representation is at full speed when dealing with BIG
DATA!!!
35 / 64
64. The Dream has been expanding beyond Physics
In the last ten years
1 Signal Processing
2 Numerical Linear Algebra
3 Computer Vision
4 Data Mining
5 Graph analysis
6 Neurosciences
7 etc
And we are going further
The Dream of Representation is at full speed when dealing with BIG
DATA!!!
35 / 64
65. The Dream has been expanding beyond Physics
In the last ten years
1 Signal Processing
2 Numerical Linear Algebra
3 Computer Vision
4 Data Mining
5 Graph analysis
6 Neurosciences
7 etc
And we are going further
The Dream of Representation is at full speed when dealing with BIG
DATA!!!
35 / 64
66. The Dream has been expanding beyond Physics
In the last ten years
1 Signal Processing
2 Numerical Linear Algebra
3 Computer Vision
4 Data Mining
5 Graph analysis
6 Neurosciences
7 etc
And we are going further
The Dream of Representation is at full speed when dealing with BIG
DATA!!!
35 / 64
67. The Dream has been expanding beyond Physics
In the last ten years
1 Signal Processing
2 Numerical Linear Algebra
3 Computer Vision
4 Data Mining
5 Graph analysis
6 Neurosciences
7 etc
And we are going further
The Dream of Representation is at full speed when dealing with BIG
DATA!!!
35 / 64
68. The Dream has been expanding beyond Physics
In the last ten years
1 Signal Processing
2 Numerical Linear Algebra
3 Computer Vision
4 Data Mining
5 Graph analysis
6 Neurosciences
7 etc
And we are going further
The Dream of Representation is at full speed when dealing with BIG
DATA!!!
35 / 64
69. The Dream has been expanding beyond Physics
In the last ten years
1 Signal Processing
2 Numerical Linear Algebra
3 Computer Vision
4 Data Mining
5 Graph analysis
6 Neurosciences
7 etc
And we are going further
The Dream of Representation is at full speed when dealing with BIG
DATA!!!
35 / 64
70. The Dream has been expanding beyond Physics
In the last ten years
1 Signal Processing
2 Numerical Linear Algebra
3 Computer Vision
4 Data Mining
5 Graph analysis
6 Neurosciences
7 etc
And we are going further
The Dream of Representation is at full speed when dealing with BIG
DATA!!!
35 / 64
71. Decomposition of Tensors
Hitchcock Proposed such decomposition first... then the deluge
Name Proposed by
Polyadic form of a tensor Hitchcock, 1927
Three-mode Tucker 1966
factor analysis
PARAFAC (parallel factors) Harshman, 1970
CANDECOMP or CAND Carroll and Chang, 1970
(canonical decomposition)
Topographic components Möcks, 1988
model
CP (CANDECOMP/PARAFAC) Kiers, 2000
36 / 64
72. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
37 / 64
73. Look at the most modern on, 17 years ago...
The CP decomposition factorizes a tensor into a sum of component
rank-one tensors (Vectors!!!)
X ≈
R
r=1
ar ◦ br ◦ cr with X ∈ RI×J×K
Where
R is a positive integer
ar ∈ RI
br
∈ RJ
cr ∈ RK
38 / 64
74. Look at the most modern on, 17 years ago...
The CP decomposition factorizes a tensor into a sum of component
rank-one tensors (Vectors!!!)
X ≈
R
r=1
ar ◦ br ◦ cr with X ∈ RI×J×K
Where
R is a positive integer
ar ∈ RI
br
∈ RJ
cr ∈ RK
38 / 64
75. Then, Point Wise
We have the following
xijk =
R
r=1
airbjrccr
Graphically
39 / 64
76. Then, Point Wise
We have the following
xijk =
R
r=1
airbjrccr
Graphically
39 / 64
77. Therefore
The rank of a tensor X, rank(X)
It is defined as the smallest number of rank-one tensors that generate X
as their sum!!!
Problem!!!
The problem is NP-hard
But that has not stopped us because
We can use many of the methods in optimization to try to figure out
the magical number R!!!
From Approximation Techniques...
To Branch and Bound...
Even Naive techniques...
40 / 64
78. Therefore
The rank of a tensor X, rank(X)
It is defined as the smallest number of rank-one tensors that generate X
as their sum!!!
Problem!!!
The problem is NP-hard
But that has not stopped us because
We can use many of the methods in optimization to try to figure out
the magical number R!!!
From Approximation Techniques...
To Branch and Bound...
Even Naive techniques...
40 / 64
79. Therefore
The rank of a tensor X, rank(X)
It is defined as the smallest number of rank-one tensors that generate X
as their sum!!!
Problem!!!
The problem is NP-hard
But that has not stopped us because
We can use many of the methods in optimization to try to figure out
the magical number R!!!
From Approximation Techniques...
To Branch and Bound...
Even Naive techniques...
40 / 64
80. Why so much effort?
A Big Difference with SVD
It is never unique unless we have a orthogonality between the columns or
rows in the matrix.
We have then
That Tensors are way more general and less prone to problems!!!
41 / 64
81. Why so much effort?
A Big Difference with SVD
It is never unique unless we have a orthogonality between the columns or
rows in the matrix.
We have then
That Tensors are way more general and less prone to problems!!!
41 / 64
82. Now
We introduce a little bit of more notation
X ≈
R
r=1
ar ◦ br ◦ cr = A, B, C
CP Decompose the Tensor using the following Optimization
min
X
X − X
s.t. X =
R
r=1
λar ◦ br ◦ cr = λ; A, B, C
42 / 64
83. Now
We introduce a little bit of more notation
X ≈
R
r=1
ar ◦ br ◦ cr = A, B, C
CP Decompose the Tensor using the following Optimization
min
X
X − X
s.t. X =
R
r=1
λar ◦ br ◦ cr = λ; A, B, C
42 / 64
84. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
43 / 64
85. Here is why...
Here a simulation by direct numerical simulation
It can easily produce 100 GB to 1000 GB per DAY
The data came from (CIRCA 2016)
It a is called S3D, a massively parallel compressible reacting flow solver
developed at Sandia National Laboratories...
For example, data came from
1 Autoignitive premixture of air and ethanol in Homogeneous Charge
Compression Ignition (HCCI)
1 Each time step requires 111 MB of storage, and the entire dataset is 70
GB.
2 A temporally-evolving planar slot jet flame with DME (dimethyl
ether) as the fuel
1 Each time step requires 32 GB storage, so the entire dataset is 520 GB
44 / 64
86. Here is why...
Here a simulation by direct numerical simulation
It can easily produce 100 GB to 1000 GB per DAY
The data came from (CIRCA 2016)
It a is called S3D, a massively parallel compressible reacting flow solver
developed at Sandia National Laboratories...
For example, data came from
1 Autoignitive premixture of air and ethanol in Homogeneous Charge
Compression Ignition (HCCI)
1 Each time step requires 111 MB of storage, and the entire dataset is 70
GB.
2 A temporally-evolving planar slot jet flame with DME (dimethyl
ether) as the fuel
1 Each time step requires 32 GB storage, so the entire dataset is 520 GB
44 / 64
87. Here is why...
Here a simulation by direct numerical simulation
It can easily produce 100 GB to 1000 GB per DAY
The data came from (CIRCA 2016)
It a is called S3D, a massively parallel compressible reacting flow solver
developed at Sandia National Laboratories...
For example, data came from
1 Autoignitive premixture of air and ethanol in Homogeneous Charge
Compression Ignition (HCCI)
1 Each time step requires 111 MB of storage, and the entire dataset is 70
GB.
2 A temporally-evolving planar slot jet flame with DME (dimethyl
ether) as the fuel
1 Each time step requires 32 GB storage, so the entire dataset is 520 GB
44 / 64
88. Here is why...
Here a simulation by direct numerical simulation
It can easily produce 100 GB to 1000 GB per DAY
The data came from (CIRCA 2016)
It a is called S3D, a massively parallel compressible reacting flow solver
developed at Sandia National Laboratories...
For example, data came from
1 Autoignitive premixture of air and ethanol in Homogeneous Charge
Compression Ignition (HCCI)
1 Each time step requires 111 MB of storage, and the entire dataset is 70
GB.
2 A temporally-evolving planar slot jet flame with DME (dimethyl
ether) as the fuel
1 Each time step requires 32 GB storage, so the entire dataset is 520 GB
44 / 64
89. Even in Machines like
a Cray XC30 super- computer
5,576 dual-socket 12-core Intel “Ivy Bridge” (2.4 GHz) compute
nodes.
The peak flop rate of each core is 19.2 GFLOPS.
Each node has 64 GB of memory.
This machines will go down
Because the data representation is not efficient...
45 / 64
90. Even in Machines like
a Cray XC30 super- computer
5,576 dual-socket 12-core Intel “Ivy Bridge” (2.4 GHz) compute
nodes.
The peak flop rate of each core is 19.2 GFLOPS.
Each node has 64 GB of memory.
This machines will go down
Because the data representation is not efficient...
45 / 64
92. Furthermore...
We have that for 550 Gigabytes compression’s as
1 5 Times 100 Gigs
2 16 Times 34 Gigs
3 55 Times 10 Gig
4 etc
Improving Running times like crazy... from 3 seconds to 70 seconds
when processing 15 TB of data...
47 / 64
93. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
48 / 64
94. We have a huge problem in Deep Neural Networks
Modern Architectures
They are consuming from 89% to 100% of the memory at host GPU and
Machines
Depending on the place the calculations are done!!!
49 / 64
95. Problem with such Architectures
Recent studies show
The weight matrix of the fully-connected layer is highly redundant.
if you reduce the number of parameters, you could achieve
A similar predictive power
Possible making them less prone to over-fitting or under-fitting
50 / 64
96. Problem with such Architectures
Recent studies show
The weight matrix of the fully-connected layer is highly redundant.
if you reduce the number of parameters, you could achieve
A similar predictive power
Possible making them less prone to over-fitting or under-fitting
50 / 64
97. Thus
In the Paper
Novikov, A., Podoprikhin, D., Osokin, A. and Vetrov, D.P., 2015.
Tensorizing neural networks. In Advances in Neural Information
Processing Systems (pp. 442-450).
They Proposed the TT-Representation
Where in a d−dimensional array (Tensor) A
If for a each dimension k = 1, ..., d and each possible value of the kth
dimension index jk = 1, ..., nk
There exists a matrix Gk [jk] such that all the elements of A can be
computed as a product of matrices.
51 / 64
98. Thus
In the Paper
Novikov, A., Podoprikhin, D., Osokin, A. and Vetrov, D.P., 2015.
Tensorizing neural networks. In Advances in Neural Information
Processing Systems (pp. 442-450).
They Proposed the TT-Representation
Where in a d−dimensional array (Tensor) A
If for a each dimension k = 1, ..., d and each possible value of the kth
dimension index jk = 1, ..., nk
There exists a matrix Gk [jk] such that all the elements of A can be
computed as a product of matrices.
51 / 64
99. Thus
In the Paper
Novikov, A., Podoprikhin, D., Osokin, A. and Vetrov, D.P., 2015.
Tensorizing neural networks. In Advances in Neural Information
Processing Systems (pp. 442-450).
They Proposed the TT-Representation
Where in a d−dimensional array (Tensor) A
If for a each dimension k = 1, ..., d and each possible value of the kth
dimension index jk = 1, ..., nk
There exists a matrix Gk [jk] such that all the elements of A can be
computed as a product of matrices.
51 / 64
100. Then
The TT-Representation
A (j1, j2 · · · , jd) = G1 [j1] G2 [j2] · · · Gd [jd]
All matrices Gk [jk] related to the same dimension k are restricted to
be of the same size rk−1 × rk.
52 / 64
101. Here a problem, we do not have a unique representation
We then go for the lowest rank
A (j1, j2 · · · , jd) =
α0,...,αd
G1 [j1] (α0, α1) · · · Gd [jd] (αd−1, αd)
Where
Gk [jk] (αk−1, αk) represent the element of the matrix Gk [jk] at position
(α0, α1)
53 / 64
102. Here a problem, we do not have a unique representation
We then go for the lowest rank
A (j1, j2 · · · , jd) =
α0,...,αd
G1 [j1] (α0, α1) · · · Gd [jd] (αd−1, αd)
Where
Gk [jk] (αk−1, αk) represent the element of the matrix Gk [jk] at position
(α0, α1)
53 / 64
103. With Memory Usage
For full representation
d
k=1
nk
and the TT-Representation
d
k=1
nkrk−1rk
54 / 64
104. With Memory Usage
For full representation
d
k=1
nk
and the TT-Representation
d
k=1
nkrk−1rk
54 / 64
105. Then
They propose to store each layer in a TT-Representation W
Where W are the weight of a fully connected layer
Then, using our old back-propagation
y = Wx + b
With W ∈ RN×M and b ∈ RM
In TT-Representation
Y (i1, i2 · · · , id) =
j1,...,jd
G1 [i1, j1] ...Gd [id, jd] X (j1, j2 · · · , jd) + B (i1, i2 · · · , id)
55 / 64
106. Then
They propose to store each layer in a TT-Representation W
Where W are the weight of a fully connected layer
Then, using our old back-propagation
y = Wx + b
With W ∈ RN×M and b ∈ RM
In TT-Representation
Y (i1, i2 · · · , id) =
j1,...,jd
G1 [i1, j1] ...Gd [id, jd] X (j1, j2 · · · , jd) + B (i1, i2 · · · , id)
55 / 64
107. Then
They propose to store each layer in a TT-Representation W
Where W are the weight of a fully connected layer
Then, using our old back-propagation
y = Wx + b
With W ∈ RN×M and b ∈ RM
In TT-Representation
Y (i1, i2 · · · , id) =
j1,...,jd
G1 [i1, j1] ...Gd [id, jd] X (j1, j2 · · · , jd) + B (i1, i2 · · · , id)
55 / 64
108. This has the following complexity
The previous representation allows to handle a larger number of
parameters
Without too much overhead...
With the following complexities
Operation Time Memory
FC forward pass O(MN) O(MN)
TT forward pass O dr2m max {M, N} O dr2 max {M, N}
FC backward pass O(MN) O(MN)
TT backward pass O dr2m max {M, N} O dr3 max {M, N}
56 / 64
109. This has the following complexity
The previous representation allows to handle a larger number of
parameters
Without too much overhead...
With the following complexities
Operation Time Memory
FC forward pass O(MN) O(MN)
TT forward pass O dr2m max {M, N} O dr2 max {M, N}
FC backward pass O(MN) O(MN)
TT backward pass O dr2m max {M, N} O dr3 max {M, N}
56 / 64
110. Applications for this
Manage Better
The amount of memory being used in the devices
Increase the size of the Deep Networks
Although I have some thoughts about this...
Implement CNN Networks into mobile devices
Kim, Yong-Deok, Eunhyeok Park, Sungjoo Yoo, Taelim Choi, Lu
Yang, and Dongjun Shin. "Compression of deep convolutional neural
networks for fast and low power mobile applications." arXiv preprint
arXiv:1511.06530 (2015).
57 / 64
111. Applications for this
Manage Better
The amount of memory being used in the devices
Increase the size of the Deep Networks
Although I have some thoughts about this...
Implement CNN Networks into mobile devices
Kim, Yong-Deok, Eunhyeok Park, Sungjoo Yoo, Taelim Choi, Lu
Yang, and Dongjun Shin. "Compression of deep convolutional neural
networks for fast and low power mobile applications." arXiv preprint
arXiv:1511.06530 (2015).
57 / 64
112. Applications for this
Manage Better
The amount of memory being used in the devices
Increase the size of the Deep Networks
Although I have some thoughts about this...
Implement CNN Networks into mobile devices
Kim, Yong-Deok, Eunhyeok Park, Sungjoo Yoo, Taelim Choi, Lu
Yang, and Dongjun Shin. "Compression of deep convolutional neural
networks for fast and low power mobile applications." arXiv preprint
arXiv:1511.06530 (2015).
57 / 64
113. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
58 / 64
114. Given that
Something Notable
Sparse tensors appear in many large-scale applications with
multidimensional and sparse data.
What support do we have for such situations?
Liu, Bangtian, Chengyao Wen, Anand D. Sarwate, and Maryam Mehri
Dehnavi. "A Unified Optimization Approach for Sparse Tensor
Operations on GPUs." arXiv preprint arXiv:1705.09905 (2017).
59 / 64
115. Given that
Something Notable
Sparse tensors appear in many large-scale applications with
multidimensional and sparse data.
What support do we have for such situations?
Liu, Bangtian, Chengyao Wen, Anand D. Sarwate, and Maryam Mehri
Dehnavi. "A Unified Optimization Approach for Sparse Tensor
Operations on GPUs." arXiv preprint arXiv:1705.09905 (2017).
59 / 64
116. They pointed out different resources that you have around
Shared memory systems
The Tensor Toolbox [21], [4] and N-way Toolbox [22] are two widely
used MATLAB
The Cyclops Tensor Framework (CTF) is a C++ library which
provides automatic parallelization for sparse tensor operations.
etc
Distributed memory systems
Gigatensor handles tera-scale tensors using the MapReduce
framework.
Hypertensor is a sparse tensor library for SpMTTKRP on
distributed-memory environments.
etc
60 / 64
117. They pointed out different resources that you have around
Shared memory systems
The Tensor Toolbox [21], [4] and N-way Toolbox [22] are two widely
used MATLAB
The Cyclops Tensor Framework (CTF) is a C++ library which
provides automatic parallelization for sparse tensor operations.
etc
Distributed memory systems
Gigatensor handles tera-scale tensors using the MapReduce
framework.
Hypertensor is a sparse tensor library for SpMTTKRP on
distributed-memory environments.
etc
60 / 64
118. They pointed out different resources that you have around
Shared memory systems
The Tensor Toolbox [21], [4] and N-way Toolbox [22] are two widely
used MATLAB
The Cyclops Tensor Framework (CTF) is a C++ library which
provides automatic parallelization for sparse tensor operations.
etc
Distributed memory systems
Gigatensor handles tera-scale tensors using the MapReduce
framework.
Hypertensor is a sparse tensor library for SpMTTKRP on
distributed-memory environments.
etc
60 / 64
119. They pointed out different resources that you have around
Shared memory systems
The Tensor Toolbox [21], [4] and N-way Toolbox [22] are two widely
used MATLAB
The Cyclops Tensor Framework (CTF) is a C++ library which
provides automatic parallelization for sparse tensor operations.
etc
Distributed memory systems
Gigatensor handles tera-scale tensors using the MapReduce
framework.
Hypertensor is a sparse tensor library for SpMTTKRP on
distributed-memory environments.
etc
60 / 64
120. They pointed out different resources that you have around
Shared memory systems
The Tensor Toolbox [21], [4] and N-way Toolbox [22] are two widely
used MATLAB
The Cyclops Tensor Framework (CTF) is a C++ library which
provides automatic parallelization for sparse tensor operations.
etc
Distributed memory systems
Gigatensor handles tera-scale tensors using the MapReduce
framework.
Hypertensor is a sparse tensor library for SpMTTKRP on
distributed-memory environments.
etc
60 / 64
121. They pointed out different resources that you have around
Shared memory systems
The Tensor Toolbox [21], [4] and N-way Toolbox [22] are two widely
used MATLAB
The Cyclops Tensor Framework (CTF) is a C++ library which
provides automatic parallelization for sparse tensor operations.
etc
Distributed memory systems
Gigatensor handles tera-scale tensors using the MapReduce
framework.
Hypertensor is a sparse tensor library for SpMTTKRP on
distributed-memory environments.
etc
60 / 64
122. And the Grial
GPU
Li proposes a parallel algorithm and implementation of on GPUs via
parallelizing certain algorithms on fibers.
TensorFlow... actually supports certain version of Tensor
representation...
Something Notable
Efforts to solve more problems are on the way
The future looks promising
61 / 64
123. And the Grial
GPU
Li proposes a parallel algorithm and implementation of on GPUs via
parallelizing certain algorithms on fibers.
TensorFlow... actually supports certain version of Tensor
representation...
Something Notable
Efforts to solve more problems are on the way
The future looks promising
61 / 64
124. And the Grial
GPU
Li proposes a parallel algorithm and implementation of on GPUs via
parallelizing certain algorithms on fibers.
TensorFlow... actually supports certain version of Tensor
representation...
Something Notable
Efforts to solve more problems are on the way
The future looks promising
61 / 64
125. And the Grial
GPU
Li proposes a parallel algorithm and implementation of on GPUs via
parallelizing certain algorithms on fibers.
TensorFlow... actually supports certain version of Tensor
representation...
Something Notable
Efforts to solve more problems are on the way
The future looks promising
61 / 64
126. Outline
1 Introduction
The Dream of Tensors
A Short Story on Compression
A Short History
What a Heck are Tensors?
2 The Tensor Models for Data Science
Decomposition for Compression
CANDECOMP/PARAFAC Decomposition
The Dream of Compression and BIG DATA
Tensorizing Neural Networks
Hardware Support for the Dream
3 Conclusions
The Dream Will Follow....
62 / 64
127. As Always
We need people able to dream these new ways of doing stuff...
Therefore, a series of pieces of advise...
Learn more than a simple framework...
Learn the mathematics
And more importantly
Learn how to Model the Reality using such
Mathematical Tools...
63 / 64
128. As Always
We need people able to dream these new ways of doing stuff...
Therefore, a series of pieces of advise...
Learn more than a simple framework...
Learn the mathematics
And more importantly
Learn how to Model the Reality using such
Mathematical Tools...
63 / 64
129. As Always
We need people able to dream these new ways of doing stuff...
Therefore, a series of pieces of advise...
Learn more than a simple framework...
Learn the mathematics
And more importantly
Learn how to Model the Reality using such
Mathematical Tools...
63 / 64