This document proposes a stochastic model and design scenario analysis approach for managing uncertainties in hardware-software codesign projects. It describes representing a system using hierarchical, sequencing, and development views. Teams provide probabilistic estimates for tasks. A stochastic integer linear program formulation incorporates the views and estimates to analyze design scenarios varying objectives, risks, and constraints. Examples analyzing partitioning, risks, and convergence are presented for evaluating the approach on sample systems.
this is the forth slide for machine learning workshop in Hulu. Machine learning methods are summarized in the beginning of this slide, and boosting tree is introduced then. You are commended to try boosting tree when the feature number is not too much (<1000)
Implementation of the fully adaptive radar framework: Practical limitationsLuis Úbeda Medina
The document discusses the practical limitations of implementing a fully adaptive radar framework (FAR). It begins by outlining the key components of the FAR, including how the sensor parameters can be adaptively changed by a controller to better fit the system's needs based on information about the environment. It then presents the notation used and provides an example use case of tracking a target moving in a 2D environment using a sensor network with limited resources. Finally, it states that the last section will discuss the practical limitations of the FAR framework.
딥러닝 중급 - AlexNet과 VggNet (Basic of DCNN : AlexNet and VggNet)Hansol Kang
The document summarizes the basics of Deep Convolutional Neural Networks (DCNNs) including AlexNet and VGGNet. It discusses how AlexNet introduced improvements like ReLU activation and dropout to address overfitting issues. It then focuses on the VGGNet, noting that it achieved good performance through increasing depth using small 3x3 filters and adding convolutional layers. The document shares details of VGGNet configurations ranging from 11 to 19 weight layers and their performance on image classification tasks.
The document discusses algorithms for solving various optimization problems related to knapsack problems and scheduling problems. It begins by describing an efficient linear-time algorithm to find the largest subrectangle of 1s in a binary matrix using dynamic programming. It then discusses improvements to the space complexity of the 0/1 knapsack problem and algorithms for variants where items have unlimited quantities or values. Finally, it proposes algorithms for problems involving scheduling jobs on a single machine to maximize profit while meeting deadlines and partitioning a list into subsets with minimal sum difference.
This document provides an overview of deep deterministic policy gradient (DDPG), which combines aspects of DQN and policy gradient methods to enable deep reinforcement learning with continuous action spaces. It summarizes DQN and its limitations for continuous domains. It then explains policy gradient methods like REINFORCE, actor-critic, and deterministic policy gradient (DPG) that can handle continuous action spaces. DDPG adopts key elements of DQN like experience replay and target networks, and models the policy as a deterministic function like DPG, to apply deep reinforcement learning to complex continuous control tasks.
It's the deck for one Hulu internal machine learning workshop, which introduces the background, theory and application of expectation propagation method.
This document discusses speaker diarization, which is the process of segmenting an audio stream into homogeneous segments according to speaker identity. It covers feature extraction methods like MFCCs, segmentation using Bayesian Information Criteria to compare Gaussian mixture models, and clustering algorithms like k-means and hierarchical agglomerative clustering. Dendrogram visualizations are used to identify natural speaker clusters. The overall goal is to partition audio recordings of discussions or debates into homogeneous segments to attribute speech segments to individual speakers.
Hyperparameter optimization with approximate gradientFabian Pedregosa
This document discusses hyperparameter optimization using approximate gradients. It introduces the problem of optimizing hyperparameters along with model parameters. While model parameters can be estimated from data, hyperparameters require methods like cross-validation. The document proposes using approximate gradients to optimize hyperparameters more efficiently than costly methods like grid search. It derives the gradient of the objective with respect to hyperparameters and presents an algorithm called HOAG that approximates this gradient using inexact solutions. The document analyzes HOAG's convergence and provides experimental results comparing it to other hyperparameter optimization methods.
this is the forth slide for machine learning workshop in Hulu. Machine learning methods are summarized in the beginning of this slide, and boosting tree is introduced then. You are commended to try boosting tree when the feature number is not too much (<1000)
Implementation of the fully adaptive radar framework: Practical limitationsLuis Úbeda Medina
The document discusses the practical limitations of implementing a fully adaptive radar framework (FAR). It begins by outlining the key components of the FAR, including how the sensor parameters can be adaptively changed by a controller to better fit the system's needs based on information about the environment. It then presents the notation used and provides an example use case of tracking a target moving in a 2D environment using a sensor network with limited resources. Finally, it states that the last section will discuss the practical limitations of the FAR framework.
딥러닝 중급 - AlexNet과 VggNet (Basic of DCNN : AlexNet and VggNet)Hansol Kang
The document summarizes the basics of Deep Convolutional Neural Networks (DCNNs) including AlexNet and VGGNet. It discusses how AlexNet introduced improvements like ReLU activation and dropout to address overfitting issues. It then focuses on the VGGNet, noting that it achieved good performance through increasing depth using small 3x3 filters and adding convolutional layers. The document shares details of VGGNet configurations ranging from 11 to 19 weight layers and their performance on image classification tasks.
The document discusses algorithms for solving various optimization problems related to knapsack problems and scheduling problems. It begins by describing an efficient linear-time algorithm to find the largest subrectangle of 1s in a binary matrix using dynamic programming. It then discusses improvements to the space complexity of the 0/1 knapsack problem and algorithms for variants where items have unlimited quantities or values. Finally, it proposes algorithms for problems involving scheduling jobs on a single machine to maximize profit while meeting deadlines and partitioning a list into subsets with minimal sum difference.
This document provides an overview of deep deterministic policy gradient (DDPG), which combines aspects of DQN and policy gradient methods to enable deep reinforcement learning with continuous action spaces. It summarizes DQN and its limitations for continuous domains. It then explains policy gradient methods like REINFORCE, actor-critic, and deterministic policy gradient (DPG) that can handle continuous action spaces. DDPG adopts key elements of DQN like experience replay and target networks, and models the policy as a deterministic function like DPG, to apply deep reinforcement learning to complex continuous control tasks.
It's the deck for one Hulu internal machine learning workshop, which introduces the background, theory and application of expectation propagation method.
This document discusses speaker diarization, which is the process of segmenting an audio stream into homogeneous segments according to speaker identity. It covers feature extraction methods like MFCCs, segmentation using Bayesian Information Criteria to compare Gaussian mixture models, and clustering algorithms like k-means and hierarchical agglomerative clustering. Dendrogram visualizations are used to identify natural speaker clusters. The overall goal is to partition audio recordings of discussions or debates into homogeneous segments to attribute speech segments to individual speakers.
Hyperparameter optimization with approximate gradientFabian Pedregosa
This document discusses hyperparameter optimization using approximate gradients. It introduces the problem of optimizing hyperparameters along with model parameters. While model parameters can be estimated from data, hyperparameters require methods like cross-validation. The document proposes using approximate gradients to optimize hyperparameters more efficiently than costly methods like grid search. It derives the gradient of the objective with respect to hyperparameters and presents an algorithm called HOAG that approximates this gradient using inexact solutions. The document analyzes HOAG's convergence and provides experimental results comparing it to other hyperparameter optimization methods.
The document presents a deep reinforcement learning approach that uses a convolutional neural network trained with Q-learning to learn control policies directly from raw pixel input in complex Atari games. The network is trained to estimate the optimal action-value function by minimizing temporal-difference errors between its predictions and targets generated from rewards and subsequent predictions. Experience replay is used to alleviate problems with correlated data and non-stationary distributions by randomly sampling from a replay memory of past transitions. When applied to seven Atari games, the approach outperforms previous methods on six games and surpasses a human expert on three games using a single network architecture and hyperparameters across all games.
Deep Convolutional GANs - meaning of latent spaceHansol Kang
DCGAN은 GAN에 단순히 conv net을 적용했을 뿐만 아니라, latent space에서도 의미를 찾음.
DCGAN 논문 리뷰 및 PyTorch 기반의 구현.
VAE 세미나 이슈 사항에 대한 리뷰.
my github : https://github.com/messy-snail/GAN_PyTorch
[참고]
https://github.com/znxlwm/pytorch-MNIST-CelebA-GAN-DCGAN
https://github.com/taeoh-kim/Pytorch_DCGAN
Radford, Alec, Luke Metz, and Soumith Chintala. "Unsupervised representation learning with deep convolutional generative adversarial networks." arXiv preprint arXiv:1511.06434 (2015).
Generative Adversarial Networks (GANs) are a class of machine learning frameworks where two neural networks contest with each other in a game. A generator network generates new data instances, while a discriminator network evaluates them for authenticity, classifying them as real or generated. This adversarial process allows the generator to improve over time and generate highly realistic samples that can pass for real data. The document provides an overview of GANs and their variants, including DCGAN, InfoGAN, EBGAN, and ACGAN models. It also discusses techniques for training more stable GANs and escaping issues like mode collapse.
Ryan White presented work on analyzing stochastic strategic networks. He developed models where networks experience attacks over time that remove nodes and weight. The models consider observing the network process only at renewal times, and finding functionals of interest like the distribution of the first observed threshold crossing. For a special case, he derived tractable results for the functional and compared to simulations. He discussed extensions like additional auxiliary thresholds.
쉽게 설명하는 GAN (What is this? Gum? It's GAN.)Hansol Kang
The document discusses generative adversarial networks (GANs). It begins with an introduction to GANs, describing their concept and training process. It then reviews a seminal GAN paper, discussing its mathematical formulation of GAN training as a minimax game and theoretical results showing global optimality can be achieved. The document concludes by outlining the configuration, implementation, and flowchart for a GAN experiment.
Photo-realistic Single Image Super-resolution using a Generative Adversarial ...Hansol Kang
* Ledig, Christian, et al. "Photo-realistic single image super-resolution using a generative adversarial network." Proceedings of the IEEE conference on computer vision and pattern recognition. 2017.
PyTorch is an open-source machine learning library for Python. It is primarily developed by Facebook's AI research group. The document discusses setting up PyTorch, including installing necessary packages and configuring development environments. It also provides examples of core PyTorch concepts like tensors, common datasets, and constructing basic neural networks.
Data science involves extracting insights from large volumes of data. It is an interdisciplinary field that uses techniques from statistics, machine learning, and other domains. The document provides examples of classification algorithms like k-nearest neighbors, naive Bayes, and perceptrons that are commonly used in data science to build models for tasks like spam filtering or sentiment analysis. It also discusses clustering, frequent pattern mining, and other machine learning concepts.
EuroPython 2017 - PyData - Deep Learning your Broadband Network @ HOMEHONGJOO LEE
45 min talk about collecting home network performance measures, analyzing and forecasting time series data, and building anomaly detection system.
In this talk, we will go through the whole process of data mining and knowledge discovery. Firstly we write a script to run speed test periodically and log the metric. Then we parse the log data and convert them into a time series and visualize the data for a certain period.
Next we conduct some data analysis; finding trends, forecasting, and detecting anomalous data. There will be several statistic or deep learning techniques used for the analysis; ARIMA (Autoregressive Integrated Moving Average), LSTM (Long Short Term Memory).
Gradient descent optimization with simple examples. covers sgd, mini-batch, momentum, adagrad, rmsprop and adam.
Made for people with little knowledge of neural network.
Adversarial learning for neural dialogue generationKeon Kim
This document summarizes an adversarial learning approach for neural dialogue generation. The model uses a generator and discriminator, where the generator produces responses and the discriminator determines if they are human-like. The generator is trained to maximize rewards from the discriminator using policy gradients. Two methods are introduced to assign rewards at each generation step to address issues with the baseline approach. Teacher forcing is also used to directly expose the generator to human responses during training. The results showed this adversarial training approach generates higher quality responses than previous baselines.
The document discusses distributed linear classification on Apache Spark. It describes using Spark to train logistic regression and linear support vector machine models on large datasets. Spark improves on MapReduce by conducting communications in-memory and supporting fault tolerance. The paper proposes using a trust region Newton method to optimize the objective functions for logistic regression and linear SVM. Conjugate gradient is used to approximate the Hessian matrix and solve the Newton system without explicitly storing the large Hessian.
파이콘 코리아 2018년도 튜토리얼 세션의 "RL Adventure : DQN 부터 Rainbow DQN까지"의 발표 자료입니다.
2017년도 Deepmind에서 발표한 value based 강화학습 모형인 Rainbow의 이해를 돕기 위한 튜토리얼로 DQN부터 Rainbow까지 순차적으로 중요한 점만 요약된 내용이 들어있습니다.
파트 1 : DQN, Double & Dueling DQN - 성태경
파트 2 : PER and NoisyNet - 양홍선
파트 3 : Distributed RL - 이의령
파트 4 : RAINBOW - 김예찬
관련된 코드와 구현체를 확인하고 싶으신 분들은
https://github.com/hongdam/pycon2018-RL_Adventure
에서 확인하실 수 있습니다
The document outlines topics covered in a NetworkX tutorial, including installation, basic classes, generating graphs, analyzing graphs, saving/loading graphs, and plotting graphs with Matplotlib. Specific sections cover local and cluster installation of NetworkX, adding nodes and edges to graphs along with attributes, basic graph properties like number of nodes/edges and neighbors, simple graph generators, random graph generators, and the algorithms package.
TensorFlow Dev Summit 2018 Extended: TensorFlow Eager ExecutionTaegyun Jeon
TensorFlow's eager execution allows running operations immediately without building graphs. This makes debugging easier and improves the development workflow. Eager execution can be enabled with tf.enable_eager_execution(). Common operations like variables, gradients, control flow work the same in eager and graph modes. Code written with eager execution in mind is compatible with graph-based execution for deployment. Eager execution provides benefits for iteration and is useful alongside TensorFlow's high-level APIs.
is anyone_interest_in_auto-encoding_variational-bayesNAVER Engineering
Deep generative model 중 하나인 VAE의 Framework은 컴퓨터 비전, 자연어 처리 등 머신러닝의 전반에서 generative model의 변화를 가져왔다.
VAE를 처음 접하는 연구자들을 위해 대부분의 VAE tutorial은 구현을 목적으로 Neural Network구조와 Loss function에 초점을 맞추고 있다. 본 세미나는 Variational Inference 관점에서 Auto-encoding variational bayes에 나오는 수식들을 살펴보고자 한다. 본 수식들이 구현에서는 어떻게 적용되는지도 살펴보고자 한다.
[GAN by Hung-yi Lee]Part 1: General introduction of GANNAVER Engineering
Generative Adversarial Network and its Applications on Speech and Natural Language Processing, Part 1.
발표자: Hung-yi Lee(국립 타이완대 교수)
발표일: 18.7.
Generative adversarial network (GAN) is a new idea for training models, in which a generator and a discriminator compete against each other to improve the generation quality. Recently, GAN has shown amazing results in image generation, and a large amount and a wide variety of new ideas, techniques, and applications have been developed based on it. Although there are only few successful cases, GAN has great potential to be applied to text and speech generations to overcome limitations in the conventional methods.
In the first part of the talk, I will first give an introduction of GAN and provide a thorough review about this technology. In the second part, I will focus on the applications of GAN to speech and natural language processing. I will demonstrate the applications of GAN on voice I will also talk about the research directions towards unsupervised speech recognition by GAN.conversion, unsupervised abstractive summarization and sentiment controllable chat-bot.
QMIX: monotonic value function factorization paper review민재 정
QMIX is a deep multi-agent reinforcement learning method that allows for centralized training with decentralized execution. It represents the joint action-value function as a factored and monotonic combination of individual agent value functions. This ensures greedy policies over the individual value functions correspond to greedy policies over the joint value function. Experiments in StarCraft II micromanagement tasks show QMIX outperforms independent learners and value decomposition networks by effectively learning cooperative behaviors while ensuring scalability.
Este documento discute os desafios enfrentados por universidades brasileiras para serem relevantes e inovadoras. Apresenta exemplos de projetos desenvolvidos na UFRPE que resolveram problemas reais, mas enfrentaram barreiras burocráticas. Defende que a computação deve se afastar de abordagens rígidas e incentivar a solução inovadora de problemas complexos de verdade para a sociedade.
The document presents a deep reinforcement learning approach that uses a convolutional neural network trained with Q-learning to learn control policies directly from raw pixel input in complex Atari games. The network is trained to estimate the optimal action-value function by minimizing temporal-difference errors between its predictions and targets generated from rewards and subsequent predictions. Experience replay is used to alleviate problems with correlated data and non-stationary distributions by randomly sampling from a replay memory of past transitions. When applied to seven Atari games, the approach outperforms previous methods on six games and surpasses a human expert on three games using a single network architecture and hyperparameters across all games.
Deep Convolutional GANs - meaning of latent spaceHansol Kang
DCGAN은 GAN에 단순히 conv net을 적용했을 뿐만 아니라, latent space에서도 의미를 찾음.
DCGAN 논문 리뷰 및 PyTorch 기반의 구현.
VAE 세미나 이슈 사항에 대한 리뷰.
my github : https://github.com/messy-snail/GAN_PyTorch
[참고]
https://github.com/znxlwm/pytorch-MNIST-CelebA-GAN-DCGAN
https://github.com/taeoh-kim/Pytorch_DCGAN
Radford, Alec, Luke Metz, and Soumith Chintala. "Unsupervised representation learning with deep convolutional generative adversarial networks." arXiv preprint arXiv:1511.06434 (2015).
Generative Adversarial Networks (GANs) are a class of machine learning frameworks where two neural networks contest with each other in a game. A generator network generates new data instances, while a discriminator network evaluates them for authenticity, classifying them as real or generated. This adversarial process allows the generator to improve over time and generate highly realistic samples that can pass for real data. The document provides an overview of GANs and their variants, including DCGAN, InfoGAN, EBGAN, and ACGAN models. It also discusses techniques for training more stable GANs and escaping issues like mode collapse.
Ryan White presented work on analyzing stochastic strategic networks. He developed models where networks experience attacks over time that remove nodes and weight. The models consider observing the network process only at renewal times, and finding functionals of interest like the distribution of the first observed threshold crossing. For a special case, he derived tractable results for the functional and compared to simulations. He discussed extensions like additional auxiliary thresholds.
쉽게 설명하는 GAN (What is this? Gum? It's GAN.)Hansol Kang
The document discusses generative adversarial networks (GANs). It begins with an introduction to GANs, describing their concept and training process. It then reviews a seminal GAN paper, discussing its mathematical formulation of GAN training as a minimax game and theoretical results showing global optimality can be achieved. The document concludes by outlining the configuration, implementation, and flowchart for a GAN experiment.
Photo-realistic Single Image Super-resolution using a Generative Adversarial ...Hansol Kang
* Ledig, Christian, et al. "Photo-realistic single image super-resolution using a generative adversarial network." Proceedings of the IEEE conference on computer vision and pattern recognition. 2017.
PyTorch is an open-source machine learning library for Python. It is primarily developed by Facebook's AI research group. The document discusses setting up PyTorch, including installing necessary packages and configuring development environments. It also provides examples of core PyTorch concepts like tensors, common datasets, and constructing basic neural networks.
Data science involves extracting insights from large volumes of data. It is an interdisciplinary field that uses techniques from statistics, machine learning, and other domains. The document provides examples of classification algorithms like k-nearest neighbors, naive Bayes, and perceptrons that are commonly used in data science to build models for tasks like spam filtering or sentiment analysis. It also discusses clustering, frequent pattern mining, and other machine learning concepts.
EuroPython 2017 - PyData - Deep Learning your Broadband Network @ HOMEHONGJOO LEE
45 min talk about collecting home network performance measures, analyzing and forecasting time series data, and building anomaly detection system.
In this talk, we will go through the whole process of data mining and knowledge discovery. Firstly we write a script to run speed test periodically and log the metric. Then we parse the log data and convert them into a time series and visualize the data for a certain period.
Next we conduct some data analysis; finding trends, forecasting, and detecting anomalous data. There will be several statistic or deep learning techniques used for the analysis; ARIMA (Autoregressive Integrated Moving Average), LSTM (Long Short Term Memory).
Gradient descent optimization with simple examples. covers sgd, mini-batch, momentum, adagrad, rmsprop and adam.
Made for people with little knowledge of neural network.
Adversarial learning for neural dialogue generationKeon Kim
This document summarizes an adversarial learning approach for neural dialogue generation. The model uses a generator and discriminator, where the generator produces responses and the discriminator determines if they are human-like. The generator is trained to maximize rewards from the discriminator using policy gradients. Two methods are introduced to assign rewards at each generation step to address issues with the baseline approach. Teacher forcing is also used to directly expose the generator to human responses during training. The results showed this adversarial training approach generates higher quality responses than previous baselines.
The document discusses distributed linear classification on Apache Spark. It describes using Spark to train logistic regression and linear support vector machine models on large datasets. Spark improves on MapReduce by conducting communications in-memory and supporting fault tolerance. The paper proposes using a trust region Newton method to optimize the objective functions for logistic regression and linear SVM. Conjugate gradient is used to approximate the Hessian matrix and solve the Newton system without explicitly storing the large Hessian.
파이콘 코리아 2018년도 튜토리얼 세션의 "RL Adventure : DQN 부터 Rainbow DQN까지"의 발표 자료입니다.
2017년도 Deepmind에서 발표한 value based 강화학습 모형인 Rainbow의 이해를 돕기 위한 튜토리얼로 DQN부터 Rainbow까지 순차적으로 중요한 점만 요약된 내용이 들어있습니다.
파트 1 : DQN, Double & Dueling DQN - 성태경
파트 2 : PER and NoisyNet - 양홍선
파트 3 : Distributed RL - 이의령
파트 4 : RAINBOW - 김예찬
관련된 코드와 구현체를 확인하고 싶으신 분들은
https://github.com/hongdam/pycon2018-RL_Adventure
에서 확인하실 수 있습니다
The document outlines topics covered in a NetworkX tutorial, including installation, basic classes, generating graphs, analyzing graphs, saving/loading graphs, and plotting graphs with Matplotlib. Specific sections cover local and cluster installation of NetworkX, adding nodes and edges to graphs along with attributes, basic graph properties like number of nodes/edges and neighbors, simple graph generators, random graph generators, and the algorithms package.
TensorFlow Dev Summit 2018 Extended: TensorFlow Eager ExecutionTaegyun Jeon
TensorFlow's eager execution allows running operations immediately without building graphs. This makes debugging easier and improves the development workflow. Eager execution can be enabled with tf.enable_eager_execution(). Common operations like variables, gradients, control flow work the same in eager and graph modes. Code written with eager execution in mind is compatible with graph-based execution for deployment. Eager execution provides benefits for iteration and is useful alongside TensorFlow's high-level APIs.
is anyone_interest_in_auto-encoding_variational-bayesNAVER Engineering
Deep generative model 중 하나인 VAE의 Framework은 컴퓨터 비전, 자연어 처리 등 머신러닝의 전반에서 generative model의 변화를 가져왔다.
VAE를 처음 접하는 연구자들을 위해 대부분의 VAE tutorial은 구현을 목적으로 Neural Network구조와 Loss function에 초점을 맞추고 있다. 본 세미나는 Variational Inference 관점에서 Auto-encoding variational bayes에 나오는 수식들을 살펴보고자 한다. 본 수식들이 구현에서는 어떻게 적용되는지도 살펴보고자 한다.
[GAN by Hung-yi Lee]Part 1: General introduction of GANNAVER Engineering
Generative Adversarial Network and its Applications on Speech and Natural Language Processing, Part 1.
발표자: Hung-yi Lee(국립 타이완대 교수)
발표일: 18.7.
Generative adversarial network (GAN) is a new idea for training models, in which a generator and a discriminator compete against each other to improve the generation quality. Recently, GAN has shown amazing results in image generation, and a large amount and a wide variety of new ideas, techniques, and applications have been developed based on it. Although there are only few successful cases, GAN has great potential to be applied to text and speech generations to overcome limitations in the conventional methods.
In the first part of the talk, I will first give an introduction of GAN and provide a thorough review about this technology. In the second part, I will focus on the applications of GAN to speech and natural language processing. I will demonstrate the applications of GAN on voice I will also talk about the research directions towards unsupervised speech recognition by GAN.conversion, unsupervised abstractive summarization and sentiment controllable chat-bot.
QMIX: monotonic value function factorization paper review민재 정
QMIX is a deep multi-agent reinforcement learning method that allows for centralized training with decentralized execution. It represents the joint action-value function as a factored and monotonic combination of individual agent value functions. This ensures greedy policies over the individual value functions correspond to greedy policies over the joint value function. Experiments in StarCraft II micromanagement tasks show QMIX outperforms independent learners and value decomposition networks by effectively learning cooperative behaviors while ensuring scalability.
Este documento discute os desafios enfrentados por universidades brasileiras para serem relevantes e inovadoras. Apresenta exemplos de projetos desenvolvidos na UFRPE que resolveram problemas reais, mas enfrentaram barreiras burocráticas. Defende que a computação deve se afastar de abordagens rígidas e incentivar a solução inovadora de problemas complexos de verdade para a sociedade.
Codigos maquinas linguagens... E triathlon ;-) TUDO é software?Jones Albuquerque
O documento discute a evolução dos códigos e linguagens ao longo da história humana, desde os primeiros sistemas de escrita até as linguagens de programação modernas. Também aborda como as máquinas atuais, como computadores, são programadas usando esses códigos e linguagens. Por fim, analisa dois estudos de caso - esquistossomose e triatlo - para ilustrar como até mesmo sistemas biológicos podem ser vistos como "software".
Codes languages, machines and synthetic biology and one case LIKA-CESAR-BRAZILJones Albuquerque
This document discusses codes, languages, machines, and synthetic biology. It begins by comparing genetic engineering to computer science, with genetic codes and grammars analogous to programming languages. It then describes the iGEM-LIKA-CESAR synthetic biology project, which aims to develop a robotic system linked to genetic engineering to detect breast cancer. The document argues that synthetic biology combines molecular engineering and computer science to program DNA like software. It questions whether there could be universal genetic machines or grammars, similar to concepts in computer science, as synthetic biology continues to merge the biological and digital worlds.
This document discusses the parallels between computer science and synthetic biology from a perspective that "everything is software". It notes how early humans developed coding systems like mathematics to record and compute information. Over time, more advanced coding systems like writing systems and languages emerged, alongside machines that could process these codes. The document then discusses how computer science developed coding languages and machines that can recognize and execute these languages. It argues that synthetic biology may be viewed similarly, with DNA acting as a coding language that can be programmed like software. The document concludes by discussing how initiatives like iGEM are attempting to engineer genetic codes and grammars to build biological machines, bringing together concepts from computer science, engineering and synthetic biology.
Introdução a Gerência de Configuração de SoftwareCamilo Almendra
Este documento apresenta os principais conceitos e benefícios da Gerência de Configuração, incluindo problemas comuns no desenvolvimento de software que a GC pode ajudar a resolver, como quebra de comunicação entre equipes e atualização simultânea de componentes compartilhados. A GC é definida como o processo de identificar, organizar e controlar modificações ao software sendo construído.
ISI TICs - SENAI INSTITUTES and status - english version v3Jones Albuquerque
This document provides information on SENAI's Innovation Institute for Information and Communication Technologies (ISI-TICs) located in Pernambuco, Brazil. It discusses the ISI-TICs vision of establishing ICT/software as the main factor of Brazilian industry competitiveness through innovation. It outlines the ISI-TICs mission to promote industry competitiveness through knowledge transfer, applied research, and innovation. It also provides details on ISI-TICs service areas and the strong ICT ecosystem in Pernambuco that it is a part of, including organizations like Porto Digital, CIn/UFPE, C.E.S.A.R, and INES.
Learning to Project and Binarise for Hashing-based Approximate Nearest Neighb...Sean Moran
In this paper we focus on improving the effectiveness of hashing-based approximate nearest neighbour search. Generating similarity preserving hashcodes for images has been shown to be an effective and efficient method for searching through large datasets. Hashcode generation generally involves two steps: bucketing the input feature space with a set of hyperplanes, followed by quantising the projection of the data-points onto the normal vectors to those hyperplanes. This procedure results in the makeup of the hashcodes depending on the positions of the data-points with respect to the hyperplanes in the feature space, allowing a degree of locality to be encoded into the hashcodes. In this paper we study the effect of learning both the hyperplanes and the thresholds as part of the same model. Most previous research either learn the hyperplanes assuming a fixed set of thresholds, or vice-versa. In our experiments over two standard image datasets we find statistically significant increases in retrieval effectiveness versus a host of state-of-the-art data-dependent and independent hashing models.
Gradient Boosted Regression Trees in Scikit Learn by Gilles Louppe & Peter Pr...PyData
This document discusses gradient boosted regression trees (GBRT) and their implementation in scikit-learn. It begins with an introduction to machine learning concepts like classification, regression, and decision trees. It then covers the basics of boosting and gradient boosting, describing how GBRT works by sequentially fitting trees to residuals. The rest of the document demonstrates scikit-learn's GBRT implementation, provides tips on regularization and hyperparameters, and presents a case study on house price prediction.
Software Defect Prediction on Unlabeled DatasetsSung Kim
The document describes techniques for software defect prediction when labeled training data is unavailable. It proposes Transfer Defect Learning (TCA+) to improve cross-project defect prediction by normalizing data distributions between source and target projects. For projects with heterogeneous metrics, it introduces Heterogeneous Defect Prediction (HDP) which matches similar metrics between source and target to build cross-project prediction models. It also discusses CLAMI for defect prediction using only unlabeled data without human effort. The techniques are evaluated on open source projects to demonstrate their effectiveness compared to traditional cross-project and within-project prediction.
Transfer Learning for Improving Model Predictions in Highly Configurable Soft...Pooyan Jamshidi
Modern software systems are now being built to be used in dynamic environments utilizing configuration capabilities to adapt to changes and external uncertainties. In a self-adaptation context, we are often interested in reasoning about the performance of the systems under different configurations. Usually, we learn a black-box model based on real measurements to predict the performance of the system given a specific configuration. However, as modern systems become more complex, there are many configuration parameters that may interact and, therefore, we end up learning an exponentially large configuration space. Naturally, this does not scale when relying on real measurements in the actual changing environment. We propose a different solution: Instead of taking the measurements from the real system, we learn the model using samples from other sources, such as simulators that approximate performance of the real system at low cost.
Yulia Honcharenko "Application of metric learning for logo recognition"Fwdays
Typical approaches of solving classification problems require the collection of a dataset for each new class and retraining of the model. Metric learning allows you to train a model once and then easily add new classes with 5-10 reference images.
So we’ll talk about metric learning based on YouScan experience: task, data, different losses and approaches, metrics we used, pitfalls and peculiarities, things that worked and didn’t.
Abstract : For many years, Machine Learning has focused on a key issue: the design of input features to solve prediction tasks. In this presentation, we show that many learning tasks from structured output prediction to zero-shot learning can benefit from an appropriate design of output features, broadening the scope of regression. As an illustration, I will briefly review different examples and recent results obtained in my team.
The document introduces two approaches to chemical prediction: quantum simulation based on density functional theory and machine learning based on data. It then discusses using graph-structured neural networks for chemical prediction on datasets like QM9. It presents Neural Fingerprint (NFP) and Gated Graph Neural Network (GGNN) models for predicting molecular properties from graph-structured data. Chainer Chemistry is introduced as a library for chemical and biological machine learning that implements these graph convolutional networks.
This document describes a machine learning project that uses support vector machines (SVM) and k-nearest neighbors (k-NN) algorithms to segment gesture phases based on radial basis function (RBF) kernels and k-nearest neighbors. The project aims to classify frames of movement data into five gesture phases (rest, preparation, stroke, hold, retraction) using two classifiers. The SVM approach achieved 53.27% accuracy on test data while the k-NN approach achieved significantly higher accuracy of 92.53%. The document provides details on the dataset, feature extraction methods, model selection process and results of applying each classifier to the test data.
This document contains slides from a lecture on pattern recognition. It discusses several topics:
- Maximum likelihood estimation and how it can be used to estimate parameters of Gaussian distributions from sample data.
- The problem of dimensionality when applying pattern recognition techniques - as the number of features or dimensions increases, classification accuracy may decrease and computational complexity increases.
- Component analysis techniques like PCA and LDA that aim to reduce dimensionality by projecting data onto a lower-dimensional space.
- An assignment involving generating an image with multiple classes, estimating class parameters with MLE, and classifying pixels with Bayesian decision theory.
Gradient Boosted Regression Trees in scikit-learnDataRobot
Slides of the talk "Gradient Boosted Regression Trees in scikit-learn" by Peter Prettenhofer and Gilles Louppe held at PyData London 2014.
Abstract:
This talk describes Gradient Boosted Regression Trees (GBRT), a powerful statistical learning technique with applications in a variety of areas, ranging from web page ranking to environmental niche modeling. GBRT is a key ingredient of many winning solutions in data-mining competitions such as the Netflix Prize, the GE Flight Quest, or the Heritage Health Price.
I will give a brief introduction to the GBRT model and regression trees -- focusing on intuition rather than mathematical formulas. The majority of the talk will be dedicated to an in depth discussion how to apply GBRT in practice using scikit-learn. We will cover important topics such as regularization, model tuning and model interpretation that should significantly improve your score on Kaggle.
Hanjun Dai, PhD Student, School of Computational Science and Engineering, Geo...MLconf
Graph Representation Learning with Deep Embedding Approach:
Graphs are commonly used data structure for representing the real-world relationships, e.g., molecular structure, knowledge graphs, social and communication networks. The effective encoding of graphical information is essential to the success of such applications. In this talk I’ll first describe a general deep learning framework, namely structure2vec, for end to end graph feature representation learning. Then I’ll present the direct application of this model on graph problems on different scales, including community detection and molecule graph classification/regression. We then extend the embedding idea to temporal evolving user-product interaction graph for recommendation. Finally I’ll present our latest work on leveraging the reinforcement learning technique for graph combinatorial optimization, including vertex cover problem for social influence maximization and traveling salesman problem for scheduling management.
The document discusses machine learning optimization problems and linear/logistic regression algorithms. It notes that machine learning can be viewed as an optimization problem with constraints, a function to optimize, and an optimization algorithm. Linear regression aims to minimize prediction error by finding the best fitting linear model, while logistic regression predicts class probabilities using a sigmoid function. Both use gradient descent to optimize their error functions and learn model parameters from data.
- The document presents a neural network model for recognizing handwritten digits. It uses a dataset of 20x20 pixel grayscale images of digits 0-9.
- The proposed neural network has an input layer of 400 nodes, a hidden layer of 25 nodes, and an output layer of 10 nodes. It is trained using backpropagation to classify images.
- The model achieves an accuracy of over 96.5% on test data after 200 iterations of training, outperforming a logistic regression model which achieved 91.5% accuracy. Future work could involve classifying more complex natural images.
This document discusses object detection using Adaboost and various techniques. It begins with an overview of the Adaboost algorithm and provides a toy example to illustrate how it works. Next, it describes how Viola and Jones used Adaboost with Haar-like features and an integral image representation for rapid face detection in images. It achieved high detection rates with very low false positives. The document also discusses how Schneiderman and Kanade used a parts-based representation with localized wavelet coefficients as features for object detection and used statistical independence of parts to obtain likelihoods for classification.
The document proposes using bandit structured prediction to train neural machine translation models with weak feedback in the form of task loss evaluations instead of full labeled data. It applies this approach to domain adaptation by training on one domain and evaluating on another. Control variates are used to reduce variance and improve generalization. Experimental results show the approach outperforms linear models and prior work, successfully adapting a model from Europarl to News Commentary and TED talks with improved BLEU scores over baselines on both in-domain and out-of-domain test sets.
This document presents a comparative study of two genetic algorithm-based task allocation models in distributed computing systems. It aims to minimize turnaround time, where the previous model aimed to maximize reliability. The models are implemented on two example cases, with the minimum turnaround time model finding an allocation with a turnaround of 14 units and slightly lower reliability than the maximum reliability model's allocation of 20 units. In conclusion, minimizing turnaround time leads to slightly reduced reliability compared to maximizing reliability.
[PR12] PR-036 Learning to Remember Rare EventsTaegyun Jeon
This document summarizes a paper on learning to remember rare events using a memory-augmented neural network. The paper proposes a memory module that stores examples from previous tasks to help learn new rare tasks from only a single example. The memory module is trained end-to-end with the neural network on two tasks: one-shot learning on Omniglot characters and machine translation of rare words. The implementation uses a TensorFlow memory module that stores key-value pairs to retrieve examples similar to a query. Experiments show the memory module improves one-shot learning performance and handles rare words better than baselines.
The asynchronous parallel algorithms are developed to solve massive optimization problems in a distributed data system, which can be run in parallel on multiple nodes with little or no synchronization. Recently they have been successfully implemented to solve a range of difficult problems in practice. However, the existing theories are mostly based on fairly restrictive assumptions on the delays, and cannot explain the convergence and speedup properties of such algorithms. In this talk we will give an overview on distributed optimization, and discuss some new theoretical results on the convergence of asynchronous parallel stochastic gradient algorithm with unbounded delays. Simulated and real data will be used to demonstrate the practical implication of these theoretical results.
Transfer Learning for Performance Analysis of Highly-Configurable SoftwarePooyan Jamshidi
This document discusses using transfer learning to analyze the performance of configurable software systems. It begins by noting that today's most popular software systems are highly configurable and that their increasing configurability makes understanding performance behavior difficult. The author then describes using transfer learning to enable learning performance models more efficiently by reusing data from related source domains. This allows developers and users to better understand performance tradeoffs and find optimal configurations.
The document discusses various methods for estimating effort for software projects. It describes that estimating costs is crucial for project planning and should be done early. It identifies the main types of costs as facilities, methods/tools, and staff effort. It then examines causes of inaccurate estimates such as changes, overlooked tasks, lack of requirements understanding, and insufficient analysis. Key influences on estimates are also listed such as complexity, integration needs, team experience, and standards. The document advocates repeating estimates and describes expert judgment, analogy, Delphi technique, and Wolverton models as estimation methods.
Similar to Managing Uncertainties in Hardware-Software Codesign Projects (20)
TOPIC OF DISCUSSION: CENTRIFUGATION SLIDESHARE.pptxshubhijain836
Centrifugation is a powerful technique used in laboratories to separate components of a heterogeneous mixture based on their density. This process utilizes centrifugal force to rapidly spin samples, causing denser particles to migrate outward more quickly than lighter ones. As a result, distinct layers form within the sample tube, allowing for easy isolation and purification of target substances.
Signatures of wave erosion in Titan’s coastsSérgio Sacani
The shorelines of Titan’s hydrocarbon seas trace flooded erosional landforms such as river valleys; however, it isunclear whether coastal erosion has subsequently altered these shorelines. Spacecraft observations and theo-retical models suggest that wind may cause waves to form on Titan’s seas, potentially driving coastal erosion,but the observational evidence of waves is indirect, and the processes affecting shoreline evolution on Titanremain unknown. No widely accepted framework exists for using shoreline morphology to quantitatively dis-cern coastal erosion mechanisms, even on Earth, where the dominant mechanisms are known. We combinelandscape evolution models with measurements of shoreline shape on Earth to characterize how differentcoastal erosion mechanisms affect shoreline morphology. Applying this framework to Titan, we find that theshorelines of Titan’s seas are most consistent with flooded landscapes that subsequently have been eroded bywaves, rather than a uniform erosional process or no coastal erosion, particularly if wave growth saturates atfetch lengths of tens of kilometers.
Candidate young stellar objects in the S-cluster: Kinematic analysis of a sub...Sérgio Sacani
Context. The observation of several L-band emission sources in the S cluster has led to a rich discussion of their nature. However, a definitive answer to the classification of the dusty objects requires an explanation for the detection of compact Doppler-shifted Brγ emission. The ionized hydrogen in combination with the observation of mid-infrared L-band continuum emission suggests that most of these sources are embedded in a dusty envelope. These embedded sources are part of the S-cluster, and their relationship to the S-stars is still under debate. To date, the question of the origin of these two populations has been vague, although all explanations favor migration processes for the individual cluster members. Aims. This work revisits the S-cluster and its dusty members orbiting the supermassive black hole SgrA* on bound Keplerian orbits from a kinematic perspective. The aim is to explore the Keplerian parameters for patterns that might imply a nonrandom distribution of the sample. Additionally, various analytical aspects are considered to address the nature of the dusty sources. Methods. Based on the photometric analysis, we estimated the individual H−K and K−L colors for the source sample and compared the results to known cluster members. The classification revealed a noticeable contrast between the S-stars and the dusty sources. To fit the flux-density distribution, we utilized the radiative transfer code HYPERION and implemented a young stellar object Class I model. We obtained the position angle from the Keplerian fit results; additionally, we analyzed the distribution of the inclinations and the longitudes of the ascending node. Results. The colors of the dusty sources suggest a stellar nature consistent with the spectral energy distribution in the near and midinfrared domains. Furthermore, the evaporation timescales of dusty and gaseous clumps in the vicinity of SgrA* are much shorter ( 2yr) than the epochs covered by the observations (≈15yr). In addition to the strong evidence for the stellar classification of the D-sources, we also find a clear disk-like pattern following the arrangements of S-stars proposed in the literature. Furthermore, we find a global intrinsic inclination for all dusty sources of 60 ± 20◦, implying a common formation process. Conclusions. The pattern of the dusty sources manifested in the distribution of the position angles, inclinations, and longitudes of the ascending node strongly suggests two different scenarios: the main-sequence stars and the dusty stellar S-cluster sources share a common formation history or migrated with a similar formation channel in the vicinity of SgrA*. Alternatively, the gravitational influence of SgrA* in combination with a massive perturber, such as a putative intermediate mass black hole in the IRS 13 cluster, forces the dusty objects and S-stars to follow a particular orbital arrangement. Key words. stars: black holes– stars: formation– Galaxy: center– galaxies: star formation
Dr. Firoozeh Kashani-Sabet is an innovator in Middle Eastern Studies and approaches her work, particularly focused on Iran, with a depth and commitment that has resulted in multiple book publications. She is notable for her work with the University of Pennsylvania, where she serves as the Walter H. Annenberg Professor of History.
BIRDS DIVERSITY OF SOOTEA BISWANATH ASSAM.ppt.pptxgoluk9330
Ahota Beel, nestled in Sootea Biswanath Assam , is celebrated for its extraordinary diversity of bird species. This wetland sanctuary supports a myriad of avian residents and migrants alike. Visitors can admire the elegant flights of migratory species such as the Northern Pintail and Eurasian Wigeon, alongside resident birds including the Asian Openbill and Pheasant-tailed Jacana. With its tranquil scenery and varied habitats, Ahota Beel offers a perfect haven for birdwatchers to appreciate and study the vibrant birdlife that thrives in this natural refuge.
Compositions of iron-meteorite parent bodies constrainthe structure of the pr...Sérgio Sacani
Magmatic iron-meteorite parent bodies are the earliest planetesimals in the Solar System,and they preserve information about conditions and planet-forming processes in thesolar nebula. In this study, we include comprehensive elemental compositions andfractional-crystallization modeling for iron meteorites from the cores of five differenti-ated asteroids from the inner Solar System. Together with previous results of metalliccores from the outer Solar System, we conclude that asteroidal cores from the outerSolar System have smaller sizes, elevated siderophile-element abundances, and simplercrystallization processes than those from the inner Solar System. These differences arerelated to the formation locations of the parent asteroids because the solar protoplane-tary disk varied in redox conditions, elemental distributions, and dynamics at differentheliocentric distances. Using highly siderophile-element data from iron meteorites, wereconstruct the distribution of calcium-aluminum-rich inclusions (CAIs) across theprotoplanetary disk within the first million years of Solar-System history. CAIs, the firstsolids to condense in the Solar System, formed close to the Sun. They were, however,concentrated within the outer disk and depleted within the inner disk. Future modelsof the structure and evolution of the protoplanetary disk should account for this dis-tribution pattern of CAIs.
Evidence of Jet Activity from the Secondary Black Hole in the OJ 287 Binary S...Sérgio Sacani
Wereport the study of a huge optical intraday flare on 2021 November 12 at 2 a.m. UT in the blazar OJ287. In the binary black hole model, it is associated with an impact of the secondary black hole on the accretion disk of the primary. Our multifrequency observing campaign was set up to search for such a signature of the impact based on a prediction made 8 yr earlier. The first I-band results of the flare have already been reported by Kishore et al. (2024). Here we combine these data with our monitoring in the R-band. There is a big change in the R–I spectral index by 1.0 ±0.1 between the normal background and the flare, suggesting a new component of radiation. The polarization variation during the rise of the flare suggests the same. The limits on the source size place it most reasonably in the jet of the secondary BH. We then ask why we have not seen this phenomenon before. We show that OJ287 was never before observed with sufficient sensitivity on the night when the flare should have happened according to the binary model. We also study the probability that this flare is just an oversized example of intraday variability using the Krakow data set of intense monitoring between 2015 and 2023. We find that the occurrence of a flare of this size and rapidity is unlikely. In machine-readable Tables 1 and 2, we give the full orbit-linked historical light curve of OJ287 as well as the dense monitoring sample of Krakow.
Anti-Universe And Emergent Gravity and the Dark UniverseSérgio Sacani
Recent theoretical progress indicates that spacetime and gravity emerge together from the entanglement structure of an underlying microscopic theory. These ideas are best understood in Anti-de Sitter space, where they rely on the area law for entanglement entropy. The extension to de Sitter space requires taking into account the entropy and temperature associated with the cosmological horizon. Using insights from string theory, black hole physics and quantum information theory we argue that the positive dark energy leads to a thermal volume law contribution to the entropy that overtakes the area law precisely at the cosmological horizon. Due to the competition between area and volume law entanglement the microscopic de Sitter states do not thermalise at sub-Hubble scales: they exhibit memory effects in the form of an entropy displacement caused by matter. The emergent laws of gravity contain an additional ‘dark’ gravitational force describing the ‘elastic’ response due to the entropy displacement. We derive an estimate of the strength of this extra force in terms of the baryonic mass, Newton’s constant and the Hubble acceleration scale a0 = cH0, and provide evidence for the fact that this additional ‘dark gravity force’ explains the observed phenomena in galaxies and clusters currently attributed to dark matter.
Presentation of our paper, "Towards Quantitative Evaluation of Explainable AI Methods for Deepfake Detection", by K. Tsigos, E. Apostolidis, S. Baxevanakis, S. Papadopoulos, V. Mezaris. Presented at the ACM Int. Workshop on Multimedia AI against Disinformation (MAD’24) of the ACM Int. Conf. on Multimedia Retrieval (ICMR’24), Thailand, June 2024. https://doi.org/10.1145/3643491.3660292 https://arxiv.org/abs/2404.18649
Software available at https://github.com/IDT-ITI/XAI-Deepfakes
Managing Uncertainties in Hardware-Software Codesign Projects
1. Managing Uncertainties in
Hardware-Software Codesign Projects
Jones Albuquerque (DEINFO-UFRPE)
Ferrer-Savall, J. (ESAB-UPC)
Bocanegra, S. (DEINFO-UFRPE)
Ferreira, T. (DEINFO-UFRPE)
Lopez-Codina, D. (ESAB-UPC)
Coelho Jr, C. (DCC-UFMG)
joa@deinfo.ufrpe.br
July 17, 2014
2. Outline
• The problem
• Related works
• Stochastic model based on system views
• Design scenario analysis
• Some results
• Conclusions and future works
1
3. Conceptual Level Design
Conceptual System Description
System Specification
Internal System Representation
Codesign Partitioning
Software Behavior Hardware Behavior
RTL, ASIC, DSP, Standard Cell, ...Assembly, C, Java, ...
Traditional
Methodologies
Codesign
2
5. Team Space
DESIGN
TOOLS
TARGET
DESIGN
SPACE
general−purpose
DSP system
C
C++
JAVA
TECHNOLOGIES
core−basic ASIC
Standard cell
Assembly
Multiprocessor
FPGA
Min cut
Simulated Annealing
manual partitioning
ILP
Scheduling Blocks
Formal verification
CONSTRAINTS
OPTIMIZATIONCost
Area
Power Development time
Reuse
Latency
Re−programability
TEAMS
Development risk
Changes in teams
Task assignment
Team load
Design times
Group communication
Skills
Technologies + Constraints + Tools + TEAMS
4
6. The Problem
Analysis + Partitioning
Conceptual Design
Task Assignment Task Implementation
Team Technology Technology Implementation
(HW/SW Codesign)(Team Assignment)
5
7. Related Works/Tools
• Castle, Chinook, COMET, Cosmos, COSYMA, CoWare, LY-
COS, Polis, PISH, Ptolemy, Symphony, Vulcan
• COSYMA, Chinook, and LYCOS present estimated analysis
and quality profiles
• PISH makes formal verification of the synthesis process
• CASTLE considers teams as designers
• COMET does not require a functionally complete description
in its specification
6
8. Neglected Aspects
• System is implemented by teams detaining specific technolo-
gies
• No single person has a complete knowledge of the system
• Early design scenarios and risk analysis
• Team load and task assignment
• Process is not adjusted neither tuned during development
7
9. • IEEE/ACM CODES, group discussion:
”This part of system design is very guru-intense and
tool support is scarse”
10. Outline
• The problem
• Related works
• Stochastic model based on system views
• Design scenario analysis
• Some results
• Conclusions and future works
8
11. Team Estimates
• PSP (Personal Software Process, 1995), TSP (Team Soft-
ware Process, 2000)
• PROBE for past task tracking based on linear regression
(range and variance)
• Formally, given a set of historical data for variables x and y,
to determine a likely value yk based on a known or estimated
new value xk we use
yk = β0 + xkβ1,
where the estimating parameters β0 and β1 are calculated
from the historical data
9
12. Probabilistic Characterization
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
-4 -3 -2 -1 0 1 2 3 4
mu = 0 & sigma = 2/3
mu = 0 & sigma = 1
mu = 0 & sigma = 2
Confidence Degree Standard
Deviation
(σ)
Statistical Seman-
tic
Very High (VH) 3σ = M − µ 99.7% of the values
are in the estimated
range.
High (H) 2σ = M − µ 95.5% of the values
are in the estimated
range.
Low(L) σ = M − µ 68.3% of the values
are in the estimated
range.
10
13. System Described by Views
• Hierarchical view: static objects and dependencies
• Sequencing view: dynamic constraints
• Development view: development dependencies
11
21. The Design Process using System Views
1. The system is described as a collection of interconnected objects (hierarchical
view) in compliance with system’s functional specification;
2. The sequential and development views of the system are determined by the project
manager;
3. The main constraints of the system design are determined and annotated in
their respective views;
4. The system-level model is given to the development teams;
5. The development teams estimate values for each object and method considering
the technology options and the team’s skill;
6. For each design scenario:
The automatic tool will choose the best implementation con-
sidering the
objective function, that must be optimized (cost function), and
the risk
of success for satisfying the design constraints (inference de-
grees).
18
23. Design Scenarios
• Early design options
• Risk analysis without implementation costs
• Time-to-market
• Implementation cost
• Execution time
• Team load
Mathematical model incorporating system views, team estimates, and risk analysis
20
24. Mathematical Formulation
SILP - Stochastic Integer Linear Problem
min{
∑
i,j
cijxij}
s.t.
Prob{
∑
j
aijxij ≤ bi} ≥ 1 − αi,
where xij ∈ {0, 1}, aij, bi and cij are random variables, and 0 ≤ αi ≤ 1 represents the
uncertainty that the equation will be satisfied.
21
25. Constraints: only one initial/final week
• There is only one initial (γijk) and final (Γijk) week for each
object i implemented by team j, i.e.
∀i ∈ Objects, j ∈ Teams
∑
k∈Weeks
γijk = xij;
∀i ∈ Objects, j ∈ Teams
∑
k∈Weeks
Γijk = xij.
22
26. Constraints: path execution time
• Path execution time. For each path (p) in SG Graph (se-
quencing view), the estimated execution time of the meth-
ods implemented by teams (dmj) in the path must satisfy a
maximum execution time (DIp) with a defined probability of
certainty (1 − αpath). Each method m of object i in path
p ∈ PathsE is represented by a 3-tuple (i, p, m), i.e.
∀P ∈ PathsE, j ∈ Teams
Prob{
∑
(i,p,m)∈P
dmjxij ≤ DIp} ≥ 1 − αpath
23
27. Constraints: team load
• Maximum load for team j (k1 is the first week). At week k1: team j can be implementing
an object i or not, i.e.
∀i ∈ Objects, j ∈ Teams aijk1 = γijk1 − Γijk1
where, ∀ijΓijk1 = 0. At other weeks: team j can be implementing object i if it begins
in this week (γijk = 1) or it is already implementing it (aij(k−1) = 1) and it did not end
yet (Γijk = 0), i.e.
∀i ∈ Objects, j ∈ Teams, k ∈ Weeks aijk = aij(k−1) + γijk − Γijk
At each week, the maximum load of a team (Λj) must be satisfied, i.e. the entire effort
wasted on all objects by a team at a week must be less or equal to the maximum load
of this team with a defined probability of certainty (1 − αload). Thus,
∀j ∈ Teams, k ∈ Weeks Prob{
∑
i∈Objects
λijaijk ≤ Λj} ≥ 1 − αload
24
28. Constraints: object assignment
• Object assignment. The final week (Γijk) of an implemen-
tation is greater or equal the development time (tij) plus
the initial week (γijk) with a defined probability of certainty
(1 − αassignment), i.e.
∀i ∈ Objects, j ∈ Teams
Prob{
∑
k∈Weeks
kΓijk ≥ tijxij +
∑
k∈Weeks
kγijk} ≥ 1 − αassignment.
25
29. Solving SILP
• Exact solution: uniform (0-1 ILP with quadratic number of original variables), gaussian
(non-linear)
• Traditional approaches: expected value formulation (EVF) [Ermolieve,1988]
• All estimates will be under/overestimated in the same way
– each team tends to mantain the same direction on estimates
– uncertainty is high on early stages of design
26
30. Solving SILP: Extending EVF
• Extended expected value formulation for scenario represen-
tation
• The approximation is
min{
∑
i,j
Prob{cij ≥ 1 − αo}xij}
s.t.
∑
j
Prob{aij ≥ 1 − αi}xij ≤ bi
27
32. The Scenario Formulation
• Extending the approximation for design scenarios, and con-
sidering a1 and a2 normally distributed, the constraint
Prob{a1x1 + a2x2 ≤ b} ≥ 1 − α
can be replaced by
(µa1 + F−1(1 − α)σa1)x1 + (µa2 + F−1(1 − α)σa2)x2 ≤ b
29
33. Design Scenarios Revisited
• Early system design views based on graphs (hierarchical, sequencing - SG, development
- DG)
• Team estimates using PROBE/PSP and confidence degrees (µ, σ)
• Design scenarios varying the objective function (min{
∑
i,j
cijxij})
• Risk analysis varying the α parameter in aij = (µaij + F−1(1 − αi)σaij )
• Partitioning analysis varying constraint inequations and values
30
34. Outline
• The problem
• Related works
• Stochastic model based on system views
• Design scenario analysis
• Some results
• Conclusions and future works
31
40. Clustering
Risk NetWork Controller - Objects (min Dev Time)
rcvd xmit enqueue exec bit buffer frame xmit b xmit f
0.5 tm1 tm1 tm1 tm1 tm1 tm2 tm3 tm4 tm4
0.8 tm3 tm3 tm3 tm3 tm3 tm6 tm3 tm4 tm4
0.97 tm3 tm3 tm3 tm3 tm3 tm5 tm3 tm4 tm4
0.99 tm6 tm3 tm3 tm4 tm3 tm4 tm4 tm4 tm4
Clustering team assignment when minimizing System Cost
37
41. Clustering
Risk NetWork Controller - Objects (min Dev Time)
rcvd xmit enqueue exec bit buffer frame xmit b xmit f
0.5 tm6 tm1 tm6 tm4 tm1 tm6 tm6 tm6 tm1
0.8 tm6 tm1 tm6 tm4 tm5 tm6 tm5 tm4 tm5
0.97 tm3 tm1 tm6 tm4 tm3 tm6 tm5 tm5 tm5
0.99 tm6 tm3 tm3 tm4 tm3 tm6 tm5 tm5 tm6
Clustering team assignment when minimizing Execution Time
38
42. Convergence Analysis
1 2 6 9 11 i
P(i)
i
P(i)
2 2 3 4 1195
Minimizing Development Time and Execution Time
39
43. Computational Parameters
Times for first iteration
Case-Study Num.variables Num.constraints Iterations USER Time (h:m:s)
MRF 2852 1089 13890 00:01:02
NET 3965 1462 38576 00:03:32
SAR 7260 2690 187808 01:16:21
Constraints = O(objects × weeks × teams)
40
44. Conclusions
• A stochastic model to system design by teams and multiple
technologies on conceptual level
• Interactive partitioning using design manager decisions based
on clustering scenarios
• Clustering scenarios to refine the process and minimize the
SILP problem at each iteration, reducing the computational
time
41
45. Contributions
• A stochastic model for system design by teams
• An approach with confidence degrees to model risk analysis and team aspects
• System description by views on early stages (conceptual level design)
• A mathematical formulation for partitioning with multiple technologies
• An extended expected value formulation to represent design scenarios
• Interactive and iterative partitioning using clustering to refine the process and to reduce
the SILP problem
• Design scenarios representing partitioning options, uncertainty, team aspects and design
constraints
42
46. At Work
• Applying in real industry: PINTO FILHO, K. V. ; BOCANEGRA, S. ; ALBUQUERQUE,
J. . Uma proposta de modelagem para o calculo de reservas de contingencia e gerencial
em projetos de software usando programacao linear inteira estocastica DOI 10.5752/P.2316-
9451.2012v1n1p50. Abakos, v. 1, p. 50-74, 2012.
• Integrating conceptual level design in a commercial tool (FACEPE - 2013-2015)
• Solving SILP: other approximations, Interior Points, and parallel strategies
• Other mathematical formulations for the conceptual level design problem and heuristics
43
47. Aknowledgements
• This work was [partially] supported by the National Insti-
tute of Science and Technology for Software Engineering
(www.ines.org.br), funded by CNPq and FACEPE, grants
573964/2008-4 and APQ-1037-1.03/08.
44