The document describes a new multi-vehicle trajectory generator (MTG) that can generate realistic vehicle encounter scenarios by encoding real encounter data into an interpretable latent space representation. The MTG uses a bi-directional encoder with GRU modules to extract temporal features from encounter trajectories, and a multi-branch decoder to separately generate multiple vehicle trajectories while sharing hidden states. A new disentanglement metric is proposed to evaluate deep generative models based on the variance of latent codes under different noise levels, without requiring classifiers. The MTG is shown to generate more rational vehicle encounters than baseline models and allow controlling encounters through the disentangled latent codes.
Effective Multi-Stage Training Model for Edge Computing Devices in Intrusion ...IJCNCJournal
Intrusion detection poses a significant challenge within expansive and persistently interconnected environments. As malicious code continues to advance and sophisticated attack methodologies proliferate, various advanced deep learning-based detection approaches have been proposed. Nevertheless, the complexity and accuracy of intrusion detection models still need further enhancement to render them more adaptable to diverse system categories, particularly within resource-constrained devices, such as those embedded in edge computing systems. This research introduces a three-stage training paradigm, augmented by an enhanced pruning methodology and model compression techniques. The objective is to elevate the system's effectiveness, concurrently maintaining a high level of accuracy for intrusion detection. Empirical assessments conducted on the UNSW-NB15 dataset evince that this solution notably reduces the model's dimensions, while upholding accuracy levels equivalent to similar proposals.
Effective Multi-Stage Training Model for Edge Computing Devices in Intrusion ...IJCNCJournal
Intrusion detection poses a significant challenge within expansive and persistently interconnected environments. As malicious code continues to advance and sophisticated attack methodologies proliferate, various advanced deep learning-based detection approaches have been proposed. Nevertheless, the complexity and accuracy of intrusion detection models still need further enhancement to render them more adaptable to diverse system categories, particularly within resource-constrained devices, such as those embedded in edge computing systems. This research introduces a three-stage training paradigm, augmented by an enhanced pruning methodology and model compression techniques. The objective is to elevate the system's effectiveness, concurrently maintaining a high level of accuracy for intrusion detection. Empirical assessments conducted on the UNSW-NB15 dataset evince that this solution notably reduces the model's dimensions, while upholding accuracy levels equivalent to similar proposals.
Comparative study of optimization algorithms on convolutional network for aut...IJECEIAES
This document compares the performance of 8 optimization algorithms (SGD, RMSprop, Adagrad, Adadelta, Adam, Adamax, Nadam, Ftrl) when training a ResNet convolutional neural network on an autonomous driving dataset with 11 categories of vehicle locations. Preliminary results found SGD performed best while Ftrl performed worst, though more analysis is needed to determine the optimal algorithm. The network was trained on 20 epochs with images from the PandaSet database to classify vehicle position using features extracted from front camera images.
Predictive maintenance of electromechanical systems based on enhanced generat...IAESIJAI
Predictive maintenance (PdM) is a cost-cutting method that involves avoiding breakdowns and production losses. Deep learning (DL) algorithms can be used for defect prediction and diagnostics due to the huge amount of data generated by the integration of analog and digital systems in manufacturing operations. To improve the predictive maintenance strategy, this study uses a hybrid of the convolutional neural network (CNN) and conditional generative adversarial neural network (CGAN) model. The proposed CNN-CGAN algorithm improves forecast accuracy while substantially reducing model complexity. A comparison with standalone CGAN utilizing a public dataset is performed to evaluate the proposed model. The results show that the proposed CNN-CGAN model outperforms the conditional GAN (CGAN) in terms of prediction accuracy. The average F-Score is increased from 97.625% for the CGAN to 100% for the CNN-CGAN.
Machine learning in Dynamic Adaptive Streaming over HTTP (DASH)Eswar Publications
Recently machine learning has been introduced into the area of adaptive video streaming. This paper explores a novel taxonomy that includes six state of the art techniques of machine learning that have been applied to Dynamic Adaptive Streaming over HTTP (DASH): (1) Q-learning, (2) Reinforcement learning, (3) Regression, (4) Classification, (5) Decision Tree learning, and (6) Neural networks.
Residual balanced attention network for real-time traffic scene semantic segm...IJECEIAES
Intelligent transportation systems (ITS) are among the most focused research in this century. Actually, autonomous driving provides very advanced tasks in terms of road safety monitoring which include identifying dangers on the road and protecting pedestrians. In the last few years, deep learning (DL) approaches and especially convolutional neural networks (CNNs) have been extensively used to solve ITS problems such as traffic scene semantic segmentation and traffic signs classification. Semantic segmentation is an important task that has been addressed in computer vision (CV). Indeed, traffic scene semantic segmentation using CNNs requires high precision with few computational resources to perceive and segment the scene in real-time. However, we often find related work focusing only on one aspect, the precision, or the number of computational parameters. In this regard, we propose RBANet, a robust and lightweight CNN which uses a new proposed balanced attention module, and a new proposed residual module. Afterward, we have simulated our proposed RBANet using three loss functions to get the best combination using only 0.74M parameters. The RBANet has been evaluated on CamVid, the most used dataset in semantic segmentation, and it has performed well in terms of parameters’ requirements and precision compared to related work.
Performance Analysis of Bfsk Multi-Hop Communication Systems Over K-μ Fading ...ijwmn
Multi-hop communication systems gained popularity in wireless communications; they can be used to
extend the coverage of the network and reduce the transmitted power. The transmission of data from the
source node to the destination node in multi-hop communications undergoes through intermediate relay
nodes. In this paper, we study the performance of multi-hop communication systems, in terms of average bit
error rate (BER) with Binary frequency shift keying assuming the κ-µ fading channel model. Due to the
difficulty in finding the probability density function (PDF) of the end-to-end signal to noise ratio (SNR) and
hence for the performance metrics, we use Gaussian Mixture (GM) approximation technique to
approximate the PDF of the end to end SNR assuming the κ-µ fading models as weighted sums of Gaussian
distributions. Numerical results are provided for the BER of binary frequency shift keying (BFSK) of
amplify and forward (AF) multi-hop communication systems assuming different values for the fading
parameters (, ) and for different number of hops. Numerical results are validated by comparing them
with simulation results.
RunPool: A Dynamic Pooling Layer for Convolution Neural NetworkPutra Wanda
Deep learning (DL) has achieved a significant performance in computer vision problems, mainly in automatic feature extraction and representation. However, it is not easy to determine the best pooling method in a different case study. For instance, experts can implement the best types of pooling in image processing cases, which might not be optimal for various tasks. Thus, it is
required to keep in line with the philosophy of DL. In dynamic neural network architecture, it is not practically possible to find
a proper pooling technique for the layers. It is the primary reason why various pooling cannot be applied in the dynamic and multidimensional dataset. To deal with the limitations, it needs to construct an optimal pooling method as a better option than max pooling and average pooling. Therefore, we introduce a dynamic pooling layer called RunPool to train the convolutional
neuralnetwork(CNN)architecture.RunPoolpoolingisproposedtoregularizetheneuralnetworkthatreplacesthedeterministic
pooling functions. In the final section, we test the proposed pooling layer to address classification problems with online social network (OSN) dataset
Effective Multi-Stage Training Model for Edge Computing Devices in Intrusion ...IJCNCJournal
Intrusion detection poses a significant challenge within expansive and persistently interconnected environments. As malicious code continues to advance and sophisticated attack methodologies proliferate, various advanced deep learning-based detection approaches have been proposed. Nevertheless, the complexity and accuracy of intrusion detection models still need further enhancement to render them more adaptable to diverse system categories, particularly within resource-constrained devices, such as those embedded in edge computing systems. This research introduces a three-stage training paradigm, augmented by an enhanced pruning methodology and model compression techniques. The objective is to elevate the system's effectiveness, concurrently maintaining a high level of accuracy for intrusion detection. Empirical assessments conducted on the UNSW-NB15 dataset evince that this solution notably reduces the model's dimensions, while upholding accuracy levels equivalent to similar proposals.
Effective Multi-Stage Training Model for Edge Computing Devices in Intrusion ...IJCNCJournal
Intrusion detection poses a significant challenge within expansive and persistently interconnected environments. As malicious code continues to advance and sophisticated attack methodologies proliferate, various advanced deep learning-based detection approaches have been proposed. Nevertheless, the complexity and accuracy of intrusion detection models still need further enhancement to render them more adaptable to diverse system categories, particularly within resource-constrained devices, such as those embedded in edge computing systems. This research introduces a three-stage training paradigm, augmented by an enhanced pruning methodology and model compression techniques. The objective is to elevate the system's effectiveness, concurrently maintaining a high level of accuracy for intrusion detection. Empirical assessments conducted on the UNSW-NB15 dataset evince that this solution notably reduces the model's dimensions, while upholding accuracy levels equivalent to similar proposals.
Comparative study of optimization algorithms on convolutional network for aut...IJECEIAES
This document compares the performance of 8 optimization algorithms (SGD, RMSprop, Adagrad, Adadelta, Adam, Adamax, Nadam, Ftrl) when training a ResNet convolutional neural network on an autonomous driving dataset with 11 categories of vehicle locations. Preliminary results found SGD performed best while Ftrl performed worst, though more analysis is needed to determine the optimal algorithm. The network was trained on 20 epochs with images from the PandaSet database to classify vehicle position using features extracted from front camera images.
Predictive maintenance of electromechanical systems based on enhanced generat...IAESIJAI
Predictive maintenance (PdM) is a cost-cutting method that involves avoiding breakdowns and production losses. Deep learning (DL) algorithms can be used for defect prediction and diagnostics due to the huge amount of data generated by the integration of analog and digital systems in manufacturing operations. To improve the predictive maintenance strategy, this study uses a hybrid of the convolutional neural network (CNN) and conditional generative adversarial neural network (CGAN) model. The proposed CNN-CGAN algorithm improves forecast accuracy while substantially reducing model complexity. A comparison with standalone CGAN utilizing a public dataset is performed to evaluate the proposed model. The results show that the proposed CNN-CGAN model outperforms the conditional GAN (CGAN) in terms of prediction accuracy. The average F-Score is increased from 97.625% for the CGAN to 100% for the CNN-CGAN.
Machine learning in Dynamic Adaptive Streaming over HTTP (DASH)Eswar Publications
Recently machine learning has been introduced into the area of adaptive video streaming. This paper explores a novel taxonomy that includes six state of the art techniques of machine learning that have been applied to Dynamic Adaptive Streaming over HTTP (DASH): (1) Q-learning, (2) Reinforcement learning, (3) Regression, (4) Classification, (5) Decision Tree learning, and (6) Neural networks.
Residual balanced attention network for real-time traffic scene semantic segm...IJECEIAES
Intelligent transportation systems (ITS) are among the most focused research in this century. Actually, autonomous driving provides very advanced tasks in terms of road safety monitoring which include identifying dangers on the road and protecting pedestrians. In the last few years, deep learning (DL) approaches and especially convolutional neural networks (CNNs) have been extensively used to solve ITS problems such as traffic scene semantic segmentation and traffic signs classification. Semantic segmentation is an important task that has been addressed in computer vision (CV). Indeed, traffic scene semantic segmentation using CNNs requires high precision with few computational resources to perceive and segment the scene in real-time. However, we often find related work focusing only on one aspect, the precision, or the number of computational parameters. In this regard, we propose RBANet, a robust and lightweight CNN which uses a new proposed balanced attention module, and a new proposed residual module. Afterward, we have simulated our proposed RBANet using three loss functions to get the best combination using only 0.74M parameters. The RBANet has been evaluated on CamVid, the most used dataset in semantic segmentation, and it has performed well in terms of parameters’ requirements and precision compared to related work.
Performance Analysis of Bfsk Multi-Hop Communication Systems Over K-μ Fading ...ijwmn
Multi-hop communication systems gained popularity in wireless communications; they can be used to
extend the coverage of the network and reduce the transmitted power. The transmission of data from the
source node to the destination node in multi-hop communications undergoes through intermediate relay
nodes. In this paper, we study the performance of multi-hop communication systems, in terms of average bit
error rate (BER) with Binary frequency shift keying assuming the κ-µ fading channel model. Due to the
difficulty in finding the probability density function (PDF) of the end-to-end signal to noise ratio (SNR) and
hence for the performance metrics, we use Gaussian Mixture (GM) approximation technique to
approximate the PDF of the end to end SNR assuming the κ-µ fading models as weighted sums of Gaussian
distributions. Numerical results are provided for the BER of binary frequency shift keying (BFSK) of
amplify and forward (AF) multi-hop communication systems assuming different values for the fading
parameters (, ) and for different number of hops. Numerical results are validated by comparing them
with simulation results.
RunPool: A Dynamic Pooling Layer for Convolution Neural NetworkPutra Wanda
Deep learning (DL) has achieved a significant performance in computer vision problems, mainly in automatic feature extraction and representation. However, it is not easy to determine the best pooling method in a different case study. For instance, experts can implement the best types of pooling in image processing cases, which might not be optimal for various tasks. Thus, it is
required to keep in line with the philosophy of DL. In dynamic neural network architecture, it is not practically possible to find
a proper pooling technique for the layers. It is the primary reason why various pooling cannot be applied in the dynamic and multidimensional dataset. To deal with the limitations, it needs to construct an optimal pooling method as a better option than max pooling and average pooling. Therefore, we introduce a dynamic pooling layer called RunPool to train the convolutional
neuralnetwork(CNN)architecture.RunPoolpoolingisproposedtoregularizetheneuralnetworkthatreplacesthedeterministic
pooling functions. In the final section, we test the proposed pooling layer to address classification problems with online social network (OSN) dataset
COMPARATIVE PERFORMANCE ANALYSIS OF RNSC AND MCL ALGORITHMS ON POWER-LAW DIST...acijjournal
Cluster analysis of graph related problems is an important issue now-a-day. Different types of graph
clustering techniques are appeared in the field but most of them are vulnerable in terms of effectiveness
and fragmentation of output in case of real-world applications in diverse systems. In this paper, we will
provide a comparative behavioural analysis of RNSC (Restricted Neighbourhood Search Clustering) and
MCL (Markov Clustering) algorithms on Power-Law Distribution graphs. RNSC is a graph clustering
technique using stochastic local search. RNSC algorithm tries to achieve optimal cost clustering by
assigning some cost functions to the set of clusterings of a graph. This algorithm was implemented by A.
D. King only for undirected and unweighted random graphs. Another popular graph clustering
algorithm MCL is based on stochastic flow simulation model for weighted graphs. There are plentiful
applications of power-law or scale-free graphs in nature and society. Scale-free topology is stochastic i.e.
nodes are connected in a random manner. Complex network topologies like World Wide Web, the web of
human sexual contacts, or the chemical network of a cell etc., are basically following power-law
distribution to represent different real-life systems. This paper uses real large-scale power-law
distribution graphs to conduct the performance analysis of RNSC behaviour compared with Markov
clustering (MCL) algorithm. Extensive experimental results on several synthetic and real power-law
distribution datasets reveal the effectiveness of our approach to comparative performance measure of
these algorithms on the basis of cost of clustering, cluster size, modularity index of clustering results and
normalized mutual information (NMI).
Fault-Tolerance Aware Multi Objective Scheduling Algorithm for Task Schedulin...csandit
Computational Grid (CG) creates a large heterogeneous and distributed paradigm to manage and execute the applications which are computationally intensive. In grid scheduling tasks are assigned to the proper processors in the grid system to for its execution by considering the execution policy and the optimization objectives. In this paper, makespan and the faulttolerance of the computational nodes of the grid which are the two important parameters for the task execution, are considered and tried to optimize it. As the grid scheduling is considered to be NP-Hard, so a meta-heuristics evolutionary based techniques are often used to find a solution for this. We have proposed a NSGA II for this purpose. The performance estimation ofthe proposed Fault tolerance Aware NSGA II (FTNSGA II) has been done by writing program in Matlab. The simulation results evaluates the performance of the all proposed algorithm and the results of proposed model is compared with existing model Min-Min and Max-Min algorithm which proves effectiveness of the model.
PREDICTION OF AVERAGE TOTAL PROJECT DURATION USING ARTIFICIAL NEURAL NETWORKS...IAEME Publication
The prediction of project‘s expectancy life is an important issue for entrepreneurs since it helps them to avoid the expiration time of projects. To properly address this issue, Neural Network-based approach, fuzzy logic and regression methods are used to predict the necessary time that can be consumed to put an end to the targeted project. Before applying the three aforementioned approaches, the modeling and simulation of the activities network are introduced for calculating the total average time of project. Then, comparatively speaking, the neural network, fuzzy logic and regression method approach are compared in terms prediction’s accuracy. The generated error from the three methods is compared, namely different types of errors are calculated. Basically, the input variables consist of the probability of success (PS), the coefficient of improvement (Coef_PS) and the coefficient of learning (CofA), while the output variable is the average total duration of the project (DTTm). The Predicted mean square error (MSE) values are purposefully used to compare the three models. Interestingly, the results show that the optimum prediction model is the fuzzy logic model with accurate results. It is noteworthy to say that the application in this paper can be applied on a real case study.
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...IAEME Publication
This paper presents an approach based on applying an aggregated predictor formed by multiple versions of a multilayer neural network with a back-propagation optimization algorithm for helping the engineer to get a list of the most appropriate well-test interpretation models for a given set of pressure/ production data. The proposed method consists of three stages: (1) data decorrelation through principal component analysis to reduce the covariance between the variables and the dimension of the input layer in the artificial neural network, (2) bootstrap replicates of the learning set where the data is repeatedly sampled with a random split of the data into train sets and using these as new learning sets, and (3) automatic reservoir model identification through aggregated predictor formed by a plurality vote when predicting a new class. This method is described in detail to ensure successful replication of results. The required training and test dataset were generated by using analytical solution models. In our case, there were used 600 samples: 300 for training, 100 for cross-validation, and 200 for testing. Different network structures were tested during this study to arrive at optimum network design. We notice that the single net methodology always brings about confusion in selecting the correct model even though the training results for the constructed networks are close to 1. We notice also that the principal component analysis is an effective strategy in reducing the number of input features, simplifying the network structure, and lowering the training time of the ANN. The results obtained show that the proposed model provides better performance when predicting new data with a coefficient of correlation approximately equal to 95% Compared to a previous approach 80%, the combination of the PCA and ANN is more stable and determine the more accurate results with lesser computational complexity than was feasible previously. Clearly, the aggregated predictor is more stable and shows less bad classes compared to the previous approach.
ESTIMATING PROJECT DEVELOPMENT EFFORT USING CLUSTERED REGRESSION APPROACHcscpconf
Due to the intangible nature of “software”, accurate and reliable software effort estimation is a challenge in the software Industry. It is unlikely to expect very accurate estimates of software
development effort because of the inherent uncertainty in software development projects and the complex and dynamic interaction of factors that impact software development. Heterogeneity exists in the software engineering datasets because data is made available from diverse sources.
This can be reduced by defining certain relationship between the data values by classifying them into different clusters. This study focuses on how the combination of clustering and
regression techniques can reduce the potential problems in effectiveness of predictive efficiency due to heterogeneity of the data. Using a clustered approach creates the subsets of data having a degree of homogeneity that enhances prediction accuracy. It was also observed in this study that ridge regression performs better than other regression techniques used in the analysis.
Estimating project development effort using clustered regression approachcsandit
Due to the intangible nature of “software”, accurate and reliable software effort estimation is a
challenge in the software Industry. It is unlikely to expect very accurate estimates of software
development effort because of the inherent uncertainty in software development projects and the
complex and dynamic interaction of factors that impact software development. Heterogeneity
exists in the software engineering datasets because data is made available from diverse sources.
This can be reduced by defining certain relationship between the data values by classifying
them into different clusters. This study focuses on how the combination of clustering and
regression techniques can reduce the potential problems in effectiveness of predictive efficiency
due to heterogeneity of the data. Using a clustered approach creates the subsets of data having
a degree of homogeneity that enhances prediction accuracy. It was also observed in this study
that ridge regression performs better than other regression techniques used in the analysis.
Here we describe federated learning based traffic flow prediction system. In federated learning we solve the problem of data security and also provide collaborative learning. model parameter are shared here ,not data
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDMOD...ijgca
This document discusses modeling cloud computing data centers as queuing systems to analyze performance factors. It begins with background on cloud computing and queuing theory. It then models a cloud data center as an [(M/G/1) : (∞/GDMODEL)] queuing system with single task arrivals and infinite task buffer capacity. Key performance factors analyzed include mean number of tasks in the system. Analytical results are obtained by solving the model to estimate response time distribution and other metrics. The modeling approach allows determining the relationship between performance and number of servers/buffer size.
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDMOD...ijgca
This document discusses modeling cloud computing data centers as queuing systems to analyze performance factors. It presents an analytical model of a cloud data center as a [(M/G/1) : (∞/GDMODEL)] queuing system with single task arrivals and infinite task buffer capacity. The model is solved to obtain important performance metrics like mean number of tasks in the system. Prior work on modeling cloud systems and queuing theory concepts are also reviewed. Key assumptions of the proposed model include tasks following a Poisson arrival process and service times having a general probability distribution.
A simple framework for contrastive learning of visual representationsDevansh16
Link: https://machine-learning-made-simple.medium.com/learnings-from-simclr-a-framework-contrastive-learning-for-visual-representations-6c145a5d8e99
If you'd like to discuss something, text me on LinkedIn, IG, or Twitter. To support me, please use my referral link to Robinhood. It's completely free, and we both get a free stock. Not using it is literally losing out on free money.
Check out my other articles on Medium. : https://rb.gy/zn1aiu
My YouTube: https://rb.gy/88iwdd
Reach out to me on LinkedIn. Let's connect: https://rb.gy/m5ok2y
My Instagram: https://rb.gy/gmvuy9
My Twitter: https://twitter.com/Machine01776819
My Substack: https://devanshacc.substack.com/
Live conversations at twitch here: https://rb.gy/zlhk9y
Get a free stock on Robinhood: https://join.robinhood.com/fnud75
This paper presents SimCLR: a simple framework for contrastive learning of visual representations. We simplify recently proposed contrastive self-supervised learning algorithms without requiring specialized architectures or a memory bank. In order to understand what enables the contrastive prediction tasks to learn useful representations, we systematically study the major components of our framework. We show that (1) composition of data augmentations plays a critical role in defining effective predictive tasks, (2) introducing a learnable nonlinear transformation between the representation and the contrastive loss substantially improves the quality of the learned representations, and (3) contrastive learning benefits from larger batch sizes and more training steps compared to supervised learning. By combining these findings, we are able to considerably outperform previous methods for self-supervised and semi-supervised learning on ImageNet. A linear classifier trained on self-supervised representations learned by SimCLR achieves 76.5% top-1 accuracy, which is a 7% relative improvement over previous state-of-the-art, matching the performance of a supervised ResNet-50. When fine-tuned on only 1% of the labels, we achieve 85.8% top-5 accuracy, outperforming AlexNet with 100X fewer labels.
Comments: ICML'2020. Code and pretrained models at this https URL
Subjects: Machine Learning (cs.LG); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (stat.ML)
Cite as: arXiv:2002.05709 [cs.LG]
(or arXiv:2002.05709v3 [cs.LG] for this version)
Submission history
From: Ting Chen [view email]
[v1] Thu, 13 Feb 2020 18:50:45 UTC (5,093 KB)
[v2] Mon, 30 Mar 2020 15:32:51 UTC (5,047 KB)
[v3] Wed, 1 Jul 2020 00:09:08 UTC (5,829 KB)
Deep Convolutional Neural Networks (CNNs) have achieved impressive performance in
edge detection tasks, but their large number of parameters often leads to high memory and energy
costs for implementation on lightweight devices. In this paper, we propose a new architecture, called
Efficient Deep-learning Gradients Extraction Network (EDGE-Net), that integrates the advantages of Depthwise Separable Convolutions and deformable convolutional networks (DeformableConvNet) to address these inefficiencies. By carefully selecting proper components and utilizing
network pruning techniques, our proposed EDGE-Net achieves state-of-the-art accuracy in edge
detection while significantly reducing complexity. Experimental results on BSDS500 and NYUDv2
datasets demonstrate that EDGE-Net outperforms current lightweight edge detectors with only
500k parameters, without relying on pre-trained weights.
Deep Convolutional Neural Networks (CNNs) have achieved impressive performance in
edge detection tasks, but their large number of parameters often leads to high memory and energy
costs for implementation on lightweight devices. In this paper, we propose a new architecture, called
Efficient Deep-learning Gradients Extraction Network (EDGE-Net), that integrates the advantages of Depthwise Separable Convolutions and deformable convolutional networks (DeformableConvNet) to address these inefficiencies. By carefully selecting proper components and utilizing
network pruning techniques, our proposed EDGE-Net achieves state-of-the-art accuracy in edge
detection while significantly reducing complexity. Experimental results on BSDS500 and NYUDv2
datasets demonstrate that EDGE-Net outperforms current lightweight edge detectors with only
500k parameters, without relying on pre-trained weights.
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLINGijaia
Imbalanced data is a significant challenge in classification with machine learning algorithms. This is particularly important with log message data as negative logs are sparse so this data is typically imbalanced. In this paper, a model to generate text log messages is proposed which employs a SeqGAN network. An Autoencoder is used for feature extraction and anomaly detection is done using a GRU network. The proposed model is evaluated with three imbalanced log data sets, namely BGL, OpenStack, and Thunderbird. Results are presented which show that appropriate oversampling and data balancing
improves anomaly detection accuracy.
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLINGgerogepatton
Imbalanced data is a significant challenge in classification with machine learning algorithms. This is particularly important with log message data as negative logs are sparse so this data is typically imbalanced. In this paper, a model to generate text log messages is proposed which employs a SeqGAN network. An Autoencoder is used for feature extraction and anomaly detection is done using a GRU network. The proposed model is evaluated with three imbalanced log data sets, namely BGL, OpenStack, and Thunderbird. Results are presented which show that appropriate oversampling and data balancing improves anomaly detection accuracy.
A h k clustering algorithm for high dimensional data using ensemble learningijitcs
The document summarizes a proposed clustering algorithm for high dimensional data that combines hierarchical (H-K) clustering, subspace clustering, and ensemble clustering. It begins with background on challenges of clustering high dimensional data and related work applying dimension reduction, subspace clustering, ensemble clustering, and H-K clustering individually. The proposed model first applies subspace clustering to identify clusters within subsets of features. It then performs H-K clustering on each subspace cluster. Finally, it applies ensemble clustering techniques to integrate the results into a single clustering. The goal is to leverage each technique's strengths to improve clustering performance for high dimensional data compared to using a single approach.
This document describes an implementation of fast image convolution using Winograd's minimal filtering algorithm for 3x3 filters. The implementation combines C code with BLAS calls for GEMM. It is optimized for Intel Xeon Phi processors and uses Intel MKL for BLAS calls. Benchmarking shows the implementation achieves 10% greater overall performance than MKL convolution and can be up to 1.5x faster for some layers and up to 4x slower for others, indicating potential for a hybrid approach. High-bandwidth memory on Intel Xeon Phi significantly improves efficiency of fast convolution.
Massive parallelism with gpus for centrality ranking in complex networksijcsit
Many problems in Computer Science can be modelled using graphs. Evaluating node centrality in complex
networks, which can be considered equivalent to undirected graphs, provides an useful metric of the
relative importance of each node inside the evaluated network. The knowledge on which the most central
nodes are, has various applications, such as improving information spreading in diffusion networks. In this
case, most central nodes can be considered to have higher influence rates over other nodes in the network.
The main purpose in this work is developing a GPU based and massively parallel application so as to
evaluate the node centrality in complex networks using the Nvidia CUDA programming model. The main
contribution of this work is the strategies for the development of an algorithm to evaluate the node
centrality in complex networks using Nvidia CUDA parallel programming model. We show that the
strategies improves algorithm´s speed-up in two orders of magnitude on one NVIDIA Tesla k20 GPU
cluster node, when compared to the hybrid OpenMP/MPI algorithm version, running in the same cluster,
with 4 nodes 2 Intel(R) Xeon(R) CPU E5-2660 each, for radius zero
A frame work for clustering time evolving dataiaemedu
The document proposes a framework for clustering time-evolving categorical data using a sliding window technique. It uses an existing clustering algorithm (Node Importance Representative) and a Drifting Concept Detection algorithm to detect changes in cluster distributions between the current and previous data windows. If a threshold difference in clusters is exceeded, reclustering is performed on the new window. Otherwise, the new clusters are added to the previous results. The framework aims to improve on prior work by handling drifting concepts in categorical time-series data.
AN ENTROPIC OPTIMIZATION TECHNIQUE IN HETEROGENEOUS GRID COMPUTING USING BION...ijcsit
This document summarizes a research paper that proposes a new method for improving both fault tolerance and load balancing in grid computing networks. The method converts the tree structure of grid computing nodes into a distributed R-tree index structure and then applies an entropy estimation technique. This entropy estimation helps discard nodes with high entropy from the tree, reducing complexity. The method then uses thresholding and control algorithms to select optimal route paths based on load balance and fault tolerance. Various optimization techniques like genetic algorithms, ant colony optimization, and particle swarm optimization are also applied to reach better solutions. Experimental results showed the proposed method improved performance over other existing methods.
The impact of channel model on the performance of distance-based schemes in v...IJECEIAES
This document discusses three distance-based schemes - RNDN, EVNDN, and OIFP - for data dissemination in vehicular named data networks (VNDNs). It evaluates the impact of different propagation models (Friis, Nakagami, random) on the performance of these schemes. Simulation results show that the Nakagami model significantly degrades network performance and increases retransmissions for RNDN. EVNDN sees more retransmissions under Friis and random models. OIFP exhibits higher interest satisfaction ratios and lower delays compared to the other schemes under all models. The document concludes that propagation models have a noticeable effect on distance-based scheme performance in VNDNs.
COMPARATIVE PERFORMANCE ANALYSIS OF RNSC AND MCL ALGORITHMS ON POWER-LAW DIST...acijjournal
Cluster analysis of graph related problems is an important issue now-a-day. Different types of graph
clustering techniques are appeared in the field but most of them are vulnerable in terms of effectiveness
and fragmentation of output in case of real-world applications in diverse systems. In this paper, we will
provide a comparative behavioural analysis of RNSC (Restricted Neighbourhood Search Clustering) and
MCL (Markov Clustering) algorithms on Power-Law Distribution graphs. RNSC is a graph clustering
technique using stochastic local search. RNSC algorithm tries to achieve optimal cost clustering by
assigning some cost functions to the set of clusterings of a graph. This algorithm was implemented by A.
D. King only for undirected and unweighted random graphs. Another popular graph clustering
algorithm MCL is based on stochastic flow simulation model for weighted graphs. There are plentiful
applications of power-law or scale-free graphs in nature and society. Scale-free topology is stochastic i.e.
nodes are connected in a random manner. Complex network topologies like World Wide Web, the web of
human sexual contacts, or the chemical network of a cell etc., are basically following power-law
distribution to represent different real-life systems. This paper uses real large-scale power-law
distribution graphs to conduct the performance analysis of RNSC behaviour compared with Markov
clustering (MCL) algorithm. Extensive experimental results on several synthetic and real power-law
distribution datasets reveal the effectiveness of our approach to comparative performance measure of
these algorithms on the basis of cost of clustering, cluster size, modularity index of clustering results and
normalized mutual information (NMI).
Fault-Tolerance Aware Multi Objective Scheduling Algorithm for Task Schedulin...csandit
Computational Grid (CG) creates a large heterogeneous and distributed paradigm to manage and execute the applications which are computationally intensive. In grid scheduling tasks are assigned to the proper processors in the grid system to for its execution by considering the execution policy and the optimization objectives. In this paper, makespan and the faulttolerance of the computational nodes of the grid which are the two important parameters for the task execution, are considered and tried to optimize it. As the grid scheduling is considered to be NP-Hard, so a meta-heuristics evolutionary based techniques are often used to find a solution for this. We have proposed a NSGA II for this purpose. The performance estimation ofthe proposed Fault tolerance Aware NSGA II (FTNSGA II) has been done by writing program in Matlab. The simulation results evaluates the performance of the all proposed algorithm and the results of proposed model is compared with existing model Min-Min and Max-Min algorithm which proves effectiveness of the model.
PREDICTION OF AVERAGE TOTAL PROJECT DURATION USING ARTIFICIAL NEURAL NETWORKS...IAEME Publication
The prediction of project‘s expectancy life is an important issue for entrepreneurs since it helps them to avoid the expiration time of projects. To properly address this issue, Neural Network-based approach, fuzzy logic and regression methods are used to predict the necessary time that can be consumed to put an end to the targeted project. Before applying the three aforementioned approaches, the modeling and simulation of the activities network are introduced for calculating the total average time of project. Then, comparatively speaking, the neural network, fuzzy logic and regression method approach are compared in terms prediction’s accuracy. The generated error from the three methods is compared, namely different types of errors are calculated. Basically, the input variables consist of the probability of success (PS), the coefficient of improvement (Coef_PS) and the coefficient of learning (CofA), while the output variable is the average total duration of the project (DTTm). The Predicted mean square error (MSE) values are purposefully used to compare the three models. Interestingly, the results show that the optimum prediction model is the fuzzy logic model with accurate results. It is noteworthy to say that the application in this paper can be applied on a real case study.
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...IAEME Publication
This paper presents an approach based on applying an aggregated predictor formed by multiple versions of a multilayer neural network with a back-propagation optimization algorithm for helping the engineer to get a list of the most appropriate well-test interpretation models for a given set of pressure/ production data. The proposed method consists of three stages: (1) data decorrelation through principal component analysis to reduce the covariance between the variables and the dimension of the input layer in the artificial neural network, (2) bootstrap replicates of the learning set where the data is repeatedly sampled with a random split of the data into train sets and using these as new learning sets, and (3) automatic reservoir model identification through aggregated predictor formed by a plurality vote when predicting a new class. This method is described in detail to ensure successful replication of results. The required training and test dataset were generated by using analytical solution models. In our case, there were used 600 samples: 300 for training, 100 for cross-validation, and 200 for testing. Different network structures were tested during this study to arrive at optimum network design. We notice that the single net methodology always brings about confusion in selecting the correct model even though the training results for the constructed networks are close to 1. We notice also that the principal component analysis is an effective strategy in reducing the number of input features, simplifying the network structure, and lowering the training time of the ANN. The results obtained show that the proposed model provides better performance when predicting new data with a coefficient of correlation approximately equal to 95% Compared to a previous approach 80%, the combination of the PCA and ANN is more stable and determine the more accurate results with lesser computational complexity than was feasible previously. Clearly, the aggregated predictor is more stable and shows less bad classes compared to the previous approach.
ESTIMATING PROJECT DEVELOPMENT EFFORT USING CLUSTERED REGRESSION APPROACHcscpconf
Due to the intangible nature of “software”, accurate and reliable software effort estimation is a challenge in the software Industry. It is unlikely to expect very accurate estimates of software
development effort because of the inherent uncertainty in software development projects and the complex and dynamic interaction of factors that impact software development. Heterogeneity exists in the software engineering datasets because data is made available from diverse sources.
This can be reduced by defining certain relationship between the data values by classifying them into different clusters. This study focuses on how the combination of clustering and
regression techniques can reduce the potential problems in effectiveness of predictive efficiency due to heterogeneity of the data. Using a clustered approach creates the subsets of data having a degree of homogeneity that enhances prediction accuracy. It was also observed in this study that ridge regression performs better than other regression techniques used in the analysis.
Estimating project development effort using clustered regression approachcsandit
Due to the intangible nature of “software”, accurate and reliable software effort estimation is a
challenge in the software Industry. It is unlikely to expect very accurate estimates of software
development effort because of the inherent uncertainty in software development projects and the
complex and dynamic interaction of factors that impact software development. Heterogeneity
exists in the software engineering datasets because data is made available from diverse sources.
This can be reduced by defining certain relationship between the data values by classifying
them into different clusters. This study focuses on how the combination of clustering and
regression techniques can reduce the potential problems in effectiveness of predictive efficiency
due to heterogeneity of the data. Using a clustered approach creates the subsets of data having
a degree of homogeneity that enhances prediction accuracy. It was also observed in this study
that ridge regression performs better than other regression techniques used in the analysis.
Here we describe federated learning based traffic flow prediction system. In federated learning we solve the problem of data security and also provide collaborative learning. model parameter are shared here ,not data
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDMOD...ijgca
This document discusses modeling cloud computing data centers as queuing systems to analyze performance factors. It begins with background on cloud computing and queuing theory. It then models a cloud data center as an [(M/G/1) : (∞/GDMODEL)] queuing system with single task arrivals and infinite task buffer capacity. Key performance factors analyzed include mean number of tasks in the system. Analytical results are obtained by solving the model to estimate response time distribution and other metrics. The modeling approach allows determining the relationship between performance and number of servers/buffer size.
PERFORMANCE FACTORS OF CLOUD COMPUTING DATA CENTERS USING [(M/G/1) : (∞/GDMOD...ijgca
This document discusses modeling cloud computing data centers as queuing systems to analyze performance factors. It presents an analytical model of a cloud data center as a [(M/G/1) : (∞/GDMODEL)] queuing system with single task arrivals and infinite task buffer capacity. The model is solved to obtain important performance metrics like mean number of tasks in the system. Prior work on modeling cloud systems and queuing theory concepts are also reviewed. Key assumptions of the proposed model include tasks following a Poisson arrival process and service times having a general probability distribution.
A simple framework for contrastive learning of visual representationsDevansh16
Link: https://machine-learning-made-simple.medium.com/learnings-from-simclr-a-framework-contrastive-learning-for-visual-representations-6c145a5d8e99
If you'd like to discuss something, text me on LinkedIn, IG, or Twitter. To support me, please use my referral link to Robinhood. It's completely free, and we both get a free stock. Not using it is literally losing out on free money.
Check out my other articles on Medium. : https://rb.gy/zn1aiu
My YouTube: https://rb.gy/88iwdd
Reach out to me on LinkedIn. Let's connect: https://rb.gy/m5ok2y
My Instagram: https://rb.gy/gmvuy9
My Twitter: https://twitter.com/Machine01776819
My Substack: https://devanshacc.substack.com/
Live conversations at twitch here: https://rb.gy/zlhk9y
Get a free stock on Robinhood: https://join.robinhood.com/fnud75
This paper presents SimCLR: a simple framework for contrastive learning of visual representations. We simplify recently proposed contrastive self-supervised learning algorithms without requiring specialized architectures or a memory bank. In order to understand what enables the contrastive prediction tasks to learn useful representations, we systematically study the major components of our framework. We show that (1) composition of data augmentations plays a critical role in defining effective predictive tasks, (2) introducing a learnable nonlinear transformation between the representation and the contrastive loss substantially improves the quality of the learned representations, and (3) contrastive learning benefits from larger batch sizes and more training steps compared to supervised learning. By combining these findings, we are able to considerably outperform previous methods for self-supervised and semi-supervised learning on ImageNet. A linear classifier trained on self-supervised representations learned by SimCLR achieves 76.5% top-1 accuracy, which is a 7% relative improvement over previous state-of-the-art, matching the performance of a supervised ResNet-50. When fine-tuned on only 1% of the labels, we achieve 85.8% top-5 accuracy, outperforming AlexNet with 100X fewer labels.
Comments: ICML'2020. Code and pretrained models at this https URL
Subjects: Machine Learning (cs.LG); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (stat.ML)
Cite as: arXiv:2002.05709 [cs.LG]
(or arXiv:2002.05709v3 [cs.LG] for this version)
Submission history
From: Ting Chen [view email]
[v1] Thu, 13 Feb 2020 18:50:45 UTC (5,093 KB)
[v2] Mon, 30 Mar 2020 15:32:51 UTC (5,047 KB)
[v3] Wed, 1 Jul 2020 00:09:08 UTC (5,829 KB)
Deep Convolutional Neural Networks (CNNs) have achieved impressive performance in
edge detection tasks, but their large number of parameters often leads to high memory and energy
costs for implementation on lightweight devices. In this paper, we propose a new architecture, called
Efficient Deep-learning Gradients Extraction Network (EDGE-Net), that integrates the advantages of Depthwise Separable Convolutions and deformable convolutional networks (DeformableConvNet) to address these inefficiencies. By carefully selecting proper components and utilizing
network pruning techniques, our proposed EDGE-Net achieves state-of-the-art accuracy in edge
detection while significantly reducing complexity. Experimental results on BSDS500 and NYUDv2
datasets demonstrate that EDGE-Net outperforms current lightweight edge detectors with only
500k parameters, without relying on pre-trained weights.
Deep Convolutional Neural Networks (CNNs) have achieved impressive performance in
edge detection tasks, but their large number of parameters often leads to high memory and energy
costs for implementation on lightweight devices. In this paper, we propose a new architecture, called
Efficient Deep-learning Gradients Extraction Network (EDGE-Net), that integrates the advantages of Depthwise Separable Convolutions and deformable convolutional networks (DeformableConvNet) to address these inefficiencies. By carefully selecting proper components and utilizing
network pruning techniques, our proposed EDGE-Net achieves state-of-the-art accuracy in edge
detection while significantly reducing complexity. Experimental results on BSDS500 and NYUDv2
datasets demonstrate that EDGE-Net outperforms current lightweight edge detectors with only
500k parameters, without relying on pre-trained weights.
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLINGijaia
Imbalanced data is a significant challenge in classification with machine learning algorithms. This is particularly important with log message data as negative logs are sparse so this data is typically imbalanced. In this paper, a model to generate text log messages is proposed which employs a SeqGAN network. An Autoencoder is used for feature extraction and anomaly detection is done using a GRU network. The proposed model is evaluated with three imbalanced log data sets, namely BGL, OpenStack, and Thunderbird. Results are presented which show that appropriate oversampling and data balancing
improves anomaly detection accuracy.
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLINGgerogepatton
Imbalanced data is a significant challenge in classification with machine learning algorithms. This is particularly important with log message data as negative logs are sparse so this data is typically imbalanced. In this paper, a model to generate text log messages is proposed which employs a SeqGAN network. An Autoencoder is used for feature extraction and anomaly detection is done using a GRU network. The proposed model is evaluated with three imbalanced log data sets, namely BGL, OpenStack, and Thunderbird. Results are presented which show that appropriate oversampling and data balancing improves anomaly detection accuracy.
A h k clustering algorithm for high dimensional data using ensemble learningijitcs
The document summarizes a proposed clustering algorithm for high dimensional data that combines hierarchical (H-K) clustering, subspace clustering, and ensemble clustering. It begins with background on challenges of clustering high dimensional data and related work applying dimension reduction, subspace clustering, ensemble clustering, and H-K clustering individually. The proposed model first applies subspace clustering to identify clusters within subsets of features. It then performs H-K clustering on each subspace cluster. Finally, it applies ensemble clustering techniques to integrate the results into a single clustering. The goal is to leverage each technique's strengths to improve clustering performance for high dimensional data compared to using a single approach.
This document describes an implementation of fast image convolution using Winograd's minimal filtering algorithm for 3x3 filters. The implementation combines C code with BLAS calls for GEMM. It is optimized for Intel Xeon Phi processors and uses Intel MKL for BLAS calls. Benchmarking shows the implementation achieves 10% greater overall performance than MKL convolution and can be up to 1.5x faster for some layers and up to 4x slower for others, indicating potential for a hybrid approach. High-bandwidth memory on Intel Xeon Phi significantly improves efficiency of fast convolution.
Massive parallelism with gpus for centrality ranking in complex networksijcsit
Many problems in Computer Science can be modelled using graphs. Evaluating node centrality in complex
networks, which can be considered equivalent to undirected graphs, provides an useful metric of the
relative importance of each node inside the evaluated network. The knowledge on which the most central
nodes are, has various applications, such as improving information spreading in diffusion networks. In this
case, most central nodes can be considered to have higher influence rates over other nodes in the network.
The main purpose in this work is developing a GPU based and massively parallel application so as to
evaluate the node centrality in complex networks using the Nvidia CUDA programming model. The main
contribution of this work is the strategies for the development of an algorithm to evaluate the node
centrality in complex networks using Nvidia CUDA parallel programming model. We show that the
strategies improves algorithm´s speed-up in two orders of magnitude on one NVIDIA Tesla k20 GPU
cluster node, when compared to the hybrid OpenMP/MPI algorithm version, running in the same cluster,
with 4 nodes 2 Intel(R) Xeon(R) CPU E5-2660 each, for radius zero
A frame work for clustering time evolving dataiaemedu
The document proposes a framework for clustering time-evolving categorical data using a sliding window technique. It uses an existing clustering algorithm (Node Importance Representative) and a Drifting Concept Detection algorithm to detect changes in cluster distributions between the current and previous data windows. If a threshold difference in clusters is exceeded, reclustering is performed on the new window. Otherwise, the new clusters are added to the previous results. The framework aims to improve on prior work by handling drifting concepts in categorical time-series data.
AN ENTROPIC OPTIMIZATION TECHNIQUE IN HETEROGENEOUS GRID COMPUTING USING BION...ijcsit
This document summarizes a research paper that proposes a new method for improving both fault tolerance and load balancing in grid computing networks. The method converts the tree structure of grid computing nodes into a distributed R-tree index structure and then applies an entropy estimation technique. This entropy estimation helps discard nodes with high entropy from the tree, reducing complexity. The method then uses thresholding and control algorithms to select optimal route paths based on load balance and fault tolerance. Various optimization techniques like genetic algorithms, ant colony optimization, and particle swarm optimization are also applied to reach better solutions. Experimental results showed the proposed method improved performance over other existing methods.
The impact of channel model on the performance of distance-based schemes in v...IJECEIAES
This document discusses three distance-based schemes - RNDN, EVNDN, and OIFP - for data dissemination in vehicular named data networks (VNDNs). It evaluates the impact of different propagation models (Friis, Nakagami, random) on the performance of these schemes. Simulation results show that the Nakagami model significantly degrades network performance and increases retransmissions for RNDN. EVNDN sees more retransmissions under Friis and random models. OIFP exhibits higher interest satisfaction ratios and lower delays compared to the other schemes under all models. The document concludes that propagation models have a noticeable effect on distance-based scheme performance in VNDNs.
Reimagining Your Library Space: How to Increase the Vibes in Your Library No ...Diana Rendina
Librarians are leading the way in creating future-ready citizens – now we need to update our spaces to match. In this session, attendees will get inspiration for transforming their library spaces. You’ll learn how to survey students and patrons, create a focus group, and use design thinking to brainstorm ideas for your space. We’ll discuss budget friendly ways to change your space as well as how to find funding. No matter where you’re at, you’ll find ideas for reimagining your space in this session.
How to Build a Module in Odoo 17 Using the Scaffold MethodCeline George
Odoo provides an option for creating a module by using a single line command. By using this command the user can make a whole structure of a module. It is very easy for a beginner to make a module. There is no need to make each file manually. This slide will show how to create a module using the scaffold method.
A workshop hosted by the South African Journal of Science aimed at postgraduate students and early career researchers with little or no experience in writing and publishing journal articles.
How to Manage Your Lost Opportunities in Odoo 17 CRMCeline George
Odoo 17 CRM allows us to track why we lose sales opportunities with "Lost Reasons." This helps analyze our sales process and identify areas for improvement. Here's how to configure lost reasons in Odoo 17 CRM
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...Dr. Vinod Kumar Kanvaria
Exploiting Artificial Intelligence for Empowering Researchers and Faculty,
International FDP on Fundamentals of Research in Social Sciences
at Integral University, Lucknow, 06.06.2024
By Dr. Vinod Kumar Kanvaria
LAND USE LAND COVER AND NDVI OF MIRZAPUR DISTRICT, UPRAHUL
This Dissertation explores the particular circumstances of Mirzapur, a region located in the
core of India. Mirzapur, with its varied terrains and abundant biodiversity, offers an optimal
environment for investigating the changes in vegetation cover dynamics. Our study utilizes
advanced technologies such as GIS (Geographic Information Systems) and Remote sensing to
analyze the transformations that have taken place over the course of a decade.
The complex relationship between human activities and the environment has been the focus
of extensive research and worry. As the global community grapples with swift urbanization,
population expansion, and economic progress, the effects on natural ecosystems are becoming
more evident. A crucial element of this impact is the alteration of vegetation cover, which plays a
significant role in maintaining the ecological equilibrium of our planet.Land serves as the foundation for all human activities and provides the necessary materials for
these activities. As the most crucial natural resource, its utilization by humans results in different
'Land uses,' which are determined by both human activities and the physical characteristics of the
land.
The utilization of land is impacted by human needs and environmental factors. In countries
like India, rapid population growth and the emphasis on extensive resource exploitation can lead
to significant land degradation, adversely affecting the region's land cover.
Therefore, human intervention has significantly influenced land use patterns over many
centuries, evolving its structure over time and space. In the present era, these changes have
accelerated due to factors such as agriculture and urbanization. Information regarding land use and
cover is essential for various planning and management tasks related to the Earth's surface,
providing crucial environmental data for scientific, resource management, policy purposes, and
diverse human activities.
Accurate understanding of land use and cover is imperative for the development planning
of any area. Consequently, a wide range of professionals, including earth system scientists, land
and water managers, and urban planners, are interested in obtaining data on land use and cover
changes, conversion trends, and other related patterns. The spatial dimensions of land use and
cover support policymakers and scientists in making well-informed decisions, as alterations in
these patterns indicate shifts in economic and social conditions. Monitoring such changes with the
help of Advanced technologies like Remote Sensing and Geographic Information Systems is
crucial for coordinated efforts across different administrative levels. Advanced technologies like
Remote Sensing and Geographic Information Systems
9
Changes in vegetation cover refer to variations in the distribution, composition, and overall
structure of plant communities across different temporal and spatial scales. These changes can
occur natural.
This slide is special for master students (MIBS & MIFB) in UUM. Also useful for readers who are interested in the topic of contemporary Islamic banking.
1. See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/335140862
A Multi-Vehicle Trajectories Generator to Simulate Vehicle-to-Vehicle
Encountering Scenarios
Conference Paper · May 2019
DOI: 10.1109/ICRA.2019.8793776
CITATIONS
27
READS
1,063
3 authors:
Some of the authors of this publication are also working on these related projects:
Modeling and Understanding Human Driver Behaviors for Designing a Personalized ADAS View project
Autonomous vehicle security View project
Wenhao Ding
Carnegie Mellon University
23 PUBLICATIONS 134 CITATIONS
SEE PROFILE
Wenshuo Wang
University of California, Berkeley
70 PUBLICATIONS 1,372 CITATIONS
SEE PROFILE
Ding Zhao
University of Michigan
123 PUBLICATIONS 1,684 CITATIONS
SEE PROFILE
All content following this page was uploaded by Wenhao Ding on 18 February 2020.
The user has requested enhancement of the downloaded file.
2. A New Multi-vehicle Trajectory Generator to Simulate
Vehicle-to-Vehicle Encounters
Wenhao Ding1, Wenshuo Wang2, and Ding Zhao2
Abstract— Generating multi-vehicle trajectories from existing
limited data can provide rich resources for autonomous vehi-
cle development and testing. This paper introduces a multi-
vehicle trajectory generator (MTG) that can encode multi-
vehicle interaction scenarios (called driving encounters) into an
interpretable representation from which new driving encounter
scenarios are generated by sampling. The MTG consists of
a bi-directional encoder and a multi-branch decoder. A new
disentanglement metric is then developed for model analyses
and comparisons in terms of model robustness and the in-
dependence of the latent codes. Comparison of our proposed
MTG with β-VAE and InfoGAN demonstrates that the MTG
has stronger capability to purposely generate rational vehicle-
to-vehicle encounters through operating the disentangled latent
codes. Thus the MTG could provide more data for engineers
and researchers to develop testing and evaluation scenarios for
autonomous vehicles.
I. INTRODUCTION
Autonomous driving is being considered as a powerful
tool to bring a series of revolutionary changes in human
life. However, efficiently interacting with surrounding ve-
hicles in an uncertain environment continue to challenge
the deployment of autonomous vehicles because of scenario
diversity. Classifying the range of scenarios and separately
designing associated appropriate solutions appears to be an
alternative to overcome the challenge, but the limited prior
knowledge about the complex driving scenarios pose an
equally serious problem [1]. Some studies employed deep
learning technologies, such as learning controllers via end-
to-end neural networks [2] to handle the large amounts of
high-quality data without requiring full recovery of the multi-
vehicle interactions. Deep learning technologies, however, is
limited to scenarios that have never been shown up in the
training data set.
Most released databases [3] do not offer sufficient in-
formation on multi-vehicle interaction scenarios because of
technical limitations and the required costs of collecting data
[4]. One alternative is to generate new scenarios that are
similar to real world by modeling the limited data available,
as shown in Fig. 1. The basic underlying concept is inspired
by image style transformation [6], [7] and the functions of
deep generative models: projecting the encounter trajectories
into a latent space from which new trajectories can be then
generated using sampling techniques. Some similar concepts
1Wenhao Ding is with the Department of Electronic
Engineering, Tsinghua University, Beijing 10084, China
dingwenhao95@gmail.com
2Wenshuo Wang and Ding Zhao are with the Department of Mechanical
Engineering, Carnegie Mellon University (CMU), Pittsburgh, PA 15213,
USA. wwsbit@gmail.com, dingzhao@cmu.edu
Fig. 1. Procedure of generating multi-vehicle encounter trajectories.
have been developed, for example, variational autoencoders
(VAE) [8] which characterize the generated data more ex-
plicitly with a Gaussian distribution prior [9] [10]. However,
original VAE cannot fully capture the temporal and spatial
features of multi-vehicle encounter trajectories because it
only handle information over spatial space. As a remedy,
the neural networks in a recurrent frame of accounting the
history such as long short-term memory (LSTM) [11] and
gated recurrent units (GRU) [12] are able to tackle sequences,
like vehicle encounter trajectories, ordered by time.
In this paper, we develop a deep generative framework
integrated with a GRU module to generate multi-vehicle
encounter trajectories. Fig. 2 illustrates our proposed MTG
which encodes driving encounters into interpretable repre-
sentations with a bi-directional GRU module (green), and
generates the sequences through a multi-branch decoder
(red) separately. The reparameterization process (purple) is
introduced between the encoder and the decoder [8].
On the other hand, model performance evaluation is chal-
lenging because of lack of ground truth for the generated
multi-vehicle trajectories. Here we propose a new disentan-
glement metric for model performance evaluation in terms
of interpretability and stability. Compared with previous
disentanglement metric [10], our metric is not sensitive to
hyper-parameters and can also quantify the relevance among
the latent codes.
This paper contributes to the published literature by
• Introducing a deep generative model that uses latent
codes to characterize the dynamic interaction of multi-
vehicle trajectories.
arXiv:1809.05680v5
[cs.CV]
24
Feb
2019
3. Fig. 2. Scheme of Multi-Vehicle Trajectory Generator, which consists of three parts: encoder (green), sampling process (purple) and decoder (red).
• Generating multi-vehicle trajectories that are consistent
with the real data in the spatio-temporal space.
• Proposing a new disentanglement metric that can com-
prehensively analyze deep generative models in terms
of interpretability and stability.
II. RELATED WORK
This section briefly describes four types of deep gener-
ative models, an evaluation metric, time series processing
methods, and multiple trajectories generation.
A. Deep Generative Models and Evaluation Metric
Generative Adversarial Networks and InfoGAN. Gen-
erative Adversarial Networks (GAN) consists of two modules
[5]: a generator and a discriminator. These two modules
compete each other to reach a saddle balance point where
the generator can deceive the discriminator by generating
high-quality samples. Thus the object function of GAN is
formulated as
min
G
max
D
V (G, D) =
Ex∼pdata(x)[log D(x)] + Ez∼pz(z)[log(1 − D(G(z)))]
(1)
where x represents the real data, z represents the noise, and
G and D are the generator and discriminator, respectively.
In practice, the inputs of GAN are usually chosen as random
noises and thus the output of GAN are unpredictable. In order
to generate predictable samples, maximizing the mutual
information between the observation and the latent codes is
introduced, forming as an InfoGAN [13], formulated as
min
G
max
D
[V (G, D) − λI(z; x̄ = G(z))] (2)
VAE and β-VAE. The optimization of VAE is usually
converted to the evidence lower bound (ELBO), since the
estimation of the marginal log-likelihood is computationally
intractable because of the curse of dimensionality. The opti-
mization function of VAE and β-VAE is formulated in (3).
When β = 1, the original VAE [8] is obtained, and when
β > 1, the β-VAE [9], [14] was obtained.
log pθ(x) = Eqφ(z|x)[log pθ(x|z)] − βDKL(qφ||p(z)) (3)
where qφ represents the encoder parameterized by φ, pθ
represents the decoder parameterized by θ. β is used to
reduces the distance between pθ(z|x) and qφ(z|x), thereby
increasing β helps obtain more disentanglement. In (3), The
first term can be interpreted as a reconstruction error between
real and generated trajectories, and the second term is a
tightness condition to ensure pθ(z|x) = qφ(z|x).
Evaluation Metric. Disentanglement is a key factor for
deep generative model evaluation. Some research used the
accuracy of classifiers to represent the disentangled ability
[9], [10]. For example, Higgins et al. [9] acquired the
training data by calculating the difference 1/L
PL
l=1 |z1
k−z2
k|
between the latent codes z1
k and z2
k with a fixed k dimension.
Kim et al. [10] further considered the relationship between
the latent variables and the latent codes. The extreme sen-
sitivity of simple classifiers to hyper-parameters, however,
can skew the evaluation result. Moreover, the metrics in [9],
[10] cannot be directly used to analyze the stability and
dependency of the latent codes. In Section III-C, we propose
a new disentanglement metric capable of comprehensively
evaluating model performance without using any classifier-
based approaches.
B. Time Series Processing
Dealing with time series is challenging because of the
dependency among adjacent states. Both LSTM [11] and
GRU [12] are commonly used to tackle this issue because the
forget gate in them allows to control the amount of influence
the previous state can lay on the later state. GRU, as a
simplified version of LSTM, is more efficient with almost
equal performance [15], [16].
Combining RNN/LSTM/GRU and deep generative models
has been investigated. In [15], β-VAE framework with RNN
modules was designed to generate simple sketch drawings
with interpretable latent codes. All lines in the sketch were
considered as one sequence to avoid the interaction between
multiple sequences. The success of [15] partly depends on
the use of bi-directional RNN [17], which extracts more
sequential features with a forward-backward data flow [18].
4. Fig. 3. Trajectories generated from different models. (a), (b), (c) and (d) represent trajectories of real data, InfoGAN, VAE, and MTG, respectively.
TABLE I
ANALYSIS OF TRAJECTORIES IN THE TIME DOMAIN
Generated results from MTG with 4 values of z1 (0.1, 0.3, 0.5, 0.7) Description
Two vehicles change from driving in the same
direction to encountering with each other.
Two vehicles change from driving in the oppo-
site direction to encountering with each other.
The trajectories of two vehicles rotate in the
global coordinate.
C. Multiple Trajectories Generation
Multiple trajectories generation has been used to predict
subsequent trajectories under a supervised condition. For
example, a multi-LSTM module was used to deal with
different sequences and attain concatenated outputs through a
tensor [19]. The social LSTM [20] was proposed to generate
multi-pedestrian trajectories. The hidden states of different
sequences share information to increase the correlation,
allowing the agents to interact with their neighborhoods.
Later, an improved Social LSTM (or GAN) was proposed to
directly create trajectories from the generator of GAN [21].
III. PROPOSED METHODS
This section will introduce two baselines for comparison,
then describe MTG and the new disentanglement metric.
A. Baselines
In order to to explore the benefits of the modified struc-
ture, Baseline 1 is developed using a single-directional
GRU encoder of β-VAE. The encoder processes multiple
sequences simultaneously with one GRU module, and the
outputs (µ and σ) of the encoder are re-sampled through the
reparameterization trick [8]. The process is formulated as:
henc = GRUenc([S1; S2]) (4)
µ = Wµhenc + bµ , σ = exp(
Wσhenc + bσ
2
) (5)
z = µ + σ × N(0, I) (6)
where S1 and S2 are two input sequences (driving en-
counter), and z is the latent codes in dimension K. The
decoder takes z as the initial state and outputs sequence co-
ordinates one after another. The two sequences are generated
from the decoder at the same time by (7). We select Tanh
as the last activation function to make sure the output of the
decoder is in [-1,1].
[ ¯
S1; ¯
S2] = GRUdec(Pstart, z) (7)
In order to test another prevalent deep generative frame-
work, Baseline 2 is built on InfoGAN and has the same
GRU modules as our MTG. The generator in InfoGAN
shares the hidden states among multiple sequences, and
the discriminator encompasses a bi-directional GRU. The
specific structure of InfoGAN is detailed in Appendix.
B. Multi-Vehicle Trajectory Generator (MTG)
Compared to the Baseline 1, our MTG has two im-
provements. First, the bi-directional counterparts replace the
single-directional GRU module, which enables the encoder to
extract deeper representations from the trajectories, because
the traffic trajectories are still practically reasonable after
being reversed in the time domain. The pipeline of the
encoder of MTG is formulated as:
h→
enc = GRU→
enc([S1; S2]) , h←
enc = GRU←
enc([S1; S2]) (8)
henc = [h→
enc; h←
enc] (9)
Second, we separate the decoder into multiple branches and
share the hidden states among them. In this way, the hidden
5. Algorithm 1 Our disentanglement metric
1: Σ = [0.1, 0.4, 0.7, 1.0, 1.3, 1.6, 1.9, 2.2, 2.5, 2.8]
2: initiate Ω = []
3: for i in range(1, dim(z)) do
4: for σ in Σ do
5: initiate Θ = []
6: for j in range(1, dim(z)) do
7: when i 6= j, zj ∼ Normal(0, σ);
8: end for
9: for l in range(1, L) do
10: zi ∼ Normal(0, σ);
11: Z = concat(zk, k = 1 · · · dim(z));
12: Encoder(Z) =⇒ S, Decoder(S) =⇒ Ẑ;
13: append Ẑ to Θ;
14: end for
15: ωi,σ = V ar(Θ), append ωi,σ to Ω;
16: end for
17: end for
18: display σ for each i in Ω;
state retains all the information over past positions and
provides guidance to generate the other sequence. We note
that generating two sequences independently avoids mutual
influence. The pipeline of the decoder of MTG is formulated
as:
[St
1, ht
1] = GRUdec,1(St−1
1 , ht−1
2 ) (10)
[St
2, ht
2] = GRUdec,2(St−1
2 , ht−1
1 ) (11)
Then the objective function is concluded as:
L = F(S1, ¯
S1) + F(S2, ¯
S2) + β × DKL(qφ||p(z)) (12)
with F(·, ·) as the mean square error to calculate the recon-
struction error, and S̄i represents the reconstructive trajectory.
C. A New Disentanglement Metric
The metric in [10] holds one dimension of the latent
code fixed and selects other dimensions randomly, then
calculates the variance of the output of the encoder under
the assumption that the fixed dimension should have the less
variance, which is easily recognized by a linear classifier.
As a contrast, our metric (see Algorithm 1) is more stable.
We divide the input data into several groups with different
variances (each group has L samples) zk,σm , in which k ∈ K
is the index of the latent code, and m ∈ M is the group
index of the different variances. Only one code was selected
and sampled for each group with the others fixed. We input
these artificial latent codes z into the decoder to generate the
trajectories and then feed them into the encoder to obtain new
latent codes ẑ again. Finally, we calculate the variances of ẑ
for each group, revealing the dependence among latent codes
and model stability.
IV. EXPERIMENTS
A. Dataset and Preprocessing
We used driving encounter data collected by the University
of Michigan Transportation Research Institute (UMTRI) [4]
Fig. 4. Comparison of two evaluation metrics on Autoencoder and VAE.
from about 3,500 vehicles equipped with on-board GPS.
The latitude and longitude data was recorded by the GPS
device to represent vehicle positions. All data was collected
with a sampling frequency of 10 Hz. We then linearly
upscale and downscale the trajectory vectors to the size of
50. Considering the bias sensitivity of neural networks, we
normalize the value of the sequence into the range of [-1, 1]
by (13), where Si = {(xi, yi)|i = 1...50}.
{S̃i =
Si − mean(Si)
max(S1, S2)
|i = 1, 2} (13)
B. Experiment Settings and Evaluation
In all experiments, we set the dimension of z to 10. The
dimension selection could be different since we do not have
prior knowledge about the dimension of the latent space.
To test the capability of each code separately, we keep other
dimension fixed and change the value of one dimension from
-1 to 1 with a step size of 0.1. This step size can be smaller
for more detailed analysis. We conduct experiments from two
different aspects to compare our MTG with the two baselines.
The first aspect is to evaluate the generative model ac-
cording to traffic rationality. As shown in the first column of
Fig. 5, we analyze the generated trajectories in time domain
with three criteria:
• The distance between two sequences, which represents
the variation of the distance between two vehicles.
• The variation of speed expressed by the distance be-
tween two adjacent points.
• The variation of trajectory direction for smoothness
evaluation, where the angle between two consecutive
vectors represents the variation of moving direction.
The second aspect is to use our proposed metric to evaluate
models in the latent space. For each input variance, we
calculate a variance of the output trajectories and display it
as a bar as shown in Fig. 6 and Fig. 7. Each group consists
of ten different variances distinguished by color.
V. RESULTS ANALYSIS
A. Generative Trajectories Overview
Fig. 3 shows the generated trajectories from the two
baselines and our MTG. Each row shows the variation of
one latent code with all others fixed. For the InfoGAN
6. Fig. 5. Results of three models with traffic rationality evaluation. Both speed and direction results are from one vehicle.
baseline, the last three rows are almost the same, i.e., the
codes do not affect any features. This can be explained by
the unstable training of InfoGAN. Generating trajectories
capable of deceiving the discriminator makes the generator
difficult to obtain diversity, since the generator tends to
generate similar trajectories that are more likely to mislead
the discriminator. As a contrast, the VAE baseline and MTG
obtain more diverse trajectories.
Fig. 3 also shows that our MTG attains smoother and
more interpretable trajectories (i.e., no circles or sharp turns
appear) than the two baselines. The two baselines output
some trajectories that are unlikely to appear in the real world.
Table I lists some ‘zoom-in’ figures for more detailed
analysis of the generated trajectories of MTG. We connect
the associated points in the two sequences, from the starting
point (green) to the end point (red), with lines. In each
row, the four figures derive from four different values of
one latent code with a continuous change from left to right.
These trajectories indicate that MTG is able to control some
properties of generated trajectories (e.g., the location where
two vehicles meet and their directions) through the latent
codes.
B. Traffic Rationality Analysis
Fig. 5 shows some typical generated results based on the
three criteria introduced in Section IV-B. Different colors
represent different values of the latent code z1, wherein black
dashed lines represent the real traffic data for comparison.
The corresponding distance indicates that the InfoGAN
outputs trajectories with a small variance even for different
input values of z. This is in line with the problem of mode
collapse that is common in GAN family. The VAE baseline
and MTG obtain distances closer to the real trajectory. For
MTG, the distance gradually decreases and then increases
with z changing from -1 to 1, and vehicle speed changes
along the latent code value. Comparing the last two vertical
columns in Fig. 5 indicates that MTG can generate a much
smoother trajectory than VAE. In real world, vehicles cannot
take a sharp turning within a very short period of time
because of the physical constraints. Therefore, a high value
of consecutive angle will reduce the validity.
C. Disentanglement Analysis
Fig 4(a) and (b) with z6 and others fixed explain why our
metric outperforms previous one [10]. We obtain Fig 4(a)
by using the metric in [10] with an autoencoder. After being
normalized by dividing the standard deviation (left plot in
Fig 4(a)), the right part in Fig 4(a) shows the output variances
of ẑ. Although there is little difference in all codes, z6 still
attains the lowest value. Certainly, if all results are close to
this case, we can obtain a high accuracy of the classifier
while evaluating an autoencoder without any disentangled
ability; thus, the metric in [10] has problem evaluating
the disentanglement. As a contrast, our metric (Fig 4(b))
can identify the code that dominates and also reveal the
dependency among the codes. High values of latent codes
except z6 indicate that a strong dependency among all codes
zi.
We then evaluated and compared two metrics on the VAE.
Although z6 attains a much low value (Fig 4(c)) because of
the capability of VAE to disentangle the latent code, it is
still unsure if the other codes are influenced by z6. Fig 4(d)
shows the nearly zero value of the remaining codes (except
z5), i.e. the independence among the codes. Besides proving
the disentanglement, we find that z5 is a normal distribution
without any information (output variance equal to 1).
At last, we use the proposed metric to compare the VAE
baseline with our MTG. Fig. 6 shows (1) that only the
sampling code obtains an increasing output variance when
increasing the input variance, and (2) that the other codes are
close to zero. In other words, there is almost no dependency
among the latent codes in MTG, or, changing one code
does not influence other features. Fig. 7 shows two normal
distributions in the positions of z6 and z9, which indicates
7. Fig. 6. Results of MTG with our disentanglement metric.
that the VAE baseline obtains two latent codes without any
useful information. The plot of Code 8 in Fig. 7 also shows
that z8 influences z1 and z5 because their output variances
are non-zero.
The subplots inside Fig. 6 and Fig. 7 show the ratio of
output variance and input variance. A more robust approach
will force the value close to 1. The values in both figures,
however, are much greater than 1, which indicates that both
the VAE baseline and MTG are not robust enough. A robust
generative model requires that the encoder recognize all
trajectories generated from the decoder, i.e., the variances
of the outputs and the input should be the same.
VI. CONCLUSION
Developing safety policies and best practices for au-
tonomous vehicle deployment and low-cost self-driving ap-
plications will always depend on the data provided by
the high-quality generation of multi-vehicle and pedestrian
encounters. This paper proposed a novel method to generate
the multi-vehicles trajectories by using publicly available
data. Toward the end of extracting the features of two
spatiotemporal sequences, a separate generator architecture
with shared information was implemented. A new disentan-
glement metric capable of comprehensively analyzing the
generated trajectories and model robustness was also pro-
posed. An evaluation of traffic rationality using the proposed
disentanglement metric found that the MTG obtained more
stable latent codes and generated high-quality trajectories.
Fig. 7. Results of VAE baseline with our disentanglement metric.
While this paper only considered trajectories with short
lengths, the starting point of trajectories can be set arbitrar-
ily and the trajectories cascaded carefully. Future research
will account road profiles with conditional deep generative
models. The generated encounter data are expected to aid
in training automatic driving algorithms and providing indi-
vidual automatic vehicles with more low-cost data to ’learn’
complex traffic scenarios that are rarely encountered in the
real world.
APPENDIX
For hyper-parameter settings, network architectures and
more experiment results, please refer to the supplemen-
tary material: https://wenhao.pub/publication/
trajectory-supplement.pdf.
REFERENCES
[1] T. Appenzeller, “The scientists’ apprentice,” American Association for
the Advancement of Science, vol. 357, no. 6346, pp. 16–17, 2017.
[2] S. Yang, W. Wang, C. Liu, and W. Deng, “Scene understanding in
deep learning based end-to-end controllers for autonomous vehicles,”
IEEE Transactions on Systems, Man, and Cybernetics: Systems, 2018.
[3] W. Wang and D. Zhao, “Extracting traffic primitives directly from nat-
uralistically logged data for self-driving applications,” IEEE Robotics
and Automation Letters, vol. 3, no. 2, pp. 1223–1229, April 2018.
[4] W. Wang, C. Liu, and D. Zhao, “How much data are enough? a
statistical approach with case study on longitudinal driving behavior,”
IEEE Transactions on Intelligent Vehicles, vol. 2, no. 2, p. 8598, 2017.
[5] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley,
S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in
Advances in neural information processing systems, 2014, pp. 2672–
2680.
8. [6] M. Mirza and S. Osindero, “Conditional generative adversarial nets,”
arXiv preprint arXiv:1411.1784, 2014.
[7] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A. C.
Courville, “Improved training of wasserstein gans,” in Advances in
Neural Information Processing Systems, 2017, pp. 5767–5777.
[8] D. P. Kingma and M. Welling, “Auto-encoding variational bayes,”
arXiv preprint arXiv:1312.6114, 2013.
[9] I. Higgins, L. Matthey, A. Pal, C. Burgess, X. Glorot, M. Botvinick,
S. Mohamed, and A. Lerchner, “beta-vae: Learning basic visual
concepts with a constrained variational framework,” 2016.
[10] H. Kim and A. Mnih, “Disentangling by factorising,” arXiv preprint
arXiv:1802.05983, 2018.
[11] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural
computation, vol. 9, no. 8, pp. 1735–1780, 1997.
[12] K. Cho, B. Van Merriënboer, C. Gulcehre, D. Bahdanau, F. Bougares,
H. Schwenk, and Y. Bengio, “Learning phrase representations using
rnn encoder-decoder for statistical machine translation,” arXiv preprint
arXiv:1406.1078, 2014.
[13] X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever, and
P. Abbeel, “Infogan: Interpretable representation learning by informa-
tion maximizing generative adversarial nets,” in Advances in neural
information processing systems, 2016, pp. 2172–2180.
[14] C. P. Burgess, I. Higgins, A. Pal, L. Matthey, N. Watters, G. Des-
jardins, and A. Lerchner, “Understanding disentangling in beta-vae,”
arXiv preprint arXiv:1804.03599, 2018.
[15] D. Ha and D. Eck, “A neural representation of sketch drawings,” arXiv
preprint arXiv:1704.03477, 2017.
[16] X. Y. Zhang, F. Yin, Y. M. Zhang, C. L. Liu, and Y. Bengio, “Drawing
and recognizing chinese characters with recurrent neural network,”
IEEE Transactions on Pattern Analysis and Machine Intelligence,
vol. PP, no. 99, pp. 1–1, 2018.
[17] M. Schuster and K. K. Paliwal, “Bidirectional recurrent neural net-
works,” IEEE Transactions on Signal Processing, vol. 45, no. 11, pp.
2673–2681, 1997.
[18] J. Cross and L. Huang, “Incremental parsing with minimal features
using bi-directional lstm,” arXiv preprint arXiv:1606.06406, 2016.
[19] N. Deo and M. M. Trivedi, “Convolutional social pooling for vehicle
trajectory prediction,” arXiv preprint arXiv:1805.06771, 2018.
[20] A. Alahi, K. Goel, V. Ramanathan, A. Robicquet, L. Fei-Fei, and
S. Savarese, “Social lstm: Human trajectory prediction in crowded
spaces,” in Proceedings of the IEEE Conference on Computer Vision
and Pattern Recognition, 2016, pp. 961–971.
[21] A. Gupta, J. Johnson, L. Fei-Fei, S. Savarese, and A. Alahi, “So-
cial gan: Socially acceptable trajectories with generative adversarial
networks,” in IEEE Conference on Computer Vision and Pattern
Recognition (CVPR), no. CONF, 2018.
View publication stats
View publication stats