Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Filters








15,023 Hits in 2.1 sec

Architecture Disentanglement for Deep Neural Networks [article]

Jie Hu, Liujuan Cao, Qixiang Ye, Tong Tong, ShengChuan Zhang, Ke Li, Feiyue Huang, Rongrong Ji, Ling Shao
2021 arXiv   pre-print
Understanding the inner workings of deep neural networks (DNNs) is essential to provide trustworthy artificial intelligence techniques for practical applications.  ...  In this paper, we introduce neural architecture disentanglement (NAD) to fill the gap.  ...  Introduction A fundamental problem in using deep neural networks (DNNs) is our inability to understand their inner workings, which is crucial in many real-world applications, including healthcare, criminal  ... 
arXiv:2003.13268v2 fatcat:ge33yia6vvg7pcf7fawv2lrina

Differentiable Disentanglement Filter: an Application Agnostic Core Concept Discovery Probe [article]

Guntis Barzdins, Eduards Sidorovics
2019 arXiv   pre-print
Meanwhile disentangling the actual core concepts engrained in the word embeddings (like word2vec or BERT) or deep convolutional image recognition neural networks (like PG-GAN) is difficult and some success  ...  In this paper we propose a novel neural network nonlinearity named Differentiable Disentanglement Filter (DDF) which can be transparently inserted into any existing neural network layer to automatically  ...  Successful disentanglement of the core concepts in each layer is only the first step towards understanding the internal logic of the deep neural network.  ... 
arXiv:1907.07507v2 fatcat:a275ltqghjg2bpkmm254cjstwm

A Deeper Look at the Unsupervised Learning of Disentangled Representations in β-VAE from the Perspective of Core Object Recognition [article]

Harshvardhan Sikka
2020 arXiv   pre-print
Artificial Neural Networks, computational graphs consisting of weighted edges and mathematical operations at vertices, are loosely inspired by neural networks in the brain and have proven effective at  ...  (Pinto et al., 2008) (DiCarlo et al., 2012) For many data analysis tasks, learning representations where each dimension is statistically independent and thus disentangled from the others is useful.  ...  (Geron, 2017) Human Performance vs Deep Neural Networks With our understanding of core object recognition in the brain and of general deep neural network architectures, it is useful to examine explicit  ... 
arXiv:2005.07114v1 fatcat:6754eedw2nbexawoblecfbuuda

Joint Parameter Discovery and Generative Modeling of Dynamic Systems [article]

Gregory Barber, Mulugeta A. Haile, Tzikang Chen
2021 arXiv   pre-print
The neural framework uses a deep latent variable model to disentangle the system physical parameters from canonical coordinate observations.  ...  We present a neural framework for estimating physical parameters in a manner consistent with the underlying physics.  ...  We propose a neural framework for join parameter discovery and generative modeling of Hamiltonian systems that merges deep latent variable models and physical constraint embedded neural networks, 2.  ... 
arXiv:2103.10905v1 fatcat:tbdxwzr7lbdljcyrnml6pvlnvi

Multi-input Architecture and Disentangled Representation Learning for Multi-dimensional Modeling of Music Similarity [article]

Sebastian Ribecky, Jakob Abeßer, Hanna Lukashevich
2021 arXiv   pre-print
To achieve this, we propose a multi-input deep neural network architecture, which simultaneously processes mel-spectrogram, CENS-chromagram and tempogram in order to extract informative features for the  ...  Current representation learning strategies pursue the disentanglement of such factors from deep representations, resulting in highly interpretable models.  ...  Acknowledgements This research was partially supported by H2020 EU project AI4Media -A European Excellence Centre for Media, Society and Democracy -under Grand Agreement 951911.  ... 
arXiv:2111.01710v1 fatcat:lq33o4ka3jhzfhoaxuhvrdzgbq

Differentiable Frequency-based Disentanglement for Aerial Video Action Recognition [article]

Divya Kothandaraman, Ming Lin, Dinesh Manocha
2022 arXiv   pre-print
We use this differentiable mask prior to enable the neural network to intrinsically learn disentangled feature representations via an identity loss function.  ...  Our formulation empowers the network to inherently compute disentangled salient features within its layers.  ...  Deep neural networks for human activity recognition [2] , [3] have been widely used for ground-camera scenes [4] , [5] , where the human actors take a high fraction of the pixels in the video scenes  ... 
arXiv:2209.09194v2 fatcat:3llvgn4ewbbkvenuc7tyqt7vpa

Knowledge Consistency between Neural Networks and Beyond [article]

Ruofan Liang, Tianlin Li, Longfei Li, Jing Wang, Quanshi Zhang
2020 arXiv   pre-print
This paper aims to analyze knowledge consistency between pre-trained deep neural networks.  ...  We propose a generic definition for knowledge consistency between neural networks at different fuzziness levels.  ...  INTRODUCTION Deep neural networks (DNNs) have shown promise in many tasks of artificial intelligence.  ... 
arXiv:1908.01581v2 fatcat:3d236kykxbav3gjco6qlanz25q

Unsupervised deep learning identifies semantic disentanglement in single inferotemporal neurons [article]

Irina Higgins, Le Chang, Victoria Langston, Demis Hassabis, Christopher Summerfield, Doris Tsao, Matthew Botvinick
2020 arXiv   pre-print
Deep supervised neural networks trained to classify objects have emerged as popular models of computation in the primate ventral stream.  ...  Unlike deep classifiers, beta-VAE "disentangles" sensory data into interpretable latent factors, such as gender or hair length.  ...  Acknowledgments We would like to thank Raia Hadsell, Zeb Kurth-Nelson and Koray Kavukcouglu for comments on the manuscript.  ... 
arXiv:2006.14304v1 fatcat:xt2cluzicbbgbc2quo2oeofmj4

An Explicit Local and Global Representation Disentanglement Framework with Applications in Deep Clustering and Unsupervised Object Detection [article]

Rujikorn Charakorn, Yuttapong Thawornwattana, Sirawaj Itthipuripat, Nick Pawlowski, Poramate Manoonpong, Nat Dilokthanakul
2020 arXiv   pre-print
The code for our experiments is at https://github.com/51616/split-vae .  ...  Finally, we establish connections between SPLIT and recent research in cognitive neuroscience regarding the disentanglement in human visual perception.  ...  The authors would like to thank Murray Shanahan for discussions on the preliminary ideas and results, which built the foundation for this work.  ... 
arXiv:2001.08957v2 fatcat:2qjbomhn5zdntdojaq5y4jivqe

Neural Network Renormalization Group

Shuo-Hui Li, Lei Wang
2018 Physical Review Letters  
We present a variational renormalization group (RG) approach using a deep generative model based on normalizing flows.  ...  Conversely, the neural net directly maps independent Gaussian noises to physical configurations following the inverse RG flow.  ...  Figure 2 . 2 (a) A reference neural network architecture with only disentanglers.  ... 
doi:10.1103/physrevlett.121.260601 fatcat:4kolmbywu5cnrk2aipge2kh5qa

A Short Survey of Systematic Generalization [article]

Yuanpeng Li
2022 arXiv   pre-print
We hope this paper provides a background and is beneficial for discoveries in future work.  ...  Acknowledgments We thank Liang Zhao for beneficial discussions, suggestions, and adding information. We also thank Yi Yang for the helpful advice.  ...  Transformers (Vaswani et al. 2017 ) are modern neural network architectures with self-attention.  ... 
arXiv:2211.11956v1 fatcat:n5kdiovowjdplpc6olyjgpbjna

Disentangling Physical Dynamics From Unknown Factors for Unsupervised Video Prediction

Vincent Le Guen, Nicolas Thome
2020 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)  
Since physics is too restrictive for describing the full visual content of generic videos, we introduce PhyDNet, a two-branch deep architecture, which explicitly disentangles PDE dynamics from unknown  ...  Ablation studies also highlight the important gain brought out by both disentanglement and PDE-constrained prediction.  ...  Deep video prediction Deep neural networks have recently achieved state-of-the-art performances for datadriven video prediction.  ... 
doi:10.1109/cvpr42600.2020.01149 dblp:conf/cvpr/GuenT20a fatcat:uqz3dewzrfhb7fr5y6o5xge3aq

Adversarial Learning of Deepfakes in Accounting [article]

Marco Schreyer, Timur Sattarov, Bernd Reimer, Damian Borth
2019 arXiv   pre-print
In this work, we show an adversarial attack against CAATs using deep neural networks.  ...  Second, we show that adversarial autoencoder neural networks are capable of learning a human interpretable model of journal entries that disentangles the entries latent generative factors.  ...  Acknowledgements We thank the members of the statistics department at Deutsche Bundesbank for their valuable review and remarks.  ... 
arXiv:1910.03810v1 fatcat:7lzr63ahcbh7vm4n2y4xxr33bq

Disentangling Physical Dynamics from Unknown Factors for Unsupervised Video Prediction [article]

Vincent Le Guen, Nicolas Thome
2020 arXiv   pre-print
Since physics is too restrictive for describing the full visual content of generic videos, we introduce PhyDNet, a two-branch deep architecture, which explicitly disentangles PDE dynamics from unknown  ...  Ablation studies also highlight the important gain brought out by both disentanglement and PDE-constrained prediction.  ...  Deep video prediction Deep neural networks have recently achieved state-of-the-art performances for datadriven video prediction.  ... 
arXiv:2003.01460v2 fatcat:fviqde6rczf4hjxxnb2tzf53uq

Analyzing and Reducing the Damage of Dataset Bias to Face Recognition With Synthetic Data

Adam Kortylewski, Bernhard Egger, Andreas Schneider, Thomas Gerig, Andreas Morel-Forster, Thomas Vetter
2019 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)  
Furthermore, our study shows that current neural network architectures cannot disentangle face pose and facial identity, which limits their generalization ability. 2) We pre-train neural networks with  ...  Our analysis reveals that deeper neural network architectures can generalize better to unseen face poses.  ...  Deep networks cannot disentangle face pose from facial identity.  ... 
doi:10.1109/cvprw.2019.00279 dblp:conf/cvpr/KortylewskiESGM19 fatcat:iuzo4xqjrfgrtoe74tbso7qesq
« Previous Showing results 1 — 15 out of 15,023 results