Figure 1 - uploaded by Swapnil Nitin Shah
Content may be subject to copyright.
| Graphical representation of the restricted Boltzmann machine. Visible layer (observed variables) is comprised of units í µí±£ and the hidden layer (latent variables) is comprised of units ℎ. The layers are fully connected, and the connections W are bidirectional.

| Graphical representation of the restricted Boltzmann machine. Visible layer (observed variables) is comprised of units í µí±£ and the hidden layer (latent variables) is comprised of units ℎ. The layers are fully connected, and the connections W are bidirectional.

Contexts in source publication

Context 1
... one associates a) the average field í µí¼™ B and effective action Π[í µí¼™ B ] with the layer of visible spins í µí±£ of an RBM ( Fig. 1) like structure and b) í µí¼™ and bare action í µí±†[í µí¼™] with the layer of hidden spins ℎ thereof, after suitable discretization, from (9) and (17), This model, however, suffers from a major issue computationally. In general, the effective functional Π is not solvable perturbatively in í µí¼™ B owing to divergent integrals (similar ...
Context 2
... one associates a) the average field í µí¼™ B and effective action Π[í µí¼™ B ] with the layer of visible spins í µí±£ of an RBM ( Fig. 1) like structure and b) í µí¼™ and bare action í µí±†[í µí¼™] with the layer of hidden spins ℎ thereof, after suitable discretization, from (9) and (17), This model, however, suffers from a major issue computationally. In general, the effective functional Π is not solvable perturbatively in í µí¼™ B owing to divergent integrals (similar ...

Similar publications

Chapter
Full-text available
Defining and determining personal privacy is difficult; people jeopardize their personal lives and careers due to a lack of discernment. In this paper we propose an automated self-learning privacy assistant without a central entity determining the privacy and without pre-determined or labelled privacy data. We use the web as a data trove to determi...
Preprint
Full-text available
This work presents improvements in monocular hand shape estimation by building on top of recent advances in unsupervised learning. We extend momentum contrastive learning and contribute a structured collection of hand images, well suited for visual representation learning, which we call HanCo. We find that the representation learned by established...
Article
Full-text available
The rapid development of city makes it complicated to analyse urban structure. It’s difficult to learn city ecosystem using traditional methods including interview and survey. A city generates lots of data every day, which could tell city’s dynamic fabric. The moving patterns of citizens could be illustrated through analysing traffic data on their...
Conference Paper
Full-text available
Real world data is mostly unlabeled or only few instances are labeled. Manually labeling data is a very expensive and daunting task. This calls for unsupervised learning techniques that are powerful enough to achieve comparable results as semi-supervised/supervised techniques. Contrastive self-supervised learning has emerged as a powerful direction...

Citations

... Since then, the use of ML in condensed matter physics has rapidly expanded to include a variety of techniques [3][4][5]. These and their applications can be broadly grouped into two categories: supervised ML (SML), in which the input data to train the machine is labeled [6][7][8][9][10][11][12][13][14][15][16][17][18][19][20][21]; and unsupervised ML (UML), in which the input data is unlabeled and the machine proposes its own classification scheme [10,[19][20][21][22][23][24][25][26][27][28][29][30][31]. Across these two categories, the identification of thermodynamic phases in various models has remained a central theme. ...
Article
Full-text available
One of the most prominent tasks of machine learning (ML) methods within the field of condensed matter physics has been to classify phases of matter. Given their many successes in performing this task, one may ask whether these methods—particularly unsupervised ones—can go beyond learning the thermodynamic behavior of a system. This question is especially intriguing when considering systems that have a “hidden order”. In this work we study two random spin systems with a hidden ferromagnetic order that can be exposed by applying a Mattis gauge transformation. We demonstrate that the principal component analysis, perhaps the simplest unsupervised ML method, can detect the hidden order, quantify the corresponding gauge variables, and map the original random models onto simpler gauge-transformed ferromagnetic ones, all without any prior knowledge of the underlying gauge transformation. Our work illustrates that ML algorithms can in principle identify not manifestly obvious symmetries of a system.
... This is where ML can be of immense assistance [7][8][9][10][11]. On the other hand, physicists have also been contributing to ML, drawing inspiration from the theories and techniques developed for computational physics, as well as providing insights into the foundation of AI and ML [12][13][14][15]. One such insight manifests as we study deep learning side by side with the renormalization group (Fig.1). ...
Preprint
Full-text available
Physicists have had a keen interest in the areas of Artificial Intelligence (AI) and Machine Learning (ML) for some time now. With a special inclination on unraveling the mechanism at the core of the process of learning. In particular, exploring the underlying mathematical structure of a neural net (NN) is expected to not only help us in understanding the epistemological meaning of `Learning' but also has the potential to unravel the secrets behind the working of the brain. Here, it is worthwhile to establish correspondences and draw parallels between methods developed in core areas of Physics and the techniques developed at the forefront of AI and ML. Although recent explorations indicating a mapping between the Renormalisation Group(RG) and Deep Learning(DL) have shown valuable insights, we intend to investigate the relationship between RG and Autoencoders (AE) in particular. We will use Transfer Learning(TL) to embed the procedure of coarse-graining in a NN and compare it with the underlying mechanism of encoding-decoding through a series of tests.
... Post-hoc local explanations and feature relevance techniques are increasingly the most adopted methods for explaining DNNs 1 . Despite having some success, these techniques suffer from two major limitations -(1) Lack of consensus on an objective definition of interpretability for deep learning models and (2) Trade-off between interpretability and performance of these models 1 . A resolution to both of these problems can be achieved if the interpretation of such models can be provided by a formal logic system (as used by symbolic AI approaches) which is directly obtained from the underlying model. ...
Preprint
Full-text available
Deep learning models are widely used for various industrial and scientific applications. Even though these models have achieved considerable success in recent years, there exists a lack of understanding of the rationale behind decisions made by such systems in the machine learning community. This problem of interpretability is further aggravated by the increasing complexity of such models. This paper utilizes concepts from machine learning, quantum computation and quantum field theory to demonstrate how a many valued quantum logic system naturally arises in a specific class of generative deep learning models called Convolutional Deep Belief Networks. It provides a robust theoretical framework for constructing deep learning models equipped with the interpretability of many valued quantum logic systems without compromising their computing efficiency.
... Among these several learning algorithms one of the most popular is K-Means , another clustering methodology. In clustering the given datasets are divided in several parts or clusters based on their similarities and properties [8] [12]. In the above mentioned figure in the unsupervised section an image dataset is given including several image of different fruits. ...
... 12: AE b (b) cluster obtained at epoch 3The colors red, blue, green, cyan, magenta, lawn green, orange and grey represents 8 different classes.Due to higher number of data points in the extended dataset, all the classes are clearly visible inFigure 4.12 above. This time the clusters are visibly separable. ...
Thesis
Full-text available
In this thesis, self-supervised learning is used to enhance process data monitoring with the help of ML. Industrial, process dataset is not easy to acquire and often the dataset might not be large enough to train the system as desired. Therefore, Self-Supervised Learning (SSL) method is used to build a model in order to overcome this small dataset problem. In the first phase, the proposed model will learn about the given process dataset and will classify it based on the normal and fault case operation properties. In the next phase, the model will try to generate data points which will mimic the original data points that existed in the original dataset. This generated dataset will be combined later with the original dataset, enabling the system to learn about all the fault cases and non fault operation. The extended dataset, which is a combination of the original and generated dataset, can be used in terms of predictive maintenance, process monitoring and optimization.
... Since then, this field has exploded with a variety of ML applications [3][4][5]. These techniques and applications can be broadly grouped into two categories: supervised ML (SML), in which the input data is labelled to train the machine [6][7][8][9][10][11][12][13][14][15][16][17][18][19][20][21]; and unsupervised ML (UML), in which the input data is unlabelled and the machine proposes its own classification scheme [19][20][21][22][23][24][25][26][27][28][29][30][31][32]. As a major task of the condensed matter physicist, the classification of phases in various models has remained central among these applications. ...
Preprint
Full-text available
In condensed matter physics, one of the goals of machine learning is the classification of phases of matter. The consideration of a system's symmetries can significantly assist the machine in this goal. We demonstrate the ability of an unsupervised machine learning protocol, the Principal Component Analysis method, to detect hidden quenched gauge symmetries introduced via the so-called Mattis gauge transformation. Our work reveals that unsupervised machine learning can identify hidden properties of a model and may therefore provide new insights into the models themselves.
... At this point, input from theoretical physics can be proven beneficial. Among the various ideas invoked in the interface between the theoretical description of physical systems and machine learning to interpret and improve (deep) learning algorithms geometrization [16], variational approaches [17][18][19] and classical thermodynamics [20,21] have been proposed. In [22][23][24][25][26][27] ideas from the Renormalization Group flow are used to comprehend the flow of configurations triggered by generative models after training on systems from condensed matter physics. ...
Preprint
Focusing on the grand-canonical extension of the ordinary restricted Boltzmann machine, we suggest an energy-based model for feature extraction that uses a layer of hidden units with varying size. By an appropriate choice of the chemical potential and given a sufficiently large number of hidden resources the generative model is able to efficiently deduce the optimal number of hidden units required to learn the target data with exceedingly small generalization error. The formal simplicity of the grand-canonical ensemble combined with a rapidly converging ansatz in mean-field theory enable us to recycle well-established numerical algothhtims during training, like contrastive divergence, with only minor changes. As a proof of principle and to demonstrate the novel features of grand-canonical Boltzmann machines, we train our generative models on data from the Ising theory and MNIST.
... Dahası, denetimsiz öğrenme çok çeşitli yöntem ve algoritmaları kapsamaktadır. Cümle gömme için kontrastlı öğrenmeden temsil öğrenimi için varyasyonel oto kodlayıcılara kadar, denetimsiz teknikler gelişmeye ve görüntü füzyonu, anormallik tespiti ve arıza teşhisi gibi çeşitli alanlarda uygulama bulmaya devam etmektedir (Shah, 2019). Bu yöntemler, anlamlı temsilleri öğrenmek için verilerin doğal yapısını kullanarak, açık etiketler olmadan denoising, özellik çıkarma ve transfer öğrenme gibi görevleri mümkün kılmaktadır (Ferzo, 2024;Zhao & Wang, 2022). ...
Article
Physicists have had a keen interest in the areas of Artificial Intelligence (AI) and Machine Learning (ML) for a while, with a special inclination toward unraveling the fundamental mechanism behind the process of learning. In particular, exploring the underlying mathematical structure of a neural net (NN) is expected to not only help us understand the epistemological meaning of ‘Learning’ but also has the potential to unravel the secrets behind the workings of the brain. Here, it is worthwhile to establish correspondences and draw parallels between methods developed in core areas of Physics and the techniques developed at the forefront of AI and ML. Although recent explorations indicating a mapping between the Renormalisation Group (RG) and Deep Learning (DL) have shown valuable insights, we intend to investigate the relationship between RG and Autoencoders (AE) in particular. We will use Transfer Learning (TL) to embed the coarse-graining procedure in a NN and compare it with the underlying mechanism of encoding-decoding through a series of tests.