Leverage Turing Intelligence capabilities to integrate AI into your operations, enhance automation, and optimize cloud migration for scalable impact.
Advance foundation model research and improve LLM reasoning, coding, and multimodal capabilities with Turing AGI Advancement.
Access a global network of elite AI professionals through Turing Jobs—vetted experts ready to accelerate your AI initiatives.
How do social networking platforms like Facebook and Instagram generate friend or follower suggestions? What’s more, we tend to know many of these suggestions. A common technology used for this is network analysis.
This article will explore network analysis and geometric deep learning in detail and examine the differences between them.
A network is a symbolic representation of the essential characteristics of a group of objects/people. It is also known as a graph in mathematics.
Let’s understand this better with an example: say you’re a project manager and are asked to use four machines with the help of two men. What do you do? You try to figure out how to make the most out of the men by drafting a schedule detailing what time they will work and on which machines. You then need to check if the work assigned is being performed to par. All this seems like a lot of effort and it is. One way to save time is with the help of networks. Look at the following diagram:
Image source: Author
Image source: Author
Here, the arrows are known as ‘edges’ and represent the relationships between the objects and the circles, known as ‘vertices’ or ‘nodes’. They represent the objects we will analyze. In this way, we can create a network to make work easier.
For example, if we study a social relationship between Instagram users, nodes will be the target users and edges will be the relationship, such as the friendships between users.
Network analysis is beneficial for many applications. It helps in understanding the complex relationships in social networks or in analyzing the biological systems of organisms. It also helps in comprehending networks like those in banking, airlines, and supply chain. Here’s how:
Geometric deep learning was first introduced by Bronstein et al in their 2017 paper titled ‘Geometric Deep Learning: Going beyond Euclidean data’. It is defined as “an umbrella term for emerging techniques attempting to generalize (structured) deep neural models to non-Euclidean domains, such as graphs and manifolds”.
Geometric deep learning posits i) whether we can give the world a base where different architectures like convolutional neural networks (CNNs), recurrent neural networks (RNNs), transformers, etc. can have a common mathematical framework, and ii) if we can have certain prior physical knowledge that can be embedded in any architecture.
A new field of ML, geometric deep learning can learn from complex data like graphs and multi-dimensional points. In recent years, algorithms like CNN, long short-term memory (LSTM), generative adversarial network (GAN), etc. have helped achieve amazing accuracy on different types of problems.
According to Wikipedia, “A Euclidean space is a finite-dimensional vector space over the reals R, with an inner product”. Simply put, it involves the functional of 1D, 2D to n number of dimensions.
Euclidean looks for a flat surface whereas non-Euclidean looks for a curved surface. Some examples of non-Euclidean space are graphs/networks, manifolds, and similar complex structures. A few examples of Euclidean space are text, audio, images, etc.
Many algorithms used in ML applications are old and only work on Euclidean data. There is also a wide range of 3D-shaped representations, one of which is manifolds. These can be explained as multi-dimensional spaces. Here, a shape composed of any point is represented by a single point in this new space, and similar shapes are close to each other.
In order to apply deep neural networks on these types of datasets, it’s imperative to use other techniques by keeping the limits of non-Euclidean data in mind. These are:
A problem with learning in higher dimensions is that as the dimensions increase, so does the volume. For instance, if a solid ball is taken to infinite dimensions, the volume will shift from the center to the circumference and the ball will become hollow. But, if the ball is brought into a lower dimension, the sphere will be refilled. What this implies is that higher dimensional data is sparse and it becomes difficult to learn from sparse data.
Owing to this sparsity, one cannot find something that is close to the other dataset. However, geometric deep learning takes advantage of the fact that most tasks or functions that need to be estimated have underlying regularities through geometric principles in lower dimensions.
Another problem with graph-structured data is that standard deep neural networks are not able to interpret or learn it. This is because most networks are based on convolutions, which run smoothly on Euclidean data. As mentioned, datasets like graphs, manifolds, etc. are considered non-Euclidean data, meaning that they do not have shift-invariance and a common system of coordinates.
In order to deal with these problems and generalize deep neural networks to non-Euclidean domains, geometric deep learning has emerged as a separate research field.
Moving from two to three-dimensional volumes is essential for the future of deep learning. It would allow technology to achieve efficiency closer to the human brain using machine learning and deep learning techniques. Due to the increase in non-Euclidean data, geometric deep learning is gaining more ground. Through it, technology can match billions of atoms to track down new drug options to treat existing diseases, for example. This may be particularly useful in situations where chronic diseases are difficult to treat.