Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
  • AI
  • Tech
  • Cybersecurity
  • Finance
  • DeFi & Blockchain
  • Startups
  • Gaming
Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
Dataconomy
No Result
View All Result

Research: A periodic table for machine learning

I-Con reframes machine learning as mapping relationships between data points, unifying 23 methods and enabling new discoveries like label-free image classifiers through shared structure.

byKerem Gülen
April 24, 2025
in Research
Home Research
Share on FacebookShare on TwitterShare on LinkedInShare on WhatsAppShare on e-mail

In machine learning, few ideas have managed to unify complexity the way the periodic table once did for chemistry. Now, researchers from MIT, Microsoft, and Google are attempting to do just that with I-Con, or Information Contrastive Learning. The idea is deceptively simple: represent most machine learning algorithms—classification, regression, clustering, and even large language models—as special cases of one general principle: learning the relationships between data points.

Just like chemical elements fall into predictable groups, the researchers claim that machine learning algorithms also form a pattern. By mapping those patterns, I-Con doesn’t just clarify old methods. It predicts new ones. One such prediction? A state-of-the-art image classification algorithm requiring zero human labels.

Imagine a ballroom dinner. Each guest (data point) finds a seat (cluster) ideally near friends (similar data). Some friends sit together, others spread across tables. This metaphor, called the clustering gala, captures how I-Con treats clustering: optimizing how closely data points group based on inherent relationships. It’s not just about who’s next to whom, but what types of bonds matter; be it visual similarity, shared class labels, or graph connections.

Stay Ahead of the Curve!

Don't miss out on the latest insights, trends, and analysis in the world of data, technology, and startups. Subscribe to our newsletter and get exclusive content delivered straight to your inbox.

This ballroom analogy extends to all of machine learning. The I-Con framework shows that algorithms differ mainly in how they define those relationships. Change the guest list or seating logic, and you get dimensionality reduction, self-supervised learning, or spectral clustering. It all boils down to preserving certain relationships while simplifying others.

research-a-periodic-table-for-machine-learning-0_03

The architecture behind I-Con

At its core, I-Con is built on an information-theoretic foundation. The objective: minimize the difference (KL divergence) between a target distribution, what the algorithm thinks relationships should be, and a learned distribution, the actual model output. Formally, this is written as:

L(θ, ϕ) = ∑ DKL(pθ(·|i) || qϕ(·|i))

Different learning techniques arise from how the two distributions, pθ and qϕ, are constructed. When pθ groups images by visual closeness and qϕ groups them by label similarity, the result is supervised classification. When pθ relies on graph structure, and qϕ approximates it through clusters, we get spectral clustering. Even language modeling fits in, treating token co-occurrence as a relationship to be preserved.

research-a-periodic-table-for-machine-learning-0_03

The table that organizes everything

Inspired by chemistry’s periodic table, the I-Con team built a grid categorizing algorithms based on their connection types. Each square in the table represents a unique way data points relate in the input versus output space. Once all known techniques were placed, surprising gaps remained. These gaps didn’t point to missing data—they hinted at methods that could exist but hadn’t been discovered yet.

To test this, the researchers filled in one such gap by combining clustering with debiased contrastive learning. The result: a new method that outperformed existing unsupervised image classifiers on ImageNet by 8%. It worked by injecting a small amount of noise—“universal friendship” among data points—that made the clustering process more stable and less biased toward overconfident assignments.

research-a-periodic-table-for-machine-learning-0_03

Debiasing plays a central role in this discovery. Traditional contrastive learning penalizes dissimilar samples too harshly, even when those samples might not be truly unrelated. I-Con introduces a better approach: mixing in a uniform distribution that softens overly rigid assumptions about data separations. It’s a conceptually clean tweak with measurable gains in performance.

Another method involves expanding the definition of neighborhood itself. Instead of looking only at direct nearest neighbors, I-Con propagates through the neighborhood graph—taking “walks” to capture more global structure. These walks simulate how information spreads across nodes, improving the clustering process. Tests on DiNO vision transformers confirm that small-scale propagation (walk length of 1 or 2) yields the most gain without overwhelming the model.


Research: Google’s AI eats your clicks


Performance and payoff

The I-Con framework isn’t just theory. On ImageNet-1K, it beat previous state-of-the-art clustering models like TEMI and SCAN using simpler, self-balancing loss functions. Unlike its predecessors, I-Con doesn’t need manually tuned penalties or size constraints. It just works—across DiNO ViT-S, ViT-B, and ViT-L backbones.

Debiased InfoNCE Clustering (I-Con) improved Hungarian accuracy by:

  • +4.5% on ViT-B/14
  • +7.8% on ViT-L/14

It also outperformed k-Means, Contrastive Clustering, and SCAN consistently. The key lies in its clean unification of methods and adaptability—cluster probabilities, neighbor graphs, class labels, all fall under one umbrella.

I-Con isn’t just a unifier; it’s a blueprint for invention. By showing that many algorithms are just different ways of choosing neighborhood distributions, it empowers researchers to invent new combinations. Swap one connection type for another. Mix in debiasing. Tune neighborhood depth. Each tweak corresponds to a new entry in the table—a new algorithm ready to be tested.

As MIT’s Shaden Alshammari put it, machine learning is starting to feel less like an art of guesswork and more like a structured design space. I-Con turns learning into exploration—less alchemy, more engineering.

What I-Con really offers is a deeper philosophy of machine learning. It reveals that beneath the vast diversity of models and methods, a common structure may exist—one built not on rigid formulas, but on relational logic. In that sense, I-Con doesn’t solve intelligence. It maps it. And like the first periodic table, it gives us a glimpse of what’s still waiting to be discovered.


Featured image credit

Tags: GoogleMachine LearningMicrosoftMIT

Related Posts

Apple Watch data can predict your health with 92% accuracy

Apple Watch data can predict your health with 92% accuracy

July 11, 2025
How AI platforms rank on data privacy in 2025

How AI platforms rank on data privacy in 2025

July 9, 2025
This AI lab wants to automate scientific discovery

This AI lab wants to automate scientific discovery

July 8, 2025
Counterpoint data shows the global smartwatch market is now shrinking

Counterpoint data shows the global smartwatch market is now shrinking

July 4, 2025
ChatGPT referrals can’t save publishers from AI search

ChatGPT referrals can’t save publishers from AI search

July 3, 2025
AI made executives worse at stock picking

AI made executives worse at stock picking

July 2, 2025

LATEST NEWS

Is Google’s foldable phone already dead on arrival?

Google simplifies Lens to make room for its Gemini AI

Ex-Intel CEO Pat Gelsinger has a new mission for AI

Amazon Kindle finally lets you filter lockscreen ads

Samsung preps foldable screens for Apple’s iPhone Fold

This is what a Windows crash looks like now

Dataconomy

COPYRIGHT © DATACONOMY MEDIA GMBH, ALL RIGHTS RESERVED.

  • About
  • Imprint
  • Contact
  • Legal & Privacy

Follow Us

  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
No Result
View All Result
Subscribe

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy Policy.