Trending
Jablonka Lab (Lab for AI for Materials)'s Avatar

Jablonka Lab (Lab for AI for Materials)

@jablonkagroup

Team-run account for the group led by @kjablonka.com

29
Followers
3
Following
20
Posts
22.11.2024
Joined
Posts Following

Latest posts by Jablonka Lab (Lab for AI for Materials) @jablonkagroup

Just as human chemists learn through diverse materials and experiences—textbooks, laboratory work, research papers, and problem-solving —ChemPile’s varied content types aim to provide a comprehensive learning
arXiv: arxiv.org/pdf/2505.12534
read more: chempile.lamalab.org

20.05.2025 15:48 👍 0 🔁 0 💬 0 📌 0
Post image

We introduce the ChemPile, the largest natural language chemistry dataset (>75B tokens).
dataset: huggingface.co/collections/...

20.05.2025 15:48 👍 0 🔁 0 💬 1 📌 0
Post image

Training large language models for chemistry is bottlenecked by one critical problem: there is no unified dataset that connects all chemical domains.

20.05.2025 15:48 👍 3 🔁 1 💬 1 📌 0
Post image

We're excited to present our posters today at the AI4Mat workshop at #ICLR25 #AI4Mat #Singapore

28.04.2025 00:38 👍 0 🔁 0 💬 0 📌 0
Post image

LAMA Lab at ICLR in Singapore!
#iclr2025 #singapore #AI #ML #chemistry #iclr

25.04.2025 11:57 👍 2 🔁 0 💬 1 📌 0
Post image Post image Post image

we're ready for spring! team building is always more fun when it's outside ☀️

21.04.2025 09:55 👍 1 🔁 0 💬 0 📌 0
Post image

Day 1 of the Foundation Models workshop hosted by the ELLIS Winter School!

18.03.2025 13:35 👍 0 🔁 0 💬 0 📌 0
Post image

 Not sure where to start? Our documentation has step-by-step guides for every scenario
lamalab-org.github.io/chembench/

11.03.2025 16:52 👍 0 🔁 0 💬 0 📌 0
Post image

✨Public Datasets & Leaderboard – All datasets are live on HuggingFace, alongside a real-time performance leaderboard! huggingface.co/datasets/jab...

11.03.2025 16:52 👍 0 🔁 0 💬 1 📌 0

What's new?
✨Multimodal Support – Handle text, data, and chemistry-specific inputs seamlessly
✨Redesigned API – Now standardized on LiteLLM messages for effortless integration
✨Custom System Prompts – Tailor benchmarks to your unique use case

11.03.2025 16:52 👍 0 🔁 0 💬 1 📌 0

🚀ChemBench just leveled up!
We’re thrilled to announce the latest release of ChemBench—now smarter and smoother! Dive into benchmarking any chemistry AI model with our revamped framework, designed for flexibility and ease.
#ChemistryAI #MachineLearning #OpenScience #Innovation

11.03.2025 16:52 👍 1 🔁 1 💬 1 📌 0
Video thumbnail

🌟LLM limitations persist: Still lagging in 3D molecular spatial reasoning
#LLMs #MachineLearning #OpenScience

06.03.2025 07:46 👍 0 🔁 0 💬 0 📌 0
Post image

🌟System prompt insights: Ablation studies show no effect on evaluation outcomes
🌟VLLMs dominate: Outperform specialized models like Decimer in benchmarks

06.03.2025 07:46 👍 0 🔁 0 💬 1 📌 0
Fig A: Bar plot of model performance comparison with error bars
Fig B: Radar plot of relative performance for each model for each subtopic

Fig A: Bar plot of model performance comparison with error bars Fig B: Radar plot of relative performance for each model for each subtopic

MaCBench leaderboard hosted on HuggingFace spaces

MaCBench leaderboard hosted on HuggingFace spaces

🚀Our revised MaCBench paper is now on arxiv! arxiv.org/pdf/2411.16955

Key updates!
🌟Robust reproducibility: 5x experiment runs + error bars for statistical confidence
🌟Full dataset & leaderboard: Now live on HuggingFace with model comparisons huggingface.co/spaces/jablo...

06.03.2025 07:43 👍 1 🔁 0 💬 1 📌 1
Preview
Probing the limitations of multimodal language models for chemistry and materials research Recent advancements in artificial intelligence have sparked interest in scientific assistants that could support researchers across the full spectrum of scientific workflows, from literature review to...

Supported by Carl Zeiss Foundation, Intel, Merck, Alexander von Humboldt Foundation, Friedrich-Schiller-Universität Jena, IIT Delhi.

📜Manuscript: arxiv.org/abs/2411.16955
👩‍💻GitHub: github.com/lamalab-org/...

27.11.2024 16:46 👍 2 🔁 0 💬 0 📌 0
Post image

For instance, one would expect vision models to perform very well and better than text models on spatial reasoning - such as identifying the correct isomeric relation between two compounds.

But this is not the case!

27.11.2024 16:46 👍 1 🔁 0 💬 1 📌 0
Post image

But we did not stop there! We dug deeper with ablations to understand the bottlenecks in applicability.
We compared different modalities, multi-step vs single step reasoning, guided prompting, etc.

27.11.2024 16:46 👍 1 🔁 0 💬 1 📌 0
Post image

We observed a striking disparity in performance across tasks. Models can identify lab equipment but struggle with identifying safety violations in real-life laboratory scenarios.

27.11.2024 16:46 👍 1 🔁 0 💬 1 📌 0
Post image

We and M3RG-Group from IIT Delhi created MaCBench: a multimodal materials and chemistry benchmark. (2137 questions)

We focus on the tasks we consider crucial for scientific development, practical lab scenarios, Spectral Analysis, US patents, and more.

27.11.2024 16:46 👍 1 🔁 0 💬 1 📌 0
Video thumbnail

Are Vision Language Models ready for scientific research?
🧑‍🔬🧪

We compared leading VLLMs on the three pillars of chemical and material science discovery: data extraction, lab experimentation and data interpretation.
arxiv.org/abs/2411.16955

27.11.2024 16:46 👍 11 🔁 4 💬 1 📌 1