Just as human chemists learn through diverse materials and experiences—textbooks, laboratory work, research papers, and problem-solving —ChemPile’s varied content types aim to provide a comprehensive learning
arXiv: arxiv.org/pdf/2505.12534
read more: chempile.lamalab.org
20.05.2025 15:48
👍 0
🔁 0
💬 0
📌 0
We introduce the ChemPile, the largest natural language chemistry dataset (>75B tokens).
dataset: huggingface.co/collections/...
20.05.2025 15:48
👍 0
🔁 0
💬 1
📌 0
Training large language models for chemistry is bottlenecked by one critical problem: there is no unified dataset that connects all chemical domains.
20.05.2025 15:48
👍 3
🔁 1
💬 1
📌 0
We're excited to present our posters today at the AI4Mat workshop at #ICLR25 #AI4Mat #Singapore
28.04.2025 00:38
👍 0
🔁 0
💬 0
📌 0
LAMA Lab at ICLR in Singapore!
#iclr2025 #singapore #AI #ML #chemistry #iclr
25.04.2025 11:57
👍 2
🔁 0
💬 1
📌 0
we're ready for spring! team building is always more fun when it's outside ☀️
21.04.2025 09:55
👍 1
🔁 0
💬 0
📌 0
Day 1 of the Foundation Models workshop hosted by the ELLIS Winter School!
18.03.2025 13:35
👍 0
🔁 0
💬 0
📌 0
Not sure where to start? Our documentation has step-by-step guides for every scenario
lamalab-org.github.io/chembench/
11.03.2025 16:52
👍 0
🔁 0
💬 0
📌 0
✨Public Datasets & Leaderboard – All datasets are live on HuggingFace, alongside a real-time performance leaderboard! huggingface.co/datasets/jab...
11.03.2025 16:52
👍 0
🔁 0
💬 1
📌 0
What's new?
✨Multimodal Support – Handle text, data, and chemistry-specific inputs seamlessly
✨Redesigned API – Now standardized on LiteLLM messages for effortless integration
✨Custom System Prompts – Tailor benchmarks to your unique use case
11.03.2025 16:52
👍 0
🔁 0
💬 1
📌 0
🚀ChemBench just leveled up!
We’re thrilled to announce the latest release of ChemBench—now smarter and smoother! Dive into benchmarking any chemistry AI model with our revamped framework, designed for flexibility and ease.
#ChemistryAI #MachineLearning #OpenScience #Innovation
11.03.2025 16:52
👍 1
🔁 1
💬 1
📌 0
🌟LLM limitations persist: Still lagging in 3D molecular spatial reasoning
#LLMs #MachineLearning #OpenScience
06.03.2025 07:46
👍 0
🔁 0
💬 0
📌 0
🌟System prompt insights: Ablation studies show no effect on evaluation outcomes
🌟VLLMs dominate: Outperform specialized models like Decimer in benchmarks
06.03.2025 07:46
👍 0
🔁 0
💬 1
📌 0
Fig A: Bar plot of model performance comparison with error bars
Fig B: Radar plot of relative performance for each model for each subtopic
MaCBench leaderboard hosted on HuggingFace spaces
🚀Our revised MaCBench paper is now on arxiv! arxiv.org/pdf/2411.16955
Key updates!
🌟Robust reproducibility: 5x experiment runs + error bars for statistical confidence
🌟Full dataset & leaderboard: Now live on HuggingFace with model comparisons huggingface.co/spaces/jablo...
06.03.2025 07:43
👍 1
🔁 0
💬 1
📌 1
For instance, one would expect vision models to perform very well and better than text models on spatial reasoning - such as identifying the correct isomeric relation between two compounds.
But this is not the case!
27.11.2024 16:46
👍 1
🔁 0
💬 1
📌 0
But we did not stop there! We dug deeper with ablations to understand the bottlenecks in applicability.
We compared different modalities, multi-step vs single step reasoning, guided prompting, etc.
27.11.2024 16:46
👍 1
🔁 0
💬 1
📌 0
We observed a striking disparity in performance across tasks. Models can identify lab equipment but struggle with identifying safety violations in real-life laboratory scenarios.
27.11.2024 16:46
👍 1
🔁 0
💬 1
📌 0
We and M3RG-Group from IIT Delhi created MaCBench: a multimodal materials and chemistry benchmark. (2137 questions)
We focus on the tasks we consider crucial for scientific development, practical lab scenarios, Spectral Analysis, US patents, and more.
27.11.2024 16:46
👍 1
🔁 0
💬 1
📌 0
Are Vision Language Models ready for scientific research?
🧑🔬🧪
We compared leading VLLMs on the three pillars of chemical and material science discovery: data extraction, lab experimentation and data interpretation.
arxiv.org/abs/2411.16955
27.11.2024 16:46
👍 11
🔁 4
💬 1
📌 1