Connor Lawless's Avatar

Connor Lawless

@lawlessopt

Stanford MS&E Postdoc | Human-Centered AI & OR Prev: @CornellORIE @MSFTResearch, @IBMResearch, @uoftmie 🌈

448
Followers
304
Following
27
Posts
15.11.2024
Joined
Posts Following

Latest posts by Connor Lawless @lawlessopt

Post image Post image Post image Post image

@lawlessopt.bsky.social and I are excited to present our #AAAI2026 tutorial on β€œLLMs for Optimization: Modeling, Solving, and Validating with Generative AI.”

When: Tuesday, Jan 20, 2026, 8:30am–12:30pm SGT
Where: Garnet 216 (Singapore EXPO)

(Connor’s intro slides are shown here.)
CC @aaai.org

20.01.2026 01:50 πŸ‘ 8 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0

Thank you!!

18.11.2025 00:33 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

It's been an absolute pleasure working with Ellen, Madeleine, and their amazing PhD students for the past year on making optimization more accessible with generative AI!

I am on the job market this year - check out my website (conlaw.github.io) for more details on what I've been up to.

16.11.2025 21:04 πŸ‘ 7 πŸ” 3 πŸ’¬ 1 πŸ“Œ 0
Post image

In our final session for the day, we're focused on a hot topic: machine learning and mixed integer programming. Connor Lawless will start the session and tells us how to use LLMs for cold-start cutting plane separator configuration.
doi.org/10.1007/978-...

12.11.2025 04:33 πŸ‘ 4 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Preview
DiffCoALG@NeurIPS 2025 About this Workshop Combinatorial algorithms are fundamental across a wide range of domains, owing to their ability to model optimization and decision-making tasks under complex constraints. These alg...

πŸ”₯ New workshop at @neuripsconf.bsky.social!
DiffCoALG bridges the gap between classic algorithms & differentiable learning.
Think: LLM reasoning, routing, SAT, MIP β€” neurally optimized.
πŸ“Œ Submit by Aug 22! πŸ€–πŸ§ 
πŸ”— sites.google.com/view/diffcoa...
#NeurIPS2025

12.08.2025 14:16 πŸ‘ 10 πŸ” 4 πŸ’¬ 1 πŸ“Œ 0
Preview
EquivaMap: Leveraging LLMs for Automatic Equivalence Checking of Optimization Formulations A fundamental problem in combinatorial optimization is identifying equivalent formulations. Despite the growing need for automated equivalence checks -- driven, for example, by optimization copilots, ...

πŸ“•: EquivaMap: Leveraging LLMs for Automatice Equivalence Checking of Optimization formulations
(Joint work with @ellen-v.bsky.social @hzhai.bsky.social and @leqiliu.bsky.social )
πŸ”—: arxiv.org/abs/2502.14760

16.07.2025 18:51 πŸ‘ 3 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

There's been a lot of work using LLMs to formulate MILPs, but how do we know that the formulations are correct?

Come chat with Haotian at poster W-515 to learn about our work on automatic equivalence checking for optimization models!

16.07.2025 18:49 πŸ‘ 6 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Our empirical results highlight that existing pointwise approaches for recourse can fail to catch potential fixed predictions, whereas our approach (provably) succeeds!

14.07.2025 16:15 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

We model the problem as a mixed-integer quadratically constrained program that runs in seconds on real-world datasets.

14.07.2025 16:15 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

This paradigm lets us spot fixed predictions before deploying a model, lets us audit public models for recourse (even if we don't have any available data!), and gives interpretable summaries of regions with fixed predictions to help with debugging.

14.07.2025 16:14 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

In this paper, we introduce a new paradigm for algorithmic recourse that aims to certify recourse over an entire region of the feature space!

14.07.2025 16:13 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Existing approaches to algorithmic recourse focus on verifying recourse on an individual-by-individual basis, which can cause model developers to miss potential fixed predictions, requires a lot of data, and makes it difficult to debug recourse issues!

14.07.2025 16:12 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Machine learning models can assign fixed predictions that preclude individuals from changing their outcome. Think credit applicants that can never get a loan approved, or young patients that can never get an organ transplant - no matter how sick they are!

14.07.2025 16:11 πŸ‘ 1 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Preview
Understanding Fixed Predictions via Confined Regions Machine learning models can assign fixed predictions that preclude individuals from changing their outcome. Existing approaches to audit fixed predictions do so on a pointwise basis, which requires ac...

Excited to be chatting about our new paper "Understanding Fixed Predictions via Confined Regions" (joint work with @berkustun.bsky.social, Lily Weng, and Madeleine Udell) at #ICML2025!

πŸ• Wed 16 Jul 4:30 p.m. PDT β€” 7 p.m. PDT
πŸ“East Exhibition Hall A-B #E-1104
πŸ”— arxiv.org/abs/2502.16380

14.07.2025 16:08 πŸ‘ 5 πŸ” 3 πŸ’¬ 1 πŸ“Œ 0
Post image

Our ✨spotlight paper✨ "Primal-Dual Neural Algorithmic Reasoning" is coming to #ICML2025!

We bring Neural Algorithmic Reasoning (NAR) to the NP-hard frontier πŸ’₯

πŸ—“ Poster session: Tuesday 11:00–13:30
πŸ“ East Exhibition Hall A-B, # E-3003
πŸ”— openreview.net/pdf?id=iBpkz...

🧡

13.07.2025 21:34 πŸ‘ 6 πŸ” 2 πŸ’¬ 1 πŸ“Œ 0

This is my first time at an HCI conference - come say hi if you're around!

25.03.2025 06:59 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

In addition to a bunch of quantitative experiments, we ran a user study with a prototype system to inform design recommendations for future interactive optimization systems. Check out the paper for more details!

25.03.2025 06:59 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We built a hybrid LLM and CP system that uses LLMs to translate user requests in chat into operations on an underlying CP optimization model to schedule a new meeting. This gets the best of both worlds - the flexibility of LLMs with the decision making power of optimization!

25.03.2025 06:58 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Building optimization models in practice involves a ton of back and forth between optimization and domain experts to understand a decision making problem. Can we enable domain experts to craft their own optimization models instead? We study this through the lens of scheduling.

25.03.2025 06:57 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
"I Want It That Way": Enabling Interactive Decision Support Using Large Language Models and Constraint Programming A critical factor in the success of decision support systems is the accurate modeling of user preferences. Psychology research has demonstrated that users often develop their preferences during the el...

Excited to be chatting about our ACM TIIS paper at IUI today:

"I Want it That Way": Enabling Interactive Decision Support via Large Language Models and Constraint Programming

πŸ”—: arxiv.org/abs/2312.06908

25.03.2025 06:55 πŸ‘ 3 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

In case youre wondering why this thread looks suspiciously like a bunch of screenshots from a presentation...

I'll be chatting about this project at the INFORMs Computing Society Conference in the debate room at 3. Come say hi!

16.03.2025 17:50 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

More broadly, this is a first step towards a new paradigm where we can exploit natural language information to do better algorithm configuration and design! There's tons of exciting open problems towards this goal (reach out if you're interested!).

16.03.2025 17:49 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

Surprisingly, we can get high performing configurations from our framework - outperforming solver defaults on a number of real world problems, without solving a single MILP!

16.03.2025 17:48 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We introduce a LLM based framework with some algorithmic bells and whistles (ensembling, solver specific context...) to capitalize on LLM strengths while addressing these challenges.

16.03.2025 17:47 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Unfortunately, LLMs aren't a natural fit for configuration. Parameters are problem specific, LLMs have stochastic outputs, and frankly - it's a tough problem!

16.03.2025 17:46 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image Post image Post image

Can we get better problem-specific solver configurations without the big computational price tag?

In this paper we show that we can thanks to Large Language Models! Why LLMs? They can identify useful optimization structure and have a lot of built in math programming knowledge!

16.03.2025 17:44 πŸ‘ 1 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Post image

MILP solvers ship with a ton of parameters that can have a massive impact on solver performance (over 70% for separator configuration alone!), but are notoriously difficult to set.

Existing approaches for algorithm configuration require solving a ton of MILPs leading to days of compute.

16.03.2025 17:41 πŸ‘ 1 πŸ” 0 πŸ’¬ 2 πŸ“Œ 0
Preview
LLMs for Cold-Start Cutting Plane Separator Configuration Mixed integer linear programming (MILP) solvers ship with a staggering number of parameters that are challenging to select a priori for all but expert optimization users, but can have an outsized impa...

Super excited about this new work with Yingxi Li, Anders Wikun, @ellen-v.bsky.social, and Madeleine Udell forthcoming at CPAIOR2025:

LLMs for Cold-Start Cutting Plane Separator Configuration

πŸ”—: arxiv.org/abs/2412.12038

16.03.2025 17:38 πŸ‘ 11 πŸ” 5 πŸ’¬ 1 πŸ“Œ 0
Preview
The foundations of America’s prosperity are being dismantled Federal scientists warn that Americans could feel the effects of the new administration's devastating cuts for decades to come

For decades, the US government has painstakingly kept American science #1 globallyβ€”and every facet of American life has improved because of it. The internet? Flu shot? Ozempic? All grew out of federally-funded research. Now all that's being dismantled. 1/ www.technologyreview.com/2025/02/21/1...

21.02.2025 13:00 πŸ‘ 3021 πŸ” 1512 πŸ’¬ 79 πŸ“Œ 117

"Not a step back"
Possibly --- even a step _forward_?
/s

12.02.2025 21:10 πŸ‘ 3 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0