Free Quiz
Write for Us
Learn Artificial Intelligence and Machine Learning
  • Artificial Intelligence
  • Data Science
    • Language R
    • Deep Learning
    • Tableau
  • Machine Learning
  • Python
  • Blockchain
  • Crypto
  • Big Data
  • NFT
  • Technology
  • Interview Questions
  • Others
    • News
    • Startups
    • Books
  • Artificial Intelligence
  • Data Science
    • Language R
    • Deep Learning
    • Tableau
  • Machine Learning
  • Python
  • Blockchain
  • Crypto
  • Big Data
  • NFT
  • Technology
  • Interview Questions
  • Others
    • News
    • Startups
    • Books
Learn Artificial Intelligence and Machine Learning
No Result
View All Result

Home » Researchers teach LLMs to solve complex planning challenges

Researchers teach LLMs to solve complex planning challenges

Tarun Khanna by Tarun Khanna
April 3, 2025
in Machine Learning
Reading Time: 4 mins read
0
Researchers teach LLMs to solve complex planning challenges

Photo Credit: https://news.mit.edu/

Share on FacebookShare on TwitterShare on LinkedInShare on WhatsApp

This new framework leverages a model’s reasoning capabilities to create a “smart assistant” that discovers the most exceptional solution to multistep problems

Imagine a coffee enterprise trying to enhance its supply chain. The enterprise assets beans from 3 suppliers, roasts them at 2 facilities into both dark or mild coffee, and then transport the roasted coffee to 3 retail locations. The providers have different constant capacity, and roasting fees and shipping expenses vary from location to location.

The enterprise seeks to limit cost while meeting a 23% growth in demand.

Also Read:

Accelerating Machine Learning Model Deployment with MLOps Tools

Machine Learning Prediction Examples

Top 10 Real World Applications of Machine Learning

Machine Learning Role In Paraphrasing Tools To Avoid Plagiarism

Wouldn’t it be easier for the enterprise to just ask ChatGPT to come up with an exceptional plan? In reality, for all their fantastic capabilities, large language fashions (LLMs) often perform poorly while tasked with directly solving such complicated planning issues on their very own.

Rather than looking to alternate the model to make an LLM a better planner, MIT researchers gone through a different technique. They launched a framework that guides an LLM to interrupt down the hassle like a human would, after which automatically resolve it the use of a powerful software program tool.

A user only needs to explain the trouble in natural language — no task-specific examples are had to train or prompt the LLM. The model encodes a user’s text prompt into a format that may be unraveled by the optimization solver designed to correctly extremely tough planning challenges.

During the formulation technique, the LLM checks its work at more than one intermediate step to make sure the plan is depicted successfully to the solver. If it spots an error, as opposed to giving up, the LLM attempts to fix the broken part of the formulation.

When the researchers tested their framework on 9 complex challenges, together with minimizing the distance warehouse robots have to travel to finish tasks, it attained an 85% achievement rate, while the best baseline simplest attained a 39% achievement rate.

The versatile framework could be implemented to various multistep planning tasks, consisting of scheduling airline crews or dealing with machine time in a factory.

“Our research launches a framework that basically acts as a smart assistant for planning issues. It can determine the best plan that meets all of the requirements you have, even though the regulations are complicated or unusual,” stated Yilun Hao, a graduate student in the MIT Laboratory for Information and Decision Systems (LIDS) and lead author of a paper on this research.

She is joined at the paper by Yang Zhang, a research scientist on the MIT-IBM Watson AI Lab; and senior creator Chuchu Fan, a partner professor of aeronautics and astronautics and LIDS principal investigator. The research may be presented at the International Conference on Learning Representations.

 

Optimization 101

The Fan group develops algorithms that automatically solve up what are referred to as combinatorial optimization issues. These vast issues have many interrelated choice variables, each with more than one alternative that hastily add up to billions of potential choices.

Humans solve such issues with by narrowing them down to few alternatives after which figuring out which one leads to the best overall plan. The researchers’ algorithmic solvers follow the same principles to optimization problems that are a way too complex for a human to crack.

But the solvers they develop tend to have steep learning curves and are commonly most effective utilized by professionals.

“We thought that LLMs could allow nonexperts to use those solving algorithms. In our lab, we take a domain professional’s issues and formalize it right into an issue our solver can solve. Could we teach an LLM to do the same things?” Fan stated.

Using the framework the researchers evolved, known as LLM-Based Formalized Programming (LLMFP), a person gives a natural language description of the issue, background information at the task, and a query that describes their aim.
Then LLMFP prompts an LLM to cause about the problem and decide the choice variables and key constraints a good way to shape the ideal solution.

LLMFP asks the LLM to detail the necessities of each variable before encoding the records right into a mathematical formulation of an optimization issues. It writes code that encodes the problem and calls the attached optimization solver, which reaches at an ideal solution.

“It is similar to how we teach undergrads about optimization issues at MIT. We don’t teach them just one domain. We teach them the technique,” Fan adds.

As long as the inputs to the solver are correct, it’ll give the right solution. Any errors in the solution come from errors in the formulation method.

To make sure it has located an operating plan, LLMFP analyzes the solution and modifies any wrong steps inside the issue formulation. Once the plan passes this self-evaluation, the solution is described to the user in natural language.

 

Perfecting the plan

This self-assessment module also permits the LLM to add any implicit constraints it misses the primary time around, Hao says.
For example, if the framework is optimizing a supply chain to limit charges for a coffeeshop, a human is aware of the coffeeshop can’t deliver a poor quantity of roasted beans, however an LLM won’t realize that.

The self-assessment step would flag that error and prompt the model to fix it.

“Plus, an LLM can adapt to the options of the user. If the model realizes a selected user does no longer like to change the time or price range of their tour plans, it may suggest converting things that fit the user’s requirements,” Fan stated.

In a sequence of tests, their framework attained an average achievement rate between 83 and 87% across 9 diverse planning issues using numerous LLMs. While a few baseline models have been better at certain issues, LLMFP attained an overall achievement rate about two times as high as the baseline strategies.

Unlike those different approaches, LLMFP does not require domain-specific examples for training. It can locate the solution technique to a making planning problem proper out of the box.

In addition, the user can adapt LLMFP for different optimization solvers by adjusting the prompts fed to the LLM.

“With LLMs, we’ve got an possibility to create an interface that permits humans to use tools from other domains to solve issues in ways they may not have been thinking about earlier than before,” Fan stated.

In the future, the researchers need to enable LLMFP to take images as input to supplement the descriptions of a planning issues. This could help the framework solve tasks which are especially difficult to fully describe with natural language.
This work was funded, in part, via the Office of Naval Research and the MIT-IBM Watson AI Lab.

ShareTweetShareSend
Previous Post

Anthropic invented ‘AI microscope’ to show how large language models think

Next Post

Vana is letting customers own a piece of the AI models trained on their data

Tarun Khanna

Tarun Khanna

Founder DeepTech Bytes - Data Scientist | Author | IT Consultant
Tarun Khanna is a versatile and accomplished Data Scientist, with expertise in IT Consultancy as well as Specialization in Software Development and Digital Marketing Solutions.

Related Posts

How-To-Kick-Start-Your-Machine-Learning-Career
Machine Learning

How To Kick Start Your Machine Learning Career?

April 14, 2022
AI Paraphrasing Tools
Artificial Intelligence

Working Of Machine Learning In AI Paraphrasing Tools

March 31, 2022
Machine Learning

Machine Learning Life Cycle Management

March 10, 2022
top-artificial-intelligence-technologies
Artificial Intelligence

Top Artificial Intelligence Technologies

January 4, 2022
Next Post
Vana is letting users own a piece of the AI models trained on their data

Vana is letting customers own a piece of the AI models trained on their data

TRENDING

Patient Identity Verification Process Explained – The Anti-fraud Pill

by Tarun Khanna
March 31, 2022
0
ShareTweetShareSend

Emerging Artificial Intelligence And Machine Learning Trends

ai-and-ml-trends
by Tarun Khanna
February 6, 2021
0
ShareTweetShareSend

AI Breakthrough: Scientists Transform Everyday Transistor Into an Artificial Neuron

AI Breakthrough: Scientists Transform Everyday Transistor Into an Artificial Neuron

Photo Credit: https://scitechdaily.com/

by Tarun Khanna
April 16, 2025
0
ShareTweetShareSend

Deep Learning vs Machine Learning – What’s the difference?

Deep Learning vs Machine Learning

Deep Learning vs Machine Learning – What’s the difference?

by Tarun Khanna
February 1, 2021
0
ShareTweetShareSend

Optimizing Costs in Kubernetes Environments: A Comprehensive Guide

Kubernetes-Environments
by Tarun Khanna
January 30, 2024
0
ShareTweetShareSend

5 Ways to Improve the Conversion Rate of Your Website’s Service Pages

conversion-rate website pages
by Tarun Khanna
January 25, 2023
0
ShareTweetShareSend

DeepTech Bytes

Deep Tech Bytes is a global standard digital zine that brings multiple facets of deep technology including Artificial Intelligence (AI), Machine Learning (ML), Data Science, Blockchain, Robotics,Python, Big Data, Deep Learning and more.
Deep Tech Bytes on Google News

Quick Links

  • Home
  • Affiliate Programs
  • About Us
  • Write For Us
  • Submit Startup Story
  • Advertise With Us
  • Terms of Service
  • Disclaimer
  • Cookies Policy
  • Privacy Policy
  • DMCA
  • Contact Us

Topics

  • Artificial Intelligence
  • Data Science
  • Python
  • Machine Learning
  • Deep Learning
  • Big Data
  • Blockchain
  • Tableau
  • Cryptocurrency
  • NFT
  • Technology
  • News
  • Startups
  • Books
  • Interview Questions

Connect

For PR Agencies & Content Writers:

connect@deeptechbytes.com

Facebook Twitter Linkedin Instagram
Listen on Apple Podcasts
Listen on Google Podcasts
Listen on Google Podcasts
Listen on Google Podcasts
DMCA.com Protection Status

© 2024 Designed by AK Network Solutions

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In
No Result
View All Result
  • Artificial Intelligence
  • Data Science
    • Language R
    • Deep Learning
    • Tableau
  • Machine Learning
  • Python
  • Blockchain
  • Crypto
  • Big Data
  • NFT
  • Technology
  • Interview Questions
  • Others
    • News
    • Startups
    • Books

© 2023. Designed by AK Network Solutions