no code implementations • 26 May 2024 • Shangding Gu, Bilgehan Sel, Yuhao Ding, Lu Wang, QIngwei Lin, Alois Knoll, Ming Jin
In numerous reinforcement learning (RL) problems involving safety-critical systems, a key challenge lies in balancing multiple objectives while simultaneously meeting all stringent safety constraints.
no code implementations • 26 May 2024 • Vanshaj Khattar, Yuhao Ding, Bilgehan Sel, Javad Lavaei, Ming Jin
Meta-reinforcement learning has widely been used as a learning-to-learn framework to solve unseen tasks with limited experience.
no code implementations • 21 May 2024 • Bilgehan Sel, Priya Shanmugasundaram, Mohammad Kachuee, Kun Zhou, Ruoxi Jia, Ming Jin
Large Language Models (LLMs) have shown remarkable capabilities in tasks such as summarization, arithmetic reasoning, and question answering.
2 code implementations • 2 May 2024 • Shangding Gu, Bilgehan Sel, Yuhao Ding, Lu Wang, QIngwei Lin, Ming Jin, Alois Knoll
Ensuring the safety of Reinforcement Learning (RL) is crucial for its deployment in real-world applications.
no code implementations • 20 Aug 2023 • Ming Jin, Bilgehan Sel, Fnu Hardeep, Wotao Yin
This paper outlines a natural conversational approach to solving personalized energy-related problems using large language models (LLMs).
no code implementations • 20 Aug 2023 • Bilgehan Sel, Ahmad Al-Tawaha, Vanshaj Khattar, Ruoxi Jia, Ming Jin
Current literature, aiming to surpass the "Chain-of-Thought" approach, often resorts to an external modus operandi involving halting, modifying, and then resuming the generation process to boost Large Language Models' (LLMs) reasoning capacities.
no code implementations • 2 Dec 2022 • Ming Jin, Vanshaj Khattar, Harshal Kaushik, Bilgehan Sel, Ruoxi Jia
We study the expressibility and learnability of convex optimization solution functions and their multi-layer architectural extension.