Chromatographic problem solving, commonly referred to as method development (MD), is hugely complex, given the many operational parameters that must be optimized and their large effect on the elution times of individual sample compounds. Recently, the use of reinforcement learning has been proposed to automate and expedite this process for liquid chromatography (LC). This study further explores deep reinforcement learning (RL) for LC method development. Given the large training budgets required, an in-silico approach was taken to train several Proximal Policy Optimization (PPO) agents. High-performing PPO agents were trained using sparse rewards (=rewarding only when all sample components were fully separated) and large experimental budgets. Strategies like frame stacking or long short-term memory networks were also investigated to improve the agents further. The trained agents were benchmarked against a Bayesian Optimization (BO) algorithm using a set of 1000 randomly-composed samples. Both algorithms were tasked with finding gradient programs that fully resolved all compounds in the samples, using a minimal number of experiments. When the number of parameters to tune was limited (single-segment gradient programs) PPO required on average, 1 to 2 fewer experiments, but did not outperform BO with respect to the number of solutions found, with PPO and BO solving 17% and 19% of the most challenging samples, respectively. However, PPO excelled at more complex tasks involving a higher number of parameters. As an example, when optimizing a five-segment gradient PPO solved 31% of samples, while BO solved 24% of samples.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1016/j.chroma.2025.465845 | DOI Listing |
Bull Math Biol
March 2025
Department of Mathematics, University of Illinois Urbana-Champaign, Urbana, IL, USA.
Two mechanisms that have been used to study the evolution of cooperative behavior are altruistic punishment, in which cooperative individuals pay additional costs to punish defection, and multilevel selection, in which competition between groups can help to counteract individual-level incentives to cheat. Boyd, Gintis, Bowles, and Richerson have used simulation models of cultural evolution to suggest that altruistic punishment and pairwise group-level competition can work in concert to promote cooperation, even when neither mechanism can do so on its own. In this paper, we formulate a PDE model for multilevel selection motivated by the approach of Boyd and coauthors, modeling individual-level birth-death competition with a replicator equation based on individual payoffs and describing group-level competition with pairwise conflicts based on differences in the average payoffs of the competing groups.
View Article and Find Full Text PDFJ Exp Anal Behav
March 2025
Behavioral Neuroscience Laboratory, Department of Psychology, Boğaziçi University, Istanbul, Turkey.
Robots are increasingly used alongside Skinner boxes to train animals in operant conditioning tasks. Similarly, animals are being employed in artificial intelligence research to train various algorithms. However, both types of experiments rely on unidirectional learning, where one partner-the animal or the robot-acts as the teacher and the other as the student.
View Article and Find Full Text PDFJ Hum Lact
March 2025
Jersey Shore University Medical Center, Hackensack Meridian Health, Neptune, NJ, USA.
Maintaining Baby-Friendly Hospital Initiative (BFHI) standards within a complex healthcare system presents unique challenges. This case study from a regional perinatal center in the northeast United States details the design and implementation of a program to address BFHI Step 2, which requires ongoing competency assessment and team member training to ensure breastfeeding support. The shift of BFHI competencies to continuous professional development introduced logistical challenges, compounded by staff turnover and budget constraints.
View Article and Find Full Text PDFFront Artif Intell
February 2025
Department of Computer Science & Engineering, Indian Institute of Technology Ropar, Rupnagar, India.
Machine learning techniques have emerged as a promising tool for efficient cache management, helping optimize cache performance and fortify against security threats. The range of machine learning is vast, from reinforcement learning-based cache replacement policies to Long Short-Term Memory (LSTM) models predicting content characteristics for caching decisions. Diverse techniques such as imitation learning, reinforcement learning, and neural networks are extensively useful in cache-based attack detection, dynamic cache management, and content caching in edge networks.
View Article and Find Full Text PDFInd Eng Chem Res
March 2025
Department of Chemical Engineering, Imperial College London, London, South Kensington SW7 2AZ, U.K.
This work proposes a control-informed reinforcement learning (CIRL) framework that integrates proportional-integral-derivative (PID) control components into the architecture of deep reinforcement learning (RL) policies, incorporating prior knowledge from control theory into the learning process. CIRL improves performance and robustness by combining the best of both worlds: the disturbance-rejection and set point-tracking capabilities of PID control and the nonlinear modeling capacity of deep RL. Simulation studies conducted on a continuously stirred tank reactor system demonstrate the improved performance of CIRL compared to both conventional model-free deep RL and static PID controllers.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!