Newest reinforcement learning Solutions for 2024

Explore cutting-edge reinforcement learning tools launched in 2024. Perfect for staying ahead in your field.

reinforcement learning

  • Open-source framework offering reinforcement learning-based cryptocurrency trading agents with backtesting, live trading integration, and performance tracking.
    0
    0
    What is CryptoTrader Agents?
    CryptoTrader Agents provides a comprehensive toolkit for designing, training, and deploying AI-driven trading strategies in cryptocurrency markets. It includes a modular environment for data ingestion, feature engineering, and custom reward functions. Users can leverage preconfigured reinforcement learning algorithms or integrate their own models. The platform offers simulated backtesting on historical price data, risk management controls, and detailed metric tracking. When ready, agents can connect to live exchange APIs for automated execution. Built on Python, the framework is fully extensible, enabling users to prototype new tactics, run parameter sweeps, and monitor performance in real time.
  • A high-performance Python framework delivering fast, modular reinforcement learning algorithms with multi-environment support.
    0
    0
    What is Fast Reinforcement Learning?
    Fast Reinforcement Learning is a specialized Python framework designed to accelerate the development and execution of reinforcement learning agents. It offers out-of-the-box support for popular algorithms such as PPO, A2C, DDPG and SAC, combined with high-throughput vectorized environment management. Users can easily configure policy networks, customize training loops and leverage GPU acceleration for large-scale experiments. The library’s modular design ensures seamless integration with OpenAI Gym environments, enabling researchers and practitioners to prototype, benchmark and deploy agents across a variety of control, game and simulation tasks.
  • DeepSeek R1 is an advanced, open-source AI model specializing in reasoning, math, and coding.
    0
    0
    What is Deepseek R1?
    DeepSeek R1 represents a significant breakthrough in artificial intelligence, delivering top-tier performance in reasoning, mathematics, and coding tasks. Utilizing a sophisticated MoE (Mixture of Experts) architecture with 37B activated parameters and 671B total parameters, DeepSeek R1 implements advanced reinforcement learning techniques to achieve state-of-the-art benchmarks. The model offers robust performance, including 97.3% accuracy on MATH-500 and a 96.3% percentile ranking on Codeforces. Its open-source nature and cost-effective deployment options make it accessible for a wide range of applications.
  • Python-based RL framework implementing deep Q-learning to train an AI agent for Chrome's offline dinosaur game.
    0
    0
    What is Dino Reinforcement Learning?
    Dino Reinforcement Learning offers a comprehensive toolkit for training an AI agent to play the Chrome dinosaur game via reinforcement learning. By integrating with a headless Chrome instance through Selenium, it captures real-time game frames and processes them into state representations optimized for deep Q-network inputs. The framework includes modules for replay memory, epsilon-greedy exploration, convolutional neural network models, and training loops with customizable hyperparameters. Users can monitor training progress via console logs and save checkpoints for later evaluation. Post-training, the agent can be deployed to play live games autonomously or benchmarked against different model architectures. The modular design allows easy substitution of RL algorithms, making it a flexible platform for experimentation.
  • Open source TensorFlow-based Deep Q-Network agent that learns to play Atari Breakout using experience replay and target networks.
    0
    0
    What is DQN-Deep-Q-Network-Atari-Breakout-TensorFlow?
    DQN-Deep-Q-Network-Atari-Breakout-TensorFlow provides a complete implementation of the DQN algorithm tailored for the Atari Breakout environment. It uses a convolutional neural network to approximate Q-values, applies experience replay to break correlations between sequential observations, and employs a periodically updated target network to stabilize training. The agent follows an epsilon-greedy policy for exploration and can be trained from scratch on raw pixel input. The repository includes configuration files, training scripts to monitor reward growth over episodes, evaluation scripts to test trained models, and TensorBoard utilities for visualizing training metrics. Users can adjust hyperparameters such as learning rate, replay buffer size, and batch size to experiment with different setups.
  • Open-source PyTorch framework for multi-agent systems to learn and analyze emergent communication protocols in cooperative reinforcement learning tasks.
    0
    0
    What is Emergent Communication in Agents?
    Emergent Communication in Agents is an open-source PyTorch framework designed for researchers exploring how multi-agent systems develop their own communication protocols. The library offers flexible implementations of cooperative reinforcement learning tasks, including referential games, combination games, and object identification challenges. Users define speaker and listener agent architectures, specify message channel properties like vocabulary size and sequence length, and select training strategies such as policy gradients or supervised learning. The framework includes end-to-end scripts for running experiments, analyzing communication efficiency, and visualizing emergent languages. Its modular design allows easy extension with new game environments or custom loss functions. Researchers can reproduce published studies, benchmark new algorithms, and probe compositionality and semantics of emergent agent languages.
  • Gym-Recsys provides customizable OpenAI Gym environments for scalable training and evaluation of reinforcement learning recommendation agents
    0
    0
    What is Gym-Recsys?
    Gym-Recsys is a toolbox that wraps recommendation tasks into OpenAI Gym environments, allowing reinforcement learning algorithms to interact with simulated user-item matrices step by step. It provides synthetic user behavior generators, supports loading popular datasets, and delivers standard recommendation metrics like Precision@K and NDCG. Users can customize reward functions, user models, and item pools to experiment with different RL-based recommendation strategies in a reproducible manner.
  • FlowRL AI enables real-time metric-driven UI personalization using reinforcement learning.
    0
    0
    What is flowRL?
    FlowRL AI is a powerful platform that provides real-time UI personalization using reinforcement learning. By tailoring the user interface to meet individual user needs and preferences, FlowRL drives significant improvements in key business metrics. The platform is designed to dynamically adjust UI elements based on live data, enabling businesses to deliver highly personalized user experiences that increase engagement and conversion rates.
  • A collection of customizable grid-world environments compatible with OpenAI Gym for reinforcement learning algorithm development and testing.
    0
    0
    What is GridWorldEnvs?
    GridWorldEnvs offers a comprehensive suite of grid-world environments to support the design, testing, and benchmarking of reinforcement learning and multi-agent systems. Users can easily configure grid dimensions, agent start positions, goal locations, obstacles, reward structures, and action spaces. The library includes ready-to-use templates such as classic grid navigation, obstacle avoidance, and cooperative tasks, while also allowing custom scenario definitions via JSON or Python classes. Seamless integration with the OpenAI Gym API means that standard RL algorithms can be applied directly. Additionally, GridWorldEnvs supports single-agent and multi-agent experiments, logging, and visualization utilities for tracking agent performance.
  • gym-fx provides a customizable OpenAI Gym environment to train and evaluate reinforcement learning agents for Forex trading strategies.
    0
    0
    What is gym-fx?
    gym-fx is an open-source Python library that implements a simulated Forex trading environment using the OpenAI Gym interface. It offers support for multiple currency pairs, integration of historical price feeds, technical indicators, and fully customizable reward functions. By providing a standardized API, gym-fx simplifies the process of benchmarking and developing reinforcement learning algorithms for algorithmic trading. Users can configure market slippage, transaction costs, and observation spaces to closely mimic live trading scenarios, facilitating robust strategy development and evaluation.
  • gym-llm offers Gym-style environments for benchmarking and training LLM agents on conversational and decision-making tasks.
    0
    0
    What is gym-llm?
    gym-llm extends the OpenAI Gym ecosystem to large language models by defining text-based environments where LLM agents interact through prompts and actions. Each environment follows Gym’s step, reset, and render conventions, emitting observations as text and accepting model-generated responses as actions. Developers can craft custom tasks by specifying prompt templates, reward calculations, and termination conditions, enabling sophisticated decision-making and conversational benchmarks. Integration with popular RL libraries, logging tools, and configurable evaluation metrics facilitates end-to-end experimentation. Whether assessing an LLM’s ability to solve puzzles, manage dialogues, or navigate structured tasks, gym-llm provides a standardized, reproducible framework for research and development of advanced language agents.
  • A Python-based OpenAI Gym environment offering customizable multi-room gridworlds for reinforcement learning agents’ navigation and exploration research.
    0
    0
    What is gym-multigrid?
    gym-multigrid provides a suite of customizable gridworld environments designed for multi-room navigation and exploration tasks in reinforcement learning. Each environment consists of interconnected rooms populated with objects, keys, doors, and obstacles. Users can adjust grid size, room configurations, and object placements programmatically. The library supports both full and partial observation modes, offering RGB or matrix state representations. Actions include movement, object interaction, and door manipulation. By integrating it as a Gym environment, researchers can leverage any Gym-compatible agent, seamlessly training and evaluating algorithms on tasks like key-door puzzles, object retrieval, and hierarchical planning. gym-multigrid’s modular design and minimal dependencies make it ideal for benchmarking new AI strategies.
  • HFO_DQN is a reinforcement learning framework that applies Deep Q-Network to train soccer agents in RoboCup Half Field Offense environment.
    0
    0
    What is HFO_DQN?
    HFO_DQN combines Python and TensorFlow to deliver a complete pipeline for training soccer agents using Deep Q-Networks. Users can clone the repository, install dependencies including the HFO simulator and Python libraries, and configure training parameters in YAML files. The framework implements experience replay, target network updates, epsilon-greedy exploration, and reward shaping tailored for the half field offense domain. It features scripts for agent training, performance logging, evaluation matches, and plotting results. Modular code structure allows integration of custom neural network architectures, alternative RL algorithms, and multi-agent coordination strategies. Outputs include trained models, performance metrics, and behavior visualizations, facilitating research in reinforcement learning and multi-agent systems.
  • Jason-RL equips Jason BDI agents with reinforcement learning, enabling Q-learning and SARSA-based adaptive decision making through reward experience.
    0
    0
    What is jason-RL?
    jason-RL adds a reinforcement learning layer to the Jason multi-agent framework, allowing AgentSpeak BDI agents to learn action-selection policies via reward feedback. It implements Q-learning and SARSA algorithms, supports configuration of learning parameters (learning rate, discount factor, exploration strategy), and logs training metrics. By defining reward functions in agent plans and running simulations, developers can observe agents improve decision making over time, adapting to changing environments without manual policy coding.
  • MARFT is an open-source multi-agent RL fine-tuning toolkit for collaborative AI workflows and language model optimization.
    0
    0
    What is MARFT?
    MARFT is a Python-based LLMs, enabling reproducible experiments and rapid prototyping of collaborative AI systems.
  • An open-source Minecraft-inspired RL platform enabling AI agents to learn complex tasks in customizable 3D sandbox environments.
    0
    0
    What is MineLand?
    MineLand provides a flexible 3D sandbox environment inspired by Minecraft for training reinforcement learning agents. It features Gym-compatible APIs for seamless integration with existing RL libraries such as Stable Baselines, RLlib, and custom implementations. Users gain access to a library of tasks, including resource collection, navigation, and construction challenges, each with configurable difficulty and reward structures. Real-time rendering, multi-agent scenarios, and headless modes allow for scalable training and benchmarking. Developers can design new maps, define custom reward functions, and plugin additional sensors or controls. MineLand’s open-source codebase fosters reproducible research, collaborative development, and rapid prototyping of AI agents in complex virtual worlds.
  • Simplified PyTorch implementation of AlphaStar, enabling StarCraft II RL agent training with modular network architecture and self-play.
    0
    0
    What is mini-AlphaStar?
    mini-AlphaStar demystifies the complex AlphaStar architecture by offering an accessible, open-source PyTorch framework for StarCraft II AI development. It features spatial feature encoders for screen and minimap inputs, non-spatial feature processing, LSTM memory modules, and separate policy and value networks for action selection and state evaluation. Using imitation learning to bootstrap and reinforcement learning with self-play for fine-tuning, it supports environment wrappers compatible with StarCraft II via pysc2, logging through TensorBoard, and configurable hyperparameters. Researchers and students can generate datasets from human gameplay, train models on custom scenarios, evaluate agent performance, and visualize learning curves. The modular codebase enables easy experimentation with network variants, training schedules, and multi-agent setups. Designed for education and prototyping rather than production deployment.
  • A Unity ML-Agents based environment for training cooperative multi-agent inspection tasks in customizable 3D virtual scenarios.
    0
    0
    What is Multi-Agent Inspection Simulation?
    Multi-Agent Inspection Simulation provides a comprehensive framework for simulating and training multiple autonomous agents to perform inspection tasks cooperatively within Unity 3D environments. It integrates with the Unity ML-Agents toolkit, offering configurable scenes with inspection targets, adjustable reward functions, and agent behavior parameters. Researchers can script custom environments, define the number of agents, and set training curricula via Python APIs. The package supports parallel training sessions, TensorBoard logging, and customizable observations including raycasts, camera feeds, and positional data. By adjusting hyperparameters and environment complexity, users can benchmark reinforcement learning algorithms on coverage, efficiency, and coordination metrics. The open-source codebase encourages extension for robotics prototyping, cooperative AI research, and educational demonstrations in multi-agent systems.
  • Open-source Python environment for training AI agents to cooperatively surveil and detect intruders in grid-based scenarios.
    0
    0
    What is Multi-Agent Surveillance?
    Multi-Agent Surveillance offers a flexible simulation framework where multiple AI agents act as predators or evaders in a discrete grid world. Users can configure environment parameters such as grid dimensions, number of agents, detection radii, and reward structures. The repository includes Python classes for agent behavior, scenario generation scripts, built-in visualization via matplotlib, and seamless integration with popular reinforcement learning libraries. This makes it easy to benchmark multi-agent coordination, develop custom surveillance strategies, and conduct reproducible experiments.
  • An open-source Python simulation environment for training cooperative drone swarm control with multi-agent reinforcement learning.
    0
    0
    What is Multi-Agent Drone Environment?
    Multi-Agent Drone Environment is a Python package offering a customizable multi-agent simulation for UAV swarms, built on OpenAI Gym and PyBullet. Users define multiple drone agents with kinematic and dynamic models to explore cooperative tasks such as formation flying, target tracking, and obstacle avoidance. The environment supports modular task configuration, realistic collision detection, and sensor emulation, while allowing custom reward functions and decentralized policies. Developers can integrate their own reinforcement learning algorithms, evaluate performance under varied scenarios, and visualize agent trajectories and metrics in real time. Its open-source design encourages community contributions, making it ideal for research, teaching, and prototyping advanced multi-agent control solutions.
Featured
Flowith
Flowith is a canvas-based agentic workspace which offers free 🍌Nano Banana Pro and other effective models...
Refly.ai
Refly.AI empowers non-technical creators to automate workflows using natural language and a visual canvas.
BGRemover
Easily remove image backgrounds online with SharkFoto BGRemover.
FineVoice
Clone, Design, and Create Expressive AI Voices in Seconds, with Perfect Sound Effects and Music.
Elser AI
All-in-one AI video creation studio that turns any text and images into full videos up to 30 minutes.
FixArt AI
FixArt AI offers free, unrestricted AI tools for image and video generation without sign-up.
Qoder
Qoder is an agentic coding platform for real software, Free to use the best model in preview.
Skywork.ai
Skywork AI is an innovative tool to enhance productivity using AI.
Yollo AI
Chat & create with your AI companion. Image to Video, AI Image Generator.
VoxDeck
Next-gen AI presentation maker,Turn your ideas & docs into attention-grabbing slides with AI.
Funy AI
AI bikini & kiss videos from images or text. Try the AI Clothes Changer & Image Generator!
SharkFoto
SharkFoto is an all-in-one AI-powered platform for creating and editing videos, images, and music efficiently.
ThumbnailCreator.com
AI-powered tool for creating stunning, professional YouTube thumbnails quickly and easily.
Pippit
Elevate your content creation with Pippit's powerful AI tools!
SuperMaker AI Video Generator
Create stunning videos, music, and images effortlessly with SuperMaker.
AnimeShorts
Create stunning anime shorts effortlessly with cutting-edge AI technology.
Img2.AI
AI platform that converts photos into stylized images and short animated videos with fast, high-quality results and one-click upscaling.
Nana Banana: Advanced AI Image Editor
AI-powered image editor turning photos and text prompts into high-quality, consistent, commercial-ready images for creators and brands.
Van Gogh Free Video Generator
An AI-powered free video generator that creates stunning videos from text and images effortlessly.
Create WhatsApp Link
Free WhatsApp link and QR generator with analytics, branded links, routing, and multi-agent chat features.
Gobii
Gobii lets teams create 24/7 autonomous digital workers to automate web research and routine tasks.
AI FIRST
Conversational AI assistant automating research, browser tasks, web scraping, and file management through natural language.
Kling 3.0
Kling 3.0 is an AI-powered 4K video generator with native audio, advanced motion control, and Canvas Agent.
TextToHuman
Free AI humanizer that instantly rewrites AI text into natural, human-like writing. No signup required.
GLM Image
GLM Image combines hybrid AR and diffusion models to generate high-fidelity AI images with exceptional text rendering.
AirMusic
AirMusic.ai generates high-quality AI music tracks from text prompts with style, mood customization, and stems export.
Manga Translator AI
AI Manga Translator instantly translates manga images into multiple languages online.
LTX-2 AI
Open-source LTX-2 generates 4K videos with native audio sync from text or image prompts, fast and production-ready.
WhatsApp Warmup Tool
AI-powered WhatsApp warmup tool automates bulk messaging while preventing account bans.
Qwen-Image-2512 AI
Qwen-Image-2512 is a fast, high-resolution AI image generator with native Chinese text support.
FalcoCut
FalcoCut: web-based AI platform for video translation, avatar videos, voice cloning, face-swap and short video generation.
ai song creator
Create full-length, royalty-free AI-generated music up to 8 minutes with commercial license.
SOLM8
AI girlfriend you call, and chat with. Real voice conversations with memory. Every moment feels special with her.
Telegram Group Bot
TGDesk is an all-in-one Telegram Group Bot to capture leads, boost engagement, and grow communities.
Remy - Newsletter Summarizer
Remy automates newsletter management by summarizing emails into digestible insights.
APIMart
APIMart offers unified access to 500+ AI models including GPT-5 and Claude 4.5 with cost savings.
Vertech Academy
Vertech offers AI prompts designed to help students and teachers learn and teach effectively.
RSW Sora 2 AI Studio
Remove Sora watermark instantly with AI-powered tool for zero quality loss and fast downloads.
PoYo API
PoYo.ai is a unified AI API platform for image, video, music and chat generation, built for developers.
Explee
Start outreach RIGHT NOW with single-line description of your ICP
Seedance 1.5 Pro
Seedance 1.5 Pro is an AI-powered cinematic video generator with perfect lip-sync and real-time audio-video sync.
Lease A Brain
AI-powered team of expert virtual professionals ready to assist in diverse business tasks. Sign-up for a free trial.
Rebelgrowth
Grow your revenue from organic traffic on autopilot: Keyword research. SEO optimized articles and EVEN backlinks.
Edensign
Edensign is an AI-driven virtual staging platform transforming real estate photos quickly and realistically.
NanoPic
NanoPic offers fast, high-quality conversational image editing powered by AI with 2K/4K output.
codeflying
CodeFlying – Vibe Coding App Builder | Create Full-Stack Apps by Chatting with AI
remio - Personal AI Assistant
remio is an AI-powered personal knowledge hub that captures and organizes all your digital info automatically.
TattooAI AI Tattoo Generator
AI Tattoo Generator creates personalized, high-quality tattoo designs quickly with advanced AI technology.
Camtasia online
Camtasia Online is a free tool for screen recording and video editing, all from your web browser.
Avoid.so
Avoid.so offers advanced AI humanizer technology to bypass AI detection algorithms seamlessly.
Wollo.ai
Wollo allows you to create, explore, and chat with AI characters using advanced, emotionally aware AI technology.
Chatronix
LLM aggregator that connects multiple AI models in one platform for comparison, integration, and automation.
Vadu AI
All-in-one AI video & image generator with Sora 2, Veo 3, Kling, and 10+ top models.