PyXAB - A Python Library for X-Armed Bandit and Online Blackbox Optimization Algorithms
-
Updated
Oct 24, 2024 - Python
PyXAB - A Python Library for X-Armed Bandit and Online Blackbox Optimization Algorithms
Yahoo! news article recommendation system by linUCB
Bandit algorithms
Python implementation of UCB, EXP3 and Epsilon greedy algorithms
Personal reimplementation of some ML algorithms for learning purposes
A benchmark to test decision-making algorithms for contextual-bandits. The library implements a variety of algorithms (many of them based on approximate Bayesian Neural Networks and Thompson sampling), and a number of real and syntethic data problems exhibiting a diverse set of properties.
This repository aims at learning most popular MAB and CMAB algorithms and watch how they run. It is interesting for those wishing to start learning these topics.
🐯REPLICA of "Auction-based combinatorial multi-armed bandit mechanisms with strategic arms"
Pricing and advertising strategy for the e-commerce of an airline company, based on Multi-Armed Bandits (MABs) algorithms and Gaussian Processes. Simulations include non-stationary environments.
Python library of bandits and RL agents in different real-world environments
A short implementation of bandit algorithms - ETC, UCB, MOSS and KL-UCB
Non-stationary Bandits and Meta-Learning with a Small Set of Optimal Arms
Python implementation for Reinforcement Learning algorithms -- Bandit algorithms, MDP, Dynamic Programming (value/policy iteration), Model-free Control (off-policy Monte Carlo, Q-learning)
Multi-Objective Multi-Armed Bandit
Implementation for NeurIPS 2020 paper "Locally Differentially Private (Contextual) Bandits Learning" (https://arxiv.org/abs/2006.00701)
A comprehensive Python library implementing a variety of contextual and non-contextual multi-armed bandit algorithms—including LinUCB, Epsilon-Greedy, Upper Confidence Bound (UCB), Thompson Sampling, KernelUCB, NeuralLinearBandit, and DecisionTreeBandit—designed for reinforcement learning applications
The official code repo for HyperAgent for neural bandits and GPT-HyperAgent for content moderation.
DPE code - Code used in "Optimal Algorithms for Multiplayer Multi-Armed Bandits" (AISTATS 2020)
An illustrative project including some multi-armed bandit algorithms and contextual bandit algorithms
Add a description, image, and links to the bandit-algorithms topic page so that developers can more easily learn about it.
To associate your repository with the bandit-algorithms topic, visit your repo's landing page and select "manage topics."