A letter from May 05, 2025

Time Travelled — 12 months

Peaceful right?

Dear FutureMe, Over the past term, diving into the Gymnasium API and the Atari Video Pinball environment transformed my understanding of reinforcement learning from abstract theory into hands-on practice. I went from exploring simple Q-tables and SARSA on toy problems to building deep function approximators with DQN and its Double-DQN variant. I implemented policy-gradient methods—A2C and A3C—to see how on-policy learning trades off sample efficiency and stability. Pushing even further, I coded Proximal Policy Optimization (PPO) from scratch, witnessing firsthand how its clipped surrogate objective tames policy updates and accelerates convergence. Working across these algorithms sharpened my Python programming skills—I became fluent with vectorized NumPy operations, custom Gymnasium wrappers, and PyTorch/TensorFlow model definitions. I developed robust training pipelines: automated checkpointing, learning-rate schedules, and logging with TensorBoard. Debugging high-variance returns forced me to become a better problem solver, learning to isolate sources of instability (network initialization, reward scaling, exploration schedules) and apply rigorous hyperparameter sweeps. Above all, I gained an appreciation for reproducible research: version-controlled code, seed management, and clear experiment tracking in MLflow. Skills I’ve Gained Deep RL Algorithms: From tabular Q-learning and SARSA to replay-based DQN/DDQN, and advanced actor-critic methods (A2C/A3C) through to PPO. Environment Engineering: Mastery of Gymnasium wrappers (frame skip, grayscale, frame-stack) and custom preprocessing for 84×84×4 state representations. Modeling & Optimization: Building CNN backbones, tuning A***/RMSProp optimizers, implementing entropy bonuses, and crafting clipped objectives. Debugging & Analysis: Diagnosing exploding gradients, reward sparsity issues, and implementing learning-curve visualizations to guide iterative improvements. Software Best Practices: Modular code design, unit tests for custom environments, reproducible experiment tracking (Git, MLflow), and cloud-based training orchestration. Collaboration & Communication: Sharing clear Jupyter Notebook reports, drafting LaTeX write-ups of algorithmic details, and presenting results both in slides and in writing. Goals for the Coming Year Explore Advanced Architectures: Implement Rainbow DQN, Soft Actor-Critic (SAC), and Twin Delayed DDPG (TD3) to compare performance on sparse-reward tasks. Real-World Applications: Apply RL to real-time robotics control or resource allocation problems, bridging simulation and hardware. Scalable Training Pipelines: Learn distributed RL frameworks (e.g., RLlib, Acme) and deploy experiments at scale using Kubernetes and Docker. Research & Publication: Write up a case study on sample-efficient RL in sparse environments and aim for submission to a workshop or conference. Continued Professional Growth: Deepen my understanding of theoretical foundations—convergence proofs, policy iteration theory—and mentor peers in study groups or workshops. Through these steps, I plan to build on the solid foundation this course provided, pushing both my practical skills and theoretical insight in reinforcement learning.

Load more comments

Sign in to FutureMe

or use your email address

Don't know your password? Sign in with an email link instead.

By signing in to FutureMe you agree to the Terms of use.

Create an account

or use your email address

You will receive a confirmation email

By signing in to FutureMe you agree to the Terms of use.

Share this FutureMe letter

Copy the link to your clipboard:

Or share directly via social media:

Why is this inappropriate?