AI Research & Reflections

✦ Exploring Life, RL, model alignment, and diffusion models ✦

🧠 Research 🤖 AI Safety ✨ Blog ✨ Resume

👋 About Me

🎯

I'm a researcher exploring the intersection of AI safety and capability. My work focuses on understanding and mitigating LLM hallucinations, improving model alignment, and advancing diffusion model techniques.

📝

This blog documents my research, projects, and daily observations from the frontiers of AI development. Join me on this journey through the complexities of artificial intelligence.

📚 Featured Articles

#1

Understanding LLM Hallucinations

An exploration of why language models generate false information and what we can do about it.

#llm#alignment
Read More →
#2

Diffusion Models Explained

A deep dive into how diffusion models work and their applications in AI generation.

#diffusion#generative-ai
Read More →
#3

Model Alignment Challenges

The technical and philosophical challenges in aligning AI systems with human values.

#alignment#safety
Read More →
View all articles

🚀 Current Projects

Hallucination Detection Tool

In Progress

Building a system to detect and flag potential hallucinations in LLM outputs.

Alignment Metrics Dashboard

In Progress

Visualizing alignment metrics across different model architectures.

Reinforcement Learning Trading Agents

Exection

Developing RL agents that can adapt to dynamic market conditions for optimized trading strategies.

💭 Recent Thoughts

The gap between capability and alignment is growing faster than we anticipated.

🕐 Nov 9

Interesting paper on constitutional AI - game changer for scalable oversight?

🕐 Nov 8

Spent the day debugging a diffusion model. The loss curves never lie.

🕐 Nov 7
More thoughts