Tim Bula's picture

Tim Bula

timrbula

AI & ML interests

LLMs for language and code

Recent Activity

liked a model 5 days ago
mistralai/Mistral-Small-24B-Base-2501
reacted to singhsidhukuldeep's post with ❤️ 8 days ago
It's not every day you see a research paper named "Alice's Adventures in a Differentiable Wonderland," and when you open it, it's a 281-page book! I haven't completed it yet, but this amazing work, written by Simone Scardapane, is a fascinating introduction to deep neural networks and differentiable programming. Some key technical highlights: • Covers core concepts like automatic differentiation, stochastic optimization, and activation functions in depth • Explains modern architectures like convolutional networks, transformers, and graph neural networks • Provides mathematical foundations including linear algebra, gradients, and probability theory • Discusses implementation details in PyTorch and JAX • Explores advanced topics like Bayesian neural networks and neural scaling laws The book takes a unique approach, framing neural networks as compositions of differentiable primitives rather than biological analogs. It provides both theoretical insights and practical coding examples. I especially enjoyed the sections on: • Vector-Jacobian products and reverse-mode autodiff • Stochastic gradient descent and mini-batch optimization • ReLU, GELU, and other modern activation functions • Universal approximation capabilities of MLPs Whether you're new to deep learning or an experienced practitioner, this book offers valuable insights into the fundamentals and latest developments. Highly recommended for anyone working with neural networks!
liked a model 19 days ago
ibm-granite/granite-3.1-3b-a800m-base
View all activity

Organizations

IBM's profile picture

timrbula's activity

reacted to singhsidhukuldeep's post with ❤️ 8 days ago
view post
Post
2199
It's not every day you see a research paper named "Alice's Adventures in a Differentiable Wonderland," and when you open it, it's a 281-page book!

I haven't completed it yet, but this amazing work, written by Simone Scardapane, is a fascinating introduction to deep neural networks and differentiable programming.

Some key technical highlights:

• Covers core concepts like automatic differentiation, stochastic optimization, and activation functions in depth

• Explains modern architectures like convolutional networks, transformers, and graph neural networks

• Provides mathematical foundations including linear algebra, gradients, and probability theory

• Discusses implementation details in PyTorch and JAX

• Explores advanced topics like Bayesian neural networks and neural scaling laws

The book takes a unique approach, framing neural networks as compositions of differentiable primitives rather than biological analogs. It provides both theoretical insights and practical coding examples.

I especially enjoyed the sections on:

• Vector-Jacobian products and reverse-mode autodiff
• Stochastic gradient descent and mini-batch optimization
• ReLU, GELU, and other modern activation functions
• Universal approximation capabilities of MLPs

Whether you're new to deep learning or an experienced practitioner, this book offers valuable insights into the fundamentals and latest developments. Highly recommended for anyone working with neural networks!
reacted to MoritzLaurer's post with 🔥 27 days ago
view post
Post
1708
The TRL v0.13 release is 🔥! My highlight are the new process reward trainer to train models similar to o1 and tool call support:

🧠 Process reward trainer: Enables training of Process-supervised Reward Models (PRMs), which reward the quality of intermediate steps, promoting structured reasoning. Perfect for tasks like stepwise reasoning.

🔀 Model merging: A new callback leverages mergekit to merge models during training, improving performance by blending reference and policy models - optionally pushing merged models to the Hugging Face Hub.

🛠️ Tool call support: TRL preprocessing now supports tool integration, laying the groundwork for agent fine-tuning with examples like dynamic temperature fetching in prompts.

⚖️ Mixture of judges: The new AllTrueJudge combines decisions from multiple binary judges for more nuanced evaluation.

Read the release notes and other resources here 👇
Release: https://github.com/huggingface/trl/releases/tag/v0.13.0
Mergekit: https://github.com/arcee-ai/mergekit
Mixture of judges paper: The Perfect Blend: Redefining RLHF with Mixture of Judges (2409.20370)