Skip to content
calculated | content

calculated | content

  • About
  • Contact
  • Home

Thoughts on Data Science, Machine Learning, and AI

  • About
  • Contact
  • Home

Category: Uncategorized

WeightWatcher: Empirical Quality Metrics for Deep Neural Networks

We introduce the weightwatcher (ww) , a python tool for a python tool for computing quality metrics of trained, and … More

AI, DATA SCIENCE, Deep Learning, KERAS, machine learning, NLP, PYTORCH, TENSORFLOW

Towards a new Theory of Learning: Statistical Mechanics of Deep Neural Networks

Introduction For the past few years, we have talked a lot about how we can understand the properties of Deep … More

This Week in Machine Learning and AI: Implicit Self-Regularization

Big thanks to and the team at This Week in Machine Learning and AI for my recent interview: Implicit Self-Regularization … More

SF Bay ACM Talk: Heavy Tailed Self Regularization in Deep Neural Networks

My Collaborator did a great job giving a talk on our research at the local San Francisco Bay ACM Meetup … More

Heavy Tailed Self Regularization in Deep Neural Nets: 1 year of research

My talk at ICSI-the International Computer Science Institute at UC Berkeley. ICSI is a leading independent, nonprofit center for research … More

AI, Deep Learning, machine learning, UC Berkeley

Don’t Peek part 2: Predictions without Test Data

This is a followup to a previous post: DON’T PEEK: DEEP LEARNING WITHOUT LOOKING … AT TEST DATA The idea…suppose … More

Machine Learning and AI for the Lean Start Up

Machine Learning and AI for the Lean Start Up My recent talk at the French Tech Hub Startup Accelerator

Don’t Peek: Deep Learning without looking … at test data

What is the purpose of a theory ?  To explain why something works.  Sure.  But what good is a theory … More

Rank Collapse in Deep Learning

We can learn a lot about Why Deep Learning Works by studying the properties of the layer weight matrices of … More

Power Laws in Deep Learning 2: Universality

Power Law Distributions in Deep Learning In a previous post, we saw that the Fully Connected (FC) layers of the … More

Posts navigation

Older posts
Newer posts

Recent Posts

  • WW-PGD: Projected Gradient Descent optimizer
  • WeightWatcher, HTSR theory, and the Renormalization Group
  • Fine-Tuned Llama3.2: Bad Instructions ?
  • What’s instructive about Instruct Fine-Tuning: a weightwatcher analysis
  • Describing Double Descent with WeightWatcher

Archives

  • December 2025
  • December 2024
  • October 2024
  • March 2024
  • February 2024
  • January 2024
  • March 2023
  • February 2023
  • July 2022
  • June 2022
  • October 2021
  • August 2021
  • July 2021
  • April 2021
  • November 2020
  • September 2020
  • February 2020
  • December 2019
  • April 2019
  • December 2018
  • November 2018
  • October 2018
  • September 2018
  • June 2018
  • April 2018
  • December 2017
  • September 2017
  • July 2017
  • June 2017
  • February 2017
  • January 2017
  • October 2016
  • September 2016
  • June 2016
  • February 2016
  • December 2015
  • April 2015
  • March 2015
  • January 2015
  • November 2014
  • September 2014
  • August 2014
  • November 2013
  • October 2013
  • August 2013
  • May 2013
  • April 2013
  • December 2012
  • November 2012
  • October 2012
  • September 2012
  • April 2012
  • February 2012
Blog at WordPress.com.
  • Subscribe Subscribed
    • calculated | content
    • Join 732 other subscribers
    • Already have a WordPress.com account? Log in now.
    • calculated | content
    • Subscribe Subscribed
    • Sign up
    • Log in
    • Report this content
    • View site in Reader
    • Manage subscriptions
    • Collapse this bar
 

Loading Comments...