Skip to content
calculated | content

calculated | content

  • About
  • Contact
  • Home

Thoughts on Data Science, Machine Learning, and AI

  • About
  • Contact
  • Home

WeightWatcher, HTSR theory, and the Renormalization Group

There is a deep connection between the open-source weightwatcher tool, which implements ideas from the theory of Heavy Tailed Self-Regularization…

AI, artificial-intelligence, Deep Learning, llms, physics

Fine-Tuned Llama3.2: Bad Instructions ?

Recently, Meta released LLama3.2 1B and 3B Instruct Fine Tuned LLM. To mixed reviews. On the one hand, it’s ranking…

AI, artificial-intelligence, llm, machine learning, technology

What’s instructive about Instruct Fine-Tuning: a weightwatcher analysis

Are you Fine-Tuning an open-source LLMs ? Like Llama, Mistral, or Qwen? A That is, Instruct Fine Tuning. Whether you…

Describing Double Descent with WeightWatcher

Double Descent (DD) is something that has surprised statisticians, computer scientists, and deep learning practitioners–but it was known in the…

SVDSmoothing LLM Layers with WeightWatcher

Recently, Microsoft Research published the LASER method: ”Layer-Selective Rank Reduction” in this recent, very popular paper The Truth is in There:…

AI, artificial-intelligence, DATA SCIENCE, llm, machine learning

Evaluating LLMs with WeightWatcher Part III: The Magic of Mistral, a Story of Dragon Kings

Recently, the Mistral models have taken the LLM world by storm. The Mistral Mixture of Experts (MOE) 8x7b model outperforms other…

AI, artificial-intelligence, llm, machine learning, python

Evaluating Fine-Tuned LLMs with WeightWatcher Part II: PEFT / LoRa Models

Evaluating LLMs is hard. Especially when you don’t have a lot of test data.In the last post, we saw how to…

Deep Learning, Fine Tune, Fine Tuning, llm, LORA, PEFT

Evaluating Fine-Tuned LLMs with WeightWatcher

if you are fine-tuning your own LLMs, you need a way to evaluate them. And while there are over a dozen…

AI, artificial-intelligence, generative-ai, llm, machine learning

WeightWatcher new feature: fix_fingers=’clip_xmax’

WeightWatcher 0.7 has just been released, and it includes the new and improved advanced feature for analyzing Deep Neural Networks…

Deep Learning, Deep Neural Networks

WeightWatcher 0.7: March 2023

First, let me say thanks to all the users in our great community — we have reached over 93K downloads…

Deep Learning, Deep Neural Networks

Posts navigation

Older posts

Recent Posts

  • WeightWatcher, HTSR theory, and the Renormalization Group
  • Fine-Tuned Llama3.2: Bad Instructions ?
  • What’s instructive about Instruct Fine-Tuning: a weightwatcher analysis
  • Describing Double Descent with WeightWatcher
  • SVDSmoothing LLM Layers with WeightWatcher

Archives

  • December 2024
  • October 2024
  • March 2024
  • February 2024
  • January 2024
  • March 2023
  • February 2023
  • July 2022
  • June 2022
  • October 2021
  • August 2021
  • July 2021
  • April 2021
  • November 2020
  • September 2020
  • February 2020
  • December 2019
  • April 2019
  • December 2018
  • November 2018
  • October 2018
  • September 2018
  • June 2018
  • April 2018
  • December 2017
  • September 2017
  • July 2017
  • June 2017
  • February 2017
  • January 2017
  • October 2016
  • September 2016
  • June 2016
  • February 2016
  • December 2015
  • April 2015
  • March 2015
  • January 2015
  • November 2014
  • September 2014
  • August 2014
  • November 2013
  • October 2013
  • August 2013
  • May 2013
  • April 2013
  • December 2012
  • November 2012
  • October 2012
  • September 2012
  • April 2012
  • February 2012
Blog at WordPress.com.
  • Subscribe Subscribed
    • calculated | content
    • Already have a WordPress.com account? Log in now.
    • calculated | content
    • Subscribe Subscribed
    • Sign up
    • Log in
    • Report this content
    • View site in Reader
    • Manage subscriptions
    • Collapse this bar
 

Loading Comments...