bar
heidloff.net - Building is my Passion
Niklas Heidloff
Cancel

Instruction Fine-tuning of Large Language Models

This post explains what ‘instruct’ versions of Large Language Models are and how instructions can be used for efficient fine-tuning. Often there are instruct versions of popular Large Language Mod...

Efficient Fine-tuning with PEFT and LoRA

Classic fine-tuning of Large Language Models typically changes most weights of the models which requires a lot of resources. LoRA based fine-tuning freezes the original weights and only trains a sm...

Running Python locally

There are several ways to run Python code locally. Often it is desired to run it in virtual environments and containers to be able to run multiple configurations in parallel and to easily remove co...

Open Source LLMs in Watsonx.ai

Watsonx.ai is IBM’s offering to train, validate, tune, and deploy generative AI based on foundation models. It comes with several open source models which are briefly described in this post. As a ...

Causal LLMs and Seq2Seq Architectures

The Hugging Face libraries have become the de-facto standard how to access foundation models from Python, both for inference and fine-tuning. This post describes how to use the Hugging Face APIs fo...

Decoding Methods for Generative AI

With watsonx.ai, you can train, validate, tune, and deploy generative AI based on foundation models. This post explains some of the strategies to instruct Large Language Models how to generate text...

Fine-tuning FLAN-T5 for Summarization

FLAN-T5 is a Large Language Model which was open sourced by Google at the end of 2022. It has been fine-tuned on multiple tasks, but can be further fine-tuned. This post explains how to do this via...

IBM announces new Foundation Model Capabilities

At IBM Think 2023 several exciting new Foundation Model capabilities have been announced. Below are some of my highlights. Most of the content below is from the great talk from Dr. Darío Gil, IBM ...

Running Generative AI Experiments for Question Answering

To find the best possible models and parameters for Question Answering via Generative AI, a lot of experiments need to be run. While some techniques have been proven successful, other approaches ne...

Optimizing Generative AI for Question Answering

Transformer based AI models can generate amazing answers to users’ questions. While the underlaying Large Language Models are not retrained, the performance of Question Answering AI can be improved...

Disclaimer
The postings on this site are my own and don’t necessarily represent IBM’s positions, strategies or opinions.
Trending Tags