AMD OLMo 1B Language Models Performance In Benchmarks

AMD OLMo is a set of 1 billion parameter language models that have been pre-trained on 16 nodes with four (4) AMD Instinct MI250 GPUs and 1.3 trillion tokens

AMD OLMo 1B SFT DPO: Using the UltraFeedback dataset and Direct Preference Optimization (DPO), this model is in line with human preferences

AMD OLMo 1B is based on the model architecture and training configuration of the completely open source 1 billion version of OLMo

Next-token prediction is used to train the AMD OLMo models, which are transformer language models that solely use decoders

It contrast AMD OLMo models with other completely open-source models of comparable scale that have made their training code

How alignment training enables it AMD OLMo 1B SFT DPO model to function similarly to other conversation baselines on responsible AI assessment benchmarks

Additionally, AMD Ryzen AI PCs with NPUs, which may assist allow a wide range of edge use cases, were equipped with the language model