AMD OLMo 1B Language Models Performance In Benchmarks
AMD OLMo is a set of 1 billion parameter language models that have been pre-trained on 16 nodes with four (4) AMD Instinct MI250 GPUs and 1.3 trillion tokens
AMD OLMo 1B SFT DPO: Using the UltraFeedback dataset and Direct Preference Optimization (DPO), this model is in line with human preferences
AMD OLMo 1B is based on the model architecture and training configuration of the completely open source 1 billion version of OLMo
Next-token prediction is used to train the AMD OLMo models, which are transformer language models that solely use decoders
It contrast AMD OLMo models with other completely open-source models of comparable scale that have made their training code
How alignment training enables it AMD OLMo 1B SFT DPO model to function similarly to other conversation baselines on responsible AI assessment benchmarks
Additionally, AMD Ryzen AI PCs with NPUs, which may assist allow a wide range of edge use cases, were equipped with the language model