This technology has been used in various research projects such as “Training a helpful and harmless assistant with reinforcement learning from human feedback” (Bai et al., 2023), “Red teaming language models to reduce harms: methods, scaling behaviors, and lessons learned” (Ganguli et al., 2022), and “Booksum: a collection of datasets for long-form narrative summarization” (Kryściński et al., 2022). Other notable projects include “CAMEL: communicative agents for mind exploration of large scale language model society” (Li et al., 2024), “The CoT collection: improving zero-shot and few-shot learning of language models via chain-of-thought fine-tuning” (Kim et al., 2023), “Octopack: instruction tuning code large language models” (Muennighoff et al., 2024), “Free Dolly: introducing the worlds first truly open instruction-tuned LLM” (Conover et al., 2023), and “Instruction tuning with GPT-4” (Peng et al., 2023).
Habana Gaudi is a specialized chip that can significantly reduce costs and time required for pre-training large-scale models like BERT by up to 25%. This technology has been used in various research projects, such as “Training a helpful and harmless assistant with reinforcement learning from human feedback” (Bai et al., 2023), “Red teaming language models to reduce harms: methods, scaling behaviors, and lessons learned” (Ganguli et al., 2022), and “Booksum: a collection of datasets for long-form narrative summarization” (Kryściński et al., 2022). Other notable projects include “CAMEL: communicative agents for mind exploration of large scale language model society” (Li et al., 2024), “The CoT collection: improving zero-shot and few-shot learning of language models via chain-of-thought fine-tuning” (Kim et al., 2023), “Octopack: instruction tuning code large language models” (Muennighoff et al., 2024), “Free Dolly: introducing the worlds first truly open instruction-tuned LLM” (Conover et al., 2023), and “Instruction tuning with GPT-4” (Peng et al., 2023). By optimizing hardware specifically for this task, Habana Gaudi improves the accuracy and performance of models overall.