Correct model architecture name to match allenai/OLMo-1B
#2
by
PierreLepagnol
- opened
In this file: https://huggingface.co/allenai/OLMo-1B/blob/main/config.json the architecture is : "OLMoForCausalLM"
In this repo the architecture is : "OlmoForCausalLM"
This makes vllm pipeline fail.
Hi
@PierreLepagnol
, because this model (OLMo-1B-hf) is integrated directly with the transformers library (unlike its original model, OLMo-1B), the change in casing (from "OLMo" to "Olmo") was necessary for that integration.
Could you please detail the nature of the failure in the vllm pipeline you mentioned?
shanearora
changed pull request status to
closed