From 68670aeaee394453224d113936d6b8e4ccef7473 Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Sat, 13 Dec 2025 22:22:25 +0200 Subject: [PATCH 1/2] convert : fix gpt-oss --- convert_hf_to_gguf.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/convert_hf_to_gguf.py b/convert_hf_to_gguf.py index 151608d56b8..41a97428238 100755 --- a/convert_hf_to_gguf.py +++ b/convert_hf_to_gguf.py @@ -9612,11 +9612,12 @@ def set_gguf_parameters(self): self.gguf_writer.add_sliding_window(self.hparams["sliding_window"]) self.gguf_writer.add_expert_feed_forward_length(self.hparams["intermediate_size"]) - rope_scaling = self.hparams.get("rope_scaling") or {} + rope_scaling = self.hparams.get("rope_scaling") or (self.hparams.get("rope_parameters" or {})) rope_type = rope_scaling.get("rope_type", rope_scaling.get("type")) assert rope_type == "yarn", f"GPT-OSS only supports yarn rope scaling, got {rope_type}" self.gguf_writer.add_rope_scaling_type(gguf.RopeScalingType.YARN) self.gguf_writer.add_rope_scaling_factor(rope_scaling["factor"]) + self.gguf_writer.add_rope_freq_base(rope_scaling.get("rope_theta")) self.gguf_writer.add_rope_scaling_orig_ctx_len(rope_scaling.get("original_max_position_embeddings", 4096)) From 4b4264f25aa07c73080ebb44450f58e16b78a10c Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Sat, 13 Dec 2025 22:55:00 +0200 Subject: [PATCH 2/2] Update convert_hf_to_gguf.py MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Co-authored-by: Sigbjørn Skjæret --- convert_hf_to_gguf.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/convert_hf_to_gguf.py b/convert_hf_to_gguf.py index 41a97428238..3c0b94fdb32 100755 --- a/convert_hf_to_gguf.py +++ b/convert_hf_to_gguf.py @@ -9612,7 +9612,7 @@ def set_gguf_parameters(self): self.gguf_writer.add_sliding_window(self.hparams["sliding_window"]) self.gguf_writer.add_expert_feed_forward_length(self.hparams["intermediate_size"]) - rope_scaling = self.hparams.get("rope_scaling") or (self.hparams.get("rope_parameters" or {})) + rope_scaling = self.hparams.get("rope_scaling") or self.hparams.get("rope_parameters") or {} rope_type = rope_scaling.get("rope_type", rope_scaling.get("type")) assert rope_type == "yarn", f"GPT-OSS only supports yarn rope scaling, got {rope_type}" self.gguf_writer.add_rope_scaling_type(gguf.RopeScalingType.YARN)