From b25681e7a56ae4b6d457627b796e949359fcc44f Mon Sep 17 00:00:00 2001 From: h-guo18 <67671475+h-guo18@users.noreply.github.com> Date: Tue, 28 Apr 2026 20:37:30 +0000 Subject: [PATCH 1/2] fix example Signed-off-by: h-guo18 <67671475+h-guo18@users.noreply.github.com> --- tools/launcher/examples/Qwen/Qwen3-8B/hf_online_eagle3.yaml | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/tools/launcher/examples/Qwen/Qwen3-8B/hf_online_eagle3.yaml b/tools/launcher/examples/Qwen/Qwen3-8B/hf_online_eagle3.yaml index 5b55e9a2221..6c4743cb588 100644 --- a/tools/launcher/examples/Qwen/Qwen3-8B/hf_online_eagle3.yaml +++ b/tools/launcher/examples/Qwen/Qwen3-8B/hf_online_eagle3.yaml @@ -24,7 +24,7 @@ pipeline: task_0: script: common/eagle3/make_dataset.sh args: - - -f modules/Model-Optimizer/examples/speculative_decoding/prepare_input_conversations/example_data_config.yaml + - -f modules/Model-Optimizer/examples/dataset/example_data_config.yaml - --full-conversations slurm_config: _factory_: "slurm_factory" @@ -45,7 +45,6 @@ pipeline: - training.ar_validate_steps=500000 - training.num_train_epochs=1 environment: - - HOME: /tmp - TORCHINDUCTOR_CACHE_DIR: /tmp/torch_cache slurm_config: _factory_: "slurm_factory" @@ -68,7 +67,6 @@ pipeline: - --concurrency 1 environment: - HF_MODEL_CKPT: <> - - HOME: /tmp - TORCHINDUCTOR_CACHE_DIR: /tmp/torch_cache slurm_config: _factory_: "slurm_factory" From 776f71bdbc6eb8f34469f890d95fa7d7e1313b9e Mon Sep 17 00:00:00 2001 From: h-guo18 <67671475+h-guo18@users.noreply.github.com> Date: Wed, 29 Apr 2026 00:58:29 +0000 Subject: [PATCH 2/2] fix example Signed-off-by: h-guo18 <67671475+h-guo18@users.noreply.github.com> --- tools/launcher/examples/Qwen/Qwen3-8B/hf_online_eagle3.yaml | 3 --- 1 file changed, 3 deletions(-) diff --git a/tools/launcher/examples/Qwen/Qwen3-8B/hf_online_eagle3.yaml b/tools/launcher/examples/Qwen/Qwen3-8B/hf_online_eagle3.yaml index 6c4743cb588..969a865f35f 100644 --- a/tools/launcher/examples/Qwen/Qwen3-8B/hf_online_eagle3.yaml +++ b/tools/launcher/examples/Qwen/Qwen3-8B/hf_online_eagle3.yaml @@ -44,8 +44,6 @@ pipeline: - training.disable_tqdm=true - training.ar_validate_steps=500000 - training.num_train_epochs=1 - environment: - - TORCHINDUCTOR_CACHE_DIR: /tmp/torch_cache slurm_config: _factory_: "slurm_factory" nodes: 1 @@ -67,7 +65,6 @@ pipeline: - --concurrency 1 environment: - HF_MODEL_CKPT: <> - - TORCHINDUCTOR_CACHE_DIR: /tmp/torch_cache slurm_config: _factory_: "slurm_factory" nodes: 1