Skip to content

dynamic_vram: fix pinning with model defined dtype#12208

Merged
comfyanonymous merged 1 commit intoComfy-Org:masterfrom
rattus128:prs/dynamic-vram-fixes/fix-model-dtype-pins
Feb 1, 2026
Merged

dynamic_vram: fix pinning with model defined dtype#12208
comfyanonymous merged 1 commit intoComfy-Org:masterfrom
rattus128:prs/dynamic-vram-fixes/fix-model-dtype-pins

Conversation

@rattus128
Copy link
Contributor

@rattus128 rattus128 commented Feb 1, 2026

pinned memory was converted back to pinning the CPU side weight without any changes. Fix the pinner to use the CPU weight and not the model defined geometry. This will either save RAM or stop buffer overruns when the types mismatch.

Fix the model defined weight caster to use the [ s.weight, s.bias ] interpretation, as xfer_dest might be the flattened pin now. Fix the detection of needing to cast to not be conditional on !pin.

#12205

Example Test conditions:

RTX3060, Linux, --fast dynamic_vram

LTX2 VAE decode with on-demand model:

image

Before:

First crash:

  File "/home/rattus/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/rattus/ComfyUI/comfy/ops.py", line 414, in forward
    return self.forward_comfy_cast_weights(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/rattus/ComfyUI/comfy/ops.py", line 406, in forward_comfy_cast_weights
    weight, bias, offload_stream = cast_bias_weight(self, input, offloadable=True)
                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/rattus/ComfyUI/comfy/ops.py", line 207, in cast_bias_weight
    return cast_bias_weight_with_vbar(s, dtype, device, bias_dtype, non_blocking, compute_dtype)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/rattus/ComfyUI/comfy/ops.py", line 131, in cast_bias_weight_with_vbar
    comfy.model_management.cast_to_gathered(xfer_source, xfer_dest, non_blocking=non_blocking, stream=offload_stream)
  File "/home/rattus/ComfyUI/comfy/model_management.py", line 1189, in cast_to_gathered
    dest_views = comfy.memory_management.interpret_gathered_like(tensors, r)
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/rattus/ComfyUI/comfy/memory_management.py", line 70, in interpret_gathered_like
    raise ValueError(f"Buffer too small: needs {offset + size} bytes, but only has {gathered.numel()}. ")
ValueError: Buffer too small: needs 1770496 bytes, but only has 885760. 

Second crash (matches issue OP):

  File "/home/rattus/ComfyUI/comfy/ldm/lightricks/vae/causal_conv3d.py", line 72, in forward
    return self.conv(x) if x.shape[2] >= self.time_kernel_size else x[:, :, :0, :, :]
           ^^^^^^^^^^^^
  File "/home/rattus/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/rattus/venv/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/rattus/ComfyUI/comfy/ops.py", line 414, in forward
    return self.forward_comfy_cast_weights(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/rattus/ComfyUI/comfy/ops.py", line 406, in forward_comfy_cast_weights
    weight, bias, offload_stream = cast_bias_weight(self, input, offloadable=True)
                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/rattus/ComfyUI/comfy/ops.py", line 207, in cast_bias_weight
    return cast_bias_weight_with_vbar(s, dtype, device, bias_dtype, non_blocking, compute_dtype)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/rattus/ComfyUI/comfy/ops.py", line 138, in cast_bias_weight_with_vbar
    post_cast.copy_(pre_cast)
RuntimeError: The size of tensor a (3) must match the size of tensor b (7078912) at non-singleton dimension 4

Prompt executed in 1.60 seconds

After:

Model VideoVAE prepared for dynamic VRAM loading. 4663MB Staged. 0 patches attached.
Prompt executed in 17.79 seconds

pinned memory was converted back to pinning the CPU side weight without
any changes. Fix the pinner to use the CPU weight and not the model defined
geometry. This will either save RAM or stop buffer overruns when the types
mismatch.

Fix the model defined weight caster to use the [ s.weight, s.bias ]
interpretation, as xfer_dest might be the flattened pin now. Fix the detection
of needing to cast to not be conditional on !pin.
@comfyanonymous comfyanonymous merged commit 361b9a8 into Comfy-Org:master Feb 1, 2026
12 checks passed
@jkrauss82
Copy link

This still did not fix the issue with the division by 4 error for me.

!!! Exception during processing !!! self.storage_offset() must be divisible by 4 to view Byte as Float (different element sizes), but got 25702
Traceback (most recent call last):
  File "/home/user/workspace/ComfyUI/execution.py", line 527, in execute
    output_data, output_ui, has_subgraph, has_pending_tasks = await get_output_data(prompt_id, unique_id, obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, v3_data=v3_data)
                                                              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/user/workspace/ComfyUI/execution.py", line 331, in get_output_data
    return_values = await _async_map_node_over_list(prompt_id, unique_id, obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, v3_data=v3_data)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/user/workspace/ComfyUI/execution.py", line 305, in _async_map_node_over_list
    await process_inputs(input_dict, i)
  File "/home/user/workspace/ComfyUI/execution.py", line 293, in process_inputs
    result = f(**inputs)
             ^^^^^^^^^^^
  File "/home/user/workspace/ComfyUI/nodes.py", line 825, in load_vae
    sd, metadata = comfy.utils.load_torch_file(vae_path, return_metadata=True)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/user/workspace/ComfyUI/comfy/utils.py", line 128, in load_torch_file
    raise e
  File "/home/user/workspace/ComfyUI/comfy/utils.py", line 108, in load_torch_file
    sd, metadata = load_safetensors(ckpt)
                   ^^^^^^^^^^^^^^^^^^^^^^
  File "/home/user/workspace/ComfyUI/comfy/utils.py", line 96, in load_safetensors
    sd[name] = data_area[start:end].view(_TYPES[info["dtype"]]).view(info["shape"])
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: self.storage_offset() must be divisible by 4 to view Byte as Float (different element sizes), but got 25702

Reverting Comfy to 873de5f makes it work again

@rattus128
Copy link
Contributor Author

This still did not fix the issue with the division by 4 error for me.

!!! Exception during processing !!! self.storage_offset() must be divisible by 4 to view Byte as Float (different element sizes), but got 25702
Traceback (most recent call last):
  File "/home/user/workspace/ComfyUI/execution.py", line 527, in execute
    output_data, output_ui, has_subgraph, has_pending_tasks = await get_output_data(prompt_id, unique_id, obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, v3_data=v3_data)
                                                              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/user/workspace/ComfyUI/execution.py", line 331, in get_output_data
    return_values = await _async_map_node_over_list(prompt_id, unique_id, obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, v3_data=v3_data)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/user/workspace/ComfyUI/execution.py", line 305, in _async_map_node_over_list
    await process_inputs(input_dict, i)
  File "/home/user/workspace/ComfyUI/execution.py", line 293, in process_inputs
    result = f(**inputs)
             ^^^^^^^^^^^
  File "/home/user/workspace/ComfyUI/nodes.py", line 825, in load_vae
    sd, metadata = comfy.utils.load_torch_file(vae_path, return_metadata=True)
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/user/workspace/ComfyUI/comfy/utils.py", line 128, in load_torch_file
    raise e
  File "/home/user/workspace/ComfyUI/comfy/utils.py", line 108, in load_torch_file
    sd, metadata = load_safetensors(ckpt)
                   ^^^^^^^^^^^^^^^^^^^^^^
  File "/home/user/workspace/ComfyUI/comfy/utils.py", line 96, in load_safetensors
    sd[name] = data_area[start:end].view(_TYPES[info["dtype"]]).view(info["shape"])
               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: self.storage_offset() must be divisible by 4 to view Byte as Float (different element sizes), but got 25702

Reverting Comfy to 873de5f makes it work again

It's a different issue to the bug report, but thanks for the log! What is the safetensors file it is loading? Your trace suggests a misaligned sft weight which we can fix, but helps if I have a handle on the safetensors file.

@jkrauss82
Copy link

it is an SDXL checkpoint, here are the safetensors meta:

{
    "format": "pt",
    "sd_merge_recipe": {
        "type": "webui",
        "primary_model_hash": "d9d414bd112afb0db7363bb7b976da34d840d80e7d1ba3b3943c0dfbeb70499c",
        "secondary_model_hash": null,
        "tertiary_model_hash": null,
        "interp_method": "Weighted sum",
        "multiplier": 0.3,
        "save_as_half": false,
        "custom_name": "ilustmix_v60",
        "config_source": 0,
        "bake_in_vae": "sdxlVAE_sdxlVAE_264202.safetensors",
        "discard_weights": "",
        "is_inpainting": false,
        "is_instruct_pix2pix": false
    },
    "sd_merge_models": {
        "d9d414bd112afb0db7363bb7b976da34d840d80e7d1ba3b3943c0dfbeb70499c": {
            "name": "ilustmix_v40.safetensors",
            "legacy_hash": "a56eb33a",
            "sd_merge_recipe": {
                "type": "webui",
                "primary_model_hash": "cb3a162f3af00b00dc500e815953d5983dc33a4be3c480a855a27ff4db31caac",
                "secondary_model_hash": "a3a562d381e7c54f326e5f68a280a660b0f76571033ec625c90b64a9ea29066f",
                "tertiary_model_hash": null,
                "interp_method": "Weighted sum",
                "multiplier": 0.4,
                "save_as_half": false,
                "custom_name": "ilustmix_v40",
                "config_source": 0,
                "bake_in_vae": "sdxlVAE_sdxlVAE_264202.safetensors",
                "discard_weights": "",
                "is_inpainting": false,
                "is_instruct_pix2pix": false
            }
        },
        "cb3a162f3af00b00dc500e815953d5983dc33a4be3c480a855a27ff4db31caac": {
            "name": "ilustmix_v30_1357311.safetensors",
            "legacy_hash": "88426183",
            "sd_merge_recipe": {
                "type": "webui",
                "primary_model_hash": "1620b6d4794f996d38dae6b0b522d05a2e92d5c938791d86973bbf404c9b0b7d",
                "secondary_model_hash": null,
                "tertiary_model_hash": null,
                "interp_method": "Weighted sum",
                "multiplier": 0.2,
                "save_as_half": false,
                "custom_name": "ilustmix",
                "config_source": 0,
                "bake_in_vae": "None",
                "discard_weights": "",
                "is_inpainting": false,
                "is_instruct_pix2pix": false
            }
        },
        "1620b6d4794f996d38dae6b0b522d05a2e92d5c938791d86973bbf404c9b0b7d": {
            "name": "ilustmix_v20_1221746.safetensors",
            "legacy_hash": "97fb5c13",
            "sd_merge_recipe": {
                "type": "webui",
                "primary_model_hash": "cbd74f1aeb3a6f9c6d0bca71c658a9fb01f6e192a9027f5d099b986ea0950f64",
                "secondary_model_hash": "7a48fb8eae2bbb1b4ecdfe479fdede83bf252f4841cc1f2dde278c6c39636b91",
                "tertiary_model_hash": null,
                "interp_method": "Weighted sum",
                "multiplier": 0.5,
                "save_as_half": false,
                "custom_name": "ilustmix_v20",
                "config_source": 0,
                "bake_in_vae": "None",
                "discard_weights": "",
                "is_inpainting": false,
                "is_instruct_pix2pix": false
            }
        },
        "cbd74f1aeb3a6f9c6d0bca71c658a9fb01f6e192a9027f5d099b986ea0950f64": {
            "name": "illustrij_v50_1153542.safetensors",
            "legacy_hash": "4cb125f4",
            "sd_merge_recipe": null
        },
        "7a48fb8eae2bbb1b4ecdfe479fdede83bf252f4841cc1f2dde278c6c39636b91": {
            "name": "cyberillustrious_v10_1169622.safetensors",
            "legacy_hash": "683959ce",
            "sd_merge_recipe": null
        },
        "3a2df45ef8a243818be8aaa312dee2e187bfe6851a9cc6b57022b3a1440279bb": {
            "name": "prefectIllustriousXL_v10_1282816.safetensors",
            "legacy_hash": "832a96e1",
            "sd_merge_recipe": null
        },
        "a3a562d381e7c54f326e5f68a280a660b0f76571033ec625c90b64a9ea29066f": {
            "name": "illustrij_v10_1441661.safetensors",
            "legacy_hash": "c0656ffa",
            "sd_merge_recipe": null
        },
        "e10c2b912e3adeff666e8e56d44b1e9d294cb04bff335b552a63d00f58a0346b": {
            "name": "jedpointreal_v1ILVae_1541657.safetensors",
            "legacy_hash": "288f91ff",
            "sd_merge_recipe": null
        }
    }
}

just in case my start up options, but I do not think they cause the issue:

python main.py --listen 0.0.0.0 --port 7860 --fp8_e4m3fn-text-enc --fp8_e4m3fn-unet --bf16-vae --preview-method taesd --fast --use-sage-attention --reserve-vram 1.25 --disable-cuda-malloc

python venv

$ pip freeze
absl-py==2.3.0
accelerate==1.10.1
addict==2.4.0
aiohappyeyeballs==2.6.1
aiohttp==3.12.13
aiosignal==1.3.2
albucore==0.0.24
albumentations==2.0.8
alembic==1.16.2
annotated-types==0.7.0
antlr4-python3-runtime==4.9.3
anyio==4.9.0
attrs==25.3.0
av==14.4.0
cachetools==6.1.0
certifi==2025.6.15
cffi==1.17.1
charset-normalizer==3.4.2
cmake==4.0.3
coloredlogs==15.0.1
colorlog==6.9.0
comfy-aimdo==0.1.6
comfy-kitchen==0.2.7
comfyui-embedded-docs==0.4.0
comfyui-workflow-templates-core==0.3.115
comfyui-workflow-templates-media-api==0.3.47
comfyui-workflow-templates-media-image==0.3.75
comfyui-workflow-templates-media-other==0.3.98
comfyui-workflow-templates-media-video==0.3.42
comfyui_frontend_package==1.32.9
comfyui_workflow_templates==0.8.27
compressed-tensors==0.12.2
contourpy==1.3.2
cssselect2==0.8.0
cstr @ git+https://github.com/WASasquatch/cstr@0520c29a18a7a869a6e5983861d6f7a4c86f8e9b
cycler==0.12.1
Cython==3.1.2
decord==0.6.0
diffusers==0.35.1
dill==0.4.0
easydict==1.13
einops==0.8.1
embreex==2.17.7.post6
facexlib==0.3.0
fairscale==0.4.13
ffmpy @ git+https://github.com/WASasquatch/ffmpy.git@f000737698b387ffaeab7cd871b0e9185811230d
filelock==3.18.0
filterpy==1.4.5
flatbuffers==25.2.10
fonttools==4.58.4
frozenlist==1.7.0
fsspec==2025.5.1
ftfy==6.1.1
fvcore==0.1.5.post20221221
gguf==0.17.1
gitdb==4.0.12
GitPython==3.1.44
greenlet==3.2.3
h11==0.16.0
hf-xet==1.1.5
httpcore==1.0.9
httpx==0.28.1
huggingface-hub==0.35.3
humanfriendly==10.0
idna==3.10
imageio==2.37.0
imageio-ffmpeg==0.6.0
img2texture @ git+https://github.com/WASasquatch/img2texture.git@d6159abea44a0b2cf77454d3d46962c8b21eb9d3
importlib_metadata==8.7.0
insightface==0.7.3
iopath==0.1.10
jax==0.6.2
jaxlib==0.6.2
Jinja2==3.1.6
joblib==1.5.1
jsonschema==4.24.0
jsonschema-specifications==2025.4.1
kiwisolver==1.4.8
kornia==0.8.1
kornia_rs==0.1.9
lark==1.2.2
lazy_loader==0.4
llvmlite==0.44.0
loguru==0.7.3
lxml==5.4.0
Mako==1.3.10
manifold3d==3.1.1
mapbox_earcut==1.0.3
MarkupSafe==3.0.2
matplotlib==3.10.3
mediapipe==0.10.21
ml_dtypes==0.5.1
mpmath==1.3.0
multidict==6.5.0
networkx==3.5
ninja==1.11.1.4
numba==0.61.2
numpy==1.26.4
nunchaku @ https://github.com/nunchaku-tech/nunchaku/releases/download/v1.0.1dev20250930/nunchaku-1.0.1.dev20250930+torch2.8-cp312-cp312-linux_x86_64.whl#sha256=2040c22106079401b507653e76b3d05a6c8af9931cf6b727574acae700cc4014
nvidia-cublas-cu12==12.8.4.1
nvidia-cuda-cupti-cu12==12.8.90
nvidia-cuda-nvrtc-cu12==12.8.93
nvidia-cuda-runtime-cu12==12.8.90
nvidia-cudnn-cu12==9.10.2.21
nvidia-cufft-cu12==11.3.3.83
nvidia-cufile-cu12==1.13.1.3
nvidia-curand-cu12==10.3.9.90
nvidia-cusolver-cu12==11.7.3.90
nvidia-cusparse-cu12==12.5.8.93
nvidia-cusparselt-cu12==0.7.1
nvidia-nccl-cu12==2.27.3
nvidia-nvjitlink-cu12==12.8.93
nvidia-nvtx-cu12==12.8.90
omegaconf==2.3.0
onnx==1.18.0
onnxruntime==1.23.1
onnxruntime-gpu==1.23.0
opencv-contrib-python==4.11.0.86
opencv-contrib-python-headless==4.11.0.86
opencv-python==4.11.0.86
opencv-python-headless==4.11.0.86
opt_einsum==3.4.0
packaging==25.0
pandas==2.3.0
peft==0.17.0
piexif==1.1.3
pilgram==1.2.1
pillow==11.2.1
platformdirs==4.3.8
pooch==1.8.2
portalocker==3.2.0
prettytable==3.16.0
propcache==0.3.2
protobuf==4.25.8
psutil==7.0.0
py-cpuinfo==9.0.0
pycocotools==2.0.10
pycollada==0.9
pycparser==2.22
pydantic==2.11.7
pydantic-settings==2.9.1
pydantic_core==2.33.2
PyMatting==1.1.14
pyparsing==3.2.3
python-dateutil==2.9.0.post0
python-dotenv==1.1.0
pytorch-extension==0.2
pytz==2025.2
PyYAML==6.0.2
referencing==0.36.2
regex==2024.11.6
rembg==2.0.66
reportlab==4.4.2
requests==2.32.4
rotary-embedding-torch==0.8.9
rpds-py==0.25.1
rtree==1.4.0
safetensors==0.5.3
sageattention==2.1.1
sam3 @ file:///home/krauss/workspace/ComfyUI/custom_nodes/comfyui_sam3/src/sam3
scikit-image==0.25.2
scikit-learn==1.7.0
scipy==1.15.3
segment-anything==1.0
sentencepiece==0.2.0
setuptools==80.9.0
shapely==2.1.1
simsimd==6.4.9
six==1.17.0
smmap==5.0.2
sniffio==1.3.1
sounddevice==0.5.2
soundfile==0.13.1
spandrel==0.4.1
SQLAlchemy==2.0.41
stringzilla==3.12.5
svg.path==6.3
svglib==1.5.1
sympy==1.14.0
tabulate==0.9.0
termcolor==3.1.0
threadpoolctl==3.6.0
tifffile==2025.6.11
timm==1.0.22
tinycss2==1.4.0
tokenizers==0.22.1
tomli==2.2.1
torch==2.8.0
torchaudio==2.8.0
torchsde==0.2.6
torchvision==0.23.0
tqdm==4.67.1
trampoline==0.1.2
transformers==4.57.0
trimesh==4.6.12
triton==3.4.0
typing-inspection==0.4.1
typing_extensions==4.14.0
tzdata==2025.2
ultralytics==8.3.157
ultralytics-thop==2.0.14
urllib3==2.5.0
vhacdx==0.0.8.post2
wcwidth==0.2.13
webcolors==24.11.1
webencodings==0.5.1
xxhash==3.5.0
yacs==0.1.8
yapf==0.43.0
yarl==1.20.1
zipp==3.23.0

simonri pushed a commit to simonri/ComfyUI-flash-attention-3 that referenced this pull request Feb 2, 2026
pinned memory was converted back to pinning the CPU side weight without
any changes. Fix the pinner to use the CPU weight and not the model defined
geometry. This will either save RAM or stop buffer overruns when the types
mismatch.

Fix the model defined weight caster to use the [ s.weight, s.bias ]
interpretation, as xfer_dest might be the flattened pin now. Fix the detection
of needing to cast to not be conditional on !pin.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants