Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

graph_prepare.cc:742:ERROR:error during serialize: memory usage too large #16

Open
zqxuturbo opened this issue Nov 13, 2024 · 0 comments
Labels
question Further information is requested

Comments

@zqxuturbo
Copy link

Hi,
When I convert llama2-7b with target device 8gen2、8295、8775 (htp v73, soc 43),memory error occured:

2024-11-13 14:49:27,866 - INFO - qnn-model-lib-generator: Target: x86_64-linux-clang Library: /tmp/db824256-3d3a-4b91-a94a-4a841d7d04e5fygbm454/tmp_7_tx2cu/x86_64-linux-clang/libqnn_model.so

[2024-11-13 14:49:58,453] [INFO] Saving model
[2024-11-13 14:50:10,731] [INFO] Graph name: prompt_part1
[2024-11-13 14:50:10,733] [INFO] -=- QNN Model Libraries to QNN Context Binary (qnn-context-binary-generator) -=-
[2024-11-13 14:50:10,734] [INFO] Contents of HTP Settings: {'graphs': [{'graph_names': ['prompt_part1'], 'fp16_relaxed_precision': 1, 'vtcm_mb': 0, 'O': 3}], 'devices': [{'dsp_arch': 'v73', 'soc_model': 43}]}
[2024-11-13 14:50:10,734] [INFO] Contents of HTP Config file used: {'backend_extensions': {'shared_library_path': 'libQnnHtpNetRunExtensions.so', 'config_file_path': '/tmp/db824256-3d3a-4b91-a94a-4a841d7d04e5fygbm454/tmpyvclsp5b/htp_setting.json'}, 'context_configs': {'enable_graphs': ['prompt_part1']}, 'graph_configs': [{'graph_name': 'prompt_part1'}], 'memory': {'mem_type': 'shared_buffer'}}
[2024-11-13 14:50:10,734] [INFO] Running /qnn_sdk/bin/x86_64-linux-clang/qnn-context-binary-generator --backend /qnn_sdk/lib/x86_64-linux-clang/libQnnHtp.so --model /tmp/db824256-3d3a-4b91-a94a-4a841d7d04e5fygbm454/tmpmuz4pbse.so --output_dir /tmp/db824256-3d3a-4b91-a94a-4a841d7d04e5fygbm454/tmpyvclsp5b --binary_file qnn_model --config_file /tmp/db824256-3d3a-4b91-a94a-4a841d7d04e5fygbm454/tmpyvclsp5b/htp_context.json
[2024-11-13 14:56:26,382] [INFO] qnn-context-binary-generator pid:15974
0.0ms [ ERROR ] fa_alloc.cc:3866:ERROR:graph requires estimated allocation of 2473687 KB, limit is 2097152 KB

 0.0ms [ ERROR ] graph_prepare.cc:742:ERROR:error during serialize: memory usage too large

 0.0ms [ ERROR ] graph_prepare.cc:6095:ERROR:Serialize error: memory usage too large

 0.0ms [ ERROR ] QnnDsp <E> Graph prompt_part1 serialization failed

 0.0ms [ ERROR ] QnnDsp <E> Failed to serialize graph prompt_part1

 0.0ms [ ERROR ] QnnDsp <E> Context binary serialization failed

 0.0ms [ ERROR ] QnnDsp <E> Get context blob failed.

 0.0ms [ ERROR ] QnnDsp <E> Failed to get serialized binary

 0.0ms [ ERROR ] QnnDsp <E> Failed to get context binary with err 0x138f

374134.5ms [ ERROR ] Could not get binary.
Graph Finalize failure

[2024-11-13 14:56:26,560] [ERROR] Conversion to context binary failed with exit code 15

@mestrona-3 mestrona-3 added the question Further information is requested label Nov 14, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants