Error when using the Android ChatApp with llama3.2-3B
model exported with 4096 context length
#52
Labels
question
Further information is requested
This model works:
python3 -m qai_hub_models.models.llama_v3_2_3b_chat_quantized.export --output-dir genie_bundle --device "Snapdragon 8 Elite QRD" --skip-inferencing --skip-profiling --context-length 2048
This does not:
python3 -m qai_hub_models.models.llama_v3_2_3b_chat_quantized.export --output-dir genie_bundle --device "Snapdragon 8 Elite QRD" --skip-inferencing --skip-profiling --context-length 4096
With the following error:
Device specs:
The text was updated successfully, but these errors were encountered: