Releases: sdpython/onnx-diagnostic
Releases · sdpython/onnx-diagnostic
0.8.1
- #290: adds one prompt for text2text-generation
- #289: adds command line options
--exppoto give the exporter additional options - #287: adds input
'inputs_prompt'to test a LLM, meant to be used during validation - #288: add .contiguous in torch.cond branch (attention patch for sdpa implementation)
- #286: adds variable to track random nodes in models
0.8.0
- #283: fix historical aggregation when multiple input sets are used
- #282: add tools to understand better which functions were patched
- #280: fixes patches for sdpa_attention_forward for different version of transformers
- #278: implements
onnx_generate_with_genai - #277: changes the serialization for all caches to reorder the model outputs (key_1, value_1, key_2, ...)
- #276: implements
onnx_generatewhich implements method generate for an onnx model, - #275: fixes function
patched_vmap
0.7.16
- #273: enables export with FakeTensor
- #272: makes patches work with FakeTensor
- #270: add export sample code to export a specific model id with the appropriate inputs
- #269: adds one unit test to track a patch fixing broadcast output shape
- #267: patches
sdpa_attention_forwardbecause of a control flow (transformers>=5.0) - #266: makes
patch_torchan integer intorch_export_patchesto enable more patches
0.7.15
0.7.14
0.7.13
0.7.12
0.7.11
0.7.11
- #224: support model_id with // to specify a subfolder
- #223: adds task image-to-video
- #220: adds option --ort-logs to display onnxruntime logs when creating the session
- #220: adds a patch for PR #40791 huggingface/transformers#40791_ in transformers
0.7.10
- #218: patches used sdpa_mask_recent_torch used from _vmap_for_bhqkv
0.7.9
0.7.8
0.7.7
- #205: add in_channels in image_text_to_text
- #204: switch default num_hidden_layers to 4
- #203: Add option to disable patches for torch in command line validate
- #202: add models DeepseekV3ForCausalLM, Gemma3ForCausalLM, Glm4vMoeForConditionalGeneration
- #201: switch CI to 4.55.4
- #200: fixes patches for 4.55.1+, DynamicCache is no longer registered by default, this code moved to executorch.py in transformers
- #199: delete hidden_size and num_attention_heads modification in a config
- #198: support gpt-oss
- #197: updates CI for torch 2.8
- #196: implements a patch to rewrite a loop in modeling_qwen2_vl.VisionAttention