Upload 3 files
#5
by
LutherXD
- opened
Key Modifications
processing_opencua.py
Implemented theOpenCUAProcessorclass with interfaces fully aligned with theQwenVLProcessor(Qwen-VL processor) to ensure consistent usage patterns across similar multimodal models.
Followed the same method signatures, input/output formats, and core logic as Qwen-VL processor for seamless integration.config.json
Updated theauto-processormapping to point toOpenCUAProcessor, allowing theAutoProcessorclass to correctly load the OpenCUA processor without manual specification.modeling_opencua.py- FSDP Sharding Support: Set
_base_model_prefixandno_split_moduleattributes to enable proper FSDP (Fully Sharded Data Parallel) sharding during training (previously missing these attributes caused FSDP sharding failures). - Property Additions: Added
model,lm_head, and_support_sdpaproperties to support logits calculation in downstream frameworks (e.g., evaluation pipelines, loss computation). - Forward Function Enhancement:
- Adapted the forward method to comply with the latest transformers library interface standards.
- Added handling logic for cases where attention_mask is None to prevent runtime errors.
-Integrated placeholder_mask logic to support multimodal input masking requirements.
- FSDP Sharding Support: Set
xywang626
changed pull request status to
merged
@xywang626 It seems like you confused with 7B, 32B, 72B config.json. 7B and 72B config.json changed without changing safetensors.
@xywang626 It seems like you confused with 7B, 32B, 72B config.json. 7B and 72B config.json changed without changing safetensors.
Thank you for your kind remind! I mistakenly uploaded the wrong config.json. I will update config.json immediately!
Thanks! Replaced files all merged.