- Added
custom_models.jsonsystem — users can now add their own models without modifying the core repository. #19 - Supports both Hugging Face (
hf_models) and GGUF (gguf_models) formats. - Fully optional — if the file is missing or empty, JoyCaption continues as normal.
- Safe and mergeable — your custom definitions are loaded dynamically at runtime.
- Includes example file:
custom_models_example.json - Added detailed documentation: 📘 custom_models.md
- Fixed Module Loading Issues: Resolved "ModuleNotFoundError: No module named 'JC'" by implementing sorted file loading in
__init__.py - Fixed CUDA Memory Issues: Improved error handling for CUDA memory allocation failures
- Inline Utility Functions: Moved error handling and utility functions back into main modules to avoid import conflicts
- Enhanced Error Handling: Better error messages and resource cleanup for model loading failures
- Improved Memory Management: More robust GPU memory cleanup and error recovery
- Added Progress Feedback: Console output now shows processing status for GGUF models
- Enhanced Processing Mode: Added "Auto" option to processing_mode (Auto/GPU/CPU) for better hardware detection
- Better Tooltips: Improved parameter descriptions and user guidance
- Fixed Installation Guide: Corrected path error in Chinese installation guide (
llama_cpp_install_zh.md) - Updated Model Path Instructions: Added clearer guidance for model placement in README
- Configuration Centralization: Added
gguf_settingstojc_data.jsonfor better configuration management - Stabilized Module Loading: Implemented deterministic file loading order to prevent import race conditions
https://github.com/1038lab/ComfyUI-JoyCaption/blob/main/example_workflows/JoyCaption-GGUF.json
- GGUF Model Support: Added comprehensive support for quantized GGUF models for better performance and lower memory usage
- New
JoyCaption GGUFnode for basic GGUF model usage - New
JoyCaption GGUF (Advanced)node with full parameter control - Support for multiple quantization levels (Q2_K, Q3_K_S, Q3_K_M, Q3_K_L, IQ4_XS, Q4_K_S, Q4_K_M, Q5_K_S, Q5_K_M, Q6_K, Q8_0, F16)
- Automatic model and vision projection model downloading
- llama_cpp_install folder: Added comprehensive installation guides and scripts for llama-cpp-python
llama_cpp_install.py- Automated installation script with CUDA support detectionllama_cpp_install.md- English installation guide for ComfyUI portablellama_cpp_install_zh.md- Chinese installation guide for ComfyUI portable- Supports Windows, macOS, and Linux platforms
- Automatic CUDA/GPU detection and compilation
- Pre-compiled wheel support for faster installation
- New
- Fixed TypeError: Resolved
'NoneType' object cannot be interpreted as an integererror with top_k parameter - Fixed Image Format: Updated image processing to use correct base64 data URI format for llama-cpp-python
- Fixed Missing Variables: Resolved
NameError: name 'prompt_text' is not definedin JC_GGUF class - Enhanced Error Handling: Improved llama-cpp-python dependency detection and fallback mechanisms
- Clean Console Output: Eliminated base64 image data spam in backend console
- Enhanced Error Handling: Improved error messages and parameter validation
- Better Performance: GGUF models provide 2-4x speed improvement with lower memory usage
- Simplified Installation: One-click llama-cpp-python installation with automatic CUDA support
- Memory Efficiency: GGUF models use 50-80% less memory than standard models
- Processing Speed: Faster inference with quantized models
- Output Suppression: Clean console output during model loading and generation
- Optimized Loading: Improved model loading times and memory management
- Parameter Handling: Proper top_k parameter validation (only included when > 0)
- Image Processing: Correct PIL to base64 conversion for vision models
- Code Quality: Cleaned up production code, removed unnecessary comments
- Stability: Enhanced error recovery and model management
- Installation Automation: Streamlined llama-cpp-python installation process
- Standard Models: Continue to support HuggingFace format models
- GGUF Models: Comprehensive support for efficient quantized models
- Q2_K: 3.18GB, ~4GB VRAM, Good quality, Low VRAM systems (6GB+)
- Q3_K_S: 3.66GB, ~5GB VRAM, Good+ quality, Budget systems (8GB+)
- Q3_K_M: 4.02GB, ~5GB VRAM, Better quality, Balanced performance
- Q3_K_L: 4.32GB, ~6GB VRAM, Better+ quality, Good quality/size ratio
- IQ4_XS: 4.48GB, ~6GB VRAM, Very Good quality, Recommended (8GB+)
- Q4_K_S: 4.69GB, ~6GB VRAM, Very Good quality, Quality focused
- Q4_K_M: 4.92GB, ~7GB VRAM, Very Good+ quality, Balanced choice
- Q5_K_S: 5.60GB, ~7GB VRAM, Excellent quality, High quality (10GB+)
- Q5_K_M: 5.73GB, ~8GB VRAM, Excellent+ quality, Premium quality
- Q6_K: 6.60GB, ~8GB VRAM, Near Original quality, Maximum quality (12GB+)
- Q8_0: 8.54GB, ~10GB VRAM, Original- quality, Full precision alternative
- F16: 16.1GB, ~18GB VRAM, Original quality, Full precision (24GB+)
- Automated Installation:
llama_cpp_install/llama_cpp_install.py- One-click installation script - English Guide:
llama_cpp_install/llama_cpp_install.md- Detailed English installation instructions - Chinese Guide:
llama_cpp_install/llama_cpp_install_zh.md- 中文安装指南 - Cross-Platform Support: Windows, macOS, and Linux installation guides
- CUDA Support: Automatic GPU detection and CUDA compilation
- Pre-compiled Wheels: Faster installation with pre-built binaries
- Enhanced CUDA performance for high-end GPUs
- Optimized model loading and caching system
- Improved memory management strategies
- Added
Global Cachemode for faster processing with sufficient VRAM - Enhanced
Keep in Memorymode for balanced performance - Optimized
Clear After Runmode for limited VRAM scenarios
- Added
- Implemented efficient memory cleanup after processing
- Added automatic memory usage monitoring
- Optimized CUDA memory allocation with max_split_size_mb configuration
- Enhanced model caching system with validation checks
- Improved memory cleanup during model switching
- Added automatic GPU memory management for different VRAM capacities
- Enhanced error handling and recovery mechanisms
- Improved model loading validation
- Optimized image processing pipeline
- Added better tooltips and documentation for node parameters
- Fixed CaptionTool nodes not registering in ComfyUI interface
- Added multi-language support for node interfaces
- English (en) - Default language
- French (fr) - Support for French interface
- Japanese (ja) - 日本語インターフェース対応
- Korean (ko) - 한국어 인터페이스 지원
- Russian (ru) - Поддержка русского интерфейса
- Chinese (zh) - 中文界面支持
- Initial release of ComfyUI-JoyCaption
- Added JoyCaption node for image captioning
- Integrated memory management system
- Added caption tools for text processing
- Implemented efficient memory handling for large image processing
- Added automatic memory cleanup after processing
- Optimized memory usage during batch operations
- Added memory usage monitoring
- Added Image Batch Path node (🖼️) for batch image loading
- Support for sequential, reverse, and random image loading
- Configurable batch size and start position
- Automatic EXIF orientation correction
- Support for jpg, jpeg, png, and webp formats
- Added Caption Saver node (📝) for caption management
-
Flexible output path configuration
-
Custom filename support
-
Optional image copying with captions
-
Automatic file overwrite protection
-
UTF-8 encoding support
-
Batch processing capability
-

