CT-GLIP: 3D Grounded Language-Image Pretraining with CT Scans and Radiology Reports for Full-Body Scenarios

arXiv — cs.CVWednesday, December 3, 2025 at 5:00:00 AM
  • CT-GLIP, a new 3D Grounded Language-Image Pretrained model, has been introduced to enhance the alignment of CT scans with radiology reports, addressing limitations in existing methods that rely on global embeddings. This model constructs fine-grained CT-report pairs to improve cross-modal contrastive learning, enabling better identification of organs and abnormalities in a zero-shot manner.
  • The development of CT-GLIP is significant as it enhances the accuracy of medical imaging analysis, potentially improving diagnostic capabilities and patient outcomes by allowing for more precise organ recognition and abnormality detection without the need for extensive retraining.
  • This advancement reflects a broader trend in the integration of AI in healthcare, where models like CT-GLIP and others are increasingly being developed to automate and improve the efficiency of medical processes, such as report generation and segmentation of findings, thereby addressing the challenges posed by manual analysis and the need for high-quality data in medical imaging.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
AlignBench: Benchmarking Fine-Grained Image-Text Alignment with Synthetic Image-Caption Pairs
NeutralArtificial Intelligence
AlignBench has been introduced as a benchmark for evaluating fine-grained image-text alignment using synthetic image-caption pairs, addressing limitations in existing models like CLIP that rely on rule-based perturbations or short captions. This benchmark allows for a more detailed assessment of visual-language models (VLMs) by annotating each sentence for correctness.
Mixture of Ranks with Degradation-Aware Routing for One-Step Real-World Image Super-Resolution
PositiveArtificial Intelligence
A new Mixture-of-Ranks (MoR) architecture has been proposed for one-step real-world image super-resolution (Real-ISR), integrating sparse Mixture-of-Experts (MoE) to enhance the adaptability of models in reconstructing high-resolution images from degraded samples. This approach utilizes a fine-grained expert partitioning strategy, treating each rank in Low-Rank Adaptation (LoRA) as an independent expert, thereby improving the model's ability to capture heterogeneous characteristics of real-world images.
Random forest-based out-of-distribution detection for robust lung cancer segmentation
PositiveArtificial Intelligence
A new study has introduced a random forest-based method for out-of-distribution detection in lung cancer segmentation, utilizing a Swin Transformer model pretrained on over 10,000 3D CT scans. This approach aims to enhance the accuracy of identifying cancerous lesions in CT images, particularly in scenarios where data may not conform to expected distributions.
VIVAT: Virtuous Improving VAE Training through Artifact Mitigation
PositiveArtificial Intelligence
A new paper introduces VIVAT, a systematic approach designed to mitigate common artifacts in the training of Variational Autoencoders (VAEs), which are crucial for generative computer vision. The study identifies five prevalent artifacts and proposes modifications to improve VAE performance, achieving state-of-the-art results in image reconstruction metrics and enhancing text-to-image generation quality.
Prompt-OT: An Optimal Transport Regularization Paradigm for Knowledge Preservation in Vision-Language Model Adaptation
PositiveArtificial Intelligence
A new framework named Prompt-OT has been introduced to enhance the adaptation of vision-language models (VLMs) like CLIP, addressing challenges related to overfitting and zero-shot generalization during fine-tuning. This optimal transport-guided approach preserves the structural consistency of feature distributions between pre-trained and fine-tuned models, ensuring effective prompt learning.
LORE: Lagrangian-Optimized Robust Embeddings for Visual Encoders
PositiveArtificial Intelligence
The introduction of Lagrangian-Optimized Robust Embeddings (LORE) presents a new unsupervised adversarial fine-tuning framework aimed at enhancing the robustness of visual encoders against adversarial perturbations. This framework addresses critical limitations in existing fine-tuning strategies, particularly their instability and suboptimal trade-offs between robustness and accuracy on clean data.