
| Chang Shu1* | Baian Chen2* | Fangyu Liu1 | Zihao Fu1 | Ehsan Shareghi 3 | Nigel Collier1 |
| 1University of Cambridge 2Ruiping Health 3University of Monash |
Demo (insert GIF here) (Baian)
Overview
Domain-specific foundation models are extremely useful in the biomedical domain as biomedical text is highly specialized and contains many domain-specific terms and concepts that are not present in general domain text corpora such as Wikipedia and Books. Pre-training on large volumes of biomedical text has shown to improve the performance of language models on several biomedical text mining tasks when compared to existing publicly available biomedical PLMs. However, to the best of our knowldege, there is not exisiting multimodal foundationmodel Therefore, we develop the Visual Med-Alpaca, Resources:
We apologize for the inconvenience, but this project is currently undergoing internal ethical screening at Cambridge University. We anticipate releasing the following assets within the next 1-2 weeks. You are more than welcome to Join Our Waitlist, and we'll notify you as soon as they become available.
Model Architecture and Training Recipe
Overview of the model architecture and training procedure.
Domain Adaptation: Self-Instruct in Biomedical Domain (Baian)
How to generate the instruct-tuning set
Visual Adaptation: Deplot and Medical VQA (Baian)
We also build a large-scale, high-quality video dataset, Vimeo90K. This dataset consists of 89,800 video clips downloaded from vimeo.com, which covers large variaty of scenes and actions. It is designed for the following four video processing tasks: temporal frame interpolation, video denoising, video deblocking, and video super-resolution.
Sampled Frames (Full-resolution samples are here):
Implementation Details
Hyper-parameter
Training time
Comparison with Other Methods
Compare with ChatGPT / Alpaca / Galactica
Future Work
Compare with ChatGPT / Alpaca / Galactica
Limitations
Visual Med-Alpaca, is intended for academic research purposes only. Any commercial or clinical use of the model is strictly prohibited. This decision is based on the non-commercial license inherited from LLaMA, on which the model is built. Additionally, Visual Med-Alpaca is not legally approved for medical use in any country. Users should be aware of the model's limitations in terms of medical knowledge and the possibility of misinformation. Therefore, any reliance on Visual Med-Alpaca for medical decision-making is at the user's own risk.
Note: The developers and owners of the model, the Language Technology Lab at Cambridge University, do not assume any liability for the accuracy or completeness of the information provided by Visual Med-Alpaca, nor will they be responsible for any potential harm caused by the misuse of the model.
Acknowledgement
We are deeply grateful for the contributions made by open-source projects:
LLaMA,
Stanford Alpaca,
Alpaca-LoRA,
Deplot,
BigBio,
ROCO,
Visual-ChatGPT,
GenerativeImage2Text.
|