This repository is dedicated to following trends that lies at the intersection between medical image analysis and foundation models.
- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (2018)
- ALBERT: A Lite BERT for Self-supervised Learning of Language Representations (2019)
- Improving Language Understanding by Generative Pre-Training (GPT-1) (2018)
- Language Models are Unsupervised Multitask Learners (GPT-2) (2019)
- Language Models are Few-Shot Learners (GPT-3) (2020)
- Training language models to follow instructions with human feedback (InstructGPT/ChatGPT) (November 30, 2022)
- GPT-4 (March 14, 2023)
- Auto-GPT (March 19, 2023)
- Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension (2019)
- Training Compute-Optimal Large Language Models (Chinchilla) (2022)
- PaLM (2022)
- BLOOM: A 176B-Parameter Open-Access Multilingual Language Model (2022)
- OPT: Open Pre-trained Transformer Language Models (2022)
- LaMDA: Language Models for Dialog Applications (2022)
- BioMedLM GPT (2022)
- Vicuna (March, 2023)
- LLaMA: Open and Efficient Foundation Language Models (February 27, 2023)
- Alpaca (March 14, 2023)
- Scaling Transformer to 1M tokens and beyond with RMT (April 19, 2023)
- PaLM 2
- LLaMA 2
- BLOOMZ
- Luotuo
- Ziya
- MOSS
- Claude
- DoctorGLM
- ChatGLM
- ChatGLM 2
- VQA: Visual Question Answering (2015)
- ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks (2019)
- UNITER: UNiversal Image-TExt Representation Learning (2020)
- Scaling up visual and visionlanguage representation learning with noisy text supervision (ALGIN) (2021)
- Learning transferable visual models from natural language supervision (CLIP) (2021)
- VLMo: Unified Vision-Language Pre-Training with Mixture-of-Modality-Experts (2022)
- An Empirical Study of Training End-to-End Vision-and-Language Transformers (METER) (2022)
- Flamingo: a Visual Language Model for Few-Shot Learning (2022)
- Multi-Granularity Cross-modal Alignment for Generalized Medical Visual Representation Learning (MGCA) (2022)
- Generalized radiograph representation learning via cross-supervision between images and free-text radiology reports (2022)
- PaLM-E: An Embodied Multimodal Language Model (March 2023)
- Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models (March 8, 2023)
- HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in HuggingFace (March 30, 2023)
- InstructBLIP: Towards General-purpose Vision-Language Models with Instruction Tuning (May 11, 2023)
- Medical Visual Question Answering via Conditional Reasoning and Contrastive Learning (May 2023)
- Visual Programming: Compositional visual reasoning without training (2023)
- Scaling Down to Scale Up: A Guide to Parameter-Efficient Fine-Tuning (March 28, 2023)
- PMC-LLaMA: Further Finetuning LLaMA on Medical Papers (April 27, 2023)
- Multitask Prompt Tuning Enables Parameter-Efficient Transfer Learning (March 6, 2023)
- CODA-Prompt: COntinual Decomposed Attention-based Prompting for Rehearsal-Free Continual Learning (November 22, 2022)
- The Power of Scale for Parameter-Efficient Prompt Tuning (April 18, 2021)
- Prefix-Tuning: Optimizing Continuous Prompts for Generation (January 1, 2021)
- How to Fine-Tune BERT for Text Classification? (October 13, 2019)
- Lora: Low-rank adaptation of large language models (2021)
- GPT Understands, Too (2021) (P-Tuing)
- AutoPrompt: Eliciting Knowledge from Language Models with Automatically Generated Prompts (2020)
- Exploiting Cloze Questions for Few Shot Text Classification and Natural Language Inference (PET) (2021)
- It's Not Just Size That Matters: Small Language Models Are Also Few-Shot Learners (2021)
- Calibrate Before Use: Improving Few-shot Performance of Language Models (ICL) (2021)
- The Power of Scale for Parameter-Efficient Prompt Tuning (2021)
- WARP: Word-level Adversarial ReProgramming (2021)
- Learning to Prompt for Vision-Language Models (CoOp) (2022)
- Chain-of-Thought Prompting Elicits Reasoning in Large Language Models (CoT) (January 2022)
- React: Synergizing reasoning and acting in language models (2022)
- How Does In-Context Learning Help Prompt Tuning? (March 2, 2023)
- Enhancing Chain-of-Thoughts Prompting with Iterative Bootstrapping in Large Language Models (April 23, 2023)
- Prompt Engineering for Healthcare: Methodologies and Applications (April 28, 2023)
- P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks (March 20, 2022)
- SegGPT (April 6, 2023)
- Scaling Vision Transformers to 22 Billion Parameters (February 10, 2023)
- Painter (2022)
- ChatAug: Leveraging ChatGPT for Text Data Augmentation (February 25, 2023)
- A Survey on In-context Learning (2022)
- Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing (January 2023)
- VLP: A Survey on Vision-language Pre-training (February 2023)
- ChatGPT: the future of discharge summaries? (February 6, 2023)
- ChatCAD: Interactive Computer-Aided Diagnosis on Medical Image using Large Language Models (February 14, 2023)
- The Diagnostic and Triage Accuracy of the GPT-3 Artificial Intelligence Model (February 1, 2023)
- Foundation models for generalist medical artificial intelligence (April 12, 2023)
- Multilingual translation for zero-shot biomedical classification using BioTranslator (February 10, 2023)
- Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining (April 28, 2023)
- Large language models encode clinical knowledge (July 12, 2023)
- Language Models (Mostly) Know What They Know (2022)
- Teaching Models to Express Their Uncertainty in Words (2022)
- On the Planning Abilities of Large Language Models (A Critical Investigation with a Proposed Benchmark) (February 13, 2023)
- Evaluating ChatGPT's Information Extraction Capabilities: An Assessment of Performance, Explainability, Calibration, and Faithfulness (April 23, 2023)
- Sparks of Artificial General Intelligence: Early experiments with GPT-4 (March 22, 2023)
- Appropriateness of Recommendations Provided by ChatGPT to Interventional Radiologists (April 13, 2023)
- Using the Veil of Ignorance to align AI systems with principles of justice (April 24, 2023)
- Inducing anxiety in large language models increases exploration and bias (April 21, 2023)
- Are Emergent Abilities of Large Language Models a Mirage? (April 28, 2023)
- LIMA: Less Is More for Alignment (May 18, 2023)
- How Is ChatGPT’s Behavior Changing over Time? (July 18, 2023)
- Challenges and Applications of Large Language Models (July 19, 2023)
- Augmenting the national institutes of health chest radiograph dataset with expert annotations of possible pneumonia (2019)
- Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison (2019)
- MIMIC-CXR, a de-identified publicly available database of chest radiographs with free-text reports (2019)
- Covid-net: A tailored deep convolutional neural network design for detection of covid-19 cases from chest x-ray images (2020)
- Finetuning Large Language Models (April 22, 2023)
- Understanding Large Language Models (April 16, 2023)
- Understanding Parameter-Efficient Finetuning (April 12, 2023)
- Using LLM's output to circumvent data moat? Technical, business, and legal concerns discussed by Andrew Ng.
- What you'd like to share if you join a reading group on LLMs? Question rasied by Andrew g in LinkedIn.
- The Age of AI has begun (March 21, 2023)
- GPT-3: Its Nature, Scope, Limits, and Consequences (2020)
- ShareGPT: Share your wildest conversations with ChatGPT
- OpenAI API
- Transformers from Scratch (2021)
- Choose Your Weapon: Survival Strategies for Depressed AI Academics (March 31, 2023)
- Visualizing A Neural Machine Translation Model (Mechanics of Seq2seq Models With Attention) (2018)
- The Illustrated Transformer (2018)
- Jay Alammar's Homepage (Jay is keeping producing quality tutorials and talks for AI researchers and practitioners!)
- How GPT3 Works - Visualizations and Animations (July 2020)
- A Visual Guide to Using BERT for the First Time (November 2019)
- The Illustrated GPT-2 (Visualizing Transformer Language Models) (August 2019)
- The Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning) (December 2018)
- ChatGPT Prompt Engineering for Developers (April 27, 2023)
- OpenAI Cookbook
- What Is ChatGPT Doing … and Why Does It Work? by Stephen Wolfram (Frebruary 14, 2023)
- Chatbot Arena