Large Language Models (LLMs)

Articles

  1. Model size vs Computer overhead - The trade-off between model size and compute overhead and reveal there is significant room to reduce the compute-optimal model size with minimal compute overhead.

Papers

  1. LLM are few shot learners - scaling LLMs with data is enough to make them few shot.

Models

  1. StabilityLM

    1. Vicuna

    2. LLaMA

Instructor

  1. Instructor model - "We introduce InstructorπŸ‘¨β€πŸ«, an instruction-finetuned text embedding model that can generate text embeddings tailored to any task (e.g., classification, retrieval, clustering, text evaluation, etc.) and domains (e.g., science, finance, etc.) by simply providing the task instruction, without any finetuning. Instructor achieves sota on 70 diverse embedding tasks!"

Datasets

Tools

  1. LangChain

    1. An amazing tutorial in Youtube by Patrick Loeber about

      • LLMs

        • Prompt Templates

        • Chains

        • Agents and Tools

        • Memory

        • Document Loaders

        • Indexes

  2. LangFlow, Medium, HuggingFace - is a UI for LangChain, designed with react-flow to provide an effortless way to experiment and prototype flows.

  3. PandasAI - PandasAI, asking data Qs using LLMs on Panda's DFs with two code lines. πš™πšŠπš—πšπšŠπšœ_πšŠπš’ = π™ΏπšŠπš—πšπšŠπšœπ™°π™Έ(πš•πš•πš–) & πš™πšŠπš—πšπšŠπšœ_πšŠπš’.πš›πšžπš—(𝚍𝚏, πš™πš›πš˜πš–πš™πš='πš†πš‘πš’πšŒπš‘ πšŠπš›πšŽ πšπš‘πšŽ 𝟻 πš‘πšŠπš™πš™πš’πšŽπšœπš πšŒπš˜πšžπš—πšπš›πš’πšŽπšœ?')

  4. LLaMa Index - LlamaIndex (GPT Index) is a project that provides a central interface to connect your LLM's with external data.

  5. LLM-foundry - LLM training code for Databricks foundation models using MoasicML

  6. MinGPT - A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training

  7. NanoGPT - The simplest, fastest repository for training/finetuning medium-sized GPTs.

  8. Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)

Guardrails

  1. Databricks GR - Implementing LLM Guardrails for Safe and Responsible Generative AI Deployment on Databricks

Best Practices

Reinforcement Learning for LLM

  1. John Schulman - Reinforcement Learning from Human Feedback: Progress and Challenges

Metrics

  1. Understanding ROUGE - a family of metrics that evaluate the performance of a LLM in text summarization, i.e., ROUGE-1, ROUGE-2, ROUGE-L, for unigrams, bi grams, LCS, respectively.

Use Cases

Last updated