
Lil'Log
Date: June 23, 2023 | Estimated Reading Time: 31 min | Author: Lilian Weng Prompt Engineering Prompt Engineering, also known as In-Context Prompting, refers to methods for how to communicate with LLM to steer its behavior for desired outcomes without updating the …
What are Diffusion Models? | Lil'Log - GitHub Pages
2021年7月11日 · [Updated on 2021-09-19: Highly recommend this blog post on score-based generative modeling by Yang Song (author of several key papers in the references)]. [Updated on 2022-08-27: Added classifier-free guidance, GLIDE, unCLIP and Imagen. [Updated on 2022-08-31: Added latent diffusion model. [Updated on 2024-04-13: Added progressive distillation, consistency models, and the Model Architecture ...
The Transformer Family Version 2.0 | Lil'Log - GitHub Pages
2023年1月27日 · Many new Transformer architecture improvements have been proposed since my last post on “The Transformer Family” about three years ago. Here I did a big refactoring and enrichment of that 2020 post — restructure the hierarchy of sections and improve many sections with more recent papers. Version 2.0 is a superset of the old version, about twice the length.
LLM Powered Autonomous Agents | Lil'Log - GitHub Pages
2023年6月23日 · Building agents with LLM (large language model) as its core controller is a cool concept. Several proof-of-concepts demos, such as AutoGPT, GPT-Engineer and BabyAGI, serve as inspiring examples. The potentiality of LLM extends beyond generating well-written copies, stories, essays and programs; it can be framed as a powerful general problem solver. Agent System Overview In a LLM-powered ...
Prompt Engineering | Lil'Log - GitHub Pages
2023年3月15日 · @article{weng2023prompt, title = "Prompt Engineering", author = "Weng, Lilian", journal = "lilianweng.github.io", year = "2023", month = "Mar", url = "https://lilianweng.github.io/posts/2023-03-15-prompt-engineering/" } Useful Resources# OpenAI Cookbook has many in-depth examples for how to utilize LLM efficiently.
The Transformer Family | Lil'Log - GitHub Pages
2020年4月7日 · See my old post for other types of attention if interested.. Multi-Head Self-Attention#. The multi-head self-attention module is a key component in Transformer. Rather than only computing the attention once, the multi-head mechanism splits the inputs into smaller chunks and then computes the scaled dot-product attention over each subspace in parallel.
Controllable Neural Text Generation | Lil'Log - GitHub Pages
[Updated on 2021-02-01: Updated to version 2.0 with several work added and many typos fixed.] [Updated on 2021-05-26: Add P-tuning and Prompt Tuning in the “prompt design” section.] [Updated on 2021-09-19: Add “unlikelihood training”.]. There is a gigantic amount of free text on the Web, several magnitude more than labelled benchmark datasets.
From GAN to WGAN | Lil'Log - GitHub Pages
[Updated on 2018-09-30: thanks to Yoonju, we have this post translated in Korean!] [Updated on 2019-04-18: this post is also available on arXiv.] Generative adversarial network (GAN) has shown great results in many generative tasks to replicate the real-world rich content such as images, human language, and music. It is inspired by …
Reinforcement-Learning | Lil'Log - GitHub Pages
2024年11月28日 · Date: May 5, 2019 | Estimated Reading Time: 15 min | Author: Lilian Weng Implementing Deep Reinforcement Learning Models with Tensorflow + OpenAI Gym The full implementation is available in lilianweng/deep-reinforcement-learning-gym In the previous two posts, I have introduced the algorithms of many deep reinforcement learning models.
Large Transformer Model Inference Optimization | Lil'Log - GitHub …
[Updated on 2023-01-24: add a small section on Distillation.] Large transformer models are mainstream nowadays, creating SoTA results for a variety of tasks. They are powerful but very expensive to train and use. The extremely high inference cost, in both time and memory, is a big bottleneck for adopting a powerful transformer for solving real-world tasks at scale.