753:
742:
731:
56:?", a command such as "write a poem about leaves falling", or a longer statement including context, instructions, and conversation history. Prompt engineering may involve phrasing a query, specifying a style, providing relevant context or assigning a role to the AI such as "Act as a native French speaker". A prompt may include a few examples for a model to learn from, such as asking the model to complete "maison → house, chat → cat, chien →" (the expected response being
5207:
4295:
541:
retrieval, the LLM generates an output that incorporates information from both the query and the retrieved documents. This method is particularly beneficial for handling proprietary or dynamic information that was not included in the initial training or fine-tuning phases of the model. RAG is also notable for its use of "few-shot" learning, where the model uses a small number of examples, often automatically retrieved from a database, to inform its outputs.
5187:
3988:
510:
550:
456:
815:(TTV) generation is an emerging technology enabling the creation of videos directly from textual descriptions. This field holds potential for transforming video production, animation, and storytelling. By utilizing the power of artificial intelligence, TTV allows users to bypass traditional video editing tools and translate their ideas into moving images.
3040:
Madaan, Aman; Tandon, Niket; Gupta, Prakhar; Hallinan, Skyler; Gao, Luyu; Wiegreffe, Sarah; Alon, Uri; Dziri, Nouha; Prabhumoye, Shrimai; Yang, Yiming; Gupta, Shashank; Prasad
Majumder, Bodhisattwa; Hermann, Katherine; Welleck, Sean; Yazdanbakhsh, Amir (2023-03-01). "Self-Refine: Iterative Refinement
1893:
model (such as an LLM) which was trained to follow human-given instructions to follow instructions provided by a malicious user. This stands in contrast to the intended operation of instruction-following systems, wherein the ML model is intended only to follow trusted instructions (prompts) provided
540:
In response to a query, a document retriever selects the most relevant documents. This relevance is typically determined by first encoding both the query and the documents into vectors, then identifying documents whose vectors are closest in
Euclidean distance to the query vector. Following document
436:
But if one cannot access such scores (such as when one is accessing the model through a restrictive API), uncertainty can still be estimated and incorporated into the model output. One simple method is to prompt the model to use words to estimate uncertainty. Another is to prompt the model to refuse
205:
For example, given the question "Q: The cafeteria had 23 apples. If they used 20 to make lunch and bought 6 more, how many apples do they have?", a CoT prompt might induce the LLM to answer "A: The cafeteria had 23 apples originally. They used 20 to make lunch. So they had 23 - 20 = 3. They bought 6
2498:
Bach, Stephen H.; Sanh, Victor; Yong, Zheng-Xin; Webson, Albert; Raffel, Colin; Nayak, Nihal V.; Sharma, Abheesht; Kim, Taewoon; M Saiful Bari; Fevry, Thibault; Alyafeai, Zaid; Dey, Manan; Santilli, Andrea; Sun, Zhiqing; Ben-David, Srulik; Xu, Canwen; Chhablani, Gunjan; Wang, Han; Jason Alan Fries;
697:
documentation encourages short, descriptive prompts: instead of "Show me a picture of lots of blooming
California poppies, make them bright, vibrant orange, and draw them in an illustrated style with colored pencils", an effective prompt might be "Bright orange California poppies drawn with colored
2796:
Chung, Hyung Won; Hou, Le; Longpre, Shayne; Zoph, Barret; Tay, Yi; Fedus, William; Li, Yunxuan; Wang, Xuezhi; Dehghani, Mostafa; Brahma, Siddhartha; Webson, Albert; Gu, Shixiang Shane; Dai, Zhuyun; Suzgun, Mirac; Chen, Xinyun; Chowdhery, Aakanksha; Castro-Ros, Alex; Pellat, Marie; Robinson, Kevin;
340:
Self-refine prompts the LLM to solve the problem, then prompts the LLM to critique its solution, then prompts the LLM to solve the problem again in view of the problem, solution, and critique. This process is repeated until stopped, either by running out of tokens, time, or by the LLM outputting a
2291:
Wei, Jason; Tay, Yi; Bommasani, Rishi; Raffel, Colin; Zoph, Barret; Borgeaud, Sebastian; Yogatama, Dani; Bosma, Maarten; Zhou, Denny; Metzler, Donald; Chi, Ed H.; Hashimoto, Tatsunori; Vinyals, Oriol; Liang, Percy; Dean, Jeff; Fedus, William (31 August 2022). "Emergent
Abilities of Large Language
2252:
Wei, Jason; Tay, Yi; Bommasani, Rishi; Raffel, Colin; Zoph, Barret; Borgeaud, Sebastian; Yogatama, Dani; Bosma, Maarten; Zhou, Denny; Metzler, Donald; Chi, Ed H.; Hashimoto, Tatsunori; Vinyals, Oriol; Liang, Percy; Dean, Jeff; Fedus, William (31 August 2022). "Emergent
Abilities of Large Language
276:
Input: There are a set of bricks. The yellow brick C is on top of the brick E. The yellow brick D is on top of the brick A. The yellow brick E is on top of the brick D. The white brick A is on top of the brick B. For the brick B, the color is white. Now we have to get a specific brick. The bricks
557:
GraphRAG, coined by
Microsoft Research, extends RAG such that instead of relying solely on vector similarity (as in most RAG approaches), GraphRAG uses the LLM-generated knowledge graph. This graph allows the model to connect disparate pieces of information, synthesize insights, and holistically
71:
or a text-to-audio model, a typical prompt is a description of a desired output such as "a high-quality photo of an astronaut riding a horse" or "Lo-fi slow BPM electro chill with organic samples". Prompting a text-to-image model may involve adding, removing, emphasizing and re-ordering words to
383:
prompting is similar to tree-of-thought. The model is prompted to answer a question with an explanation. The model is then prompted to explain parts of the explanation, and so on. Inconsistent explanation trees are pruned or discarded. This improves performance on complex commonsense reasoning.
2265:
In prompting, a pre-trained language model is given a prompt (e.g. a natural language instruction) of a task and completes the response without any further training or gradient updates to its parameters... The ability to perform a task via few-shot prompting is emergent when a model has random
561:
Researchers have demonstrated GraphRAG's effectiveness using datasets like the
Violent Incident Information from News Articles (VIINA). By combining LLM-generated knowledge graphs with graph machine learning, GraphRAG substantially improves both the comprehensiveness and diversity of generated
269:
Chain-of-Symbol prompting in conjunction with CoT prompting assists LLMs with its difficulty of spatial reasoning in text. In other words, using arbitrary symbols such as ' / ' assist the LLM to interpret spacing in text. This assists in reasoning and increases the performance of the LLM.
616:. Questions nearest to the centroids of each cluster are selected. An LLM does zero-shot CoT on each question. The resulting CoT examples are added to the dataset. When prompted with a new question, CoT examples to the nearest questions can be retrieved and added to the prompt.
2797:
Valter, Dasha; Narang, Sharan; Mishra, Gaurav; Yu, Adams; Zhao, Vincent; Huang, Yanping; Dai, Andrew; Yu, Hongkun; Petrov, Slav; Chi, Ed H.; Dean, Jeff; Devlin, Jacob; Roberts, Adam; Zhou, Denny; Le, Quoc V.; Wei, Jason (2022). "Scaling
Instruction-Finetuned Language Models".
140:
tasks (using 62 datasets, as each task can have multiple datasets). The model showed good performance on new tasks, surpassing models trained directly on just performing one task (without pretraining). To solve a task, T0 is given the task in a structured prompt, for example
132:
could be cast as a question answering problem over a context. In addition, they trained a first single, joint, multi-task model that would answer any task-related question like "What is the sentiment" or "Translate this sentence to German" or "Who is the president?"
2499:
Al-shaibani, Maged S.; Sharma, Shanya; Thakker, Urmish; Almubarak, Khalid; Tang, Xiangru; Radev, Dragomir; Mike Tian-Jian Jiang; Rush, Alexander M. (2022). "PromptSource: An
Integrated Development Environment and Repository for Natural Language Prompts".
2948:
Zhou, Denny; Schärli, Nathanael; Hou, Le; Wei, Jason; Scales, Nathan; Wang, Xuezhi; Schuurmans, Dale; Cui, Claire; Bousquet, Olivier; Le, Quoc; Chi, Ed (2022-05-01). "Least-to-Most
Prompting Enables Complex Reasoning in Large Language Models".
186:(LLMs) to solve a problem as a series of intermediate steps before giving a final answer. Chain-of-thought prompting improves reasoning ability by inducing the model to answer a multi-step problem with steps of reasoning that mimic a
2370:
By the time you type a query into ChatGPT, the network should be fixed; unlike humans, it should not continue to learn. So it came as a surprise that LLMs do, in fact, learn from their users' prompts—an ability known as in-context
107:
for each specific task, which are not temporary, what has been learnt during in-context learning is of a temporary nature. It does not carry the temporary contexts or biases, except the ones already present in the (pre)training
3755:
Kirillov, Alexander; Mintun, Eric; Ravi, Nikhila; Mao, Hanzi; Rolland, Chloe; Gustafson, Laura; Xiao, Tete; Whitehead, Spencer; Berg, Alexander C.; Lo, Wan-Yen; Dollár, Piotr; Girshick, Ross (2023-04-01). "Segment Anything".
528:
and answer formulation by a Large Language Model (LLM). The initial phase utilizes dense embeddings to retrieve documents. This retrieval can be based on a variety of database formats depending on the use case, such as a
3179:
3431:
Guo, Qingyan; Wang, Rui; Guo, Junliang; Li, Bei; Song, Kaitao; Tan, Xu; Liu, Guoqing; Bian, Jiang; Yang, Yujiu (2023). "Connecting Large Language Models with Evolutionary Algorithms Yields Powerful Prompt Optimizers".
1720:
An earlier result uses the same idea of gradient descent search, but is designed for masked language models like BERT, and searches only over token sequences, rather than numerical vectors. Formally, it searches for
1083:
1019:
955:
1132:
323:
performs several chain-of-thought rollouts, then selects the most commonly reached conclusion out of all the rollouts. If the rollouts disagree by a lot, a human can be queried for the correct chain of thought.
3730:
Gal, Rinon; Alaluf, Yuval; Atzmon, Yuval; Patashnik, Or; Bermano, Amit H.; Chechik, Gal; Cohen-Or, Daniel (2022). "An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion".
288:
first prompts the model to generate relevant facts for completing the prompt, then proceed to complete the prompt. The completion quality is usually higher, as the model can be conditioned on relevant facts.
3230:
Lewis, Patrick; Perez, Ethan; Piktus, Aleksandra; Petroni, Fabio; Karpukhin, Vladimir; Goyal, Naman; Küttler, Heinrich; Lewis, Mike; Yih, Wen-tau; Rocktäschel, Tim; Riedel, Sebastian; Kiela, Douwe (2020).
3843:
In this work, we explore "prompt tuning," a simple yet effective mechanism for learning "soft prompts"...Unlike the discrete text prompts used by GPT-3, soft prompts are learned through back-propagation
2973:
Wang, Xuezhi; Wei, Jason; Schuurmans, Dale; Le, Quoc; Chi, Ed; Narang, Sharan; Chowdhery, Aakanksha; Zhou, Denny (2022-03-01). "Self-Consistency Improves Chain of Thought Reasoning in Language Models".
569:
for question answering using text-to-query generation. These techniques can be combined to perform search across both unstructured and structured data, providing expanded context and improved ranking.
277:
must now be grabbed from top to bottom, and if the lower brick is to be grabbed, the upper brick must be removed first. How to get brick D? B/A/D/E/C C/E E/D D Output: So we get the result as C, E, D.
3107:
Jung, Jaehun; Qin, Lianhui; Welleck, Sean; Brahman, Faeze; Bhagavatula, Chandra; Le Bras, Ronan; Choi, Yejin (2022). "Maieutic Prompting: Logically Consistent Reasoning with Recursive Explanations".
598:
Each of the generated instructions is used to prompt the target LLM, followed by each of the inputs. The log-probabilities of the outputs are computed and added. This is the score of the instruction.
1827:
1460:
3083:
Yao, Shunyu; Yu, Dian; Zhao, Jeffrey; Shafran, Izhak; Griffiths, Thomas L.; Cao, Yuan; Narasimhan, Karthik (2023-05-17). "Tree of Thoughts: Deliberate Problem Solving with Large Language Models".
2383:
Johannes von Oswald; Niklasson, Eyvind; Randazzo, Ettore; Sacramento, JoĂŁo; Mordvintsev, Alexander; Zhmoginov, Andrey; Vladymyrov, Max (2022). "Transformers learn in-context by gradient descent".
595:
Prompting LLM is presented with example input-output pairs, and asked to generate instructions that could have caused a model following the instructions to generate the outputs, given the inputs.
1164:
to prompt-specific parameters: in prefix-tuning, they are parameters associated with the prompt tokens at each layer; in prompt tuning, they are merely the soft tokens added to the vocabulary.
3779:
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)
3143:
The directional stimulus serves as hints or cues for each input query to guide LLMs toward the desired output, such as keywords that the desired summary should include for summarization.
306:
prompts a model to first list the sub-problems to a problem, then solve them in sequence, such that later sub-problems can be solved with the help of answers to previous sub-problems.
421:
By default, the output of language models may not contain estimates of uncertainty. The model may output text that appears confident, though the underlying token predictions have low
332:
Complexity-based prompting performs several CoT rollouts, then select the rollouts with the longest chains of thought, then select the most commonly reached conclusion out of those.
1343:
3474:
Zhou, Yongchao; Ioan Muresanu, Andrei; Han, Ziwen; Paster, Keiran; Pitis, Silviu; Chan, Harris; Ba, Jimmy (2022-11-01). "Large Language Models Are Human-Level Prompt Engineers".
3743:
Using only 3-5 images of a user-provided concept, like an object or a style, we learn to represent it through new "words" in the embedding space of a frozen text-to-image model.
3194:
If you explicitly indicate in your prompt that you want the generative AI to emit a certainty or uncertainty qualification then you will almost certainly get such an indication.
261:
Chain-of-thought prompting is just one of many prompt-engineering techniques. Various other techniques have been proposed. At least 29 distinct techniques have been published.
3347:
Fernando, Chrisantha; Banarse, Dylan; Michalewski, Henryk; Osindero, Simon; Rocktäschel, Tim (2023). "Promptbreeder: Self-Referential Self-Improvement Via Prompt Evolution".
4027:
1224:
30:
This article is about means of interacting (command prompts) with an artificial intelligence system. For general computer command line interfaces and commando entries, see
1856:
1715:
1663:
1158:
5081:
1085:
be the token embeddings of the input and output respectively. During training, the tunable embeddings, input, and output tokens are concatenated into a single sequence
1634:
2147:
Brown, Tom; Mann, Benjamin; Ryder, Nick; Subbiah, Melanie; Kaplan, Jared D.; Dhariwal, Prafulla; Neelakantan, Arvind (2020). "Language models are few-shot learners".
1544:
1598:
1571:
364:
generalizes chain-of-thought by prompting the model to generate one or more "possible next steps", and then running the model on each of the possible next steps by
858:
based on a set of example images. This embedding vector acts as a "pseudo-word" which can be included in a prompt to express the content or style of the examples.
433:
likelihood scores in their token predictions, and so the model output uncertainty can be directly estimated by reading out the token prediction likelihood scores.
3389:
Pryzant, Reid; Iter, Dan; Li, Jerry; Lee, Yin Tat; Zhu, Chenguang; Zeng, Michael (2023). "Automatic Prompt Optimization with "Gradient Descent" and Beam Search".
2961:...least-to-most prompting. The key idea in this strategy is to break down a complex problem into a series of simpler subproblems and then solve them in sequence.
752:
741:
413:
Article: {article} Q: Write a short summary of the article in 2-4 sentences that accurately incorporates the provided keywords. Keywords: {keywords} A:
1683:
1284:
1264:
1244:
3131:
Li, Zekun; Peng, Baolin; He, Pengcheng; Galley, Michel; Gao, Jianfeng; Yan, Xifeng (2023). "Guiding Large Language Models via Directional Stimulus Prompting".
792:
Text-to-image models do not natively understand negation. The prompt "a party with no cake" is likely to produce an image including a cake. As an alternative,
3588:
2432:
Garg, Shivam; Tsipras, Dimitris; Liang, Percy; Valiant, Gregory (2022). "What Can Transformers Learn In-Context? A Case Study of Simple Function Classes".
2126:
Garg, Shivam; Tsipras, Dimitris; Liang, Percy; Valiant, Gregory (2022). "What Can Transformers Learn In-Context? A Case Study of Simple Function Classes".
4329:
2325:
2456:
McCann, Bryan; Shirish, Nitish; Xiong, Caiming; Socher, Richard (2018). "The Natural Language Decathlon: Multitask Learning as Question Answering".
4923:
878:
by prompting. As an alternative to text prompts, Segment Anything can accept bounding boxes, segmentation masks, and foreground/background points.
714:
2628:
2599:
1024:
960:
896:
2324:
Wei, Jason; Wang, Xuezhi; Schuurmans, Dale; Bosma, Maarten; Ichter, Brian; Xia, Fei; Chi, Ed H.; Le, Quoc V.; Zhou, Denny (31 October 2022).
1088:
3306:
A Benchmark to Understand the Role of Knowledge Graphs on Large Language Model's Accuracy for Question Answering on Enterprise SQL Databases
3264:
2444:
Training a model to perform in-context learning can be viewed as an instance of the more general learning-to-learn or meta-learning paradigm
2177:
2749:
Kojima, Takeshi; Shixiang Shane Gu; Reid, Machel; Matsuo, Yutaka; Iwasawa, Yusuke (2022). "Large Language Models are Zero-Shot Reasoners".
534:
5271:
4298:
3517:
347:
I have some code. Give one suggestion to improve readability. Don't fix the code, just give a suggestion. Code: {code} Suggestion:
2995:
Diao, Shizhe; Wang, Pengcheng; Lin, Yong; Zhang, Tong (2023-02-01). "Active Prompting with Chain-of-Thought for Large Language Models".
2080:
1989:
We demonstrate language models can perform down-stream tasks in a zero-shot setting – without any parameter or architecture modification
1981:
625:
4097:
3016:
Fu, Yao; Peng, Hao; Sabharwal, Ashish; Clark, Peter; Khot, Tushar (2022-10-01). "Complexity-Based Prompting for Multi-Step Reasoning".
2048:"Basic prompt: 'Write a poem about leaves falling.' Better prompt: 'Write a poem in the style of Edgar Allan Poe about leaves falling.'
3180:"Latest Prompt Engineering Technique Aims To Get Certainty And Uncertainty Of Generative AI Directly On The Table And Out In The Open"
2771:
2214:
Next, I gave a more complicated prompt to attempt to throw MusicGen for a loop: "Lo-fi slow BPM electro chill with organic samples."
4107:
3939:
3799:
In this paper, we propose prefix-tuning, a lightweight alternative to fine-tuning... Prefix-tuning draws inspiration from prompting
3495:
Zhang, Zhuosheng; Zhang, Aston; Li, Mu; Smola, Alex (2022-10-01). "Automatic Chain of Thought Prompting in Large Language Models".
2721:
608:
CoT examples can be generated by LLM themselves. In "auto-CoT", a library of questions are converted to vectors by a model such as
213:
prompting technique. However, simply appending the words "Let's think step-by-step", has also proven effective, which makes CoT a
4439:
4034:
2898:
Liu, Jiacheng; Liu, Alisa; Lu, Ximing; Welleck, Sean; West, Peter; Le Bras, Ronan; Choi, Yejin; Hajishirzi, Hannaneh (May 2022).
2615:'Chain-of-thought prompting allows us to describe multistep problems as a series of intermediate steps,' Google CEO Sundar Pichai
93:
85:
3232:
1928:
4322:
701:
Word order affects the output of a text-to-image prompt. Words closer to the start of a prompt may be emphasized more heavily.
217:
prompting technique. This allows for better scaling as a user no longer needs to formulate many specific CoT Q&A examples.
100:
in downstream scaling laws occur such that its efficacy increases at a different rate in larger models than in smaller models.
713:
has been used in Stable Diffusion and Midjourney prompts to generate images in the distinctive style of Polish digital artist
1724:
1348:
3855:
Sun, Simeng; Liu, Yang; Iter, Dan; Zhu, Chenguang; Iyyer, Mohit (2023). "How Does In-Context Learning Help Prompt Tuning?".
3280:
Edge, Darren; Trinh, Ha; Cheng, Newman; Bradley, Joshua; Chao, Alex; Mody, Apurva; Truitt, Steven; Larson, Jonathan (2024),
5251:
5112:
730:
152:
A repository for prompts reported that over 2,000 public prompts for around 170 datasets were available in February 2022.
5213:
4764:
4501:
2395:
Thus we show how trained Transformers become mesa-optimizers i.e. learn models by gradient descent in their forward pass
4072:
4043:
243:. It is possible to fine-tune models on CoT reasoning datasets to enhance this capability further and stimulate better
17:
2420:
Mesa-Optimization is the situation that occurs when a learned model (such as a neural network) is itself an optimizer.
5025:
4652:
4459:
4315:
2202:
496:
709:
Some text-to-image models are capable of imitating the style of particular artists by name. For example, the phrase
4980:
3707:"Understanding OpenAI's Sora: A Revolutionary Leap | PromptSora: Discover Prompts and Videos for Sora from Open AI"
2669:
117:
3811:
Lester, Brian; Al-Rfou, Rami; Constant, Noah (2021). "The Power of Scale for Parameter-Efficient Prompt Tuning".
3206:
2654:
2352:
4003:
2819:
2695:
2566:
886:
In "prefix-tuning", "prompt tuning" or "soft prompting", floating-point-valued vectors are searched directly by
5256:
5167:
5107:
4705:
4092:
4067:
2904:
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
2521:
521:
229:
113:
1959:
4700:
4389:
3706:
2061:
1903:
2844:
Sahoo, Pranab; Singh, Ayush Kumar; Saha, Sriparna; Jain, Vinija; Mondal, Samrat; Chadha, Aman (2024-02-05),
84:, defined as a model's ability to temporarily learn from prompts. The ability for in-context learning is an
5142:
4496:
4449:
4444:
2547:
1289:
650:
images. Text-to-image models typically do not understand grammar and sentence structure in the same way as
3681:
471:. In particular, it dives into the technical vector implementation before positioning the overall concept.
5193:
4489:
4415:
137:
129:
104:
353:
Code: {code} Let's use this suggestion to improve the code. Suggestion: {suggestion} New Code:
4817:
4752:
4353:
4216:
2477:
Sanh, Victor; et al. (2021). "Multitask Prompted Training Enables Zero-Shot Task Generalization".
834:
553:
GraphRAG with a knowledge graph combining access patterns for unstructured, structured and mixed data.
404:
includes a hint or cue, such as desired keywords, to guide a language model toward the desired output.
53:
5246:
5218:
5076:
4715:
4546:
4369:
1886:
97:
3777:
Li, Xiang Lisa; Liang, Percy (2021). "Prefix-Tuning: Optimizing Continuous Prompts for Generation".
3460:
3417:
3375:
1170:
228:, CoT prompting significantly aided the model, allowing it to perform comparably with task-specific
5261:
5117:
4374:
4102:
3876:
Shin, Taylor; Razeghi, Yasaman; Logan IV, Robert L.; Wallace, Eric; Singh, Sameer (November 2020).
1942:
Prompt engineering is the process of structuring words that can be interpreted and understood by a
474:
430:
3613:
1832:
1691:
1639:
1141:
190:. It allows large language models to overcome difficulties with some reasoning tasks that require
5162:
5147:
4800:
4795:
4695:
4563:
4344:
4150:
4145:
4112:
577:
Large language models (LLM) themselves can be used to compose prompts for large language models.
478:
2655:"Pathways Language Model (PaLM): Scaling to 540 Billion Parameters for Breakthrough Performance"
5241:
5122:
4882:
4601:
4596:
4170:
4020:
2846:
A Systematic Survey of Prompt Engineering in Large Language Models: Techniques and Applications
1946:
model. Think of it as the language you need to speak in order to tell an AI model what to draw.
854:
For text-to-image models, "Textual inversion" performs an optimization process to create a new
800:
appear in the resulting image. A common approach is to include generic undesired terms such as
237:
31:
3882:
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)
3328:
Singh, Chandan; Morris, John; Aneja, Jyoti; Rush, Alexander; Gao, Jianfeng (October 4, 2022).
1980:
Radford, Alec; Wu, Jeffrey; Child, Rewon; Luan, David; Amodei, Dario; Sutskever, Ilya (2019).
1603:
646:
were released to the public. These models take text prompts as input and use them to generate
5152:
5137:
5102:
4790:
4690:
4558:
3447:
3404:
3362:
2105:
1465:
651:
609:
240:
199:
183:
5020:
3964:
604:
Repeat until some stopping criteria is reached, then output the highest-scored instructions.
295:
Generate some knowledge about the concepts in the input. Input: {question} Knowledge:
5266:
5172:
5127:
4573:
4518:
4364:
4359:
4077:
3878:"AutoPrompt: Eliciting Knowledge from Language Models with Automatically Generated Prompts"
1877:
1576:
1549:
365:
89:
49:
2871:
Hu, Hanxu; Lu, Hongyuan; Zhang, Huajian; Song, Yun-Ze; Lam, Wai; Zhang, Yue (2023-10-03),
662:
A text-to-image prompt commonly includes a description of the subject of the art (such as
8:
4747:
4725:
4474:
4469:
4427:
4379:
4208:
4180:
3914:
2357:
2266:
performance until a certain scale, after which performance increases to well-above random
812:
466:
422:
166:
In 2023 several text-to-text and text-to-image prompt databases were publicly available.
146:
40:
is the process of structuring an instruction that can be interpreted and understood by a
3993:
3657:
209:
As originally proposed, each CoT prompt included a few Q&A examples. This made it a
136:
In 2021, researchers fine-tuned one generatively pretrained model (T0) on performing 12
5132:
4710:
4117:
3895:
3856:
3834:
3816:
3790:
3757:
3732:
3496:
3475:
3433:
3390:
3348:
3309:
3285:
3244:
3156:
3132:
3108:
3084:
3063:
3042:
3017:
2996:
2975:
2950:
2927:
2907:
2876:
2849:
2798:
2750:
2500:
2478:
2457:
2433:
2407:
2384:
2331:
2293:
2254:
2156:
2127:
1668:
1269:
1249:
1229:
875:
525:
3890:
3830:
3813:
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing
846:
Some approaches augment or replace natural language text prompts with non-text input.
601:
The highest-scored instructions are given to the prompting LLM for further variations.
5198:
5186:
4990:
4642:
4513:
4506:
3899:
3838:
3794:
2931:
2574:
191:
3786:
2922:
2178:"This horse-riding astronaut is a milestone on AI's long road towards understanding"
1286:
is the rest of the model. In prefix-tuning, one provide a set of input-output pairs
4943:
4933:
4740:
4534:
4484:
4479:
4422:
4410:
4198:
3885:
3826:
3782:
2917:
1890:
1867:
887:
822:
762:
639:
613:
244:
233:
187:
45:
3330:"Explaining Patterns in Data with Language Models via Interpretable Autoprompting"
2330:. Advances in Neural Information Processing Systems (NeurIPS 2022). Vol. 35.
5056:
5000:
4822:
4464:
4384:
4262:
3566:
1161:
871:
566:
530:
380:
112:, from one conversation to the other. This result of "mesa-optimization" within
48:
text describing the task that an AI should perform: a prompt for a text-to-text
5030:
4995:
4985:
4810:
4568:
4394:
4122:
4082:
3877:
867:
855:
514:
225:
3614:"Stable Diffusion 2.0 and the Importance of Negative Prompts for Good Results"
2382:
2226:
143:
If {{premise}} is true, is it also true that {{hypothesis}}? ||| {{entailed}}.
5235:
4975:
4955:
4872:
4551:
4062:
2899:
2578:
2278:
2081:"How to Prime and Prompt ChatGPT for More Reliable Contract Drafting Support"
1873:
1135:
631:
68:
41:
592:
There are two LLMs. One is the target LLM, and another is the prompting LLM.
5061:
4892:
4307:
4267:
3546:
1968:
Prompt engineering is the art of communicating with a generative AI model.
437:
to answer in a standardized way if the input does not satisfy conditions.
128:
In 2018, researchers first proposed that all previously separate tasks in
5157:
4928:
4837:
4832:
4454:
4432:
4057:
3884:. Online: Association for Computational Linguistics. pp. 4222–4235.
3282:
From Local to Global: A Graph RAG Approach to Query-Focused Summarization
3062:
Long, Jieyi (2023-05-15). "Large Language Model Guided Tree-of-Thought".
2906:. Dublin, Ireland: Association for Computational Linguistics: 3154–3169.
1078:{\displaystyle \mathbf {Y} =\{\mathbf {y_{1}} ,\dots ,\mathbf {y_{n}} \}}
1014:{\displaystyle \mathbf {X} =\{\mathbf {x_{1}} ,\dots ,\mathbf {x_{m}} \}}
950:{\displaystyle \mathbf {E} =\{\mathbf {e_{1}} ,\dots ,\mathbf {e_{k}} \}}
761:
Demonstration of the effect of negative prompts on images generated with
585:
369:
3658:"Introducing Make-A-Video: An AI system that generates videos from text"
3589:"This Artist Is Dominating AI-Generated Art and He's Not Happy About It"
837:– As yet unreleased, Sora purportedly can produce high-resolution videos
5051:
5010:
5005:
4918:
4827:
4735:
4647:
4627:
4282:
4193:
2696:"Google's Chain of Thought Prompting Can Boost Today's Best Algorithms"
2227:"How to Write AI Photoshoot Prompts: A Guide for Better Product Photos"
1127:{\displaystyle {\text{concat}}(\mathbf {E} ;\mathbf {X} ;\mathbf {Y} )}
694:
643:
312:
Input: Q: {question} A: Let's break down this problem: 1.
195:
2277:
Caballero, Ethan; Gupta, Kshitij; Rish, Irina; Krueger, David (2022).
2033:
5046:
5015:
4913:
4757:
4720:
4657:
4611:
4606:
4591:
4257:
2722:"Amazon's Alexa scientists demonstrate bigger AI isn't always better"
2652:
2327:
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
831:
Make-a-Video – Focuses on creating detailed and diverse video outputs
558:
understand summarized semantic concepts over large data collections.
2281:. International Conference on Learning Representations (ICLR), 2023.
825:– Offers a user-friendly interface and supports various video styles
4948:
4780:
3940:"What's Old Is New Again: GPT-3 Prompt Injection Attack Affects AI"
3861:
3821:
3762:
3737:
3633:
3501:
3480:
3438:
3395:
3353:
3329:
3314:
3290:
3249:
3161:
3137:
3113:
3089:
3068:
3047:
3022:
3001:
2980:
2955:
2912:
2881:
2873:
Chain-of-Symbol Prompting Elicits Planning in Large Language Models
2854:
2803:
2772:"LLMs have not learned our language — we're trying to learn theirs"
2755:
2505:
2483:
2462:
2438:
2389:
2336:
2298:
2259:
2161:
2132:
2062:"The ultimate guide to prompt engineering your GPT-3.5-Turbo model"
796:
allow a user to indicate, in a separate prompt, which terms should
635:
109:
72:
achieve a desired subject, style, layout, lighting, and aesthetic.
3346:
3233:"Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks"
2001:
5071:
4908:
4862:
4785:
4685:
4680:
4632:
4277:
4232:
2748:
1717:
is preappended to the hidden states in every layer of the model.
626:
Artificial intelligence art § Prompt engineering and sharing
4012:
5086:
5066:
4938:
4730:
4272:
4188:
3183:
1167:
More formally, this is prompt tuning. Let an LLM be written as
881:
647:
160:
2629:"Google's Latest AI Model Can Be Taught How to Solve Problems"
2600:"Google's Latest AI Model Can Be Taught How to Solve Problems"
4887:
4867:
4857:
4852:
4847:
4842:
4805:
4637:
4165:
4160:
4155:
3518:"Dall-E2 VS Stable Diffusion: Same Prompt, Different Results"
509:
426:
1973:
572:
4877:
4237:
3266:
GraphRAG: Unlocking LLM discovery on narrative private data
1927:
Diab, Mohamad; Herrera, Julian; Chernow, Bob (2022-10-28).
957:
be a set of soft prompt tokens (tunable embeddings), while
549:
221:
3875:
3473:
1688:
For prefix tuning, it is similar, but the "prefix vector"
3039:
2900:"Generated Knowledge Prompting for Commonsense Reasoning"
3965:"GPT-3 'prompt injection' attack causes bot bad manners"
3729:
2893:
2891:
2522:"Language Models Perform Reasoning via Chain of Thought"
2455:
3229:
2431:
2125:
1822:{\displaystyle \arg \max _{\tilde {X}}\sum _{i}\log Pr}
1455:{\displaystyle \arg \max _{\tilde {Z}}\sum _{i}\log Pr}
828:
Lumiere – Designed for high-resolution video generation
654:, and require a different set of prompting techniques.
544:
206:
more apples, so they have 3 + 6 = 9. The answer is 9."
2670:"Harnessing the power of GPT-3 in scientific research"
2323:
2245:
1858:
is ranges over token sequences of a specified length.
3810:
3754:
3279:
3106:
2888:
2567:"Get the Best From ChatGPT With These Golden Prompts"
2290:
2251:
2146:
1982:"Language Models are Unsupervised Multitask Learners"
1835:
1727:
1694:
1671:
1642:
1606:
1579:
1552:
1468:
1351:
1292:
1272:
1252:
1232:
1173:
1144:
1091:
1027:
963:
899:
517:
and Large Language Model (LLM) for answer formulation
3327:
3304:
Sequeda, Juan; Allemang, Dean; Jacob, Bryon (2023),
3303:
3015:
2653:
Sharan Narang and Aakanksha Chowdhery (2022-04-04).
2476:
1960:"A developer's guide to prompt engineering and LLMs"
513:
Two-phase process of document retrieval using dense
416:
3035:
3033:
2972:
2795:
2744:
2742:
1636:, then prepend the vector with the "prefix vector"
849:
2843:
2497:
1979:
1926:
1850:
1821:
1709:
1677:
1657:
1628:
1592:
1565:
1538:
1454:
1337:
1278:
1258:
1238:
1218:
1152:
1134:, and fed to the large language models (LLM). The
1126:
1077:
1013:
949:
3494:
3237:Advances in Neural Information Processing Systems
3055:
2947:
2203:"Meta open sources an AI-powered music generator"
2149:Advances in Neural Information Processing Systems
2034:"How to write an effective GPT-3 or GPT-4 prompt"
565:Earlier work showed the effectiveness of using a
445:
396:
5233:
3388:
3082:
3030:
2994:
2943:
2941:
2820:"Better Language Models Without Massive Compute"
2739:
2027:
2025:
1958:Ziegler, Albert; Berryman, John (17 July 2023).
1735:
1359:
280:
3207:"How Each Index Works - LlamaIndex 🦙 v0.10.17"
3155:OpenAI (2023-03-27). "GPT-4 Technical Report".
3102:
3100:
2897:
1957:
3130:
1345:, and then use gradient descent to search for
440:
253:Q: {question} A: Let's think step by step.
4323:
4028:
3854:
3682:"Video generation models as world simulators"
3567:"Stable Diffusion prompt: a definitive guide"
2938:
2870:
2078:
2059:
2022:
890:, to maximize the log-likelihood on outputs.
327:
4337:
3430:
3097:
2548:"How to Turn Your Chatbot Into a Life Coach"
1326:
1293:
1072:
1036:
1008:
972:
944:
908:
882:Using gradient descent to search for prompts
870:'s AI research released Segment Anything, a
477:. There might be a discussion about this on
315:
3126:
3124:
2319:
2317:
2315:
2313:
2311:
2309:
182:(CoT) prompting is a technique that allows
4330:
4316:
4035:
4021:
3962:
3541:
3539:
3537:
562:answers for global sensemaking questions.
298:
27:Structuring text as input to generative AI
3963:Vigliarolo, Brandon (19 September 2022).
3889:
3860:
3820:
3761:
3736:
3611:
3500:
3479:
3437:
3394:
3352:
3313:
3289:
3248:
3160:
3136:
3112:
3088:
3067:
3046:
3021:
3000:
2979:
2954:
2921:
2911:
2880:
2853:
2802:
2754:
2504:
2482:
2461:
2437:
2388:
2335:
2297:
2258:
2160:
2131:
573:Using language models to generate prompts
497:Learn how and when to remove this message
4539:
3915:"Prompt injection attacks against GPT-3"
3912:
3748:
3586:
3121:
2818:Wei, Jason; Tay, Yi (29 November 2022).
2763:
2306:
2060:Gouws-Stewart, Natasha (June 16, 2023).
2031:
548:
508:
3776:
3534:
2769:
2519:
2200:
1573:, if the model first encodes the input
524:(RAG) is a two-phase process involving
372:, or some other method of tree search.
145:is the prompt used for making T0 solve
14:
5234:
3243:. Curran Associates, Inc.: 9459–9474.
3154:
2817:
2693:
2176:Heaven, Will Douglas (April 6, 2022).
2175:
1922:
1920:
1918:
75:
4311:
4016:
3913:Willison, Simon (12 September 2022).
3515:
3177:
2866:
2864:
2626:
2597:
1338:{\displaystyle \{(X^{i},Y^{i})\}_{i}}
1266:is the token-to-vector function, and
804:in the negative prompt for an image.
375:
194:and multiple steps to solve, such as
5168:Generative adversarial network (GAN)
3937:
3061:
2667:
2564:
2545:
2353:"How AI Knows Things No One Told It"
2079:Greenberg, J., Laura (31 May 2023).
1546:is the log-likelihood of outputting
1246:is a sequence of linguistic tokens,
545:Graph retrieval-augmented generation
449:
159:prompting technique was proposed by
2017:what is the fermat's little theorem
1915:
1861:
841:
720:
425:scores. Large language models like
393:Q: {question} A: False, because
256:
232:models on several tasks, achieving
174:
92:. In-context learning itself is an
24:
5272:Generative artificial intelligence
4073:Generative pre-trained transformer
2861:
2350:
861:
390:Q: {question} A: True, because
356:
25:
5283:
4042:
2032:Robinson, Reid (August 3, 2023).
657:
417:Prompting to disclose uncertainty
236:results at the time on the GSM8K
80:Prompt engineering is enabled by
5206:
5205:
5185:
4294:
4293:
3986:
3704:
3587:Heikkilä, Melissa (2022-09-16).
2520:Wei, Jason; Zhou (11 May 2022).
1146:
1117:
1109:
1101:
1066:
1062:
1045:
1041:
1029:
1002:
998:
981:
977:
965:
938:
934:
917:
913:
901:
850:Textual inversion and embeddings
807:
751:
740:
729:
704:
619:
454:
410:Article: {article} Keywords:
94:emergent property of model scale
52:can be a query such as "what is
3956:
3931:
3906:
3891:10.18653/v1/2020.emnlp-main.346
3869:
3848:
3831:10.18653/V1/2021.EMNLP-MAIN.243
3804:
3770:
3723:
3698:
3674:
3650:
3626:
3605:
3580:
3559:
3516:Monge, Jim Clyde (2022-08-25).
3509:
3488:
3467:
3424:
3382:
3340:
3321:
3297:
3273:
3257:
3223:
3199:
3171:
3148:
3076:
3009:
2988:
2966:
2837:
2811:
2789:
2770:Dickson, Ben (30 August 2022).
2719:
2713:
2687:
2661:
2646:
2620:
2591:
2558:
2539:
2513:
2491:
2470:
2449:
2425:
2400:
2376:
2344:
2284:
2271:
2219:
2194:
2169:
2140:
666:), the desired medium (such as
265:Chain-of-Symbol (CoS) Prompting
169:
5118:Recurrent neural network (RNN)
5108:Differentiable neural computer
4068:Generative adversarial network
2668:Dang, Ekta (8 February 2023).
2119:
2098:
2072:
2053:
1994:
1951:
1929:"Stable Diffusion Prompt Book"
1842:
1816:
1797:
1787:
1773:
1744:
1701:
1649:
1623:
1610:
1533:
1530:
1517:
1505:
1495:
1481:
1449:
1446:
1433:
1421:
1411:
1397:
1368:
1322:
1296:
1219:{\displaystyle LLM(X)=F(E(X))}
1213:
1210:
1204:
1198:
1189:
1183:
1121:
1097:
711:in the style of Greg Rutkowski
588:over prompts for another LLM:
522:Retrieval-augmented generation
446:Retrieval-augmented generation
402:Directional-stimulus prompting
397:Directional-stimulus prompting
335:
13:
1:
5163:Variational autoencoder (VAE)
5123:Long short-term memory (LSTM)
4390:Computational learning theory
3787:10.18653/V1/2021.ACL-LONG.353
2923:10.18653/v1/2022.acl-long.225
2694:Montti, Roger (13 May 2022).
2565:Chen, Brian X. (2023-05-25).
2546:Chen, Brian X. (2023-06-23).
1909:
1904:Social engineering (security)
783:: "round stones, round rocks"
286:Generated knowledge prompting
281:Generated knowledge prompting
5143:Convolutional neural network
2279:"Broken Neural Scaling Laws"
2201:Wiggers, Kyle (2023-06-12).
1894:by the ML model's operator.
1851:{\displaystyle {\tilde {X}}}
1710:{\displaystyle {\tilde {Z}}}
1658:{\displaystyle {\tilde {Z}}}
1153:{\displaystyle \mathbf {Y} }
103:In contrast to training and
7:
5252:Natural language processing
5138:Multilayer perceptron (MLP)
3938:Papp, Donald (2022-09-17).
3634:"Lumiere - Google Research"
3178:Eliot, Lance (2023-08-18).
1897:
612:. The question vectors are
441:Automatic prompt generation
10:
5288:
5214:Artificial neural networks
5128:Gated recurrent unit (GRU)
4354:Differentiable programming
1887:computer security exploits
1871:
1865:
1160:tokens; the gradients are
623:
584:algorithm uses one LLM to
537:, or keyword table index.
328:Complexity-based prompting
123:
67:When communicating with a
29:
5181:
5095:
5039:
4968:
4901:
4773:
4673:
4666:
4620:
4584:
4547:Artificial neural network
4527:
4403:
4370:Automatic differentiation
4343:
4291:
4250:
4225:
4207:
4179:
4138:
4131:
4050:
3638:Lumiere - Google Research
1889:carried out by getting a
802:ugly, boring, bad anatomy
582:automatic prompt engineer
362:Tree-of-thought prompting
321:Self-consistency decoding
316:Self-consistency decoding
4375:Neuromorphic engineering
4338:Differentiable computing
4103:Self-supervised learning
3612:Max Woolf (2022-11-28).
1629:{\displaystyle E(X^{i})}
120:or "learning to learn".
5148:Residual neural network
4564:Artificial Intelligence
4113:Variational autoencoder
1885:is a family of related
1539:{\displaystyle \log Pr}
874:model that can perform
304:Least-to-most prompting
299:Least-to-most prompting
54:Fermat's little theorem
3815:. pp. 3045–3059.
3781:. pp. 4582–4597.
3455:Cite journal requires
3412:Cite journal requires
3370:Cite journal requires
1852:
1823:
1711:
1679:
1659:
1630:
1594:
1567:
1540:
1456:
1339:
1280:
1260:
1240:
1220:
1154:
1138:are computed over the
1128:
1079:
1015:
951:
690:), color and texture.
554:
518:
238:mathematical reasoning
60:), an approach called
32:Command-line interface
5257:Unsupervised learning
5103:Neural Turing machine
4691:Human image synthesis
3593:MIT Technology Review
3041:with Self-Feedback".
2700:Search Engine Journal
2182:MIT Technology Review
2002:"Introducing ChatGPT"
1853:
1824:
1712:
1680:
1660:
1631:
1595:
1593:{\displaystyle X^{i}}
1568:
1566:{\displaystyle Y^{i}}
1541:
1457:
1340:
1281:
1261:
1241:
1221:
1155:
1129:
1080:
1016:
952:
682:), lighting (such as
664:bright orange poppies
652:large language models
552:
512:
200:commonsense reasoning
184:large language models
116:layers, is a form of
90:large language models
5194:Computer programming
5173:Graph neural network
4748:Text-to-video models
4726:Text-to-image models
4574:Large language model
4559:Scientific computing
4365:Statistical manifold
4360:Information geometry
4078:Large language model
2627:McAuliffe, Zachary.
2598:McAuliffe, Zachary.
2106:"GPT Best Practices"
1878:Cross-site scripting
1833:
1725:
1692:
1669:
1640:
1604:
1577:
1550:
1466:
1349:
1290:
1270:
1250:
1230:
1171:
1142:
1089:
1025:
961:
897:
771:: no negative prompt
467:confusing or unclear
429:can have accurately
350:Example refinement:
44:model. A prompt is
4540:In-context learning
4380:Pattern recognition
3969:www.theregister.com
2408:"Mesa-Optimization"
2358:Scientific American
475:clarify the section
224:, a 540B parameter
82:in-context learning
76:In-context learning
5133:Echo state network
5021:JĂĽrgen Schmidhuber
4716:Facial recognition
4711:Speech recognition
4621:Software libraries
4118:Vision transformer
4088:Prompt engineering
4004:Prompt engineering
3705:Team, PromptSora.
3211:docs.llamaindex.ai
2571:The New York Times
2552:The New York Times
1848:
1819:
1760:
1750:
1707:
1675:
1655:
1626:
1590:
1563:
1536:
1452:
1384:
1374:
1335:
1276:
1256:
1236:
1216:
1150:
1124:
1075:
1011:
947:
876:image segmentation
674:), style (such as
555:
526:document retrieval
519:
376:Maieutic prompting
344:Example critique:
38:Prompt engineering
18:Few-shot prompting
5229:
5228:
4991:Stephen Grossberg
4964:
4963:
4305:
4304:
4246:
4245:
3919:simonwillison.net
2824:ai.googleblog.com
2526:ai.googleblog.com
2085:contractnerds.com
1845:
1800:
1751:
1747:
1734:
1704:
1678:{\displaystyle F}
1652:
1508:
1424:
1375:
1371:
1358:
1279:{\displaystyle F}
1259:{\displaystyle E}
1239:{\displaystyle X}
1095:
533:, summary index,
507:
506:
499:
62:few-shot learning
16:(Redirected from
5279:
5247:Machine learning
5219:Machine learning
5209:
5208:
5189:
4944:Action selection
4934:Self-driving car
4741:Stable Diffusion
4706:Speech synthesis
4671:
4670:
4535:Machine learning
4411:Gradient descent
4332:
4325:
4318:
4309:
4308:
4297:
4296:
4199:Stable Diffusion
4136:
4135:
4037:
4030:
4023:
4014:
4013:
3990:
3989:
3979:
3978:
3976:
3975:
3960:
3954:
3953:
3951:
3950:
3935:
3929:
3928:
3926:
3925:
3910:
3904:
3903:
3893:
3873:
3867:
3866:
3864:
3852:
3846:
3845:
3824:
3808:
3802:
3801:
3774:
3768:
3767:
3765:
3752:
3746:
3745:
3740:
3727:
3721:
3720:
3718:
3717:
3702:
3696:
3695:
3693:
3692:
3678:
3672:
3671:
3669:
3668:
3654:
3648:
3647:
3645:
3644:
3630:
3624:
3623:
3621:
3620:
3609:
3603:
3602:
3600:
3599:
3584:
3578:
3577:
3575:
3574:
3563:
3557:
3556:
3554:
3553:
3543:
3532:
3531:
3529:
3528:
3513:
3507:
3506:
3504:
3492:
3486:
3485:
3483:
3471:
3465:
3464:
3458:
3453:
3451:
3443:
3441:
3428:
3422:
3421:
3415:
3410:
3408:
3400:
3398:
3386:
3380:
3379:
3373:
3368:
3366:
3358:
3356:
3344:
3338:
3337:
3325:
3319:
3318:
3317:
3301:
3295:
3294:
3293:
3277:
3271:
3270:
3261:
3255:
3254:
3252:
3227:
3221:
3220:
3218:
3217:
3203:
3197:
3196:
3191:
3190:
3175:
3169:
3166:
3164:
3152:
3146:
3145:
3140:
3128:
3119:
3118:
3116:
3104:
3095:
3094:
3092:
3080:
3074:
3073:
3071:
3059:
3053:
3052:
3050:
3037:
3028:
3027:
3025:
3013:
3007:
3006:
3004:
2992:
2986:
2985:
2983:
2970:
2964:
2963:
2958:
2945:
2936:
2935:
2925:
2915:
2895:
2886:
2885:
2884:
2868:
2859:
2858:
2857:
2841:
2835:
2834:
2832:
2830:
2815:
2809:
2808:
2806:
2793:
2787:
2786:
2784:
2782:
2767:
2761:
2760:
2758:
2746:
2737:
2736:
2734:
2732:
2717:
2711:
2710:
2708:
2706:
2691:
2685:
2684:
2682:
2680:
2665:
2659:
2658:
2650:
2644:
2643:
2641:
2639:
2624:
2618:
2617:
2612:
2610:
2595:
2589:
2588:
2586:
2585:
2562:
2556:
2555:
2543:
2537:
2536:
2534:
2532:
2517:
2511:
2510:
2508:
2495:
2489:
2488:
2486:
2474:
2468:
2467:
2465:
2453:
2447:
2446:
2441:
2429:
2423:
2422:
2417:
2415:
2404:
2398:
2397:
2392:
2380:
2374:
2373:
2367:
2365:
2351:Musser, George.
2348:
2342:
2341:
2339:
2321:
2304:
2303:
2301:
2288:
2282:
2275:
2269:
2268:
2262:
2249:
2243:
2242:
2240:
2238:
2223:
2217:
2216:
2211:
2210:
2198:
2192:
2191:
2189:
2188:
2173:
2167:
2166:
2164:
2144:
2138:
2137:
2135:
2123:
2117:
2116:
2114:
2113:
2102:
2096:
2095:
2093:
2091:
2076:
2070:
2069:
2066:masterofcode.com
2057:
2051:
2050:
2045:
2044:
2029:
2020:
2019:
2014:
2013:
1998:
1992:
1991:
1986:
1977:
1971:
1970:
1955:
1949:
1948:
1939:
1938:
1933:
1924:
1891:machine learning
1883:Prompt injection
1868:Prompt injection
1862:Prompt injection
1857:
1855:
1854:
1849:
1847:
1846:
1838:
1828:
1826:
1825:
1820:
1815:
1814:
1802:
1801:
1793:
1790:
1785:
1784:
1759:
1749:
1748:
1740:
1716:
1714:
1713:
1708:
1706:
1705:
1697:
1684:
1682:
1681:
1676:
1664:
1662:
1661:
1656:
1654:
1653:
1645:
1635:
1633:
1632:
1627:
1622:
1621:
1600:into the vector
1599:
1597:
1596:
1591:
1589:
1588:
1572:
1570:
1569:
1564:
1562:
1561:
1545:
1543:
1542:
1537:
1529:
1528:
1510:
1509:
1501:
1498:
1493:
1492:
1461:
1459:
1458:
1453:
1445:
1444:
1426:
1425:
1417:
1414:
1409:
1408:
1383:
1373:
1372:
1364:
1344:
1342:
1341:
1336:
1334:
1333:
1321:
1320:
1308:
1307:
1285:
1283:
1282:
1277:
1265:
1263:
1262:
1257:
1245:
1243:
1242:
1237:
1225:
1223:
1222:
1217:
1159:
1157:
1156:
1151:
1149:
1133:
1131:
1130:
1125:
1120:
1112:
1104:
1096:
1093:
1084:
1082:
1081:
1076:
1071:
1070:
1069:
1050:
1049:
1048:
1032:
1020:
1018:
1017:
1012:
1007:
1006:
1005:
986:
985:
984:
968:
956:
954:
953:
948:
943:
942:
941:
922:
921:
920:
904:
888:gradient descent
842:Non-text prompts
818:Models include:
794:negative prompts
763:Stable Diffusion
755:
744:
733:
721:Negative prompts
688:crepuscular rays
668:digital painting
640:Stable Diffusion
502:
495:
491:
488:
482:
458:
457:
450:
257:Other techniques
245:interpretability
234:state of the art
220:When applied to
192:logical thinking
188:train of thought
180:Chain-of-thought
175:Chain-of-thought
157:chain-of-thought
144:
86:emergent ability
46:natural language
21:
5287:
5286:
5282:
5281:
5280:
5278:
5277:
5276:
5262:2022 neologisms
5232:
5231:
5230:
5225:
5177:
5091:
5057:Google DeepMind
5035:
5001:Geoffrey Hinton
4960:
4897:
4823:Project Debater
4769:
4667:Implementations
4662:
4616:
4580:
4523:
4465:Backpropagation
4399:
4385:Tensor calculus
4339:
4336:
4306:
4301:
4287:
4263:Google DeepMind
4242:
4221:
4203:
4175:
4127:
4046:
4041:
4011:
4010:
4009:
3991:
3987:
3982:
3973:
3971:
3961:
3957:
3948:
3946:
3936:
3932:
3923:
3921:
3911:
3907:
3874:
3870:
3853:
3849:
3809:
3805:
3775:
3771:
3753:
3749:
3728:
3724:
3715:
3713:
3703:
3699:
3690:
3688:
3680:
3679:
3675:
3666:
3664:
3656:
3655:
3651:
3642:
3640:
3632:
3631:
3627:
3618:
3616:
3610:
3606:
3597:
3595:
3585:
3581:
3572:
3570:
3565:
3564:
3560:
3551:
3549:
3545:
3544:
3535:
3526:
3524:
3514:
3510:
3493:
3489:
3472:
3468:
3456:
3454:
3445:
3444:
3429:
3425:
3413:
3411:
3402:
3401:
3387:
3383:
3371:
3369:
3360:
3359:
3345:
3341:
3326:
3322:
3302:
3298:
3278:
3274:
3263:
3262:
3258:
3228:
3224:
3215:
3213:
3205:
3204:
3200:
3188:
3186:
3176:
3172:
3153:
3149:
3129:
3122:
3105:
3098:
3081:
3077:
3060:
3056:
3038:
3031:
3014:
3010:
2993:
2989:
2971:
2967:
2946:
2939:
2896:
2889:
2869:
2862:
2842:
2838:
2828:
2826:
2816:
2812:
2794:
2790:
2780:
2778:
2768:
2764:
2747:
2740:
2730:
2728:
2718:
2714:
2704:
2702:
2692:
2688:
2678:
2676:
2666:
2662:
2651:
2647:
2637:
2635:
2625:
2621:
2608:
2606:
2596:
2592:
2583:
2581:
2563:
2559:
2544:
2540:
2530:
2528:
2518:
2514:
2496:
2492:
2475:
2471:
2454:
2450:
2430:
2426:
2413:
2411:
2406:
2405:
2401:
2381:
2377:
2363:
2361:
2349:
2345:
2322:
2307:
2289:
2285:
2276:
2272:
2250:
2246:
2236:
2234:
2233:. June 12, 2023
2225:
2224:
2220:
2208:
2206:
2199:
2195:
2186:
2184:
2174:
2170:
2145:
2141:
2124:
2120:
2111:
2109:
2104:
2103:
2099:
2089:
2087:
2077:
2073:
2058:
2054:
2042:
2040:
2030:
2023:
2011:
2009:
2000:
1999:
1995:
1984:
1978:
1974:
1964:The GitHub Blog
1956:
1952:
1936:
1934:
1931:
1925:
1916:
1912:
1900:
1880:
1870:
1864:
1837:
1836:
1834:
1831:
1830:
1810:
1806:
1792:
1791:
1786:
1780:
1776:
1755:
1739:
1738:
1726:
1723:
1722:
1696:
1695:
1693:
1690:
1689:
1670:
1667:
1666:
1644:
1643:
1641:
1638:
1637:
1617:
1613:
1605:
1602:
1601:
1584:
1580:
1578:
1575:
1574:
1557:
1553:
1551:
1548:
1547:
1524:
1520:
1500:
1499:
1494:
1488:
1484:
1467:
1464:
1463:
1440:
1436:
1416:
1415:
1410:
1404:
1400:
1379:
1363:
1362:
1350:
1347:
1346:
1329:
1325:
1316:
1312:
1303:
1299:
1291:
1288:
1287:
1271:
1268:
1267:
1251:
1248:
1247:
1231:
1228:
1227:
1172:
1169:
1168:
1145:
1143:
1140:
1139:
1116:
1108:
1100:
1092:
1090:
1087:
1086:
1065:
1061:
1060:
1044:
1040:
1039:
1028:
1026:
1023:
1022:
1001:
997:
996:
980:
976:
975:
964:
962:
959:
958:
937:
933:
932:
916:
912:
911:
900:
898:
895:
894:
884:
872:computer vision
864:
862:Image prompting
852:
844:
810:
790:
789:
788:
787:
786:
777:: "green trees"
758:
757:
756:
747:
746:
745:
736:
735:
734:
723:
707:
660:
628:
622:
575:
567:knowledge graph
547:
531:vector database
503:
492:
486:
483:
472:
459:
455:
448:
443:
419:
414:
411:
399:
394:
391:
378:
359:
357:Tree-of-thought
354:
348:
338:
330:
318:
313:
301:
296:
283:
278:
259:
254:
177:
172:
142:
126:
78:
35:
28:
23:
22:
15:
12:
11:
5:
5285:
5275:
5274:
5269:
5264:
5259:
5254:
5249:
5244:
5227:
5226:
5224:
5223:
5222:
5221:
5216:
5203:
5202:
5201:
5196:
5182:
5179:
5178:
5176:
5175:
5170:
5165:
5160:
5155:
5150:
5145:
5140:
5135:
5130:
5125:
5120:
5115:
5110:
5105:
5099:
5097:
5093:
5092:
5090:
5089:
5084:
5079:
5074:
5069:
5064:
5059:
5054:
5049:
5043:
5041:
5037:
5036:
5034:
5033:
5031:Ilya Sutskever
5028:
5023:
5018:
5013:
5008:
5003:
4998:
4996:Demis Hassabis
4993:
4988:
4986:Ian Goodfellow
4983:
4978:
4972:
4970:
4966:
4965:
4962:
4961:
4959:
4958:
4953:
4952:
4951:
4941:
4936:
4931:
4926:
4921:
4916:
4911:
4905:
4903:
4899:
4898:
4896:
4895:
4890:
4885:
4880:
4875:
4870:
4865:
4860:
4855:
4850:
4845:
4840:
4835:
4830:
4825:
4820:
4815:
4814:
4813:
4803:
4798:
4793:
4788:
4783:
4777:
4775:
4771:
4770:
4768:
4767:
4762:
4761:
4760:
4755:
4745:
4744:
4743:
4738:
4733:
4723:
4718:
4713:
4708:
4703:
4698:
4693:
4688:
4683:
4677:
4675:
4668:
4664:
4663:
4661:
4660:
4655:
4650:
4645:
4640:
4635:
4630:
4624:
4622:
4618:
4617:
4615:
4614:
4609:
4604:
4599:
4594:
4588:
4586:
4582:
4581:
4579:
4578:
4577:
4576:
4569:Language model
4566:
4561:
4556:
4555:
4554:
4544:
4543:
4542:
4531:
4529:
4525:
4524:
4522:
4521:
4519:Autoregression
4516:
4511:
4510:
4509:
4499:
4497:Regularization
4494:
4493:
4492:
4487:
4482:
4472:
4467:
4462:
4460:Loss functions
4457:
4452:
4447:
4442:
4437:
4436:
4435:
4425:
4420:
4419:
4418:
4407:
4405:
4401:
4400:
4398:
4397:
4395:Inductive bias
4392:
4387:
4382:
4377:
4372:
4367:
4362:
4357:
4349:
4347:
4341:
4340:
4335:
4334:
4327:
4320:
4312:
4303:
4302:
4292:
4289:
4288:
4286:
4285:
4280:
4275:
4270:
4265:
4260:
4254:
4252:
4248:
4247:
4244:
4243:
4241:
4240:
4235:
4229:
4227:
4223:
4222:
4220:
4219:
4213:
4211:
4205:
4204:
4202:
4201:
4196:
4191:
4185:
4183:
4177:
4176:
4174:
4173:
4168:
4163:
4158:
4153:
4148:
4142:
4140:
4133:
4129:
4128:
4126:
4125:
4123:Word embedding
4120:
4115:
4110:
4105:
4100:
4095:
4090:
4085:
4083:Neural network
4080:
4075:
4070:
4065:
4060:
4054:
4052:
4048:
4047:
4040:
4039:
4032:
4025:
4017:
3992:
3985:
3984:
3983:
3981:
3980:
3955:
3930:
3905:
3868:
3847:
3803:
3769:
3747:
3722:
3697:
3673:
3649:
3625:
3604:
3579:
3558:
3533:
3508:
3487:
3466:
3457:|journal=
3423:
3414:|journal=
3381:
3372:|journal=
3339:
3320:
3296:
3272:
3256:
3222:
3198:
3170:
3147:
3120:
3096:
3075:
3054:
3029:
3008:
2987:
2965:
2937:
2887:
2860:
2836:
2810:
2788:
2762:
2738:
2720:Ray, Tiernan.
2712:
2686:
2660:
2645:
2619:
2590:
2557:
2538:
2512:
2490:
2469:
2448:
2424:
2399:
2375:
2343:
2305:
2283:
2270:
2244:
2218:
2193:
2168:
2139:
2118:
2097:
2071:
2052:
2021:
1993:
1972:
1950:
1913:
1911:
1908:
1907:
1906:
1899:
1896:
1866:Main article:
1863:
1860:
1844:
1841:
1818:
1813:
1809:
1805:
1799:
1796:
1789:
1783:
1779:
1775:
1772:
1769:
1766:
1763:
1758:
1754:
1746:
1743:
1737:
1733:
1730:
1703:
1700:
1674:
1651:
1648:
1625:
1620:
1616:
1612:
1609:
1587:
1583:
1560:
1556:
1535:
1532:
1527:
1523:
1519:
1516:
1513:
1507:
1504:
1497:
1491:
1487:
1483:
1480:
1477:
1474:
1471:
1451:
1448:
1443:
1439:
1435:
1432:
1429:
1423:
1420:
1413:
1407:
1403:
1399:
1396:
1393:
1390:
1387:
1382:
1378:
1370:
1367:
1361:
1357:
1354:
1332:
1328:
1324:
1319:
1315:
1311:
1306:
1302:
1298:
1295:
1275:
1255:
1235:
1215:
1212:
1209:
1206:
1203:
1200:
1197:
1194:
1191:
1188:
1185:
1182:
1179:
1176:
1162:backpropagated
1148:
1123:
1119:
1115:
1111:
1107:
1103:
1099:
1074:
1068:
1064:
1059:
1056:
1053:
1047:
1043:
1038:
1035:
1031:
1010:
1004:
1000:
995:
992:
989:
983:
979:
974:
971:
967:
946:
940:
936:
931:
928:
925:
919:
915:
910:
907:
903:
893:Formally, let
883:
880:
863:
860:
856:word embedding
851:
848:
843:
840:
839:
838:
832:
829:
826:
809:
806:
785:
784:
778:
772:
765:
760:
759:
750:
749:
748:
739:
738:
737:
728:
727:
726:
725:
724:
722:
719:
715:Greg Rutkowski
706:
703:
676:hyperrealistic
659:
658:Prompt formats
656:
621:
618:
606:
605:
602:
599:
596:
593:
574:
571:
546:
543:
505:
504:
462:
460:
453:
447:
444:
442:
439:
418:
415:
412:
409:
398:
395:
392:
389:
377:
374:
358:
355:
352:
346:
341:"stop" token.
337:
334:
329:
326:
317:
314:
311:
300:
297:
294:
282:
279:
275:
258:
255:
252:
226:language model
176:
173:
171:
168:
125:
122:
77:
74:
50:language model
26:
9:
6:
4:
3:
2:
5284:
5273:
5270:
5268:
5265:
5263:
5260:
5258:
5255:
5253:
5250:
5248:
5245:
5243:
5242:Deep learning
5240:
5239:
5237:
5220:
5217:
5215:
5212:
5211:
5204:
5200:
5197:
5195:
5192:
5191:
5188:
5184:
5183:
5180:
5174:
5171:
5169:
5166:
5164:
5161:
5159:
5156:
5154:
5151:
5149:
5146:
5144:
5141:
5139:
5136:
5134:
5131:
5129:
5126:
5124:
5121:
5119:
5116:
5114:
5111:
5109:
5106:
5104:
5101:
5100:
5098:
5096:Architectures
5094:
5088:
5085:
5083:
5080:
5078:
5075:
5073:
5070:
5068:
5065:
5063:
5060:
5058:
5055:
5053:
5050:
5048:
5045:
5044:
5042:
5040:Organizations
5038:
5032:
5029:
5027:
5024:
5022:
5019:
5017:
5014:
5012:
5009:
5007:
5004:
5002:
4999:
4997:
4994:
4992:
4989:
4987:
4984:
4982:
4979:
4977:
4976:Yoshua Bengio
4974:
4973:
4971:
4967:
4957:
4956:Robot control
4954:
4950:
4947:
4946:
4945:
4942:
4940:
4937:
4935:
4932:
4930:
4927:
4925:
4922:
4920:
4917:
4915:
4912:
4910:
4907:
4906:
4904:
4900:
4894:
4891:
4889:
4886:
4884:
4881:
4879:
4876:
4874:
4873:Chinchilla AI
4871:
4869:
4866:
4864:
4861:
4859:
4856:
4854:
4851:
4849:
4846:
4844:
4841:
4839:
4836:
4834:
4831:
4829:
4826:
4824:
4821:
4819:
4816:
4812:
4809:
4808:
4807:
4804:
4802:
4799:
4797:
4794:
4792:
4789:
4787:
4784:
4782:
4779:
4778:
4776:
4772:
4766:
4763:
4759:
4756:
4754:
4751:
4750:
4749:
4746:
4742:
4739:
4737:
4734:
4732:
4729:
4728:
4727:
4724:
4722:
4719:
4717:
4714:
4712:
4709:
4707:
4704:
4702:
4699:
4697:
4694:
4692:
4689:
4687:
4684:
4682:
4679:
4678:
4676:
4672:
4669:
4665:
4659:
4656:
4654:
4651:
4649:
4646:
4644:
4641:
4639:
4636:
4634:
4631:
4629:
4626:
4625:
4623:
4619:
4613:
4610:
4608:
4605:
4603:
4600:
4598:
4595:
4593:
4590:
4589:
4587:
4583:
4575:
4572:
4571:
4570:
4567:
4565:
4562:
4560:
4557:
4553:
4552:Deep learning
4550:
4549:
4548:
4545:
4541:
4538:
4537:
4536:
4533:
4532:
4530:
4526:
4520:
4517:
4515:
4512:
4508:
4505:
4504:
4503:
4500:
4498:
4495:
4491:
4488:
4486:
4483:
4481:
4478:
4477:
4476:
4473:
4471:
4468:
4466:
4463:
4461:
4458:
4456:
4453:
4451:
4448:
4446:
4443:
4441:
4440:Hallucination
4438:
4434:
4431:
4430:
4429:
4426:
4424:
4421:
4417:
4414:
4413:
4412:
4409:
4408:
4406:
4402:
4396:
4393:
4391:
4388:
4386:
4383:
4381:
4378:
4376:
4373:
4371:
4368:
4366:
4363:
4361:
4358:
4356:
4355:
4351:
4350:
4348:
4346:
4342:
4333:
4328:
4326:
4321:
4319:
4314:
4313:
4310:
4300:
4290:
4284:
4281:
4279:
4276:
4274:
4271:
4269:
4266:
4264:
4261:
4259:
4256:
4255:
4253:
4249:
4239:
4236:
4234:
4231:
4230:
4228:
4224:
4218:
4215:
4214:
4212:
4210:
4206:
4200:
4197:
4195:
4192:
4190:
4187:
4186:
4184:
4182:
4178:
4172:
4169:
4167:
4164:
4162:
4159:
4157:
4154:
4152:
4149:
4147:
4144:
4143:
4141:
4137:
4134:
4130:
4124:
4121:
4119:
4116:
4114:
4111:
4109:
4106:
4104:
4101:
4099:
4096:
4094:
4091:
4089:
4086:
4084:
4081:
4079:
4076:
4074:
4071:
4069:
4066:
4064:
4063:Deep learning
4061:
4059:
4056:
4055:
4053:
4049:
4045:
4044:Generative AI
4038:
4033:
4031:
4026:
4024:
4019:
4018:
4015:
4007:
4006:
4005:
3999:
3995:
3970:
3966:
3959:
3945:
3941:
3934:
3920:
3916:
3909:
3901:
3897:
3892:
3887:
3883:
3879:
3872:
3863:
3858:
3851:
3844:
3840:
3836:
3832:
3828:
3823:
3818:
3814:
3807:
3800:
3796:
3792:
3788:
3784:
3780:
3773:
3764:
3759:
3751:
3744:
3739:
3734:
3726:
3712:
3708:
3701:
3687:
3683:
3677:
3663:
3659:
3653:
3639:
3635:
3629:
3615:
3608:
3594:
3590:
3583:
3568:
3562:
3548:
3542:
3540:
3538:
3523:
3519:
3512:
3503:
3498:
3491:
3482:
3477:
3470:
3462:
3449:
3440:
3435:
3427:
3419:
3406:
3397:
3392:
3385:
3377:
3364:
3355:
3350:
3343:
3335:
3331:
3324:
3316:
3311:
3307:
3300:
3292:
3287:
3283:
3276:
3268:
3267:
3260:
3251:
3246:
3242:
3238:
3234:
3226:
3212:
3208:
3202:
3195:
3185:
3181:
3174:
3168:
3163:
3158:
3151:
3144:
3139:
3134:
3127:
3125:
3115:
3110:
3103:
3101:
3091:
3086:
3079:
3070:
3065:
3058:
3049:
3044:
3036:
3034:
3024:
3019:
3012:
3003:
2998:
2991:
2982:
2977:
2969:
2962:
2957:
2952:
2944:
2942:
2933:
2929:
2924:
2919:
2914:
2909:
2905:
2901:
2894:
2892:
2883:
2878:
2874:
2867:
2865:
2856:
2851:
2847:
2840:
2825:
2821:
2814:
2805:
2800:
2792:
2777:
2773:
2766:
2757:
2752:
2745:
2743:
2727:
2723:
2716:
2701:
2697:
2690:
2675:
2671:
2664:
2656:
2649:
2634:
2630:
2623:
2616:
2605:
2601:
2594:
2580:
2576:
2572:
2568:
2561:
2553:
2549:
2542:
2527:
2523:
2516:
2507:
2502:
2494:
2485:
2480:
2473:
2464:
2459:
2452:
2445:
2440:
2435:
2428:
2421:
2410:. 31 May 2019
2409:
2403:
2396:
2391:
2386:
2379:
2372:
2360:
2359:
2354:
2347:
2338:
2333:
2329:
2328:
2320:
2318:
2316:
2314:
2312:
2310:
2300:
2295:
2287:
2280:
2274:
2267:
2261:
2256:
2248:
2232:
2228:
2222:
2215:
2204:
2197:
2183:
2179:
2172:
2163:
2158:
2155:: 1877–1901.
2154:
2150:
2143:
2134:
2129:
2122:
2107:
2101:
2086:
2082:
2075:
2067:
2063:
2056:
2049:
2039:
2035:
2028:
2026:
2018:
2007:
2003:
1997:
1990:
1983:
1976:
1969:
1965:
1961:
1954:
1947:
1945:
1944:text-to-image
1930:
1923:
1921:
1919:
1914:
1905:
1902:
1901:
1895:
1892:
1888:
1884:
1879:
1875:
1874:SQL injection
1869:
1859:
1839:
1811:
1807:
1803:
1794:
1781:
1777:
1770:
1767:
1764:
1761:
1756:
1752:
1741:
1731:
1728:
1718:
1698:
1686:
1672:
1665:, then apply
1646:
1618:
1614:
1607:
1585:
1581:
1558:
1554:
1525:
1521:
1514:
1511:
1502:
1489:
1485:
1478:
1475:
1472:
1469:
1441:
1437:
1430:
1427:
1418:
1405:
1401:
1394:
1391:
1388:
1385:
1380:
1376:
1365:
1355:
1352:
1330:
1317:
1313:
1309:
1304:
1300:
1273:
1253:
1233:
1207:
1201:
1195:
1192:
1186:
1180:
1177:
1174:
1165:
1163:
1137:
1113:
1105:
1057:
1054:
1051:
1033:
993:
990:
987:
969:
929:
926:
923:
905:
891:
889:
879:
877:
873:
869:
859:
857:
847:
836:
835:OpenAI's Sora
833:
830:
827:
824:
821:
820:
819:
816:
814:
813:Text-to-video
808:Text-to-video
805:
803:
799:
795:
782:
779:
776:
773:
770:
767:
766:
764:
754:
743:
732:
718:
716:
712:
705:Artist styles
702:
699:
696:
691:
689:
685:
681:
677:
673:
669:
665:
655:
653:
649:
645:
641:
637:
633:
632:text-to-image
627:
620:Text-to-image
617:
615:
611:
603:
600:
597:
594:
591:
590:
589:
587:
583:
578:
570:
568:
563:
559:
551:
542:
538:
536:
532:
527:
523:
516:
511:
501:
498:
490:
480:
479:the talk page
476:
470:
468:
463:This section
461:
452:
451:
438:
434:
432:
428:
424:
408:
405:
403:
388:
385:
382:
373:
371:
367:
366:breadth-first
363:
351:
345:
342:
333:
325:
322:
310:
307:
305:
293:
290:
287:
274:
271:
267:
266:
262:
251:
248:
246:
242:
239:
235:
231:
227:
223:
218:
216:
212:
207:
203:
201:
197:
193:
189:
185:
181:
167:
164:
163:researchers.
162:
158:
153:
150:
148:
139:
134:
131:
121:
119:
118:meta-learning
115:
111:
106:
101:
99:
95:
91:
87:
83:
73:
70:
69:text-to-image
65:
63:
59:
55:
51:
47:
43:
42:generative AI
39:
33:
19:
5062:Hugging Face
5026:David Silver
4674:Audio–visual
4528:Applications
4507:Augmentation
4352:
4268:Hugging Face
4087:
4002:
4001:
4000:profile for
3997:
3972:. Retrieved
3968:
3958:
3947:. Retrieved
3943:
3933:
3922:. Retrieved
3918:
3908:
3881:
3871:
3850:
3842:
3812:
3806:
3798:
3778:
3772:
3750:
3742:
3725:
3714:. Retrieved
3710:
3700:
3689:. Retrieved
3685:
3676:
3665:. Retrieved
3661:
3652:
3641:. Retrieved
3637:
3628:
3617:. Retrieved
3607:
3596:. Retrieved
3592:
3582:
3571:. Retrieved
3569:. 2023-05-14
3561:
3550:. Retrieved
3525:. Retrieved
3522:MLearning.ai
3521:
3511:
3490:
3469:
3448:cite journal
3426:
3405:cite journal
3384:
3363:cite journal
3342:
3333:
3323:
3305:
3299:
3281:
3275:
3265:
3259:
3240:
3236:
3225:
3214:. Retrieved
3210:
3201:
3193:
3187:. Retrieved
3173:
3167:
3150:
3142:
3078:
3057:
3011:
2990:
2968:
2960:
2903:
2872:
2845:
2839:
2827:. Retrieved
2823:
2813:
2791:
2779:. Retrieved
2775:
2765:
2729:. Retrieved
2725:
2715:
2703:. Retrieved
2699:
2689:
2677:. Retrieved
2673:
2663:
2648:
2636:. Retrieved
2632:
2622:
2614:
2607:. Retrieved
2603:
2593:
2582:. Retrieved
2570:
2560:
2551:
2541:
2529:. Retrieved
2525:
2515:
2493:
2472:
2451:
2443:
2427:
2419:
2412:. Retrieved
2402:
2394:
2378:
2369:
2362:. Retrieved
2356:
2346:
2326:
2286:
2273:
2264:
2247:
2235:. Retrieved
2230:
2221:
2213:
2207:. Retrieved
2205:. TechCrunch
2196:
2185:. Retrieved
2181:
2171:
2152:
2148:
2142:
2121:
2110:. Retrieved
2100:
2088:. Retrieved
2084:
2074:
2065:
2055:
2047:
2041:. Retrieved
2037:
2016:
2010:. Retrieved
2008:. 2022-11-30
2005:
1996:
1988:
1975:
1967:
1963:
1953:
1943:
1941:
1935:. Retrieved
1882:
1881:
1719:
1687:
1462:. In words,
1166:
892:
885:
865:
853:
845:
823:Runway Gen-2
817:
811:
801:
797:
793:
791:
780:
774:
768:
710:
708:
700:
692:
687:
684:rim lighting
683:
679:
675:
671:
667:
663:
661:
634:models like
629:
607:
581:
579:
576:
564:
560:
556:
539:
520:
493:
484:
473:Please help
464:
435:
420:
406:
401:
400:
386:
379:
361:
360:
349:
343:
339:
331:
320:
319:
308:
303:
302:
291:
285:
284:
272:
268:
264:
263:
260:
249:
219:
214:
210:
208:
204:
179:
178:
170:Text-to-text
165:
156:
155:In 2022 the
154:
151:
135:
127:
102:
81:
79:
66:
61:
57:
37:
36:
5267:Linguistics
5210:Categories
5158:Autoencoder
5113:Transformer
4981:Alex Graves
4929:OpenAI Five
4833:IBM Watsonx
4455:Convolution
4433:Overfitting
4108:Transformer
4058:Autoencoder
3662:ai.meta.com
2776:VentureBeat
2674:VentureBeat
2006:OpenAI Blog
672:photography
586:beam search
336:Self-refine
202:questions.
114:transformer
105:fine-tuning
5236:Categories
5199:Technology
5052:EleutherAI
5011:Fei-Fei Li
5006:Yann LeCun
4919:Q-learning
4902:Decisional
4828:IBM Watson
4736:Midjourney
4628:TensorFlow
4475:Activation
4428:Regression
4423:Clustering
4283:Mistral AI
4194:Midjourney
3974:2023-02-09
3949:2023-02-09
3924:2023-02-09
3862:2302.11521
3822:2104.08691
3763:2304.02643
3738:2208.01618
3716:2024-02-25
3711:PromptSora
3691:2024-02-25
3686:openai.com
3667:2024-02-25
3643:2024-02-25
3619:2023-08-14
3598:2023-08-14
3573:2023-08-14
3552:2023-08-14
3527:2022-08-31
3502:2210.03493
3481:2211.01910
3439:2309.08532
3396:2305.03495
3354:2309.16797
3315:2311.07509
3291:2404.16130
3250:2005.11401
3216:2024-04-08
3189:2024-08-31
3162:2303.08774
3138:2302.11520
3114:2205.11822
3090:2305.10601
3069:2305.08291
3048:2303.17651
3023:2210.00720
3002:2302.12246
2981:2203.11171
2956:2205.10625
2913:2110.08387
2882:2305.10276
2855:2402.07927
2804:2210.11416
2756:2205.11916
2584:2023-08-16
2506:2202.01279
2484:2110.08207
2463:1806.08730
2439:2208.01066
2390:2212.07677
2337:2201.11903
2299:2206.07682
2260:2206.07682
2209:2023-08-15
2187:2023-08-14
2162:2005.14165
2133:2208.01066
2112:2023-08-16
2043:2023-08-14
2012:2023-08-16
1987:. OpenAI.
1937:2023-08-07
1910:References
1872:See also:
698:pencils".
695:Midjourney
644:Midjourney
624:See also:
535:tree index
515:embeddings
469:to readers
431:calibrated
423:likelihood
230:fine-tuned
196:arithmetic
147:entailment
96:, meaning
5082:MIT CSAIL
5047:Anthropic
5016:Andrew Ng
4914:AlphaZero
4758:VideoPoet
4721:AlphaFold
4658:MindSpore
4612:SpiNNaker
4607:Memristor
4514:Diffusion
4490:Rectifier
4470:Batchnorm
4450:Attention
4445:Adversary
4258:Anthropic
4251:Companies
3900:226222232
3839:233296808
3795:230433941
3547:"Prompts"
2932:239016123
2579:0362-4331
2371:learning.
2292:Models".
2253:Models".
1843:~
1804:∗
1798:~
1765:
1753:∑
1745:~
1732:
1702:~
1650:~
1512:∗
1506:~
1473:
1428:∗
1422:~
1389:
1377:∑
1369:~
1356:
1055:…
991:…
927:…
866:In 2023,
630:In 2022,
614:clustered
487:July 2024
407:Example:
387:Example:
309:Example:
292:Example:
273:Example:
250:Example:
241:benchmark
215:zero-shot
5190:Portals
4949:Auto-GPT
4781:Word2vec
4585:Hardware
4502:Datasets
4404:Concepts
4299:Category
4051:Concepts
3944:Hackaday
2829:10 March
2781:10 March
2731:10 March
2705:10 March
2679:10 March
2638:10 March
2609:10 March
2531:10 March
2237:June 12,
2231:claid.ai
2108:. OpenAI
1898:See also
1226:, where
636:DALL-E 2
381:Maieutic
211:few-shot
5072:Meta AI
4909:AlphaGo
4893:PanGu-ÎŁ
4863:ChatGPT
4838:Granite
4786:Seq2seq
4765:Whisper
4686:WaveNet
4681:AlexNet
4653:Flux.jl
4633:PyTorch
4485:Sigmoid
4480:Softmax
4345:General
4278:Meta AI
4233:Suno AI
3994:Scholia
2090:24 July
680:pop-art
465:may be
124:History
110:dataset
5087:Huawei
5067:OpenAI
4969:People
4939:MuZero
4801:Gemini
4796:Claude
4731:DALL-E
4643:Theano
4273:OpenAI
4226:Musics
4209:Videos
4189:DALL-E
4181:Images
4151:Gemini
4146:Claude
4132:Models
3996:has a
3898:
3837:
3793:
3269:, 2024
3184:Forbes
2930:
2577:
2414:17 May
2364:17 May
2038:Zapier
1829:where
1136:losses
1094:concat
781:Bottom
775:Centre
648:AI art
642:, and
161:Google
98:breaks
5153:Mamba
4924:SARSA
4888:LLaMA
4883:BLOOM
4868:GPT-J
4858:GPT-4
4853:GPT-3
4848:GPT-2
4843:GPT-1
4806:LaMDA
4638:Keras
4171:LLaMA
4166:GPT-4
4161:GPT-3
4156:GPT-2
3998:topic
3896:S2CID
3857:arXiv
3835:S2CID
3817:arXiv
3791:S2CID
3758:arXiv
3733:arXiv
3497:arXiv
3476:arXiv
3434:arXiv
3391:arXiv
3349:arXiv
3334:arXiv
3310:arXiv
3286:arXiv
3245:arXiv
3157:arXiv
3133:arXiv
3109:arXiv
3085:arXiv
3064:arXiv
3043:arXiv
3018:arXiv
2997:arXiv
2976:arXiv
2951:arXiv
2928:S2CID
2908:arXiv
2877:arXiv
2850:arXiv
2799:arXiv
2751:arXiv
2726:ZDNET
2501:arXiv
2479:arXiv
2458:arXiv
2434:arXiv
2385:arXiv
2332:arXiv
2294:arXiv
2255:arXiv
2157:arXiv
2128:arXiv
1985:(PDF)
1932:(PDF)
427:GPT-4
5077:Mila
4878:PaLM
4811:Bard
4791:BERT
4774:Text
4753:Sora
4238:Udio
4217:Sora
4139:Text
4098:RLHF
3461:help
3418:help
3376:help
2831:2023
2783:2023
2733:2023
2707:2023
2681:2023
2640:2023
2633:CNET
2611:2023
2604:CNET
2575:ISSN
2533:2023
2416:2023
2366:2023
2239:2023
2092:2023
1876:and
1021:and
868:Meta
693:The
610:BERT
580:The
370:beam
222:PaLM
4818:NMT
4701:OCR
4696:HWR
4648:JAX
4602:VPU
4597:TPU
4592:IPU
4416:SGD
4093:RAG
3886:doi
3827:doi
3783:doi
2918:doi
1762:log
1736:max
1729:arg
1470:log
1386:log
1360:max
1353:arg
798:not
769:Top
686:or
678:or
670:or
198:or
138:NLP
130:NLP
88:of
58:dog
5238::
3967:.
3942:.
3917:.
3894:.
3880:.
3841:.
3833:.
3825:.
3797:.
3789:.
3741:.
3709:.
3684:.
3660:.
3636:.
3591:.
3536:^
3520:.
3452::
3450:}}
3446:{{
3409::
3407:}}
3403:{{
3367::
3365:}}
3361:{{
3332:.
3308:,
3284:,
3241:33
3239:.
3235:.
3209:.
3192:.
3182:.
3141:.
3123:^
3099:^
3032:^
2959:.
2940:^
2926:.
2916:.
2902:.
2890:^
2875:,
2863:^
2848:,
2822:.
2774:.
2741:^
2724:.
2698:.
2672:.
2631:.
2613:.
2602:.
2573:.
2569:.
2550:.
2524:.
2442:.
2418:.
2393:.
2368:.
2355:.
2308:^
2263:.
2229:.
2212:.
2180:.
2153:33
2151:.
2083:.
2064:.
2046:.
2036:.
2024:^
2015:.
2004:.
1966:.
1962:.
1940:.
1917:^
1685:.
717:.
638:,
368:,
247:.
149:.
64:.
4331:e
4324:t
4317:v
4036:e
4029:t
4022:v
4008:.
3977:.
3952:.
3927:.
3902:.
3888::
3865:.
3859::
3829::
3819::
3785::
3766:.
3760::
3735::
3719:.
3694:.
3670:.
3646:.
3622:.
3601:.
3576:.
3555:.
3530:.
3505:.
3499::
3484:.
3478::
3463:)
3459:(
3442:.
3436::
3420:)
3416:(
3399:.
3393::
3378:)
3374:(
3357:.
3351::
3336:.
3312::
3288::
3253:.
3247::
3219:.
3165:.
3159::
3135::
3117:.
3111::
3093:.
3087::
3072:.
3066::
3051:.
3045::
3026:.
3020::
3005:.
2999::
2984:.
2978::
2953::
2934:.
2920::
2910::
2879::
2852::
2833:.
2807:.
2801::
2785:.
2759:.
2753::
2735:.
2709:.
2683:.
2657:.
2642:.
2587:.
2554:.
2535:.
2509:.
2503::
2487:.
2481::
2466:.
2460::
2436::
2387::
2340:.
2334::
2302:.
2296::
2257::
2241:.
2190:.
2165:.
2159::
2136:.
2130::
2115:.
2094:.
2068:.
1840:X
1817:]
1812:i
1808:X
1795:X
1788:|
1782:i
1778:Y
1774:[
1771:r
1768:P
1757:i
1742:X
1699:Z
1673:F
1647:Z
1624:)
1619:i
1615:X
1611:(
1608:E
1586:i
1582:X
1559:i
1555:Y
1534:]
1531:)
1526:i
1522:X
1518:(
1515:E
1503:Z
1496:|
1490:i
1486:Y
1482:[
1479:r
1476:P
1450:]
1447:)
1442:i
1438:X
1434:(
1431:E
1419:Z
1412:|
1406:i
1402:Y
1398:[
1395:r
1392:P
1381:i
1366:Z
1331:i
1327:}
1323:)
1318:i
1314:Y
1310:,
1305:i
1301:X
1297:(
1294:{
1274:F
1254:E
1234:X
1214:)
1211:)
1208:X
1205:(
1202:E
1199:(
1196:F
1193:=
1190:)
1187:X
1184:(
1181:M
1178:L
1175:L
1147:Y
1122:)
1118:Y
1114:;
1110:X
1106:;
1102:E
1098:(
1073:}
1067:n
1063:y
1058:,
1052:,
1046:1
1042:y
1037:{
1034:=
1030:Y
1009:}
1003:m
999:x
994:,
988:,
982:1
978:x
973:{
970:=
966:X
945:}
939:k
935:e
930:,
924:,
918:1
914:e
909:{
906:=
902:E
500:)
494:(
489:)
485:(
481:.
34:.
20:)
Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.