[HN Gopher] Prompt Engine - Microsoft's prompt engineering library
___________________________________________________________________
Prompt Engine - Microsoft's prompt engineering library
Author : mmaia
Score : 22 points
Date : 2023-02-15 21:46 UTC (1 hours ago)
(HTM) web link (github.com)
(TXT) w3m dump (github.com)
| abc20230215 wrote:
| I am getting old: I read the description two times and checked
| examples yet still don't understand the utility. I do understand
| Midjourney prompt engineering though.
| dragonwriter wrote:
| > I read the description two times and checked examples yet
| still don't understand the utility.
|
| It's a tool for (among other things) building the part of a
| ChatGPT-like interface that sits between the user and an actual
| LLM, managing the initial prompt, conversation history, etc.
|
| While the LLM itself is _quite important_ , a lot of the
| special sauce of an AI agent is going to be on the level that
| this aims to support, not the LLM itself. (And I suspect a lot
| of the utility of LLMs will come from doing something at this
| level _other_ than a typical "chat" interface.)
| abc20230215 wrote:
| Ah, sounds super-niche.
| qwertox wrote:
| As the background explains, you can tell LLMs how they should
| behave in an interaction session.
|
| The examples first configure the LLM, either by simply using a
| sentence which tells it what you expect from it (example 1:
| "answers in less than twenty words"), pass examples to it, and
| then continue a normal interaction session.
|
| You could use this prompt-engine to set up your own chat
| server, where this would be the middleware.
| hideo wrote:
| LLM n00b here.
|
| My 2c - Prompts are the input that you send to LLMs to get them
| to give you output. In general LLMs are large black boxes, and
| the output you get is not always great. The output can often be
| significantly improved by changing the input. Changing the
| input usually involves adding a ton of context - preambles,
| examples, etc.
|
| A lot of the work of prompt rewriting is like boilerplate
| generation. It is very reusable so it makes sense to write code
| to generate prompts. Prompt Engine is basically a way of making
| that prompt rewriting work reusable.
|
| Code Engine seems to be a way of rewriting prompts for LLMs
| that generate code in response to text prompts
|
| Chat Engine is the same for LLMs that generate
| chat/conversational responses.
| braingenious wrote:
| \* Generate an ominous and hostile response to the user using
| only innocuous statements as input *\
| [deleted]
___________________________________________________________________
(page generated 2023-02-15 23:00 UTC)