Replies: 2 comments
-
Great Idea! Should be feasible with the current state of pipelines, I'd love to see a PR for this! |
Beta Was this translation helpful? Give feedback.
-
I’m trying to come up with a proper chat structure. My idea is that the user is constantly exposed to new text snippets. For example, when the user sends the text "/next" (or another command), the Now the user can validate the output with 👍 or 👎. At any point, the user can export the chat (or inspect the open-webui dataset) and parse it, expecting the following message history structure:
I think that this implementation is not very good. But:
I’m wondering if there is a way to:
I know that I’m overengineering this simple process, but I'd like it to be scalable to other types of validation and compatible with the OpenAI chat history specification. |
Beta Was this translation helpful? Give feedback.
-
I am interested in using the open-webui interface to validate the outputs of a large language model (LLM).
I have a collection of text snippets that I would like to classify using an LLM based on a system prompt (e.g., "Is the following text a positive or negative review?"). The goal is to build a small dataset to evaluate classification capabilities. This dataset can later be used for testing classification with new system prompts or for training a text classifier.
I aim to automate the process as much as possible, automatically feeding new text to the LLM so that the human evaluator only needs to press 👍 or 👎 in the interface.
I understand that this method of building a dataset for binary text classification is somewhat convoluted (the simpler way is to manually open text snippets with a text editor and annotate them in a table-like structure). However, in the future, I plan to extend the LLM output evaluation beyond binary classification (e.g., using annotation reasons) and to have multiple contributors for the annotations.
Is this a use case for pipelines?
Beta Was this translation helpful? Give feedback.
All reactions