Skip to main content
To evaluate the feedback, a basic technical understanding of large language models, the Genow application, and existing and future features and limitations of Genow is necessary. The feedback received can be evaluated particularly well if it is of high quality.
We would recommend encouraging users to provide continuous, comprehensive feedback via the feedback function in order to further develop the use case in terms of quality. This feedback can then be viewed via the admin panel. For a structured analysis of the feedback, we recommend processing it via CSV. To do this, you can use the ‘Download Feedback’ function in the Admin Panel.

Formation of Clusters

Clusters are useful for evaluating feedback. Similar feedback is grouped together. The following table provides an overview of common clusters and how to address them. First, it’s a good idea to cluster negative and positive feedback. Positive user feedback describes how well users have adopted the technology and how well they have been trained. Negative feedback provides insight into potential for improvement. Please note that the following table refers to the evaluation of negative feedback:
Cluster NameDescriptionResponsibilityTakeaway
CorrectGenow’s response was actually correct based on the connected data and the technical capabilities of language models, but feedback was provided.Use CaseUser training on expectation management or on the capabilities and limitations of language models may be necessary.
Information missingThe information is not available to the language model, but should be added.Use CaseUpdate information
Missing Feature / will be fixed by updatesThe issue will be corrected by Genow (Roadmap!)Use Case / GenowUpdate informationUse Case contacts Genow to submit a feature request. If there is a high degree of urgency, a project contract can be concluded for implementation. Sometimes the feature is already on Genow’s roadmap or it is a known problem that will be resolved with a short-term release.
Information inaccurateInformation in existing documents inaccurate / insufficiently formulated, which has a negative impact on the quality of the response.Use CaseUpdate document, make sure the guidelines and recommendations for your data was matched (see Data Maintanace article).
Imprecise question / interaction in need of improvementUser input is sometimes crucial when complex information is to be extracted from documents.Use CaseUser training necessary
Information not output correctlyInformation not extracted correctly, although it is available.Use Case / GenowDepending on the individual case. The files and their names should always be checked. In complex cases, pipeline optimisation is necessary. This can be carried out by Genow.
Available feature or optimisation of an existing feature would provide improvementIssue or process during use could be fixed or improved by an available feature.Use Case / GenowDepending on the individual case.
BugError message, demonstrably incorrect behaviour by Genow; ideally traceable.GenowError message, time of the error message and description as well as screenshot to [email protected] for error correction.
In individual cases there are further clusters …

Review Feedback

All feedback should be reviewed. For this purpose, either a spreadsheet (e.g., an Excel file) or the existing feedback dashboard in the Genow admin panel can be used. For the evaluation of feedback, the query, Genow’s response, and the user comment should be recorded. Subsequently, the question should be assigned to one, or at most two, clusters. Additionally, a comment can be added to justify the cluster assignment. A possible structure results from the following example table:
QuestionAnswer from GenowGenow thread historyTarget answer / comment by test userPositive / negativeComment ClusterCluster 1Cluster 2
The user request sent to Genow that was evaluatedGenow’s answer to the graded questionHistory of previous communication before the graded responseComment by the evaluator; contains information about what answer would have been expected or whether information is missingComment from test evaluation

Create a summarized Evaluation

The clusters created and the frequency with which they occur should be placed in an overall context with the total number of feedbacks, positive and negative feedbacks and the total number of requests. The summarized evaluation can be shown in the following table:
EvaluationAmountTakeawayOptimisation steps
Total number of requests (see Analytics Dashboard in the Admin Panel)I.e. used more or less than expectedI.e.: motivate user to increase usage
Positive FeedbacksI.e. comparison to total number of requestsI.e: Go live or next project phase possible?
Negative FeedbackI.e. comparison to total number of requestsI.e.: Optimisation necessary?
Responsibility Use CaseSum of clusters that can be solved internally by the use case (e.g. updating information). Indicator of how much work still needs to be done internally.
Responsibility GenowSum of Clusters that will be solved by Genow (i.e. Bugs)
Optimisation necesseryClusters that can be improved through optimizations in the data pipeline or through updates and features.In consultation with Genow - depending on the individual case. E.g. glossaries, fallback strategy, individual base prompt, metadata, …
Cluster 1
Cluster 2…
The table can be supplemented with any other relevant clusters. For the sake of clarity, a graphical representation is recommended at the end.