Congratulations on the First Prize!!
General: Great setup, good introduction, good splitting into sections that clearly represent your project.
Fazl: Very good depth of prior work. The definition of truthfulness could benefit from TruthfulQA.
Alignment: Is about truthfulness, a quite important factor in alignment. Shows clear ways that pre-prompting for helpfulness (a desired trait) makes the model less truthful.
AI psychology: A multi-factor experimental design that very well represents how we might have to look at AI models in the future. Very interesting results as well. Represents naturalistic chatbot interactions (e.g. Alexa) very well.
Novelty: This has not been seen before, as far as the judges know.
Generality: Seems like a generalizable effect that have quite a few percentage point impacts on model truthfulness. Replicability: Has a Github without documentation but seems simple to run based on the experimental design. Parameters chosen are clear.
Leave a comment
Log in with itch.io to leave a comment.