This story on HackerNoon has a decentralized backup on Sia.
Transaction ID: iZKYMoo-BIVe6w2Uu3wnw0LvIK-KmGkJoamLXu_6Too
Cover

Input Format for Fine-tuning GPT-3.5 for Praise Evaluation

Written by @highlighter | Published on 2025/5/31

TL;DR
Explore the structured input format (System, User, Assistant roles) used to fine-tune GPT-3.5 models for identifying effort- and outcome-based praise in tutor responses via JSON output.

Abstract and 1 Introduction

2. Background

2.1 Effective Tutoring Practice

2.2 Feedback for Tutor Training

2.3 Sequence Labeling for Feedback Generation

2.4 Large Language Models in Education

3. Method

3.1 Dataset and 3.2 Sequence Labeling

3.3 GPT Facilitated Sequence Labeling

3.4 Metrics

4. Results

4.1 Results on RQ1

4.2 Results on RQ2

5. Discussion

6. Limitation and Future Works

7. Conclusion

8. Acknowledgments

9. References

APPENDIX

A. Lesson Principles

B. Input for Fine-Tunning GPT-3.5

C. Scatter Matric of the Correlation on the Outcome-based Praise

D. Detailed Results of Fine-Tuned GPT-3.5 Model's Performance

B. INPUT FOR FINE-TUNING GPT-3.5

Table 6: Input for fine-tuning GPT 3.5 model

Note: Praise Type and Content: This part simulates an interactive environment where the model plays the role of a response evaluator. The conversation flow is designed to mimic a real-world interaction, with system and user roles alternately providing context, instruction, and input (the tutor response) for processing.

This paper is available on arxiv under CC BY 4.0 DEED license.

Authors:

(1) Jionghao Lin, Carnegie Mellon University (jionghal@cs.cmu.edu);

(2) Eason Chen, Carnegie Mellon University (easonc13@cmu.edu);

(3) Zeifei Han, University of Toronto (feifei.han@mail.utoronto.ca);

(4) Ashish Gurung, Carnegie Mellon University (agurung@andrew.cmu.edu);

(5) Danielle R. Thomas, Carnegie Mellon University (drthomas@cmu.edu);

(6) Wei Tan, Monash University (wei.tan2@monash.edu);

(7) Ngoc Dang Nguyen, Monash University (dan.nguyen2@monash.edu);

(8) Kenneth R. Koedinger, Carnegie Mellon University (koedinger@cmu.edu).

[story continues]


Written by
@highlighter
Shining light on key points, making the vital stand out, guiding eyes to what matters most.

Topics and
tags
gpt-3.5-fine-tuning|praise-evaluation|tutor-responses|json-output|prompt-engineering|conversational-ai|model-training-data|educational-ai
This story on HackerNoon has a decentralized backup on Sia.
Transaction ID: iZKYMoo-BIVe6w2Uu3wnw0LvIK-KmGkJoamLXu_6Too