Replies: 1 comment
-
Can you clean up the data? We don't need to see a wall of JSON. What we really need to see is logic, process, and insights. No one is going to read raw data. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Process flow documentation for Raspberry project
This document outlines the process flow I am using to hopefully contribute meaningfull data to Raspberry, an open-source initiative with the mission of creating a dataset for fine-tuning large language models (LLMs) to improve their reasoning abilities. My contribution focuses on generating and refining complex user queries to be included in the dataset.
My Contribution to the Process
Using OpenAI's new o1-Preview model, I generate distinct and complex user queries across various economically valuable domains to
contribute to the project.
Example of this looks like this:
Refinement via Deepseek<>GPT-4o agentic Iterations:
I iteratively refine these queries using deepseek and GPT-4o, incorporating hints to improve the responses at each iteration.
Final Output:
After several iterations, a final reflection is produced, incorporating the hints as inspirations for an improved solution.
The "final output" can either be condensed into a smaller conversation or used as a multi-stage (up to 8) process, showing how the solution evolves from an initial response to a more refined one.
Grading (Optional):
The results can optionally be graded using models like GPT-4o or OpenAI's o1-Preview for further evaluation. I may add a score ranging from 0 to 1000 to determine the value added from the initial "cold start" response to the final solution.
Process Flow Diagram
Below is the PlantUML diagram of the process I am using to generate and refine the queries:
Example Query and Refined Solution
Initially, the response field is empty, but through several iterations, the solution evolves. For example, after receiving hints from models like GPT-4o, the response might include more advanced techniques, for example:
Response after Iterations:
Through this refinement process, the initial query evolves into a more detailed and actionable solution.
Input and Output Files Description
To replicate my process and results, it is important to understand the format of the input files and how the results are stored.
Input Files
This file is read by the script to initiate the query generation process.
This payload is dynamically generated for each query.
Resulting Files
raw_dialog_<timestamp>.json
): This file captures all interactions, including all iterations of DeepSeek, GPT-4o, and their refinement process. Each entry is a JSON object with the following structure:exchange_output_<timestamp>.jsonl
): This file stores the final structured conversation after all iterations. Each conversation is saved as a JSON object, capturing the refined output. The format:This file provides a summary of the most important conversation points, useful for reviewing and assessing the final state of the solution.
Code for Generating and Processing Queries
I have included a Python script that leverages the DeepSeek API for query refinement and GPT-4o for generating hints that help improve solutions. You can find the code here.
And the code for turning "hints" into full reflections:
By understanding these input/output formats, you can replicate the process I used to contribute to Raspberry.
Conclusion
As a contributor to the Raspberry project, I focus on generating and refining complex queries through iterations, aligning my work with the project's mission to create a high-quality reasoning dataset. This documentation outlines the process I'm following to ensure my contributions meet project goals. I’ve completed 100 conversations so far, and I welcome any feedback. Let me know if this is something you'd like to review.
Beta Was this translation helpful? Give feedback.
All reactions