Project Update - Big News! #45
daveshap
announced in
Announcements
Replies: 3 comments
-
This is great news |
Beta Was this translation helpful? Give feedback.
0 replies
-
3 months on, no updates. Is this still happening? Are you doing something behind closed doors? Making any progress? Or why are you not sharing the results? |
Beta Was this translation helpful? Give feedback.
0 replies
-
@mindplay-dk There's work going on, we haven't been tracking it in this repo -- not behind closed doors, just in different public repos. Paper -> CoT pipeline: https://github.com/thehunmonkgroup/raspberry-paper-to-cot-pipeline |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Executive Summary
On Wednesday, 2024-09-18, the Raspberry team had a great meeting with folks from LAION (https://laion.ai/) to discuss potential collaboration. We identified two overarching areas of alignment between our projects.
Top Level Topics
Quick recap
The team discussed various topics including reinforcement learning, different types of intelligence, and the potential of combining code interpreters, math interpreters, and external grounding for AI models. They explored using multiple models to generate and refine data, with a focus on developing a multi-agent framework for scalable data generation and verification. The meeting also covered team introductions, ongoing projects, and potential approaches to improve AI capabilities using open-source models and automated data synthesis pipelines.
Topics Discussed
Reinforcement Learning and AI Models: The team explored the potential of combining code interpreters, math interpreters, and external grounding for AI models. They discussed how these elements could be integrated to create a more comprehensive and consistent model of reality.
Multi-Agent Framework: The team proposed developing a multi-agent framework for scalable data generation and verification. This approach would involve using multiple models to generate and refine data, with minimal human review required.
Open-Source Data Synthesis Pipeline: A key project discussed was the development of an automated data synthesis pipeline using open-source resources. This pipeline would aim to generate high-quality datasets more efficiently and cost-effectively than traditional methods.
Model Fine-Tuning Strategy: The team considered using an open-source model like Llama 3.1 as a base for fine-tuning. The strategy involved using copies of the model to generate diverse trajectories, evaluate them, and iteratively refine the dataset.
Data Validation Techniques: Methods for validating synthesized chains of thought were discussed. These included using rubrics to evaluate factors such as salience, logical soundness, and relevance. The team also considered implementing RAG (Retrieval-Augmented Generation) as part of the verification step in the data synthesis pipeline.
Semantic Clustering Challenges: The team debated the effectiveness of using multi-dimensional sentence embeddings for fine-tuning categorization. Concerns were raised about the vulnerability of sentence embeddings to bag-of-words approaches and the difficulty of clustering semantically similar ideas.
Specialized Model Development: There was a proposal to develop specialized models or prompts for different stages of the data synthesis process. This approach would allow for more targeted and efficient data generation and verification.
Compute Resource Exploration: The team planned to investigate potential compute resources, specifically mentioning European supercomputers, for large-scale data generation and self-verification processes.
Beta Was this translation helpful? Give feedback.
All reactions