README / README.md
kroonen's picture
Update README.md
34648b8 verified
metadata
title: README
emoji: πŸ“ˆ
colorFrom: gray
colorTo: red
sdk: static
pinned: false

The Lab

GitHub license MIT licensed Safety Focus

Welcome to The Lab – a vibrant, research-driven hub dedicated to advancing safe and accessible AI. Founded by Robin Kroonen, our mission is to develop AI systems that are not only powerful but also aligned with human values, through rigorous safety evaluations, transparent research, and open-source collaboration.


Table of Contents


About The Lab

The Lab is a research initiative of Kroonen AI, Inc., where we conduct specialized research at the intersection of AI capability and safety. We believe that as AI systems become more powerful, ensuring they remain aligned with human values and operate within appropriate boundaries becomes increasingly important.


What We Do

At The Lab, our work revolves around:

  • Safety Research: Developing comprehensive evaluation methodologies for language models, including ASL-3 style testing frameworks.
  • Fine-Tuning Innovation: Creating fine-tuning approaches that enhance capabilities while maintaining robust safety guardrails.
  • Open Collaboration: Partnering with researchers and organizations committed to responsible AI development.
  • Professional Consulting: Offering expert guidance on model safety, deployment strategies, and ethical AI implementation.

Our goal is to advance AI that remains beneficial, safe, and aligned with human values as it becomes increasingly capable.


Our Specialties

  • Safety Evaluation Frameworks:
    Comprehensive methodologies for testing model responses across potentially problematic domains, with special focus on maintaining safety despite various persuasion techniques.

  • Custom Fine-Tuning with Safety Guardrails:
    Utilizing techniques like Supervised Fine-Tuning (SFT), Direct Preference Optimization (DPO), and Low-Rank Adaptation (LoRA) while ensuring safety boundaries remain intact.

  • Persona & Behavioral Alignment:
    Researching how emotional fine-tuning affects safety boundaries and creating balanced approaches to persona development.

  • Advanced Reasoning with Ethical Constraints:
    Implementing Chain-of-Thought (CoT) methodologies that improve reasoning capabilities while maintaining ethical boundaries.

  • Deployment Safety:
    Guiding safe and secure model deployment in isolated, controlled environments.


Projects from The Lab

Ophelia.chat

An innovative, safety-focused conversational assistant currently in beta on TestFlight.

  • Features:
    • Cloud and local inference support via an Ollama Server.
    • Built-in safety measures and content filtering.
    • Privacy-preserving design.
    • Community-driven development on GitHub.
  • License: MIT

SafetyBench

A comprehensive benchmark for evaluating model safety across various scenarios and persuasion techniques.

ASL-3 Evaluation Framework

A sophisticated testing system for language model safety inspired by industry best practices.

Persona-Safe Models

Fine-tuned models that maintain emotional resonance and distinct personalities while preserving strong safety boundaries.


AI Safety Framework

Our approach to AI safety includes:

  • ASL-3 Style Evaluations: Testing across chemical, biological, radiological, nuclear, and explosive (CBRNE) domains to ensure models resist providing harmful information.
  • Multiple Persuasion Techniques: Evaluating model responses to direct requests, emotional coaxing, fictional scenarios, indirect framing, and thought experiments.
  • Tone & Persona Analysis: Measuring how emotional fine-tuning affects safety boundaries.
  • Risk Vector Detection: Systems trained to identify subtle patterns in model outputs that may indicate safety vulnerabilities.

All evaluations happen in isolated, offline environments with strict controls to prevent unsafe outputs from being deployed.

For more details, visit our Safety & Ethics page.


Open Source Licensing

We believe in open innovation while prioritizing responsibility. Our projects are released under open licenses:

  • The Lab Models: Apache License
  • Ophelia.chat: MIT License
  • Safety Evaluation Tools: Appropriate licensing with usage guidelines

For details, refer to our LICENSE file.


Collaboration & Custom Services

Are you looking for specialized safety evaluation or AI consulting?
We offer:

  • Tailored Safety Solutions: Custom evaluation frameworks and fine-tuning approaches that prioritize safety.
  • Clear & Competitive Pricing: Transparent pricing structures that reflect our commitment to quality.
  • Confidentiality & Security: Rigorous protocols to safeguard your data.

Reach out via kroonen.ai/thelab or email research@kroonen.ai to explore how we can work together.


Stay Connected

Join The Lab and be part of a community committed to developing AI that is not only powerful but also safe, beneficial, and aligned with human values.


Committed to safe and accessible AI research.