deepsex-34b-GGUF

Maintained By
TheBloke

Deepsex-34B GGUF

PropertyValue
Base ModelYi-34B
FormatGGUF (Various Quantizations)
LicenseMIT
LanguageEnglish

What is deepsex-34b-GGUF?

Deepsex-34B GGUF is a specialized language model derived from Yi-34B, engineered specifically for creative writing and character interactions. The model has been systematically fine-tuned through a multi-stage process involving light novel content and carefully curated roleplay datasets.

Implementation Details

The model was developed through a sophisticated training pipeline that included: initial training on 4GB of deduplicated light novels, continuous pre-training using QLora (r=64, alpha=128) for 3 epochs, and final fine-tuning on high-quality roleplay data filtered through Goliath-120B scoring.

  • Multiple GGUF quantization options (Q2_K to Q8_0)
  • Supports GPU acceleration with layer offloading
  • Compatible with various frameworks including llama.cpp and text-generation-webui
  • Uses Alpaca prompt format for consistency

Core Capabilities

  • Advanced creative writing and storytelling
  • Character-driven dialogue generation
  • Context-aware responses up to 4096 tokens
  • Multiple quantization options for different hardware requirements

Frequently Asked Questions

Q: What makes this model unique?

This model combines the power of Yi-34B with specialized training on creative writing content, making it particularly effective for narrative generation and character interactions. Its unique training pipeline includes quality-filtered roleplay data and extensive light novel content.

Q: What are the recommended use cases?

The model is optimized for creative writing applications, character-based interactions, and narrative generation. It's particularly suitable for applications requiring nuanced character dialogue and storytelling capabilities.

The first platform built for prompt engineering