Mythalion-13B-GPTQ

Maintained By
TheBloke

Mythalion-13B-GPTQ

PropertyValue
Parameter Count2.03B (Quantized)
Model TypeLLaMA2-based
LicenseLLaMA2
QuantizationGPTQ (Multiple Options)

What is Mythalion-13B-GPTQ?

Mythalion-13B-GPTQ is a quantized version of the collaborative effort between PygmalionAI and Gryphe, combining Pygmalion-2 13B and MythoMax L2 13B models. This GPTQ-quantized variant offers various compression options while maintaining performance, making it more accessible for users with different hardware capabilities.

Implementation Details

The model comes in multiple GPTQ configurations, ranging from 4-bit to 8-bit quantization with different group sizes (32g, 64g, 128g) and Act Order options. The implementation supports both Alpaca and Pygmalion formatting styles, making it versatile for different use cases.

  • Multiple quantization options (4-bit and 8-bit variants)
  • Various group sizes for VRAM optimization
  • Compatible with ExLlama (4-bit versions)
  • Supports both Transformers and AutoGPTQ frameworks

Core Capabilities

  • Optimized for roleplay and chat interactions
  • Supports multi-turn conversations
  • Enhanced fictional writing capabilities
  • Flexible prompt formatting support
  • Multiple dataset training (including PIPPA, OpenOrca, Claude multiround chat)

Frequently Asked Questions

Q: What makes this model unique?

This model combines the strengths of Pygmalion-2 and MythoMax, offering superior performance in roleplay and chat scenarios while providing multiple quantization options for different hardware requirements.

Q: What are the recommended use cases?

The model is primarily designed for fictional writing and entertainment purposes, excelling in roleplay and chat applications. It's particularly well-suited for use with platforms like SillyTavern with specific recommended generation settings.

The first platform built for prompt engineering