google/gemma-2-9b - 4b_2n4m_128bs Compression

This is a compressed model using deltazip.

Paper, Compression Tool, Inference Engine (Soon).

Compression Configuration

  • Base Model: google/gemma-2-9b
  • Compression Scheme: 4b_2n4m_128bs
  • Dataset: HuggingFaceH4/ultrachat_200k
  • Dataset Split: train_sft
  • Max Sequence Length: 2048
  • Number of Samples: 256

Sample Output

Prompt:

[{'role': 'user', 'content': 'Who is Alan Turing?'}]

Output:

<bos><start_of_turn>user
Who is Alan Turing?<end_of_turn>
*Alan Turing* (1912-1954) was a British mathematician and computer scientist who is considered the father of theoretical computer science and artificial intelligence.

**Key Contributions:**

* **Turing Machine:** In 1936, Turing proposed the concept of a "Turing machine," a theoretical model of computation that can simulate any algorithm. This laid the foundation for modern computers.
* **Breaking the Enigma Code:** During World War II, Turing played a crucial role in breaking the German Enigma code, which helped the Allied forces win the war

Evaluation

Downloads last month
4
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for deltazip/google.gemma-2-9b-it.4b_2n4m_128bs

Base model

google/gemma-2-9b
Finetuned
(225)
this model

Dataset used to train deltazip/google.gemma-2-9b-it.4b_2n4m_128bs

Collection including deltazip/google.gemma-2-9b-it.4b_2n4m_128bs