A large LLM created by combining two fine-tuned Llama 70B models into one 120B model. Combines Xwin and Euryale.
Credits to
#merge
Prompt tokens measure input size. Reasoning tokens show internal thinking before a response. Completion tokens reflect total output length.