cover

HierSpeech++: All the Amazing Things It Could Do

20 Dec 2024

In this work, we propose HierSpeech++, which achieves a human-level high-quality zero-shot speech synthesis performance.

cover

The Limitations of HierSpeech++ and a Quick Fix

20 Dec 2024

Although our model improves the zero-shot speech synthesis performance significantly, our model also synthesizes the noisy environmental information.

cover

HierSpeech++: How Does It Compare to Vall-E, Natural Speech 2, and StyleTTS2?

20 Dec 2024

We compared the zero-shot TTS performance of our model with Vall-E, NaturalSpeech 2, and StyleTTS 2.

cover

Batching Techniques for LLMs

14 Dec 2024

By reducing the queueing delay and the inefficiencies from padding, the fine-grained batching mechanisms significantly increase the throughput of LLM serving.

cover

LLM Service & Autoregressive Generation: What This Means

14 Dec 2024

Once trained, LLMs are often deployed as a conditional generation service (e.g., completion API [34] or chatbot.

cover

The Generation and Serving Procedures of Typical LLMs: A Quick Explanation

14 Dec 2024

In this section, we describe the generation and serving procedures of typical LLMs and the iteration-level scheduling used in LLM serving.

cover

PagedAttention: An Attention Algorithm Inspired By the Classical Virtual Memory in Operating Systems

14 Dec 2024

To address this problem, we propose PagedAttention, an attention algorithm inspired by the classical virtual memory and paging techniques in operating systems.

cover

The Limitations and Failure Cases of DreamLLM: How Far Can it Go?

28 Nov 2024

While DREAMLLM has made significant strides toward the development of versatile, creative, and foundational MLLMs, it still has several limitations.

cover

DreamLLM: Additional Related Works to Look Out For

28 Nov 2024

This breakthrough garnered a lot of attention and paved the way for further research and development in the field.