PulseAugur
LIVE 00:45:45
ENTITY WavLM

WavLM

PulseAugur coverage of WavLM — every cluster mentioning WavLM across labs, papers, and developer communities, ranked by signal.

Total · 30d
8
8 over 90d
Releases · 30d
0
0 over 90d
Papers · 30d
8
8 over 90d
TIER MIX · 90D
SENTIMENT · 30D

1 day(s) with sentiment data

RECENT · PAGE 1/1 · 6 TOTAL
  1. TOOL · CL_29444 ·

    New framework improves speech confidence detection using Whisper

    Researchers have developed a new semi-supervised framework for detecting speaker confidence in speech, addressing the challenge of limited labeled data. This approach combines deep semantic embeddings from OpenAI's Whis…

  2. RESEARCH · CL_22202 ·

    WavCube model unifies speech understanding and generation with compressed representation

    Researchers have developed WavCube, a novel speech representation model designed to unify speech understanding and generation tasks. This model utilizes a compact continuous latent space derived from a self-supervised l…

  3. TOOL · CL_18816 ·

    Phoneme-level analysis improves detection of emotionally manipulated synthetic speech

    Researchers have developed a new method for detecting deepfake audio by analyzing speech at the phoneme level. This approach, which uses self-supervised embeddings, proved more effective than previous methods that treat…

  4. RESEARCH · CL_15484 ·

    Researchers explore quantum and deep learning for audio deepfake detection

    Two research papers submitted to the Environment-Aware Speech and Sound Deepfake Detection Challenge (ESDD2) in 2026 propose novel deep-learning frameworks for detecting manipulated audio. The first paper introduces a d…

  5. RESEARCH · CL_16198 ·

    New GRIDS framework detects anomalies in self-supervised speech models

    Researchers have developed a new framework called GRIDS to analyze how perturbations affect the internal representations of self-supervised speech models. By using Local Intrinsic Dimensionality (LID), the framework can…

  6. RESEARCH · CL_14111 ·

    LASE model improves cross-script voice cloning by making embeddings language-uninformative

    Researchers have developed LASE, a Language-Adversarial Speaker Encoder, to improve multilingual voice cloning. Standard encoders struggle to maintain speaker identity across different scripts, particularly when project…