PulseAugur
LIVE 06:02:45
ENTITY PyTorch

PyTorch

PulseAugur coverage of PyTorch — every cluster mentioning PyTorch across labs, papers, and developer communities, ranked by signal.

Total · 30d
61
61 over 90d
Releases · 30d
0
0 over 90d
Papers · 30d
38
38 over 90d
TIER MIX · 90D
RELATIONSHIPS
SENTIMENT · 30D

3 day(s) with sentiment data

RECENT · PAGE 2/3 · 60 TOTAL
  1. RESEARCH · CL_18344 ·

    LLMs fine-tuned to predict neural network performance from code

    Researchers have developed a method to fine-tune Large Language Models (LLMs) for predicting neural network performance on image classification tasks. By analyzing neural network architecture code, an LLM can determine …

  2. TOOL · CL_15855 ·

    Researchers use BiLSTM with attention to improve game review sentiment analysis

    Researchers have developed an attention-based Bidirectional Long Short-Term Memory (BiLSTM) model to improve sentiment classification of Steam game reviews. This deep learning approach, implemented in PyTorch, was train…

  3. RESEARCH · CL_16106 ·

    Kernel Ridge Regression offers new deep learning architecture, Cubit

    Researchers have introduced Cubit, a novel architecture that replaces the attention mechanism in Transformers with Kernel Ridge Regression (KRR). This approach, detailed in a recent arXiv paper, offers a potentially str…

  4. TOOL · CL_16004 ·

    New CUDA implementation speeds up optimal transport calculations on GPUs

    Researchers have developed FastSinkhorn, a new CUDA implementation for the Sinkhorn algorithm used in optimal transport computations. This method operates entirely in the log-domain, ensuring numerical stability even wi…

  5. RESEARCH · CL_14450 ·

    Researchers explore novel attention mechanisms and optimization techniques for LLMs

    Researchers are exploring novel attention mechanisms to overcome the quadratic complexity of standard self-attention in transformers, particularly for long-context processing. Several papers introduce methods like Light…

  6. RESEARCH · CL_14340 ·

    AI model uses copula-enhanced Vision Transformer for myopia diagnosis

    Researchers have developed a novel approach using a copula-enhanced Vision Transformer to improve the diagnosis of high myopia from ultra-widefield fundus images. This method addresses the challenges of capturing inter-…

  7. TOOL · CL_14019 ·

    AI assists programmer in creating Pascal Numeric Library, rivaling NumPy

    A programmer, assisted by GitHub Copilot, has developed a comprehensive implementation of BLAS levels 1-3 in Pascal. This project aims to create a Pascal Numeric Library (PNL) that rivals the functionality of Python lib…

  8. RESEARCH · CL_13675 ·

    AI model recovers keystrokes with 85% accuracy using laptop microphone audio

    Researchers have developed a method to recover typed text by analyzing laptop microphone audio. A convolutional neural network (CNN) was trained on log-mel spectrograms of individual keystrokes, achieving approximately …

  9. RESEARCH · CL_13517 ·

    CuTeDSL emerges as new GPU kernel path for LLM inference, challenging CUTLASS

    The landscape of GPU kernel engineering for LLM inference is shifting, with CuTeDSL emerging as a potential successor to C++ CuTe/CUTLASS. This evolution is highlighted by industry trends in technologies like FlashAtten…

  10. COMMENTARY · CL_13081 ·

    Free Pascal and BLAS offer faster matrix multiplication for AI development

    A user explored the performance of Python for AI tasks, noting its slowness but acknowledging the extensive AI ecosystem as its primary advantage. They conducted a test comparing Free Pascal and BLAS for matrix multipli…

  11. RESEARCH · CL_12339 ·

    AI agents automate data prep, while new Python ML compiler speeds LLM compression

    Researchers have developed a new open-source machine learning compiler stack written in just 5,000 lines of Python. This stack offers unprecedented transparency by lowering large language models to CUDA with six interme…

  12. RESEARCH · CL_11895 ·

    New algorithm speeds up EigenDecomposition for large matrices in deep learning

    Researchers have developed a new batch-efficient algorithm for EigenDecomposition (ED), a critical computation in computer vision and deep learning. This divide-and-conquer approach aims to overcome the computational bo…

  13. RESEARCH · CL_11790 ·

    Neural ODEs advance with mixed precision training and causal forecasting methods

    Researchers have developed a new mixed-precision training framework for Neural Ordinary Differential Equations (Neural ODEs) to reduce computational costs. This framework uses low-precision computations for evaluating n…

  14. RESEARCH · CL_11904 ·

    New C++ engine HASE achieves 33M steps/sec for multi-agent RL training

    Researchers have developed a new C++ engine called Hide-And-Seek-Engine (HASE) designed to significantly improve the efficiency of training reinforcement learning agents in decentralized, partially observable environmen…

  15. RESEARCH · CL_14104 ·

    VkSplat pipeline boosts 3D Gaussian Splatting training with Vulkan compute

    Researchers have developed VkSplat, a novel training pipeline for 3D Gaussian Splatting (3DGS) that utilizes Vulkan compute for enhanced performance and broader compatibility. This new approach offers a significant spee…

  16. SIGNIFICANT · CL_07248 ·

    DeepSeek V4 First Release Adaptation Behind: Why does Ascend insist on not doing a CUDA compatibility layer?

    Huawei's Ascend AI accelerators are forging a unique path by eschewing CUDA compatibility to build an independent ecosystem. This strategy focuses on deep architectural changes in their latest Ascend 950 chips to addres…

  17. RESEARCH · CL_06254 ·

    Studies benchmark AutoML and BiLSTM for NLP tasks, showing mixed results

    Researchers have compared traditional machine learning methods with deep learning models for various natural language processing tasks, including fine-grained emotion classification and sentiment analysis. Studies utili…

  18. RESEARCH · CL_06307 ·

    New HDET method explores hyperparameters for large model training

    Researchers have introduced Hyperparameter-Divergent Ensemble Training (HDET), a novel method designed to optimize the training of large neural networks. HDET repurposes data-parallel replicas to simultaneously explore …

  19. TOOL · CL_05620 ·

    IBM Research integrates vLLM into its RITS Platform for AI development

    IBM Research has integrated vLLM, an open-source library for fast LLM inference, into its RITS Platform. This integration aims to enhance the platform's capabilities by leveraging vLLM's efficient processing for large l…

  20. RESEARCH · CL_06196 ·

    PointTransformerX offers portable, efficient 3D point cloud processing without sparse algorithms

    Researchers have developed PointTransformerX (PTX), a new vision transformer backbone for processing 3D point clouds that eliminates the need for custom CUDA operators. This PyTorch-native model achieves competitive acc…