PulseAugur / Pulse
LIVE 09:09:11

Pulse

last 48h
[12/1912] 89 sources

What AI is actually talking about — clusters surfacing on Bluesky, Reddit, HN, Mastodon and Lobsters, re-ranked to elevate originality and crush noise.

  1. DataScience SG Meetup - How we got top 3% in Kaggle

    Eugene Yan shared insights from his experience placing in the top 3% of a Kaggle competition at a DataScience SG Meetup. The presentation covered various aspects of the competition, including evaluation metrics, feature engineering, machine learning techniques, and ensembling methods. The talk, held at SMU, drew a large audience interested in practical data science applications. AI

  2. Thoughts on Functional Programming in Scala Course (Coursera)

    Eugene Yan shares his experience taking a Coursera course on functional programming in Scala, taught by the language's designer, Martin Odersky. The six-week course covered Scala fundamentals, functional programming concepts, and emphasized software engineering practices like unit testing with ScalaTest. Yan found that while he may not frequently use recursive solutions in his data science work, the course improved his understanding of Scala and problem-solving through tail recursion, ultimately making his code more robust and efficient. AI

  3. SortMySkills is now live!

    Eugene Yan has launched SortMySkills, a web-based card sorting game designed to help individuals identify their core passions and career inclinations. The game prompts users to rank 50 general skills based on their enjoyment, with a constraint of only five skills being designated as AI

  4. Strata x Hadoop 2016 - How Lazada Ranks Products

    Eugene Yan presented at Strata x Hadoop Singapore in 2016 about Lazada's approach to product ranking within their catalog and search results. The goal of this ranking system was to enhance customer experience and boost conversion rates. Yan shared the presentation deck and invited feedback on both the ranking framework and his presentation skills. AI

  5. How Prototyping Can Help You to Get Buy-In

    Eugene Yan details a multi-part process for building a product classification API, emphasizing the importance of prototyping to gain stakeholder buy-in. He explains how to acquire and prepare data, including cleaning titles and handling encoding issues, before training a machine learning model. The series also covers developing the API itself and demonstrates image search capabilities, though the API was later discontinued due to cloud costs. AI

    IMPACT Provides a practical guide to end-to-end data product development, useful for engineers building similar classification systems.

  6. Title: P9: Survey of Open-Source Machine Learning and Data Sciecne in [2024-10-03 Thu] a python library for bandit algorithms and off-policy evaluation 8) AIRI

    OpenAI has released Triton 1.0, an open-source programming language designed to make GPU programming more accessible for researchers. Triton allows users to write efficient GPU code, comparable to expert-level performance, with significantly less code than traditional methods. This release aims to simplify the development of complex neural network operations and improve performance by automating low-level GPU optimizations. AI

  7. Variational lossy autoencoder

    OpenAI has published research on a Variational Autoencoder (VAE) that combines VAEs with autoregressive models like RNNs and PixelCNNs. This new VAE architecture allows for control over what the latent code learns, enabling it to discard irrelevant information such as texture in images. The model achieves state-of-the-art results on density estimation tasks for MNIST, OMNIGLOT, and Caltech-101 Silhouettes. AI

    Variational lossy autoencoder
  8. Transfer of adversarial robustness between perturbation types

    OpenAI researchers are exploring the transferability of adversarial robustness across different types of perturbations in neural networks. Their findings indicate that robustness against one perturbation type does not always guarantee robustness against others and can sometimes be detrimental. They recommend evaluating adversarial defenses using a diverse range of perturbation types and sizes to ensure comprehensive security. Additionally, OpenAI is investigating adversarial examples as a concrete AI safety problem, noting their potential to cause significant issues, such as tricking autonomous vehicles. AI

    Transfer of adversarial robustness between perturbation types

    IMPACT Highlights the ongoing challenges in securing AI systems against sophisticated adversarial attacks, necessitating robust evaluation and defense strategies.

  9. Evolution Strategies

    OpenAI researchers have found that evolution strategies (ES), a decades-old optimization technique, can rival the performance of modern reinforcement learning (RL) methods on benchmarks like Atari and MuJoCo. ES offers advantages such as simpler implementation without backpropagation, easier scalability in distributed settings, and better handling of sparse rewards. This approach trains agents significantly faster than traditional RL, with one experiment reducing training time for a humanoid walker from 10 hours to 10 minutes. AI

    Evolution Strategies
  10. Learning to learn deep learning 📖

    Google AI has introduced Test-Time Diffusion Deep Researcher (TTD-DR), a novel framework that mimics human research processes by iteratively drafting and revising reports using retrieved information. This approach models report writing as a diffusion process, refining initial drafts through a denoising mechanism powered by search. OpenAI has also published several articles detailing techniques for training large neural networks, including data, pipeline, and tensor parallelism, as well as exploring the nonlinear computational properties of deep linear networks due to floating-point arithmetic. Additionally, OpenAI discussed infrastructure considerations for deep learning and a reparameterization technique called weight normalization to accelerate training. AI

    Learning to learn deep learning 📖
  11. RL²: Fast reinforcement learning via slow reinforcement learning

    OpenAI has published a series of research papers detailing advancements in reinforcement learning (RL). These include achieving superhuman performance in Dota 2 with OpenAI Five, developing benchmarks for safe exploration in RL environments, and quantifying generalization capabilities with a new CoinRun environment. The research also explores novel methods for encouraging exploration through curiosity, learning policy representations in multiagent systems, and evolving loss functions for faster training on new tasks. Additionally, OpenAI is working on variance reduction techniques for policy gradients and exploring the equivalence between policy gradients and soft Q-learning. AI

    RL²: Fast reinforcement learning via slow reinforcement learning

    IMPACT These advancements in reinforcement learning, including new benchmarks and methods for generalization and exploration, could accelerate the development of more capable and safer AI systems.

  12. Generative models: exploration to deployment

    Researchers are developing new methods to improve LLM capabilities in various domains. One study introduces MemCoE, a cognition-inspired framework for LLM agents to learn how to organize and update long-term user memory, enhancing personalization. Another paper, ReLay, explores personalized LLM-generated summaries, finding that while personalization improves comprehension, it also introduces risks of bias and hallucinations. Additionally, a new benchmark called ClassEval-Pro has been created to evaluate LLMs on class-level code generation, revealing significant performance gaps among current frontier models. AI

    Generative models: exploration to deployment

    IMPACT Advances in LLM memory, personalization, and code generation benchmarks will drive further research and development in AI agents and software engineering.